DocumentCode
797664
Title
Rates of convergence of nearest neighbor estimation under arbitrary sampling
Author
Kulkarni, Sanjeev R. ; Posner, Steven E.
Author_Institution
Dept. of Electr. Eng., Princeton Univ., NJ, USA
Volume
41
Issue
4
fYear
1995
fDate
7/1/1995 12:00:00 AM
Firstpage
1028
Lastpage
1039
Abstract
Rates of convergence for nearest neighbor estimation are established in a general framework in terms of metric covering numbers of the underlying space. The first result is to find explicit finite sample upper bounds for the classical independent and identically distributed (i.i.d.) random sampling problem in a separable metric space setting. The convergence rate is a function of the covering numbers of the support of the distribution. For example, for bounded subsets of R r, the convergence rate is O(1/n2r/). The main result is to extend the problem to allow samples drawn from a completely arbitrary random process in a separable metric space and to examine the performance in terms of the individual sample sequences. The authors show that for every sequence of samples the asymptotic time-average of nearest neighbor risks equals twice the time-average of the conditional Bayes risks of the sequence. Finite sample upper bounds under arbitrary sampling are again obtained in terms of the covering numbers of the underlying space. In particular, for bounded subsets of Rr the convergence rate of the time-averaged risk is O(1/n2r/). The authors then establish a consistency result for kn-nearest neighbor estimation under arbitrary sampling and prove a convergence rate matching established rates for i.i.d. sampling. Finally, they show how their arbitrary sampling results lead to some classical i.i.d. sampling results and in fact extend them to stationary sampling. The framework and results are quite general while the proof techniques are surprisingly elementary
Keywords
convergence; entropy; estimation theory; random processes; signal sampling; arbitrary sampling; asymptotic time-average; bounded subsets; classical independent identically distributed random sampling problem; conditional Bayes risks; convergence rate; explicit finite sample upper bounds; metric covering numbers; nearest neighbor estimation; nearest neighbor risks; sample sequences; separable metric space; separable metric space setting; sequence; stationary sampling; time-averaged risk; Convergence; Entropy; Extraterrestrial measurements; Nearest neighbor searches; Neural networks; Random processes; Random variables; Sampling methods; Student members; Upper bound;
fLanguage
English
Journal_Title
Information Theory, IEEE Transactions on
Publisher
ieee
ISSN
0018-9448
Type
jour
DOI
10.1109/18.391248
Filename
391248
Link To Document