Title :
Scaling up M-estimation via sampling designs: The Horvitz-Thompson stochastic gradient descent
Author :
Clemencon, Stephan ; Bertail, Patrice ; Chautru, Emilie
Author_Institution :
LTCI, Telecom ParisTech, Paris, France
Abstract :
In certain situations that shall be undoubtedly more and more common in the Big Data era, the datasets available are so massive that computing statistics over the full sample is hardly feasible, if not unfeasible. A natural approach in this context consists in using survey schemes and substituting the “full data” statistics with their counterparts based on the resulting random samples, of manageable size. It is the purpose of this paper to investigate the impact of survey sampling with unequal inclusion probabilities on (stochastic) gradient descent-based M-estimation methods in large-scale statistical-learning problems. We prove that, in presence of some a priori information, one may significantly reduce the number of terms that must be averaged to estimate the gradient at each step with overwhelming probability, while preserving the asymptotic accuracy. These striking results are described here by limit theorems.
Keywords :
Big Data; data analysis; estimation theory; gradient methods; learning (artificial intelligence); probability; sampling methods; Big Data; Horvitz-Thompson stochastic gradient descent; M-estimation methods; a priori information; asymptotic accuracy preservation; large-scale statistical-learning problems; sampling designs; survey sampling; unequal inclusion probabilities; Accuracy; Big data; Estimation; Probability; Sociology; Statistics; Zinc; Horvitz-Thompson estimation; sampling design; statistical learning; stochastic gradient descent; survey;
Conference_Titel :
Big Data (Big Data), 2014 IEEE International Conference on
Conference_Location :
Washington, DC
DOI :
10.1109/BigData.2014.7004208