Title :
Sparse feature selection by information theory
Author :
Guangyao Zhou ; Geman, Stuart ; Buhmann, J.M.
Author_Institution :
Div. of Appl. Math., Brown Univ., Providence, RI, USA
fDate :
June 29 2014-July 4 2014
Abstract :
Learning sparse structures in high dimensions defines a combinatorial selection problem of e.g. informative feature dimensions and a subsequent estimation task to determine adequate model parameters. The arguably simplest sparse inference problem requires estimating a sparse mean in high dimensions when most dimensions should be discarded [6]. We reduce sparse mean estimation to a pure selection problem by restricting the source to binary values that are contaminated with various noise models. The model selection principle of Approximation Set Coding [2], [3] is rigorously applied, and generalization capacity is used to evaluate different algorithms. Simulation results demonstrate the effectiveness of generalization capacity compared to traditional model selection approaches. Sampling-based approximation yields insights into the behavior of algorithms in high dimensions at different noise levels.
Keywords :
Gaussian noise; compressed sensing; estimation theory; feature selection; inference mechanisms; information theory; applied capacity; approximation set coding; binary value; combinatorial selection problem; generalization capacity; information theory; informative feature dimensions; learning sparse structures; model selection principle; noise models; sampling-based approximation; sparse feature selection; sparse inference problem; sparse mean estimation reduction; subsequent estimation task; Approximation algorithms; Approximation methods; Estimation; Information theory; Mathematical model; Noise; Noise level;
Conference_Titel :
Information Theory (ISIT), 2014 IEEE International Symposium on
Conference_Location :
Honolulu, HI
DOI :
10.1109/ISIT.2014.6874968