Title :
Second-order methods for L1 regularized problems in machine learning
Author :
Hansen, Samantha ; Nocedal, Jorge
Author_Institution :
Northwestern Univ., Evanston, IL, USA
Abstract :
This paper proposes a Hessian-free Newton method for solving large-scale convex functions with an L1 regularization term. These problems arise in supervised machine learning models in which it is important to seek a sparse parameter vector. The proposed method operates in a batch setting, which is well suited for parallel computing environments, and employs sub-sampled Hessian information to accelerate progress of the iteration. The method consists of two phases, an active-set prediction phase that employs first-order and second-order information, and subspace phase that performs a Newton-like step. Numerical results on a speech recognition problem illustrate the practical behavior of the method.
Keywords :
Newton method; learning (artificial intelligence); speech recognition; Hessian-free Newton method; L1 regularized term problems; active-set prediction phase; first-order information; large-scale convex functions; machine learning; parallel computing environments; second-order information method; sparse parameter vector; speech recognition problem; subsampled Hessian information; supervised machine learning models; Machine learning; Minimization; Newton method; OWL; Optimization; Training; Vectors; Hessian-Free Newton; Iterative Shrinkage; L1 Regularization; Logistic Regression; Newton Method;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on
Conference_Location :
Kyoto
Print_ISBN :
978-1-4673-0045-2
Electronic_ISBN :
1520-6149
DOI :
10.1109/ICASSP.2012.6289101