DocumentCode :
106442
Title :
A Dirty Model for Multiple Sparse Regression
Author :
Jalali, A. ; Ravikumar, Penugonda ; Sanghavi, Sujay
Author_Institution :
Dept. of Electr. & Comput. Eng., Univ. of Texas at Austin, Austin, TX, USA
Volume :
59
Issue :
12
fYear :
2013
fDate :
Dec. 2013
Firstpage :
7947
Lastpage :
7968
Abstract :
The task of sparse linear regression consists of finding an unknown sparse vector from linear measurements. Solving this task even under “high-dimensional” settings, where the number of samples is fewer than the number of variables, is now known to be possible via methods such as the LASSO. We consider the multiple sparse linear regression problem, where the task consists of recovering several related sparse vectors at once. A simple approach to this task would involve solving independent sparse linear regression problems, but a natural question is whether one can reduce the overall number of samples required by leveraging partial sharing of the support sets, or nonzero patterns, of the signal vectors. A line of recent research has studied the use of ℓ1/ℓq norm block-regularizations with q > 1 for such problems. However, depending on the level of sharing, these could actually perform worse in sample complexity when compared to solving each problem independently. We present a new “adaptive” method for multiple sparse linear regression that can leverage support and parameter overlap when it exists, but not pay a penalty when it does not. We show how to achieve this using a very simple idea: decompose the parameters into two components and regularize these differently. We show, theoretically and empirically, that our method strictly and noticeably outperforms both ℓ1 or ℓ1/ℓq methods, over the entire range of possible overlaps (except at boundary cases, where we match the best method), even under high-dimensional scaling.
Keywords :
regression analysis; LASSO; linear measurements; multiple sparse linear regression problem; signal vectors; sparse vector; sparse vectors; Covariance matrices; Estimation; Graphical models; Linear regression; Sparse matrices; Standards; Vectors; High-dimensional statistics; multi-task learning; multiple regression;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/TIT.2013.2280272
Filename :
6588338
Link To Document :
بازگشت