DocumentCode :
39407
Title :
A Convex Formulation for Learning a Shared Predictive Structure from Multiple Tasks
Author :
Jianhui Chen ; Lei Tang ; Jun Liu ; Jieping Ye
Author_Institution :
GE Global Res., San Ramon, CA, USA
Volume :
35
Issue :
5
fYear :
2013
fDate :
May-13
Firstpage :
1025
Lastpage :
1038
Abstract :
In this paper, we consider the problem of learning from multiple related tasks for improved generalization performance by extracting their shared structures. The alternating structure optimization (ASO) algorithm, which couples all tasks using a shared feature representation, has been successfully applied in various multitask learning problems. However, ASO is nonconvex and the alternating algorithm only finds a local solution. We first present an improved ASO formulation (iASO) for multitask learning based on a new regularizer. We then convert iASO, a nonconvex formulation, into a relaxed convex one (rASO). Interestingly, our theoretical analysis reveals that rASO finds a globally optimal solution to its nonconvex counterpart iASO under certain conditions. rASO can be equivalently reformulated as a semidefinite program (SDP), which is, however, not scalable to large datasets. We propose to employ the block coordinate descent (BCD) method and the accelerated projected gradient (APG) algorithm separately to find the globally optimal solution to rASO; we also develop efficient algorithms for solving the key subproblems involved in BCD and APG. The experiments on the Yahoo webpages datasets and the Drosophila gene expression pattern images datasets demonstrate the effectiveness and efficiency of the proposed algorithms and confirm our theoretical analysis.
Keywords :
generalisation (artificial intelligence); learning (artificial intelligence); mathematical programming; APG algorithm; ASO algorithm; BCD method; Drosophila gene expression pattern images datasets; Yahoo webpages datasets; accelerated projected gradient algorithm; alternating structure optimization algorithm; block coordinate descent; improved ASO formulation; improved generalization performance; multiple related tasks; multitask learning problems; nonconvex counterpart iASO; nonconvex formulation; rASO; relaxed ASO; semidefinite program; shared feature representation; shared predictive structure; Acceleration; Algorithm design and analysis; Complexity theory; Fasteners; Optimization; Prediction algorithms; Vectors; Multitask learning; accelerated projected gradient; alternating structure optimization; shared predictive structure;
fLanguage :
English
Journal_Title :
Pattern Analysis and Machine Intelligence, IEEE Transactions on
Publisher :
ieee
ISSN :
0162-8828
Type :
jour
DOI :
10.1109/TPAMI.2012.189
Filename :
6296661
Link To Document :
بازگشت