Title :
Online multi-label learning with accelerated nonsmooth stochastic gradient descent
Author :
Sunho Park ; Seungjin Choi
Author_Institution :
Dept. of Comput. Sci. & Eng., POSTECH, Pohang, South Korea
Abstract :
Multi-label learning refers to methods for learning a set of functions that assigns a set of relevant labels to each instance. One of popular approaches to multi-label learning is label ranking, where a set of ranking functions are learned to order all the labels such that relevant labels are ranked higher than irrelevant ones. Rank-SVM is a representative method for label ranking where ranking loss is minimized in the framework of max margin. However, the dual form in Rank-SVM involves a quadratic programming which is generally solved in cubic time in the size of training data. The primal form is appealing for the development of online learning but involves a nonsmooth convex loss function. In this paper we present a method for online multi-label learning where we minimize the primal form using the accelerated nonsmooth stochastic gradient descent which has been recently developed to extend Nesterov´s smoothing method to the stochastic setting. Numerical experiments on several large-scale datasets demonstrate the computational efficiency and fast convergence of our proposed method, compared to existing methods including subgradient-based algorithms.
Keywords :
convex programming; gradient methods; learning (artificial intelligence); quadratic programming; smoothing methods; stochastic processes; support vector machines; Nesterov smoothing method; accelerated nonsmooth stochastic gradient descent method; cubic time; label ranking; max margin framework; nonsmooth convex loss function; online multilabel learning; primal form minimization; quadratic programming; rank-SVM representative method; ranking functions; ranking loss; subgradient-based algorithms; training data size; Acceleration; Approximation methods; Convergence; Minimization; Optimization; Stochastic processes; Training data; Label ranking; Nesterov´s method; multi-label learning; nonsmooth minimization; stochastic gradient descent;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on
Conference_Location :
Vancouver, BC
DOI :
10.1109/ICASSP.2013.6638273