Author :
Zhao, Lei ; Mammadov, Musa ; Yearwood, John
Author_Institution :
GSITMS, Univ. of Ballarat, Ballarat, VIC, Australia
Abstract :
Problems of data classification can be studied in the framework of regularization theory as ill-posed problems. In this framework, loss functions play an important role in the application of regularization theory to classification. In this paper, we review some important convex loss functions, including hinge loss, square loss, modified square loss, exponential loss, logistic regression loss, as well as some non-convex loss functions, such as sigmoid loss, φ-loss, ramp loss, normalized sigmoid loss, and the loss function of 2 layer neural network. Based on the analysis of these loss functions, we propose a new differentiable nonconvex loss function, called smoothed 0-1 loss function, which is a natural approximation of the 0-1 loss function. To compare the performance of different loss functions, we propose two binary classification algorithms for binary classification, one for convex loss functions, the other for non-convex loss functions. A set of experiments are launched on several binary data sets from the UCI repository. The results show that the proposed smoothed 0-1 loss function is robust, especially for those noisy data sets with many outliers.
Keywords :
concave programming; convex programming; data analysis; neural nets; pattern classification; regression analysis; φ-loss; 2 layer neural network; UCI repository; binary classification algorithms; binary data sets; data classification; differentiable nonconvex loss function; exponential loss; hinge loss; ill-posed problems; logistic regression loss; loss function analysis; modified square loss; noisy data sets; non-convex loss functions; normalized sigmoid loss; ramp loss; regularization theory; smoothed 0-1 loss function; classification; loss function; non-convex; optimization; regularization;