Title of article
A Proximal Method of Stochastic Gradient for Convex Optimization
Author/Authors
Saeidian ، Zeinab Department of Mathematics - University of Kashan , Mahmoudoghli ، Maryam K.N. Toosi University of Technology
From page
19
To page
32
Abstract
The Proximal Stochastic Average Gradient (Prox-SAG+) is a primary method used for solving optimization problems that contain the sum of two convex functions. This kind of problem usually arises in machine learning, which utilizes a large amount of data to create component functions from a dataset. A proximal operation is applied to obtain the optimal value due to its appropriate properties. The Prox-SAG+ algorithm is faster than some other methods and has a simpler algorithm than previous ones. Moreover, using this specific operator can help to reassure that the achieved result is optimal. Additionally, it has been proven that the proposed method has an approximately geometric rate of convergence. Implementing the proposed operator makes the method more practical than other algorithms found in the literature. Numerical analysis also confirms the efficiency of the proposed scheme.
Keywords
Proximal stochastic average gradient , Convergence property , Training examples , Machine learning
Journal title
Control and Optimization in Applied Mathematics
Journal title
Control and Optimization in Applied Mathematics
Record number
2769782
Link To Document