• DocumentCode
    72292
  • Title

    RES: Regularized Stochastic BFGS Algorithm

  • Author

    Mokhtari, Aryan ; Ribeiro, Alejandro

  • Author_Institution
    Dept. of Electr. & Syst. Eng., Univ. of Pennsylvania, Philadelphia, PA, USA
  • Volume
    62
  • Issue
    23
  • fYear
    2014
  • fDate
    Dec.1, 2014
  • Firstpage
    6089
  • Lastpage
    6104
  • Abstract
    RES, a regularized stochastic version of the Broyden-Fletcher-Goldfarb-Shanno (BFGS) quasi-Newton method, is proposed to solve strongly convex optimization problems with stochastic objectives. The use of stochastic gradient descent algorithms is widespread, but the number of iterations required to approximate optimal arguments can be prohibitive in high dimensional problems. Application of second-order methods, on the other hand, is impracticable because the computation of objective function Hessian inverses incurs excessive computational cost. BFGS modifies gradient descent by introducing a Hessian approximation matrix computed from finite gradient differences. RES utilizes stochastic gradients in lieu of deterministic gradients for both the determination of descent directions and the approximation of the objective function´s curvature. Since stochastic gradients can be computed at manageable computational cost, RES is realizable and retains the convergence rate advantages of its deterministic counterparts. Convergence results show that lower and upper bounds on the Hessian eigenvalues of the sample functions are sufficient to guarantee almost sure convergence of a subsequence generated by RES and convergence of the sequence in expectation to optimal arguments. Numerical experiments showcase reductions in convergence time relative to stochastic gradient descent algorithms and non-regularized stochastic versions of BFGS. An application of RES to the implementation of support vector machines is developed.
  • Keywords
    Hessian matrices; convex programming; eigenvalues and eigenfunctions; gradient methods; radio networks; stochastic processes; support vector machines; Broyden-Fletcher-Goldfarb-Shanno quasi-Newton method; Hessian approximation matrix; Hessian eigenvalues; RES; convex optimization; finite gradient differences; objective function curvature; regularized stochastic BFGS algorithm; second-order methods; stochastic gradient descent algorithm; support vector machines; Approximation algorithms; Approximation methods; Convergence; Eigenvalues and eigenfunctions; Linear programming; Optimization; Signal processing algorithms; Quasi-Newton methods; large-scale optimization; stochastic optimization; support vector machines;
  • fLanguage
    English
  • Journal_Title
    Signal Processing, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1053-587X
  • Type

    jour

  • DOI
    10.1109/TSP.2014.2357775
  • Filename
    6899692