DocumentCode :
114540
Title :
On the convergence to saddle points of concave-convex functions, the gradient method and emergence of oscillations
Author :
Holding, Thomas ; Lestas, Ioannis
Author_Institution :
Cambridge Centre for Anal., Univ. of Cambridge, Cambridge, UK
fYear :
2014
fDate :
15-17 Dec. 2014
Firstpage :
1143
Lastpage :
1148
Abstract :
It is known that for a strictly concave-convex function, the gradient method introduced by Arrow and Hurwicz [1], has guaranteed global convergence to its saddle point. Nevertheless, there are classes of problems where the function considered is not strictly concave-convex, in which case convergence to a saddle point is not guaranteed. In the paper we provide a characterization of the asymptotic behaviour of the gradient method, in the general case where this is applied to a general concave-convex function. We prove that for any initial conditions the gradient method is guaranteed to converge to a trajectory described by an explicit linear ODE. We further show that this result has a natural extension to subgradient methods, where the dynamics are constrained in a prescribed convex set. The results are used to provide simple characterizations of the limiting solutions for special classes of optimization problems, and modifications of the problem so as to avoid oscillations are also discussed.
Keywords :
concave programming; convergence; convex programming; gradient methods; set theory; asymptotic behaviour; concave-convex functions; explicit linear ODE; guaranteed global convergence; optimization problems; prescribed convex set; saddle points; subgradient methods; Convergence; Gradient methods; Manifolds; Oscillators; Stability analysis; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control (CDC), 2014 IEEE 53rd Annual Conference on
Conference_Location :
Los Angeles, CA
Print_ISBN :
978-1-4799-7746-8
Type :
conf
DOI :
10.1109/CDC.2014.7039535
Filename :
7039535
Link To Document :
بازگشت