DocumentCode
3212432
Title
A new super-memory gradient method for unconstrained optimization
Author
Tang, Jingyong ; Dong, Li
Author_Institution
Coll. of Math. & Inf. Sci., Xinyang Normal Univ., Xinyang, China
Volume
1
fYear
2010
fDate
13-14 Sept. 2010
Firstpage
93
Lastpage
96
Abstract
In this paper, we propose a new super-memory gradient method for unconstrained optimization problems. The global convergence and linear convergence rate are proved under some mild conditions. The method uses the current and previous iterative information to generate a new search direction and uses Wolfe line search to define the step-size at each iteration. It has a possibly simple structure and avoids the computation and storage of some matrices, which is suitable to solve large scale optimization problems. Numerical experiments show that the new algorithm is effective in practical computation in many situations.
Keywords
convergence; gradient methods; matrix algebra; optimisation; Wolfe line search; global convergence; iterative information; linear convergence; search direction; super memory gradient method; unconstrained optimization; Convergence; Gradient methods; Iterative methods; Minimization; Search methods; Wolfe search rule; onvergence; super-memory gradient method; unconstrained optimization;
fLanguage
English
Publisher
ieee
Conference_Titel
Computational Intelligence and Natural Computing Proceedings (CINC), 2010 Second International Conference on
Conference_Location
Wuhan
Print_ISBN
978-1-4244-7705-0
Type
conf
DOI
10.1109/CINC.2010.5643886
Filename
5643886
Link To Document