Title :
Why natural gradient?
Author :
Amari, S. ; Douglas, S.C.
Author_Institution :
RIKEN, Inst. of Phys. & Chem. Res., Saitama, Japan
Abstract :
Gradient adaptation is a useful technique for adjusting a set of parameters to minimize a cost function. While often easy to implement, the convergence speed of gradient adaptation can be slow when the slope of the cost function varies widely for small changes in the parameters. In this paper, we outline an alternative technique, termed natural gradient adaptation, that overcomes the poor convergence properties of gradient adaptation in many cases. The natural gradient is based on differential geometry and employs knowledge of the Riemannian structure of the parameter space to adjust the gradient search direction. Unlike Newton´s method, natural gradient adaptation does not assume a locally-quadratic cost function. Moreover, for maximum likelihood estimation tasks, natural gradient adaptation is asymptotically Fisher-efficient. A simple example illustrates the desirable properties of natural gradient adaptation
Keywords :
adaptive estimation; convergence of numerical methods; differential geometry; iterative methods; maximum likelihood estimation; optimisation; Riemannian structure; convergence speed; cost function; differential geometry; gradient search direction; maximum likelihood estimation tasks; natural gradient adaptation; parameter space; Adaptive filters; Cities and towns; Convergence; Cost function; Geometry; Information systems; Iterative methods; Newton method; Optimization methods; Parameter estimation;
Conference_Titel :
Acoustics, Speech and Signal Processing, 1998. Proceedings of the 1998 IEEE International Conference on
Conference_Location :
Seattle, WA
Print_ISBN :
0-7803-4428-6
DOI :
10.1109/ICASSP.1998.675489