Title :
Parameter convergence and learning curves for neural networks
Author :
Fine, Terrence L. ; Mukherjee, Sayandev
Author_Institution :
Sch. of Electr. Eng., Cornell Univ., Ithaca, NY, USA
Abstract :
We revisit the oft-studied asymptotic (in sample size) behavior of the parameter or weight estimate returned by any member of a large family of neural network training algorithms. By properly accounting for the characteristic property of neural networks that their empirical and generalization errors possess multiple minima, we establish conditions under which the parameter estimate converges strongly into the set of minima of the generalization error. These results are then used to derive learning curves for generalization and empirical errors that leads to bounds on rates of convergence
Keywords :
convergence of numerical methods; error analysis; learning (artificial intelligence); neural nets; parameter estimation; asymptotic behavior; convergence rate bounds; empirical errors; generalization errors; learning curve; neural networks; parameter convergence; parameter estimate; sample size; training algorithms; weight estimate; Convergence; Nearest neighbor searches; Neural networks; Parameter estimation; Random sequences; Random variables; Space technology;
Conference_Titel :
Information Theory, 1998. Proceedings. 1998 IEEE International Symposium on
Conference_Location :
Cambridge, MA
Print_ISBN :
0-7803-5000-6
DOI :
10.1109/ISIT.1998.708991