• DocumentCode
    2821559
  • Title

    Homotopy continuation methods for neural networks

  • Author

    Chow, J. ; Udpa, L. ; Udpa, S.S.

  • Author_Institution
    Dept. of Electr. Eng. & Comput. Eng., Iowa State Univ., Ames, IA, USA
  • fYear
    1991
  • fDate
    11-14 Jun 1991
  • Firstpage
    2483
  • Abstract
    The application of the homotopy continuation method for finding the global minimum during the training phase of a multilayer neural network is presented. A brief description of the theory of the homotopy continuation methods is given. The backward error propagation algorithm used for training neural networks is summarized. The reformulation of the error minimization problem in the learning algorithm in a framework suitable for the continuation method and the procedure for training neural networks using the homotopy continuation method are described. Results of comparing the performance of the proposed method with traditional training methods are given
  • Keywords
    errors; learning systems; minimisation; neural nets; backward error propagation algorithm; error minimization problem; global minimum; homotopy continuation method; learning algorithm; multilayer neural network; training phase; Differential equations; Erbium; Jacobian matrices; Neural networks; Polynomials;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Circuits and Systems, 1991., IEEE International Sympoisum on
  • Print_ISBN
    0-7803-0050-5
  • Type

    conf

  • DOI
    10.1109/ISCAS.1991.176030
  • Filename
    176030