DocumentCode :
349205
Title :
Undersampling for the training of feedback neural networks on large sequences; application to the modeling of an induction machine
Author :
Constant, L. ; Dagues, B. ; Rivals, I. ; Personnaz, L.
Author_Institution :
Lab. d´´Electrotech. et d´´Electron. Ind., CNRS, Toulouse, France
Volume :
2
fYear :
1999
fDate :
5-8 Sep 1999
Firstpage :
1025
Abstract :
This paper proposes an economic method for the nonlinear modeling of dynamic processes using feedback neural networks, by undersampling the training sequences. The undersampling (i) allows a better exploration of the operating range of the process for a given size of the training sequences, and (ii) it speeds up the training of the feedback networks. This method is successfully applied to the training of a neural model of the electromagnetic part of an induction machine, whose sampling period must be small enough to take fast variations of the input voltage into account, i.e, smaller than 1 μs
Keywords :
asynchronous machines; electric machine analysis computing; learning (artificial intelligence); machine theory; recurrent neural nets; dynamic processes; feedback neural networks; induction machine; input voltage variations; neural model; nonlinear modeling; operating range; sampling period; training; undersampling; Electromagnetic modeling; Electronic mail; Induction machines; Industrial economics; Industrial training; Neural networks; Neurofeedback; Sampling methods; Stators; Voltage;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Electronics, Circuits and Systems, 1999. Proceedings of ICECS '99. The 6th IEEE International Conference on
Conference_Location :
Pafos
Print_ISBN :
0-7803-5682-9
Type :
conf
DOI :
10.1109/ICECS.1999.813408
Filename :
813408
Link To Document :
بازگشت