Title of article :
Searching Most Efficient Neural Network Architecture Using Akaike’s Information Criterion (AIC)
Author/Authors :
Gaurang Panchal، نويسنده , , Amit Ganatra، نويسنده , , Kosta B. Gligorevis، نويسنده , , Devyani Panchal، نويسنده ,
Issue Information :
روزنامه با شماره پیاپی سال 2010
Pages :
4
From page :
41
To page :
44
Abstract :
The problem of model selection is considerably important for acquiring higher levels of generalization capability in supervised learning. Neural networks are commonly used networks in many engineering applications due to its better generalization property. An ensemble neural network algorithm is proposed based on the Akaike information criterion (AIC). Ecologists have long relied on hypothesis testing to include or exclude variables in models, although the conclusions often depend on the approach used. The advent of methods based on information theory, also known as information-theoretic approaches, has changed the way we look at model selection The Akaike information criterion (AIC) has been successfully used in model selection. It is not easy to decide the optimal size of the neural network because of its strong nonlinearity. We discuss problems with well used information and propose a model selection method.
Keywords :
Hidden neurons , Akaike’s information criterion (AIC) , neural network , Correct Classification Rate (CRR)
Journal title :
International Journal of Computer Applications
Serial Year :
2010
Journal title :
International Journal of Computer Applications
Record number :
659353
Link To Document :
بازگشت