DocumentCode
315237
Title
New bounds for correct generalization
Author
Mattera, Davide ; Palmieri, Francesco
Author_Institution
Dipartimento di Ingegneria Elettronica, Naples Univ., Italy
Volume
2
fYear
1997
fDate
9-12 Jun 1997
Firstpage
1051
Abstract
A theoretical approach to the determination of the number of training examples for a neural network architecture is provided by the theory of Vapnik and Chervonenkis on the minimization of the empirical risk. We report here a new bound on the joint probability that both the approximation error between the binary function learned by the input/output examples and the target binary function is larger than ε and the empirical error on the examples is smaller than a fixed non-null fraction of ε. The given bounds are independent of the probability distribution on the input space and improve some existing results on the generalization abilities of an adaptive binary function
Keywords
error statistics; generalisation (artificial intelligence); learning (artificial intelligence); minimisation; neural net architecture; probability; Vapnik-Chervonenkis theory; approximation error; binary function; bounds; empirical risk; error probability; generalization; minimization; neural network architecture; probability; Error correction; Error probability; Frequency measurement; Neural networks; Probability distribution; Risk analysis;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks,1997., International Conference on
Conference_Location
Houston, TX
Print_ISBN
0-7803-4122-8
Type
conf
DOI
10.1109/ICNN.1997.616173
Filename
616173
Link To Document