Title :
Capacity control in classifiers for pattern recognition
Author_Institution :
AT&T Bell Lab., Holmdel, NJ, USA
fDate :
31 Aug-2 Sep 1992
Abstract :
Achieving good performance in statistical pattern recognition requires matching the capacity of the classifier to the size of the available training set. A classifier with too many adjustable parameters (large capacity) is likely to learn the training set without difficulty, but be unable to generalize properly to new patterns. If the capacity is too small, even the training set might not be learned without appreciable error. There is thus an intermediate, optimal classifier capacity which guarantees the best expected generalization for the given training set size. The method of structural risk minimization provides a theoretical tool for tuning the capacity of the classifier to this optimal match. It is noted that the capacity can be controlled through a variety of methods involving not only the structure of the classifier itself, but also properties of the input space that can be modified through preprocessing, as well as modifications of the learning algorithm which regularize the search for solutions to the problem of learning the training set. Experiments performed on a benchmark problem of handwritten digit recognition are discussed
Keywords :
neural nets; pattern recognition; handwritten digit recognition; input space; layered neural networks; learning algorithm; optimal classifier capacity; preprocessing; statistical pattern recognition; structural risk minimization; training set; Adaptive systems; Data preprocessing; Extraterrestrial measurements; Neural networks; Optimal matching; Pattern matching; Pattern recognition; Risk management; Supervised learning; Transfer functions;
Conference_Titel :
Neural Networks for Signal Processing [1992] II., Proceedings of the 1992 IEEE-SP Workshop
Conference_Location :
Helsingoer
Print_ISBN :
0-7803-0557-4
DOI :
10.1109/NNSP.1992.253687