Title :
On the probabilistic interpretation of neural network classifiers and discriminative training criteria
Author_Institution :
Lehrstuhl fuer Inf. VI, Tech. Hochschule Aachen, Germany
fDate :
2/1/1995 12:00:00 AM
Abstract :
A probabilistic interpretation is presented for two important issues in neural network based classification, namely the interpretation of discriminative training criteria and the neural network outputs as well as the interpretation of the structure of the neural network. The problem of finding a suitable structure of the neural network can be linked to a number of well established techniques in statistical pattern recognition. Discriminative training of neural network outputs amounts to approximating the class or posterior probabilities of the classical statistical approach. This paper extends these links by introducing and analyzing novel criteria such as maximizing the class probability and minimizing the smoothed error rate. These criteria are defined in the framework of class conditional probability density functions. We show that these criteria can be interpreted in terms of weighted maximum likelihood estimation. In particular, this approach covers widely used techniques such as corrective training, learning vector quantization, and linear discriminant analysis
Keywords :
maximum likelihood estimation; neural nets; pattern classification; probability; speech recognition; statistical analysis; discriminative training criteria; neural network classifiers; probabilistic interpretation; probability density functions; smoothed error rate; speech recognition; statistical pattern recognition; weighted maximum likelihood estimation; Automatic speech recognition; Error analysis; Helium; Kernel; Linear discriminant analysis; Neural networks; Pattern recognition; Probability density function; Speech recognition; Vector quantization;
Journal_Title :
Pattern Analysis and Machine Intelligence, IEEE Transactions on