DocumentCode :
1115815
Title :
The Relationship of the Bayes Risk to Certain Separability Measures in Normal Classification
Author :
Yablon, Marvin ; Chu, J.T.
Author_Institution :
MEMBER, IEEE, Department of Mathematics, John Jay College of Criminal Justice, The City University of New York, New York, NY 10019.
Issue :
2
fYear :
1980
fDate :
3/1/1980 12:00:00 AM
Firstpage :
97
Lastpage :
100
Abstract :
For the problem of classifying an element (e.g., an unknown pattern) into one of two given categories where the associated observables are distributed according to one of two known multivariate normal populations having a common covariance matrix, it is shown that the minimum Bayes risk is a strict monotonic function of certain separability or statistical distance measures regardless of the a priori probabilities and the assigned loss function. However, for the associated conditional expected losses, strict monotonicity holds, if and only if a certain condition dependent on these probabilities and the given loss function is satisfied. These results remain valid for classification problems in which the observable can be transformed by a one-to-one differentiable mapping to normality.
Keywords :
Bayesian methods; Covariance matrix; Error probability; Input variables; Loss measurement; Mathematics; Parameter estimation; Pattern recognition; Probability density function; Probability distribution; Bhattacharyya coefficient and distance; Chernoff bound; Kolmogorov´s variational distance; Matusita distance; divergence; error probability in pattern recognition; minimum Bayes risk; multivariate normal classification; two classificatory categories;
fLanguage :
English
Journal_Title :
Pattern Analysis and Machine Intelligence, IEEE Transactions on
Publisher :
ieee
ISSN :
0162-8828
Type :
jour
DOI :
10.1109/TPAMI.1980.4766987
Filename :
4766987
Link To Document :
بازگشت