• DocumentCode
    969688
  • Title

    Subclass discriminant analysis

  • Author

    Manli Zhu ; Martinez, A.M.

  • Author_Institution
    Dept. of Electr. & Comput. Eng., Ohio State Univ., Columbus, OH
  • Volume
    28
  • Issue
    8
  • fYear
    2006
  • Firstpage
    1274
  • Lastpage
    1286
  • Abstract
    Over the years, many discriminant analysis (DA) algorithms have been proposed for the study of high-dimensional data in a large variety of problems. Each of these algorithms is tuned to a specific type of data distribution (that which best models the problem at hand). Unfortunately, in most problems the form of each class pdf is a priori unknown, and the selection of the DA algorithm that best fits our data is done over trial-and-error. Ideally, one would like to have a single formulation which can be used for most distribution types. This can be achieved by approximating the underlying distribution of each class with a mixture of Gaussians. In this approach, the major problem to be addressed is that of determining the optimal number of Gaussians per class, i.e., the number of subclasses. In this paper, two criteria able to find the most convenient division of each class into a set of subclasses are derived. Extensive experimental results are shown using five databases. Comparisons are given against linear discriminant analysis (LDA), direct LDA (DLDA), heteroscedastic LDA (HLDA), nonparametric DA (NDA), and kernel-based LDA (K-LDA). We show that our method is always the best or comparable to the best
  • Keywords
    Gaussian distribution; feature extraction; Gaussian mixture; data distribution; direct LDA; heteroscedastic LDA; high-dimensional data; kernel-based LDA; linear discriminant analysis; nonparametric DA; subclass discriminant analysis; trial-and-error selection; Algorithm design and analysis; Data mining; Databases; Eigenvalues and eigenfunctions; Gaussian approximation; Gaussian distribution; Gaussian processes; Linear discriminant analysis; Pattern analysis; Pattern recognition; Feature extraction; classification; discriminant analysis; eigenvalue decomposition; mixture of Gaussians.; pattern recognition; stability criterion; Algorithms; Artificial Intelligence; Data Interpretation, Statistical; Discriminant Analysis; Image Enhancement; Image Interpretation, Computer-Assisted; Imaging, Three-Dimensional; Information Storage and Retrieval; Models, Biological; Models, Statistical; Pattern Recognition, Automated; Reproducibility of Results; Sensitivity and Specificity;
  • fLanguage
    English
  • Journal_Title
    Pattern Analysis and Machine Intelligence, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0162-8828
  • Type

    jour

  • DOI
    10.1109/TPAMI.2006.172
  • Filename
    1642662