• DocumentCode
    1166413
  • Title

    Soft nearest prototype classification

  • Author

    Seo, Sambu ; Bode, Mathias ; Obermayer, Klaus

  • Author_Institution
    Dept. of Electr. Eng. & Comput. Sci., Tech. Univeristy of Berlin, Germany
  • Volume
    14
  • Issue
    2
  • fYear
    2003
  • fDate
    3/1/2003 12:00:00 AM
  • Firstpage
    390
  • Lastpage
    398
  • Abstract
    We propose a new method for the construction of nearest prototype classifiers which is based on a Gaussian mixture ansatz and which can be interpreted as an annealed version of learning vector quantization (LVQ). The algorithm performs a gradient descent on a cost-function minimizing the classification error on the training set. We investigate the properties of the algorithm and assess its performance for several toy data sets and for an optical letter classification task. Results show 1) that annealing in the dispersion parameter of the Gaussian kernels improves classification accuracy; 2) that classification results are better than those obtained with standard learning vector quantization (LVQ 2.1, LVQ 3) for equal numbers of prototypes; and 3) that annealing of the width parameter improved the classification capability. Additionally, the principled approach provides an explanation of a number of features of the (heuristic) LVQ methods.
  • Keywords
    learning (artificial intelligence); pattern classification; vector quantisation; Gaussian mixture ansatz; classification error; learning vector quantization; multiclass classification; nearest prototype classifiers; training set; Annealing; Bayesian methods; Decision theory; Euclidean distance; Kernel; Loss measurement; Optical sensors; Prototypes; Training data; Vector quantization;
  • fLanguage
    English
  • Journal_Title
    Neural Networks, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1045-9227
  • Type

    jour

  • DOI
    10.1109/TNN.2003.809407
  • Filename
    1189636