• DocumentCode
    27017
  • Title

    Dictionary Training for Sparse Representation as Generalization of K-Means Clustering

  • Author

    Sahoo, Sujit Kumar ; Makur, Anuran

  • Author_Institution
    Sch. of Electr. Electron. Eng., Nanyang Technol. Univ., Singapore, Singapore
  • Volume
    20
  • Issue
    6
  • fYear
    2013
  • fDate
    Jun-13
  • Firstpage
    587
  • Lastpage
    590
  • Abstract
    Recent dictionary training algorithms for sparse representation like K-SVD, MOD, and their variation are reminiscent of K-means clustering, and this letter investigates such algorithms from that viewpoint. It shows: though K-SVD is sequential like K-means, it fails to simplify to K-means by destroying the structure in the sparse coefficients. In contrast, MOD can be viewed as a parallel generalization of K-means, which simplifies to K-means without perturbing the sparse coefficients. Keeping memory usage in mind, we propose an alternative to MOD; a sequential generalization of K-means (SGK). While experiments suggest a comparable training performances across the algorithms, complexity analysis shows MOD and SGK to be faster under a dimensionality condition.
  • Keywords
    pattern clustering; signal representation; -SVD; K-means clustering generalization; MOD; SGK; complexity analysis; dictionary training; dimensionality condition; sparse coefficient; sparse representation; Approximation methods; Complexity theory; Dictionaries; Encoding; Minimization; Signal processing algorithms; Training; Dictionary training; K-SVD; K-means; MOD;
  • fLanguage
    English
  • Journal_Title
    Signal Processing Letters, IEEE
  • Publisher
    ieee
  • ISSN
    1070-9908
  • Type

    jour

  • DOI
    10.1109/LSP.2013.2258912
  • Filename
    6504716