• DocumentCode
    2081003
  • Title

    Accelerated Kernel Feature Analysis

  • Author

    Jiang, Xianhua ; Motai, Yuichi ; Snapp, Robert R. ; Zhu, Xingquan

  • Author_Institution
    University of Vermont, Burlington,VT
  • Volume
    1
  • fYear
    2006
  • fDate
    17-22 June 2006
  • Firstpage
    109
  • Lastpage
    116
  • Abstract
    A fast algorithm, Accelerated Kernel Feature Analysis (AKFA), that discovers salient features evidenced in a sample of n unclassified patterns, is presented. Like earlier kernel-based feature selection algorithms, AKFA implicitly embeds each pattern into a Hilbert space, H, induced by a Mercer kernel. An ell-dimensional linear subspace of H is iteratively constructed by maximizing a variance condition for the nonlinearly transformed sample. This linear subspace can then be used to define more efficient data representations and pattern classifiers. AKFA requires O(elln2) operations, as compared to 0(n^3) for Sch¨olkof, Smola, and M¨uller’s Kernel Principal Component Analysis (KPCA), and O(ell^2 n^2) for Smola, Mangasarian, and Sch¨olkopf’s Sparse Kernel Feature Analysis (SKFA). Numerical experiments show that AKFA can generate more concise feature representations than both KPCA and SKFA, and demonstrate that AKFA obtains similar classification performance as KPCA for a face recognition problem.
  • Keywords
    Acceleration; Algorithm design and analysis; Face recognition; Hilbert space; Iterative algorithms; Kernel; Pattern analysis; Pattern classification; Pattern recognition; Principal component analysis;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Computer Vision and Pattern Recognition, 2006 IEEE Computer Society Conference on
  • ISSN
    1063-6919
  • Print_ISBN
    0-7695-2597-0
  • Type

    conf

  • DOI
    10.1109/CVPR.2006.43
  • Filename
    1640748