• DocumentCode
    1307553
  • Title

    Clustered Nyström Method for Large Scale Manifold Learning and Dimension Reduction

  • Author

    Zhang, Kai ; Kwok, James T.

  • Author_Institution
    Life Sci. Div., Lawrence Berkeley Nat. Lab., Berkeley, CA, USA
  • Volume
    21
  • Issue
    10
  • fYear
    2010
  • Firstpage
    1576
  • Lastpage
    1587
  • Abstract
    Kernel (or similarity) matrix plays a key role in many machine learning algorithms such as kernel methods, manifold learning, and dimension reduction. However, the cost of storing and manipulating the complete kernel matrix makes it infeasible for large problems. The Nyström method is a popular sampling-based low-rank approximation scheme for reducing the computational burdens in handling large kernel matrices. In this paper, we analyze how the approximating quality of the Nyström method depends on the choice of landmark points, and in particular the encoding powers of the landmark points in summarizing the data. Our (non-probabilistic) error analysis justifies a “clustered Nyström method” that uses the k-means clustering centers as landmark points. Our algorithm can be applied to scale up a wide variety of algorithms that depend on the eigenvalue decomposition of kernel matrix (or its variant), such as kernel principal component analysis, Laplacian eigenmap, spectral clustering, as well as those involving kernel matrix inverse such as least-squares support vector machine and Gaussian process regression. Extensive experiments demonstrate the competitive performance of our algorithm in both accuracy and efficiency.
  • Keywords
    Gaussian processes; approximation theory; error analysis; learning (artificial intelligence); matrix algebra; pattern clustering; principal component analysis; regression analysis; support vector machines; Gaussian process regression; Laplacian eigenmap; clustered Nyström method; dimension reduction; error analysis; k-means clustering centers; kernel matrix; kernel methods; kernel principal component analysis; large scale manifold learning; least-squares support vector machine; machine learning algorithms; sampling-based low-rank approximation scheme; spectral clustering; Approximation algorithms; Approximation error; Eigenvalues and eigenfunctions; Kernel; Manifolds; Matrix decomposition; Dimension reduction; Nyström method; eigenvalue decomposition; kernel matrix; low-rank approximation; manifold learning; sampling; Algorithms; Artificial Intelligence; Cluster Analysis; Least-Squares Analysis;
  • fLanguage
    English
  • Journal_Title
    Neural Networks, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1045-9227
  • Type

    jour

  • DOI
    10.1109/TNN.2010.2064786
  • Filename
    5559473