Title :
High Dimensional Matrix Relevance Learning
Author :
Schleif, Frank-Michael ; Villmann, Thomas ; Xibin Zhu
Author_Institution :
Sch. of Comput. Sci., Univ. of Birmingham, Birmingham, UK
Abstract :
In supervised learning the parameters of a parametric Euclidean distance or mahalanobis distance can be effectively learned by so called Matrix Relevance Learning. This adaptation is not only useful to improve the discrimination capabilities of the model, but also to identify relevant features or relevant correlated features in the input data. Classical Matrix Relevance Learning scales quadratic with the number of input dimensions M and becomes prohibitive if M exceeds some thousand input features. We address Matrix Relevance Learning for data with a very large number of input dimensions. Such high dimensional data occur frequently in the life sciences domain e.g. For microarray or spectral data. We derive two respective approximation schemes and show exemplarily the implementation in Generalized Matrix Relevance Learning (GMLVQ) for classification problems. The first approximation scheme is based on Limited Rank Matrix Approximation (LiRaM) LiRaM is a random subspace projection technique which was formerly mainly considered for visualization purposes. The second novel approximation scheme is based on the Nystroem approximation and is exact if the number of Eigen values equals the rank of the Relevance Matrix. Using multiple benchmark problems, we demonstrate that the training process yields fast low rank approximations of the relevance matrices without harming the generalization ability. The approaches can be used to identify discriminative features for high dimensional data sets.
Keywords :
approximation theory; eigenvalues and eigenfunctions; learning (artificial intelligence); matrix algebra; pattern classification; GMLVQ; LiRaM; Mahalanobis distance; Nystroem approximation; approximation schemes; classification problems; discrimination capabilities; eigenvalues; features identification; generalized matrix relevance learning; high dimensional data sets; high dimensional matrix relevance learning; limited rank matrix approximation; parametric Euclidean distance; random subspace projection technique; supervised learning; Approximation methods; Complexity theory; Eigenvalues and eigenfunctions; Kernel; Matrix decomposition; Prototypes; Training; discriminative embedding; feature reduction; high dimensional; low rank approximation; relevance learning;
Conference_Titel :
Data Mining Workshop (ICDMW), 2014 IEEE International Conference on
Conference_Location :
Shenzhen
Print_ISBN :
978-1-4799-4275-6
DOI :
10.1109/ICDMW.2014.15