DocumentCode :
3710118
Title :
Random Matrices: l1 Concentration and Dictionary Learning with Few Samples
Author :
Kyle Luh;Van Vu
Author_Institution :
Dept. of Math., Yale Univ., New Haven, CT, USA
fYear :
2015
Firstpage :
1409
Lastpage :
1425
Abstract :
Let X be a sparse random matrix of size n × p (p ≫ n). We prove that if p ≥ Cn log4 n, then with probability 1 - o(1), ∥XT v∥1 is close to its expectation for all vectors v ∈ Rn (simultaneously). The bound on p is sharp up to the polylogarithmic factor. The study of this problem is directly motivated by an application. Let A be an n×n matrix, X be an n×p matrix and Y = AX. A challenging and important problem in data analysis, motivated by dictionary learning and other practical problems, is to recover both A and X, given Y . Under normal circumstances, it is clear that this problem is underdetermined. However, in the case when X is sparse and random, Spielman, Wang and Wright showed that one can recover both A and X efficiently from Y with high probability, given that p (the number of samples) is sufficiently large. Their method works for p ≥ Cn2 log2 n and they conjectured that p ≥ Cn log n suffices. The bound n log n is sharp for an obvious information theoretical reason. The matrix concentration result verifies the Spielman et. al. conjecture up to a log3 n factor. Our proof of the concentration result is based on two ideas. The first is an economical way to apply the union bound. The second is a refined version of Bernstein´s concentration inequality for a sum of independent variables. Both have nothing to do with random matrices and are applicable in general settings.
Keywords :
"Sparse matrices","Dictionaries","Yttrium","Linear matrix inequalities","Random variables","Algorithm design and analysis","Standards"
Publisher :
ieee
Conference_Titel :
Foundations of Computer Science (FOCS), 2015 IEEE 56th Annual Symposium on
ISSN :
0272-5428
Type :
conf
DOI :
10.1109/FOCS.2015.90
Filename :
7354464
Link To Document :
بازگشت