DocumentCode
3250696
Title
Complexity and rate-distortion tradeoff via successive refinement
Author
No, Albert ; Ingber, Amir ; Weissman, Tsachy
Author_Institution
Dept. of Electr. Eng., Stanford Univ., Stanford, CA, USA
fYear
2013
fDate
2-4 Oct. 2013
Firstpage
1531
Lastpage
1536
Abstract
We demonstrate how successive refinement ideas can be used in point-to-point lossy compression problems in order to reduce complexity. We show two examples, the binary-Hamming and quadratic-Gaussian cases, in which a layered code construction results in a low complexity scheme that attains optimal performance. For example, when the number of layers grows with the block length n, we show how to design an O(nlog(n)) algorithm that asymptotically achieves the rate distortion bound. We then show that with the same scheme, used with a fixed number of layers, successive refinement is achieved in the classical sense, and at the same time the second order performance (i.e. dispersion) is also tight.
Keywords
Hamming codes; binary codes; computational complexity; data compression; binary Hamming case; complexity reduction; layered code construction; point-to-point lossy compression problem; quadratic Gaussian case; rate distortion; successive refinement; Channel coding; Complexity theory; Decoding; Dispersion; Rate-distortion; Source coding; Binary source; Gaussian source; complexity; rate-distortion; refined strong covering lemma; source dispersion; sparse regression code; successive refinement;
fLanguage
English
Publisher
ieee
Conference_Titel
Communication, Control, and Computing (Allerton), 2013 51st Annual Allerton Conference on
Conference_Location
Monticello, IL
Print_ISBN
978-1-4799-3409-6
Type
conf
DOI
10.1109/Allerton.2013.6736709
Filename
6736709
Link To Document