Title :
An overhead reduction technique for mega-state compression schemes
Author :
Bookstein, A. ; Klein, S.T. ; Raita, T.
Author_Institution :
Chicago Univ., IL, USA
Abstract :
Many of the most effective compression methods involve complicated models. Unfortunately, as model complexity increases, so does the cost of storing the model itself. This paper examines a method to reduce the amount of storage needed to represent a Markov model with an extended alphabet, by applying a clustering scheme that brings together similar states. Experiments run on a variety of large natural language texts show that much of the overhead of storing the model can be saved at the cost of a very small loss of compression efficiency
Keywords :
Markov processes; data compression; natural languages; word processing; Markov model; clustering scheme; compression efficiency; experiments; extended alphabet; megastate compression schemes; model complexity; natural language texts; overhead reduction; storage reduction; Character generation; Context modeling; Costs; Data compression; Decoding; Frequency; History; Natural languages; State-space methods; Text recognition;
Conference_Titel :
Data Compression Conference, 1997. DCC '97. Proceedings
Conference_Location :
Snowbird, UT
Print_ISBN :
0-8186-7761-9
DOI :
10.1109/DCC.1997.582061