DocumentCode :
3129920
Title :
Adaptive cache compression for high-performance processors
Author :
Alameldeen, Alaa R. ; Wood, David A.
Author_Institution :
Dept. of Comput. Sci., Wisconsin Univ., Madison, WI, USA
fYear :
2004
fDate :
19-23 June 2004
Firstpage :
212
Lastpage :
223
Abstract :
Modern processors use two or more levels of cache memories to bridge the rising disparity between processor and memory speeds. Compression can improve cache performance by increasing effective cache capacity and eliminating misses. However, decompressing cache lines also increases cache access latency, potentially degrading performance. In this paper, we develop an adaptive policy that dynamically adapts to the costs and benefits of cache compression. We propose a two-level cache hierarchy where the L1 cache holds uncompressed data and the L2 cache dynamically selects between compressed and uncompressed storage. The L2 cache is 8-way set-associative with LRU replacement, where each set can store up to eight compressed lines but has space for only four uncompressed lines. On each L2 reference, the LRU stack depth and compressed size determine whether compression (could have) eliminated a miss or incurs an unnecessary decompression overhead. Based on this outcome, the adaptive policy updates a single global saturating counter, which predicts whether to allocate lines in compressed or uncompressed form. We evaluate adaptive cache compression using full-system simulation and a range of benchmarks. We show that compression can improve performance for memory-intensive commercial workloads by up to 17%. However, always using compression hurts performance for low-miss-rate benchmarks - due to unnecessary decompression overhead - degrading performance by up to 18%. By dynamically monitoring workload behavior, the adaptive policy achieves comparable benefits from compression, while never degrading performance by more than 0.4%.
Keywords :
cache storage; data compression; memory architecture; memory protocols; microprocessor chips; parallel architectures; 8-way set-associative; LRU replacement; LRU stack depth; adaptive cache compression; adaptive policy; cache access latency; cache capacity; cache line decompression; cache memories; data compression; decompression overhead; high-performance processors; low-miss-rate benchmarks; memory-intensive commercial workloads; two-level cache hierarchy; workload behavior monitoring; Bridges; Cache memory; Cache storage; Counting circuits; Degradation; Delay; Microarchitecture; Monitoring; Moore´s Law; Technological innovation;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computer Architecture, 2004. Proceedings. 31st Annual International Symposium on
ISSN :
1063-6897
Print_ISBN :
0-7695-2143-6
Type :
conf
DOI :
10.1109/ISCA.2004.1310776
Filename :
1310776
Link To Document :
بازگشت