Welcome to P K Kelkar Library, Online Public Access Catalogue (OPAC)

Normal view MARC view ISBD view

A primer on compression in the memory hierarchy /

By: Sardashti, Somayeh [author.].
Contributor(s): Arelakis, Angelos [author.] | Stenström, Per [author.] | Wood, David Allen [author.].
Material type: materialTypeLabelBookSeries: Synthesis digital library of engineering and computer science: ; Synthesis lectures in computer architecture: # 36.Publisher: San Rafael, California (1537 Fourth Street, San Rafael, CA 94901 USA) : Morgan & Claypool, 2016.Description: 1 PDF (xvii, 68 pages) : illustrations.Content type: text Media type: electronic Carrier type: online resourceISBN: 9781627057042.Subject(s): Data compression (Computer science) | Memory hierarchy (Computer science) | cache design | memory system | memory hierarchy | data compression | performance | energy efficiencyDDC classification: 005.746 Online resources: Abstract with links to resource Also available in print.
Contents:
3. Cache compression -- 3.1 Cache compaction taxonomy -- 3.2 Cache compaction mechanisms -- 3.2.1 Simple compaction mechanisms -- 3.2.2 Supporting variable size compression -- 3.2.3 Decoupled compressed caches -- 3.2.4 Skewed compressed caches -- 3.3 Policies to manage compressed caches -- 3.4 Cache compression to improve cache power and area -- 3.5 Summary --
2. Compression algorithms -- 2.1 Value locality -- 2.2 Compression algorithm taxonomy -- 2.3 Classification of compression algorithms -- 2.3.1 Run-length encoding -- 2.3.2 Lempel-Ziv (LZ) coding -- 2.3.3 Huffman coding -- 2.3.4 Frequent value compression (FVC) -- 2.3.5 Frequent pattern compression (FPC) -- 2.3.6 Base-delta-immediate (BDI) -- 2.3.7 Cache packer (C-PACK) -- 2.3.8 Deduplication -- 2.3.9 Instruction compression -- 2.3.10 Floating-point compression -- 2.3.11 Hybrid compression -- 2.4 Metrics to evaluate the success of a compression algorithm -- 2.5 Summary --
1. Introduction --
4. Memory compression -- 4.1 Baseline system architecture of a compressed memory system -- 4.2 Compression algorithms -- 4.3 Compressed memory organizations -- 4.3.1 The IBM MXT approach -- 4.3.2 The Ekman/Stenstrom approach -- 4.3.3 The decoupled zero-compression approach -- 4.3.4 The linear compressed pages approach -- 4.4 Summary --
5. Cache/memory link compression -- 5.1 Link compression for narrow value locality -- 5.2 Link compression for clustered value locality -- 5.3 Link compression for temporal value locality -- 5.3.1 The citron scheme -- 5.3.2 Frequent value encoding -- 5.4 Link compression methods applied to compressed memory data -- 5.5 Summary --
6. Concluding remarks -- References -- Authors' biographies.
Abstract: This synthesis lecture presents the current state-of-the-art in applying low-latency, lossless hardware compression algorithms to cache, memory, and the memory/cache link. There are many non-trivial challenges that must be addressed to make data compression work well in this context. First, since compressed data must be decompressed before it can be accessed, decompression latency ends up on the critical memory access path. This imposes a significant constraint on the choice of compression algorithms. Second, while conventional memory systems store fixed-size entities like data types, cache blocks, and memory pages, these entities will suddenly vary in size in a memory system that employs compression. Dealing with variable size entities in a memory system using compression has a significant impact on the way caches are organized and how to manage the resources in main memory. We systematically discuss solutions in the open literature to these problems. Chapter 2 provides the foundations of data compression by first introducing the fundamental concept of value locality. We then introduce a taxonomy of compression algorithms and show how previously proposed algorithms fit within that logical framework. Chapter 3 discusses the different ways that cache memory systems can employ compression, focusing on the tradeoffs between latency, capacity, and complexity of alternative ways to compact compressed cache blocks. Chapter 4 discusses issues in applying data compression to main memory and Chapter 5 covers techniques for compressing data on the cache-to-memory links. This book should help a skilled memory system designer understand the fundamental challenges in applying compression to the memory hierarchy and introduce him/her to the state-of-the-art techniques in addressing them.
    average rating: 0.0 (0 votes)
Item type Current location Call number Status Date due Barcode Item holds
E books E books PK Kelkar Library, IIT Kanpur
Available EBKE674
Total holds: 0

Mode of access: World Wide Web.

System requirements: Adobe Acrobat Reader.

Part of: Synthesis digital library of engineering and computer science.

Includes bibliographical references (pages 55-66).

3. Cache compression -- 3.1 Cache compaction taxonomy -- 3.2 Cache compaction mechanisms -- 3.2.1 Simple compaction mechanisms -- 3.2.2 Supporting variable size compression -- 3.2.3 Decoupled compressed caches -- 3.2.4 Skewed compressed caches -- 3.3 Policies to manage compressed caches -- 3.4 Cache compression to improve cache power and area -- 3.5 Summary --

2. Compression algorithms -- 2.1 Value locality -- 2.2 Compression algorithm taxonomy -- 2.3 Classification of compression algorithms -- 2.3.1 Run-length encoding -- 2.3.2 Lempel-Ziv (LZ) coding -- 2.3.3 Huffman coding -- 2.3.4 Frequent value compression (FVC) -- 2.3.5 Frequent pattern compression (FPC) -- 2.3.6 Base-delta-immediate (BDI) -- 2.3.7 Cache packer (C-PACK) -- 2.3.8 Deduplication -- 2.3.9 Instruction compression -- 2.3.10 Floating-point compression -- 2.3.11 Hybrid compression -- 2.4 Metrics to evaluate the success of a compression algorithm -- 2.5 Summary --

1. Introduction --

4. Memory compression -- 4.1 Baseline system architecture of a compressed memory system -- 4.2 Compression algorithms -- 4.3 Compressed memory organizations -- 4.3.1 The IBM MXT approach -- 4.3.2 The Ekman/Stenstrom approach -- 4.3.3 The decoupled zero-compression approach -- 4.3.4 The linear compressed pages approach -- 4.4 Summary --

5. Cache/memory link compression -- 5.1 Link compression for narrow value locality -- 5.2 Link compression for clustered value locality -- 5.3 Link compression for temporal value locality -- 5.3.1 The citron scheme -- 5.3.2 Frequent value encoding -- 5.4 Link compression methods applied to compressed memory data -- 5.5 Summary --

6. Concluding remarks -- References -- Authors' biographies.

Abstract freely available; full-text restricted to subscribers or individual document purchasers.

Compendex

INSPEC

Google scholar

Google book search

This synthesis lecture presents the current state-of-the-art in applying low-latency, lossless hardware compression algorithms to cache, memory, and the memory/cache link. There are many non-trivial challenges that must be addressed to make data compression work well in this context. First, since compressed data must be decompressed before it can be accessed, decompression latency ends up on the critical memory access path. This imposes a significant constraint on the choice of compression algorithms. Second, while conventional memory systems store fixed-size entities like data types, cache blocks, and memory pages, these entities will suddenly vary in size in a memory system that employs compression. Dealing with variable size entities in a memory system using compression has a significant impact on the way caches are organized and how to manage the resources in main memory. We systematically discuss solutions in the open literature to these problems. Chapter 2 provides the foundations of data compression by first introducing the fundamental concept of value locality. We then introduce a taxonomy of compression algorithms and show how previously proposed algorithms fit within that logical framework. Chapter 3 discusses the different ways that cache memory systems can employ compression, focusing on the tradeoffs between latency, capacity, and complexity of alternative ways to compact compressed cache blocks. Chapter 4 discusses issues in applying data compression to main memory and Chapter 5 covers techniques for compressing data on the cache-to-memory links. This book should help a skilled memory system designer understand the fundamental challenges in applying compression to the memory hierarchy and introduce him/her to the state-of-the-art techniques in addressing them.

Also available in print.

Title from PDF title page (viewed on December 29, 2015).

There are no comments for this item.

Log in to your account to post a comment.

Powered by Koha