Parallelization of variable rate decompression through metadata

Conference Paper (2020)
Author(s)

Lennart Noordsij (ASML)

Steven Van Der van der Vlugt (ASML)

Mohamed A. Bamakhrama (Synopsys Corporation, Universiteit Leiden)

Z. Al-Ars (TU Delft - Computer Engineering)

Peter Lindstrom (Lawrence Livermore National Laboratory)

Research Group
Computer Engineering
Copyright
© 2020 Lennart Noordsij, Steven Van Der Vlugt, Mohamed A. Bamakhrama, Z. Al-Ars, Peter Lindstrom
DOI related publication
https://doi.org/10.1109/PDP50117.2020.00045
More Info
expand_more
Publication Year
2020
Language
English
Copyright
© 2020 Lennart Noordsij, Steven Van Der Vlugt, Mohamed A. Bamakhrama, Z. Al-Ars, Peter Lindstrom
Research Group
Computer Engineering
Pages (from-to)
245-252
ISBN (electronic)
9781728165820
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Data movement has long been identified as the biggest challenge facing modern computer systems' designers. To tackle this challenge, many novel data compression algorithms have been developed. Often variable rate compression algorithms are favored over fixed rate. However, variable rate decompression is difficult to parallelize. Most existing algorithms adopt a single parallelization strategy suited for a particular HW platform. Such an approach fails to harness the parallelism found in diverse modern HW architectures. We propose a parallelization method for tiled variable rate compression algorithms that consists of multiple strategies that can be applied interchangeably. This allows an algorithm to apply the strategy most suitable for a specific HW platform. Our strategies are based on generating metadata during encoding, which is used to parallelize the decoding process. To demonstrate the effectiveness of our strategies, we implement them in a state-of-the-art compression algorithm called ZFP. We show that the strategies suited for multicore CPUs are different from the ones suited for GPUs. On a CPU, we achieve a near optimal decoding speedup and an overhead size which is consistently less than 0.04% of the compressed data size. On a GPU, we achieve average decoding rates of up to 100 GiB/s. Our strategies allow the user to make a trade-off between decoding throughput and metadata size overhead.

Files

License info not available