2021 IEEE International Parallel and Distributed Processing Symposium (IPDPS) 2021
DOI: 10.1109/ipdps49936.2021.00095
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating Multigrid-based Hierarchical Scientific Data Refactoring on GPUs

Abstract: Rapid growth in scientific data and a widening gap between computational speed and I/O bandwidth makes it increasingly infeasible to store and share all data produced by scientific simulations. Instead, we need methods for reducing data volumes: ideally, methods that can scale data volumes adaptively so as to enable negotiation of performance and fidelity tradeoffs in different situations. Multigrid-based hierarchical data representations hold promise as a solution to this problem, allowing for flexible conver… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
8
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(8 citation statements)
references
References 28 publications
0
8
0
Order By: Relevance
“…Compared with cuSZ and cuMGARD, cuZFP provides slightly higher compression throughput, but it only supports fixedrate mode [19], limiting its adoption in practice. Both cuSZ and cuMGARD use Huffman encoding to achieve high compression ratios and their decompression throughput is greatly limited by slow Huffman decoding on GPUs, but cuSZ has a much higher throughput than cuMGARD [38,5]. Thus, in this work, we focus on optimizing Huffman decoding for cuSZ.…”
Section: Error-bounded Lossy Compression On Gpumentioning
confidence: 99%
See 2 more Smart Citations
“…Compared with cuSZ and cuMGARD, cuZFP provides slightly higher compression throughput, but it only supports fixedrate mode [19], limiting its adoption in practice. Both cuSZ and cuMGARD use Huffman encoding to achieve high compression ratios and their decompression throughput is greatly limited by slow Huffman decoding on GPUs, but cuSZ has a much higher throughput than cuMGARD [38,5]. Thus, in this work, we focus on optimizing Huffman decoding for cuSZ.…”
Section: Error-bounded Lossy Compression On Gpumentioning
confidence: 99%
“…SZ, ZFP, and MGARD were first developed for CPU architectures, and all started rolling out their GPU-based lossy compression recently. The SZ team, the ZFP team, and the MGARD team released their CUDA versions, called cuSZ [39], cuZFP [7], and cuMGARD [5], respectively. All the versions provide much higher throughputs for compression and decompression compared with their CPU versions [39,19,38].…”
Section: Error-bounded Lossy Compression On Gpumentioning
confidence: 99%
See 1 more Smart Citation
“…As the type of processor that contributes the most of the computing parallelism in many current and future HPC systems, Graphics Processing Units (GPUs), equipped with thousands of low-power cores, offer high computational power and energy efficiency. Many applications and libraries have been designed and optimized for GPU accelerators [1,3,8,9,13,25,34,36,42,43]. Benefiting from the fact that GPUs are designed for highly parallelizable computations while CPUs are more efficient with serial computations, CPUs and GPUs that are linked through fast interconnections [30,31] are usually used together to form heterogeneous systems that can efficiently handle a large spectrum of scientific computing workloads.…”
Section: Introductionmentioning
confidence: 99%
“…Limitations of state-of-the-art approaches. Existing error-bounded lossy compressors for GPUs (such as cuSZ [14], cuZFP [15], and MGARD-GPU [16]) suffer from either low throughputs or low compression ratios. Specifically, although cuZFP has slightly higher throughput compared with cuSZ and MGARD-GPU, it supports only the fixed-rate mode [17], which suffers much lower compression quality than the fixed-accuracy mode (a.k.a error-bounded mode) [18], significantly limiting its adoption in practice.…”
Section: Introductionmentioning
confidence: 99%