2021 IEEE International Parallel and Distributed Processing Symposium (IPDPS) 2021
DOI: 10.1109/ipdps49936.2021.00097
|View full text |Cite
|
Sign up to set email alerts
|

Revisiting Huffman Coding: Toward Extreme Performance on Modern GPU Architectures

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
7
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

3
5

Authors

Journals

citations
Cited by 23 publications
(11 citation statements)
references
References 26 publications
1
7
0
Order By: Relevance
“…3) Test Datasets: We conduct our evaluation and comparison based on eight typical 1D∼4D real-world HPC simulation datasets, including six from Scientific Data Reduction Benchmarks [34]: 1D HACC cosmology simulation [12], 2D LAMMPS (part of the EXAALT ECP project) molecular dynamics simulation [24], 3D CESM-ATM climate simulation [6], 3D Nyx cosmology simulation [31], 4D Hurricane ISABEL simulation [16], and 4D QMCPack quantum simulation [32]. They have been widely used in much prior work [37,26,27,47,46,38,40,39,20,4] and are good representatives of production-level simulation datasets. Additionally, we also evaluate two datasets that highlight our decoders' potential to be used as in-memory compressors as discussed in §I, including 3D RTM simulation data for petroleum exploration [17] and 1D GAMESS data for quantum chemistry simulation [10].…”
Section: Performance Evaluationmentioning
confidence: 99%
See 2 more Smart Citations
“…3) Test Datasets: We conduct our evaluation and comparison based on eight typical 1D∼4D real-world HPC simulation datasets, including six from Scientific Data Reduction Benchmarks [34]: 1D HACC cosmology simulation [12], 2D LAMMPS (part of the EXAALT ECP project) molecular dynamics simulation [24], 3D CESM-ATM climate simulation [6], 3D Nyx cosmology simulation [31], 4D Hurricane ISABEL simulation [16], and 4D QMCPack quantum simulation [32]. They have been widely used in much prior work [37,26,27,47,46,38,40,39,20,4] and are good representatives of production-level simulation datasets. Additionally, we also evaluate two datasets that highlight our decoders' potential to be used as in-memory compressors as discussed in §I, including 3D RTM simulation data for petroleum exploration [17] and 1D GAMESS data for quantum chemistry simulation [10].…”
Section: Performance Evaluationmentioning
confidence: 99%
“…For example, Lal et al proposed a Huffman-based entropy encoding system (E 2 MC) for GPUs [23]. More recently, Tian et al proposed a fast parallel Huffman codebook construction algorithm and a parallel Huffman encoder for modern GPU architectures [40]. Since much work has already been focused on optimizing Huffman encoding, we do not presently consider optimizing encoding in our work.…”
Section: Use-case Of Our Two Decodersmentioning
confidence: 99%
See 1 more Smart Citation
“…Note that our design is different from other Huffman coding works in terms of adaptivity. For example, Tian et al [50] proposed a reduction-based scheme for GPUs that iteratively merges the encoded symbols and adaptively determines the number of merge iterations. However, CEAZ only builds a new codebook for the data chunk when the change of its histogram exceeds a threshold in order to target FPGA with limited resources and low clock frequency.…”
Section: Adaptive Online Codewords Updatesmentioning
confidence: 99%
“…Currently, several GPU-based error-controlled lossy compressors (such as CUSZ [13] and cuZFP [14]) have been developed, but they suffer from either sub-optimal compression throughput or low compression ratios. For instance, CUSZ can achieve much higher compression ratios than cuZFP, but its performance is substantially limited by the Huffman encoding stage and dictionary encoding step [15]. However, the high compression ratios of SZ/CUSZ significantly depend on Huffman encoding and dictionary encoding, because the output of the prediction-and-quantization step in SZ/CUSZ is often composed of a large amount of repeated symbols.…”
Section: Introductionmentioning
confidence: 99%