2020 ACM/IEEE 47th Annual International Symposium on Computer Architecture (ISCA) 2020
DOI: 10.1109/isca45697.2020.00075
|View full text |Cite
|
Sign up to set email alerts
|

JPEG-ACT: Accelerating Deep Learning via Transform-based Lossy Compression

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
31
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 42 publications
(31 citation statements)
references
References 32 publications
0
31
0
Order By: Relevance
“…Note that convolutional layers also dominate the computational time during training process, which benefits us for apply compression with low overhead. Similar to many previous studies [13,24,42,58], our research goal is to develop an efficient and generic strategy to achieve a high reduction in memory consumption for CNN training. Our work can increase the batch size limit and convergence speed or enable training on the hardware with lower memory capacity for the same CNN model.…”
Section: Research Goals and Challengesmentioning
confidence: 99%
See 1 more Smart Citation
“…Note that convolutional layers also dominate the computational time during training process, which benefits us for apply compression with low overhead. Similar to many previous studies [13,24,42,58], our research goal is to develop an efficient and generic strategy to achieve a high reduction in memory consumption for CNN training. Our work can increase the batch size limit and convergence speed or enable training on the hardware with lower memory capacity for the same CNN model.…”
Section: Research Goals and Challengesmentioning
confidence: 99%
“…Moreover, memory compression approaches based on lossless compression of activation data [49] suffer from the limited compression ratio (e.g., only around 2:1 for most floatingpoint data). Alternatively, recent works [6,13] proposed to develop compression offloading accelerators for reducing the activation data size before transferring it to the CPU DRAM. However, adding a new dedicated hardware component to the existing GPU architecture requires tremendous industry efforts and is not ready for immediate deployment.…”
Section: Introductionmentioning
confidence: 99%
“…[11][12][13][14][15] focus on designing hardware-friendly compression encoders and decompression decoders or reducing the calculation cost of sparse data by using a special storage format. [16][17][18][19] design algorithms to compress data transferred between GPU and CPU memory during training.…”
Section: Introductionmentioning
confidence: 99%
“…We observe the specific 2-D frequency domain information of activations, which demonstrates the fact that we can use 2D-DCT to compress feature maps in earlier layers of CNN. Then we combine Approximate Sparsity Preprocessing (ASP) and lowbit quantization to compress feature maps in the later layers, which achieves a compression ratio of 2.6× (+30% over [16]).…”
mentioning
confidence: 99%
See 1 more Smart Citation