2021
DOI: 10.1109/tpami.2021.3066410
|View full text |Cite
|
Sign up to set email alerts
|

Discrimination-aware Network Pruning for Deep Model Compression

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
49
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 69 publications
(50 citation statements)
references
References 54 publications
1
49
0
Order By: Relevance
“…Aside from quantization, there are also several other popular techniques to compress PLMs. Pruning removes unimportant parameters or connections in a well-trained model [36], [37], [38], [39], [40], and is widely explored in PLMs [41], [42]. A direct way is to remove connections with small magnitudes during the pre-training and adds them back when necessary in downstream tasks [41].…”
Section: Network Compression For Pre-trained Language Modelsmentioning
confidence: 99%
“…Aside from quantization, there are also several other popular techniques to compress PLMs. Pruning removes unimportant parameters or connections in a well-trained model [36], [37], [38], [39], [40], and is widely explored in PLMs [41], [42]. A direct way is to remove connections with small magnitudes during the pre-training and adds them back when necessary in downstream tasks [41].…”
Section: Network Compression For Pre-trained Language Modelsmentioning
confidence: 99%
“…In the case of ThiNet [32], x is a point sampled from the input feature map and reconstruction(x) is its reconstruction after pruning. Metrics used by ThiNet [32], He et al [33], and Liu et al [34] choose channels based on the least error incurred to output feature maps. Hence, the layer-wise feature maps error is used as a saliency metric.…”
Section: Weight and Input Images Based Saliency Metricsmentioning
confidence: 99%
“…The main difference between the approach explored by ThiNet [32] and He et al [33] is how they estimate the damaged feature map. Liu et al [34] also introduce a layer-wise loss alongside the reconstruction error.…”
Section: Weight and Input Images Based Saliency Metricsmentioning
confidence: 99%
See 1 more Smart Citation
“…However, a typical deep model is difficult to be deployed on such devices. Due to its limited resource, the model is required to be compressed according to its storage space [9,10], which might reduce the performance. Meanwhile, the inference speed is slow, since its computational capacity is much worse than the external devices having dedicated graphics cards.…”
Section: Introductionmentioning
confidence: 99%