2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.01206
|View full text |Cite
|
Sign up to set email alerts
|

DiSparse: Disentangled Sparsification for Multitask Model Compression

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(4 citation statements)
references
References 26 publications
0
4
0
Order By: Relevance
“…Then the non-local attention mechanism is applied for feature alignment: where ⊗ refers to dot product operation. In practice, the embedded features have half of the channel compared to the original feature and we adopt a neighborhood attention mechanism 47 , which can be regarded as a more efficient implementation of nonlocal attention in our experiments, achieving similar SR performance.…”
Section: Methodsmentioning
confidence: 99%
“…Then the non-local attention mechanism is applied for feature alignment: where ⊗ refers to dot product operation. In practice, the embedded features have half of the channel compared to the original feature and we adopt a neighborhood attention mechanism 47 , which can be regarded as a more efficient implementation of nonlocal attention in our experiments, achieving similar SR performance.…”
Section: Methodsmentioning
confidence: 99%
“…However, their efficacy has been largely reliant on substantial computational and storage resources, partly due to the extraction of redundant features by convolutional layers. Various model compression strategies and network designs used to improve network efficiency in the past [23]- [25], including network pruning [26]- [27],weight quantization [28] low-rank decomposition [29], and knowledge distillation [30]- [31], etc. However, these methods are all considered as post-processing steps, and thus their performance is usually bounded by the upper bound of a given initial model.…”
Section: Sc_c2f Modulementioning
confidence: 99%
“…Model-level resource-efficient learning techniques, such as pruning, quantization, and ensembling of local and global features [2], [20], [21], effectively reduce memory usage while maintaining performance comparable to that of the original dense networks. Recent studies propose resourceefficient MTL methods targeting three aspects: sharing mechanisms [5], [22], interference reduction [6], and task grouping [23]. Sharing mechanisms enable tasks to share model components, reducing model size and resource usage.…”
Section: B Resource-efficient Learningmentioning
confidence: 99%
“…Some tasks might need resource-intensive deep neural networks, while others can achieve satisfactory results with simpler, shallower models. To address these challenges, various MTL approaches [5], [6] have been proposed. These studies focus on network compression through hard parameter sharing, addressing parameter entanglement between tasks.…”
Section: Introductionmentioning
confidence: 99%