2023
DOI: 10.1007/s11263-023-01871-1
|View full text |Cite
|
Sign up to set email alerts
|

Towards Language-Guided Visual Recognition via Dynamic Convolutions

Gen Luo,
Yiyi Zhou,
Xiaoshuai Sun
et al.
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
2
2
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 75 publications
0
4
0
Order By: Relevance
“…Using less than 65,000 of the original 85 million trainable parameters, SAFT outperforms both full fine-tuning and linear probing as most known PEFT methods, including VPT-Shallow [10], VPT-Deep [10], Adapt-Former [14], LoRA [9], and FacT [11] in terms of absolute performance (accuracy). When compared to SSF [15] and RepAdapter [16], SAFT achieves competitive performance whilst using significantly fewer parameters (roughly 25%). Notably, SAFT achieves new state-of-the-art performance on three of the 19 datasets (Flower102, Pets, and EuroSAT).…”
Section: Resultsmentioning
confidence: 99%
See 3 more Smart Citations
“…Using less than 65,000 of the original 85 million trainable parameters, SAFT outperforms both full fine-tuning and linear probing as most known PEFT methods, including VPT-Shallow [10], VPT-Deep [10], Adapt-Former [14], LoRA [9], and FacT [11] in terms of absolute performance (accuracy). When compared to SSF [15] and RepAdapter [16], SAFT achieves competitive performance whilst using significantly fewer parameters (roughly 25%). Notably, SAFT achieves new state-of-the-art performance on three of the 19 datasets (Flower102, Pets, and EuroSAT).…”
Section: Resultsmentioning
confidence: 99%
“…RepAdapter [16] introduces a re-parameterizable, linear projection adapter block for Vision Transformers (ViT), primarily integrated into Multi-Head Attention (MHA) and Feed-Forward Network (FFN) layers. Its formulation, devoid of non-linear activation for simplicity, is given as:…”
Section: Repadaptermentioning
confidence: 99%
See 2 more Smart Citations