2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00891
|View full text |Cite
|
Sign up to set email alerts
|

Attribute Surrogates Learning and Spectral Tokens Pooling in Transformers for Few-shot Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 32 publications
(7 citation statements)
references
References 35 publications
0
5
0
Order By: Relevance
“…Some researchers utilize long-range relationships of local image patches and transformer networks with the selfattention mechanism to achieve better image recognition performance. Recent works [6] [7] [8] have proposed either combining or replacing CNNs with transformer networks. These efforts often entail employing long token sequence learning within transformers to mitigate catastrophic forgetting.…”
Section: Introductionmentioning
confidence: 99%
“…Some researchers utilize long-range relationships of local image patches and transformer networks with the selfattention mechanism to achieve better image recognition performance. Recent works [6] [7] [8] have proposed either combining or replacing CNNs with transformer networks. These efforts often entail employing long token sequence learning within transformers to mitigate catastrophic forgetting.…”
Section: Introductionmentioning
confidence: 99%
“…Moreover, token sparsification can also remove spurious correlations to improve the testing accuracy (Likhomanenko et al, 2021;Zhu et al, 2021a). This insight provides a guideline in designing token sparsification and few-shot learning methods for Transformer (He et al, 2022;Guibas et al, 2022).…”
Section: Introductionmentioning
confidence: 99%
“…In addition, transformer-based methods often need significant training iterations and are hard to converge during the training process. [103] proposed hierarchically cascaded transformers that exploit spectral tokens pooling and surrogates learning to improve data efficiency. However, it used multiple transformers and caused the instability of training.…”
Section: Few-shot Classification Methodsmentioning
confidence: 99%
“…In this section, we evaluate our method on popular benchmark datasets for few-shot classifi- to 80 epochs, while previous methods [103,34] needs more than 400 epochs to converge.…”
Section: Implementation Detailsmentioning
confidence: 99%
See 1 more Smart Citation