2022
DOI: 10.1109/access.2022.3221138
|View full text |Cite
|
Sign up to set email alerts
|

Transformers Meet Small Datasets

Abstract: The research and application areas of transformers have been extensively enlarged due to the success of vision transformers (ViTs). However, due to the lack of local content acquisition capabilities, the pure transformer architectures cannot be trained directly on small datasets. In this work, we first propose a new hybrid model by combining the transformer and convolution neural network (CNN). The proposed model improves the classification ability on small datasets. This is accomplished by introducing more co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 14 publications
(2 citation statements)
references
References 38 publications
0
2
0
Order By: Relevance
“…After the last T-Block that focuses more on spectral attention, we further enhance the spatial features using spatialspectral domain learning (SDL) module 37 , whose output is the desired deblurred multispectral image e Y = f À1 B ðY Þ. It is quite interesting to notice that numerous recent articles have proposed and successfully demonstrated the training of the Transformer with just small data [38][39][40][41] . The CODE addresses the challenge of small data learning using a completely different philosophy.…”
Section: Code-based Small-data Learning Theorymentioning
confidence: 99%
See 1 more Smart Citation
“…After the last T-Block that focuses more on spectral attention, we further enhance the spatial features using spatialspectral domain learning (SDL) module 37 , whose output is the desired deblurred multispectral image e Y = f À1 B ðY Þ. It is quite interesting to notice that numerous recent articles have proposed and successfully demonstrated the training of the Transformer with just small data [38][39][40][41] . The CODE addresses the challenge of small data learning using a completely different philosophy.…”
Section: Code-based Small-data Learning Theorymentioning
confidence: 99%
“…The CODE addresses the challenge of small data learning using a completely different philosophy. Simply speaking, typical techniques [38][39][40][41] have to force the deep network to return a good deep solution (as the final solution), while CODE just accepts the weak DE solution. CODE assumes that though the small scale of data results in such a weak solution, the solution itself still contains useful information.…”
Section: Code-based Small-data Learning Theorymentioning
confidence: 99%