2023 IEEE 9th Intl Conference on Big Data Security on Cloud (BigDataSecurity), IEEE Intl Conference on High Performance and Sma 2023
DOI: 10.1109/bigdatasecurity-hpsc-ids58521.2023.00033
|View full text |Cite
|
Sign up to set email alerts
|

Vision Transformer (ViT)-based Applications in Image Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 10 publications
(2 citation statements)
references
References 44 publications
0
2
0
Order By: Relevance
“…The ViT model primarily comprises three key modules: Linear Projection of Flattened Patches, Transformer Encoder, and MLP Head. ViT exhibits robust feature extraction capabilities and demonstrates resilience to various forms of noise and disturbances 33 . Pre‐trained ViT models on extensive datasets showcase remarkable generalization abilities.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The ViT model primarily comprises three key modules: Linear Projection of Flattened Patches, Transformer Encoder, and MLP Head. ViT exhibits robust feature extraction capabilities and demonstrates resilience to various forms of noise and disturbances 33 . Pre‐trained ViT models on extensive datasets showcase remarkable generalization abilities.…”
Section: Methodsmentioning
confidence: 99%
“…ViT exhibits robust feature extraction capabilities and demonstrates resilience to various forms of noise and disturbances. 33 Pre-trained ViT models on extensive datasets showcase remarkable generalization abilities.…”
Section: Vision Transformermentioning
confidence: 99%