2022
DOI: 10.1016/j.ymssp.2022.109174
|View full text |Cite
|
Sign up to set email alerts
|

Contrastive-weighted self-supervised model for long-tailed data classification with vision transformer augmented

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
2
0
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 39 publications
(3 citation statements)
references
References 22 publications
0
2
0
1
Order By: Relevance
“…A preliminary study investigating masked pretraining for transformers demonstrates its fine-tuning performance on a data-scarce, 4-way bearing classification task. [33] Contrastive pretraining was also considered for downstream finetuning on long-tailed fault distributions [34], and for learning more robust latent features [35]. These studies verify the use of pre-training for bearing classification; however, they pretrain and fine-tune on the same classes and datasets.…”
Section: Introductionmentioning
confidence: 90%
“…A preliminary study investigating masked pretraining for transformers demonstrates its fine-tuning performance on a data-scarce, 4-way bearing classification task. [33] Contrastive pretraining was also considered for downstream finetuning on long-tailed fault distributions [34], and for learning more robust latent features [35]. These studies verify the use of pre-training for bearing classification; however, they pretrain and fine-tune on the same classes and datasets.…”
Section: Introductionmentioning
confidence: 90%
“…Vision Transformer (ViT), which is based on self-attention, can capture long-distance dependencies in images and extract global features of images, and it is becoming a new direction in the field of computer vision [24]. ViT has been applied in various fields including industry [25], medicine [26] and agriculture [27] but currently has fewer applications in the field of livestock phenotypic measurements. The outstanding performance of ViT is based on a huge data size and sacrifices a large number of computational resources, making it difficult to apply to small datasets [28].…”
Section: Introductionmentioning
confidence: 99%
“…Essas abordagens enfrentam consideráveis quedas de eficácia quando avaliadas sob as perspectivas do desbalanceamento e da qualidade, especialmente em relac ¸ão aos rótulos menos frequentes, doravante referidos como rótulos tail. De fato, os rótulos tail correspondem a cerca de 80% do espac ¸o de rótulos em vários cenários do mundo real [Hou et al 2022, Ge et al 2022, Huang and Wu 2019.…”
Section: Introduc ¸ãOunclassified