2022
DOI: 10.1007/978-3-031-19830-4_19
|View full text |Cite
|
Sign up to set email alerts
|

Attention Diversification for Domain Generalization

Abstract: Convolutional neural networks (CNNs) have demonstrated gratifying results at learning discriminative features. However, when applied to unseen domains, state-of-the-art models are usually prone to errors due to domain shift. After investigating this issue from the perspective of shortcut learning, we find the devils lie in the fact that models trained on different domains merely bias to different domain-specific features yet overlook diverse task-related features. Under this guidance, a novel Attention Diversi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 20 publications
(5 citation statements)
references
References 74 publications
0
5
0
Order By: Relevance
“…To develop a generalizable AI model, learning PPI from the protein structures instead of leveraging the topology that inflates the performance, we need to mitigate shortcut learning and make protein representation learning independent of the PPI training data. Furthermore, regularization has been proven to be an important tool for improving generalizability in computer vision [18]. Yet, regularization is based on the assumption that similarity between a pair of entities implies interaction, an assumption that is true in image processing and social sciences [19], yet similar molecules do not necessarily always interact in PPI and other biological systems [20].…”
Section: Supplementary Informationmentioning
confidence: 99%
“…To develop a generalizable AI model, learning PPI from the protein structures instead of leveraging the topology that inflates the performance, we need to mitigate shortcut learning and make protein representation learning independent of the PPI training data. Furthermore, regularization has been proven to be an important tool for improving generalizability in computer vision [18]. Yet, regularization is based on the assumption that similarity between a pair of entities implies interaction, an assumption that is true in image processing and social sciences [19], yet similar molecules do not necessarily always interact in PPI and other biological systems [20].…”
Section: Supplementary Informationmentioning
confidence: 99%
“…CycleMAE (Yang et al 2023) uses a Masked Auto-Encoder (MAE) model to create a cycle image reconstruction task that results in more realistic and unique image pairs. Other typical domain generalization methods generally include model ensemble (Zhou et al 2021;Qu et al 2022), feature normalization (Zhu et al 2022;Meng et al 2022) or contrastive learning techniques (Yao et al 2022;Li et al 2023a).…”
Section: Domain Generalizationmentioning
confidence: 99%
“…Given this interesting setting and the promises of domain generalization in studying machine learning robustness, the community has developed a torrent of methods. Most of the existing methods fall into two categories: one is to build explicit regularization that pushes a model to learn representations that are invariant to the "style" across these domains [54][55][56][57][58][59][60][61][62][63][64][65][66][67][68][69][70][71][72]; the other one is to perform data augmentation that can introduce more diverse data to enrich the data of certain "semantic" information with the "style" from other domains [73][74][75][76][77][78][79][80], and also aims to train a model that is invariant to these "styles". More recently, there has been a line of approaches that aims to distill knowledge from pre-trained models into a smaller model to improve generalization performance [81][82][83][84][85].…”
Section: Domain Generalizationmentioning
confidence: 99%