2023
DOI: 10.48550/arxiv.2301.13803
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fairness-aware Vision Transformer via Debiased Self-Attention

Abstract: Vision Transformer (ViT) has recently gained significant interest in solving computer vision (CV) problems due to its capability of extracting informative features and modeling long-range dependencies through the self-attention mechanism. To fully realize the advantages of ViT in realworld applications, recent works have explored the trustworthiness of ViT, including its robustness and explainability. However, another desiderata, fairness has not yet been adequately addressed in the literature. We establish th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 46 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?