2024
DOI: 10.1609/aaai.v38i3.28047
|View full text |Cite
|
Sign up to set email alerts
|

Rethinking Robustness of Model Attributions

Sandesh Kamath,
Sankalp Mittal,
Amit Deshpande
et al.

Abstract: For machine learning models to be reliable and trustworthy, their decisions must be interpretable. As these models find increasing use in safety-critical applications, it is important that not just the model predictions but also their explanations (as feature attributions) be robust to small human-imperceptible input perturbations. Recent works have shown that many attribution methods are fragile and have proposed improvements in either these methods or the model training. We observe two main causes for fragil… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
references
References 31 publications
(63 reference statements)
0
0
0
Order By: Relevance