Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence 2020
DOI: 10.24963/ijcai.2020/417
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating and Aggregating Feature-based Model Explanations

Abstract: A feature-based model explanation denotes how much each input feature contributes to a model's output for a given data point. As the number of proposed explanation functions grows, we lack quantitative evaluation criteria to help practitioners know when to use which explanation function. This paper proposes quantitative evaluation criteria for feature-based explanations: low sensitivity, high faithfulness, and low complexity. We devise a framework for aggregating explanation functions. We develop a pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
72
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 91 publications
(72 citation statements)
references
References 1 publication
0
72
0
Order By: Relevance
“…As this is rarely available, a typical approach for measuring the faithfulness of the contributions produced by interpretable models is to rely on the proxy notion of the contributions: observing the effect of removing features on the model's prediction. Following previous studies [11], [50], we computed the faithfulness score by removing features one-by-one, measuring the differences between the original predictions and the predictions from the inputs without the removed features, and calculating the correlation between the differences and the contributions of the removed features. Formally, the faithfulness score for set of test samples D te = {(x * , z * )} is calculated as follows:…”
Section: Results On Real Datasetsmentioning
confidence: 99%
“…As this is rarely available, a typical approach for measuring the faithfulness of the contributions produced by interpretable models is to rely on the proxy notion of the contributions: observing the effect of removing features on the model's prediction. Following previous studies [11], [50], we computed the faithfulness score by removing features one-by-one, measuring the differences between the original predictions and the predictions from the inputs without the removed features, and calculating the correlation between the differences and the contributions of the removed features. Formally, the faithfulness score for set of test samples D te = {(x * , z * )} is calculated as follows:…”
Section: Results On Real Datasetsmentioning
confidence: 99%
“…Consequently, approximate methods have been proposed, e.g. aggregation-based methods [ 40 ], Monte Carlo sampling [ 41 ] and approaches for graph-structured data, e.g. language and image data [ 42 ].…”
Section: Explainability Methodsmentioning
confidence: 99%
“…Narayanan et al [92] compared different types of output complexity for how they affected human performance. Bhatt et al [12] designed a complexity metric to quantify "feature importance" explanations.…”
Section: Evaluating Meaningfulnessmentioning
confidence: 99%
“…They then compared the system's scores on the new images. Bhatt et al [12] evaluated the explanation accuracy of "feature importance" explanations by both checking sensitivity, meaning similar inputs have similar feature importance explanations, and faithfulness, meaning the change in the explanations should correlate to the change in inputs.…”
Section: Evaluating Explanation Accuracymentioning
confidence: 99%