2019
DOI: 10.48550/arxiv.1903.00519
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Aggregating explanation methods for stable and robust explainability

Abstract: Despite a growing literature on explaining neural networks, no consensus has been reached on how to explain a neural network decision or how to evaluate an explanation. Our contributions in this paper are twofold. First, we investigate schemes to combine explanation methods and reduce model uncertainty to obtain a single aggregated explanation. We provide evidence that the aggregation is better at identifying important features, than on individual methods. Adversarial attacks on explanations is a recent active… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 28 publications
(81 reference statements)
0
3
0
Order By: Relevance
“…In this work we address the quantification of the surrogate explanation uncertainty by aggregating multiple surrogate coefficients and measuring the consensus among the surrogate explainers. The use of a consensus mechanism to obtain explanations that are less sensitive to sampling variance (further discussed in Section 3) has been proposed in [4,26]. Specifically, [6] and [5] consider aggregating surrogate coefficients in the form of simple ranking schemes inspired from the social sciences and economics.…”
Section: Related Workmentioning
confidence: 99%
“…In this work we address the quantification of the surrogate explanation uncertainty by aggregating multiple surrogate coefficients and measuring the consensus among the surrogate explainers. The use of a consensus mechanism to obtain explanations that are less sensitive to sampling variance (further discussed in Section 3) has been proposed in [4,26]. Specifically, [6] and [5] consider aggregating surrogate coefficients in the form of simple ranking schemes inspired from the social sciences and economics.…”
Section: Related Workmentioning
confidence: 99%
“…Lakkaraju and Bastani (2020) conducted a thought-provoking study on misleading effects of manipulated model explanations which provide arguments for why such research becomes crucial to achieve responsibility in machine learning use. Rieger and Hansen (2019) present a defence strategy against the attack via data change of Dombrowski et al (2019). The main idea is to aggregate various model explanations, which produces robust results without changing the model.…”
Section: Related Workmentioning
confidence: 99%
“…In this work we address the quantification of the surrogate explanation uncertainty by aggregating multiple surrogate coefficients. The use of a consensus mechanism to obtain explanations that are less sensitive to sampling variance (further discussed in Section 3) has been proposed in [5,24]. Specifically, [6] and [4] consider aggregating surrogate coefficients in the form of simple ranking schemes inspired from the social sciences and economics.…”
Section: Related Workmentioning
confidence: 99%