2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP) 2019
DOI: 10.1109/globalsip45357.2019.8969491
|View full text |Cite
|
Sign up to set email alerts
|

Generative Counterfactual Introspection for Explainable Deep Learning

Abstract: In this work, we propose an introspection technique for deep neural networks that relies on a generative model to instigate salient editing of the input image for model interpretation. Such modification provides the fundamental interventional operation that allows us to obtain answers to counterfactual inquiries, i.e., what meaningful change can be made to the input image in order to alter the prediction. We demonstrate how to reveal interesting properties of the given classifiers by utilizing the proposed int… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
40
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 63 publications
(40 citation statements)
references
References 23 publications
0
40
0
Order By: Relevance
“…This generative framework has been adopted by several recent studies [366,367] mainly as an attribution method to relate a particular output of a Deep Learning model to their input variables. Another interesting research direction is the use of generative models for the creation of counterfactuals, i.e., modifications to the input data that could eventually alter the original prediction of the model [368]. Counterfactual prototypes help the user understand the performance boundaries of the model under consideration for his/her improved trust and informed criticism.…”
Section: Explanations For Ai Security: Xai and Adversarial Machine Lementioning
confidence: 99%
“…This generative framework has been adopted by several recent studies [366,367] mainly as an attribution method to relate a particular output of a Deep Learning model to their input variables. Another interesting research direction is the use of generative models for the creation of counterfactuals, i.e., modifications to the input data that could eventually alter the original prediction of the model [368]. Counterfactual prototypes help the user understand the performance boundaries of the model under consideration for his/her improved trust and informed criticism.…”
Section: Explanations For Ai Security: Xai and Adversarial Machine Lementioning
confidence: 99%
“…Indeed, contfactuals are particularly suitable for informing the end-user why a given data example is assigned a particular class label. Thus, the outlined classification-oriented frameworks are evaluated on classifiers based on logistic regression [55], [136], [153], [158], decision trees [46], [80], [122], [140], [150], [155], [159], gradient boosted decision trees [147], support vector machines [131], [138], [146], random forests [81], [86], [142]- [144], neural networks [6], [48], [49], [91], [129], [130], [133], [135], [139], [141], [145], [148], [151], or combinations of these [100], [105], [134], [152], [154], [160]. In three studies [67], [128], [137], the classifiers used in the experiments are not specified.…”
Section: ) Ai Problemmentioning
confidence: 99%
“…For instance, Gomez et al visualize the generated explanations in the form of bar plots combining them with explicitly stated numerical values [138]. Alternatively, Liu et al combine feature importance bar plots with visual input and output [145]. While a textual explanation summarizes the degree of importance of the selected features, a visual explanation may present contextual in-method metrics that justify the classifier's reasoning [129].…”
Section: ) Output Representationmentioning
confidence: 99%
“…Therefore, we propose to predict and ground attributes for both clean and perturbed images to provide visual as well as attribute-based interpretations. Counterfactual explanations Explanations which consider counterdecisions or counteroutcomes are known as counterfactual explanations [25]. An interesting approach in a recent paper [13] proposes to generate counterfactual expla-Fig.…”
Section: Explainabilitymentioning
confidence: 99%