2019 IEEE/CVF International Conference on Computer Vision (ICCV) 2019
DOI: 10.1109/iccv.2019.00304
|View full text |Cite
|
Sign up to set email alerts
|

Understanding Deep Networks via Extremal Perturbations and Smooth Masks

Abstract: The problem of attribution is concerned with identifying the parts of an input that are responsible for a model's output. An important family of attribution methods is based on measuring the effect of perturbations applied to the input. In this paper, we discuss some of the shortcomings of existing approaches to perturbation analysis and address them by introducing the concept of extremal perturbations, which are theoretically grounded and interpretable. We also introduce a number of technical innovations to c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
363
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 350 publications
(365 citation statements)
references
References 23 publications
2
363
0
Order By: Relevance
“…A present trend in the ML community is a migration to the PyTorch framework with its eager execution paradigm, away from other back ends. Both the TorchRay [46] and Captum [95] packages for Python and PyTorch enable the use of interpretability methods for neural network models defined in the context of PyTorch's high-level neural network description modules. Captum can be understood as a rich selection of XAI methods based on modified backprop and is part of the PyTorch project itself.…”
Section: Appendix C E X P L a N A T I O N S O F T W A R Ementioning
confidence: 99%
“…A present trend in the ML community is a migration to the PyTorch framework with its eager execution paradigm, away from other back ends. Both the TorchRay [46] and Captum [95] packages for Python and PyTorch enable the use of interpretability methods for neural network models defined in the context of PyTorch's high-level neural network description modules. Captum can be understood as a rich selection of XAI methods based on modified backprop and is part of the PyTorch project itself.…”
Section: Appendix C E X P L a N A T I O N S O F T W A R Ementioning
confidence: 99%
“…Specifically, using APARENT (Bogard et al, 2019) -a CNN capable of predicting cleavage and polyadenylation from primary sequence -we trained an Inclusion-Scrambler to reconstruct isoform predictions for the original APARENT training data (t bits = 0.25, λ = 1; Figure 3A). As we anticipated important polyadenylation features like RNA binding protein (RBP) recognition motifs to consist of short subsequences, we regularized the Scrambler by fixing the final layer of the network to a Gaussian filter (reminiscent of a technique proposed by Fong et al, 2019) to encourage the selection of contiguous nucleotides for masking. We found that the regularized Scrambler learned to recognize known regulatory binding factors associated with alternative polyadenylation (Supplementary Fig.…”
Section: Identifying Multiple Salient Feature Sets With Dropout and Bias Layersmentioning
confidence: 99%
“…In RISE [25], the importance of a pixel is computed as the expectation over all local perturbations conditioned on the event that the pixel is observed. More recently, the concept of "extreme perturbations" has been introduced to improve the perturbation analysis by the extremal algorithm [6].…”
Section: Related Workmentioning
confidence: 99%
“…We now present the experimental results. We tested DeepCover on a variety of DNN models for ImageNet and we compare DeepCover with the most popular and most recent work in AI explanation: lime [27], shap [18], gradcam [29], rise [25] and extremal [6]. 6…”
Section: Experimental Evaluationmentioning
confidence: 99%
See 1 more Smart Citation