2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00886
|View full text |Cite
|
Sign up to set email alerts
|

There and Back Again: Revisiting Backpropagation Saliency Methods

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
49
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
1
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 92 publications
(50 citation statements)
references
References 18 publications
1
49
0
Order By: Relevance
“…In genomics, where motifs can have different effects depending on whether they occur at the center vs the flanks, and also co-bind with other motifs, the assumption of positional and contextual independence may not be valid. In fact, experiments in computer vision have reported that this aspect of GradCAM leads it to fail “anywhere besides the last convolutional layer” [27], and the GradCAM paper itself states that “localization becomes progressively worse as we move to shallower convolutional layers. This is because the later convolutional layers capture high-level semantic information and at the same time retain spatial information, while the shallower layers have smaller receptive fields and only concentrate on local features that are important for the next layers” (section 5.2 of the supplement).…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…In genomics, where motifs can have different effects depending on whether they occur at the center vs the flanks, and also co-bind with other motifs, the assumption of positional and contextual independence may not be valid. In fact, experiments in computer vision have reported that this aspect of GradCAM leads it to fail “anywhere besides the last convolutional layer” [27], and the GradCAM paper itself states that “localization becomes progressively worse as we move to shallower convolutional layers. This is because the later convolutional layers capture high-level semantic information and at the same time retain spatial information, while the shallower layers have smaller receptive fields and only concentrate on local features that are important for the next layers” (section 5.2 of the supplement).…”
Section: Resultsmentioning
confidence: 99%
“…4 ). This suggests that a promising future direction may be to improve motif identification by combining importance scores computed at multiple different layers, similar to the approaches taken by [11], [27] and [33]. It is also worth investigating why importance scores can become worse at highlighting key features when computed closer to the input layer.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Although gradient-based methods might not be the optimal solution for visual explanation (e.g., saturation, zero-gradient image regions, and false confidence in the output score phenomena [156]), the computational cost of Grad-CAMs is negligible when compared to other methods that require multiple network forward-passes per image [156,157]. Moreover, Grad-CAM is considered the reference method in several recent works [157][158][159][160][161].…”
Section: Making the Model Interpretablementioning
confidence: 99%
“…For post-hoc interpretation, the prediction results are obtained by a black-box model while the interpretation is obtained separately to explain the predictive mechanism of the black box. Among the different post-hoc interpretation techniques, backpropagation methods (Zhou et al, 2016;Selvaraju et al, 2017;Wang et al, 2020a;Shrikumar et al, 2017;Rebuffi et al, 2020;Bach et al, 2015), trace from the output back to the input to determine how the different elements in the input contribute to the prediction result. Class Activation Mapping (CAM) (Zhou et al, 2016) visualizes the feature importance in convolutional neural networks by mapping the weights in the last fully connected layer to the input layer via upsampling.…”
Section: Related Workmentioning
confidence: 99%