2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.01469
|View full text |Cite
|
Sign up to set email alerts
|

Neural Prototype Trees for Interpretable Fine-grained Image Recognition

Abstract: Figure 1: A ProtoTree is a globally interpretable model faithfully explaining its entire reasoning (left, partially shown).Additionally, the decision making process for a single prediction can be followed (right): the presence of a red chest and black wing, and the absence of a black stripe near the eye, identifies a Scarlet Tanager. A pruned ProtoTree learns roughly 200 prototypes for CUB (dataset with 200 bird species), making only 8 local decisions on average for one test image.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
93
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

2
7

Authors

Journals

citations
Cited by 166 publications
(115 citation statements)
references
References 50 publications
(59 reference statements)
1
93
0
Order By: Relevance
“…Representative examples, including concepts [127], influential training instances [90], prototypical parts [36,179], nearest neighbors and criticisms [125].…”
Section: Prototypes (Parts Of)mentioning
confidence: 99%
“…Representative examples, including concepts [127], influential training instances [90], prototypical parts [36,179], nearest neighbors and criticisms [125].…”
Section: Prototypes (Parts Of)mentioning
confidence: 99%
“…The post-hoc methods are Grad-CAM [23], Grad-CAM++ [4], RISE [22], Score-CAM [31], Ablation CAM [7]. The architectures with native attention are B-CNN [13], BR-NPA [10], the model from [14] which we call IBP (short for Interpretability By Parts), ProtoPNet [5], and ProtoTree [20]. These attention models generate several saliency maps (or attention maps) per input image but the metrics are designed for a single saliency map per image.…”
Section: Benchmarkmentioning
confidence: 99%
“…Building inherently interpretable models, beyond post hoc approaches, is our key challenge here [34]. There have been several recent efforts [6,18,28,46,54], but most of them concentrate on enhancing interpretability only in the last layers of the neural network. In [46], the final linear layer is replaced with a differentiable decision tree, and in [54] a loss is used to make each filter of the very high-level convolutional layer represent a specific object part.…”
Section: Related Workmentioning
confidence: 99%