2021
DOI: 10.1109/tpami.2020.2982882
|View full text |Cite
|
Sign up to set email alerts
|

Interpretable CNNs for Object Classification

Abstract: This paper proposes a generic method to learn interpretable convolutional filters in a deep convolutional neural network (CNN) for object classification, where each interpretable filter encodes features of a specific object part. Our method does not require additional annotations of object parts or textures for supervision. Instead, we use the same training data as traditional CNNs. Our method automatically assigns each interpretable filter in a high conv-layer with an object part of a certain category during … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
19
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2
2

Relationship

1
9

Authors

Journals

citations
Cited by 46 publications
(19 citation statements)
references
References 44 publications
0
19
0
Order By: Relevance
“…(B) Whereas in the second scenario (B), CX-ToM thinks that users do not trust the model's ability in correctly classifying Person, and therefore shows a faultline explanation using categories Woman and Deer. Biran and Cotton, 2017;Darlington, 2013;Kim, 2017a, 2017b;Goodman and Flaxman, 2017;Hoffman, 2017;Keil, 2006;Kulesza et al, 2010Kulesza et al, , 2011Moore and Swartout, 1990;Walton, 2004;Douglas, 2007;Walton, 2011;Sheh, 2017;Sheh and Monteath, 2018;Tapaswi et al, 2016;Williams et al, 2016;Agarwal et al, 2018;Akula et al, 2018Akula et al, , 2019aAkula et al, , 2019bAkula et al, , 2019cAkula et al, , 2021dAkula and Zhu, 2019;Gupta et al, 2016;Bivens et al, 2017;Zhang et al, 2019aZhang et al, , 2020aZhang et al, , 2020b. Most prior work in explaining CNN's predictions has focused on generating explanations using feature visualization and attribution.…”
Section: Related Workmentioning
confidence: 99%
“…(B) Whereas in the second scenario (B), CX-ToM thinks that users do not trust the model's ability in correctly classifying Person, and therefore shows a faultline explanation using categories Woman and Deer. Biran and Cotton, 2017;Darlington, 2013;Kim, 2017a, 2017b;Goodman and Flaxman, 2017;Hoffman, 2017;Keil, 2006;Kulesza et al, 2010Kulesza et al, , 2011Moore and Swartout, 1990;Walton, 2004;Douglas, 2007;Walton, 2011;Sheh, 2017;Sheh and Monteath, 2018;Tapaswi et al, 2016;Williams et al, 2016;Agarwal et al, 2018;Akula et al, 2018Akula et al, , 2019aAkula et al, , 2019bAkula et al, , 2019cAkula et al, , 2021dAkula and Zhu, 2019;Gupta et al, 2016;Bivens et al, 2017;Zhang et al, 2019aZhang et al, , 2020aZhang et al, , 2020b. Most prior work in explaining CNN's predictions has focused on generating explanations using feature visualization and attribution.…”
Section: Related Workmentioning
confidence: 99%
“…Throughout the past decades, the notion of interpretability increasingly gained attention by the Machine Learning (ML) community [ 10 , 11 , 12 , 13 , 14 , 15 , 16 , 17 ]. According to Kim et al [ 13 ], interpretability is particularly important for systems whose decisions have a significant impact such as in healthcare, criminal justice and finance.…”
Section: Related Workmentioning
confidence: 99%
“…We now direct our attention to local explanatory methods. One relevant class of locally explainable models is convolutional networks augmented with attention [ 27 ]. In such a case, the convolutional networks are trained to limit their vision to specific parts of the image, using network architecture which purposefully removes irrelevant information.…”
Section: Related Workmentioning
confidence: 99%