2020
DOI: 10.1007/978-3-030-38227-8_6
|View full text |Cite
|
Sign up to set email alerts
|

Generating Neural Archetypes to Instruct Fast and Interpretable Decisions

Abstract: Deep learning methods are highly accurate, yet their opaque decision process prevents them from earning full human trust. Concept-based models aim to address this issue by learning tasks based on a set of human-understandable concepts. However, state-of-the-art concept-based models rely on high-dimensional concept embedding representations which lack a clear semantic meaning, thus questioning the interpretability of their decision process. To overcome this limitation, we propose the Deep Concept Reasoner (DCR)… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 50 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?