2023
DOI: 10.48550/arxiv.2301.09611
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Explaining Deep Learning Hidden Neuron Activations using Concept Induction

Abstract: One of the current key challenges in Explainable AI is in correctly interpreting activations of hidden neurons. It seems evident that accurate interpretations thereof would provide insights into the question what a deep learning system has internally detected as relevant on the input, thus lifting some of the black box character of deep learning systems.The state of the art on this front indicates that hidden node activations appear to be interpretable in a way that makes sense to humans, at least in some case… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 31 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?