2020
DOI: 10.48550/arxiv.2010.00711
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Survey of the State of Explainable AI for Natural Language Processing

Abstract: Recent years have seen important advances in the quality of state-of-the-art models, but this has come at the expense of models becoming less interpretable. This survey presents an overview of the current state of Explainable AI (XAI), considered within the domain of Natural Language Processing (NLP). We discuss the main categorization of explanations, as well as the various ways explanations can be arrived at and visualized. We detail the operations and explainability techniques currently available for genera… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
44
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 36 publications
(44 citation statements)
references
References 66 publications
0
44
0
Order By: Relevance
“…Explainability is important since we want to be able to trust that our agents are beneficial before deploying them. For a recent survey of explainable natural language processing (NLP), see Danilevsky et al (2020). Note that explainability doesn't come for free -there still needs to be incentives for language agents to give true and useful explanations of their behaviour.…”
Section: Language Agentsmentioning
confidence: 99%

Alignment of Language Agents

Kenton,
Everitt,
Weidinger
et al. 2021
Preprint
“…Explainability is important since we want to be able to trust that our agents are beneficial before deploying them. For a recent survey of explainable natural language processing (NLP), see Danilevsky et al (2020). Note that explainability doesn't come for free -there still needs to be incentives for language agents to give true and useful explanations of their behaviour.…”
Section: Language Agentsmentioning
confidence: 99%

Alignment of Language Agents

Kenton,
Everitt,
Weidinger
et al. 2021
Preprint
“…A local explanation justifies a model or method's output for a specific input. A global explanation provides a justification on average performance of a model, independently of any particular input [20]. There are techniques for explainability using visualization of neuronal activity in layers of a deep architecture, for examples, based on sparsity and heatmaps [21], [22], [23].…”
Section: A Relevant Literaturementioning
confidence: 99%
“…We refer to the forward mapping as the transformation from the data space to the latent space, and the reverse mapping as the transformation from the latent space to the data space. The log-determinant is usually computed for the forward mapping at each of the constituent layers shown in (20). Each of the individual layers of a single flow-step are described as follows: a) Activation normalization: It is assumed that an input vector is represented as x and is of the shape…”
Section: Generative 1 × 1 Convolution Flow (Glow)mentioning
confidence: 99%
See 1 more Smart Citation
“…In this context, a flourishing literature proposing interpretability methods emerged. We refer to the survey papers of Guidotti et al (2018) and Adadi and Berrada (2018) for an overview, and to Danilevsky et al (2020) for a focus on natural language processing. With the notable exception of SHAP (Lundberg and Lee, 2017), these methods do not come with any guarantees.…”
Section: Introductionmentioning
confidence: 99%