Proceedings of the Third BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP 2020
DOI: 10.18653/v1/2020.blackboxnlp-1.10
|View full text |Cite
|
Sign up to set email alerts
|

The Explanation Game: Towards Prediction Explainability through Sparse Communication

Abstract: Explainability is a topic of growing importance in NLP. In this work, we provide a unified perspective of explainability as a communication problem between an explainer and a layperson about a classifier's decision. We use this framework to compare several explainers, including gradient methods, erasure, and attention mechanisms, in terms of their communication success. In addition, we reinterpret these methods in the light of classical feature selection, and use this as inspiration for new embedded explainers… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
37
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 23 publications
(37 citation statements)
references
References 46 publications
0
37
0
Order By: Relevance
“…For all rationalizers, we map each input word to 300D-pretrained GloVe embeddings from 840B release (Pennington et al, 2014) that are kept frozen. We instantiate all encoder networks as bidirectional LSTM (Hochreiter and Schmidhuber, 1997) layers (BiLSTM) (w/ hidden size 200) similarly to Lei et al (2016); Bastings et al (2019); Treviso and Martins (2020). Although other works (Jain et al, 2020;Paranjape et al, 2020) use more powerful BERT-based representations, we firstly experimented with BiLSTM layers and noticed our results were competitive with those reported in Jain et al (2020).…”
Section: C1 Rationalizers Experimental Setupmentioning
confidence: 99%
See 3 more Smart Citations
“…For all rationalizers, we map each input word to 300D-pretrained GloVe embeddings from 840B release (Pennington et al, 2014) that are kept frozen. We instantiate all encoder networks as bidirectional LSTM (Hochreiter and Schmidhuber, 1997) layers (BiLSTM) (w/ hidden size 200) similarly to Lei et al (2016); Bastings et al (2019); Treviso and Martins (2020). Although other works (Jain et al, 2020;Paranjape et al, 2020) use more powerful BERT-based representations, we firstly experimented with BiLSTM layers and noticed our results were competitive with those reported in Jain et al (2020).…”
Section: C1 Rationalizers Experimental Setupmentioning
confidence: 99%
“…These approaches are often brittle and fragile for the high sensitivity that they show to changes in the hyperparameters and to variability due to sampling. On the other hand, existing rationalizers that use sparse attention mechanisms (Treviso and Martins, 2020) such as sparsemax attention, while being deterministic and end-to-end differentiable, do not have a direct handle to constrain the rationale in terms of sparsity and contiguity. We endow them with these capabilities in this paper as shown in Table 1, where we position our work in the literature for highlights extraction.…”
Section: Rationalization For Highlights Extractionmentioning
confidence: 99%
See 2 more Smart Citations
“…Chen and Ji (2020) propose learning a variational word mask to improve model interpretability. Finally, extracting a short snippet from the original input text (rationale) and using it to make a prediction has been recently proposed (Lei et al, 2016;Bastings et al, 2019;Treviso and Martins, 2020;Jain et al, 2020;Chalkidis et al, 2021). Nguyen (2018) and Atanasova et al (2020) compare explanations produced by different approaches, showing that in most cases gradientbased approaches outperform sparse linear metamodels.…”
Section: Model Interpretabilitymentioning
confidence: 99%