Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.408
|View full text |Cite
|
Sign up to set email alerts
|

ERASER: A Benchmark to Evaluate Rationalized NLP Models

Abstract: State-of-the-art models in NLP are now predominantly based on deep neural networks that are opaque in terms of how they come to make predictions. This limitation has increased interest in designing more interpretable deep models for NLP that reveal the 'reasoning' behind model outputs. But work in this direction has been conducted on different datasets and tasks with correspondingly unique aims and metrics; this makes it difficult to track progress. We propose the Evaluating Rationales And Simple English Reaso… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
421
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 319 publications
(425 citation statements)
references
References 65 publications
3
421
1
Order By: Relevance
“…Relation to Automatic Tests. Prior works have proposed automatic metrics for feature importance estimates (Nguyen, 2018;Hooker et al, 2019;DeYoung et al, 2020). Typically these operate by checking that model behavior follows reasonable patterns on counterfactual inputs constructed using the explanation, e.g., by masking "important" features and checking that a class score drops.…”
Section: Evaluating Interpretabilitymentioning
confidence: 99%
“…Relation to Automatic Tests. Prior works have proposed automatic metrics for feature importance estimates (Nguyen, 2018;Hooker et al, 2019;DeYoung et al, 2020). Typically these operate by checking that model behavior follows reasonable patterns on counterfactual inputs constructed using the explanation, e.g., by masking "important" features and checking that a class score drops.…”
Section: Evaluating Interpretabilitymentioning
confidence: 99%
“…We aim at faithful explanations -the identification of the actual reason for the model's prediction, which is essential for accountability, fairness, and credibility (Chakraborty et al, 2017;Wu and Mooney, 2019) to evaluate whether a model's prediction is based on the correct evidence. The recently published ERASER benchmark (DeYoung et al, 2020) provides multiple datasets with annotated rationales, i.e., parts of the input document, which are essential for correct predictions of the target variable (Zaidan et al, 2007). By contrast to post-hoc techniques to identify relevant input parts such as LIME (Ribeiro et al, 2016) or input reduction (Feng et al, 2018), we focus on models that are faithful by design, in which the selected rationale matches the full underlying evidence used for the prediction.…”
Section: Selectmentioning
confidence: 99%
“…Existing strategies mostly rely on REINFORCE (Williams, 1992) style learning (Lei et al, 2016; or on training two disjoint models (Lehman et al, 2019;DeYoung et al, 2020), in the latter case depending on rationale supervision. This poses critical limitations as rationale annotations are costly to obtain and, in many cases, not available.…”
Section: Selectmentioning
confidence: 99%
See 1 more Smart Citation
“…How should these different approaches be compared? Several diagnostic tests have been proposed: Jain and Wallace (2019) assessed the explanatory power of attention weights by measuring their correlation with input gradients; Wiegreffe and Pinter (2019) and DeYoung et al (2020) developed more informative tests, including a combination of comprehensiveness and sufficiency metrics and the correlation with human rationales; Jacovi and Goldberg (2020) proposed a set of evaluation recommendations and a graded notion of faithfulness. Most proposed frameworks rely on correlations and counterfactual simulation, sidestepping the main practical goal of prediction explainability-the ability to communicate an explanation to a human user.…”
Section: Introductionmentioning
confidence: 99%