Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.807
|View full text |Cite
|
Sign up to set email alerts
|

Rationales for Sequential Predictions

Abstract: Sequence models are a critical component of modern NLP systems, but their predictions are difficult to explain. We consider model explanations though rationales, subsets of context that can explain individual model predictions. We find sequential rationales by solving a combinatorial optimization: the best rationale is the smallest subset of input tokens that would predict the same output as the full sequence. Enumerating all subsets is intractable, so we propose an efficient greedy algorithm to approximate th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
26
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 13 publications
(26 citation statements)
references
References 23 publications
0
26
0
Order By: Relevance
“…For this example, the transformer used for CAREER follows the architecture described in Radford et al (2018). We find the rationale using the greedy rationalization method described in Vafa et al (2021). Greedy rationalization requires fine-tuning the model for compatibility; we do this by fine-tuning with "job dropout", where with 50% probability, we drop out a uniformly random amount of observations in the history.…”
Section: E Experimental Detailsmentioning
confidence: 99%
See 1 more Smart Citation
“…For this example, the transformer used for CAREER follows the architecture described in Radford et al (2018). We find the rationale using the greedy rationalization method described in Vafa et al (2021). Greedy rationalization requires fine-tuning the model for compatibility; we do this by fine-tuning with "job dropout", where with 50% probability, we drop out a uniformly random amount of observations in the history.…”
Section: E Experimental Detailsmentioning
confidence: 99%
“…To understand CAREER's prediction, we show the model's rationale, or the jobs in this individual's history that are sufficient for explaining the model's prediction. (We adapt the greedy rationalization method fromVafa et al (2021); refer to Appendix E for more details.) In this example, CAREER only needs three previous jobs to predict biological technician: animal caretaker, engineering technician, and student.…”
mentioning
confidence: 99%
“…We focus on extractive rationalization (Lei et al, 2016), which generates a subset of inputs or highlights as "rationales" such that the model can condition predictions on them. Recent development has been focusing on improving joint training of rationalizer and predictor components (Bastings et al, 2019;Yu et al, 2019;Paranjape et al, 2020;Guerreiro and Martins, 2021;Sha et al, 2021), or extensions to text matching (Swanson et al, 2020) and sequence generation (Vafa et al, 2021). These rationale models are mainly compared based on predictive performance, as well as agreement with human annotations (DeYoung et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Intuitively, our approach trains two classifiers: an explainability classifier (EC), which labels words in the textual context where the relation is expressed as important or not for the relation to be extracted, and a relation classifier (RC), which predicts the relation that holds between two given entities using only the words deemed as important. As such, our approach is self-explanatory because of inter-dependency between RC and EC, and generates faithful explanations that correctly depict how the relation classifier makes a decision (Vafa et al 2021).…”
Section: Introductionmentioning
confidence: 99%
“…In this situation, we measure the overlap between the words identified by the EC as important and the words used by rules using standard precision, recall, and F1 scores. The second strategy relies on plausability, i.e., can the machine explanations be understood and interpreted by humans (Wiegreffe and Pinter 2019a;Vafa et al 2021)? To this end, we compare the tokens identified by the EC against human annotations of the context words marked as important for the relation.…”
Section: Introductionmentioning
confidence: 99%