Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.269
|View full text |Cite
|
Sign up to set email alerts
|

Is Attention Explanation? An Introduction to the Debate

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 31 publications
(23 citation statements)
references
References 0 publications
0
15
0
Order By: Relevance
“…Finally, our work may contribute to the "attention as explanation" debate (Jain and Wallace, 2019;Serrano and Smith, 2019;Wiegreffe and Pinter, 2019;Bibal et al, 2022). By showing that some PLMs can perform reasonably well with constant matrices, we suggest that explanations arising from the attention matrices might not be crucial for models' success.…”
Section: Introductionmentioning
confidence: 79%
“…Finally, our work may contribute to the "attention as explanation" debate (Jain and Wallace, 2019;Serrano and Smith, 2019;Wiegreffe and Pinter, 2019;Bibal et al, 2022). By showing that some PLMs can perform reasonably well with constant matrices, we suggest that explanations arising from the attention matrices might not be crucial for models' success.…”
Section: Introductionmentioning
confidence: 79%
“…For example, in the GRN inference tasks defined by scGPT, we cannot treat a coexpression network based on gene embeddings the same as a gene regulatory network. Also from scGPT, using attention we can infer gene-gene correlation strength with direction, but the relation between the correlation of features and the value in the attention map is in debate [78][79][80]. Moreover, we should consider more meaningful and representative tasks related to single-cell and spatial data for single-cell LLMs.…”
Section: Discussionmentioning
confidence: 99%
“…Despite the debate about attention being a reliable intelligibility method (Jain and Wallace, 2019;Wiegreffe and Pinter, 2019;Serrano and Smith, 2019;Bibal, Cardon, Alfter, Wilkens, Wang, François and Watrin, 2022), it remains a popular method in existing deep neural network approaches in fact-checking. Attentionbased explanations are provided in various forms:…”
Section: Attention-based Intelligibilitymentioning
confidence: 99%