Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-short.19
|View full text |Cite
|
Sign up to set email alerts
|

Relative Importance in Sentence Processing

Abstract: Determining the relative importance of the elements in a sentence is a key factor for effortless natural language understanding. For human language processing, we can approximate patterns of relative importance by measuring reading fixations using eye-tracking technology. In neural language models, gradientbased saliency methods indicate the relative importance of a token for the target objective. In this work, we compare patterns of relative importance in English language processing by humans and models and a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2
1

Relationship

3
7

Authors

Journals

citations
Cited by 15 publications
(8 citation statements)
references
References 43 publications
0
8
0
Order By: Relevance
“…This new eye tracking dataset also allows us to analyze and interpret language models or task-specific NLP models. For example, we can investigate machine-learning based explainability mechanisms such as attention and saliency in Danish language models, as suggested by Hollenstein and Beinborn (2021) or Sood et al (2020) for English.…”
Section: Discussionmentioning
confidence: 99%
“…This new eye tracking dataset also allows us to analyze and interpret language models or task-specific NLP models. For example, we can investigate machine-learning based explainability mechanisms such as attention and saliency in Danish language models, as suggested by Hollenstein and Beinborn (2021) or Sood et al (2020) for English.…”
Section: Discussionmentioning
confidence: 99%
“…In addition, we compute averages over last-layer attention head vectors ('Attention-last') as proposed in Hollenstein and Beinborn [2021] as well as 'Rollout' and attention flow ('A-flow') Abnar and Zuidema [2020] which capture the layerwise structure of deep Transformer models in comparison to raw attention head analysis.…”
Section: Benchmark Methodsmentioning
confidence: 99%
“…For the present benchmark, this is possible by leveraging a naturalistic dataset of reading English sentences, the Zurich Cognitive Language Processing Corpus (Hollenstein et al, 2018 , 2020 ). The ZuCo dataset is publicly available and has recently been used in a variety of applications including leveraging EEG and eye-tracking data to improve NLP tasks (Barrett et al, 2018 ; Mathias et al, 2020 ; McGuire and Tomuro, 2021 ), evaluating the cognitive plausibility of computational language models (Hollenstein et al, 2019b ; Hollenstein and Beinborn, 2021 ), investigating the neural dynamics of reading (Pfeiffer et al, 2020 ), developing models of human reading (Bautista and Naval, 2020 ; Bestgen, 2021 ).…”
Section: Introductionmentioning
confidence: 99%