Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2022
DOI: 10.18653/v1/2022.naacl-main.235
|View full text |Cite
|
Sign up to set email alerts
|

Neural Language Taskonomy: Which NLP Tasks are the most Predictive of fMRI Brain Activity?

Abstract: Several popular Transformer based language models have been found to be successful for text-driven brain encoding. However, existing literature leverages only pretrained text Transformer models and has not explored the efficacy of task-specific learned Transformer representations. In this work, we explore transfer learning from representations learned for ten popular natural language processing tasks (two syntactic and eight semantic) for predicting brain responses from two diverse datasets: Pereira (subjects … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
17
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(17 citation statements)
references
References 21 publications
0
17
0
Order By: Relevance
“…A number of independent studies have recently shown that representations from state-of-the-art ANN models—especially unidirectional transformer models—align well with brain responses of humans processing linguistic input (e.g., Caucheteux & King, 2022 ; Gauthier & Levy, 2019 ; Goldstein et al, 2022 ; Jain & Huth, 2018 ; Kumar et al, 2022 ; Merlin & Toneva, 2022 ; Millet et al, 2022 ; Oota et al, 2022 ; Pasquiou et al, 2022 ; Pereira et al, 2018 ; Schrimpf et al, 2021 ; Toneva & Wehbe, 2019 ). However, what makes ANN representations align with human neural responses to language has been little explored (cf.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…A number of independent studies have recently shown that representations from state-of-the-art ANN models—especially unidirectional transformer models—align well with brain responses of humans processing linguistic input (e.g., Caucheteux & King, 2022 ; Gauthier & Levy, 2019 ; Goldstein et al, 2022 ; Jain & Huth, 2018 ; Kumar et al, 2022 ; Merlin & Toneva, 2022 ; Millet et al, 2022 ; Oota et al, 2022 ; Pasquiou et al, 2022 ; Pereira et al, 2018 ; Schrimpf et al, 2021 ; Toneva & Wehbe, 2019 ). However, what makes ANN representations align with human neural responses to language has been little explored (cf.…”
Section: Discussionmentioning
confidence: 99%
“…In particular, contemporary ANN language models achieve impressive performance on a variety of linguistic tasks (e.g., Brown et al, 2020 ; Chowdhery et al, 2022 ; Devlin et al, 2018 ; Liu et al, 2019 ; OpenAI, 2023 ; Rae et al, 2021 ). Furthermore, representations extracted from ANN language models—especially unidirectional attention transformer architectures like GPT2 ( Radford et al, 2019 ) can explain substantial variance in brain activity recorded from the human language network using regression-based evaluation metrics (e.g., Caucheteux & King, 2022 ; Gauthier & Levy, 2019 ; Goldstein et al, 2022 ; Hosseini et al, 2022 ; Jain & Huth, 2018 ; Kumar et al, 2022 ; Oota et al, 2022 ; Pasquiou et al, 2022 ; Schrimpf et al, 2021 ; Toneva & Wehbe, 2019 ). This correspondence has been suggested to derive, at least in part, from the convergence of the ANNs’ linguistic representations with those in the human brain ( Caucheteux & King, 2022 ; Goldstein et al, 2022 ; Hosseini et al, 2022 ; Schrimpf et al, 2021 ), despite the vast differences in their learning and architecture (e.g., Huebner & Willits, 2021 ; Warstadt & Bowman, 2022 ).…”
Section: Introductionmentioning
confidence: 99%
“…By specifically modeling representations of linguistic features, our results suggest that some of the differences observed in Regev et al (2013) could indeed be due to non-linguistic processes such as high-level control. A separate study suggesting that brain representations of language differ between modalities compared brain responses to different types of stimuli for reading and listening: the stimuli used for reading experiments consisted of isolated sentences, whereas the stimuli used for listening experiments consisted of full narratives (Oota et al, 2022). This discrepancy perhaps explains why in Oota et al (2022), language models trained on higher-level tasks (e.g., summarization, paraphrase detection) were better able to predict listening than reading data.…”
Section: Discussionmentioning
confidence: 99%
“…A separate study suggesting that brain representations of language differ between modalities compared brain responses to different types of stimuli for reading and listening: the stimuli used for reading experiments consisted of isolated sentences, whereas the stimuli used for listening experiments consisted of full narratives (Oota et al, 2022). This discrepancy perhaps explains why in Oota et al (2022), language models trained on higher-level tasks (e.g., summarization, paraphrase detection) were better able to predict listening than reading data. Our study used matched stimuli for reading and listening experiments, and the similarities we observed highlight the importance of using narrativelength, naturalistic stimuli to elicit brain representations of high-level linguistic features (Deniz et al, 2023).…”
Section: Discussionmentioning
confidence: 99%
“…We select tasks for this work based on two major principles that have also been adopted by previous work [21]. First, tasks that require a diverse set of cognitive-linguistic skills are chosen.…”
Section: Tasks For Tuningmentioning
confidence: 99%