Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 2021
DOI: 10.18653/v1/2021.findings-acl.201
|View full text |Cite
|
Sign up to set email alerts
|

On the Interplay Between Fine-tuning and Composition in Transformers

Abstract: Pre-trained transformer language models have shown remarkable performance on a variety of NLP tasks. However, recent research has suggested that phrase-level representations in these models reflect heavy influences of lexical content, but lack evidence of sophisticated, compositional phrase information (Yu and Ettinger, 2020). Here we investigate the impact of fine-tuning on the capacity of contextualized embeddings to capture phrase meaning information beyond lexical content. Specifically, we fine-tune models… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 45 publications
0
6
0
Order By: Relevance
“…It's important to note that most existing evaluation methods and datasets for systematic compositionality are only suitable for generative-based language models like SCAN (Higgins et al 2018b) and incontext prompting . Instead, our evaluations follow recent works (Yu and Ettinger 2021;Hendrycks et al 2020;Yu and Ettinger 2020), which are designed for discriminative-based language models like BERT and ours. Since each evaluation task differs, we provide the details in the corresponding subsections.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…It's important to note that most existing evaluation methods and datasets for systematic compositionality are only suitable for generative-based language models like SCAN (Higgins et al 2018b) and incontext prompting . Instead, our evaluations follow recent works (Yu and Ettinger 2021;Hendrycks et al 2020;Yu and Ettinger 2020), which are designed for discriminative-based language models like BERT and ours. Since each evaluation task differs, we provide the details in the corresponding subsections.…”
Section: Methodsmentioning
confidence: 99%
“…As suggested in previous studies (Yu and Ettinger 2021;Cartuyvels, Spinks, and Moens 2021;Hendrycks et al 2020;Yu and Ettinger 2020), our evaluation closely revolves around the characteristics of the systematic compositionality to provide detailed insights. Due to the issue of computational resources, we restrict our analysis and comparison to the widely used BERT and tentatively implement and pretrain a multi-layer CAT from scratch following BERT.…”
Section: Introductionmentioning
confidence: 93%
See 2 more Smart Citations
“…Following the research line in understanding the reasons behind the outstanding performance of pre-trained language models and their capabilities, most recent investigations on fine-tuning have been done through probing tasks and by evaluating the encoded linguistic knowledge (Merchant et al, 2020;Mosbach et al, 2020;Talmor et al, 2020;Yu and Ettinger, 2021). These studies demonstrate that most changes in fine-tuning are applied to the upper layers, such that those layers encode taskspecific knowledge, while lower layers are responsible for the core linguistic phenomenon (Durrani et al, 2021).…”
Section: Related Workmentioning
confidence: 99%