Findings of the Association for Computational Linguistics: NAACL 2022 2022
DOI: 10.18653/v1/2022.findings-naacl.64
|View full text |Cite
|
Sign up to set email alerts
|

Attention Fusion: a light yet efficient late fusion mechanism for task adaptation in NLU

Abstract: Fine-tuning a pre-trained language model using annotated data has become the de-facto standard for adapting general-purpose pretrained models like BERT to downstream tasks. However, given the trend of larger pretrained models, fine-tuning these models for each downstream task is parameter-inefficient and computationally-expensive deeming this approach sub-optimal for adoption by NLU systems. In recent years, various approaches have been proposed for parameter efficient task adaptation such as Adaptor, Bitfit, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 14 publications
0
4
0
Order By: Relevance
“…sentiment analysis (Horne et al, 2020;Xiao et al, 2021), morphosyntactic parsing (Kondratyuk and Straka, 2019), gender debiasing and coreference resolution (Abzaliev, 2019), and cross-lingual transfer learning (Chen et al, 2022)). Building on this approach, Cao et al (2022) expand the task-oriented layer aggregation to encompass both token and task aspects. This extension is achieved through the introduction of an attention fusion model that leverages the local features of a token.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…sentiment analysis (Horne et al, 2020;Xiao et al, 2021), morphosyntactic parsing (Kondratyuk and Straka, 2019), gender debiasing and coreference resolution (Abzaliev, 2019), and cross-lingual transfer learning (Chen et al, 2022)). Building on this approach, Cao et al (2022) expand the task-oriented layer aggregation to encompass both token and task aspects. This extension is achieved through the introduction of an attention fusion model that leverages the local features of a token.…”
Section: Related Workmentioning
confidence: 99%
“…This extension is achieved through the introduction of an attention fusion model that leverages the local features of a token. We extend Cao et al (2022)'s method by introducing an attention mechanism that incorporate global views of intermediate representations.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations