Findings of the Association for Computational Linguistics: ACL 2022 2022
DOI: 10.18653/v1/2022.findings-acl.49
|View full text |Cite
|
Sign up to set email alerts
|

Domain Generalisation of NMT: Fusing Adapters with Leave-One-Domain-Out Training

Abstract: Generalising to unseen domains is underexplored and remains a challenge in neural machine translation. Inspired by recent research in parameter-efficient transfer learning from pretrained models, this paper proposes a fusionbased generalisation method that learns to combine domain-specific parameters. We propose a leave-one-domain-out training strategy to avoid information leaking to address the challenge of not knowing the test domain during training time. Empirical results on three language pairs show that o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 17 publications
0
1
0
Order By: Relevance
“…It learns the attention weights of T 1 , ...T n while being fine-tuned on the data for T n+1 . Vu et al (2022) adapt this technique for fusing domains and testing on out-of-domain data. This technique has not been applied in the context of LAs so far.…”
Section: Related Workmentioning
confidence: 99%
“…It learns the attention weights of T 1 , ...T n while being fine-tuned on the data for T n+1 . Vu et al (2022) adapt this technique for fusing domains and testing on out-of-domain data. This technique has not been applied in the context of LAs so far.…”
Section: Related Workmentioning
confidence: 99%
“…Adapters, as a specific version of a MoE, were lately also used for the task of domain adaptation. The work of Vu et al [42] focuses on the domain generalisation task via Adapter leave one-out strategy. In the similar, regularisation focused way, (and additionally improving overall complexity), Rücklé et al [36] proposed AdapterDrop technique to drop out Adapter layers, similarly to removing Transformer layers [11].…”
Section: Related Workmentioning
confidence: 99%
“…One option proposed by Vu et al (2022) explicitly accounts for the future presence of new, unknown domains when tuning adapter layers. They add a 'fusion' adapter layer whose purpose is to generalise from other, domain-specific layers.…”
Section: Architecture-centric Multi-domain Adaptationmentioning
confidence: 99%