Findings of the Association for Computational Linguistics: EMNLP 2021 2021
DOI: 10.18653/v1/2021.findings-emnlp.240
|View full text |Cite
|
Sign up to set email alerts
|

Counter-Interference Adapter for Multilingual Machine Translation

Abstract: Developing a unified multilingual model has long been a pursuit for machine translation. However, existing approaches suffer from performance degradation -a single multilingual model is inferior to separately trained bilingual ones on rich-resource languages. We conjecture that such a phenomenon is due to interference caused by joint training with multiple languages. To accommodate the issue, we propose CIAT, an adapted Transformer model with a small parameter overhead for multilingual machine translation. We … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
9
1

Relationship

0
10

Authors

Journals

citations
Cited by 21 publications
(12 citation statements)
references
References 29 publications
0
12
0
Order By: Relevance
“…Prompt tuning differs from embedding adapter (Zhu et al, 2021) that aims to address the multilingual embedding deficiency. An embedding adapter transforms all tokens embeddings but do not affect transformer layers' computation, while prompt tuning does not change tokens embeddings but adds new tunable prompt tokens to the input, serving as context and affecting all following transformer layers.…”
Section: Prompt-based Tuningmentioning
confidence: 99%
“…Prompt tuning differs from embedding adapter (Zhu et al, 2021) that aims to address the multilingual embedding deficiency. An embedding adapter transforms all tokens embeddings but do not affect transformer layers' computation, while prompt tuning does not change tokens embeddings but adds new tunable prompt tokens to the input, serving as context and affecting all following transformer layers.…”
Section: Prompt-based Tuningmentioning
confidence: 99%
“…Adapters. Adapter-based methods have been shown effective in transferring to new languages in multilingual NMT (Üstün et al, 2021;Berard, 2021;Cooper Stickland et al, 2021;Zhu et al, 2021) and fast adaptation to new domains . Combining task-specific adapters with attention mechanism (Pfeiffer et al, 2021) or ensemble (Wang et al, 2021c) allows efficient transfer to low-resource natural language understanding (NLU) and NMT tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, MNMT mainly focus on: 1) designing effective parameters sharing strategy (Zhang et al, 2021a;Zhu et al, 2021;Xie et al, 2021;Lin et al, 2021); 2) obtaining language-agnostic representations (Zhu et al, 2020;Pan et al, 2021); 3) incorporating pre-training models (Siddhant et al, 2020;Wang et al, 2020b); 4) resolving the data imbalance among diverse languages (Wang et al, 2020a;Zhang et al, 2021b;Zhou et al, 2021). Different from them, LSSD is designed for bridging the gap between the overall best checkpoint and language-specific best checkpoints.…”
Section: Advances In Mnmtmentioning
confidence: 99%