Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 2021
DOI: 10.18653/v1/2021.findings-acl.248
|View full text |Cite
|
Sign up to set email alerts
|

ZmBART: An Unsupervised Cross-lingual Transfer Framework for Language Generation

Abstract: Despite the recent advancement in NLP research, cross-lingual transfer for natural language generation is relatively understudied. In this work, we transfer supervision from high resource language (HRL) to multiple lowresource languages (LRLs) for natural language generation (NLG). We consider four NLG tasks (text summarization, question generation, news headline generation, and distractor generation) and three syntactically diverse languages, i.e., English, Hindi, and Japanese. We propose an unsupervised cros… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 16 publications
(19 citation statements)
references
References 19 publications
1
18
0
Order By: Relevance
“…• Freezing of decoder and embeddings (Maurya et al, 2021): only weights in the encoder are finetuned. The motivation behind this approach is that the decoder should keep capabilities of generating in various languages while the encoder will adapt the model to the task;…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…• Freezing of decoder and embeddings (Maurya et al, 2021): only weights in the encoder are finetuned. The motivation behind this approach is that the decoder should keep capabilities of generating in various languages while the encoder will adapt the model to the task;…”
Section: Related Workmentioning
confidence: 99%
“…• Mixing-in self-supervised data for target languages (Lester et al, 2021;Maurya et al, 2021): during finetuning, task data instances in source language will be alternated with self-supervised data instances in target language. The motivation is that such a mixing will preserve model's capability of generation in target languages;…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…As is discussed in prior sections, zero-shot cross-lingual generation with mPLM [19,27] suffers much from catastrophic forgetting [41,45]. Prior work assumes it is because mPLM has never been exposed to sentences of any other languages especially the target language during its fine-tuning on supervised data of source language [45].…”
Section: Prompt Learning To Mitigate Catastrophic Forgetting In Fs-xl...mentioning
confidence: 99%
“…While mPLM has shown promise in enabling cross-lingual transfer for generation tasks, zero-shot cross-lingual transfer (ZS-XLT) with mPLM suffers much from catastrophic forgetting, where mPLM that has been fine-tuned on the source language is unable to generate fluent sentences in the target language when being evaluated on it [19,27,41,45]. In this work, we would like to investigate if catastrophic forgetting occurs in two other cross-lingual scenarios namely FS-XLT and MTL in the context of dialogue generation given the availability of just a few data examples in target language.…”
mentioning
confidence: 99%