2017
DOI: 10.48550/arxiv.1711.00043
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Unsupervised Machine Translation Using Monolingual Corpora Only

Abstract: Machine translation has recently achieved impressive performance thanks to recent advances in deep learning and the availability of large-scale parallel corpora. There have been numerous attempts to extend these successes to low-resource language pairs, yet requiring tens of thousands of parallel sentences. In this work, we take this research direction to the extreme and investigate whether it is possible to learn to translate even without any parallel data. We propose a model that takes sentences from monolin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
130
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 93 publications
(132 citation statements)
references
References 24 publications
2
130
0
Order By: Relevance
“…In recent years, especially with the success of transfer learning (Wang et al, 2018) and pretraining in NLP (Devlin et al, 2019), several techniques for improving neural MT for low-resource languages have been proposed (Sennrich et al, 2016;Fadaee et al, 2017;Xia et al, 2019;Lample et al, 2017;Lewis et al, 2019;Liu et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…In recent years, especially with the success of transfer learning (Wang et al, 2018) and pretraining in NLP (Devlin et al, 2019), several techniques for improving neural MT for low-resource languages have been proposed (Sennrich et al, 2016;Fadaee et al, 2017;Xia et al, 2019;Lample et al, 2017;Lewis et al, 2019;Liu et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Because of the different task natures, balancing the training procedure with all the tasks is a critical problem for multi-task learning. Loss-balancing strategies (Chen et al, 2018b;Kendall et al, 2018;Liu et al, 2019;Gong et al, 2019;Guo et al, 2018a;Lample et al, 2017;Yao et al, 2019) are suitable for situations in which there are multiple training objectives that can be combined via weighted summation for each data point. In contrast, for multi-task learning across different datasets, a sampling strategy should be applied to decide the mixing ratio (how many batches to sample from each task) in each epoch.…”
Section: Multi-task Learningmentioning
confidence: 99%
“…However, in the Textcap task as well as many visionlanguage applications, such large-scale annotations are not readily available, and are both time-consuming and laborintensive to acquire. In these scenarios, unsupervised/unpaired methods (Lample et al 2017;Gu et al 2019;Caron et al 2020) that can learn vision to language or conversely from unpaired training data are highly desirable. Based on this insight, our goal is to break away from single and paired captioning conventions and conduct the unpaired captioning paradigm, which encourages TextCap models to generate diverse textual descriptions to comprehend images better.…”
Section: Introductionmentioning
confidence: 99%