Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing 2016
DOI: 10.18653/v1/d16-1160
|View full text |Cite
|
Sign up to set email alerts
|

Exploiting Source-side Monolingual Data in Neural Machine Translation

Abstract: Neural Machine Translation (NMT) based on the encoder-decoder architecture has recently become a new paradigm. Researchers have proven that the target-side monolingual data can greatly enhance the decoder model of NMT. However, the source-side monolingual data is not fully explored although it should be useful to strengthen the encoder model of NMT, especially when the parallel corpus is far from sufficient. In this paper, we propose two approaches to make full use of the sourceside monolingual data in NMT. Th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

2
219
1

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 256 publications
(222 citation statements)
references
References 29 publications
2
219
1
Order By: Relevance
“…Different works have tackled the inclusion of monolingual data, either in source (Zhang and Zong, 2016b) and target language (Gulcehre et al, 2015(Gulcehre et al, , 2017.…”
Section: Related Workmentioning
confidence: 99%
“…Different works have tackled the inclusion of monolingual data, either in source (Zhang and Zong, 2016b) and target language (Gulcehre et al, 2015(Gulcehre et al, , 2017.…”
Section: Related Workmentioning
confidence: 99%
“…However, training with respect to the new loss is often computationally intensive and requires approximations. Alternatively, multi-task learning has been used to incorporate source-side (Zhang and Zong, 2016) and target-side (Domhan and Hieber, 2017) monolingual data. Another way of utilizing monolingual data in both source and target language is to warm start Seq2Seq training from pre-trained encoder and decoder networks (Ramachandran et al, 2017;Skorokhodov et al, 2018).…”
Section: Other Approachesmentioning
confidence: 99%
“…Our approach further relates to Zhang and Zong (2016), who investigate multi-task learning for sequenceto-sequence models by strengthening the encoder using source-side monolingual data. A shared encoder architecture is used to predict both, transla-tions of parallel source sentences and permutations of monolingual source sentences.…”
Section: Related Workmentioning
confidence: 99%