Proceedings of the Fourth Workshop on Discourse in Machine Translation (DiscoMT 2019) 2019
DOI: 10.18653/v1/d19-6504
|View full text |Cite
|
Sign up to set email alerts
|

Data augmentation using back-translation for context-aware neural machine translation

Abstract: A single sentence does not always convey information required to translate it into other languages; we sometimes need to add or specialize words that are omitted or ambiguous in the source languages (e.g., zero pronouns in translating Japanese to English or epicene pronouns in translating English to French). To translate such ambiguous sentences, we exploit contexts around the source sentence, and have so far explored context-aware neural machine translation (NMT). However, a large amount of parallel corpora i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
48
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 63 publications
(49 citation statements)
references
References 17 publications
1
48
0
Order By: Relevance
“…Data augmentation is a spotlight in recent years, from a limited training data will automatically generate more training data as considered semi-supervised learning. Sennrich et al [12], Sugiyama and Yoshinaga [13] used back translation technique to generate training data to improve performance of translation model. Fadaee at al.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Data augmentation is a spotlight in recent years, from a limited training data will automatically generate more training data as considered semi-supervised learning. Sennrich et al [12], Sugiyama and Yoshinaga [13] used back translation technique to generate training data to improve performance of translation model. Fadaee at al.…”
Section: Related Workmentioning
confidence: 99%
“…The biggest disadvantage of these methods is not reserving meaning concerning the context of the sentences, so we present more complex approaches retaining the meaning as the original sentence. Back translation aims to obtain more training samples based on the translators, many research teams have used to improve translation models [12][13][14][15]23]. This technique is resolved by using the translators to translate the original data to a certain language, after that taking the translated data into the independent translator to translate back to the original language.…”
Section: Data Augmentationmentioning
confidence: 99%
“…One of the techniques to get pseudo parallel corpora for context-aware NMT models is data augmentation using back-translation [17]. So, taking this approach, we assume that sentence simplification can be partially solved with the back-translation technique without fine-tuning to a downstream task or training a new model.…”
Section: Sentence Simplification Through Back-translationmentioning
confidence: 99%
“…In computer vision, data augmentation technologies are widely applied to generate auxiliary training examples [20][21][22]. In NLP, back-translation has been proven to be effective in augmenting diverse instances [23][24][25]. We borrow this idea and translate training sentences into pivot languages.…”
Section: Plos Onementioning
confidence: 99%