Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1) 2019
DOI: 10.18653/v1/w19-5327
|View full text |Cite
|
Sign up to set email alerts
|

Incorporating Word and Subword Units in Unsupervised Machine Translation Using Language Model Rescoring

Abstract: This paper describes CAiRE's submission to the unsupervised machine translation track of the WMT'19 news shared task from German to Czech. We leverage a phrase-based statistical machine translation (PBSMT) model and a pre-trained language model to combine word-level neural machine translation (NMT) and subword-level NMT models without using any parallel data. We propose to solve the morphological richness problem of languages by training byte-pair encoding (BPE) embeddings for German and Czech separately, and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 14 publications
0
4
0
Order By: Relevance
“…Today, they play a critical role in many NLP tasks, such as text classification, machine comprehension and natural language inference (Peters et al, 2018;Devlin et al, 2018;Liu et al, 2019a;, to name just a few. They serve as a pre-training objective for downstream applications and they have been used to showcase and measure the general progress in NLP (Yu et al, 2017;Liu et al, 2019b).…”
Section: Introductionmentioning
confidence: 99%
“…Today, they play a critical role in many NLP tasks, such as text classification, machine comprehension and natural language inference (Peters et al, 2018;Devlin et al, 2018;Liu et al, 2019a;, to name just a few. They serve as a pre-training objective for downstream applications and they have been used to showcase and measure the general progress in NLP (Yu et al, 2017;Liu et al, 2019b).…”
Section: Introductionmentioning
confidence: 99%
“…English and Multilingual LMs. Pre-trained LMs exploiting a self-supervised objective with masking such as BERT (Devlin et al, 2019) and RoBERTa (Liu et al, 2019b) have revolutionized NLP. Multilingual versions of these models such as mBERT and XLM-RoBERTa (Conneau et al, 2020) were also pre-trained.…”
Section: Related Workmentioning
confidence: 99%
“…Lample et al (2016) has successfully concatenated character and word embeddings to their model, showing the potential of combining multiple representations. Liu et al (2019) proposed to leverage word and subword embeddings into the application of unsupervised machine translation.…”
Section: Related Workmentioning
confidence: 99%