2020
DOI: 10.1145/3389790
|View full text |Cite
|
Sign up to set email alerts
|

Iterative Training of Unsupervised Neural and Statistical Machine Translation Systems

Abstract: Recent work achieved remarkable results in training neural machine translation (NMT) systems in a fully unsupervised way, with new and dedicated architectures that only rely on monolingual corpora. However, previous work also showed that unsupervised statistical machine translation (USMT) performs better than unsupervised NMT (UNMT), especially for distant language pairs. To take advantage of the superiority of USMT over UNMT, and considering that SMT suffers from well-known limitations overcome by NMT, we pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 22 publications
0
2
0
Order By: Relevance
“…Other dominant factors identified in the literature that limit the end performance of BLI systems include: (i) linguistic differences (ii) algorithmic mismatch, (iii) variation in data size, (iv) parameterization etc. Similar to the supervised models, the unsupervised variants of BLI are also unable to cater to the above-mentioned challenges (Kim et al, 2020;Marie and Fujita, 2020). Instead of relying on embedding spaces trained completely independent of each other, in the recent past there have been a shift in explicitly using the isomorphism measures alongside distributional training objective (Marchisio et al, 2022).…”
Section: Introductionmentioning
confidence: 99%
“…Other dominant factors identified in the literature that limit the end performance of BLI systems include: (i) linguistic differences (ii) algorithmic mismatch, (iii) variation in data size, (iv) parameterization etc. Similar to the supervised models, the unsupervised variants of BLI are also unable to cater to the above-mentioned challenges (Kim et al, 2020;Marie and Fujita, 2020). Instead of relying on embedding spaces trained completely independent of each other, in the recent past there have been a shift in explicitly using the isomorphism measures alongside distributional training objective (Marchisio et al, 2022).…”
Section: Introductionmentioning
confidence: 99%
“…Some other noteworthy aspects identified in the literature that limit the end performance of the BLI systems, include: (a) algorithmic mismatch for independently trained monolingual embeddings, (b) different parameterization, (c) variable data sizes, (d) linguistic difference, etc., (Marie and Fujita, 2020;Marchisio et al, 2022).…”
Section: Introductionmentioning
confidence: 99%