2017
DOI: 10.1109/jstsp.2017.2764273
|View full text |Cite
|
Sign up to set email alerts
|

An Empirical Analysis of NMT-Derived Interlingual Embeddings and Their Use in Parallel Sentence Identification

Abstract: End-to-end neural machine translation has overtaken statistical machine translation in terms of translation quality for some language pairs, specially those with large amounts of parallel data. Besides this palpable improvement, neural networks provide several new properties. A single system can be trained to translate between many languages at almost no additional cost other than training time. Furthermore, internal representations learned by the network serve as a new semantic representation of words or sent… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
49
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 66 publications
(49 citation statements)
references
References 23 publications
0
49
0
Order By: Relevance
“…Similar features have proved to be useful for the related task of translation quality estimation (Specia et al, 2010(Specia et al, , 2016, which aims to detect divergences introduced by MT errors, rather than human translation. Recently, sentence embeddings have also been used to detect parallelism (España-Bonet et al, 2017;Schwenk and Douze, 2017). Although embeddings capture semantic generalizations, these models are trained with neural MT objectives, which do not distinguish semantically equivalent segments from divergent parallel segments.…”
Section: Introductionmentioning
confidence: 99%
“…Similar features have proved to be useful for the related task of translation quality estimation (Specia et al, 2010(Specia et al, , 2016, which aims to detect divergences introduced by MT errors, rather than human translation. Recently, sentence embeddings have also been used to detect parallelism (España-Bonet et al, 2017;Schwenk and Douze, 2017). Although embeddings capture semantic generalizations, these models are trained with neural MT objectives, which do not distinguish semantically equivalent segments from divergent parallel segments.…”
Section: Introductionmentioning
confidence: 99%
“…This task consists of finding the gold aligned parallel sentences given two large corpora in two distinct languages. Typically, only 8 Including a 3-layer transformer trained on a constructed parallel corpus (Chidambaram et al, 2018), a bidirectional gated recurrent unit (GRU) network trained on a collection of parallel corpora using en-es, en-ar, and ar-es bitext (Espana-Bonet et al, 2017), and a 3 layer bidirectional LSTM trained on 9 languages in Europarl (Schwenk, 2018 Table 3 on the French and German mining tasks demonstrate the proposed model outperforms Schwenk (2018), although the gap is substantially smaller than on the STS tasks. The reason for this is likely the domain mismatch between the STS data (image captions) and the training data (Europarl).…”
Section: Mining Bitextmentioning
confidence: 99%
“…Zhou (2016) learned bilingual document representations by minimizing the Euclidean distance between document representations and their translations [30]. Conneau (2017) and Espa (2017) jointly trained a sequence to sequence MT system on multiple languages to learn a shared multilingual sentence embedding space [4] [9]. Our method leverages the latest breakthrough in NLP: BERT [7] as the multilingual sentence encoder.…”
Section: Multilingual Sentence Encoder-based Approach (Mse-based)mentioning
confidence: 99%