Proceedings of the Fourth Conference on Machine Translation (Volume 3: Shared Task Papers, Day 2) 2019
DOI: 10.18653/v1/w19-5410
|View full text |Cite
|
Sign up to set email alerts
|

Quality Estimation and Translation Metrics via Pre-trained Word and Sentence Embeddings

Abstract: We propose the use of pre-trained embeddings as features of a regression model for sentencelevel quality estimation of machine translation. In our work we combine freely available BERT and LASER multilingual embeddings to train a neural-based regression model. In the second proposed method we use as an input features not only pre-trained embeddings, but also log probability of any machine translation (MT) system. Both methods are applied to several language pairs and are evaluated both as a classical quality e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
2

Relationship

1
9

Authors

Journals

citations
Cited by 29 publications
(15 citation statements)
references
References 18 publications
0
15
0
Order By: Relevance
“…Finally, their systems were pre-trained on synthetic data, obtained by taking all of the WMT submissions from earlier years and using chrF (Popović, 2015) as the synthetic output. The approach is described in greater detail in (Yankovskaya et al, 2019).…”
Section: Utartumentioning
confidence: 99%
“…Finally, their systems were pre-trained on synthetic data, obtained by taking all of the WMT submissions from earlier years and using chrF (Popović, 2015) as the synthetic output. The approach is described in greater detail in (Yankovskaya et al, 2019).…”
Section: Utartumentioning
confidence: 99%
“…Our models were not trained on Gujarati (gu). For brevity, only the best QE-metric for each language pair is shown-for full results see Appendix G. a:YISI-2(Lo, 2019) b:YISI-2 SRL(Lo, 2019) c:UNI(Yankovskaya et al, 2019) d:UNI+(Yankovskaya et al, 2019).…”
mentioning
confidence: 99%
“…We compare with a range of reference-free metrics: ibm1-morpheme and ibm1-pos4gram (Popović, 2012), LASIM (Yankovskaya et al, 2019), LP (Yankovskaya et al, 2019), YiSi-2 and YiSi-2-srl (Lo, 2019), and reference-based baselines BLEU (Papineni et al, 2002), SentBLEU (Koehn et al, 2007) and ChrF++ (Popović, 2017) for MT evaluation (see §2). 6 The main results are reported on WMT17.…”
Section: Baselinesmentioning
confidence: 99%