Abstract:The goal of WMT 2018 Shared Task on Translation Quality Estimation is to investigate automatic methods for estimating the quality of machine translation results without reference translations. This paper presents the QE Brain system, which proposes the neural Bilingual Expert model as a feature extractor based on conditional target language model with a bidirectional transformer and then processes the semantic representations of source and the translation output with a Bi-LSTM predictive model for automatic qu… Show more
“…Different from most previous quality estimation studies that require feature extraction (Blatz et al, 2004;Specia et al, 2009;Salehi et al, 2014) or post-edited data (Kim et al, 2017;Wang et al, 2018;Ive et al, 2018) to train external confidence estimators, all our approach needs is the NMT model itself. Hence, it is easy to apply our approach to arbitrary NMT models trained for arbitrary language pairs.…”
While back-translation is simple and effective in exploiting abundant monolingual corpora to improve low-resource neural machine translation (NMT), the synthetic bilingual corpora generated by NMT models trained on limited authentic bilingual data are inevitably noisy. In this work, we propose to quantify the confidence of NMT model predictions based on model uncertainty. With word-and sentence-level confidence measures based on uncertainty, it is possible for back-translation to better cope with noise in synthetic bilingual corpora. Experiments on Chinese-English and English-German translation tasks show that uncertainty-based confidence estimation significantly improves the performance of backtranslation. 1
“…Different from most previous quality estimation studies that require feature extraction (Blatz et al, 2004;Specia et al, 2009;Salehi et al, 2014) or post-edited data (Kim et al, 2017;Wang et al, 2018;Ive et al, 2018) to train external confidence estimators, all our approach needs is the NMT model itself. Hence, it is easy to apply our approach to arbitrary NMT models trained for arbitrary language pairs.…”
While back-translation is simple and effective in exploiting abundant monolingual corpora to improve low-resource neural machine translation (NMT), the synthetic bilingual corpora generated by NMT models trained on limited authentic bilingual data are inevitably noisy. In this work, we propose to quantify the confidence of NMT model predictions based on model uncertainty. With word-and sentence-level confidence measures based on uncertainty, it is possible for back-translation to better cope with noise in synthetic bilingual corpora. Experiments on Chinese-English and English-German translation tasks show that uncertainty-based confidence estimation significantly improves the performance of backtranslation. 1
“…• SRC → PE: trained first on the in-domain corpus provided, then fine-tuned on the shared task data. deepQUEST is the open source system developed by Ive et al (2018), UNQE is the unpublished system from Jiangxi Normal University, described by Specia et al (2018a), and QE Brain is the system from Alibaba described by Wang et al (2018). Reported numbers for the OpenKiwi system correspond to best models in the development set: the STACKED model for prediction of MT tags, and the ENSEMBLED model for the rest.…”
Section: Benchmark Experimentsmentioning
confidence: 99%
“…• Implementation of four QE systems: QUETCH (Kreutzer et al, 2015), NUQE (Martins et al, 2016, Predictor-Estimator Wang et al, 2018), and a stacked ensemble with a linear system (Martins et al, 2016(Martins et al, , 2017;…”
We introduce OpenKiwi, a PyTorch-based open source framework for translation quality estimation. OpenKiwi supports training and testing of word-level and sentence-level quality estimation systems, implementing the winning systems of the WMT 2015-18 quality estimation campaigns. We benchmark OpenKiwi on two datasets from WMT 2018 (English-German SMT and NMT), yielding state-of-the-art performance on the word-level tasks and near state-of-the-art in the sentencelevel tasks.
“…To implement the proposed model, the authors combined a neural based word prediction model with the translation QE models of word and sentence level. Later, the above neural MT model was replaced by Wang et al [20] with a modified version of self-attention-based transformer model [21] for estimating the English-German language-based sentence level translation quality estimation.…”
This submission describes the study of linguistically motivated features to estimate the translated sentence quality at sentence level on English-Hindi language pair. Several classification algorithms are employed to build the Quality Estimation (QE) models using the extracted features. We used source language text and the MT output to extract these features. Experiments show that our proposed approach is robust and producing competitive results for the DT based QE model on neural machine translation system.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.