Proceedings of the Tenth Workshop on Statistical Machine Translation 2015
DOI: 10.18653/v1/w15-3012
|View full text |Cite
|
Sign up to set email alerts
|

The KIT-LIMSI Translation System for WMT 2015

Abstract: This paper presented the joined submission of KIT and LIMSI to the English to German translation task of WMT 2015. In this year submission, we integrated a neural network-based translation model into a phrase-based translation model by rescoring the n-best lists.Since the computation complexity is one of the main issues for continuous space models, we compared two techniques to reduce the computation cost. We investigated models using a structured output layer as well as models trained with noise contrastive e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
3
1
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 22 publications
0
3
0
Order By: Relevance
“…Inspired by Refs. 37 and 38, the “[cls]” token used to indicate the beginning of a sequence in BERT is replaced by a language token to distinguish the different encoding methods. The sequential and reverse-sequential tokenizers share the same vocabulary of 26 lowercase letters and 10 numbers, differing in terms of their respective language token and order of text sequence.…”
Section: Methodsmentioning
confidence: 99%
“…Inspired by Refs. 37 and 38, the “[cls]” token used to indicate the beginning of a sequence in BERT is replaced by a language token to distinguish the different encoding methods. The sequential and reverse-sequential tokenizers share the same vocabulary of 26 lowercase letters and 10 numbers, differing in terms of their respective language token and order of text sequence.…”
Section: Methodsmentioning
confidence: 99%
“…In particular, they leveraged high-quality post-edits performed by professional translators on 600 output sentences from both PB-SMT and NMT systems for English-to-German translation of TED talks. The NMT system of Luong and Manning (2015) was compared against a standard PB-SMT system (Ha et al 2015), a hierarchical SMT system (Jehl et al 2015) and a system combining PB-SMT and syntax-based SMT (Huck and Birch 2015). HTERessentially, TER with a human-in-the-loop (Snover et al 2006)and multi-reference TER was used, and results showed that NMT outperformed the other approaches in all metrics at a statistical significance level of p=0.01.…”
Section: The Changing Nature Of Mt System Designmentioning
confidence: 99%
“…Conditioning a multilingual model on the input language has been studied in NMT (Ha et al, 2016b;Johnson et al, 2017), syntactic parsing (Ammar et al, 2016) and language modeling (Östling and Tiedemann, 2017). The goal is to embed language information in real-valued vectors in order to enrich internal representations with input language for multilingual models.…”
Section: Language Embeddings and Typologymentioning
confidence: 99%