Proceedings of the Second Conference on Machine Translation 2017
DOI: 10.18653/v1/w17-4704
|View full text |Cite
|
Sign up to set email alerts
|

Modeling Target-Side Inflection in Neural Machine Translation

Abstract: NMT systems have problems with large vocabulary sizes.Byte-pair encoding (BPE) is a popular approach to solving this problem, but while BPE allows the system to generate any target-side word, it does not enable effective generalization over the rich vocabulary in morphologically rich languages with strong inflectional phenomena. We introduce a simple approach to overcome this problem by training a system to produce the lemma of a word and its morphologically rich POS tag, which is then followed by a determinis… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
29
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 53 publications
(29 citation statements)
references
References 10 publications
0
29
0
Order By: Relevance
“…Morphological generation of previously unencountered word forms is a crucial problem in many areas of natural language processing (NLP). High performance can lead to better systems for downstream tasks, e.g., machine translation (Tamchyna et al, 2017). Since existing lexicons have limited coverage, learning morphological inflection patterns from labeled data is an important mission and has recently been the subject of multiple shared tasks (Cotterell et al, 2016(Cotterell et al, , 2017a.…”
Section: Introductionmentioning
confidence: 99%
“…Morphological generation of previously unencountered word forms is a crucial problem in many areas of natural language processing (NLP). High performance can lead to better systems for downstream tasks, e.g., machine translation (Tamchyna et al, 2017). Since existing lexicons have limited coverage, learning morphological inflection patterns from labeled data is an important mission and has recently been the subject of multiple shared tasks (Cotterell et al, 2016(Cotterell et al, , 2017a.…”
Section: Introductionmentioning
confidence: 99%
“…The output of the model is converted into surface forms in a separate, deterministic post-processing step. A similar two-step approach has been found to improve English to Czech NMT (Tamchyna et al, 2017), probably due to alleviating data sparsity caused by morphological complexity. As Finnish is also a morphologically complex language, adapting this approach to Finnish should result in a similar improvement.…”
Section: Nmt With Morphological Analysis and Generationmentioning
confidence: 93%
“…The annotation format used differs from the one in Tamchyna et al (2017) in several aspects, the most important of which is that the morphological tags are not complex, multicategory tags that are interleaved one-to-one with lemmas. Instead, each lemma token can be followed by zero or more morphological tags, each corresponding to a nondefault value in a single morphological category: The first lemma komissio is the only one without any morphological tags, the rest of the lemmas are trailed by one or more tags.…”
Section: Nmt With Morphological Analysis and Generationmentioning
confidence: 99%
“…Moreover, the generated morphological tags following slot placeholders can be used to limit the scope of possible surface forms during lexicalization (see Section 3.3). This approach is inspired by similar approaches in phrase-based MT (Bojar, 2007;Toutanova et al, 2008;Fraser, 2009) and was developed in parallel to recent similar experiments with two-step neural MT (Nadejde et al, 2017;Tamchyna et al, 2017). We compare the lemma-tag generation mode against the TGen default direct word-form generation mode in our experiments.…”
Section: Lemma-tag Generation Modementioning
confidence: 99%