2011
DOI: 10.1007/s10590-011-9104-y
|View full text |Cite
|
Sign up to set email alerts
|

Chunk-lattices for verb reordering in Arabic–English statistical machine translation

Abstract: Syntactic disfluencies in Arabic-to-English phrase-based SMT output are often due to incorrect verb reordering in Verb-Subject-Object sentences. As a solution, we propose a chunk-based reordering technique to automatically displace clauseinitial verbs in the Arabic side of a word-aligned parallel corpus. This method is used to preprocess the training data, and to collect statistics about verb movements. From this analysis we build specific verb reordering lattices on the test sentences before decoding, and tes… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2011
2011
2020
2020

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 8 publications
(17 citation statements)
references
References 23 publications
0
17
0
Order By: Relevance
“…SDSM and OSM have been proven optimal for language pairs where high distortion limits are required to capture long-range reordering phenomena (Durrani, Schmid, and Fraser 2011;Bisazza and Federico 2013b;…”
Section: Figurementioning
confidence: 99%
“…SDSM and OSM have been proven optimal for language pairs where high distortion limits are required to capture long-range reordering phenomena (Durrani, Schmid, and Fraser 2011;Bisazza and Federico 2013b;…”
Section: Figurementioning
confidence: 99%
“…Green, Galley, and Manning (2010) additionally proposed discriminative distortion models to achieve better translation accuracy than the baseline phrase-based system for a distortion limit of 15 words. Bisazza and Federico (2013) recently proposed a novel method to dynamically select which longrange reorderings to consider during the hypothesis extension process in a phrasebased decoder and showed an improvement in a German-English task by increasing the distortion limit to 18.…”
Section: Phrase-based Smtmentioning
confidence: 99%
“…Using language models for reordering is not something new (Feng et al, 2010), (Durrani et al, 2011), (Bisazza and Federico, 2013), but instead of using a more or less standard n-gram language model, we are going to base our model on recurrent neural network language models (Mikolov et al, 2010).…”
Section: Recurrent Neural Network Reordering Modelsmentioning
confidence: 99%