Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1304
|View full text |Cite
|
Sign up to set email alerts
|

Neural Machine Translation with Source Dependency Representation

Abstract: Source dependency information has been successfully introduced into statistical machine translation.However, there are only a few preliminary attempts for Neural Machine Translation (NMT), such as concatenating representations of source word and its dependency label together. In this paper, we propose a novel attentional NMT with source dependency representation to improve translation performance of NMT, especially on long sentences. Empirical results on NIST Chinese-toEnglish translation task show that our me… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
41
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
2
2

Relationship

4
5

Authors

Journals

citations
Cited by 63 publications
(41 citation statements)
references
References 16 publications
0
41
0
Order By: Relevance
“…Their work was further advanced in [5], [6], which encoded the entire source sentence. Chen et al [38], [43] flatten local dependency structures including parent, children, and sibling words into a word tuple, which is represeted as a compositonal vector by a convotional NN to model the long-distance dependency constraints. In spite of their success, their approaches center around capturing relations between each word and its local dependency words.…”
Section: Discussionmentioning
confidence: 99%
“…Their work was further advanced in [5], [6], which encoded the entire source sentence. Chen et al [38], [43] flatten local dependency structures including parent, children, and sibling words into a word tuple, which is represeted as a compositonal vector by a convotional NN to model the long-distance dependency constraints. In spite of their success, their approaches center around capturing relations between each word and its local dependency words.…”
Section: Discussionmentioning
confidence: 99%
“…It has greatly improved the performance of translation. On this basis, there are many interesting and effective methods [5], [16], [26], [34], [35] which have been proposed in improving attention mechanism of the NMT system. Luong et al [5] proposed global attention model and local attention model, further compare several different scoring functions of the attention weight.…”
Section: Attention Mechanismmentioning
confidence: 99%
“…Tu et al [26] presented a coverage vector to keep track of the attention history and promote the attention mechanism to focus on more untranslated source words. Chen et al [16], [34] proposed a double context method by two attention mechanism to capture more source context information for translation prediction. Our work has the same source attention mechanism, compared with the above models, the forward and the reverse target attention are also imported, which can help to produce a more smooth translation.…”
Section: Attention Mechanismmentioning
confidence: 99%
“…Operation sequence models are n-gram models that include lexical translation operations and reordering operations in a single generative story, thereby combining elements from the previous three model families (Durrani et al, 2011(Durrani et al, , 2013(Durrani et al, , 2014. Their method were further extended by source syntax information (Chen et al, 2017c(Chen et al, , 2018b to improve the performance of SMT.…”
Section: Reordering Model For Pbsmtmentioning
confidence: 99%