Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1301
|View full text |Cite
|
Sign up to set email alerts
|

Exploiting Cross-Sentence Context for Neural Machine Translation

Abstract: In translation, considering the document as a whole can help to resolve ambiguities and inconsistencies. In this paper, we propose a cross-sentence context-aware approach and investigate the influence of historical contextual information on the performance of neural machine translation (NMT). First, this history is summarized in a hierarchical way. We then integrate the historical representation into NMT in two strategies: 1) a warm-start of encoder and decoder states, and 2) an auxiliary context source for up… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

3
217
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 164 publications
(221 citation statements)
references
References 20 publications
3
217
1
Order By: Relevance
“…Encoding Discourse-Level Context Hierarchical structure networks are usually used for modelling discourse context on various natural language processing tasks such query suggestion (Sordoni et al, 2015), dialogue modeling (Serban et al, 2016) and MT (Wang et al, 2017). Therefore, we employ hierarchical encoder (Wang et al, 2017) to encoder discourse-level context for NMT. More specifically, we use the previous K source sentences X = {x −K , .…”
Section: Discourse-aware Zp Predictionmentioning
confidence: 99%
See 2 more Smart Citations
“…Encoding Discourse-Level Context Hierarchical structure networks are usually used for modelling discourse context on various natural language processing tasks such query suggestion (Sordoni et al, 2015), dialogue modeling (Serban et al, 2016) and MT (Wang et al, 2017). Therefore, we employ hierarchical encoder (Wang et al, 2017) to encoder discourse-level context for NMT. More specifically, we use the previous K source sentences X = {x −K , .…”
Section: Discourse-aware Zp Predictionmentioning
confidence: 99%
“…Inspired by these findings, we exploit intersentential context to further improve ZP prediction and thus translation. Concretely, we employ hierarchical neural networks (Sordoni et al, 2015;Wang et al, 2017) to summarize the context of previous sentences in a text, which is integrated to the joint model for ZP prediction.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…While sentence-level evaluation may be good enough when evaluating MT systems of relatively low quality, we hypothesise that with additional context, raters will be able to make more nuanced quality assessments, and will also reward translations that show more textual cohesion and coherence. We believe that this aspect should be considered in evaluation, especially when making claims about human-machine parity, since human translators can and do take inter-sentential context into account (Voigt & Jurafsky, 2012;Wang et al, 2017).…”
Section: Linguistic Contextmentioning
confidence: 99%
“…Intuitively, to generate better translation of the entire text, the model deserves considering the cross-sentence connections and dependencies, generating discourse coherent translations. Towards this demand, most previous work (Wang et al 2017;Voita et al 2018) proposed to explore additional context, generally is certain preceding adjacent sentences, to reinforce the model. However, the major goal of these models is still the quality of individual sentence while not Copyright c 2019, Association for the Advancement of Artificial Intelligence (www.aaai.org).…”
Section: Introductionmentioning
confidence: 99%