Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.530
|View full text |Cite
|
Sign up to set email alerts
|

Contextual Neural Machine Translation Improves Translation of Cataphoric Pronouns

Abstract: The advent of context-aware NMT has resulted in promising improvements in the overall translation quality and specifically in the translation of discourse phenomena such as pronouns. Previous works have mainly focused on the use of past sentences as context with a focus on anaphora translation. In this work, we investigate the effect of future sentences as context by comparing the performance of a contextual NMT model trained with the future context to the one trained with the past context. Our experiments and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(10 citation statements)
references
References 16 publications
0
8
0
Order By: Relevance
“…Coreference resolution is the task of identifying coreference relations among different mentions. As a vital natural language understanding component, a good coreference system could benefit many downstream tasks such as machine translation (Guillou, 2012;Wong et al, 2020), dialog systems (Strube and Müller, 2003), question answering (Dasigi et al, 2019), andsummarization (Steinberger et al, 2007). Due to the weak semantic meaning of pronouns (Ehrlich, 1981), grounding pronouns to their referents (PCR) has been specially studied as a more challenging task than the general coreference resolution (Mitkov, 1998;Ng, 2005).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Coreference resolution is the task of identifying coreference relations among different mentions. As a vital natural language understanding component, a good coreference system could benefit many downstream tasks such as machine translation (Guillou, 2012;Wong et al, 2020), dialog systems (Strube and Müller, 2003), question answering (Dasigi et al, 2019), andsummarization (Steinberger et al, 2007). Due to the weak semantic meaning of pronouns (Ehrlich, 1981), grounding pronouns to their referents (PCR) has been specially studied as a more challenging task than the general coreference resolution (Mitkov, 1998;Ng, 2005).…”
Section: Related Workmentioning
confidence: 99%
“…The coreference relationship between a pronoun and its referents is categorized into endophora and exophora based on whether the referred objects appear in text or out of text, and the former case can be further divided into anaphora if the referents appear in the preceding text of the pronoun and cataphora if in the following text (Halliday and Hasan, 1976;Brown and Yule, 1983). Conventional studies on the pronoun coreference resolution (PCR) task in the NLP community mainly focus on anaphora (Hobbs, 1978;NIST, 2003;Pradhan et al, 2012) and some recent work analyzes cataphora in machine translation (Wong et al, 2020), while mostly ignoring the exophoric pronouns. However, in daily dialogues or conversations, speakers may often use exophoric pronouns to refer to objects in the situational context that all speakers and listeners are aware of without introducing them in the first place.…”
Section: Introductionmentioning
confidence: 99%
“…A straightforward way to offset the sparsity of training signal is to increase the volume of training data. In fact, existing works that report strong results with targeted evaluation train their contextual parameters with millions of documentlevel sentence pairs (Bawden et al, 2018;Müller et al, 2018;Voita et al, 2019b;Zheng et al, 2020;Wong et al, 2020;Kang et al, 2020). In contrast, many works in the literature train models with the TED talks' subtitles released by the IWSLT shared tasks (Cettolo et al, 2012), which only consist of a couple of hundred thousand parallel sentences (see Table 1).…”
Section: A Sparse Training Signalmentioning
confidence: 99%
“…Most recent work on improving pronoun translations has involved building more complex architectures that incorporate contextual information (Voita et al, 2018;Wong et al, 2020). In contrast, we present a more generalized approach.…”
Section: Related Workmentioning
confidence: 99%