Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.205
|View full text |Cite
|
Sign up to set email alerts
|

Referring to what you know and do not know: Making Referring Expression Generation Models Generalize To Unseen Entities

Abstract: Data-to-text Natural Language Generation (NLG) is the computational process of generating natural language in the form of text or voice from non-linguistic data. A core micro-planning task within NLG is referring expression generation (REG), which aims to automatically generate noun phrases to refer to entities mentioned as discourse unfolds. A limitation of novel REG models is not being able to generate referring expressions to entities not encountered during the training process. To solve this problem, we pr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(13 citation statements)
references
References 24 publications
0
8
0
Order By: Relevance
“…We evaluated all the systems described in §3 on both WEBNLG and WSJ using automatic and human evaluations. We implemented the neural models based on the code of Cunha et al (2020) and Cao and Cheung (2019) 7 . For WEBNLG, we used their original parameter setting, while for WSJ, we tuned the parameters on the development set and used the best parameter set.…”
Section: Discussionmentioning
confidence: 99%
See 4 more Smart Citations
“…We evaluated all the systems described in §3 on both WEBNLG and WSJ using automatic and human evaluations. We implemented the neural models based on the code of Cunha et al (2020) and Cao and Cheung (2019) 7 . For WEBNLG, we used their original parameter setting, while for WSJ, we tuned the parameters on the development set and used the best parameter set.…”
Section: Discussionmentioning
confidence: 99%
“…ATT+Copy. Cunha et al (2020) proposed using three bidirectional LSTMs (Hochreiter and Schmidhuber, 1997) to encode a pre-context, a post-context, and the proper name of an entity (i.e., replacing underscores in entity labels with whitespaces) into three hidden vectors h (pre) , h (post) and h (r) , respectively. An auto-regressive LSTM-based decoder generates REs based on context vectors.…”
Section: Linguistically Informed Features (Ml-l)mentioning
confidence: 99%
See 3 more Smart Citations