Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.352
|View full text |Cite
|
Sign up to set email alerts
|

Data-to-text Generation by Splicing Together Nearest Neighbors

Abstract: We propose to tackle data-to-text generation tasks by directly splicing together retrieved segments of text from "neighbor" sourcetarget pairs. Unlike recent work that conditions on retrieved neighbors but generates text token-by-token, left-to-right, we learn a policy that directly manipulates segments of neighbor text, by inserting or replacing them in partially constructed generations. Standard techniques for training such a policy require an oracle derivation for each generation, and we prove that finding … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 52 publications
0
7
0
Order By: Relevance
“…We also notice that models sometimes fail to generate informative summaries while also failing to predict anything about data (see the second example in Table 6). In both Bart-Large models, we also find some cases where the models generate information that is fully irrelevant to the chart (i.e., the Hallucination effect [1,8,29]) (see the third example in Table 6).…”
Section: Error Analysis and Challengesmentioning
confidence: 92%
“…We also notice that models sometimes fail to generate informative summaries while also failing to predict anything about data (see the second example in Table 6). In both Bart-Large models, we also find some cases where the models generate information that is fully irrelevant to the chart (i.e., the Hallucination effect [1,8,29]) (see the third example in Table 6).…”
Section: Error Analysis and Challengesmentioning
confidence: 92%
“…In contrast, TempLM first finetunes a PLM on the data-to-text task and then exploits the PLM's ability in smoothing the text distribution to tackle the paraphrasing problem. (Wiseman et al, 2021), and inducing latent variables (Wiseman et al, 2018;Li and Rush, 2020;Ye et al, 2020). Much like classic template-based methods, these systems attempt to learn structured representation from diverse human-written text, which is challenging and often requires heuristics for additional supervision.…”
Section: Related Workmentioning
confidence: 99%
“…While obtaining C * is impossible, we can design approximate clusters C based on the presence of different fields, as is standard in other data-to-text methods (Wiseman et al, 2021).…”
Section: Template Extractionmentioning
confidence: 99%
See 1 more Smart Citation
“…Traditionally, nearest neighbors have been utilized in various tasks, including part-of-speech tagging [6] and example-based machine translation [23]. Recently, nearest neighbors have been successfully applied to AR models, particularly to neural machine translation [13], [15], [37], text summarization [5], [25], and datato-text generation [36]. We adopt the pioneering idea to refine and improve NAR models.…”
Section: Introductionmentioning
confidence: 99%