Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume 2021
DOI: 10.18653/v1/2021.eacl-main.64
|View full text |Cite
|
Sign up to set email alerts
|

Neural Data-to-Text Generation with LM-based Text Augmentation

Abstract: For many new application domains for datato-text generation, the main obstacle in training neural models consists of a lack of training data. While usually large numbers of instances are available on the data side, often only very few text samples are available. To address this problem, we here propose a novel fewshot approach for this setting. Our approach automatically augments the data available for training by (i) generating new text samples based on replacing specific values by alternative ones from the s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 31 publications
(15 citation statements)
references
References 34 publications
(42 reference statements)
0
15
0
Order By: Relevance
“…(Du and Black, 2018) or paraphrasing-based methods (Niu and Bansal, 2019;Li et al, 2019;Cai et al, 2020;Zhang et al, 2020a;Xie et al, 2022;Cao et al, 2022). Another line of work (Chang et al, 2021;Schick and Schütze, 2021;Zheng et al, 2022) is exploiting large-scale pretrained language models for data augmentation. However, these existing methods do not focus on creating semantically different responses.…”
Section: Further Discussionmentioning
confidence: 99%
“…(Du and Black, 2018) or paraphrasing-based methods (Niu and Bansal, 2019;Li et al, 2019;Cai et al, 2020;Zhang et al, 2020a;Xie et al, 2022;Cao et al, 2022). Another line of work (Chang et al, 2021;Schick and Schütze, 2021;Zheng et al, 2022) is exploiting large-scale pretrained language models for data augmentation. However, these existing methods do not focus on creating semantically different responses.…”
Section: Further Discussionmentioning
confidence: 99%
“…Also, to obtain the hidden state, (7) and the updated cell state tanh are used. The (8) obtains the maximum score as an output for the prediction.…”
Section: Methods and Implementation Of The Casementioning
confidence: 99%
“…They also proposed several methods for extracting context vectors. Similarly, Chang et al [8] proposed an approach that automatically augments the data available for training by generating new text samples based on the GPT-2, and proposed to match the new text samples with the data. The Buddana et al [9] created a model describing the design and operation of text generation, for which they used LSTM as an innovative solution for handling sequential data and text data in particular.…”
Section: Related Workmentioning
confidence: 99%
“…For the generation of court view, we can take them as text generation problems, such as BLEU-1, BLEU-2, BLEU-N, ROUGE-1, ROUGE-2, ROUGE-L and BERT SCORE [71,72,73,74,34,75,76,77]. The following evaluation metrics to evaluate the performance of LJP text generation are summarized, as shown in Fig.…”
Section: Evaluation Of Text Generationmentioning
confidence: 99%