2020
DOI: 10.48550/arxiv.2004.13845
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

DARE: Data Augmented Relation Extraction with GPT-2

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 18 publications
(18 citation statements)
references
References 0 publications
0
18
0
Order By: Relevance
“…Recent works show that large pretrained LMs are capable of generating data of reasonable quality (Anaby-Tavor et al, 2020;Papanikolaou and Pierleoni, 2020;Yang et al, 2020;Mohapatra et al, 2021;Kumar et al, 2020;Schick and Schütze, 2021a;Meng et al, 2022), sometimes leading to better transfer learning than human generated datasets (Liu et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…Recent works show that large pretrained LMs are capable of generating data of reasonable quality (Anaby-Tavor et al, 2020;Papanikolaou and Pierleoni, 2020;Yang et al, 2020;Mohapatra et al, 2021;Kumar et al, 2020;Schick and Schütze, 2021a;Meng et al, 2022), sometimes leading to better transfer learning than human generated datasets (Liu et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…Synthetic data generation with GPT-2 is also explored for relation extraction in Papanikolaou and Pierleoni (2020). This paper fine-tunes GPT-2 over labeled examples of the same relation type, where each sentence in the training data is marked with the two entity mentions in the corresponding relation.…”
Section: Information Extraction (Ie)mentioning
confidence: 99%
“…It has demonstrated remarkable zero-shot multitask adaptability by simply feeding the input of each task into the LM and continuing to generate words. People have also also shown that GPT-2 is able to improve classification tasks via in-domain text augmentation (Papanikolaou and Pierleoni, 2020;Sun et al, 2020). We use a similar technique by first finetuning GPT-2 in the few-shot annotations (Wolf et al, 2019), and then applying it to produce synthetic text through an iterative conditional generation process: With initial seeds being samples of T L plus new samples from information augmentation, the LM iteratively conditions on the previous output sentence to generate in-domain text 2 .…”
Section: Lm Augmentationmentioning
confidence: 99%