2020
DOI: 10.1609/aaai.v34i05.6233
|View full text |Cite
|
Sign up to set email alerts
|

Do Not Have Enough Data? Deep Learning to the Rescue!

Abstract: Based on recent advances in natural language modeling and those in text generation capabilities, we propose a novel data augmentation method for text classification tasks. We use a powerful pre-trained neural network model to artificially synthesize new labeled data for supervised learning. We mainly focus on cases with scarce labeled data. Our method, referred to as language-model-based data augmentation (LAMBADA), involves fine-tuning a state-of-the-art language generator to a specific task through an initia… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
115
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 201 publications
(116 citation statements)
references
References 20 publications
1
115
0
Order By: Relevance
“…GPT [36], GPT-2 [22] models are capable of producing grammatically correct, high-quality texts even when fine-tuned on small training data [14]. Nevertheless, the lack of ability to preserve or protect certain words from the original text cannot be assured by this method either.…”
Section: Text Generationmentioning
confidence: 99%
See 2 more Smart Citations
“…GPT [36], GPT-2 [22] models are capable of producing grammatically correct, high-quality texts even when fine-tuned on small training data [14]. Nevertheless, the lack of ability to preserve or protect certain words from the original text cannot be assured by this method either.…”
Section: Text Generationmentioning
confidence: 99%
“…Hence augmentation can improve the robustness and performance of the models. Recently, many studies have been published to tackle the problem of data augmentation in the NLP field [14][15][16]. Some approaches depend more on the language or language models [14,17], while others are (almost) independent [15,18].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Some approaches in this group replicate samples through word replacements based on embeddings of the word and its surrounding context [23,16,26]. Other group of approaches have explored translation and back-translation [20,22], auto-regressive language models [1], and auto-encoders [13].…”
Section: Introductionmentioning
confidence: 99%
“…Reconstructing the training data of the unobserved variables is quite difficult without the aid of a suitable model. Since the data in climate science, geophysics, and many other complex nonlinear systems are spatiotemporally correlated and intrinsically chaotic, the traditional data augmentation methods [4][5][6] for static data are not applicable to expanding the training data set of these problems. On the other hand, thanks to the development of many physics-based dynamical models in describing nature, long correlated time series from these models have been used for training the machine learning algorithms [7,8].…”
Section: Introductionmentioning
confidence: 99%