2019
DOI: 10.1007/978-3-030-18305-9_9
|View full text |Cite
|
Sign up to set email alerts
|

TextKD-GAN: Text Generation Using Knowledge Distillation and Generative Adversarial Networks

Abstract: Text generation is of particular interest in many NLP applications such as machine translation, language modeling, and text summarization. Generative adversarial networks (GANs) achieved a remarkable success in high quality image generation in computer vision, and recently, GANs have gained lots of interest from the NLP community as well. However, achieving similar success in NLP would be more challenging due to the discrete nature of text. In this work, we introduce a method using knowledge distillation to ef… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 35 publications
(14 citation statements)
references
References 32 publications
0
7
0
Order By: Relevance
“…Knowledge distillation is extensively studied in the field of natural language processing (NLP), in order to obtain the lightweight, efficient and effective language models. More and more KD methods are proposed for solving the numerous NLP tasks (Liu et al, 2019b;Gordon and Duh, 2019;Haidar and Rezagholizadeh, 2019;Yang et al, 2020b;Tang et al, 2019;Hu et al, 2018;Nakashole and Flauger, 2017;Jiao et al, 2019;Wang et al, 2018c;Zhou et al, 2019a;Sanh et al, 2019;Turc et al, 2019;Arora et al, 2019;Clark et al, 2019;Kim and Rush, 2016;Mou et al, 2016;Liu et al, 2019e;Hahn and Choi, 2019;Kuncoro et al, 2016;Cui et al, 2017;Wei et al, 2019;Freitag et al, 2017;Shakeri et al, 2019;Aguilar et al, 2020). The existing NLP tasks using KD contain neural machine translation (NMT) (Hahn and Choi, 2019;Zhou et al, 2019a;Kim and Rush, 2016;Wei et al, 2019;Freitag et al, 2017;Gordon and Duh, 2019), question answering system (Wang et al, 2018c;Arora et al, 2019;Yang et al, 2020b;Hu et al, 2018), document retrieval (Shakeri et al, 2019), event detection (Liu et al, 2019b), text generation (Haidar and Rezagholizadeh, 2019)...…”
Section: Kd In Nlpmentioning
confidence: 99%
“…Knowledge distillation is extensively studied in the field of natural language processing (NLP), in order to obtain the lightweight, efficient and effective language models. More and more KD methods are proposed for solving the numerous NLP tasks (Liu et al, 2019b;Gordon and Duh, 2019;Haidar and Rezagholizadeh, 2019;Yang et al, 2020b;Tang et al, 2019;Hu et al, 2018;Nakashole and Flauger, 2017;Jiao et al, 2019;Wang et al, 2018c;Zhou et al, 2019a;Sanh et al, 2019;Turc et al, 2019;Arora et al, 2019;Clark et al, 2019;Kim and Rush, 2016;Mou et al, 2016;Liu et al, 2019e;Hahn and Choi, 2019;Kuncoro et al, 2016;Cui et al, 2017;Wei et al, 2019;Freitag et al, 2017;Shakeri et al, 2019;Aguilar et al, 2020). The existing NLP tasks using KD contain neural machine translation (NMT) (Hahn and Choi, 2019;Zhou et al, 2019a;Kim and Rush, 2016;Wei et al, 2019;Freitag et al, 2017;Gordon and Duh, 2019), question answering system (Wang et al, 2018c;Arora et al, 2019;Yang et al, 2020b;Hu et al, 2018), document retrieval (Shakeri et al, 2019), event detection (Liu et al, 2019b), text generation (Haidar and Rezagholizadeh, 2019)...…”
Section: Kd In Nlpmentioning
confidence: 99%
“…It is obvious that (A.7) have 6 roots, and λ1 = λ2 = τ are two of the 6 roots. According to the convergence of formula (12), we can obtain the τ is almost small and |τ | < 1. However, in this case, whatever the values of the α, β1andγ are, the dynamic system will converge to the Nash Equilibrium, which is meaningless.…”
Section: A Proofs Inmentioning
confidence: 92%
“…GANs have a wide range of applications [13] because of their capability, which can learn to generate complex and high dimensional target distribution. The existing literature about GANs can be divided into four categories, including music generation [8,11,52], natural languages [5,12,14,25], methods of training GANs [15,33,36,38],images processing [20,44,47,55]. GANs have obtained remarkable progress in image processing, such as video generation [40,42,43], noise removal [53], deblur [18], image to image translation [16,51], image super-resolution [20], medical image processing [6,27,49].…”
Section: Introductionmentioning
confidence: 99%
“…Some recent works involving deep learning are based on CNN [50,48,6], CNN with Multi-headed Attention [46], RNN with Attention module [18], and LSTM [11,39,26]. There are two problems with this technique, a) the models are trained on lexical features such as URL Length, Count of top-level-domain, Number of punctuation symbols, and b) There are no adversarial elements present in training strategies, meaning no synthetic samples are generated or With the popularity of Generative adversarial networks (GAN), there has been a surge in NLP applications ranging from text generation [51,15,13], language models, [47,44] and text classification [12,25]. The reason being, GANs can extract and learn fine-grained information from texts and utilize that to create synthetic examples using a generator architecture.…”
Section: Literature Reviewmentioning
confidence: 99%