2019
DOI: 10.1609/aaai.v33i01.33016698
|View full text |Cite
|
Sign up to set email alerts
|

Insufficient Data Can Also Rock! Learning to Converse Using Smaller Data with Augmentation

Abstract: Recent successes of open-domain dialogue generation mainly rely on the advances of deep neural networks. The effectiveness of deep neural network models depends on the amount of training data. As it is laboursome and expensive to acquire a huge amount of data in most scenarios, how to effectively utilize existing data is the crux of this issue. In this paper, we use data augmentation techniques to improve the performance of neural dialogue models on the condition of insufficient data. Specifically, we propose … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
37
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 41 publications
(37 citation statements)
references
References 16 publications
0
37
0
Order By: Relevance
“…This task can benefit natural language interfaces to information systems by suggesting alternative invocation phrases for particular types of queries (Kumar et al, 2017). It can also bear on dialogue systems that seek to generate utterances that fit particular functional categories (Ke et al, 2018;Li et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…This task can benefit natural language interfaces to information systems by suggesting alternative invocation phrases for particular types of queries (Kumar et al, 2017). It can also bear on dialogue systems that seek to generate utterances that fit particular functional categories (Ke et al, 2018;Li et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…The transformer model for dialogue generation is configured with 512 hidden size, 8 attention heads and 6 blocks in both the encoder and decoder. The hyper-parameters in the baseline models are set following the original papers (Li et al, 2019;Shang et al, 2018;Csáky et al, 2019).…”
Section: Implementation and Reproducibilitymentioning
confidence: 99%
“…Hou et al (2018) utilize a sequence-to-sequence model to produce diverse utterances for language understanding. Li et al (2019); Niu and Bansal (2019) propose to generate sentences for dialogue augmentation. Compared with previous augmentation approaches for dialogue generation, augmented sentences in our framework are selectively generated using the pretrained models and the augmentation process is additionally fine-tuned jointly with the training of dialogue generation.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Our model is motivated by following two points: 1) existing embedding based per-sona modeling methods cannot discover the common properties among users and train the embedding for different user independently, which cause (equal to learning) a very high-dimensional persona embedding and thus have a low data utilization efficiency or require a large amount of training data for each user. 2) benefited by the semantic capturing ability of WAEs, plenitude persona information can be gathered into the continuous space (Li et al, 2019). To this end, we build our model upon the state-of-the-art conversation model WAE (Gu et al, 2019) to model utterance-level and the user-level information.…”
Section: Introductionmentioning
confidence: 99%