2019
DOI: 10.1609/aaai.v33i01.33017402
|View full text |Cite
|
Sign up to set email alerts
|

Data Augmentation for Spoken Language Understanding via Joint Variational Generation

Abstract: Data scarcity is one of the main obstacles of domain adaptation in spoken language understanding (SLU) due to the high cost of creating manually tagged SLU datasets. Recent works in neural text generative models, particularly latent variable models such as variational autoencoder (VAE), have shown promising results in regards to generating plausible and natural sentences. In this paper, we propose a novel generative architecture which leverages the generative power of latent variable models to jointly synthesi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
23
0
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 71 publications
(24 citation statements)
references
References 20 publications
0
23
0
1
Order By: Relevance
“…Therefore, the reliability of NLU becomes a precursor for the success of dialog systems. Recently, various deep neural network based NLU models are proposed and some of these models have been applied in real-world applications [1]- [3]. Most existing neural NLU modules are built by following a closed-world assumption [4], [5], i.e, the data used in the training and testing phrase are drawn from the same distribution.…”
Section: Introductionmentioning
confidence: 99%
“…Therefore, the reliability of NLU becomes a precursor for the success of dialog systems. Recently, various deep neural network based NLU models are proposed and some of these models have been applied in real-world applications [1]- [3]. Most existing neural NLU modules are built by following a closed-world assumption [4], [5], i.e, the data used in the training and testing phrase are drawn from the same distribution.…”
Section: Introductionmentioning
confidence: 99%
“…Due to the lack of high-quality SLU data, a data augmentation approach is quite essential [15]. SpecAugment [10,16] is a simple yet effective data augmentation method for E2E ASR, resulting in consistent accuracy gain.…”
Section: Data Augmentation For Slumentioning
confidence: 99%
“…The above results suggest the SGP-DST model overfits on the schema element representations and writing style in the training set, reducing its ability to generalize to new writing styles. Data augmentation is a common way to improve robustness (Hou et al 2018;Yoo, Shin, and Lee 2019). To this end, we experiment with a simple back-translation approach (Sennrich, Haddow, and Birch 2016) for augmenting the training schemas for the SGP-DST model and study its impact on model performance and schema sensitivity.…”
Section: Schema Augmentationmentioning
confidence: 99%