2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU) 2019
DOI: 10.1109/asru46091.2019.9003747
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Semi-Supervised Learning for Natural Language Understanding by Optimizing Diversity

Abstract: Expanding new functionalities efficiently is an ongoing challenge for single-turn task-oriented dialogue systems. In this work, we explore functionality-specific semi-supervised learning via self-training. We consider methods that augment training data automatically from unlabeled data sets in a functionality-targeted manner. In addition, we examine multiple techniques for efficient selection of augmented utterances to reduce training time and increase diversity. First, we consider paraphrase detection methods… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
9
1

Relationship

3
7

Authors

Journals

citations
Cited by 15 publications
(6 citation statements)
references
References 27 publications
0
6
0
Order By: Relevance
“…Selection by Submodular Optimization: Submodular data selection is used to select a diverse representative subset of samples from given dataset. This method has been applied in speech recognition (Wei et al, 2015), machine translation (Kirchhoff and Bilmes, 2014) and natural language understanding tasks (Cho et al, 2019). For SSL data selection, we use feature-based submodular selection (Kirchhoff and Bilmes, 2014), where submodular functions are given by weighted sums of nondecreasing concave functions applied to modular functions.…”
Section: Data Selection Approachesmentioning
confidence: 99%
“…Selection by Submodular Optimization: Submodular data selection is used to select a diverse representative subset of samples from given dataset. This method has been applied in speech recognition (Wei et al, 2015), machine translation (Kirchhoff and Bilmes, 2014) and natural language understanding tasks (Cho et al, 2019). For SSL data selection, we use feature-based submodular selection (Kirchhoff and Bilmes, 2014), where submodular functions are given by weighted sums of nondecreasing concave functions applied to modular functions.…”
Section: Data Selection Approachesmentioning
confidence: 99%
“…Selection by Submodular Optimization: Sub-modular data selection is used to select a diverse representative subset of samples from given dataset. This method has been applied in speech recognition (Wei et al, 2015), machine translation (Kirchhoff and Bilmes, 2014) and natural language understanding tasks (Cho et al, 2019). For SSL data selection, we use feature-based submodular selection (Kirchhoff and Bilmes, 2014), where submodular functions are given by weighted sums of nondecreasing concave functions applied to modular functions.…”
Section: Data Selection Approachesmentioning
confidence: 99%
“…We used two 100-dimension hidden layers with ReLU activation (Nair and Hinton, 2010) for the task. Further details of the embedding learning and classification model can be found in Cho et al (2019b).…”
Section: Paraphrase Classificationmentioning
confidence: 99%