2020
DOI: 10.48550/arxiv.2010.04355
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Style Attuned Pre-training and Parameter Efficient Fine-tuning for Spoken Language Understanding

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
1

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(12 citation statements)
references
References 13 publications
0
12
0
Order By: Relevance
“…Their scheme reduced the similar semantics for phonology of hardly distinguishable words with fine-tuning. Cao et al showed the pre-trained method for conversational language modeling that enabled the SLU networks to catch the linguistic representations in dialogue styles with the ASR errors [10]. Chung et al utilized a masking policy approach for the SLU task to jointly pre-train the unpaired speech and text via aligning representations [25].…”
Section: Asr-slu-based Intent Classificationmentioning
confidence: 99%
See 3 more Smart Citations
“…Their scheme reduced the similar semantics for phonology of hardly distinguishable words with fine-tuning. Cao et al showed the pre-trained method for conversational language modeling that enabled the SLU networks to catch the linguistic representations in dialogue styles with the ASR errors [10]. Chung et al utilized a masking policy approach for the SLU task to jointly pre-train the unpaired speech and text via aligning representations [25].…”
Section: Asr-slu-based Intent Classificationmentioning
confidence: 99%
“…Table 5 demonstrates the results of the proposed method and the conventional methods. Note that both Huang et al [9] and Cao et al [10] used the recognized text, and Lai et al [13] set the input as speech. For the case using recognized text from the ASR system, the proposed method outperforms other methods.…”
Section: Comparisons To Other Studiesmentioning
confidence: 99%
See 2 more Smart Citations
“…The key contributions of this paper are summarized as follows: • We introduce a semi-supervised SLU framework for learning semantics from speech to alleviate: (1) the need for a large amount of in-house, homogenous data [2,7,8,17], (2) the limitation of only intent classification [8,9,13] by predicting text, slots and intents, and (3) any additional manipulation on labels or loss, such as label projection [26], output serialization [7,18,19], ASR n-best hypothesis, or ASR-robust training losses [13,27]. Figure 2 illustrates our approach.…”
Section: Introductionmentioning
confidence: 99%