ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019
DOI: 10.1109/icassp.2019.8683086
|View full text |Cite
|
Sign up to set email alerts
|

Analyzing Uncertainties in Speech Recognition Using Dropout

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(10 citation statements)
references
References 16 publications
0
10
0
Order By: Relevance
“…Here, a stochastic pass refers to inference with a dropout realization. This technique is closely related to [24], which uses a model's prediction uncertainty computed using dropout to estimate word error rates. DUST combines ST and pseudo-label filtering based on the ASR model's uncertainty for an unlabeled speech utterance using dropout.…”
Section: Dustmentioning
confidence: 99%
See 1 more Smart Citation
“…Here, a stochastic pass refers to inference with a dropout realization. This technique is closely related to [24], which uses a model's prediction uncertainty computed using dropout to estimate word error rates. DUST combines ST and pseudo-label filtering based on the ASR model's uncertainty for an unlabeled speech utterance using dropout.…”
Section: Dustmentioning
confidence: 99%
“…In this case, the pseudo-labels generated by the teacher model for the unlabeled target domain data may be less accurate, which increases the need to apply a pseudo-label filtering strategy. To that end, we propose dropout-based uncertaintydriven self-training (DUST), which filters pseudo-labeled data based on the model's uncertainty about its prediction as measured using the degree of agreement between multiple transcriptions obtained with various realizations of dropout and a reference transcription obtained without dropout [23,24]. We show that DUST is an effective method for mismatched domain adaptation and substantially improves over the baseline model, which is trained on the source domain labeled data only, as well as over iterative ST without filtering [19], whereby the largest gain is observed when the source and target domain mismatch is most severe.…”
Section: Introductionmentioning
confidence: 99%
“…Furthermore, they employ ensembles as their primary method of confidence estimation, while we also evaluate temperature scaling and dropout methods. Dropout was previously used for obtaining confidence scores for ASR [34], but our approaches differ: in [34] multiple hypotheses are generated via dropout and then word confidences are assigned based on their frequency of appearance in the aligned hypothe-ses; in contrast, we aggregate the posterior probabilities and not the hypotheses, which simplifies the procedure as it avoids the alignment step.…”
Section: Related Workmentioning
confidence: 99%
“…We also propose to localize the uncertainty of the end-toend ASR by applying dropout mechanism. This method is motivated by the recent advances of DNN for measuring the reliability of the model [14,15]. The conventional methods do not use dropouts during the decoding time.…”
Section: Semi Supervised Learningmentioning
confidence: 99%