2019
DOI: 10.1109/access.2019.2940961
|View full text |Cite
|
Sign up to set email alerts
|

Large-Scale Semi-Supervised Training in Deep Learning Acoustic Model for ASR

Abstract: This study investigated large-scale semi-supervised training (SST) to improve acoustic models for automatic speech recognition. The conventional self-training, the recently proposed committee-based SST using heterogeneous neural networks and the lattice-based SST were examined and compared. The large-scale SST was studied in deep neural network acoustic modeling with respect to the automatic transcription quality, the importance data filtering, the training data quantity and other data attributes of a large qu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0
2

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 20 publications
(7 citation statements)
references
References 37 publications
(67 reference statements)
0
5
0
2
Order By: Relevance
“…Researchers have explored different methods of semi-supervised training for speech recognition. Long et al (2019) investigate large-scale semisupervised training to improve acoustic models for automatic speech recognition. They provide an empirical analysis of semi-supervised training with respect to transcription quality, data quality, filtering, etc.…”
Section: Related Workmentioning
confidence: 99%
“…Researchers have explored different methods of semi-supervised training for speech recognition. Long et al (2019) investigate large-scale semisupervised training to improve acoustic models for automatic speech recognition. They provide an empirical analysis of semi-supervised training with respect to transcription quality, data quality, filtering, etc.…”
Section: Related Workmentioning
confidence: 99%
“…One stream of works applies unsupervised/semi-supervised methods to tackle the problem. For example, Long et al [21] propose an improved self-training approach for semi-supervised training of DNN and RNN based acoustic models. Karita et al [22] jointly learn ASR and TTS models for semi-supervised training of ASR models.…”
Section: Related Workmentioning
confidence: 99%
“…In this work, we focus on this setting to explore how much gain can be made without undertaking the expensive and tedious task of annotating accented speech. Several semi-supervised learning techniques have been proposed to tackle this scenario by combining annotated and unannotated examples to jointly learn robust inference models [5][6][7]. We propose to leverage domain adversarial training * Work conducted as an intern at Amazon AWS AI.…”
Section: Introductionmentioning
confidence: 99%