Interspeech 2017 2017
DOI: 10.21437/interspeech.2017-1784
|View full text |Cite
|
Sign up to set email alerts
|

Exploiting Eigenposteriors for Semi-Supervised Training of DNN Acoustic Models with Sequence Discrimination

Abstract: Deep neural network (DNN) acoustic models yield posterior probabilities of senone classes. Recent studies support the existence of low-dimensional subspaces underlying senone posteriors. Principal component analysis (PCA) is applied to identify eigenposteriors and perform low-dimensional projection of the training data posteriors. The resulted enhanced posteriors are applied as soft targets for training better DNN acoustic model under the student-teacher framework. The present work advances this approach by st… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
3
0

Year Published

2018
2018
2019
2019

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 22 publications
1
3
0
Order By: Relevance
“…The success of sparse soft targets for SDM shows that the non-linear low-dimensional modeling of senone subspaces, enabled by dictionaries, is highly beneficial for mapping reverberated noisy speech acoustic features to underlying senone classes. We also note that the performance improvements using enhanced soft targets are observed in both CE and sMBR loss based systems, and we conclude that the benefits of enhanced soft targets are complementary to those of sequence training, as shown previously in [9].…”
Section: Experimental Analysissupporting
confidence: 84%
See 3 more Smart Citations
“…The success of sparse soft targets for SDM shows that the non-linear low-dimensional modeling of senone subspaces, enabled by dictionaries, is highly beneficial for mapping reverberated noisy speech acoustic features to underlying senone classes. We also note that the performance improvements using enhanced soft targets are observed in both CE and sMBR loss based systems, and we conclude that the benefits of enhanced soft targets are complementary to those of sequence training, as shown previously in [9].…”
Section: Experimental Analysissupporting
confidence: 84%
“…In terms of CS theory [11,23], while we take measurements in a very high dimensional DNN posterior space, the actual subspace where each senone belongs is very low-dimensional. Our earlier works [8,9,10, ?] on acoustic modeling explicitly took into account this multi-subspace structure of the speech data.…”
Section: Low-rank and Sparse Soft Targetsmentioning
confidence: 99%
See 2 more Smart Citations