ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9053169
|View full text |Cite
|
Sign up to set email alerts
|

Learning Noise Invariant Features Through Transfer Learning For Robust End-to-End Speech Recognition

Abstract: End-to-end models yield impressive speech recognition results on clean datasets while having inferior performance on noisy datasets. To address this, we propose transfer learning from a clean dataset (WSJ) to a noisy dataset (CHiME-4) for connectionist temporal classification models. We argue that the clean classifier (the upper layers of a neural network trained on clean data) can force the feature extractor (the lower layers) to learn the underlying noise invariant patterns in the noisy dataset. While traini… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3

Relationship

2
7

Authors

Journals

citations
Cited by 16 publications
(12 citation statements)
references
References 26 publications
0
12
0
Order By: Relevance
“…T/S learning was used to adapt a clean-trained E2E model to a noisy environment [112]. Noise-invariant feature was learned to improve robustness in [312,313]. Data augmentation [314] is another very effective way to expose more testing environments to E2E models during training.…”
Section: Miscellaneous Topicsmentioning
confidence: 99%
“…T/S learning was used to adapt a clean-trained E2E model to a noisy environment [112]. Noise-invariant feature was learned to improve robustness in [312,313]. Data augmentation [314] is another very effective way to expose more testing environments to E2E models during training.…”
Section: Miscellaneous Topicsmentioning
confidence: 99%
“…In this work, this simple strategy is used to show that the model can benefit from the unsupervised pre-training. There are more complex strategies for model fine-tuning or transfer learning in the literature, e.g., freezing or replacing layers, and adjusting the learning rate [25,26], which can be explored in our future work.…”
Section: Fine-tuningmentioning
confidence: 99%
“…In general, the reported experimental results do not indicate that these training methods surpass joint training. Belilovsky et al [18] and Nokland et al [19] Zhang et al [20] propose to use the classifier trained on the clean data to train the feature extractor on noisy data, so the noisy feature extractor is forced to learn features which fit the clean classifier and thus the method has the effect of denoising. In this work we show the classifier is transferable not only across different datasets but also within the same dataset.…”
Section: Related Workmentioning
confidence: 99%