2017 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU) 2017
DOI: 10.1109/asru.2017.8268947
|View full text |Cite
|
Sign up to set email alerts
|

Investigation of transfer learning for ASR using LF-MMI trained neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
44
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 61 publications
(45 citation statements)
references
References 21 publications
1
44
0
Order By: Relevance
“…This is particularly important, since different domains might have different amounts of data. In previous work [31], Ghahremani et al recommend that gradients of utterances from a particular domain should be scaled by the inverse of the square root of the number of utterances in the domain, thus effectively over-sampling domains with less data. In Tab.…”
Section: Multidomain Training: Impact Of Data Diversitymentioning
confidence: 99%
“…This is particularly important, since different domains might have different amounts of data. In previous work [31], Ghahremani et al recommend that gradients of utterances from a particular domain should be scaled by the inverse of the square root of the number of utterances in the domain, thus effectively over-sampling domains with less data. In Tab.…”
Section: Multidomain Training: Impact Of Data Diversitymentioning
confidence: 99%
“…However, because speech signals are high-dimensional and highly variable even for a single speaker, training deep models and learning these hierarchical representations without a large amount of training data is difficult. The computer vision [15,16], natural language processing [17][18][19][20][21], and ASR [22][23][24][25] communities have attacked the problem of limited supervised training data with great success by pre-training deep models on related tasks for which there is more training data. Following their lead, we propose an efficient ASR-based pre-training methodology in this paper and show that it may be used to improve the performance of end-toend SLU models, especially when the amount of training data is very small.…”
Section: Introductionmentioning
confidence: 99%
“…They also show that pooling data from multiple low-resource domains work better than transfer learning. Unlike [7], the current work studies domain robustness in a much larger scale, where data sparsity is not necessarily a challenge. We also study other forms of mismatch like codec, and consider many more applications domains.…”
Section: Prior Workmentioning
confidence: 99%