Lecture Notes in Computer Science
DOI: 10.1007/978-3-540-71505-4_11
|View full text |Cite
|
Sign up to set email alerts
|

SVMs for Automatic Speech Recognition: A Survey

Abstract: Hidden Markov Models (HMMs) are, undoubtedly, the most employed core technique for Automatic Speech Recognition (ASR). Nevertheless, we are still far from achieving high-performance ASR systems. Some alternative approaches, most of them based on Artificial Neural Networks (ANNs), were proposed during the late eighties and early nineties. Some of them tackled the ASR problem using predictive ANNs, while others proposed hybrid HMM/ANN systems. However, despite some achievements, nowadays, the preponderance of Ma… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0
4

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 35 publications
(25 citation statements)
references
References 42 publications
0
21
0
4
Order By: Relevance
“…However, since this was a first attempt, we only investigated a limited number of feature extraction and pattern recognition methods. There are other methods available either in terms of feature extraction such as relative spectral filtering of log domain coefficients (Shrawankar and Thakare, 2010), or in terms of pattern recognition methods such as support vector machines (Solera-Urena et al, 2007) or hybrid models of HMM and ANN (Trentin and Gori, 2001) that can be examined to identify the best possible method for achieving a higher recognition rate for speech-evoked ABR signals. Also since speech-evoked ABRs are typically the responses to the single stimuli (there is no transition of states between stimuli), it could be more beneficial to use the Gaussian Mixture Model (GMM) that is a parametric probability density function, characterized as a weighted sum of Gaussian component densities (Vyas, 2013), instead of HMM that models the data as a sequence of states.…”
Section: • • • • Using Natural Stimulimentioning
confidence: 99%
“…However, since this was a first attempt, we only investigated a limited number of feature extraction and pattern recognition methods. There are other methods available either in terms of feature extraction such as relative spectral filtering of log domain coefficients (Shrawankar and Thakare, 2010), or in terms of pattern recognition methods such as support vector machines (Solera-Urena et al, 2007) or hybrid models of HMM and ANN (Trentin and Gori, 2001) that can be examined to identify the best possible method for achieving a higher recognition rate for speech-evoked ABR signals. Also since speech-evoked ABRs are typically the responses to the single stimuli (there is no transition of states between stimuli), it could be more beneficial to use the Gaussian Mixture Model (GMM) that is a parametric probability density function, characterized as a weighted sum of Gaussian component densities (Vyas, 2013), instead of HMM that models the data as a sequence of states.…”
Section: • • • • Using Natural Stimulimentioning
confidence: 99%
“…Este algoritmo fue propuesto en 1992 y es una versión no lineal de un algoritmo lineal mucho más antiguo, la regla de decisión del hiperplano óptimo, que fue introducido en los años sesenta [6]. En la ecuación 10 se muestra la fórmula general de la SVM y en [6] se encuentran los kernels más usados.…”
Section: Svmunclassified
“…En la ecuación 10 se muestra la fórmula general de la SVM y en [6] se encuentran los kernels más usados.…”
Section: Svmunclassified
“…For a more detailed review on the use of support vector machines for ASR, including systems and difficulties, refer to [33], [34].…”
Section: Hybrid Svm/hmm Systemsmentioning
confidence: 99%