2013 IEEE International Conference on Acoustics, Speech and Signal Processing 2013
DOI: 10.1109/icassp.2013.6639145
|View full text |Cite
|
Sign up to set email alerts
|

Efficient decoding with generative score-spaces using the expectation semiring

Abstract: State-of-the-art speech recognisers are usually based on hidden Markov models (HMMs). They model a hidden symbol sequence with a Markov process, with the observations independent given that sequence. These assumptions yield efficient algorithms, but limit the power of the model. An alternative model that allows a wide range of features, including word-and phone-level features, is a log-linear model. To handle, for example, word-level variable-length features, the original feature vectors must be segmented into… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2015
2015
2016
2016

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 15 publications
0
6
0
Order By: Relevance
“…Alternatively, if GMMs are used, then derivative score spaces yields frame-level features; the derivative with respect to the component priors, for example, yields sparse GMM posterior features [44]. An interesting aspect of using structured generative models in this fashion is that feature extraction can be made efficient using an expectation semiring within the WFST framework [52]. Similar in spirit to the score-space paradigm are other methods that utilize detections of longer-term acoustic events.…”
Section: For Generative Models Adaptation To a Particular Speaker Ormentioning
confidence: 98%
See 1 more Smart Citation
“…Alternatively, if GMMs are used, then derivative score spaces yields frame-level features; the derivative with respect to the component priors, for example, yields sparse GMM posterior features [44]. An interesting aspect of using structured generative models in this fashion is that feature extraction can be made efficient using an expectation semiring within the WFST framework [52]. Similar in spirit to the score-space paradigm are other methods that utilize detections of longer-term acoustic events.…”
Section: For Generative Models Adaptation To a Particular Speaker Ormentioning
confidence: 98%
“…[ [7], [19], [20], [35] SEGMENT-LEVEL SCORE SPACES [18], [44], [51], [52] SEGMENT-LEVEL MODEL FEATURES ( , ) ( , [17], [21], [43], [56]- [58] ONE OF THE INTERESTING ASPECTS OF SUPRASEGMENTAL FEATURES IS THAT THEY CAN BE EASILY COMBINED WITH GENERATIVE MODELS FOR CLASSIFICATION. which are combined with word and pronunciation feature functions.…”
Section: Segmental Conditional Random Fieldsmentioning
confidence: 99%
“…For example, the lexicographic semiring by Shafran et al [2] is used for determinizing tagged word lattices. Also, van Dalen et al [3] report the use of expectation semiring to efficiently extract features.…”
Section: Introductionmentioning
confidence: 99%
“…Many operations in speech recognition can be elegantly described in terms of finite-state automata [1,2,3,4]. However, optimisation algorithms do not always create the desired results.…”
Section: Introductionmentioning
confidence: 99%