2004 IEEE International Conference on Acoustics, Speech, and Signal Processing
DOI: 10.1109/icassp.2004.1327060
|View full text |Cite
|
Sign up to set email alerts
|

On a practical design of a low complexity speech recognition engine

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 7 publications
0
4
0
Order By: Relevance
“…In Ortmanns et al (1997) the k-d tree and PSA techniques reduced the CPU-time required for likelihoods computations respectively down to 50.7% and to 39.9% compared to the baseline system without increase in WER. Vasilache et al (2004) present a phoneme based ASR, where HMM parameters μ and σ are represented by one out of 2 5 (for mean) and one out of 2 3 (for variance) quantizer levels. If in addition the input features will be quantized with a similar number of levels 2 5 , then the likelihoods can take only one out of 2 5 2 3 2 5 = 8192 fixed values, which can be precomputed and stored.…”
Section: Likelihood Computation Methodsmentioning
confidence: 99%
“…In Ortmanns et al (1997) the k-d tree and PSA techniques reduced the CPU-time required for likelihoods computations respectively down to 50.7% and to 39.9% compared to the baseline system without increase in WER. Vasilache et al (2004) present a phoneme based ASR, where HMM parameters μ and σ are represented by one out of 2 5 (for mean) and one out of 2 3 (for variance) quantizer levels. If in addition the input features will be quantized with a similar number of levels 2 5 , then the likelihoods can take only one out of 2 5 2 3 2 5 = 8192 fixed values, which can be precomputed and stored.…”
Section: Likelihood Computation Methodsmentioning
confidence: 99%
“…The exponent can be extracted by simple bit masking and shifting and already represents a coarse approximation of log 2 f . It can be turned into log b f to any base b by multiplication with log b 2 as shown in equation (5). The error is always lower than log b 2, but can be improved by additionally approximating the mantissa m. By using a polynomial of first (6) or second (7) degree the computation is reduced to a few multiplications and additions:…”
Section: Approximation Of Logarithmmentioning
confidence: 99%
“…Especially the evaluation of the Gaussian Mixtures uses a lot of floating point operations for computing the Mahalanobis distances. The use of look-up tables can avoid these computations [5] either in part or completely. But on a PDA the necessary additional memory access is more expensive than doing the computation by means of integer arithmetic.…”
Section: Optimization Of Decodingmentioning
confidence: 99%
See 1 more Smart Citation