Speech Recognition and Coding 1995
DOI: 10.1007/978-3-642-57745-1_29
|View full text |Cite
|
Sign up to set email alerts
|

Search Strategies For Large-Vocabulary Continuous-Speech Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

1995
1995
2000
2000

Publication Types

Select...
4
1
1

Relationship

2
4

Authors

Journals

citations
Cited by 14 publications
(12 citation statements)
references
References 12 publications
0
12
0
Order By: Relevance
“…For a bigram language model [3], a separate copy of the lexical tree is needed for each predecessor word v. When going from a bigram to a trigram language model, we have to take into account that, for a trigram language model, the probabilities are conditioned on the previous two predecessor words (u; v) rather than one predecessor word v in the bigram case [6,7]. Therefore, we have to make the copies dependent on the two predecessor words.…”
Section: Word Conditioned Searchmentioning
confidence: 99%
See 2 more Smart Citations
“…For a bigram language model [3], a separate copy of the lexical tree is needed for each predecessor word v. When going from a bigram to a trigram language model, we have to take into account that, for a trigram language model, the probabilities are conditioned on the previous two predecessor words (u; v) rather than one predecessor word v in the bigram case [6,7]. Therefore, we have to make the copies dependent on the two predecessor words.…”
Section: Word Conditioned Searchmentioning
confidence: 99%
“…To describe the time conditioned search algorithm, we define the following quantities as introduced in [6]: h(w; ;t)= probability that word w produces the acoustic vectors x+1:::xt. H(v; ) = probability that the acoustic vectors x1:::x are generated by a word/state sequence with v as the last word and as the word boundary.…”
Section: Bigram Language Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…The first estimation method considered, here called Linear Simple (LS), assumes a simple discounting constant that can be estimated by assuming either a Poisson process for new words occurring after a given context (Witten & Bell, 1991), or by applying the LOO estimation method (Nadas, 1985;Ney et al, 1994). In both cases a good approximation of the so computed estimators yields the GT estimator for novel bigrams:…”
Section: Good-turing (Gt) Formulamentioning
confidence: 99%
“…Another advantage is that a significantly smaller amount of n-grams have to be kept in storage, as most n-grams in real texts occur once or twice. By assuming instead 0< <1 and by applying the Leaving-one-out (LOO) estimation criterion, a different solution (S ) was provided by Ney et al (1994) (see Table II). The same authors also claimed that no improvements were seen by assuming as a function of the context y.…”
Section: Good-turing (Gt) Formulamentioning
confidence: 99%