1996
DOI: 10.1007/978-1-4613-1367-0_3
|View full text |Cite
|
Sign up to set email alerts
|

Maximum Mutual Information Estimation of Hidden Markov Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
34
0
2

Year Published

2001
2001
2016
2016

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 42 publications
(37 citation statements)
references
References 4 publications
1
34
0
2
Order By: Relevance
“…This criterion is often referred to as mutual information criterion in speech recognition, information theory and image object recognition [2,8].…”
Section: Classification Frameworkmentioning
confidence: 99%
“…This criterion is often referred to as mutual information criterion in speech recognition, information theory and image object recognition [2,8].…”
Section: Classification Frameworkmentioning
confidence: 99%
“…As for the case of the EB algorithm, the derivatives for reestimation of the mixture weights are replaced by smoothed versions according to (Normandin, 1996).…”
Section: Gradient Descentmentioning
confidence: 99%
“…Most applications of discriminative training methods for speech recognition use either the maximum mutual information (MMI) (Bahl et al, 1986;Brown, 1987;Cardin et al, 1993;Chow, 1990;Kapadia et al, 1993;Normandin, 1996;Normandin et al, 1994a,b;Normandin and Morgera, 1991;Reichl and Ruske, 1995;Valtchev et al, 1996Valtchev et al, , 1997 or the minimum classi®cation error (MCE) (Chou et al, 1992(Chou et al, , 1993(Chou et al, , 1994Paliwal et al, 1995;Reichl and Ruske, 1995) criterion. In MCE training, an approximation to the error rate on the training data is optimized, whereas MMI training optimizes the a posteriori probability of the training utterances and hence the class separability.…”
Section: Introductionmentioning
confidence: 99%
“…The main idea behind Discriminative training (DT) is to introduce a discriminative criterion to the training method of Hidden Markov Models (HMMs). Several discriminative training methods have been proposed for ASR, such as maximum mutual information estimation (MMIE) [2,3,4], minimum classification error (MCE) [5,6,7]; and minimum word/phone error (MWE/MPE) [8,9]. For Hidden Markov (HMM) based speech recognition, conventional discriminative training criterions directly minimize the empirical risk on the training data sample and do not focus on the model generalization.…”
Section: Introductionmentioning
confidence: 99%