ICASSP '80. IEEE International Conference on Acoustics, Speech, and Signal Processing
DOI: 10.1109/icassp.1980.1170862
|View full text |Cite
|
Sign up to set email alerts
|

Further results on the recognition of a continuously read natural corpus

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
26
0

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 45 publications
(26 citation statements)
references
References 4 publications
0
26
0
Order By: Relevance
“…Third, the acoustic model, implemented by a Deep Neural Network (DNN), transforms the MFCC features into phonemes' probabilities. Context-sensitive phonemes are the norm, triphones [18] being the most common approach. A triphone is a particular phoneme when combined with a particular predecessor and a particular successor.…”
Section: Speech Recognition With Wfstmentioning
confidence: 99%
“…Third, the acoustic model, implemented by a Deep Neural Network (DNN), transforms the MFCC features into phonemes' probabilities. Context-sensitive phonemes are the norm, triphones [18] being the most common approach. A triphone is a particular phoneme when combined with a particular predecessor and a particular successor.…”
Section: Speech Recognition With Wfstmentioning
confidence: 99%
“…Some of the possibilities include sub-phoneme units, phones with right or left context, biphones, diphones [2] and variations [3], dyads or transemes [4], avents [5], triphones [6],demisyllables [7], whole words and phrases. Current research in psychoacoustics and psycholinguistics suggest that the syllable might be a basic unit of human speech perception.…”
Section: Introductionmentioning
confidence: 99%
“…Both methods are based on the computation of emission probabilities directly on the training data during the forward-backward training algorithm. The first method characterizes the difference between two distributions by means of a divergence criteria derived from the one proposed in [4]: (1) where is the kth training occurrence of model , is the isolated emission probability of by model , and is the number of training occurrences of model . The underlying idea of our method is that a model is considered to be well trained if there is a sufficient number of examples for this model.…”
Section: The Clustering Techniquesmentioning
confidence: 99%
“…(1). To guide the clustering and to avoid having too few/too many clusters per class, the final number of clusters per class is constrained between two thresholds: Nmin and Nmax.…”
Section: Context Modeling and Clustering In Continuous Speech Recognimentioning
confidence: 99%
See 1 more Smart Citation