2008
DOI: 10.1016/j.specom.2008.04.005
|View full text |Cite
|
Sign up to set email alerts
|

The use of articulator motion information in automatic speech segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2010
2010
2014
2014

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 13 publications
(7 citation statements)
references
References 21 publications
0
7
0
Order By: Relevance
“…Although these results obtained in this paper are very encouraging, future work is required (1) to improve the recognition accuracy for shorter delay values with minimized number of user-define parameters, (2) to extend recognition and test the approach using larger datasets of more vowels, consonants, words, and even sentences, and (3) to automatically segment training data [7], [28], which is necessary when larger datasets are available in the future. …”
Section: Discussion and Future Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Although these results obtained in this paper are very encouraging, future work is required (1) to improve the recognition accuracy for shorter delay values with minimized number of user-define parameters, (2) to extend recognition and test the approach using larger datasets of more vowels, consonants, words, and even sentences, and (3) to automatically segment training data [7], [28], which is necessary when larger datasets are available in the future. …”
Section: Discussion and Future Workmentioning
confidence: 99%
“…Most published work in this domain has used only lip or facial data, so-called visual speech recognition, or automatic lip reading [8], because recording tongue motion is logistically difficult. The lip and facial data are also commonly used as an extra input source for acoustic speech recognition in so-called articulatory speech recognition [9] or audio-visual speech recognition [7], [8]. However, the tongue is a very important articulator, particularly, for vowels.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In this work, the HMM based AS system proposed in a previous study [10] was used as the first stage of segmentation. The first stage system uses the publicly available MOCHA-TIMIT database [7].…”
Section: S1mentioning
confidence: 99%
“…Results: The proposed HMM boundary refinement method was tested on the MOCHA-TIMIT database using a HMM based AS system as the first stage, [10]. At first, the method was tested on only two phoneme-to-phoneme boundaries; /y/-/uu/ boundary and /t/-(/uu/ or /o/) boundary; these boundaries have 125 and 50 occurrences in the database, respectively.…”
Section: S1mentioning
confidence: 99%