2020
DOI: 10.1038/s41598-020-58103-6
|View full text |Cite
|
Sign up to set email alerts
|

Automatic vocal tract landmark localization from midsagittal MRI data

Abstract: the various speech sounds of a language are obtained by varying the shape and position of the articulators surrounding the vocal tract. Analyzing their variations is crucial for understanding speech production, diagnosing speech disorders and planning therapy. identifying key anatomical landmarks of these structures on medical images is a pre-requisite for any quantitative analysis and the rising amount of data generated in the field calls for an automatic solution. The challenge lies in the high inter-and int… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(9 citation statements)
references
References 48 publications
0
9
0
Order By: Relevance
“…Pose estimation has recently been applied to sustained speech articulations recorded using MRI of the vocal tract [ 39 ]. A total of 256 × 256-pixel images with 1 pixel/mm resolution were analysed and RMSE accuracy results of 3.6 mm reported.…”
Section: Discussionmentioning
confidence: 99%
“…Pose estimation has recently been applied to sustained speech articulations recorded using MRI of the vocal tract [ 39 ]. A total of 256 × 256-pixel images with 1 pixel/mm resolution were analysed and RMSE accuracy results of 3.6 mm reported.…”
Section: Discussionmentioning
confidence: 99%
“…Together with the development and optimization of MR sequences, several studies have evaluated the different applications in post-imaging such as identification of "anatomical landmarks" and analysis of segmentation in different portions of the vocal tract [13]. This is done by comparing different segmentation programs and developing a protocol for the application of segmentation programs, and its correlation with the emitted sound.…”
Section: Discussionmentioning
confidence: 99%
“…With suitable physical mechanisms for the voiced excitation 31 , e.g., a reed pipe as described by Arai 32 , 33 , the 3D-printed models can be used to synthesize different vowels. in combination with other MRI or CT datasets, to study questions of morphology and anatomic development, gender differences, or inter-speaker anatomic or articulatory variability of the vocal tract 34 38 . …”
Section: Background and Summarymentioning
confidence: 99%
“…in combination with other MRI or CT datasets, to study questions of morphology and anatomic development, gender differences, or inter-speaker anatomic or articulatory variability of the vocal tract 34 38 .…”
Section: Background and Summarymentioning
confidence: 99%