2017
DOI: 10.1109/taslp.2017.2759002
|View full text |Cite
|
Sign up to set email alerts
|

Modal and Nonmodal Voice Quality Classification Using Acoustic and Electroglottographic Features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
23
0
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 26 publications
(24 citation statements)
references
References 44 publications
0
23
0
1
Order By: Relevance
“…Therefore, diverse acoustic feature–type groups were explored (e.g., glottal, prosodic, spectral) using the COVAREP speech toolkit [ 32 ]. Previously, COVAREP features have been utilized to investigate and automatically recognize voice quality [ 33 , 34 ], respiratory [ 35 ], voice [ 36 , 37 ], and psychogenic disorders [ 38 , 39 ]. The COVAREP feature set includes 73 individual glottal, prosodic, and spectral features.…”
Section: Methodsmentioning
confidence: 99%
“…Therefore, diverse acoustic feature–type groups were explored (e.g., glottal, prosodic, spectral) using the COVAREP speech toolkit [ 32 ]. Previously, COVAREP features have been utilized to investigate and automatically recognize voice quality [ 33 , 34 ], respiratory [ 35 ], voice [ 36 , 37 ], and psychogenic disorders [ 38 , 39 ]. The COVAREP feature set includes 73 individual glottal, prosodic, and spectral features.…”
Section: Methodsmentioning
confidence: 99%
“…Objective methods involve the processing of dynamic signals obtained from sensors and the calculation of metrics using a vast array of signal processing schemes. The microphone has primarily been used in a vast majority of voice quality studies [10][11][12][13]. Although microphones are convenient, easy to use in the field and have a large bandwidth, the captured voice signals are often contaminated by background noise, and they are distorted by reverberation in the surrounding spaces.…”
Section: Introductionmentioning
confidence: 99%
“…Jitter and shimmer measures of fundamental frequency and loudness perturbations were found to correlate with voice quality based on microphone data [25]. MFCCs and Cepstral Peak Prominence (CPP) have been used to discriminate between modal, breathy, strained and other voice types based on microphone data [11,26]. The difference between the amplitudes of the first two harmonic components on microphone and NSA spectra, H1-H2, was found to be correlated with perceived speech breathiness [13,14].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Hence, alternative features for the analysis and classification of phonation types are needed. Mel-frequency cepstral coefficients (MFCCs) derived from speech signals were investigated in [29,34] for classification of phonation types in speech. Similarly the authors of [35] proposed cepstral features derived from high-resolution spectrum obtained by the zerotime windowing (ZTW) method.…”
Section: Introductionmentioning
confidence: 99%