2019
DOI: 10.1016/j.ijmedinf.2019.05.023
|View full text |Cite
|
Sign up to set email alerts
|

HypernasalityNet: Deep recurrent neural network for automatic hypernasality detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
39
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 27 publications
(40 citation statements)
references
References 41 publications
1
39
0
Order By: Relevance
“…The maximum use of AI was to study hypernasality. Three studies tried to detect its presence ( 40 , 41 , 43 ) and two classified it according to severity ( 44 , 48 ). They used extracted features of speech as inputs for the classifiers.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…The maximum use of AI was to study hypernasality. Three studies tried to detect its presence ( 40 , 41 , 43 ) and two classified it according to severity ( 44 , 48 ). They used extracted features of speech as inputs for the classifiers.…”
Section: Resultsmentioning
confidence: 99%
“…These provide encouraging experimental results, but they do not reflect the practical performance of these systems. They characterize speech based on hypernasality ( 40 , 41 , 43 , 44 , 46 , 48 ), identifying phonemes ( 42 , 47 ) and misarticulations ( 39 , 45 ). These studies use features of speech data at a single point in time.…”
Section: Discussionmentioning
confidence: 99%
“…More recently, hypernasality evaluation algorithms rely on machine learning; frame-level features such as MFCCs are extracted from the segmented regions to train classifiers like support vector machine and Gaussian mixture models [7,8,9]. In similar vein, convolutional neural networks and recurrent neural networks have also been used to detect hypernasality [10,11].…”
Section: Introductionmentioning
confidence: 99%
“…In the past there have been many proposals to evaluate HN based on acoustic information. While the results in terms of accuracy are generally excellent, most studies have used only a limited number of utterance types, such as sustained vowels [6][7][8][9][10][11]. This is not clearly compatible with standard clinical recommendations [2], which strongly recommend that patients are evaluated using a variety of phonemes and utterances with varying complexity (e.g., CAPS-A protocol, [2,12]).…”
Section: Introductionmentioning
confidence: 99%
“…Classification algorithms include Random Forest (RF), Support Vector Machine (SVM) or Artificial Neural Network (ANN). Acoustic features used in previous studies include, among others, Mel Frequency Cepstrum Coefficients (MFCCs), the Voice Low Tone to High Tone Ratio (VLTHTR) and the vowel formants and their bandwidth [6][7][8][9][10][11]. In most studies using this approach the fragments analyzed were either sustained vowels [6,8] or vowel fragments that had been annotated manually in words or sentences [7,[9][10][11].…”
Section: Introductionmentioning
confidence: 99%