2022 IEEE Region 10 Symposium (TENSYMP) 2022
DOI: 10.1109/tensymp54529.2022.9864557
|View full text |Cite
|
Sign up to set email alerts
|

Detecting emotions from human speech: role of gender information

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(12 citation statements)
references
References 22 publications
0
12
0
Order By: Relevance
“…This technique involves generating new data from existing databases through slight modifications. Standard data augmentation techniques within the SER field include adding noise, stretching, altering pitch, and shifting [10], [12], [13].…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…This technique involves generating new data from existing databases through slight modifications. Standard data augmentation techniques within the SER field include adding noise, stretching, altering pitch, and shifting [10], [12], [13].…”
Section: Related Workmentioning
confidence: 99%
“…There are various methods for extracting audio characteristics from databases. The literature has employed techniques such as Mel-Frequency Cepstral Coefficients (MFCC) [13], [15], Zero Crossing Rate (ZCR) [3], Chromagram [9], Mel Spectrogram [10], and Root Mean Square (RMS) values [12]. The choice of feature extraction technique directly influences the classifier's performance and the resulting outcomes.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…In the early days of SER research, the primary focus was on probabilistic models such as Hidden Markov Models (HMMs) [6,7] and Gaussian Mixture Models (GMMs) [8,9]. Recently, with the advent of deep learning, the landscape of emotion recognition has significantly shifted towards neural-network-based approaches [4,10]. Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Deep Neural Networks (DNNs) now play a predominant role in advancing speech emotion recognition [11].…”
Section: Introductionmentioning
confidence: 99%