2021
DOI: 10.1109/access.2020.3047395
|View full text |Cite
|
Sign up to set email alerts
|

Attention-Based Convolution Skip Bidirectional Long Short-Term Memory Network for Speech Emotion Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(5 citation statements)
references
References 75 publications
0
5
0
Order By: Relevance
“…More interestingly, activity recognition can be achieved through sensors integrated into smartphones. In [48], acceleration, gyroscope, and magnetometer data from a cyclist's smartphone were obtained through a sensor logger, then a 1D-CNN-BiLSTM model based on an attention mechanism was used to detect the cyclist's activity. Although the sensor-based gesture recognition approach can accurately recognize movements, it requires users to wear special sensors, resulting in poor user comfort.…”
Section: Wearable Sensor-based Methodsmentioning
confidence: 99%
“…More interestingly, activity recognition can be achieved through sensors integrated into smartphones. In [48], acceleration, gyroscope, and magnetometer data from a cyclist's smartphone were obtained through a sensor logger, then a 1D-CNN-BiLSTM model based on an attention mechanism was used to detect the cyclist's activity. Although the sensor-based gesture recognition approach can accurately recognize movements, it requires users to wear special sensors, resulting in poor user comfort.…”
Section: Wearable Sensor-based Methodsmentioning
confidence: 99%
“…Zhang proposed a deep-learning acoustic model which used attention mechanism. It used spatiotemporal information and captured emotion-related features more effectively [31]. Feng combines word with part of speech, position and dependency syntax separately to form 3 new combined features and proposed a novel sentiment analysis model based on multichannel convolutional neural network with multi-head attention mechanism (MCNN-MA) [32].…”
Section: Attention Mechanismmentioning
confidence: 99%
“…However, some of these depend on the sequential operation of the RNNs which increases the need for resources yet the devices are resource constrained. In order to consider long-term dependencies and context in speech for emotional state prediction, the models proposed in [14], [15], [16] and [17] use attention mechanisms in combination with either LSTM or bidirectional LSTM (BiLSTM) [18]. In [19] the bidirectional gated recurrent unit (BiGRU) is used to model the long-term dependencies for SER.…”
Section: Related Workmentioning
confidence: 99%
“…In [19] the bidirectional gated recurrent unit (BiGRU) is used to model the long-term dependencies for SER. Particularly authors of [16] and [17] improve the performance of SER systems by considering spatial cues through the use of convolution layers in combination with recurrent neural networks. However, as suggested in [20] though RNNs achieve promising results, they encounter problems in convergence, sluggish training that uses a lot of memory resources due to the sequential manner in which they operate.…”
Section: Related Workmentioning
confidence: 99%