2020
DOI: 10.1609/aaai.v34i02.5491
|View full text |Cite
|
Sign up to set email alerts
|

Synch-Graph: Multisensory Emotion Recognition Through Neural Synchrony via Graph Convolutional Networks

Abstract: Human emotions are essentially multisensory, where emotional states are conveyed through multiple modalities such as facial expression, body language, and non-verbal and verbal signals. Therefore having multimodal or multisensory learning is crucial for recognising emotions and interpreting social signals. Existing multisensory emotion recognition approaches focus on extracting features on each modality, while ignoring the importance of constant interaction and co-learning between modalities. In this paper, we… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(11 citation statements)
references
References 26 publications
1
10
0
Order By: Relevance
“…After this replacement, the accuracy of the video classifier is 56.8%. This is in line with state-of-the-art results in the literature on emotion recognition from RAVDESS videos, namely 57.5% with Synchronous Graph Neural Networks (8 emotions) [50]; 61% with ConvNet-LSTM (8 emotions) [1]; 59% with an RNN (7 emotions) [9], and 82.4% with stacked autoencoders (6 emotions) [5].…”
Section: A Dataset and Model Architecturesupporting
confidence: 88%
See 1 more Smart Citation
“…After this replacement, the accuracy of the video classifier is 56.8%. This is in line with state-of-the-art results in the literature on emotion recognition from RAVDESS videos, namely 57.5% with Synchronous Graph Neural Networks (8 emotions) [50]; 61% with ConvNet-LSTM (8 emotions) [1]; 59% with an RNN (7 emotions) [9], and 82.4% with stacked autoencoders (6 emotions) [5].…”
Section: A Dataset and Model Architecturesupporting
confidence: 88%
“…V, has been studied by other authors in-the-clear, i.e. without regards for privacy protection, using a variety of deep learning architectures, with reported accuracies in the 57%-82% range, depending on the number of emotion classes included in the study (6 to 8) [5], [50], [9], [1]. The ConvNet model that we trained for our experimental results in Sec.…”
Section: Related Workmentioning
confidence: 99%
“…There are other components , such as Featureto-Label Level, Label-to-Label Level and Modality-to-Label Level, which work to pass messages among different type of nodes. Benssassi et al [101] synthesis the representation of neural synchrony graph from facial and speech features that are extracted via spiking neural networks(SNN). The neural synchrony graph is deemed as the input of Graph Convolutional Network(GCN) for emotion recognition.…”
Section: Multimodalitymentioning
confidence: 99%
“…There are other components , such as Feature-to-Label Level, Labelto-Label Level and Modality-to-Label Level, which work to pass messages among different type of nodes. Benssassi et al [98] synthesis the representation of neural synchrony graph from facial and speech features that are extracted via spiking neural networks(SNN). The neural synchrony graph is deemed as the input of Graph Convolutional Network(GCN) for emotion recognition.…”
Section: Multimodalitymentioning
confidence: 99%