2019 14th IEEE International Conference on Automatic Face &Amp; Gesture Recognition (FG 2019) 2019
DOI: 10.1109/fg.2019.8756577
|View full text |Cite
|
Sign up to set email alerts
|

A Multimodal LSTM for Predicting Listener Empathic Responses Over Time

Abstract: People naturally understand the emotions of-and often also empathize with-those around them. In this paper, we predict the emotional valence of an empathic listener over time as they listen to a speaker narrating a life story. We use the dataset provided by the OMG-Empathy Prediction Challenge, a workshop held in conjunction with IEEE FG 2019. We present a multimodal LSTM model with feature-level fusion and local attention that predicts empathic responses from audio, text, and visual features. Our best-perform… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0
2

Year Published

2019
2019
2022
2022

Publication Types

Select...
2
2
1

Relationship

2
3

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 17 publications
0
6
0
2
Order By: Relevance
“…Although attention in deep networks is not well understoode.g., it is still unclear under what theoretical conditions attention is useful-and is likely very different from how human attention is actually implemented in the brain, these attention mechanisms have proven to be surprisingly effective in improving deep neural network performance. Aside from a few very recent papers [17], [25], [26], there has not been much "attention" paid to these attention mechanisms within affective computing. We hope that our results will help to demonstrate the efficacy of such approaches and to encourage more research in this area.…”
Section: Discussionmentioning
confidence: 99%
See 3 more Smart Citations
“…Although attention in deep networks is not well understoode.g., it is still unclear under what theoretical conditions attention is useful-and is likely very different from how human attention is actually implemented in the brain, these attention mechanisms have proven to be surprisingly effective in improving deep neural network performance. Aside from a few very recent papers [17], [25], [26], there has not been much "attention" paid to these attention mechanisms within affective computing. We hope that our results will help to demonstrate the efficacy of such approaches and to encourage more research in this area.…”
Section: Discussionmentioning
confidence: 99%
“…As mentioned, many researchers have used LSTMs [11] to predict emotions over time [14]- [17]. In our SFT (Fig.…”
Section: Long Short-term Memory Networkmentioning
confidence: 99%
See 2 more Smart Citations
“…[39], [40] and [41] were some of the earlier papers that worked on comparing multimodal LSTMs with Support Vector Regressions and other approaches for valence and arousal classification recognition on the SEMAINE dataset. This subsequently led to a surge of interest in applying LSTMs, especially to time-series emotion recognition on the AVEC 2015 [42], [43], AVEC 2017 [44], [45], AVEC 2018 [46], and OMG-Empathy 2019 [47] challenges. Other noteworthy examples are [48], who investigated bidirectional LSTMs (where there is another recurrence that goes backwards in time), [49] who combined neural attention mechanisms with LSTMs, and [50] who built an LSTM with electroencephalography (EEG) input.…”
Section: Discriminative Modelsmentioning
confidence: 99%