2019
DOI: 10.1609/aaai.v33i01.33013215
|View full text |Cite
|
Sign up to set email alerts
|

Mode Variational LSTM Robust to Unseen Modes of Variation: Application to Facial Expression Recognition

Abstract: Spatio-temporal feature encoding is essential for encoding the dynamics in video sequences. Recurrent neural networks, particularly long short-term memory (LSTM) units, have been popular as an efficient tool for encoding spatio-temporal features in sequences. In this work, we investigate the effect of mode variations on the encoded spatio-temporal features using LSTMs. We show that the LSTM retains information related to the mode variation in the sequence, which is irrelevant to the task at hand (e.g. classifi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
26
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 25 publications
(30 citation statements)
references
References 30 publications
0
26
0
Order By: Relevance
“…The average classification accuracy obtained in the experiments on the CK+, AFEW, and Oulu-CASIA datasets is shown in Tables 2-4, respectively. Table 2 shows the accuracy of our proposed method and state-of-the-art methods [14,18,24,31,33] on the AFEW dataset. On the AFEW dataset, the data come from the natural environment, which is restricted by head deflection, illumination, and blur.…”
Section: Experimental Results and Analysismentioning
confidence: 99%
See 2 more Smart Citations
“…The average classification accuracy obtained in the experiments on the CK+, AFEW, and Oulu-CASIA datasets is shown in Tables 2-4, respectively. Table 2 shows the accuracy of our proposed method and state-of-the-art methods [14,18,24,31,33] on the AFEW dataset. On the AFEW dataset, the data come from the natural environment, which is restricted by head deflection, illumination, and blur.…”
Section: Experimental Results and Analysismentioning
confidence: 99%
“…Mode variational LSTM [31] 48.83% MRAN [14] 49.01% CNN-BLSTM [33] 49.09% FAN [18] 51.18% DenseNet-161 [24] 51.40%…”
Section: Methods Accuracymentioning
confidence: 99%
See 1 more Smart Citation
“…Generally in a proportion (p, 1 − p) : p ∈ [0, 1] of 80 − 20%, such as is displayed in table 2. It corresponds to a common approach found in the recurrent neural network literature [2], [3], [36], which provides a set of independent and identical distributed training samples. A property that it is suitable in the analysis of individual model results.…”
Section: Methodsmentioning
confidence: 99%
“…forcing the model to retain irrelevant information [2]. In this regard, a myriad number of hybrid models in recent literature combining LSTM with other procedures manages to reduce the impact of the noise on the prediction of their model accuracy.…”
Section: Introductionmentioning
confidence: 99%