2015 International Conference on Affective Computing and Intelligent Interaction (ACII) 2015
DOI: 10.1109/acii.2015.7344632
|View full text |Cite
|
Sign up to set email alerts
|

A multi-label convolutional neural network approach to cross-domain action unit detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
55
0
3

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 53 publications
(58 citation statements)
references
References 21 publications
0
55
0
3
Order By: Relevance
“…Performance comparison with the baseline CNN on the DISFA database [20] in terms of the average F1 score and the 2AFC score. cent approaches based on CNNs [8,9,34,28], on the two benchmark databases. As shown in Table 6, the proposed OFS-CNN achieves the state-of-the-art performance of AU recognition on the two databases 2 .…”
Section: Resultsmentioning
confidence: 99%
“…Performance comparison with the baseline CNN on the DISFA database [20] in terms of the average F1 score and the 2AFC score. cent approaches based on CNNs [8,9,34,28], on the two benchmark databases. As shown in Table 6, the proposed OFS-CNN achieves the state-of-the-art performance of AU recognition on the two databases 2 .…”
Section: Resultsmentioning
confidence: 99%
“…Spatial representation Temporal modeling AU correlation [3,23,26,27,41,43,44] × × [1,2,7,15,17,29] × × [11,12,21,22,33,35,43] × × The proposed method generative dynamic Bayesian networks (DBN) [33] was proposed to model the AU relationships and their temporal evolutions. Rather than learning, pairwise AU relations can be explicitly inferred using statistics in annotations, and then injected such relations into a multi-task learning framework to select important patches for each AU [43].…”
Section: Au Detection Methodsmentioning
confidence: 99%
“…AU-aware deep networks [21] learned representation directly from images, and then greedily picked relevant receptive fields according to a relevance measure. Ghosh et al [12] showed that a shared representation can be directly learned from input images using a multi-label CNN. However, no temporal context was involved in learning these networks.…”
Section: Au Detection Methodsmentioning
confidence: 99%
“…Given an expert-annotated label y ∈ {−1, 0, +1} L for L AUs (−1/+1 indicates absence/presence of an AU, and 0 missing label) and a prediction ŷ ∈ ℝ L , this multi-label network aims to minimize the multi-label cross entropy loss: Lfalse(boldy,y^false)=1L=1Lfalse[y>0false] log y^+false[y<0false] log false(1y^false),where [ x ] is an indicator function returning 1 if x is true, and 0 otherwise. The proposed multi-label architecture is similar to [18], which takes 40 × 40 pixel images as input. However, we used 200 × 200 pixel images in order capture more detail regarding facial texture that may aid in recognizing AUs.…”
Section: Baseline Methodsmentioning
confidence: 99%