2022
DOI: 10.48550/arxiv.2203.13046
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Facial Action Unit Recognition With Multi-models Ensembling

Abstract: The Affective Behavior Analysis in-the-wild (ABAW) 2022 Competition gives Affective Computing a large promotion. In this paper, we present our method of AU challenge in this Competition. We use improved IResnet100 as backbone. Then we train AU dataset in Aff-Wild2 on three pertained models pretrained by our private au and expression dataset, and Glint360K respectively. Finally, we ensemble the results of our models. We achieved F1 score (macro) 0.731 on AU validation set.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 15 publications
0
4
0
Order By: Relevance
“…Training. The learning objective is the sum of the weighted BCE loss and weighted multi-label loss * proposed by Jiang et al [9] which is the second place of the 3rd ABAW challenge [12]. where W 1 and W 2 are the loss weights for each AU.…”
Section: Training and Inferencementioning
confidence: 99%
See 1 more Smart Citation
“…Training. The learning objective is the sum of the weighted BCE loss and weighted multi-label loss * proposed by Jiang et al [9] which is the second place of the 3rd ABAW challenge [12]. where W 1 and W 2 are the loss weights for each AU.…”
Section: Training and Inferencementioning
confidence: 99%
“…The model is evaluated on our validation set at the end of every epoch. Following Jiang et al [9], the loss weights of the BCE loss is…”
Section: Implementation and Training Detailsmentioning
confidence: 99%
“…In the previous editions of the ABAW competition, many teams utilized multimodal features (Kollias 2022;Jiang et al 2022;Jin et al 2021;Meng et al 2022;Zhang et al 2022b,a).The model proposed by Meng et al(Meng et al 2022) leverages both audio and visual features, ultimately achieving first place in the VA track . To fully exploit the in-the-wild emotion information, Zhang et al (Zhang et al 2022b) utilizes the multimodal information from the images, audio and text and propose a unified multimodal framework for AU detection and expression recognition.The proposed framework achieved the highest scores on both tasks.…”
Section: Related Work Multimodal Featuresmentioning
confidence: 99%
“…AU detection involves the identification and tracking of subtle facial movements that correspond to specific emotional states or expressions. The improved IResnet100 introduced by Jiang et.al [8] was utilized to address the AU detection task. And Wang et.al [27] proposed a action units correlation module to learn relationships between each AU labels and proved the effectiveness of the method.…”
Section: Au Detectionmentioning
confidence: 99%