2020
DOI: 10.1007/s11263-020-01309-y
|View full text |Cite
|
Sign up to set email alerts
|

CR-Net: A Deep Classification-Regression Network for Multimodal Apparent Personality Analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
30
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 37 publications
(30 citation statements)
references
References 40 publications
0
30
0
Order By: Relevance
“…Then, a consensus strategy is employed to process all the selected frames, to produce video-level personality predictions. Li et al [29] also select a face image and a face-background image from each segment and stacks them as the clip-level stream. This approach converts the acoustic wave of an entire clip to fixedlength vectors as the clip-level audio representation.…”
Section: Related Work 21 Automatic Audio-visual Personality Analysismentioning
confidence: 99%
See 3 more Smart Citations
“…Then, a consensus strategy is employed to process all the selected frames, to produce video-level personality predictions. Li et al [29] also select a face image and a face-background image from each segment and stacks them as the clip-level stream. This approach converts the acoustic wave of an entire clip to fixedlength vectors as the clip-level audio representation.…”
Section: Related Work 21 Automatic Audio-visual Personality Analysismentioning
confidence: 99%
“…Each ERN used in our experiments is made up of two Methods Ope Con Ext Agr Neu Avg. ACC Spectral [45] 0.752 0.807 0.849 0.800 0.788 0.799 DCC [18] 0.755 0.787 0.772 0.736 0.791 0.768 NJU-LAMDA [51] 0.741 0.826 0.827 0.753 0.789 0.787 CR-Net [29] 0.830 0.876 0.904 0.887 0.903 0.880 PALs [44] 0.845 0.819 0.916 0.837 0.911 0.866 Ours (A-MModal (S)) 0.833 0.890 0.913 0.869 0.917 0.884 Ours (MModal (M)) 0.889 0.925 0.923 0.913 0.921 0.914 Ours (A-MModal (M)) 0.882 0.925 0.931 0.912 0.925 0.915 PCC Spectral [45] -0.010 0.059 0.135 0.071 0.024 0.056 DCC [18] -0.153 -0.078 0.037 -0.024 0.121 0.008 NJU-LAMDA [51] MModal denotes the graph representations of multi-modal processors. (M) and (S) represent the multi-level and singlelevel fusion, respectively.…”
Section: Implementation Detailsmentioning
confidence: 99%
See 2 more Smart Citations
“…Recent advances in machine learning (ML) have enabled the development of non-invasive automatic personality traits analysers that recognise subjects' personality traits from their audiovisual non-verbal behaviours [16,28,52,80,90,98] as there is solid psychological and biological evidence [19,27,48,95] claiming that nonverbal behaviours are reliable predictors of personality. In most of these approaches, ML models are trained with the personality labels provided by the external observers (annotators), and they therefore output their perception of the target subjects' personality.…”
Section: Introductionmentioning
confidence: 99%