2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00359
|View full text |Cite
|
Sign up to set email alerts
|

LAEO-Net: Revisiting People Looking at Each Other in Videos

Abstract: Figure 1: Intimacy or hostility? Head pose, along with body pose and facial expressions, is a rich source of information for interpreting human interactions. Being able to automatically understand the non-verbal cues provided by the relative head orientations of people in a scene enables a new level of human-centric video understanding. Green and red/orange heads represent LAEO and non-LAEO cases, respectively. Video source of second row: https://youtu.be/B3eFZMvNS1UAbstract Capturing the 'mutual gaze' of peop… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
35
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 52 publications
(35 citation statements)
references
References 25 publications
0
35
0
Order By: Relevance
“…However, these results highlight Fig. 7: Comparison between detections found with the model in the work by Marín-Jiménez et al [28] (left) and the Viola-Jones detector used by Patacchiola et al (right) Notice the difference in precision, making the CNN-based head detector much more suitable for this task than the Viola-Jones detector (best viewed on digital format).…”
Section: Comparison To Prior Workmentioning
confidence: 87%
See 2 more Smart Citations
“…However, these results highlight Fig. 7: Comparison between detections found with the model in the work by Marín-Jiménez et al [28] (left) and the Viola-Jones detector used by Patacchiola et al (right) Notice the difference in precision, making the CNN-based head detector much more suitable for this task than the Viola-Jones detector (best viewed on digital format).…”
Section: Comparison To Prior Workmentioning
confidence: 87%
“…For our testing procedure, we have used one of the models provided by the authors in the Hopenet repository [40] (300W-LP, alpha 1, robust to image quality); this model has been chosen as it should be the most suitable to be applied over real-world pictures as the ones appearing in the AFLW dataset. The input images correspond to the portion of the AFLW dataset used to test our model; they have been obtained using the head detector in the work by Marín-Jiménez et al [28], but they have been resized to 224 × 224; also, as the ResNet50 model uses color pictures as input, the pictures have not been converted to grayscale.…”
Section: Comparison To Prior Workmentioning
confidence: 99%
See 1 more Smart Citation
“…A convolutional neural network (CNN) single-shot detector (SSD) [80] is used for head detection [81] in the images (Figure 2). The model adopted in this research was developed by the authors of LAEO-Net [82]. The model's suitability for the task was evaluated by manually revising the head detection results on the collected dataset of 19 videos.…”
Section: Head Detectionmentioning
confidence: 99%
“…The numbers on the boxes specify the shape of the feature layers. The model adopted in this research is developed by the authors of LAEO-Net [82].…”
Section: Head Detectionmentioning
confidence: 99%