2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG) 2015
DOI: 10.1109/fg.2015.7163142
|View full text |Cite
|
Sign up to set email alerts
|

Dense 3D face alignment from 2D videos in real-time

Abstract: To enable real-time, person-independent 3D registration from 2D video, we developed a 3D cascade regression approach in which facial landmarks remain invariant across pose over a range of approximately 60 degrees. From a single 2D image of a person's face, a dense 3D shape is registered in real time for each frame. The algorithm utilizes a fast cascade regression framework trained on high-resolution 3D face-scans of posed and spontaneous emotion expression. The algorithm first estimates the location of a dense… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
71
1

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
3
1

Relationship

3
5

Authors

Journals

citations
Cited by 155 publications
(72 citation statements)
references
References 38 publications
0
71
1
Order By: Relevance
“…Because AAMs must be pre-trained for each participant, they are not well suited for clinical applications or large numbers of participants. We used a fully automatic, person-independent, generic approach that is comparable to AAMs to track the face and facial features, referred to as ZFace [45]. The robustness of this method for 3D registration and reconstruction from 2D video has been validated in a series of experiments (for details, see [45], [46]).…”
Section: Audiovisual Feature Extractionmentioning
confidence: 99%
See 1 more Smart Citation
“…Because AAMs must be pre-trained for each participant, they are not well suited for clinical applications or large numbers of participants. We used a fully automatic, person-independent, generic approach that is comparable to AAMs to track the face and facial features, referred to as ZFace [45]. The robustness of this method for 3D registration and reconstruction from 2D video has been validated in a series of experiments (for details, see [45], [46]).…”
Section: Audiovisual Feature Extractionmentioning
confidence: 99%
“…This is done using a combined 3D supervised descent method [47], where the shape model is defined by a 3D mesh and the 3D vertex locations of the mesh [45]. ZFace registers a dense parameterized shape model to an image such that its landmarks correspond to consistent locations on the face.…”
Section: Audiovisual Feature Extractionmentioning
confidence: 99%
“…ZFace [18], a fully person-independent, generic approach, was used to track the registered face image. For each video frame, the tracker output the 3D coordinates of 49 fiducial points and 6 degrees of freedom of rigid head movement or a failure message when a frame could not be tracked (see Figure 3).…”
Section: Methodsmentioning
confidence: 99%
“…The tracker registers a dense parameterized shape model to an image such that its landmarks correspond to consistent locations on the face. The robustness of the method for 3D registration and reconstruction from 2D video was validated in a series of experiments [for details, please see Jeni et al (2015)]. …”
Section: Automatic Tracking Of Head Orientation and Facial Landmarksmentioning
confidence: 99%
“…Occlusion, due to hands in mouth and extreme head movement, is another common problem. To overcome these challenges, we used a newly developed technique to track and align 3D features from 2D video (Jeni et al, 2015). This allowed us to ask to what extent head and facial movements communicate infants' affect.…”
mentioning
confidence: 99%