2013
DOI: 10.1145/2461912.2462019
|View full text |Cite
|
Sign up to set email alerts
|

Realtime facial animation with on-the-fly correctives

Abstract: Figure 1: Our adaptive tracking model conforms to the input expressions on-the-fly, producing a better fit to the user than state-of-the-art data driven techniques [Weise et al. 2011] which are confined to learned motion priors and generate plausible but not accurate tracking. AbstractWe introduce a real-time and calibration-free facial performance capture framework based on a sensor with video and depth input. In this framework, we develop an adaptive PCA model using shape correctives that adjust on-the-fly t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
209
0

Year Published

2014
2014
2020
2020

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 262 publications
(209 citation statements)
references
References 46 publications
0
209
0
Order By: Relevance
“…We detect dynamic occlusions caused by temporal shape and texture variations using an outlier voting scheme in superpixel space. As recently demonstrated by Li et al [31], the combination of sparse 2D facial features (e.g., eyes, eyebrows, and mouth) with dense depth maps are particularly effective in improving tracking fidelity. However, because facial landmark detection becomes significantly less reliable when the face is occluded, we synthesize plausible face textures right after our face segmentation step.…”
Section: Introductionmentioning
confidence: 89%
See 2 more Smart Citations
“…We detect dynamic occlusions caused by temporal shape and texture variations using an outlier voting scheme in superpixel space. As recently demonstrated by Li et al [31], the combination of sparse 2D facial features (e.g., eyes, eyebrows, and mouth) with dense depth maps are particularly effective in improving tracking fidelity. However, because facial landmark detection becomes significantly less reliable when the face is occluded, we synthesize plausible face textures right after our face segmentation step.…”
Section: Introductionmentioning
confidence: 89%
“…The data-driven method of Weise et al [39] uses a motion prior database to handle noise and the lowresolution depth maps from the Kinect sensor. For improved fidelity, techniques that combine depth input data with sparse facial features were introduced [31,11,13,9]. To improve accessibility with less input training, an example-based facial rigging method was introduced by Li et al [30].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…These techniques, however, cannot handle large deformations, and are not very practical for real-time applications. Real-time non-rigid reconstruction approaches have been achieved with the help of a template which is first acquired then used for tracking of non-rigidities with a good flexibility [40], [41]. Recently, we have proposed KinectDeform [11], the first nonrigid version of KinectFusion.…”
Section: Dynamic Multi-frame Approachesmentioning
confidence: 99%
“…The application-specific nature of these approaches enables their authors to show excellent performance by taking advantage of domain-specific features and constraints, but it also prevents them from serving as general tools for tracking arbitrary articulated objects. Techniques have also been developed to track fully non-rigid deformations of an underlying surface template for both specific [21] and general [14,20,28] object cases. However, the full generality of these models comes at the cost of increased model complexity, and for many objects that are well modelled as piecewise rigid bodies, such overparameterized output obscures the utility of tracking the articulated body state directly.…”
Section: Introductionmentioning
confidence: 99%