2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00364
|View full text |Cite
|
Sign up to set email alerts
|

Thin-Plate Spline Motion Model for Image Animation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
37
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 99 publications
(37 citation statements)
references
References 22 publications
0
37
0
Order By: Relevance
“…A key focus area of such works is to design appropriate motion representations for animation [35,52,55,66]. A number of improved representations have been proposed, such as those setting additional constraints on a kinematic tree [59], and thin-plate spline motion modelling [81]. A further work, titled Latent Image Animator [64], learned a latent space for possible motions.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…A key focus area of such works is to design appropriate motion representations for animation [35,52,55,66]. A number of improved representations have been proposed, such as those setting additional constraints on a kinematic tree [59], and thin-plate spline motion modelling [81]. A further work, titled Latent Image Animator [64], learned a latent space for possible motions.…”
Section: Related Workmentioning
confidence: 99%
“…Initially, such transformations were modeled using a simple set of sparse keypoints. Further works improved the motion representation [52,55], learned latent motion dictionaries [64], kinematic chains [59] or used thin-plate spline transformations [81]. However, broadly speaking, all such works propose 2D motion representations, warping the pixels or features of the input image such that they correspond to the pose of a given driving image.…”
Section: Introductionmentioning
confidence: 99%
“…MetaPortrait [6] introduces an ID-preserving talking head generation framework that leverages dense landmarks for accurate geometry-aware flow fields and adaptively fuses source identity during synthesis for better preservation of key characteristics. Besides these third-party model-based methods, some video-driven talking head generation methods [9], [10], [11], [12], [44] attempt to learn keypoints of the human face to represent the facial expression in a self-supervised manner. FOMM [9] introduces a self-supervised image animation framework that decouples appearance and motion information, and computes the motion between two faces by using their keypoints.…”
Section: Talking Head Synthesismentioning
confidence: 99%
“…Driving FOMM [9] DaGAN [12] TPSM [44] Ours Fig. 6: Qualitative comparisons on the self-reenactment experiment on the VoxCeleb1 dataset [56].…”
Section: Sourcementioning
confidence: 99%
See 1 more Smart Citation