2022
DOI: 10.1111/cgf.14640
|View full text |Cite
|
Sign up to set email alerts
|

Voice2Face: Audio‐driven Facial and Tongue Rig Animations with cVAEs

Abstract: Figure 1: System overview. During training, two cVAEs are used to encode and generate facial and tongue mesh animations conditioned on speech. During inference, fixed latent vectors are used by the decoders to generate mesh animation sequences, that are then transformed into rig space via models approximating the inverse rig function.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
references
References 35 publications
0
0
0
Order By: Relevance