Proceedings of ISCAS'95 - International Symposium on Circuits and Systems
DOI: 10.1109/iscas.1995.521548
|View full text |Cite
|
Sign up to set email alerts
|

Lip synchronization in 3-D model based coding for video-conferencing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
9
0

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(9 citation statements)
references
References 13 publications
0
9
0
Order By: Relevance
“…SynNar is similar to the Actors system [2] in that it does not rely on any form of 3D modeling or rendering (as in [3] for example) to create the visual appearance of the ''talking head''. Rather, it uses morphing techniques [4][5][6][7] to generate a photo-realistic, smoothly-interpolated video sequence from a set of keyframes which represent typical facial positions for each phoneme in the text.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…SynNar is similar to the Actors system [2] in that it does not rely on any form of 3D modeling or rendering (as in [3] for example) to create the visual appearance of the ''talking head''. Rather, it uses morphing techniques [4][5][6][7] to generate a photo-realistic, smoothly-interpolated video sequence from a set of keyframes which represent typical facial positions for each phoneme in the text.…”
Section: Introductionmentioning
confidence: 99%
“…Provine and Bruton [3] and Waters and Levergood [8] specifically studied the various positions of major facial features (visemes or visual phonemes) during speech. Provine and Bruton's results suggest that a minimum of 20 different positions of mouth and jaw should be catered for in any facial motion synthesis system, while Waters and Levergood actually used 55 in their commercial system (DECface).…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…The resulting speech-synchronized animation is of extremely good quality if good motion capture techniques and equipment are combined with a high-quality facial model. The third type of method involves using 2D imageprocessing techniques, 13,28,29 achieving good results for speech synchronization. The character is filmed speaking a corpus that includes all the necessary phonemes or triphones (a combination of three phonemes).…”
mentioning
confidence: 99%
“…Speech recognition techniques [16] can also be used for automated segmentation. The sound track can be a speech waveform [7,22,29,30,31,33] or text [22,28,36]. If required, a waveform is then created from the phonemes.…”
Section: Previous Workmentioning
confidence: 99%