Share Email Print

Proceedings Paper

Real-time facial action image synthesis system driven by speech and text
Author(s): Shigeo Morishima; Kiyoharu Aizawa; Hiroshi Harashima
Format Member Price Non-Member Price
PDF $17.00 $21.00

Paper Abstract

Automatic facial motion image synthesis schemes and a real-time system design are presented. The purpose of this schemes is to realize an intelligent human-machine interface or intelligent communication system with talking head images. Human's face is reconstructed with 3D surface model and texture mapping technique on the display of terminal. Facial motion images are synthesized naturally by transformation of the lattice points on wire frames. Two types of motion drive methods, text to image conversion and speech to image conversion are proposed in this paper. In the former manner, synthesized head can speak some given texts naturally and in the latter case, some mouth and jaw motions can be synthesized in time to speech signal of behind speaker. These schemes were implemented to a parallel image computer and a real-time image synthesizer could output facial motion images to the display as fast as video rate.

Paper Details

Date Published: 1 September 1990
PDF: 9 pages
Proc. SPIE 1360, Visual Communications and Image Processing '90: Fifth in a Series, (1 September 1990); doi: 10.1117/12.24130
Show Author Affiliations
Shigeo Morishima, Seikei Univ. (Japan)
Kiyoharu Aizawa, Univ. of Tokyo (Japan)
Hiroshi Harashima, Univ. of Tokyo (Japan)

Published in SPIE Proceedings Vol. 1360:
Visual Communications and Image Processing '90: Fifth in a Series
Murat Kunt, Editor(s)

© SPIE. Terms of Use
Back to Top
Sign in to read the full article
Create a free SPIE account to get access to
premium articles and original research
Forgot your username?