A speech driven talking head system based on a single face image

I-Chen Lin, Cheng Sheng Hung, Tzong Jer Yang, Ming Ouhyoung

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

8 Scopus citations

Abstract

In this paper, a lifelike talking head system is proposed. The talking head, which is driven by speaker independent speech recognition, requires only one single face image to synthesize lifelike facial expression. The proposed system uses speech recognition engines to get utterances and corresponding time stamps in the speech data. Associated facial expressions can be fetched from an expression pool and the synthetic facial expression can then be synchronized with speech. When applied to Internet, our web-enabled talking head system can be a vivid merchandise narrator, and only requires 50 K bytes/minute with an additional face image (about 40 Kbytes in CIF format, 24 bit-color, JPEG compression). The system can synthesize facial animation more than 30 frames/sec on a Pentium II 266 MHz PC.

Original languageEnglish
Title of host publicationProceedings - 7th Pacific Conference on Computer Graphics and Applications, Pacific Graphics 1999
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages50-58
Number of pages9
ISBN (Print)0769502938, 9780769502939
DOIs
StatePublished - 1 Jan 1999
Event7th Pacific Conference on Computer Graphics and Applications, Pacific Graphics 1999 - Seoul, Korea, Republic of
Duration: 5 Oct 19997 Oct 1999

Publication series

NameProceedings - 7th Pacific Conference on Computer Graphics and Applications, Pacific Graphics 1999

Conference

Conference7th Pacific Conference on Computer Graphics and Applications, Pacific Graphics 1999
Country/TerritoryKorea, Republic of
CitySeoul
Period5/10/997/10/99

Fingerprint

Dive into the research topics of 'A speech driven talking head system based on a single face image'. Together they form a unique fingerprint.

Cite this