Text-to-visual speech synthesis for general objects using parameter-based lip models

Ze Jing Chuang, Chung Hsien Wu

Research output: Chapter in Book/Report/Conference proceedingConference contribution

2 Citations (Scopus)

Abstract

This paper presents four parameter-based 3-dimension (3D) lip models for Chinese text-to-visual speech synthesis. This model can be applied to general objects with lip-like meshes. Three main components will be described in this paper: the generation of weighted parameter sequence of lip motions for each Mandarin syllable, the definition and construction of parameter-based lip models, and the synchronization of speech and facial animation. The result shows that the system produces a promising and encouraging speech and facial animation output.

Original languageEnglish
Title of host publicationAdvances in Multimedia Information Processing - PCM 2002 - 3rd IEEE Pacific Rim Conference on Multimedia, Proceedings
EditorsYung-Chang Chen, Long-Wen Chang, Chiou-Ting Hsu
PublisherSpringer Verlag
Pages589-597
Number of pages9
ISBN (Print)3540002626, 9783540002628
DOIs
Publication statusPublished - 2002
Event3rd IEEE Pacific Rim Conference on Multimedia, PCM 2002 - Hsinchu, Taiwan
Duration: 2002 Dec 162002 Dec 18

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume2532
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Other

Other3rd IEEE Pacific Rim Conference on Multimedia, PCM 2002
CountryTaiwan
CityHsinchu
Period02-12-1602-12-18

All Science Journal Classification (ASJC) codes

  • Theoretical Computer Science
  • Computer Science(all)

Fingerprint Dive into the research topics of 'Text-to-visual speech synthesis for general objects using parameter-based lip models'. Together they form a unique fingerprint.

Cite this