TY - GEN
T1 - Co-articulation generation using maximum direction change and apparent motion for Chinese visual speech synthesis
AU - Wu, Chung-Hsien
AU - Lee, Chung Han
AU - Chuang, Ze Jing
PY - 2010/12/1
Y1 - 2010/12/1
N2 - This study presents an approach for automated lip synchronization and smoothing for Chinese visual speech synthesis. A facial animation system with synchronization algorithm is also developed to visualize an existent Text-To-Speech system. Motion parameters for each viseme are first constructed from video footage of a human speaker. To synchronize the parameter set sequence and speech signal, a maximum direction change algorithm is also proposed to select significant parameter set sequences according to the speech duration. Moreover, to improve the smoothness of co-articulation part under a high speaking rate, four phoneme-dependent co-articulation functions are generated by integrating the Bernstein-Bézier curve and apparent motion property. A Chinese visual speech synthesis system is built to evaluate the proposed approach. The synthesis result of the proposed system is compared to the real speaker. The coarticulation generated by the proposed approach is also evaluated.
AB - This study presents an approach for automated lip synchronization and smoothing for Chinese visual speech synthesis. A facial animation system with synchronization algorithm is also developed to visualize an existent Text-To-Speech system. Motion parameters for each viseme are first constructed from video footage of a human speaker. To synchronize the parameter set sequence and speech signal, a maximum direction change algorithm is also proposed to select significant parameter set sequences according to the speech duration. Moreover, to improve the smoothness of co-articulation part under a high speaking rate, four phoneme-dependent co-articulation functions are generated by integrating the Bernstein-Bézier curve and apparent motion property. A Chinese visual speech synthesis system is built to evaluate the proposed approach. The synthesis result of the proposed system is compared to the real speaker. The coarticulation generated by the proposed approach is also evaluated.
UR - http://www.scopus.com/inward/record.url?scp=79851491711&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=79851491711&partnerID=8YFLogxK
U2 - 10.1109/COMPSYM.2010.5685459
DO - 10.1109/COMPSYM.2010.5685459
M3 - Conference contribution
AN - SCOPUS:79851491711
SN - 9781424476404
T3 - ICS 2010 - International Computer Symposium
SP - 512
EP - 517
BT - ICS 2010 - International Computer Symposium
T2 - 2010 International Computer Symposium, ICS 2010
Y2 - 16 December 2010 through 18 December 2010
ER -