Indexed by:
Abstract:
The paper introduces a speech-driven visual speech synthesis system. The loosing-coupled-mapping scheme is proposed to establish the correspondence between the acoustic speech class and the visual speech class. Employing the data-driven method in the recorded video enables one learn the mapping scheme. To enhance the correlation between the vocal and the visual speech, the articulatory-lip-correlative-speech mode is extracted by using the genetic algorithm. The results show that the extracted feature can make the corresponding lip image class have a good clustering performance. At the synthesis phase, the serial smooth lip images are received by the searching approach in accordance with the input speech. Compared with the original video, the experiment shows that synthetic visual speech achieves a good result. Moreover, further research should be done at the synthesis phase in order to correct the jerky phenomena.
Keyword:
Reprint Author's Address:
Email:
Source :
Journal of Information and Computational Science
ISSN: 1548-7741
Year: 2006
Issue: 2
Volume: 3
Page: 227-234
Cited Count:
WoS CC Cited Count: 0
SCOPUS Cited Count:
ESI Highly Cited Papers on the List: 0 Unfold All
WanFang Cited Count:
Chinese Cited Count:
30 Days PV: 6