3D lip-synch generation with data-faithful machine learning

Authors
Kim, Ig-JaeKo, Hyeong-Seok
Issue Date
2007-09
Publisher
WILEY
Citation
COMPUTER GRAPHICS FORUM, v.26, no.3, pp.295 - 301
Abstract
This paper proposes a new technique for generating three-dimensional speech animation. The proposed technique takes advantage of both data-driven and machine learning approaches. It seeks to utilize the most relevant part of the captured utterances for the synthesis of input phoneme sequences. If highly relevant data are missing or lacking, then it utilizes less relevant (but more abundant) data and relies more heavily on machine learning for the lip-synch generation. This hybrid approach produces results that are more faithful to real data than conventional machine learning approaches, while being better able to handle incompleteness or redundancy in the database than conventional data-driven approaches. Experimental results, obtained by applying the proposed technique to the utterance of various words and phrases, show that (1) the proposed technique generates lip-synchs of different qualities depending on the availability of the data, and (2) the new technique produces more realistic results than conventional machine learning approaches.
Keywords
speech animation; lip synch; facial animation
ISSN
0167-7055
URI
https://pubs.kist.re.kr/handle/201004/134141
DOI
10.1111/j.1467-8659.2007.01051.x
Appears in Collections:
KIST Article > 2007
Files in This Item:
There are no files associated with this item.
Export
RIS (EndNote)
XLS (Excel)
XML

qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE