Please use this identifier to cite or link to this item: http://drsr.daiict.ac.in//handle/123456789/121
Title: Speech driven facial animation system
Authors: Jotwani, Naresh D.
Singh, Archana
Keywords: Co-articulation
Facial animation system
Frame
Gaussian mixture model
Hidden markov model
Speech recognition
Vector quantization
Viseme
Viterbi Algorithm
Issue Date: 2006
Publisher: Dhirubhai Ambani Institute of Information and Communication Technology
Citation: Singh, Archana (2006). Speech driven facial animation system. Dhirubhai Ambani Institute of Information and Communication Technology, x, 45 p. (Acc.No: T00084)
Abstract: This thesis is concerned with the problem of synthesizing animating face driven by new audio sequence, which is not present in the previously recorded database. The main focus of the thesis is on exploring the efficient mapping of the features of speech domain to video domain. The mapping algorithms consist of two parts: building a model to fit the training data set and predicting the visual motion with the novel audio stimuli. The motivation was to construct the direct mapping mechanism from acoustic signals at low levels to visual frames. Unlike the previous efforts at higher acoustic levels (phonemes or words), the current approach skips the audio recognition phase, in which it is difficult to obtain high recognition accuracy due to speaker and language variability.
URI: http://drsr.daiict.ac.in/handle/123456789/121
Appears in Collections:M Tech Dissertations

Files in This Item:
File Description SizeFormat 
200411023.pdf
  Restricted Access
618.79 kBAdobe PDFThumbnail
View/Open Request a copy


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.