Please use this identifier to cite or link to this item:
http://drsr.daiict.ac.in//handle/123456789/121
Title: | Speech driven facial animation system |
Authors: | Jotwani, Naresh D. Singh, Archana |
Keywords: | Co-articulation Facial animation system Frame Gaussian mixture model Hidden markov model Speech recognition Vector quantization Viseme Viterbi Algorithm |
Issue Date: | 2006 |
Publisher: | Dhirubhai Ambani Institute of Information and Communication Technology |
Citation: | Singh, Archana (2006). Speech driven facial animation system. Dhirubhai Ambani Institute of Information and Communication Technology, x, 45 p. (Acc.No: T00084) |
Abstract: | This thesis is concerned with the problem of synthesizing animating face driven by new audio sequence, which is not present in the previously recorded database. The main focus of the thesis is on exploring the efficient mapping of the features of speech domain to video domain. The mapping algorithms consist of two parts: building a model to fit the training data set and predicting the visual motion with the novel audio stimuli. The motivation was to construct the direct mapping mechanism from acoustic signals at low levels to visual frames. Unlike the previous efforts at higher acoustic levels (phonemes or words), the current approach skips the audio recognition phase, in which it is difficult to obtain high recognition accuracy due to speaker and language variability. |
URI: | http://drsr.daiict.ac.in/handle/123456789/121 |
Appears in Collections: | M Tech Dissertations |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
200411023.pdf Restricted Access | 618.79 kB | Adobe PDF | View/Open Request a copy |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.