Return to search

Real Time Speech Driven Face Animation / Talstyrd Ansiktsanimering i Realtid

The goal of this project is to implement a system to analyse an audio signal containing speech, and produce a classifcation of lip shape categories (visemes) in order to synchronize the lips of a computer generated face with the speech. The thesis describes the work to derive a method that maps speech to lip move- ments, on an animated face model, in real time. The method is implemented in C++ on the PC/Windows platform. The program reads speech from pre-recorded audio files and continuously performs spectral analysis of the speech. Neural networks are used to classify the speech into a sequence of phonemes, and the corresponding visemes are shown on the screen. Some time delay between input speech and the visualization could not be avoided, but the overall visual impression is that sound and animation are synchronized.

Identiferoai:union.ndltd.org:UPSALLA1/oai:DiVA.org:liu-2015
Date January 2003
CreatorsAxelsson, Andreas, Björhäll, Erik
PublisherLinköpings universitet, Institutionen för systemteknik, Linköpings universitet, Institutionen för systemteknik, Institutionen för systemteknik
Source SetsDiVA Archive at Upsalla University
LanguageEnglish
Detected LanguageEnglish
TypeStudent thesis, info:eu-repo/semantics/bachelorThesis, text
Formatapplication/pdf
Rightsinfo:eu-repo/semantics/openAccess
RelationLiTH-ISY-Ex, ; 3389

Page generated in 0.0019 seconds