Talking Faces - Technologies and Applications
dc.contributor.author | Ostermann, Jörn | en_US |
dc.contributor.author | Weissenfeld, Axel | en_US |
dc.contributor.author | Liu, Kang | en_US |
dc.contributor.editor | Mike Chantler | en_US |
dc.date.accessioned | 2016-02-11T13:30:55Z | |
dc.date.available | 2016-02-11T13:30:55Z | |
dc.date.issued | 2005 | en_US |
dc.description.abstract | Facial animation has been combined with text-to-speech synthesis to create innovative multimodal interfaces. In this lecture, we present the technology and architecture in order to use this multimodal interface in an web-based environment to support education, entertainment and e-commerce applications. Modern text to speech synthesizers using concatenative speech synthesis are able to generate high quality speech. Face animation uses the phoneme and timing information provided by such a speech synthesizer in order to animate the mouth. There are 2 basic technologies that are used to render talking faces: 3D face models as described in MPEG-4 may be used to provide the impression of a talking cartoon or human-like character. Sample-based face models generated from recorded video enable the synthesis of a talking head that cannot be distinguished from a real person. Depending on the chosen face animation technology and latency requirements, different architectures for delivering the talking head over the Internet are required for interactive applications. Keywords: Face animation, visual speech | en_US |
dc.description.sectionheaders | Keynote 3 | en_US |
dc.description.seriesinformation | Vision, Video, and Graphics (2005) | en_US |
dc.identifier.doi | 10.2312/vvg.20051020 | en_US |
dc.identifier.isbn | 3-905673-57-6 | en_US |
dc.identifier.pages | 157-157 | en_US |
dc.identifier.uri | https://doi.org/10.2312/vvg.20051020 | en_US |
dc.publisher | The Eurographics Association | en_US |
dc.title | Talking Faces - Technologies and Applications | en_US |
Files
Original bundle
1 - 1 of 1