Creating A Speech Enabled Avatar From A Single Photograph

IEEE VIRTUAL REALITY 2008, PROCEEDINGS(2008)

引用 11|浏览12
暂无评分
摘要
This paper presents a complete framework for creating a speech-enabled avatar from a single image of a person. Our approach uses a generic facial motion model which represents deformations of a prototype face during speech. We have developed an HMM-based facial animation algorithm which takes into account both lexical stress and coarticulation. This algorithm produces realistic animations of the prototype facial surface from either text or speech. The generic facial motion model can be transformed to a novel face geometry using a set of corresponding points between the prototype face surface and the novel face. Given a face photograph, a small number of manually selected features in the photograph are used to deform the prototype face surface. The deformed surface is then used to animate the face in the photograph. We show several examples of avatars that are driven by text and speech inputs.
更多
查看译文
关键词
eye gaze,vector field,face recognition,hidden markov model,real time rendering,stress,texture synthesis,hidden markov models,speech synthesis,computer graphic,prototypes,face geometry,computer animation,computer graphics,rendering system,solid modeling,computational geometry,coarticulation,facial animation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要