Zero-Shot Style Transfer for Multimodal Data-Driven Gesture Synthesis

2023 IEEE 17th International Conference on Automatic Face and Gesture Recognition (FG)(2023)

引用 0|浏览11
暂无评分
摘要
We propose a multimodal speech driven approach to generate 2D upper-body gestures for virtual agents, in the communicative style of different speakers, seen or unseen by our model during training. Upper-body gestures of a source speaker are generated based on the content of his/her multimodal data - speech acoustics and text semantics. The synthesized source speaker's gestures are conditioned on the multimodal style representation of the target speaker. Our approach is zero-shot, and can generalize the style transfer to new unseen speakers, without any additional training. An objective evaluation is conducted to validate our approach.
更多
查看译文
关键词
synthesis,style,transfer,zero-shot,data-driven
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要