2D character animating networks - bringing static characters to move via motion transfer.

SAC(2019)

引用 1|浏览48
暂无评分
摘要
Animating a character given an image is an interesting research problem that posed to have applicational values. Namely, it could help animators to automatically animate new characters with motions similar to previous characters. The problem also has value in digital entertainment for animating characters created by players such as drawn characters. There are a few deep learning video synthesis models that could generate a video given an image. However, it is unlikely there exists a dataset that could train them for animating characters since characters for digital entertainment tend to be more novel and diverse. Nor is it practical for animators to create a large dataset for training. To this end, a 2D-CharAnimNet is proposed. Empowered by a novel motion transfer scheme for video generation, the proposed variational-autoencoder-based model could use a relatively small dataset. In addition, to improve the fidelity of video frames, dynamic skip-connections along with a polishing generative adversarial networks are also proposed. Results seem to indicate that the model has encouraging potential in adapting for applicational uses.
更多
查看译文
关键词
animation, generative model, motion transfer, neural networks, video synthesis
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要