FeatureFlow: Robust Video Interpolation via Structure-to-Texture Generation

CVPR(2020)

引用 81|浏览300
暂无评分
摘要
Video interpolation aims to synthesize non-existent frames between two consecutive frames. Although existing optical flow based methods have achieved promising results, they still face great challenges in dealing with the interpolation of complicated dynamic scenes, which include occlusion, blur or abrupt brightness change. This is mainly because these cases may break the basic assumptions of the optical flow estimation (i.e. smoothness, consistency). In this work, we devised a novel structure-to-texture generation framework which splits the video interpolation task into two stages: structure-guided interpolation and texture refinement. In the first stage, deep structure-aware features are employed to predict feature flows from two consecutive frames to their intermediate result, and further generate the structure image of the intermediate frame. In the second stage, based on the generated coarse result, a Frame Texture Compensator is trained to fill in detailed textures. To the best of our knowledge, this is the first work that attempts to directly generate the intermediate frame through blending deep features. Experiments on both the benchmark datasets and challenging occlusion cases demonstrate the superiority of the proposed framework over the state-of-the-art methods. Codes are available on https://github.com/CM-BF/FeatureFlow.
更多
查看译文
关键词
robust video interpolation,nonexistent frames,consecutive frames,optical flow based methods,dynamic scenes,abrupt brightness change,optical flow estimation,video interpolation task,structure-guided interpolation,texture refinement,deep structure-aware features,feature flows,structure image,intermediate frame,occlusion cases,FeatureFlow,structure-to-texture generation framework,frame texture compensator
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要