Learning Video-Story Composition via Recurrent Neural Network

2018 IEEE Winter Conference on Applications of Computer Vision (WACV)(2018)

引用 4|浏览54
暂无评分
摘要
In this paper, we propose a learning-based method to compose a video-story from a group of video clips that describe an activity or experience. We learn the coherence between video clips from real videos via the Recurrent Neural Network (RNN) that jointly incorporates the spatial-temporal semantics and motion dynamics to generate smooth and relevant compositions. We further rearrange the results generated by the RNN to make the overall video-story compatible with the storyline structure via a submodular ranking optimization process. Experimental results on the video-story dataset show that the proposed algorithm outperforms the state-of-the-art approach.
更多
查看译文
关键词
Recurrent Neural Network,video clips,RNN,spatial-temporal semantics,motion dynamics,smooth compositions,video-story dataset show,video-story composition learning,submodular ranking optimization process
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要