Deep Attentive and Semantic Preserving Video Summarization
Neurocomputing(2020)
摘要
Video summarization shortens a lengthy video into a succinct version, whose challenges mainly originate from the difficulties of discovering the inherent relations between the original video and its summary, meanwhile minimizing the semantic information loss. Supervised approaches, especially those in deep learning framework, have demonstrated their effectiveness in video summarization. However, these approaches mainly focus on one of the challenges, and seldom pay close attention to both challenges simultaneously. To this end, we propose to pay close attention to this deficiency by incorporating the ideas of both the encoder-decoder attention and semantic preserving loss in a deep Seq2Seq framework for video summarization. Moreover, we also introduce Huber loss to replace the popular mean square error loss to enhance the robustness of the model to outliers. Extensive experiments on two benchmark video summarization datasets demonstrate that the proposed approach consistently outperforms the state-of-the-art ones.
更多查看译文
关键词
Video summarization,Encoder-decoder,Attention mechanism,Semantic preserving
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络