VideoMem: Constructing, Analyzing, Predicting Short-term and Long-term Video Memorability

2019 IEEE/CVF International Conference on Computer Vision (ICCV)(2018)

引用 15|浏览54
暂无评分
摘要
Humans share a strong tendency to memorize/forget some of the visual information they encounter. This paper focuses on providing computational models for the prediction of the intrinsic memorability of visual content. To address this new challenge, we introduce a large scale dataset (VideoMem) composed of 10,000 videos annotated with memorability scores. In contrast to previous work on image memorability -- where memorability was measured a few minutes after memorization -- memory performance is measured twice: a few minutes after memorization and again 24-72 hours later. Hence, the dataset comes with short-term and long-term memorability annotations. After an in-depth analysis of the dataset, we investigate several deep neural network based models for the prediction of video memorability. Our best model using a ranking loss achieves a Spearman's rank correlation of 0.494 for short-term memorability prediction, while our proposed model with attention mechanism provides insights of what makes a content memorable. The VideoMem dataset with pre-extracted features is publicly available.
更多
查看译文
关键词
predicting short-term,long-term video memorability,humans share,strong tendency,visual information,intrinsic memorability,visual content,scale dataset,memorability scores,image memorability,long-term memorability annotations,deep neural network-based models,resp. long-term,VideoMem dataset,time 24.0 hour to 72.0 hour
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要