Situation Recognition: Visual Semantic Role Labeling For Image Understanding

2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)(2016)

引用 277|浏览105
暂无评分
摘要
This paper introduces situation recognition, the problem of producing a concise summary of the situation an image depicts including: (1) the main activity (e.g., clipping), (2) the participating actors, objects, substances, and locations (e.g., man, shears, sheep, wool, and field) and most importantly (3) the roles these participants play in the activity (e.g., the man is clipping, the shears are his tool, the wool is being clipped from the sheep, and the clipping is in a field). We use FrameNet, a verb and role lexicon developed by linguists, to define a large space of possible situations and collect a large-scale dataset containing over 500 activities, 1,700 roles, 11,000 objects, 125,000 images, and 200,000 unique situations. We also introduce structured prediction baselines and show that, in activity-centric images, situation-driven prediction of objects and activities outperforms independent object and activity recognition.
更多
查看译文
关键词
situation recognition,visual semantic role labeling,image understanding,FrameNet,verb and role lexicon,structured prediction baselines,activity-centric images,situation-driven object prediction,activity recognition
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要