Exemplar-based Action Recognition in Video

BMVC(2009)

引用 65|浏览43
暂无评分
摘要
In this work, we present a method for action localization and recognition using an exemplar-based approach. It starts from local dense yet scale-invariant spatio-temporal features. The most discriminative visual words are selected and used to cast bounding box hypotheses, which are then verified and further grouped into the final detections. To the best of our knowledge, we are the first to extend the exemplar-based approach using local features into the spatio-temporal domain. This allows us to avoid the problems that typically plague sliding window-based approaches - in particular the exhaustive search over spatial coordinates, time, and spatial as well as temporal scales. We report state-of- the-art results on challenging datasets, extracted from real movies, for both classification and localization.
更多
查看译文
关键词
exhaustive search,scale invariance,sliding window
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要