Action Recognition in Video Using Sparse Coding and Relative Features

2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)(2016)

引用 35|浏览62
暂无评分
摘要
This work presents an approach to category-based action recognition in video using sparse coding techniques. The proposed approach includes two main contributions: i) A new method to handle intra-class variations by decomposing each video into a reduced set of representative atomic action acts or key-sequences, and ii) A new video descriptor, ITRA: Inter-Temporal Relational Act Descriptor, that exploits the power of comparative reasoning to capture relative similarity relations among key-sequences. In terms of the method to obtain key-sequences, we introduce a loss function that, for each video, leads to the identification of a sparse set of representative key-frames capturing both, relevant particularities arising in the input video, as well as relevant generalities arising in the complete class collection. In terms of the method to obtain the ITRA descriptor, we introduce a novel scheme to quantify relative intra and inter-class similarities among local temporal patterns arising in the videos. The resulting ITRA descriptor demonstrates to be highly effective to discriminate among action categories. As a result, the proposed approach reaches remarkable action recognition performance on several popular benchmark datasets, outperforming alternative state-of-the-art techniques by a large margin.
更多
查看译文
关键词
video action recognition,category-based action recognition,sparse coding techniques,representative atomic action acts,key-sequences,video descriptor,inter-temporal relational act descriptor,comparative reasoning,loss function,representative key-frames,ITRA descriptor,inter-class similarities,intra-class similarities
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要