CStrCRL: Cross-view Contrastive Learning through Gated GCN with Strong Augmentations for Skeleton Recognition

IEEE Transactions on Circuits and Systems for Video Technology(2023)

引用 0|浏览2
暂无评分
摘要
Contrastive learning has been widely embraced for its notable success along with two augmentation methods— normal and strong augmentations—in skeleton action recognition. Existing methods gain performance largely by customizing normal augmentations while bypassing strong augmentations that riches in motion patterns. To make up for the blank, we propose a novel framework, called CStrCRL, acquiring view-invariant and discriminative features from strong augmentations by leveraging contrastive learning. Specifically, to avoid the fragility of skeleton data adversely affecting the model after applying strong augmentations, we use consistency learning to maximize the similarity between strongly and normally augmented views. Furthermore, we employ cross-view learning on strong and normal augmentations for eliminating uncertainty feature boundaries learned by the model. Moreover, we design a new backbone, termed GatedStrNet, for discriminating valid and invalid features contained in strong augmented views. Finally, extensive experiments on NTU 60/120 and PKUMMD II demonstrate that the proposed method bridges the performance gap between normal and strong augmentations on contrastive learning of skeleton recognition. Notably, with a single stream input, CStrCRL achieves accuracies of 78.93% and 84.04% on the NTU60 Xsub and Xview datasets. Our source code can be found at: https://github.com/RHu-main/CStrCRL.
更多
查看译文
关键词
Skeleton recognition,Unsupervised learning,Strong augmentations,ST-GGCN
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要