HFA-GTNet: Hierarchical Fusion Adaptive Graph Transformer network for dance action recognition

Ru Jia, Li Zhao, Rui Yang,Honghong Yang,Xiaojun Wu,Yumei Zhang, Peng Li,Yuping Su

JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION(2024)

引用 0|浏览5
暂无评分
摘要
Dance action recognition is a hot research topic in computer vision. However, current skeleton -based action recognition methods face difficulties in capturing the adequate spatial structure and temporal variations of dance actions, resulting in lower recognition accuracy. In this paper, we propose a Hierarchical Fusion Adaptive Graph Transformer network (HFA-GTNet) for dance action recognition. A Hierarchical Spatial Attention (HSAtt) module is designed to extract different levels of spatial feature information from joint to parts to group, it can effectively learn high -order dependency relationships from local joints to global poses in dance actions. Secondly, to extract the joint variations in dance actions at different speeds, we have designed a Temporal Fusion Attention (TFAtt) module. This module learns the short-term and long-term temporal dependencies among joints across frames. Additionally, to capture the variations in motion patterns and dance styles among different dancers, we introduce an Adaptive Component (AdaptC). Finally, we evaluate our model on two selfbuilt dance datasets, MSDanceAction and InDanceAction, and demonstrate its superior performance compared to other state-of-the-art methods in dance action recognition.
更多
查看译文
关键词
Dance action recognition,Self-built dance datasets,Hierarchical Spatial Attention,Temporal Fusion Attention,Adaptive Component
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要