Behavioral Recognition of Skeletal Data Based on Targeted Dual Fusion Strategy

Authors

  • Xiao Yun China University of Mining and Technology
  • Chenglong Xu China University of Mining and Technology
  • Kevin Riou Nantes Université
  • Kaiwen Dong China University of Mining and Technology
  • Yanjing Sun China University of Mining and Technology
  • Song Li China University of Mining and Technology
  • Kevin Subrin Nantes Université
  • Patrick Le Callet Nantes Université

DOI:

https://doi.org/10.1609/aaai.v38i7.28517

Keywords:

CV: Video Understanding & Activity Analysis, CV: 3D Computer Vision, CV: Biometrics, Face, Gesture & Pose, CV: Motion & Tracking

Abstract

The deployment of multi-stream fusion strategy on behavioral recognition from skeletal data can extract complementary features from different information streams and improve the recognition accuracy, but suffers from high model complexity and a large number of parameters. Besides, existing multi-stream methods using a fixed adjacency matrix homogenizes the model’s discrimination process across diverse actions, causing reduction of the actual lift for the multi-stream model. Finally, attention mechanisms are commonly applied to the multi-dimensional features, including spatial, temporal and channel dimensions. But their attention scores are typically fused in a concatenated manner, leading to the ignorance of the interrelation between joints in complex actions. To alleviate these issues, the Front-Rear dual Fusion Graph Convolutional Network (FRF-GCN) is proposed to provide a lightweight model based on skeletal data. Targeted adjacency matrices are also designed for different front fusion streams, allowing the model to focus on actions of varying magnitudes. Simultaneously, the mechanism of Spatial-Temporal-Channel Parallel Attention (STC-P), which processes attention in parallel and places greater emphasis on useful information, is proposed to further improve model’s performance. FRF-GCN demonstrates significant competitiveness compared to the current state-of-the-art methods on the NTU RGB+D, NTU RGB+D 120 and Kinetics-Skeleton 400 datasets. Our code is available at: https://github.com/sunbeam-kkt/FRF-GCN-master.

Published

2024-03-24

How to Cite

Yun, X., Xu, C., Riou, K., Dong, K., Sun, Y., Li, S., Subrin, K., & Le Callet, P. (2024). Behavioral Recognition of Skeletal Data Based on Targeted Dual Fusion Strategy. Proceedings of the AAAI Conference on Artificial Intelligence, 38(7), 6917-6925. https://doi.org/10.1609/aaai.v38i7.28517

Issue

Section

AAAI Technical Track on Computer Vision VI