Journal of System Simulation ›› 2024, Vol. 36 ›› Issue (3): 555-563.doi: 10.16182/j.issn1004731x.joss.22-1234

• Papers • Previous Articles     Next Articles

Human Action Recognition Based on Skeleton Edge Information Under Projection Subspace

Su Benyue1,2(), Zhang Peng1,2, Zhu Bangguo1,2, Guo Mengjuan1,2, Sheng Min3   

  1. 1.School of Computer and Information, Anqing Normal University, Anqing 246133, China
    2.School of Mathematics and Computer, Tongling University, Tongling 244061, China
    3.School of Mathematics and Physics, Anqing Normal University, Anqing 246133, China
  • Received:2022-10-17 Revised:2023-01-31 Online:2024-03-15 Published:2024-03-14

Abstract:

In recent years, human action recognition based on skeleton data has received a lot of attention in the fields of computer vision and human-computer interaction. Most of the existing methods focus on modeling the skeleton points in the original 3D coordinate space. However, skeleton points ignore the physical chain structure of the human body itself, which makes it difficult to portray the local correlation of human motion. In addition, due to the diversity of camera views, it is difficult to explore the comprehensive representation of actions in different views under the original point-based 3D space. In view of this, this paper proposed an action recognition method based on skeleton edge information in the projection subspace. The method defined skeleton edge information combined with the body's own connection for capturing the spatial characteristics of the action. The direction and size information of skeleton edge motion was introduced on the basis of the skeleton edge information for capturing the temporal characteristics of the action. The 2D projection subspace was used for action characterization under different subspace perspectives. A suitable feature fusion strategy was explored, and the above features were extracted comprehensively through the improved CNN framework. Experimental results on two challenging large datasets NTU-RGB+D 60 (evaluation metrics are cross-subject and cross-view) and NTU-RGB+D 120 (evaluation metrics are cross-subject and cross-set) show that compared with the benchmark method, the proposed method improves the accuracy under the four metrics by 3.2%, 2.4%, 3.1%, and 5.8%, respectively.

Key words: skeleton data, skeleton edges, edge direction, edge size, projection subspace

CLC Number: