系统仿真学报 ›› 2023, Vol. 35 ›› Issue (5): 1098-1108.doi: 10.16182/j.issn1004731x.joss.22-0087

• 论文 • 上一篇    下一篇

单视角下基于投影子空间视图的动作识别方法

苏本跃1,2(), 孙满贞3, 马庆4, 盛敏4   

  1. 1.安徽省智能感知与计算重点实验室, 安庆师范大学, 安徽 安庆 246133
    2.铜陵学院 数学与计算机学院, 安徽 铜陵 244061
    3.安庆师范大学 计算机与信息学院, 安徽 安庆 246133
    4.安庆师范大学 数理学院, 安徽 安庆 246133
  • 收稿日期:2022-01-27 修回日期:2022-04-27 出版日期:2023-05-30 发布日期:2023-05-22
  • 作者简介:苏本跃(1971-),男,教授,博士,研究方向为图形图像处理、机器学习与模式识别等。E-mail:subenyue@sohu.com
  • 基金资助:
    安徽省自然科学基金(2108085QF269);高校领军人才团队项目(皖教秘人[2019]16号)

Action Recognition Method Based on Projection Subspace Views under Single Viewing Angle

Benyue Su1,2(), Manzhen Sun3, Qing Ma4, Min Sheng4   

  1. 1.The Key Laboratory of Intelligent Perception and Computing of Anhui Province, Anqing Normal University, Anqing 246133, China
    2.School of Mathematics and Computer, Tongling University, Tongling 244061, China
    3.School of Computer and Information, Anqing Normal University, Anqing 246133, China
    4.School of Mathematics and Physics, Anqing Normal University, Anqing 246133, China
  • Received:2022-01-27 Revised:2022-04-27 Online:2023-05-30 Published:2023-05-22

摘要:

针对单视角下深度相机跟踪关节点运动存在的自遮挡问题,提出一种基于投影子空间视图的人体动作识别方法。在不增加数据采集设备的情况下,通过子空间投影,将单视角下获得的三维动作序列投影到多个二维子空间中,在二维投影空间寻求最大类间距离,以尽可能增加基于多个子空间视图融合后的3D动作类间距离。在自建AQNU数据集的识别率为99.69%,较基准方法提升1.22%。在公共NTU-RGB+D数据集子集的识别率为80.23%,较基准方法提升1.98%。实验结果表明:本文方法可在一定程度上减少单视角数据集的自遮挡问题,提高识别率和计算效率,可达到与多视角数据集相当的识别效果。

关键词: 动作识别, 单视图, 投影子空间, 图卷积网络

Abstract:

In view of the self-occlusion problem of joint action tracking by a depth camera under a single viewing angle, a new human action recognition method based on projection subspace views is proposed. Without adding data acquisition equipment, the method projects the three-dimensional(3D) action sequences obtained under a single viewing angle into multiple two-dimensional subspacesand then seeks the maximum distance between classes in the two-dimensional subspaces, so as to increase the distance between 3D actions based on the fusion of multiple subspace views as much as possible. The recognition rate in the self-built AQNU dataset is 99.69%, which is 1.22% higher than the benchmark method. The recognition rate in the public NTU-RGB+D dataset subset is 80.23%, which is 1.98% higher than the benchmark method. The experimental results show that the method proposed in this paper can alleviate the self-occlusion problem of datasets of single viewing angles to a certain extent, effectively improve the recognition rate and computational efficiency, and achieve the recognition effect equivalent to that of datasets of multiple viewing angles.

Key words: action recognition, single view, projection subspace, graph convolutional network

中图分类号: