School of Software Engineering, Xi'an Jiaotong University, Xi'an 710049, China.
Division of Information Science, Nara Institute of Science and Technology, Ikoma 630-0192, Nara, Japan.
Sensors (Basel). 2020 Jun 21;20(12):3499. doi: 10.3390/s20123499.
Skeleton-based action recognition has achieved great advances with the development of graph convolutional networks (GCNs). Many existing GCNs-based models only use the fixed hand-crafted adjacency matrix to describe the connections between human body joints. This omits the important implicit connections between joints, which contain discriminative information for different actions. In this paper, we propose an action-specific graph convolutional module, which is able to extract the implicit connections and properly balance them for each action. In addition, to filter out the useless and redundant information in the temporal dimension, we propose a simple yet effective operation named gated temporal convolution. These two major novelties ensure the superiority of our proposed method, as demonstrated on three large-scale public datasets: NTU-RGB + D, Kinetics, and NTU-RGB + D 120, and also shown in the detailed ablation studies.
基于骨架的动作识别随着图卷积网络 (GCN) 的发展取得了重大进展。许多现有的基于 GCN 的模型仅使用固定的手工制作的邻接矩阵来描述人体关节之间的连接。这忽略了关节之间重要的隐式连接,这些连接包含了不同动作的判别信息。在本文中,我们提出了一种特定于动作的图卷积模块,能够提取隐式连接,并为每个动作适当平衡它们。此外,为了在时间维度上过滤掉无用和冗余的信息,我们提出了一种简单而有效的操作,称为门控时间卷积。这两个主要的创新确保了我们提出的方法的优越性,在三个大规模的公共数据集:NTU-RGB + D、Kinetics 和 NTU-RGB + D120 上得到了验证,在详细的消融研究中也得到了展示。