Suppr超能文献

GAS-GCN:基于骨骼的动作识别的门控动作特定图卷积网络。

GAS-GCN: Gated Action-Specific Graph Convolutional Networks for Skeleton-Based Action Recognition.

机构信息

School of Software Engineering, Xi'an Jiaotong University, Xi'an 710049, China.

Division of Information Science, Nara Institute of Science and Technology, Ikoma 630-0192, Nara, Japan.

出版信息

Sensors (Basel). 2020 Jun 21;20(12):3499. doi: 10.3390/s20123499.

Abstract

Skeleton-based action recognition has achieved great advances with the development of graph convolutional networks (GCNs). Many existing GCNs-based models only use the fixed hand-crafted adjacency matrix to describe the connections between human body joints. This omits the important implicit connections between joints, which contain discriminative information for different actions. In this paper, we propose an action-specific graph convolutional module, which is able to extract the implicit connections and properly balance them for each action. In addition, to filter out the useless and redundant information in the temporal dimension, we propose a simple yet effective operation named gated temporal convolution. These two major novelties ensure the superiority of our proposed method, as demonstrated on three large-scale public datasets: NTU-RGB + D, Kinetics, and NTU-RGB + D 120, and also shown in the detailed ablation studies.

摘要

基于骨架的动作识别随着图卷积网络 (GCN) 的发展取得了重大进展。许多现有的基于 GCN 的模型仅使用固定的手工制作的邻接矩阵来描述人体关节之间的连接。这忽略了关节之间重要的隐式连接,这些连接包含了不同动作的判别信息。在本文中,我们提出了一种特定于动作的图卷积模块,能够提取隐式连接,并为每个动作适当平衡它们。此外,为了在时间维度上过滤掉无用和冗余的信息,我们提出了一种简单而有效的操作,称为门控时间卷积。这两个主要的创新确保了我们提出的方法的优越性,在三个大规模的公共数据集:NTU-RGB + D、Kinetics 和 NTU-RGB + D120 上得到了验证,在详细的消融研究中也得到了展示。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c372/7349730/b54b9a60b812/sensors-20-03499-g001.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验