• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于自组织神经架构的人类活动组合学习

Compositional Learning of Human Activities With a Self-Organizing Neural Architecture.

作者信息

Mici Luiza, Parisi German I, Wermter Stefan

机构信息

Department of Informatics, Knowledge Technology, University of Hamburg, Hamburg, Germany.

出版信息

Front Robot AI. 2019 Aug 27;6:72. doi: 10.3389/frobt.2019.00072. eCollection 2019.

DOI:10.3389/frobt.2019.00072
PMID:33501087
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7805845/
Abstract

An important step for assistive systems and robot companions operating in human environments is to learn the compositionality of human activities, i.e., recognize both activities and their comprising actions. Most existing approaches address action and activity recognition as separate tasks, i.e., actions need to be inferred before the activity labels, and are thus highly sensitive to the correct temporal segmentation of the activity sequences. In this paper, we present a novel learning approach that jointly learns human activities on two levels of semantic and temporal complexity: (1) transitive actions such as and , e.g., a cereal box, and (2) high-level activities such as . Our model consists of a hierarchy of GWR networks which process and learn inherent spatiotemporal dependencies of multiple visual cues extracted from the human body skeletal representation and the interaction with objects. The neural architecture learns and semantically segments input RGB-D sequences of high-level activities into their composing actions, without supervision. We investigate the performance of our architecture with a set of experiments on a publicly available benchmark dataset. The experimental results show that our approach outperforms the state of the art with respect to the classification of the high-level activities. Additionally, we introduce a novel top-down modulation mechanism to the architecture which uses the actions and activity labels as constraints during the learning phase. In our experiments, we show how this mechanism can be used to control the network's neural growth without decreasing the overall performance.

摘要

在人类环境中运行的辅助系统和机器人伙伴的一个重要步骤是学习人类活动的组成性,即识别活动及其组成动作。大多数现有方法将动作和活动识别视为单独的任务,也就是说,在活动标签之前需要推断动作,因此对活动序列的正确时间分割高度敏感。在本文中,我们提出了一种新颖的学习方法,该方法在语义和时间复杂性的两个层面上联合学习人类活动:(1)传递动作,如 和 ,例如一个麦片盒,以及(2)高级活动,如 。我们的模型由一个GWR网络层次结构组成,该层次结构处理并学习从人体骨骼表示和与物体的交互中提取的多个视觉线索的固有时空依赖性。该神经架构在无监督的情况下学习并将高级活动的输入RGB-D序列语义分割为其组成动作。我们在一个公开可用的基准数据集上进行了一组实验来研究我们架构的性能。实验结果表明,我们的方法在高级活动分类方面优于现有技术。此外,我们向该架构引入了一种新颖的自上而下调制机制,该机制在学习阶段将动作和活动标签用作约束。在我们的实验中,我们展示了这种机制如何用于控制网络的神经生长而不降低整体性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/6b41100b1d63/frobt-06-00072-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/6b98cf726e2b/frobt-06-00072-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/44e84fe1b29d/frobt-06-00072-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/5e3a795bdc64/frobt-06-00072-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/ad7ef206f8fe/frobt-06-00072-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/10aa25dd486e/frobt-06-00072-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/6b41100b1d63/frobt-06-00072-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/6b98cf726e2b/frobt-06-00072-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/44e84fe1b29d/frobt-06-00072-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/5e3a795bdc64/frobt-06-00072-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/ad7ef206f8fe/frobt-06-00072-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/10aa25dd486e/frobt-06-00072-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f24a/7805845/6b41100b1d63/frobt-06-00072-g0008.jpg

相似文献

1
Compositional Learning of Human Activities With a Self-Organizing Neural Architecture.基于自组织神经架构的人类活动组合学习
Front Robot AI. 2019 Aug 27;6:72. doi: 10.3389/frobt.2019.00072. eCollection 2019.
2
Self-organizing neural integration of pose-motion features for human action recognition.用于人体动作识别的姿态-运动特征自组织神经整合
Front Neurorobot. 2015 Jun 9;9:3. doi: 10.3389/fnbot.2015.00003. eCollection 2015.
3
Lifelong learning of human actions with deep neural network self-organization.基于深度神经网络自组织的人类行为终身学习。
Neural Netw. 2017 Dec;96:137-149. doi: 10.1016/j.neunet.2017.09.001. Epub 2017 Sep 20.
4
Spatio⁻Temporal Image Representation of 3D Skeletal Movements for View-Invariant Action Recognition with Deep Convolutional Neural Networks.用于深度卷积神经网络的视图不变动作识别的3D骨骼运动的时空图像表示
Sensors (Basel). 2019 Apr 24;19(8):1932. doi: 10.3390/s19081932.
5
Online recognition of unsegmented actions with hierarchical SOM architecture.基于层次 SOM 架构的非分割动作在线识别。
Cogn Process. 2021 Feb;22(1):77-91. doi: 10.1007/s10339-020-00986-4. Epub 2020 Jul 22.
6
Representation Learning of Logic Words by an RNN: From Word Sequences to Robot Actions.循环神经网络对逻辑词的表示学习:从单词序列到机器人动作
Front Neurorobot. 2017 Dec 22;11:70. doi: 10.3389/fnbot.2017.00070. eCollection 2017.
7
Lifelong Learning of Spatiotemporal Representations With Dual-Memory Recurrent Self-Organization.基于双记忆循环自组织的时空表征终身学习
Front Neurorobot. 2018 Nov 28;12:78. doi: 10.3389/fnbot.2018.00078. eCollection 2018.
8
Discriminative Relational Representation Learning for RGB-D Action Recognition.用于RGB-D动作识别的判别关系表示学习
IEEE Trans Image Process. 2016 Jun;25(6):2856-2865. doi: 10.1109/TIP.2016.2556940. Epub 2016 Apr 20.
9
Real-time multiple spatiotemporal action localization and prediction approach using deep learning.基于深度学习的实时多时空动作定位与预测方法。
Neural Netw. 2020 Aug;128:331-344. doi: 10.1016/j.neunet.2020.05.017. Epub 2020 May 19.
10
Radar image segmentation using self-adapting recurrent networks.使用自适应递归网络的雷达图像分割
Int J Neural Syst. 1997 Feb;8(1):47-54. doi: 10.1142/s0129065797000070.

本文引用的文献

1
Lifelong learning of human actions with deep neural network self-organization.基于深度神经网络自组织的人类行为终身学习。
Neural Netw. 2017 Dec;96:137-149. doi: 10.1016/j.neunet.2017.09.001. Epub 2017 Sep 20.
2
A Human Activity Recognition System Using Skeleton Data from RGBD Sensors.一种使用来自RGBD传感器的骨骼数据的人类活动识别系统。
Comput Intell Neurosci. 2016;2016:4351435. doi: 10.1155/2016/4351435. Epub 2016 Mar 16.
3
Learning Actionlet Ensemble for 3D Human Action Recognition.学习动作单元集以进行 3D 人体动作识别。
IEEE Trans Pattern Anal Mach Intell. 2014 May;36(5):914-27. doi: 10.1109/TPAMI.2013.198.
4
Self-organizing neural integration of pose-motion features for human action recognition.用于人体动作识别的姿态-运动特征自组织神经整合
Front Neurorobot. 2015 Jun 9;9:3. doi: 10.3389/fnbot.2015.00003. eCollection 2015.
5
Recognizing human-object interactions in still images by modeling the mutual context of objects and human poses.通过建模物体和人体姿势的相互上下文来识别静态图像中的人与物体交互。
IEEE Trans Pattern Anal Mach Intell. 2012 Sep;34(9):1691-703. doi: 10.1109/TPAMI.2012.67.
6
Aggregating local image descriptors into compact codes.将局部图像描述符聚合到紧凑代码中。
IEEE Trans Pattern Anal Mach Intell. 2012 Sep;34(9):1704-16. doi: 10.1109/TPAMI.2011.235.
7
From motor to sensory processing in mirror neuron computational modelling.镜像神经元计算建模中的从运动到感觉处理
Biol Cybern. 2010 Dec;103(6):471-85. doi: 10.1007/s00422-010-0415-5. Epub 2010 Dec 17.
8
Observing human-object interactions: using spatial and functional compatibility for recognition.观察人与物体的交互:利用空间和功能兼容性进行识别。
IEEE Trans Pattern Anal Mach Intell. 2009 Oct;31(10):1775-89. doi: 10.1109/TPAMI.2009.83.
9
A connectionist architecture for view-independent grip-aperture computation.一种用于视角无关抓握孔径计算的联结主义架构。
Brain Res. 2008 Aug 15;1225:133-45. doi: 10.1016/j.brainres.2008.04.076. Epub 2008 May 10.
10
Goal representation in human anterior intraparietal sulcus.人类顶内沟前部的目标表征
J Neurosci. 2006 Jan 25;26(4):1133-7. doi: 10.1523/JNEUROSCI.4551-05.2006.