• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

从新视角学习人类动作识别的深度模型。

Learning a Deep Model for Human Action Recognition from Novel Viewpoints.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):667-681. doi: 10.1109/TPAMI.2017.2691768. Epub 2017 Apr 6.

DOI:10.1109/TPAMI.2017.2691768
PMID:28410098
Abstract

Recognizing human actions from unknown and unseen (novel) views is a challenging problem. We propose a Robust Non-Linear Knowledge Transfer Model (R-NKTM) for human action recognition from novel views. The proposed R-NKTM is a deep fully-connected neural network that transfers knowledge of human actions from any unknown view to a shared high-level virtual view by finding a set of non-linear transformations that connects the views. The R-NKTM is learned from 2D projections of dense trajectories of synthetic 3D human models fitted to real motion capture data and generalizes to real videos of human actions. The strength of our technique is that we learn a single R-NKTM for all actions and all viewpoints for knowledge transfer of any real human action video without the need for re-training or fine-tuning the model. Thus, R-NKTM can efficiently scale to incorporate new action classes. R-NKTM is learned with dummy labels and does not require knowledge of the camera viewpoint at any stage. Experiments on three benchmark cross-view human action datasets show that our method outperforms existing state-of-the-art.

摘要

从未知和未见(新颖)视角识别人类动作是一个具有挑战性的问题。我们提出了一种用于新颖视角下人类动作识别的鲁棒非线性知识迁移模型(R-NKTM)。所提出的 R-NKTM 是一个深度全连接神经网络,通过找到一组将视图连接起来的非线性变换,将人类动作的知识从任何未知视图转移到共享的高级虚拟视图。R-NKTM 是从拟合真实运动捕捉数据的合成 3D 人体模型的密集轨迹的 2D 投影中学习的,并推广到真实的人类动作视频。我们的技术的优势在于,我们为所有动作和所有视点学习单个 R-NKTM,以便在不需要重新训练或微调模型的情况下,将任何真实人类动作视频的知识转移。因此,R-NKTM 可以有效地扩展以纳入新的动作类别。R-NKTM 是使用虚拟标签学习的,在任何阶段都不需要知道摄像机视点的知识。在三个基准跨视图人类动作数据集上的实验表明,我们的方法优于现有的最先进技术。

相似文献

1
Learning a Deep Model for Human Action Recognition from Novel Viewpoints.从新视角学习人类动作识别的深度模型。
IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):667-681. doi: 10.1109/TPAMI.2017.2691768. Epub 2017 Apr 6.
2
Deeply Learned View-Invariant Features for Cross-View Action Recognition.深度学习的视图不变特征用于跨视图动作识别。
IEEE Trans Image Process. 2017 Jun;26(6):3028-3037. doi: 10.1109/TIP.2017.2696786. Epub 2017 Apr 24.
3
Fight Recognition in video using Hough Forests and 2D Convolutional Neural Network.使用 Hough 森林和 2D 卷积神经网络进行视频中的目标识别。
IEEE Trans Image Process. 2018 Oct;27(10):4787-4797. doi: 10.1109/TIP.2018.2845742. Epub 2018 Jun 8.
4
Deep Attention Network for Egocentric Action Recognition.基于深度注意力网络的自我中心动作识别。
IEEE Trans Image Process. 2019 Aug;28(8):3703-3713. doi: 10.1109/TIP.2019.2901707. Epub 2019 Feb 26.
5
Cross-domain human action recognition.跨域人类动作识别
IEEE Trans Syst Man Cybern B Cybern. 2012 Apr;42(2):298-307. doi: 10.1109/TSMCB.2011.2166761. Epub 2011 Sep 26.
6
Discovering motion primitives for unsupervised grouping and one-shot learning of human actions, gestures, and expressions.发现运动基元,用于人类动作、手势和表情的无监督分组和一次性学习。
IEEE Trans Pattern Anal Mach Intell. 2013 Jul;35(7):1635-48. doi: 10.1109/TPAMI.2012.253.
7
Desktop Action Recognition From First-Person Point-of-View.基于第一人称视角的桌面行为识别。
IEEE Trans Cybern. 2019 May;49(5):1616-1628. doi: 10.1109/TCYB.2018.2806381. Epub 2018 Feb 27.
8
View-Invariant Human Action Recognition Based on a 3D Bio-Constrained Skeleton Model.基于三维生物约束骨骼模型的视图不变人体动作识别。
IEEE Trans Image Process. 2019 Aug;28(8):3959-3972. doi: 10.1109/TIP.2019.2907048. Epub 2019 Mar 22.
9
NTU RGB+D 120: A Large-Scale Benchmark for 3D Human Activity Understanding.NTU RGB+D 120:用于三维人体活动理解的大规模基准测试。
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2684-2701. doi: 10.1109/TPAMI.2019.2916873. Epub 2019 May 14.
10
Explicit modeling of human-object interactions in realistic videos.真实视频中人类-物体交互的显式建模。
IEEE Trans Pattern Anal Mach Intell. 2013 Apr;35(4):835-48. doi: 10.1109/TPAMI.2012.175.

引用本文的文献

1
Violent Video Recognition by Using Sequential Image Collage.基于序列图像拼贴的暴力视频识别
Sensors (Basel). 2024 Mar 13;24(6):1844. doi: 10.3390/s24061844.
2
Action Recognition Using Action Sequences Optimization and Two-Stream 3D Dilated Neural Network.基于动作序列优化和双流 3D 扩张神经网络的动作识别
Comput Intell Neurosci. 2022 Jun 13;2022:6608448. doi: 10.1155/2022/6608448. eCollection 2022.
3
Educational Psychology Analysis Method for Extracting Students' Facial Information Based on Image Big Data.基于图像大数据的学生面部信息提取的教育心理学分析方法。
Occup Ther Int. 2022 May 11;2022:8709591. doi: 10.1155/2022/8709591. eCollection 2022.
4
Deep Learning of Fuzzy Weighted Multi-Resolution Depth Motion Maps with Spatial Feature Fusion for Action Recognition.基于空间特征融合的模糊加权多分辨率深度运动图深度学习用于动作识别
J Imaging. 2019 Oct 21;5(10):82. doi: 10.3390/jimaging5100082.
5
Robust video content analysis schemes for human action recognition.用于人体动作识别的稳健视频内容分析方案。
Sci Prog. 2021 Apr-Jun;104(2):368504211005480. doi: 10.1177/00368504211005480.
6
Complex Human Action Recognition Using a Hierarchical Feature Reduction and Deep Learning-Based Method.基于层次特征约简和深度学习方法的复杂人类行为识别
SN Comput Sci. 2021;2(2):94. doi: 10.1007/s42979-021-00484-0. Epub 2021 Feb 13.
7
VI-Net-View-Invariant Quality of Human Movement Assessment.VI-Net- 视图不变的人体运动评估质量。
Sensors (Basel). 2020 Sep 15;20(18):5258. doi: 10.3390/s20185258.
8
Human Fall Detection Based on Body Posture Spatio-Temporal Evolution.基于人体姿态时空演变的人体跌倒检测。
Sensors (Basel). 2020 Feb 10;20(3):946. doi: 10.3390/s20030946.
9
Dynamic Spatio-Temporal Bag of Expressions (D-STBoE) Model for Human Action Recognition.用于人体动作识别的动态时空词袋(D-STBoE)模型。
Sensors (Basel). 2019 Jun 21;19(12):2790. doi: 10.3390/s19122790.
10
Estimating the Composition of Food Nutrients from Hyperspectral Signals Based on Deep Neural Networks.基于深度神经网络的高光谱信号食品营养成分估算。
Sensors (Basel). 2019 Mar 31;19(7):1560. doi: 10.3390/s19071560.