• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于合成训练数据的迁移字典学习实现任意视角动作识别

Arbitrary View Action Recognition via Transfer Dictionary Learning on Synthetic Training Data.

作者信息

Zhang Jingtian, Shum Hubert P H, Han Jungong, Shao Ling

出版信息

IEEE Trans Image Process. 2018 May 15. doi: 10.1109/TIP.2018.2836323.

DOI:10.1109/TIP.2018.2836323
PMID:29994770
Abstract

Human action recognition is crucial to many practical applications, ranging from human-computer interaction to video surveillance. Most approaches either recognize the human action from a fixed view or require the knowledge of view angle, which is usually not available in practical applications. In this paper, we propose a novel end-to-end framework to jointly learn a view-invariance transfer dictionary and a view-invariant classifier. The result of the process is a dictionary that can project real-world 2D video into a view-invariant sparse representation, as well as a classifier to recognize actions with an arbitrary view. The main feature of our algorithm is the use of synthetic data to extract view-invariance between 3D and 2D videos during the pre-training phase. This guarantees the availability of training data, and removes the hassle of obtaining real-world videos in specific viewing angles. Additionally, for better describing the actions in 3D videos, we introduce a new feature set called the 3D dense trajectories to effectively encode extracted trajectory information on 3D videos. Experimental results on the IXMAS, N-UCLA, i3DPost and UWA3DII datasets show improvements over existing algorithms.

摘要

人类动作识别对于许多实际应用至关重要,涵盖从人机交互到视频监控等领域。大多数方法要么从固定视角识别人类动作,要么需要视角知识,而这在实际应用中通常难以获取。在本文中,我们提出了一种新颖的端到端框架,用于联合学习视角不变性转移字典和视角不变分类器。该过程的结果是一个能够将真实世界的二维视频投影到视角不变稀疏表示的字典,以及一个用于识别任意视角动作的分类器。我们算法的主要特点是在预训练阶段使用合成数据来提取三维和二维视频之间的视角不变性。这保证了训练数据的可用性,并消除了获取特定视角真实世界视频的麻烦。此外,为了更好地描述三维视频中的动作,我们引入了一个名为三维密集轨迹的新特征集,以有效地编码在三维视频上提取的轨迹信息。在IXMAS、N-UCLA、i3DPost和UWA3DII数据集上的实验结果表明,该方法相较于现有算法有改进。

相似文献

1
Arbitrary View Action Recognition via Transfer Dictionary Learning on Synthetic Training Data.基于合成训练数据的迁移字典学习实现任意视角动作识别
IEEE Trans Image Process. 2018 May 15. doi: 10.1109/TIP.2018.2836323.
2
Cross-View Action Recognition via Transferable Dictionary Learning.跨视图动作识别的可迁移字典学习
IEEE Trans Image Process. 2016 May;25(6):2542-56. doi: 10.1109/TIP.2016.2548242.
3
Multi-Domain & Multi-Task Learning for Human Action Recognition.用于人类动作识别的多领域与多任务学习
IEEE Trans Image Process. 2018 Sep 28. doi: 10.1109/TIP.2018.2872879.
4
Learning a Deep Model for Human Action Recognition from Novel Viewpoints.从新视角学习人类动作识别的深度模型。
IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):667-681. doi: 10.1109/TPAMI.2017.2691768. Epub 2017 Apr 6.
5
Specificity and Latent Correlation Learning for Action Recognition Using Synthetic Multi-View Data From Depth Maps.使用深度图的合成多视角数据进行动作识别的特异性和潜在关联学习。
IEEE Trans Image Process. 2017 Dec;26(12):5560-5574. doi: 10.1109/TIP.2017.2740122. Epub 2017 Aug 14.
6
View-invariant action recognition based on artificial neural networks.基于人工神经网络的视图不变动作识别。
IEEE Trans Neural Netw Learn Syst. 2012 Mar;23(3):412-24. doi: 10.1109/TNNLS.2011.2181865.
7
Learning dictionaries of sparse codes of 3D movements of body joints for real-time human activity understanding.学习人体关节三维运动稀疏码字典,实现实时人体活动理解。
PLoS One. 2014 Dec 4;9(12):e114147. doi: 10.1371/journal.pone.0114147. eCollection 2014.
8
A discriminative model of motion and cross ratio for view-invariant action recognition.一种用于视图不变动作识别的运动和交比判别模型。
IEEE Trans Image Process. 2012 Apr;21(4):2187-97. doi: 10.1109/TIP.2011.2176346.
9
Learning sparse representations for human action recognition.学习人类动作识别的稀疏表示。
IEEE Trans Pattern Anal Mach Intell. 2012 Aug;34(8):1576-88. doi: 10.1109/TPAMI.2011.253.
10
Label consistent K-SVD: learning a discriminative dictionary for recognition.标签一致的 K-SVD:学习用于识别的判别字典。
IEEE Trans Pattern Anal Mach Intell. 2013 Nov;35(11):2651-64. doi: 10.1109/TPAMI.2013.88.

引用本文的文献

1
Neural Networks for Automatic Posture Recognition in Ambient-Assisted Living.神经网络在环境辅助生活中的自动姿势识别。
Sensors (Basel). 2022 Mar 29;22(7):2609. doi: 10.3390/s22072609.