• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

多层次深度和图像融合的人体活动检测。

Multilevel depth and image fusion for human activity detection.

出版信息

IEEE Trans Cybern. 2013 Oct;43(5):1383-94. doi: 10.1109/TCYB.2013.2276433. Epub 2013 Aug 27.

DOI:10.1109/TCYB.2013.2276433
PMID:23996589
Abstract

Recognizing complex human activities usually requires the detection and modeling of individual visual features and the interactions between them. Current methods only rely on the visual features extracted from 2-D images, and therefore often lead to unreliable salient visual feature detection and inaccurate modeling of the interaction context between individual features. In this paper, we show that these problems can be addressed by combining data from a conventional camera and a depth sensor (e.g., Microsoft Kinect). We propose a novel complex activity recognition and localization framework that effectively fuses information from both grayscale and depth image channels at multiple levels of the video processing pipeline. In the individual visual feature detection level, depth-based filters are applied to the detected human/object rectangles to remove false detections. In the next level of interaction modeling, 3-D spatial and temporal contexts among human subjects or objects are extracted by integrating information from both grayscale and depth images. Depth information is also utilized to distinguish different types of indoor scenes. Finally, a latent structural model is developed to integrate the information from multiple levels of video processing for an activity detection. Extensive experiments on two activity recognition benchmarks (one with depth information) and a challenging grayscale + depth human activity database that contains complex interactions between human-human, human-object, and human-surroundings demonstrate the effectiveness of the proposed multilevel grayscale + depth fusion scheme. Higher recognition and localization accuracies are obtained relative to the previous methods.

摘要

识别复杂的人类活动通常需要检测和建模个体的视觉特征及其之间的相互作用。当前的方法仅依赖于从 2D 图像中提取的视觉特征,因此经常导致不可靠的显著视觉特征检测和个体特征之间交互上下文的不准确建模。在本文中,我们展示了通过结合来自传统摄像机和深度传感器(例如 Microsoft Kinect)的数据可以解决这些问题。我们提出了一种新颖的复杂活动识别和定位框架,该框架有效地融合了视频处理管道多个层次的灰度和深度图像通道的信息。在个体视觉特征检测级别,将基于深度的滤波器应用于检测到的人体/物体矩形,以去除错误检测。在下一级别的交互建模中,通过整合来自灰度和深度图像的信息来提取人体/物体之间的 3D 时空上下文。深度信息还用于区分不同类型的室内场景。最后,开发了一个潜在的结构模型,以整合来自多个视频处理级别的信息,用于活动检测。在两个活动识别基准(一个具有深度信息)和一个具有复杂的人机交互、人机物交互和人与环境交互的挑战性灰度+深度人类活动数据库上进行了广泛的实验,证明了所提出的多层次灰度+深度融合方案的有效性。与之前的方法相比,获得了更高的识别和定位精度。

相似文献

1
Multilevel depth and image fusion for human activity detection.多层次深度和图像融合的人体活动检测。
IEEE Trans Cybern. 2013 Oct;43(5):1383-94. doi: 10.1109/TCYB.2013.2276433. Epub 2013 Aug 27.
2
Enhanced computer vision with Microsoft Kinect sensor: a review.增强计算机视觉的微软 Kinect 传感器:综述。
IEEE Trans Cybern. 2013 Oct;43(5):1318-34. doi: 10.1109/TCYB.2013.2265378. Epub 2013 Jun 25.
3
Real-time multiple human perception with color-depth cameras on a mobile robot.移动机器人上的彩色深度相机的实时多人感知。
IEEE Trans Cybern. 2013 Oct;43(5):1429-41. doi: 10.1109/TCYB.2013.2275291. Epub 2013 Aug 21.
4
Real-time posture reconstruction for Microsoft Kinect.基于 Microsoft Kinect 的实时姿态重建。
IEEE Trans Cybern. 2013 Oct;43(5):1357-69. doi: 10.1109/TCYB.2013.2275945. Epub 2013 Aug 22.
5
Free-viewpoint video of human actors using multiple handheld Kinects.使用多个手持 Kinect 的人体演员自由视点视频。
IEEE Trans Cybern. 2013 Oct;43(5):1370-82. doi: 10.1109/TCYB.2013.2272321. Epub 2013 Jul 22.
6
Discriminative exemplar coding for sign language recognition with Kinect.基于 Kinect 的手语识别的判别示例编码。
IEEE Trans Cybern. 2013 Oct;43(5):1418-28. doi: 10.1109/TCYB.2013.2265337. Epub 2013 Jun 19.
7
Depth-aware image seam carving.深度感知图像拼接。
IEEE Trans Cybern. 2013 Oct;43(5):1453-61. doi: 10.1109/TCYB.2013.2273270. Epub 2013 Jul 22.
8
3-D rigid body tracking using vision and depth sensors.使用视觉和深度传感器的三维刚体跟踪。
IEEE Trans Cybern. 2013 Oct;43(5):1395-405. doi: 10.1109/TCYB.2013.2272735. Epub 2013 Aug 15.
9
Rank preserving sparse learning for Kinect based scene classification.基于 Kinect 的场景分类的保序稀疏学习。
IEEE Trans Cybern. 2013 Oct;43(5):1406-17. doi: 10.1109/TCYB.2013.2264285. Epub 2013 Jul 3.
10
Accurate estimation of human body orientation from RGB-D sensors.基于 RGB-D 传感器的人体姿态精确估计。
IEEE Trans Cybern. 2013 Oct;43(5):1442-52. doi: 10.1109/TCYB.2013.2272636. Epub 2013 Jul 23.

引用本文的文献

1
HIT HAR: Human Image Threshing Machine for Human Activity Recognition Using Deep Learning Models.利用深度学习模型的人体图像过筛机(HIT HAR):用于人体活动识别。
Comput Intell Neurosci. 2022 Oct 6;2022:1808990. doi: 10.1155/2022/1808990. eCollection 2022.
2
Human activity recognition in artificial intelligence framework: a narrative review.人工智能框架中的人类活动识别:一篇叙述性综述。
Artif Intell Rev. 2022;55(6):4755-4808. doi: 10.1007/s10462-021-10116-x. Epub 2022 Jan 18.
3
Optimization of deep neural network-based human activity recognition for a wearable device.
基于深度神经网络的可穿戴设备人体活动识别优化
Heliyon. 2021 Aug 14;7(8):e07797. doi: 10.1016/j.heliyon.2021.e07797. eCollection 2021 Aug.
4
RGB-D Data-Based Action Recognition: A Review.基于 RGB-D 数据的动作识别:综述。
Sensors (Basel). 2021 Jun 21;21(12):4246. doi: 10.3390/s21124246.
5
Prediction of Human Activities Based on a New Structure of Skeleton Features and Deep Learning Model.基于骨骼特征新结构和深度学习模型的人类活动预测。
Sensors (Basel). 2020 Sep 1;20(17):4944. doi: 10.3390/s20174944.
6
A Comprehensive Survey of Vision-Based Human Action Recognition Methods.基于视觉的人体动作识别方法综述。
Sensors (Basel). 2019 Feb 27;19(5):1005. doi: 10.3390/s19051005.
7
A Human Activity Recognition Algorithm Based on Stacking Denoising Autoencoder and LightGBM.基于堆叠去噪自编码器和 LightGBM 的人体活动识别算法。
Sensors (Basel). 2019 Feb 23;19(4):947. doi: 10.3390/s19040947.
8
A Human Activity Recognition System Based on Dynamic Clustering of Skeleton Data.基于骨骼数据动态聚类的人体活动识别系统。
Sensors (Basel). 2017 May 11;17(5):1100. doi: 10.3390/s17051100.
9
Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition.基于RGB-D数据的自适应局部时空特征用于一次性学习手势识别
Sensors (Basel). 2016 Dec 17;16(12):2171. doi: 10.3390/s16122171.
10
Hierarchical Activity Recognition Using Smart Watches and RGB-Depth Cameras.使用智能手表和RGB深度相机的分层活动识别
Sensors (Basel). 2016 Oct 15;16(10):1713. doi: 10.3390/s16101713.