• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于方向和遮挡感知的深度学习网络的多人姿态估计。

Multi-Person Pose Estimation Using an Orientation and Occlusion Aware Deep Learning Network.

机构信息

College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Shiga 525-8577, Japan.

Institute of Industrial Science, The University of Tokyo, Tokyo 153-8505, Japan.

出版信息

Sensors (Basel). 2020 Mar 12;20(6):1593. doi: 10.3390/s20061593.

DOI:10.3390/s20061593
PMID:32178461
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7146407/
Abstract

Image based human behavior and activity understanding has been a hot topic in the field of computer vision and multimedia. As an important part, skeleton estimation, which is also called pose estimation, has attracted lots of interests. For pose estimation, most of the deep learning approaches mainly focus on the joint feature. However, the joint feature is not sufficient, especially when the image includes multi-person and the pose is occluded or not fully visible. This paper proposes a novel multi-task framework for the multi-person pose estimation. The proposed framework is developed based on Mask Region-based Convolutional Neural Networks (R-CNN) and extended to integrate the joint feature, body boundary, body orientation and occlusion condition together. In order to further improve the performance of the multi-person pose estimation, this paper proposes to organize the different information in serial multi-task models instead of the widely used parallel multi-task network. The proposed models are trained on the public dataset Common Objects in Context (COCO), which is further augmented by ground truths of body orientation and mutual-occlusion mask. Experiments demonstrate the performance of the proposed method for multi-person pose estimation and body orientation estimation. The proposed method can detect 84.6% of the Percentage of Correct Keypoints (PCK) and has an 83.7% Correct Detection Rate (CDR). Comparisons further illustrate the proposed model can reduce the over-detection compared with other methods.

摘要

基于图像的人体行为和活动理解一直是计算机视觉和多媒体领域的热门话题。作为其中的一个重要组成部分,骨骼估计(也称为姿势估计)吸引了很多关注。对于姿势估计,大多数深度学习方法主要关注关节特征。然而,关节特征并不充分,尤其是当图像包含多个人,并且姿势被遮挡或不完全可见时。本文提出了一种新颖的多任务框架用于多人姿势估计。所提出的框架基于掩模区域卷积神经网络(R-CNN)开发,并扩展为集成关节特征、身体边界、身体方向和遮挡条件。为了进一步提高多人姿势估计的性能,本文提出了在串行多任务模型中组织不同信息的方法,而不是广泛使用的并行多任务网络。所提出的模型在公共数据集 Common Objects in Context (COCO) 上进行训练,该数据集通过身体方向和相互遮挡掩模的真值进一步增强。实验证明了所提出的方法在多人姿势估计和身体方向估计方面的性能。所提出的方法可以检测到 84.6%的正确关键点百分比(PCK),并且具有 83.7%的正确检测率(CDR)。比较进一步表明,与其他方法相比,所提出的模型可以减少过度检测。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/44976ae6fdda/sensors-20-01593-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/b724794a9710/sensors-20-01593-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/fec0aab9e950/sensors-20-01593-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/a5eb975bbdfa/sensors-20-01593-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/06a75c5c368f/sensors-20-01593-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/caacfa20edda/sensors-20-01593-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/f4ecd8359b10/sensors-20-01593-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/a68faf8ac7d5/sensors-20-01593-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/b3b2b461bb92/sensors-20-01593-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/b261b12beed8/sensors-20-01593-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/5dd6fee4daa4/sensors-20-01593-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/aa8561d5d1d0/sensors-20-01593-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/9b19d049d66c/sensors-20-01593-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/d8f9f3011b11/sensors-20-01593-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/1d5b9f00d481/sensors-20-01593-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/0272f832c9ac/sensors-20-01593-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/8ed317de5dcb/sensors-20-01593-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/44976ae6fdda/sensors-20-01593-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/b724794a9710/sensors-20-01593-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/fec0aab9e950/sensors-20-01593-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/a5eb975bbdfa/sensors-20-01593-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/06a75c5c368f/sensors-20-01593-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/caacfa20edda/sensors-20-01593-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/f4ecd8359b10/sensors-20-01593-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/a68faf8ac7d5/sensors-20-01593-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/b3b2b461bb92/sensors-20-01593-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/b261b12beed8/sensors-20-01593-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/5dd6fee4daa4/sensors-20-01593-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/aa8561d5d1d0/sensors-20-01593-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/9b19d049d66c/sensors-20-01593-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/d8f9f3011b11/sensors-20-01593-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/1d5b9f00d481/sensors-20-01593-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/0272f832c9ac/sensors-20-01593-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/8ed317de5dcb/sensors-20-01593-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0742/7146407/44976ae6fdda/sensors-20-01593-g017.jpg

相似文献

1
Multi-Person Pose Estimation Using an Orientation and Occlusion Aware Deep Learning Network.基于方向和遮挡感知的深度学习网络的多人姿态估计。
Sensors (Basel). 2020 Mar 12;20(6):1593. doi: 10.3390/s20061593.
2
A deep learning approach for pose estimation from volumetric OCT data.基于深度学习的体 OCT 数据的姿态估计方法。
Med Image Anal. 2018 May;46:162-179. doi: 10.1016/j.media.2018.03.002. Epub 2018 Mar 10.
3
HyperFace: A Deep Multi-Task Learning Framework for Face Detection, Landmark Localization, Pose Estimation, and Gender Recognition.超人脸:一个用于人脸检测、地标定位、姿势估计和性别识别的深度多任务学习框架。
IEEE Trans Pattern Anal Mach Intell. 2019 Jan;41(1):121-135. doi: 10.1109/TPAMI.2017.2781233. Epub 2017 Dec 8.
4
Articulated Multi-Instrument 2-D Pose Estimation Using Fully Convolutional Networks.基于全卷积网络的关节式多仪器 2D 位姿估计
IEEE Trans Med Imaging. 2018 May;37(5):1276-1287. doi: 10.1109/TMI.2017.2787672.
5
In-Bed Pose Estimation: Deep Learning With Shallow Dataset.床上姿势估计:基于浅层数据集的深度学习
IEEE J Transl Eng Health Med. 2019 Jan 14;7:4900112. doi: 10.1109/JTEHM.2019.2892970. eCollection 2019.
6
LHPE-nets: A lightweight 2D and 3D human pose estimation model with well-structural deep networks and multi-view pose sample simplification method.LHPE-nets:一种具有良好结构深度网络和多视图姿态样本简化方法的轻量级 2D 和 3D 人体姿态估计模型。
PLoS One. 2022 Feb 23;17(2):e0264302. doi: 10.1371/journal.pone.0264302. eCollection 2022.
7
Learning shared template representation with augmented feature for multi-object pose estimation.利用增强特征学习共享模板表示进行多目标姿态估计。
Neural Netw. 2024 Aug;176:106352. doi: 10.1016/j.neunet.2024.106352. Epub 2024 Apr 30.
8
KSL-POSE: A Real-Time 2D Human Pose Estimation Method Based on Modified YOLOv8-Pose Framework.KSL-POSE:一种基于改进 YOLOv8-Pose 框架的实时 2D 人体姿态估计方法。
Sensors (Basel). 2024 Sep 26;24(19):6249. doi: 10.3390/s24196249.
9
CONet: Crowd and occlusion-aware network for occluded human pose estimation.CONet:用于遮挡人体姿态估计的人群和遮挡感知网络。
Neural Netw. 2024 Apr;172:106109. doi: 10.1016/j.neunet.2024.106109. Epub 2024 Jan 9.
10
Dual Networks Based 3D Multi-Person Pose Estimation From Monocular Video.基于双网络的单目视频3D多人姿态估计
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):1636-1651. doi: 10.1109/TPAMI.2022.3170353. Epub 2023 Jan 6.

本文引用的文献

1
Fully Convolutional Networks for Semantic Segmentation.全卷积网络用于语义分割。
IEEE Trans Pattern Anal Mach Intell. 2017 Apr;39(4):640-651. doi: 10.1109/TPAMI.2016.2572683. Epub 2016 May 24.
2
Articulated human detection with flexible mixtures of parts.具有灵活部件混合的关节式人体检测。
IEEE Trans Pattern Anal Mach Intell. 2013 Dec;35(12):2878-90. doi: 10.1109/TPAMI.2012.261.
3
Tracking people by learning their appearance.通过了解人们的外貌来追踪他们。
IEEE Trans Pattern Anal Mach Intell. 2007 Jan;29(1):65-81. doi: 10.1109/tpami.2007.250600.