• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于改进胶囊网络的注意力引导Huber损失用于头部姿态估计

Attention-Guided Huber Loss for Head Pose Estimation Based on Improved Capsule Network.

作者信息

Zhong Runhao, He Li, Wang Hongwei, Yuan Liang, Li Kexin, Liu Zhening

机构信息

School of Mechanical Engineering, Xinjiang University, Urumqi 830046, China.

School of Information Science and Technology, Beijing University of Chemical Technology, Beijing 100029, China.

出版信息

Entropy (Basel). 2023 Jul 5;25(7):1024. doi: 10.3390/e25071024.

DOI:10.3390/e25071024
PMID:37509971
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10378512/
Abstract

Head pose estimation is an important technology for analyzing human behavior and has been widely researched and applied in areas such as human-computer interaction and fatigue detection. However, traditional head pose estimation networks suffer from the problem of easily losing spatial structure information, particularly in complex scenarios where occlusions and multiple object detections are common, resulting in low accuracy. To address the above issues, we propose a head pose estimation model based on the residual network and capsule network. Firstly, a deep residual network is used to extract features from three stages, capturing spatial structure information at different levels, and a global attention block is employed to enhance the spatial weight of feature extraction. To effectively avoid the loss of spatial structure information, the features are encoded and transmitted to the output using an improved capsule network, which is enhanced in its generalization ability through self-attention routing mechanisms. To enhance the robustness of the model, we optimize Huber loss, which is first used in head pose estimation. Finally, experiments are conducted on three popular public datasets, 300W-LP, AFLW2000, and BIWI. The results demonstrate that the proposed method achieves state-of-the-art results, particularly in scenarios with occlusions.

摘要

头部姿态估计是分析人类行为的一项重要技术,已在人机交互和疲劳检测等领域得到广泛研究和应用。然而,传统的头部姿态估计网络存在容易丢失空间结构信息的问题,特别是在遮挡和多目标检测常见的复杂场景中,导致准确率较低。为了解决上述问题,我们提出了一种基于残差网络和胶囊网络的头部姿态估计模型。首先,使用深度残差网络从三个阶段提取特征,在不同层次上捕捉空间结构信息,并采用全局注意力模块来增强特征提取的空间权重。为了有效避免空间结构信息的丢失,通过改进的胶囊网络对特征进行编码并传输到输出,该胶囊网络通过自注意力路由机制增强了其泛化能力。为了提高模型的鲁棒性,我们优化了Huber损失,这是首次在头部姿态估计中使用。最后,在三个流行的公共数据集300W-LP、AFLW2000和BIWI上进行了实验。结果表明,所提出的方法取得了最优结果,特别是在存在遮挡的场景中。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/3936763dded3/entropy-25-01024-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/236b3dc2f57b/entropy-25-01024-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/584f86c92fb6/entropy-25-01024-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/652d27a24b55/entropy-25-01024-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/af1ec3936575/entropy-25-01024-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/4251e18d2074/entropy-25-01024-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/486aa6d7b4d0/entropy-25-01024-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/97508977bac1/entropy-25-01024-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/9be7302394bc/entropy-25-01024-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/625894f3b0f0/entropy-25-01024-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/e95f1089d328/entropy-25-01024-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/6d6f8afcbfae/entropy-25-01024-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/a2ff46b7ef3b/entropy-25-01024-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/3936763dded3/entropy-25-01024-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/236b3dc2f57b/entropy-25-01024-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/584f86c92fb6/entropy-25-01024-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/652d27a24b55/entropy-25-01024-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/af1ec3936575/entropy-25-01024-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/4251e18d2074/entropy-25-01024-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/486aa6d7b4d0/entropy-25-01024-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/97508977bac1/entropy-25-01024-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/9be7302394bc/entropy-25-01024-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/625894f3b0f0/entropy-25-01024-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/e95f1089d328/entropy-25-01024-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/6d6f8afcbfae/entropy-25-01024-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/a2ff46b7ef3b/entropy-25-01024-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0589/10378512/3936763dded3/entropy-25-01024-g013.jpg

相似文献

1
Attention-Guided Huber Loss for Head Pose Estimation Based on Improved Capsule Network.基于改进胶囊网络的注意力引导Huber损失用于头部姿态估计
Entropy (Basel). 2023 Jul 5;25(7):1024. doi: 10.3390/e25071024.
2
An Improved Tiered Head Pose Estimation Network with Self-Adjust Loss Function.一种具有自调整损失函数的改进型分层头部姿态估计网络。
Entropy (Basel). 2022 Jul 14;24(7):974. doi: 10.3390/e24070974.
3
Self-Attention Mechanism-Based Head Pose Estimation Network with Fusion of Point Cloud and Image Features.基于自注意力机制的融合点云与图像特征的头部姿态估计网络
Sensors (Basel). 2023 Dec 18;23(24):9894. doi: 10.3390/s23249894.
4
Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image.基于重建 3D 人脸模型和 2D 图像关键点匹配的头部姿势估计
Sensors (Basel). 2021 Mar 6;21(5):1841. doi: 10.3390/s21051841.
5
Deep Learning-Based 6-DoF Object Pose Estimation Considering Synthetic Dataset.基于深度学习的考虑合成数据集的六自由度物体姿态估计
Sensors (Basel). 2023 Dec 15;23(24):9854. doi: 10.3390/s23249854.
6
Detection, segmentation, and 3D pose estimation of surgical tools using convolutional neural networks and algebraic geometry.使用卷积神经网络和代数几何进行手术工具的检测、分割和三维姿态估计。
Med Image Anal. 2021 May;70:101994. doi: 10.1016/j.media.2021.101994. Epub 2021 Feb 7.
7
MSA-Net: Establishing Reliable Correspondences by Multiscale Attention Network.MSA-Net:通过多尺度注意力网络建立可靠对应关系
IEEE Trans Image Process. 2022;31:4598-4608. doi: 10.1109/TIP.2022.3186535. Epub 2022 Jul 12.
8
MH6D: Multi-Hypothesis Consistency Learning for Category-Level 6-D Object Pose Estimation.MH6D:用于类别级6D物体姿态估计的多假设一致性学习
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):4820-4833. doi: 10.1109/TNNLS.2024.3360712. Epub 2025 Feb 28.
9
DON6D: a decoupled one-stage network for 6D pose estimation.DON6D:用于6D姿态估计的解耦单阶段网络。
Sci Rep. 2024 Apr 10;14(1):8410. doi: 10.1038/s41598-024-59152-x.
10
Online Learning State Evaluation Method Based on Face Detection and Head Pose Estimation.基于人脸检测和头部姿势估计的在线学习状态评估方法。
Sensors (Basel). 2024 Feb 20;24(5):1365. doi: 10.3390/s24051365.

本文引用的文献

1
A Novel Zernike Moment-Based Real-Time Head Pose and Gaze Estimation Framework for Accuracy-Sensitive Applications.一种适用于精度敏感应用的基于新型泽尼克矩的实时头部姿态和注视估计框架。
Sensors (Basel). 2022 Nov 3;22(21):8449. doi: 10.3390/s22218449.
2
An Improved Tiered Head Pose Estimation Network with Self-Adjust Loss Function.一种具有自调整损失函数的改进型分层头部姿态估计网络。
Entropy (Basel). 2022 Jul 14;24(7):974. doi: 10.3390/e24070974.
3
Efficient-CapsNet: capsule network with self-attention routing.高效胶囊网络:具有自注意力路由的胶囊网络。
Sci Rep. 2021 Jul 19;11(1):14634. doi: 10.1038/s41598-021-93977-0.
4
Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image.基于重建 3D 人脸模型和 2D 图像关键点匹配的头部姿势估计
Sensors (Basel). 2021 Mar 6;21(5):1841. doi: 10.3390/s21051841.
5
FASHE: A FrActal Based Strategy for Head Pose Estimation.FASHE:一种基于分形的头部姿势估计策略。
IEEE Trans Image Process. 2021;30:3192-3203. doi: 10.1109/TIP.2021.3059409. Epub 2021 Feb 25.
6
Towards Robust and Accurate Multi-View and Partially-Occluded Face Alignment.面向鲁棒准确的多视角和部分遮挡人脸对齐。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):987-1001. doi: 10.1109/TPAMI.2017.2697958. Epub 2017 Apr 25.
7
The adaptive hough transform.自适应 Hough 变换。
IEEE Trans Pattern Anal Mach Intell. 1987 May;9(5):690-8. doi: 10.1109/tpami.1987.4767964.
8
Head pose estimation in computer vision: a survey.计算机视觉中的头部姿态估计:一项综述。
IEEE Trans Pattern Anal Mach Intell. 2009 Apr;31(4):607-26. doi: 10.1109/TPAMI.2008.106.