• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于双流网络的 RGB-D 人体检测中的非对称自适应融合。

Asymmetric Adaptive Fusion in a Two-Stream Network for RGB-D Human Detection.

机构信息

Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China.

出版信息

Sensors (Basel). 2021 Jan 29;21(3):916. doi: 10.3390/s21030916.

DOI:10.3390/s21030916
PMID:33572928
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7866388/
Abstract

In recent years, human detection in indoor scenes has been widely applied in smart buildings and smart security, but many related challenges can still be difficult to address, such as frequent occlusion, low illumination and multiple poses. This paper proposes an asymmetric adaptive fusion two-stream network (AAFTS-net) for RGB-D human detection. This network can fully extract person-specific depth features and RGB features while reducing the typical complexity of a two-stream network. A depth feature pyramid is constructed by combining contextual information, with the motivation of combining multiscale depth features to improve the adaptability for targets of different sizes. An adaptive channel weighting (ACW) module weights the RGB-D feature channels to achieve efficient feature selection and information complementation. This paper also introduces a novel RGB-D dataset for human detection called RGBD-human, on which we verify the performance of the proposed algorithm. The experimental results show that AAFTS-net outperforms existing state-of-the-art methods and can maintain stable performance under conditions of frequent occlusion, low illumination and multiple poses.

摘要

近年来,室内场景中的人体检测已广泛应用于智能建筑和智能安全领域,但仍有许多相关挑战难以解决,例如频繁遮挡、低光照和多种姿势。本文提出了一种用于 RGB-D 人体检测的非对称自适应融合双流网络(AAFTS-net)。该网络能够充分提取特定于人的深度特征和 RGB 特征,同时降低双流网络的典型复杂性。通过组合上下文信息构建深度特征金字塔,以结合多尺度深度特征来提高对不同大小目标的适应性。自适应通道加权(ACW)模块对 RGB-D 特征通道进行加权,以实现有效的特征选择和信息补充。本文还介绍了一个名为 RGBD-human 的新的 RGB-D 人体检测数据集,我们在该数据集上验证了所提出算法的性能。实验结果表明,AAFTS-net 优于现有的最先进方法,并且可以在频繁遮挡、低光照和多种姿势条件下保持稳定的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/232b4d6c7faa/sensors-21-00916-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/4e5bfbe0b1da/sensors-21-00916-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/89104dedc52a/sensors-21-00916-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/e697abe5c73e/sensors-21-00916-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/b7c8123ac55a/sensors-21-00916-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/c56042721f12/sensors-21-00916-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/8ae8facc2ebe/sensors-21-00916-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/d6c41a3701fc/sensors-21-00916-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/232b4d6c7faa/sensors-21-00916-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/4e5bfbe0b1da/sensors-21-00916-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/89104dedc52a/sensors-21-00916-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/e697abe5c73e/sensors-21-00916-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/b7c8123ac55a/sensors-21-00916-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/c56042721f12/sensors-21-00916-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/8ae8facc2ebe/sensors-21-00916-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/d6c41a3701fc/sensors-21-00916-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ffe9/7866388/232b4d6c7faa/sensors-21-00916-g008.jpg

相似文献

1
Asymmetric Adaptive Fusion in a Two-Stream Network for RGB-D Human Detection.基于双流网络的 RGB-D 人体检测中的非对称自适应融合。
Sensors (Basel). 2021 Jan 29;21(3):916. doi: 10.3390/s21030916.
2
CDNet: Complementary Depth Network for RGB-D Salient Object Detection.CDNet:用于RGB-D显著目标检测的互补深度网络。
IEEE Trans Image Process. 2021;30:3376-3390. doi: 10.1109/TIP.2021.3060167. Epub 2021 Mar 9.
3
IRFR-Net: Interactive Recursive Feature-Reshaping Network for Detecting Salient Objects in RGB-D Images.IRFR-Net:用于在RGB-D图像中检测显著目标的交互式递归特征重塑网络
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):4132-4144. doi: 10.1109/TNNLS.2021.3105484. Epub 2025 Feb 28.
4
FASFLNet: feature adaptive selection and fusion lightweight network for RGB-D indoor scene parsing.FASFLNet:用于 RGB-D 室内场景解析的特征自适应选择和融合轻量级网络。
Opt Express. 2023 Feb 27;31(5):8029-8041. doi: 10.1364/OE.480252.
5
Moving Object Detection Based on Fusion of Depth Information and RGB Features.基于深度信息与 RGB 特征融合的运动目标检测。
Sensors (Basel). 2022 Jun 22;22(13):4702. doi: 10.3390/s22134702.
6
Multi-Object Tracking Algorithm for RGB-D Images Based on Asymmetric Dual Siamese Networks.基于非对称双孪生网络的 RGB-D 图像多目标跟踪算法。
Sensors (Basel). 2020 Nov 25;20(23):6745. doi: 10.3390/s20236745.
7
Data-Level Recombination and Lightweight Fusion Scheme for RGB-D Salient Object Detection.用于RGB-D显著目标检测的数据级重组与轻量级融合方案
IEEE Trans Image Process. 2021;30:458-471. doi: 10.1109/TIP.2020.3037470. Epub 2020 Nov 23.
8
Hierarchical Multimodal Adaptive Fusion (HMAF) Network for Prediction of RGB-D Saliency.用于预测RGB-D显著图的分层多模态自适应融合(HMAF)网络
Comput Intell Neurosci. 2020 Nov 20;2020:8841681. doi: 10.1155/2020/8841681. eCollection 2020.
9
Three-stream Attention-aware Network for RGB-D Salient Object Detection.用于RGB-D显著目标检测的三流注意力感知网络
IEEE Trans Image Process. 2019 Jan 7. doi: 10.1109/TIP.2019.2891104.
10
Rethinking RGB-D Salient Object Detection: Models, Data Sets, and Large-Scale Benchmarks.重新思考RGB-D显著目标检测:模型、数据集和大规模基准
IEEE Trans Neural Netw Learn Syst. 2021 May;32(5):2075-2089. doi: 10.1109/TNNLS.2020.2996406. Epub 2021 May 3.

引用本文的文献

1
A Pruning Method for Deep Convolutional Network Based on Heat Map Generation Metrics.基于热图生成指标的深度卷积网络修剪方法。
Sensors (Basel). 2022 Mar 4;22(5):2022. doi: 10.3390/s22052022.

本文引用的文献

1
Exploring RGB+Depth Fusion for Real-Time Object Detection.探索 RGB+深度融合的实时目标检测。
Sensors (Basel). 2019 Feb 19;19(4):866. doi: 10.3390/s19040866.
2
RGB-D Object Recognition Using Multi-Modal Deep Neural Network and DS Evidence Theory.基于多模态深度神经网络和证据理论的 RGB-D 目标识别。
Sensors (Basel). 2019 Jan 27;19(3):529. doi: 10.3390/s19030529.
3
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
4
A general framework for tracking multiple people from a moving camera.从移动摄像机跟踪多个人的通用框架。
IEEE Trans Pattern Anal Mach Intell. 2013 Jul;35(7):1577-91. doi: 10.1109/TPAMI.2012.248.