• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

探索 RGB+深度融合的实时目标检测。

Exploring RGB+Depth Fusion for Real-Time Object Detection.

机构信息

EAVISE, KU Leuven, 2860 Sint-Katelijne-Waver, Belgium.

出版信息

Sensors (Basel). 2019 Feb 19;19(4):866. doi: 10.3390/s19040866.

DOI:10.3390/s19040866
PMID:30791476
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC6412390/
Abstract

In this paper, we investigate whether fusing depth information on top of normal RGB data for camera-based object detection can help to increase the performance of current state-of-the-art single-shot detection networks. Indeed, depth sensing is easily acquired using depth cameras such as a Kinect or stereo setups. We investigate the optimal manner to perform this sensor fusion with a special focus on lightweight single-pass convolutional neural network (CNN) architectures, enabling real-time processing on limited hardware. For this, we implement a network architecture allowing us to parameterize at which network layer both information sources are fused together. We performed exhaustive experiments to determine the optimal fusion point in the network, from which we can conclude that fusing towards the mid to late layers provides the best results. Our best fusion models significantly outperform the baseline RGB network in both accuracy and localization of the detections.

摘要

在本文中,我们研究了在基于相机的目标检测中融合深度信息(在常规 RGB 数据之上)是否有助于提高当前最先进的单次检测网络的性能。实际上,深度感应可以使用深度相机(如 Kinect 或立体设置)轻松获取。我们研究了使用专用的轻量级单通卷积神经网络(CNN)架构执行这种传感器融合的最佳方法,从而能够在有限的硬件上进行实时处理。为此,我们实现了一种网络架构,允许我们在网络的哪个层参数化来融合这两种信息源。我们进行了详尽的实验,以确定网络中最佳的融合点,由此我们可以得出结论,融合到中晚期层可以提供最佳结果。我们的最佳融合模型在检测的准确性和定位方面都明显优于基线 RGB 网络。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/d3748b3949d1/sensors-19-00866-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/7f259571628e/sensors-19-00866-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/1a741d85df9b/sensors-19-00866-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/98e529b257e6/sensors-19-00866-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/85afce13ddbd/sensors-19-00866-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/46229e55612f/sensors-19-00866-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/c564aa97da68/sensors-19-00866-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/0d7e3eaab33f/sensors-19-00866-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/d3748b3949d1/sensors-19-00866-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/7f259571628e/sensors-19-00866-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/1a741d85df9b/sensors-19-00866-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/98e529b257e6/sensors-19-00866-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/85afce13ddbd/sensors-19-00866-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/46229e55612f/sensors-19-00866-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/c564aa97da68/sensors-19-00866-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/0d7e3eaab33f/sensors-19-00866-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5a27/6412390/d3748b3949d1/sensors-19-00866-g008.jpg

相似文献

1
Exploring RGB+Depth Fusion for Real-Time Object Detection.探索 RGB+深度融合的实时目标检测。
Sensors (Basel). 2019 Feb 19;19(4):866. doi: 10.3390/s19040866.
2
Moving Object Detection Based on Fusion of Depth Information and RGB Features.基于深度信息与 RGB 特征融合的运动目标检测。
Sensors (Basel). 2022 Jun 22;22(13):4702. doi: 10.3390/s22134702.
3
Edge Preserving and Multi-Scale Contextual Neural Network for Salient Object Detection.边缘保持和多尺度上下文神经网络的显著目标检测。
IEEE Trans Image Process. 2018;27(1):121-134. doi: 10.1109/TIP.2017.2756825.
4
Data-Level Recombination and Lightweight Fusion Scheme for RGB-D Salient Object Detection.用于RGB-D显著目标检测的数据级重组与轻量级融合方案
IEEE Trans Image Process. 2021;30:458-471. doi: 10.1109/TIP.2020.3037470. Epub 2020 Nov 23.
5
Deep Attention Models for Human Tracking Using RGBD.基于 RGBD 的深度注意模型的人体跟踪
Sensors (Basel). 2019 Feb 13;19(4):750. doi: 10.3390/s19040750.
6
CNNs-Based RGB-D Saliency Detection via Cross-View Transfer and Multiview Fusion.基于卷积神经网络的跨视图迁移和多视图融合的 RGB-D 显著目标检测。
IEEE Trans Cybern. 2018 Nov;48(11):3171-3183. doi: 10.1109/TCYB.2017.2761775. Epub 2017 Oct 31.
7
3D Localization of Hand Acupoints Using Hand Geometry and Landmark Points Based on RGB-D CNN Fusion.基于 RGB-D CNN 融合的手部几何形状和特征点的手部穴位 3D 定位。
Ann Biomed Eng. 2022 Sep;50(9):1103-1115. doi: 10.1007/s10439-022-02986-1. Epub 2022 Jun 4.
8
3D object detection through fog and occlusion: passive integral imaging vs active (LiDAR) sensing.通过雾和遮挡进行3D目标检测:被动积分成像与主动(激光雷达)传感
Opt Express. 2023 Jan 2;31(1):479-491. doi: 10.1364/OE.478125.
9
CNN Deep Learning with Wavelet Image Fusion of CCD RGB-IR and Depth-Grayscale Sensor Data for Hand Gesture Intention Recognition.CNN 基于 CCD RGB-IR 与深度灰度传感器数据的子波图像融合的深度学习在手势意图识别中的应用。
Sensors (Basel). 2022 Jan 21;22(3):803. doi: 10.3390/s22030803.
10
RGB-T Salient Object Detection via Fusing Multi-level CNN Features.基于融合多级卷积神经网络特征的RGB-T显著目标检测
IEEE Trans Image Process. 2019 Dec 17. doi: 10.1109/TIP.2019.2959253.

引用本文的文献

1
L-AVATeD: The lidar and visual walking terrain dataset.L-AVATeD:激光雷达与视觉行走地形数据集。
Front Robot AI. 2024 Dec 4;11:1384575. doi: 10.3389/frobt.2024.1384575. eCollection 2024.
2
Multi-Task Foreground-Aware Network with Depth Completion for Enhanced RGB-D Fusion Object Detection Based on Transformer.基于Transformer的具有深度补全功能的多任务前景感知网络用于增强RGB-D融合目标检测
Sensors (Basel). 2024 Apr 8;24(7):2374. doi: 10.3390/s24072374.
3
Deep-Learning-Based Real-Time and Automatic Target-to-Background Ratio Calculation in Fluorescence Endoscopy for Cancer Detection and Localization.

本文引用的文献

1
Ease of disassembly of products to support circular economy strategies.产品易于拆解,以支持循环经济战略。
Resour Conserv Recycl. 2018 Aug;135:323-334. doi: 10.1016/j.resconrec.2017.06.022.
2
Fast Feature Pyramids for Object Detection.快速目标检测特征金字塔。
IEEE Trans Pattern Anal Mach Intell. 2014 Aug;36(8):1532-45. doi: 10.1109/TPAMI.2014.2300479.
3
Stereo processing by semiglobal matching and mutual information.通过半全局匹配和互信息进行立体处理。
基于深度学习的荧光内镜中实时自动目标与背景比率计算用于癌症检测和定位
Diagnostics (Basel). 2022 Aug 23;12(9):2031. doi: 10.3390/diagnostics12092031.
4
Non-destructive Plant Biomass Monitoring With High Spatio-Temporal Resolution Proximal RGB-D Imagery and End-to-End Deep Learning.利用高时空分辨率近端RGB-D图像和端到端深度学习进行无损植物生物量监测
Front Plant Sci. 2022 Apr 13;13:758818. doi: 10.3389/fpls.2022.758818. eCollection 2022.
5
sTetro-Deep Learning Powered Staircase Cleaning and Maintenance Reconfigurable Robot.基于深度学习的楼梯清洁和维护可重构机器人。
Sensors (Basel). 2021 Sep 18;21(18):6279. doi: 10.3390/s21186279.
6
Asymmetric Adaptive Fusion in a Two-Stream Network for RGB-D Human Detection.基于双流网络的 RGB-D 人体检测中的非对称自适应融合。
Sensors (Basel). 2021 Jan 29;21(3):916. doi: 10.3390/s21030916.
7
Robust 3D Hand Detection from a Single RGB-D Image in Unconstrained Environments.在非约束环境中从单个 RGB-D 图像进行鲁棒的 3D 手部检测。
Sensors (Basel). 2020 Nov 7;20(21):6360. doi: 10.3390/s20216360.
8
Double-Constraint Inpainting Model of a Single-Depth Image.单深度图像的双约束图像修复模型
Sensors (Basel). 2020 Mar 24;20(6):1797. doi: 10.3390/s20061797.
IEEE Trans Pattern Anal Mach Intell. 2008 Feb;30(2):328-41. doi: 10.1109/TPAMI.2007.1166.