• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于事件驱动的原型对象的三维空间显著性吸引机器人的注意力。

Event-driven proto-object based saliency in 3D space to attract a robot's attention.

机构信息

Event Driven Perception for Robotics, Istituto Italiano di Tecnologia, 16163, Genoa, Italy.

Electrical Engineering and Computer Science, Technische Universität Berlin, 10623, Berlin, Germany.

出版信息

Sci Rep. 2022 May 10;12(1):7645. doi: 10.1038/s41598-022-11723-6.

DOI:10.1038/s41598-022-11723-6
PMID:35538154
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9090933/
Abstract

To interact with its environment, a robot working in 3D space needs to organise its visual input in terms of objects or their perceptual precursors, proto-objects. Among other visual cues, depth is a submodality used to direct attention to visual features and objects. Current depth-based proto-object attention models have been implemented for standard RGB-D cameras that produce synchronous frames. In contrast, event cameras are neuromorphic sensors that loosely mimic the function of the human retina by asynchronously encoding per-pixel brightness changes at very high temporal resolution, thereby providing advantages like high dynamic range, efficiency (thanks to their high degree of signal compression), and low latency. We propose a bio-inspired bottom-up attention model that exploits event-driven sensing to generate depth-based saliency maps that allow a robot to interact with complex visual input. We use event-cameras mounted in the eyes of the iCub humanoid robot to directly extract edge, disparity and motion information. Real-world experiments demonstrate that our system robustly selects salient objects near the robot in the presence of clutter and dynamic scene changes, for the benefit of downstream applications like object segmentation, tracking and robot interaction with external objects.

摘要

为了与 3D 空间中的环境交互,在该空间中工作的机器人需要根据对象或其感知前体(原对象)组织其视觉输入。在其他视觉线索中,深度是一种用于将注意力引导到视觉特征和对象的子模态。当前基于深度的原对象注意模型已针对生成同步帧的标准 RGB-D 相机实现。相比之下,事件相机是一种神经形态传感器,通过以非常高的时间分辨率异步地对每个像素的亮度变化进行编码,从而松散地模拟人眼的功能,从而提供了高动态范围、效率(由于其高度的信号压缩)和低延迟等优势。我们提出了一种受生物启发的自下而上的注意模型,该模型利用事件驱动的传感来生成基于深度的显着性图,使机器人能够与复杂的视觉输入进行交互。我们使用安装在 iCub 人形机器人眼睛中的事件相机来直接提取边缘、视差和运动信息。实际实验表明,我们的系统在存在杂乱和动态场景变化的情况下,能够稳健地选择机器人附近的显着对象,从而有利于下游应用,如对象分割、跟踪和机器人与外部对象的交互。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/70eccbe25420/41598_2022_11723_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/b196d025af48/41598_2022_11723_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/19ebe6ec4213/41598_2022_11723_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/14d5f87383d6/41598_2022_11723_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/0c839a99bb0f/41598_2022_11723_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/c9ff74f46e23/41598_2022_11723_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/ba49d4b4318d/41598_2022_11723_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/70eccbe25420/41598_2022_11723_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/b196d025af48/41598_2022_11723_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/19ebe6ec4213/41598_2022_11723_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/14d5f87383d6/41598_2022_11723_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/0c839a99bb0f/41598_2022_11723_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/c9ff74f46e23/41598_2022_11723_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/ba49d4b4318d/41598_2022_11723_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c577/9090933/70eccbe25420/41598_2022_11723_Fig7_HTML.jpg

相似文献

1
Event-driven proto-object based saliency in 3D space to attract a robot's attention.基于事件驱动的原型对象的三维空间显著性吸引机器人的注意力。
Sci Rep. 2022 May 10;12(1):7645. doi: 10.1038/s41598-022-11723-6.
2
Event-driven visual attention for the humanoid robot iCub.人型机器人 iCub 的事件驱动视觉注意
Front Neurosci. 2013 Dec 13;7:234. doi: 10.3389/fnins.2013.00234. eCollection 2013.
3
A proto-object-based computational model for visual saliency.一种基于原型对象的视觉显著性计算模型。
J Vis. 2013 Nov 26;13(13):27. doi: 10.1167/13.13.27.
4
An On-chip Spiking Neural Network for Estimation of the Head Pose of the iCub Robot.一种用于估计iCub机器人头部姿态的片上脉冲神经网络。
Front Neurosci. 2020 Jun 23;14:551. doi: 10.3389/fnins.2020.00551. eCollection 2020.
5
Event-Based Vision: A Survey.基于事件的视觉:综述。
IEEE Trans Pattern Anal Mach Intell. 2022 Jan;44(1):154-180. doi: 10.1109/TPAMI.2020.3008413. Epub 2021 Dec 7.
6
Sensorimotor coordination in a "baby" robot: learning about objects through grasping.“婴儿”机器人中的感觉运动协调:通过抓握了解物体。
Prog Brain Res. 2007;164:403-24. doi: 10.1016/S0079-6123(07)64022-9.
7
Distractor-Aware Event-Based Tracking.基于干扰感知的事件跟踪
IEEE Trans Image Process. 2023;32:6129-6141. doi: 10.1109/TIP.2023.3326683. Epub 2023 Nov 8.
8
Real-time multiple human perception with color-depth cameras on a mobile robot.移动机器人上的彩色深度相机的实时多人感知。
IEEE Trans Cybern. 2013 Oct;43(5):1429-41. doi: 10.1109/TCYB.2013.2275291. Epub 2013 Aug 21.
9
Event-Based Eccentric Motion Detection Exploiting Time Difference Encoding.基于事件的利用时间差编码的偏心运动检测
Front Neurosci. 2020 May 8;14:451. doi: 10.3389/fnins.2020.00451. eCollection 2020.
10
Event-Based Motion Segmentation With Spatio-Temporal Graph Cuts.基于事件的时空图割运动分割
IEEE Trans Neural Netw Learn Syst. 2023 Aug;34(8):4868-4880. doi: 10.1109/TNNLS.2021.3124580. Epub 2023 Aug 4.

引用本文的文献

1
Event-driven figure-ground organisation model for the humanoid robot iCub.用于人形机器人iCub的事件驱动的图底组织模型。
Nat Commun. 2025 Feb 22;16(1):1874. doi: 10.1038/s41467-025-56904-9.
2
The influence of stereopsis on visual saliency in a proto-object based model of selective attention.立体知觉对视注意力选择的基于原型对象模型中视觉显著性的影响。
Vision Res. 2023 Nov;212:108304. doi: 10.1016/j.visres.2023.108304. Epub 2023 Aug 3.
3
Saliency Map and Deep Learning in Binary Classification of Brain Tumours.显著图与深度学习在脑肿瘤二分类中的应用。

本文引用的文献

1
Five Factors that Guide Attention in Visual Search.视觉搜索中引导注意力的五个因素。
Nat Hum Behav. 2017 Mar;1(3). doi: 10.1038/s41562-017-0058. Epub 2017 Mar 8.
2
Event-Based Eccentric Motion Detection Exploiting Time Difference Encoding.基于事件的利用时间差编码的偏心运动检测
Front Neurosci. 2020 May 8;14:451. doi: 10.3389/fnins.2020.00451. eCollection 2020.
3
The prominent role of perceptual salience in object discrimination: overt discrimination of graspable side does not activate grasping affordances.知觉显著性在物体辨别中的突出作用:可抓握侧的显性辨别并不激活抓握功能。
Sensors (Basel). 2023 May 7;23(9):4543. doi: 10.3390/s23094543.
Psychol Res. 2021 Apr;85(3):1234-1247. doi: 10.1007/s00426-020-01296-2. Epub 2020 Feb 8.
4
Visual salience, not the graspable part of a pictured eating utensil, grabs attention.视觉显著性而非图片中餐具可抓取的部分吸引了注意力。
Atten Percept Psychophys. 2019 Jul;81(5):1454-1463. doi: 10.3758/s13414-019-01679-7.
5
What Do Different Evaluation Metrics Tell Us About Saliency Models?不同的评估指标能告诉我们关于显著性模型的哪些信息?
IEEE Trans Pattern Anal Mach Intell. 2019 Mar;41(3):740-757. doi: 10.1109/TPAMI.2018.2815601. Epub 2018 Mar 13.
6
Graspable Objects Grab Attention More Than Images Do.可抓取物体比图像更能吸引注意力。
Psychol Sci. 2018 Feb;29(2):206-218. doi: 10.1177/0956797617730599. Epub 2017 Dec 7.
7
A spiking neural network model of 3D perception for event-based neuromorphic stereo vision systems.基于事件的神经形态立体视觉系统的三维感知尖峰神经网络模型。
Sci Rep. 2017 Jan 12;7:40703. doi: 10.1038/srep40703.
8
Eye movements between saccades: Measuring ocular drift and tremor.扫视之间的眼球运动:测量眼漂移和眼震颤。
Vision Res. 2016 May;122:93-104. doi: 10.1016/j.visres.2016.03.006. Epub 2016 Apr 17.
9
A proto-object based saliency model in three-dimensional space.一种基于原物体的三维空间显著模型。
Vision Res. 2016 Feb;119:42-9. doi: 10.1016/j.visres.2015.12.004. Epub 2016 Jan 19.
10
Border-ownership coding.边界所有权编码
Scholarpedia J. 2013;8(10):30040. doi: 10.4249/scholarpedia.30040.