• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种用于机器人应用的基于对象的视觉注意力模型。

An object-based visual attention model for robotic applications.

作者信息

Yu Yuanlong, Mann George K I, Gosine Raymond G

机构信息

Faculty of Engineering and Applied Science, Memorial University of Newfoundland, St. John's, NL A1B 3X5, Canada.

出版信息

IEEE Trans Syst Man Cybern B Cybern. 2010 Oct;40(5):1398-412. doi: 10.1109/TSMCB.2009.2038895. Epub 2010 Feb 2.

DOI:10.1109/TSMCB.2009.2038895
PMID:20129865
Abstract

By extending integrated competition hypothesis, this paper presents an object-based visual attention model, which selects one object of interest using low-dimensional features, resulting that visual perception starts from a fast attentional selection procedure. The proposed attention model involves seven modules: learning of object representations stored in a long-term memory (LTM), preattentive processing, top-down biasing, bottom-up competition, mediation between top-down and bottom-up ways, generation of saliency maps, and perceptual completion processing. It works in two phases: learning phase and attending phase. In the learning phase, the corresponding object representation is trained statistically when one object is attended. A dual-coding object representation consisting of local and global codings is proposed. Intensity, color, and orientation features are used to build the local coding, and a contour feature is employed to constitute the global coding. In the attending phase, the model preattentively segments the visual field into discrete proto-objects using Gestalt rules at first. If a task-specific object is given, the model recalls the corresponding representation from LTM and deduces the task-relevant feature(s) to evaluate top-down biases. The mediation between automatic bottom-up competition and conscious top-down biasing is then performed to yield a location-based saliency map. By combination of location-based saliency within each proto-object, the proto-object-based saliency is evaluated. The most salient proto-object is selected for attention, and it is finally put into the perceptual completion processing module to yield a complete object region. This model has been applied into distinct tasks of robots: detection of task-specific stationary and moving objects. Experimental results under different conditions are shown to validate this model.

摘要

通过扩展综合竞争假设,本文提出了一种基于对象的视觉注意力模型,该模型使用低维特征选择一个感兴趣的对象,从而使视觉感知从快速的注意力选择过程开始。所提出的注意力模型包括七个模块:存储在长期记忆(LTM)中的对象表示学习、前注意处理、自上而下的偏向、自下而上的竞争、自上而下和自下而上方式之间的调解、显著性图的生成以及感知完成处理。它分两个阶段工作:学习阶段和关注阶段。在学习阶段,当关注一个对象时,对相应的对象表示进行统计训练。提出了一种由局部编码和全局编码组成的双编码对象表示。强度、颜色和方向特征用于构建局部编码,轮廓特征用于构成全局编码。在关注阶段,模型首先使用格式塔规则将视野前注意地分割成离散的原对象。如果给出了特定任务的对象,模型从LTM中召回相应的表示,并推断出与任务相关的特征以评估自上而下的偏向。然后进行自动自下而上竞争和有意识的自上而下偏向之间的调解,以产生基于位置的显著性图。通过组合每个原对象内基于位置的显著性,评估基于原对象的显著性。选择最显著的原对象进行关注,最后将其放入感知完成处理模块以产生完整的对象区域。该模型已应用于机器人的不同任务:检测特定任务的静止和移动物体。展示了不同条件下的实验结果以验证该模型。

相似文献

1
An object-based visual attention model for robotic applications.一种用于机器人应用的基于对象的视觉注意力模型。
IEEE Trans Syst Man Cybern B Cybern. 2010 Oct;40(5):1398-412. doi: 10.1109/TSMCB.2009.2038895. Epub 2010 Feb 2.
2
Visual saliency and spike timing in the ventral visual pathway.腹侧视觉通路中的视觉显著性与峰电位时间
J Physiol Paris. 2003 Mar-May;97(2-3):365-77. doi: 10.1016/j.jphysparis.2003.09.010.
3
Top-down attention based on object representation and incremental memory for knowledge building and inference.基于对象表示和增量记忆的自上而下的注意力,用于知识构建和推理。
Neural Netw. 2013 Oct;46:9-22. doi: 10.1016/j.neunet.2013.04.002. Epub 2013 Apr 8.
4
Discriminant saliency, the detection of suspicious coincidences, and applications to visual recognition.判别显著性、可疑巧合的检测及其在视觉识别中的应用。
IEEE Trans Pattern Anal Mach Intell. 2009 Jun;31(6):989-1005. doi: 10.1109/TPAMI.2009.27.
5
Fast and robust generation of feature maps for region-based visual attention.为基于区域的视觉注意力快速且稳健地生成特征图。
IEEE Trans Image Process. 2008 May;17(5):633-44. doi: 10.1109/TIP.2008.919365.
6
Visual attention on the sphere.对球体的视觉注意力。
IEEE Trans Image Process. 2008 Nov;17(11):2000-14. doi: 10.1109/TIP.2008.2003415.
7
Rapid biologically-inspired scene classification using features shared with visual attention.利用与视觉注意力共享的特征进行快速生物启发式场景分类。
IEEE Trans Pattern Anal Mach Intell. 2007 Feb;29(2):300-12. doi: 10.1109/TPAMI.2007.40.
8
A unified spectral-domain approach for saliency detection and its application to automatic object segmentation.一种用于显著度检测的统一谱域方法及其在自动目标分割中的应用。
IEEE Trans Image Process. 2012 Mar;21(3):1272-83. doi: 10.1109/TIP.2011.2164420. Epub 2011 Aug 12.
9
Learning to recognize objects on the fly: a neurally based dynamic field approach.动态学习识别物体:基于神经的动态场方法。
Neural Netw. 2008 May;21(4):562-76. doi: 10.1016/j.neunet.2008.03.007. Epub 2008 Apr 27.
10
Focus-of-attention from local color symmetries.来自局部颜色对称性的关注焦点。
IEEE Trans Pattern Anal Mach Intell. 2004 Jul;26(7):817-30. doi: 10.1109/TPAMI.2004.29.

引用本文的文献

1
Event-driven proto-object based saliency in 3D space to attract a robot's attention.基于事件驱动的原型对象的三维空间显著性吸引机器人的注意力。
Sci Rep. 2022 May 10;12(1):7645. doi: 10.1038/s41598-022-11723-6.
2
Feedback-Driven Sensory Mapping Adaptation for Robust Speech Activity Detection.用于稳健语音活动检测的反馈驱动感官映射自适应
IEEE/ACM Trans Audio Speech Lang Process. 2017 Mar;25(3):481-492. doi: 10.1109/TASLP.2016.2639322. Epub 2016 Dec 13.
3
Combining segmentation and attention: a new foveal attention model.结合分割与注意力:一种新的中央凹注意力模型。
Front Comput Neurosci. 2014 Aug 14;8:96. doi: 10.3389/fncom.2014.00096. eCollection 2014.