• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种受机器人启发的扫描路径模型揭示了不确定性和语义对象线索对动态场景中注视引导的重要性。

A robotics-inspired scanpath model reveals the importance of uncertainty and semantic object cues for gaze guidance in dynamic scenes.

作者信息

Mengers Vito, Roth Nicolas, Brock Oliver, Obermayer Klaus, Rolfs Martin

机构信息

Technische Universität Berlin, Berlin, Germany.

Science of Intelligence, Research Cluster of Excellence, Berlin, Germany.

出版信息

J Vis. 2025 Feb 3;25(2):6. doi: 10.1167/jov.25.2.6.

DOI:10.1167/jov.25.2.6
PMID:39928323
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11812614/
Abstract

The objects we perceive guide our eye movements when observing real-world dynamic scenes. Yet, gaze shifts and selective attention are critical for perceiving details and refining object boundaries. Object segmentation and gaze behavior are, however, typically treated as two independent processes. Here, we present a computational model that simulates these processes in an interconnected manner and allows for hypothesis-driven investigations of distinct attentional mechanisms. Drawing on an information processing pattern from robotics, we use a Bayesian filter to recursively segment the scene, which also provides an uncertainty estimate for the object boundaries that we use to guide active scene exploration. We demonstrate that this model closely resembles observers' free viewing behavior on a dataset of dynamic real-world scenes, measured by scanpath statistics, including foveation duration and saccade amplitude distributions used for parameter fitting and higher-level statistics not used for fitting. These include how object detections, inspections, and returns are balanced and a delay of returning saccades without an explicit implementation of such temporal inhibition of return. Extensive simulations and ablation studies show that uncertainty promotes balanced exploration and that semantic object cues are crucial to forming the perceptual units used in object-based attention. Moreover, we show how our model's modular design allows for extensions, such as incorporating saccadic momentum or presaccadic attention, to further align its output with human scanpaths.

摘要

我们所感知的物体在观察现实世界动态场景时会引导我们的眼球运动。然而,注视转移和选择性注意对于感知细节和细化物体边界至关重要。然而,物体分割和注视行为通常被视为两个独立的过程。在此,我们提出一种计算模型,该模型以相互关联的方式模拟这些过程,并允许对不同的注意机制进行假设驱动的研究。借鉴机器人学中的信息处理模式,我们使用贝叶斯滤波器对场景进行递归分割,这也为我们用于指导主动场景探索的物体边界提供了不确定性估计。我们证明,该模型与在动态现实世界场景数据集上观察者的自由观看行为非常相似,通过扫描路径统计进行测量,包括用于参数拟合的注视持续时间和扫视幅度分布以及未用于拟合的更高层次统计。这些包括物体检测、检查和返回如何平衡,以及返回扫视的延迟,而无需明确实施这种返回的时间抑制。广泛的模拟和消融研究表明,不确定性促进了平衡探索,并且语义物体线索对于形成基于物体的注意中使用的感知单元至关重要。此外,我们展示了我们模型的模块化设计如何允许扩展,例如纳入扫视动量或扫视前注意,以进一步使其输出与人类扫描路径对齐。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/7bd7bcd8366c/jovi-25-2-6-f015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/2c388714ddd3/jovi-25-2-6-f001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/d6294c4a49a6/jovi-25-2-6-f002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/db582bbc46b2/jovi-25-2-6-f003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/4f565d691bdb/jovi-25-2-6-f004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/750a0a159d52/jovi-25-2-6-f005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/1002019125ed/jovi-25-2-6-f006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/0093a1132b21/jovi-25-2-6-f007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/118c86599a25/jovi-25-2-6-f008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/447427e21dbc/jovi-25-2-6-f009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/b307058918eb/jovi-25-2-6-f010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/ab2b2ee1754c/jovi-25-2-6-f011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/4b8aa737e9bd/jovi-25-2-6-f012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/68a795430778/jovi-25-2-6-f013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/c51e7f0d40d8/jovi-25-2-6-f014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/7bd7bcd8366c/jovi-25-2-6-f015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/2c388714ddd3/jovi-25-2-6-f001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/d6294c4a49a6/jovi-25-2-6-f002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/db582bbc46b2/jovi-25-2-6-f003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/4f565d691bdb/jovi-25-2-6-f004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/750a0a159d52/jovi-25-2-6-f005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/1002019125ed/jovi-25-2-6-f006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/0093a1132b21/jovi-25-2-6-f007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/118c86599a25/jovi-25-2-6-f008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/447427e21dbc/jovi-25-2-6-f009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/b307058918eb/jovi-25-2-6-f010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/ab2b2ee1754c/jovi-25-2-6-f011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/4b8aa737e9bd/jovi-25-2-6-f012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/68a795430778/jovi-25-2-6-f013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/c51e7f0d40d8/jovi-25-2-6-f014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d0e3/11812614/7bd7bcd8366c/jovi-25-2-6-f015.jpg

相似文献

1
A robotics-inspired scanpath model reveals the importance of uncertainty and semantic object cues for gaze guidance in dynamic scenes.一种受机器人启发的扫描路径模型揭示了不确定性和语义对象线索对动态场景中注视引导的重要性。
J Vis. 2025 Feb 3;25(2):6. doi: 10.1167/jov.25.2.6.
2
Objects guide human gaze behavior in dynamic real-world scenes.物体在动态真实场景中引导人类的注视行为。
PLoS Comput Biol. 2023 Oct 26;19(10):e1011512. doi: 10.1371/journal.pcbi.1011512. eCollection 2023 Oct.
3
The roles of scene gist and spatial dependency among objects in the semantic guidance of attention in real-world scenes.场景要点及物体间空间依存关系在现实场景中注意力语义引导方面的作用。
Vision Res. 2014 Dec;105:10-20. doi: 10.1016/j.visres.2014.08.019. Epub 2014 Sep 6.
4
Attentional synchrony and the influence of viewing task on gaze behavior in static and dynamic scenes.注意力同步以及观看任务对静态和动态场景中注视行为的影响。
J Vis. 2013 Jul 17;13(8):16. doi: 10.1167/13.8.16.
5
Does gravity matter? Effects of semantic and syntactic inconsistencies on the allocation of attention during scene perception.重力重要吗?语义和句法不一致对场景感知过程中注意力分配的影响。
J Vis. 2009 Mar 27;9(3):24.1-15. doi: 10.1167/9.3.24.
6
Semantic guidance of eye movements in real-world scenes.现实场景中眼动的语义引导
Vision Res. 2011 May 25;51(10):1192-205. doi: 10.1016/j.visres.2011.03.010. Epub 2011 Mar 21.
7
Influence of semantic consistency and perceptual features on visual attention during scene viewing in toddlers.语义一致性和感知特征对幼儿场景观看时视觉注意力的影响。
Infant Behav Dev. 2017 Nov;49:248-266. doi: 10.1016/j.infbeh.2017.09.008. Epub 2017 Oct 10.
8
The interplay between gaze and consistency in scene viewing: Evidence from visual search by young and older adults.注视与场景观察一致性之间的相互作用:来自年轻和老年成年人视觉搜索的证据。
Atten Percept Psychophys. 2021 Jul;83(5):1954-1970. doi: 10.3758/s13414-021-02242-z. Epub 2021 Mar 21.
9
Complementary effects of gaze direction and early saliency in guiding fixations during free viewing.自由观看过程中注视方向和早期显著性在引导注视方面的互补作用。
J Vis. 2014 Nov 4;14(13):3. doi: 10.1167/14.13.3.
10
Attentional competition across saccadic eye movements.扫视眼动过程中的注意竞争。
Acta Psychol (Amst). 2018 Oct;190:27-37. doi: 10.1016/j.actpsy.2018.06.011. Epub 2018 Jul 6.

本文引用的文献

1
Leveraging uncertainty in collective opinion dynamics with heterogeneity.利用具有异质性的集体意见动态中的不确定性。
Sci Rep. 2024 Nov 9;14(1):27314. doi: 10.1038/s41598-024-78856-8.
2
Motor "laziness" constrains fixation selection in real-world tasks.运动“惰性”限制了现实世界任务中的注视选择。
Proc Natl Acad Sci U S A. 2024 Mar 19;121(12):e2302239121. doi: 10.1073/pnas.2302239121. Epub 2024 Mar 12.
3
Objects guide human gaze behavior in dynamic real-world scenes.物体在动态真实场景中引导人类的注视行为。
PLoS Comput Biol. 2023 Oct 26;19(10):e1011512. doi: 10.1371/journal.pcbi.1011512. eCollection 2023 Oct.
4
Predicting Visual Fixations.预测视觉注视点。
Annu Rev Vis Sci. 2023 Sep 15;9:269-291. doi: 10.1146/annurev-vision-120822-072528. Epub 2023 Jul 7.
5
The Architecture of Object-Based Attention.基于客体的注意的架构
Psychon Bull Rev. 2023 Oct;30(5):1643-1667. doi: 10.3758/s13423-023-02281-7. Epub 2023 Apr 20.
6
DeepGaze III: Modeling free-viewing human scanpaths with deep learning.DeepGaze III:用深度学习模拟自由观看的人类扫视轨迹。
J Vis. 2022 Apr 6;22(5):7. doi: 10.1167/jov.22.5.7.
7
Humans represent the precision and utility of information acquired across fixations.人类代表了在注视过程中获取的信息的精度和实用性。
Sci Rep. 2022 Feb 14;12(1):2411. doi: 10.1038/s41598-022-06357-7.
8
Data assimilation in dynamical cognitive science.动力认知科学中的数据同化。
Trends Cogn Sci. 2022 Feb;26(2):99-102. doi: 10.1016/j.tics.2021.11.006. Epub 2021 Dec 28.
9
Capturing the objects of vision with neural networks.用神经网络捕捉视觉对象。
Nat Hum Behav. 2021 Sep;5(9):1127-1144. doi: 10.1038/s41562-021-01194-6. Epub 2021 Sep 20.
10
Guided Search 6.0: An updated model of visual search.引导式搜索 6.0:一种更新的视觉搜索模型。
Psychon Bull Rev. 2021 Aug;28(4):1060-1092. doi: 10.3758/s13423-020-01859-9. Epub 2021 Feb 5.