• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于迭代表示学习的人类扫视轨迹预测。

Predicting Human Saccadic Scanpaths Based on Iterative Representation Learning.

出版信息

IEEE Trans Image Process. 2019 Jul;28(7):3502-3515. doi: 10.1109/TIP.2019.2897966. Epub 2019 Feb 7.

DOI:10.1109/TIP.2019.2897966
PMID:30735998
Abstract

Visual attention is a dynamic process of scene exploration and information acquisition. However, existing research on attention modeling has concentrated on estimating static salient locations. In contrast, dynamic attributes presented by saccade have not been well explored in previous attention models. In this paper, we address the problem of saccadic scanpath prediction by introducing an iterative representation learning framework. Within the framework, saccade can be interpreted as an iterative process of predicting one fixation according to the current representation and updating the representation based on the gaze shift. In the predicting phase, we propose a Bayesian definition of saccade to combine the influence of perceptual residual and spatial location on the selection of fixations. In implementation, we compute the representation error of an autoencoder-based network to measure perceptual residuals of each area. Simultaneously, we integrate saccade amplitude and center-weighted mechanism to model the influence of spatial location. Based on estimating the influence of two parts, the final fixation is defined as the point with the largest posterior probability of gaze shift. In the updating phase, we update the representation pattern for the subsequent calculation by retraining the network with samples extracted around the current fixation. In the experiments, the proposed model can replicate the fundamental properties of psychophysics in visual search. In addition, it can achieve superior performance on several benchmark eye-tracking data sets.

摘要

视觉注意是一种动态的场景探索和信息获取过程。然而,现有的注意力建模研究主要集中在估计静态显著位置上。相比之下,先前的注意力模型中并没有很好地探索眼跳呈现的动态属性。在本文中,我们通过引入迭代表示学习框架来解决眼跳扫描路径预测问题。在该框架内,眼跳可以被解释为根据当前表示来预测一个注视点,并根据注视转移来更新表示的迭代过程。在预测阶段,我们提出了一种基于贝叶斯的眼跳定义,将感知残差和空间位置对注视点选择的影响结合起来。在实现方面,我们计算基于自动编码器的网络的表示误差,以衡量每个区域的感知残差。同时,我们整合眼跳幅度和中心加权机制来模拟空间位置的影响。基于估计这两部分的影响,最终的注视点被定义为注视转移后具有最大后验概率的点。在更新阶段,我们通过用当前注视点周围提取的样本重新训练网络,为后续计算更新表示模式。在实验中,所提出的模型可以复制视觉搜索中心理物理学的基本特性。此外,它在几个基准眼动数据集上也能取得优异的性能。

相似文献

1
Predicting Human Saccadic Scanpaths Based on Iterative Representation Learning.基于迭代表示学习的人类扫视轨迹预测。
IEEE Trans Image Process. 2019 Jul;28(7):3502-3515. doi: 10.1109/TIP.2019.2897966. Epub 2019 Feb 7.
2
Saccadic model of eye movements for free-viewing condition.自由观看条件下眼动的扫视模型。
Vision Res. 2015 Nov;116(Pt B):152-64. doi: 10.1016/j.visres.2014.12.026. Epub 2015 Feb 24.
3
Evaluation of Saccadic Scanpath Prediction: Subjective Assessment Database and Recurrent Neural Network Based Metric.扫视轨迹预测评估:基于主观评估数据库和递归神经网络的度量。
IEEE Trans Pattern Anal Mach Intell. 2021 Dec;43(12):4378-4395. doi: 10.1109/TPAMI.2020.3002168. Epub 2021 Nov 3.
4
Image complexity analysis with scanpath identification using remote gaze estimation model.使用远程注视估计模型进行扫描路径识别的图像复杂度分析
Multimed Tools Appl. 2020;79(33-34):24393-24412. doi: 10.1007/s11042-020-09117-9. Epub 2020 Jun 20.
5
Gravitational Laws of Focus of Attention.注意力焦点的引力法则。
IEEE Trans Pattern Anal Mach Intell. 2020 Dec;42(12):2983-2995. doi: 10.1109/TPAMI.2019.2920636. Epub 2020 Nov 3.
6
Saccadic context indicates information processing within visual fixations: evidence from event-related potentials and eye-movements analysis of the distractor effect.扫视语境指示视觉固视内的信息处理:来自相关事件电位和眼动分析的分心效应证据。
Int J Psychophysiol. 2011 Apr;80(1):54-62. doi: 10.1016/j.ijpsycho.2011.01.013. Epub 2011 Feb 1.
7
Scanpath estimation based on foveated image saliency.基于中央凹图像显著性的扫描路径估计。
Cogn Process. 2017 Feb;18(1):87-95. doi: 10.1007/s10339-016-0781-6. Epub 2016 Oct 14.
8
Saccade-synchronized rapid attention shifts in macaque visual cortical area MT.猕猴视觉皮层MT区中与扫视同步的快速注意力转移
Nat Commun. 2018 Mar 6;9(1):958. doi: 10.1038/s41467-018-03398-3.
9
Using temporally aligned event-related potentials for the investigation of attention shifts prior to and during saccades.使用时间对齐的事件相关电位来研究扫视之前和期间的注意力转移。
Neuropsychologia. 2016 Nov;92:129-141. doi: 10.1016/j.neuropsychologia.2016.03.035. Epub 2016 Apr 5.
10
Predictability of spatial and non-spatial target properties improves perception in the pre-saccadic interval.空间和非空间目标属性的可预测性可改善扫视前间隔期的感知。
Vision Res. 2013 Oct 18;91:93-101. doi: 10.1016/j.visres.2013.08.003. Epub 2013 Aug 13.

引用本文的文献

1
DeepGaze III: Modeling free-viewing human scanpaths with deep learning.DeepGaze III:用深度学习模拟自由观看的人类扫视轨迹。
J Vis. 2022 Apr 6;22(5):7. doi: 10.1167/jov.22.5.7.
2
Saccade Landing Point Prediction Based on Fine-Grained Learning Method.基于细粒度学习方法的扫视着陆点预测
IEEE Access. 2021;9:52474-52484. doi: 10.1109/access.2021.3070511. Epub 2021 Apr 1.