• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

COCO-Search18 数据集用于预测目标导向注意力控制。

COCO-Search18 fixation dataset for predicting goal-directed attention control.

机构信息

Department of Psychology, Stony Brook University, New York, USA.

Department of Computer Science, Stony Brook University, New York, USA.

出版信息

Sci Rep. 2021 Apr 22;11(1):8776. doi: 10.1038/s41598-021-87715-9.

DOI:10.1038/s41598-021-87715-9
PMID:33888734
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8062491/
Abstract

Attention control is a basic behavioral process that has been studied for decades. The currently best models of attention control are deep networks trained on free-viewing behavior to predict bottom-up attention control - saliency. We introduce COCO-Search18, the first dataset of laboratory-quality goal-directed behavior large enough to train deep-network models. We collected eye-movement behavior from 10 people searching for each of 18 target-object categories in 6202 natural-scene images, yielding [Formula: see text] 300,000 search fixations. We thoroughly characterize COCO-Search18, and benchmark it using three machine-learning methods: a ResNet50 object detector, a ResNet50 trained on fixation-density maps, and an inverse-reinforcement-learning model trained on behavioral search scanpaths. Models were also trained/tested on images transformed to approximate a foveated retina, a fundamental biological constraint. These models, each having a different reliance on behavioral training, collectively comprise the new state-of-the-art in predicting goal-directed search fixations. Our expectation is that future work using COCO-Search18 will far surpass these initial efforts, finding applications in domains ranging from human-computer interactive systems that can anticipate a person's intent and render assistance to the potentially early identification of attention-related clinical disorders (ADHD, PTSD, phobia) based on deviation from neurotypical fixation behavior.

摘要

注意控制是一个基本的行为过程,已经研究了几十年。目前,用于预测自下而上注意控制——显著性的最佳注意控制模型是基于自由观看行为训练的深度网络。我们引入了 COCO-Search18,这是第一个用于训练深度网络模型的实验室质量目标导向行为的大型数据集。我们从 10 个人的眼动行为中收集了信息,他们在 6202 张自然场景图像中搜索了 18 个目标对象类别中的每一个,产生了[公式:见文本]30 万个搜索固定点。我们对 COCO-Search18 进行了全面的描述,并使用三种机器学习方法对其进行了基准测试:ResNet50 目标检测器、基于注视密度图训练的 ResNet50 和基于行为搜索扫描路径训练的逆强化学习模型。这些模型还在转换后的图像上进行了训练/测试,以近似注视点视网膜,这是一种基本的生物约束。这些模型,每个模型对行为训练的依赖程度不同,共同构成了预测目标导向搜索固定点的最新技术。我们期望未来使用 COCO-Search18 的工作将远远超过这些初步努力,在从能够预测人的意图并提供帮助的人机交互系统到基于偏离神经典型注视行为潜在早期识别与注意力相关的临床障碍(ADHD、PTSD、恐惧症)的应用等领域找到应用。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/b845ddc2a2fb/41598_2021_87715_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/3b5374978987/41598_2021_87715_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/ea3a2823b759/41598_2021_87715_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/29d901e2d2a4/41598_2021_87715_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/a18db4cc7066/41598_2021_87715_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/b845ddc2a2fb/41598_2021_87715_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/3b5374978987/41598_2021_87715_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/ea3a2823b759/41598_2021_87715_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/29d901e2d2a4/41598_2021_87715_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/a18db4cc7066/41598_2021_87715_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9929/8062491/b845ddc2a2fb/41598_2021_87715_Fig5_HTML.jpg

相似文献

1
COCO-Search18 fixation dataset for predicting goal-directed attention control.COCO-Search18 数据集用于预测目标导向注意力控制。
Sci Rep. 2021 Apr 22;11(1):8776. doi: 10.1038/s41598-021-87715-9.
2
Predicting Goal-directed Human Attention Using Inverse Reinforcement Learning.使用逆强化学习预测目标导向的人类注意力
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2020 Jun;2020:190-199. doi: 10.1109/cvpr42600.2020.00027. Epub 2020 Aug 5.
3
Predicting Goal-directed Attention Control Using Inverse-Reinforcement Learning.使用逆强化学习预测目标导向的注意力控制
Neuron Behav Data Anal Theory. 2021;2021. doi: 10.51628/001c.22322. Epub 2021 Apr 20.
4
Target-absent Human Attention.目标缺失时的人类注意力。
Comput Vis ECCV. 2022 Oct;13664:52-68. doi: 10.1007/978-3-031-19772-7_4. Epub 2022 Oct 23.
5
Weighting the factors affecting attention guidance during free viewing and visual search: The unexpected role of object recognition uncertainty.权衡自由观看和视觉搜索过程中影响注意力引导的因素:物体识别不确定性的意外作用。
J Vis. 2022 Mar 2;22(4):13. doi: 10.1167/jov.22.4.13.
6
What stands out in a scene? A study of human explicit saliency judgment.场景中突出的是什么?一项关于人类显性显著性判断的研究。
Vision Res. 2013 Oct 18;91:62-77. doi: 10.1016/j.visres.2013.07.016. Epub 2013 Aug 15.
7
Saliency Prediction on Omnidirectional Image With Generative Adversarial Imitation Learning.基于生成对抗模仿学习的全向图像显著度预测。
IEEE Trans Image Process. 2021;30:2087-2102. doi: 10.1109/TIP.2021.3050861. Epub 2021 Jan 21.
8
Saliency-driven system models for cell analysis with deep learning.基于深度学习的细胞分析的显著驱动系统模型。
Comput Methods Programs Biomed. 2019 Dec;182:105053. doi: 10.1016/j.cmpb.2019.105053. Epub 2019 Aug 26.
9
Saccadic model of eye movements for free-viewing condition.自由观看条件下眼动的扫视模型。
Vision Res. 2015 Nov;116(Pt B):152-64. doi: 10.1016/j.visres.2014.12.026. Epub 2015 Feb 24.
10
Complementary effects of gaze direction and early saliency in guiding fixations during free viewing.自由观看过程中注视方向和早期显著性在引导注视方面的互补作用。
J Vis. 2014 Nov 4;14(13):3. doi: 10.1167/14.13.3.

引用本文的文献

1
Integrating Bayesian and neural networks models for eye movement prediction in hybrid search.整合贝叶斯模型和神经网络模型用于混合搜索中的眼动预测。
Sci Rep. 2025 May 12;15(1):16482. doi: 10.1038/s41598-025-00272-3.
2
A simultaneous EEG and eye-tracking dataset for remote sensing object detection.用于遥感目标检测的同步脑电图和眼动追踪数据集。
Sci Data. 2025 Apr 17;12(1):651. doi: 10.1038/s41597-025-04995-w.
3
Oculomotor routines for perceptual judgments.用于知觉判断的眼球运动常规。

本文引用的文献

1
Predicting Goal-directed Human Attention Using Inverse Reinforcement Learning.使用逆强化学习预测目标导向的人类注意力
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2020 Jun;2020:190-199. doi: 10.1109/cvpr42600.2020.00027. Epub 2020 Aug 5.
2
Finding any Waldo with zero-shot invariant and efficient visual search.零样本不变且高效的视觉搜索中寻找任何 Waldo。
Nat Commun. 2018 Sep 13;9(1):3730. doi: 10.1038/s41467-018-06217-x.
3
Predicting Human Eye Fixations via an LSTM-based Saliency Attentive Model.基于长短期记忆网络的显著性注意力模型预测人眼注视点
J Vis. 2024 May 1;24(5):3. doi: 10.1167/jov.24.5.3.
4
Cues for predictive eye movements in naturalistic scenes.自然场景中预测性眼球运动的线索。
J Vis. 2023 Sep 1;23(10):12. doi: 10.1167/jov.23.10.12.
5
The neuroconnectionist research programme.神经连接主义研究计划。
Nat Rev Neurosci. 2023 Jul;24(7):431-450. doi: 10.1038/s41583-023-00705-w. Epub 2023 May 30.
6
DeepGaze III: Modeling free-viewing human scanpaths with deep learning.DeepGaze III:用深度学习模拟自由观看的人类扫视轨迹。
J Vis. 2022 Apr 6;22(5):7. doi: 10.1167/jov.22.5.7.
7
Weighting the factors affecting attention guidance during free viewing and visual search: The unexpected role of object recognition uncertainty.权衡自由观看和视觉搜索过程中影响注意力引导的因素:物体识别不确定性的意外作用。
J Vis. 2022 Mar 2;22(4):13. doi: 10.1167/jov.22.4.13.
8
Predicting Goal-directed Attention Control Using Inverse-Reinforcement Learning.使用逆强化学习预测目标导向的注意力控制
Neuron Behav Data Anal Theory. 2021;2021. doi: 10.51628/001c.22322. Epub 2021 Apr 20.
IEEE Trans Image Process. 2018 Jun 29. doi: 10.1109/TIP.2018.2851672.
4
A Deep Spatial Contextual Long-Term Recurrent Convolutional Network for Saliency Detection.基于深度空间上下文的显著性检测长短期记忆卷积网络。
IEEE Trans Image Process. 2018 Jul;27(7):3264-3274. doi: 10.1109/TIP.2018.2817047.
5
Object detection through search with a foveated visual system.通过具有中央凹视觉系统的搜索进行目标检测。
PLoS Comput Biol. 2017 Oct 9;13(10):e1005743. doi: 10.1371/journal.pcbi.1005743. eCollection 2017 Oct.
6
A comparison of scanpath comparison methods.扫描路径比较方法的比较。
Behav Res Methods. 2015 Dec;47(4):1377-1392. doi: 10.3758/s13428-014-0550-3.
7
Effects of target typicality on categorical search.目标典型性对范畴搜索的影响。
J Vis. 2014 Oct 1;14(12):1. doi: 10.1167/14.12.1.
8
Guided Search 2.0 A revised model of visual search.引导搜索 2.0:一种改进的视觉搜索模型。
Psychon Bull Rev. 1994 Jun;1(2):202-38. doi: 10.3758/BF03200774.
9
Modeling guidance and recognition in categorical search: bridging human and computer object detection.分类搜索中的建模指导与识别:架起人类与计算机目标检测的桥梁
J Vis. 2013 Oct 8;13(3):30. doi: 10.1167/13.3.30.
10
Modelling eye movements in a categorical search task.在类别搜索任务中模拟眼球运动。
Philos Trans R Soc Lond B Biol Sci. 2013 Sep 9;368(1628):20130058. doi: 10.1098/rstb.2013.0058. Print 2013 Oct 19.