• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

深度 SAGA:一种基于深度学习的眼动追踪数据自动注视点标注系统。

Deep-SAGA: a deep-learning-based system for automatic gaze annotation from eye-tracking data.

机构信息

School of Sport, Exercise and Rehabilitation Sciences, The University of Birmingham, Edgbaston, Birmingham, B15 2TT, UK.

School of Psychology, The University of Birmingham, Birmingham, UK.

出版信息

Behav Res Methods. 2023 Apr;55(3):1372-1391. doi: 10.3758/s13428-022-01833-4. Epub 2022 Jun 1.

DOI:10.3758/s13428-022-01833-4
PMID:35650384
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10126076/
Abstract

With continued advancements in portable eye-tracker technology liberating experimenters from the restraints of artificial laboratory designs, research can now collect gaze data from real-world, natural navigation. However, the field lacks a robust method for achieving this, as past approaches relied upon the time-consuming manual annotation of eye-tracking data, while previous attempts at automation lack the necessary versatility for in-the-wild navigation trials consisting of complex and dynamic scenes. Here, we propose a system capable of informing researchers of where and what a user's gaze is focused upon at any one time. The system achieves this by first running footage recorded on a head-mounted camera through a deep-learning-based object detection algorithm called Masked Region-based Convolutional Neural Network (Mask R-CNN). The algorithm's output is combined with frame-by-frame gaze coordinates measured by an eye-tracking device synchronized with the head-mounted camera to detect and annotate, without any manual intervention, what a user looked at for each frame of the provided footage. The effectiveness of the presented methodology was legitimized by a comparison between the system output and that of manual coders. High levels of agreement between the two validated the system as a preferable data collection technique as it was capable of processing data at a significantly faster rate than its human counterpart. Support for the system's practicality was then further demonstrated via a case study exploring the mediatory effects of gaze behaviors on an environment-driven attentional bias.

摘要

随着便携式眼动追踪技术的不断进步,实验人员摆脱了人工实验室设计的限制,可以从真实世界的自然导航中收集注视数据。然而,该领域缺乏一种强大的方法来实现这一目标,因为过去的方法依赖于对眼动追踪数据进行耗时的手动注释,而以前的自动化尝试缺乏针对由复杂和动态场景组成的野外导航试验的必要通用性。在这里,我们提出了一个能够实时告知研究人员用户的注视焦点在哪里和注视什么的系统。该系统通过首先将头戴式摄像机记录的视频片段通过基于深度学习的对象检测算法,即掩模区域卷积神经网络(Mask R-CNN)运行来实现这一目标。该算法的输出与通过与头戴式摄像机同步的眼动追踪设备测量的逐帧注视坐标相结合,无需任何手动干预,即可检测和注释提供的视频片段中的每一帧中用户注视的内容。该方法的有效性通过系统输出与手动编码结果之间的比较得到验证。两者之间高度的一致性验证了该系统作为一种更优的数据收集技术的有效性,因为它能够以比人工编码快得多的速度处理数据。该系统的实用性通过一项案例研究得到了进一步证明,该研究探索了注视行为对环境驱动的注意偏向的中介作用。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/0099d4a1437d/13428_2022_1833_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/024fdd6d4e09/13428_2022_1833_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/748a8bf17f53/13428_2022_1833_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/3480fe3f2f1b/13428_2022_1833_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/5e8c5e2bb2be/13428_2022_1833_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/f530aedddbe6/13428_2022_1833_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/81b0369ffc1e/13428_2022_1833_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/8e35d203fea9/13428_2022_1833_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/10b94c52f734/13428_2022_1833_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/7838ee975217/13428_2022_1833_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/8dd5aacfb472/13428_2022_1833_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/0099d4a1437d/13428_2022_1833_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/024fdd6d4e09/13428_2022_1833_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/748a8bf17f53/13428_2022_1833_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/3480fe3f2f1b/13428_2022_1833_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/5e8c5e2bb2be/13428_2022_1833_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/f530aedddbe6/13428_2022_1833_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/81b0369ffc1e/13428_2022_1833_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/8e35d203fea9/13428_2022_1833_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/10b94c52f734/13428_2022_1833_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/7838ee975217/13428_2022_1833_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/8dd5aacfb472/13428_2022_1833_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8df3/10126076/0099d4a1437d/13428_2022_1833_Fig11_HTML.jpg

相似文献

1
Deep-SAGA: a deep-learning-based system for automatic gaze annotation from eye-tracking data.深度 SAGA:一种基于深度学习的眼动追踪数据自动注视点标注系统。
Behav Res Methods. 2023 Apr;55(3):1372-1391. doi: 10.3758/s13428-022-01833-4. Epub 2022 Jun 1.
2
Mobile Eye-Tracking Data Analysis Using Object Detection via YOLO v4.基于 YOLO v4 的目标检测的移动眼动追踪数据分析。
Sensors (Basel). 2021 Nov 18;21(22):7668. doi: 10.3390/s21227668.
3
Hybrid Eye-Tracking on a Smartphone with CNN Feature Extraction and an Infrared 3D Model.智能手机上的混合眼动追踪:基于 CNN 特征提取和红外 3D 模型。
Sensors (Basel). 2020 Jan 19;20(2):543. doi: 10.3390/s20020543.
4
Review and Evaluation of Eye Movement Event Detection Algorithms.眼动事件检测算法的回顾与评估。
Sensors (Basel). 2022 Nov 15;22(22):8810. doi: 10.3390/s22228810.
5
Calibration-Free Mobile Eye-Tracking Using Corneal Imaging.基于角膜成像的无标定移动眼动追踪。
Sensors (Basel). 2024 Feb 15;24(4):1237. doi: 10.3390/s24041237.
6
DGaze: CNN-Based Gaze Prediction in Dynamic Scenes.DGaze:基于 CNN 的动态场景中的注视预测。
IEEE Trans Vis Comput Graph. 2020 May;26(5):1902-1911. doi: 10.1109/TVCG.2020.2973473. Epub 2020 Feb 13.
7
Automatic Visual Attention Detection for Mobile Eye Tracking Using Pre-Trained Computer Vision Models and Human Gaze.基于预训练计算机视觉模型和人眼注视的移动眼动追踪自动视觉注意力检测。
Sensors (Basel). 2021 Jun 16;21(12):4143. doi: 10.3390/s21124143.
8
Strategies for enhancing automatic fixation detection in head-mounted eye tracking.增强头戴式眼动追踪中自动注视点检测的策略。
Behav Res Methods. 2024 Sep;56(6):6276-6298. doi: 10.3758/s13428-024-02360-0. Epub 2024 Apr 9.
9
When I Look into Your Eyes: A Survey on Computer Vision Contributions for Human Gaze Estimation and Tracking.当我凝视你的双眼:计算机视觉在人类视线估计和追踪中的应用综述。
Sensors (Basel). 2020 Jul 3;20(13):3739. doi: 10.3390/s20133739.
10
Gaze Tracking and Point Estimation Using Low-Cost Head-Mounted Devices.使用低成本头戴式设备进行注视跟踪和点估计。
Sensors (Basel). 2020 Mar 30;20(7):1917. doi: 10.3390/s20071917.

引用本文的文献

1
Algorithmic gaze annotation for mobile eye-tracking.用于移动眼动追踪的算法注视标注
Behav Res Methods. 2025 Sep 17;57(10):290. doi: 10.3758/s13428-025-02803-2.
2
eyeNotate: Interactive Annotation of Mobile Eye Tracking Data Based on Few-Shot Image Classification.EyeNotate:基于少样本图像分类的移动眼动追踪数据交互式标注
J Eye Mov Res. 2025 Jul 7;18(4):27. doi: 10.3390/jemr18040027. eCollection 2025 Aug.
3
Deep learning quantifies pathologists' visual patterns for whole slide image diagnosis.深度学习量化病理学家用于全切片图像诊断的视觉模式。

本文引用的文献

1
MYNursingHome: A fully-labelled image dataset for indoor object classification.我的养老院:用于室内物体分类的全标注图像数据集。
Data Brief. 2020 Sep 3;32:106268. doi: 10.1016/j.dib.2020.106268. eCollection 2020 Oct.
2
ilastik: interactive machine learning for (bio)image analysis.ilastik:用于(生物)图像处理的交互式机器学习。
Nat Methods. 2019 Dec;16(12):1226-1232. doi: 10.1038/s41592-019-0582-9. Epub 2019 Sep 30.
3
MCIndoor20000: A fully-labeled image dataset to advance indoor objects detection.MCIndoor20000:一个用于推进室内物体检测的全标注图像数据集。
Nat Commun. 2025 Jul 1;16(1):5493. doi: 10.1038/s41467-025-60307-1.
4
I-MPN: inductive message passing network for efficient human-in-the-loop annotation of mobile eye tracking data.I-MPN:用于移动眼动追踪数据高效人工参与标注的归纳消息传递网络。
Sci Rep. 2025 Apr 23;15(1):14192. doi: 10.1038/s41598-025-94593-y.
5
Infant EEG microstate dynamics relate to fine-grained patterns of infant attention during naturalistic play with caregivers.婴儿脑电图微状态动力学与婴儿在与照顾者自然玩耍期间的精细注意力模式相关。
Proc Natl Acad Sci U S A. 2025 Mar 18;122(11):e2414636122. doi: 10.1073/pnas.2414636122. Epub 2025 Mar 13.
6
The fundamentals of eye tracking part 4: Tools for conducting an eye tracking study.眼动追踪基础 第4部分:进行眼动追踪研究的工具。
Behav Res Methods. 2025 Jan 6;57(1):46. doi: 10.3758/s13428-024-02529-7.
7
MYFix: Automated Fixation Annotation of Eye-Tracking Videos.MYFix:眼动追踪视频的自动固定注释。
Sensors (Basel). 2024 Apr 23;24(9):2666. doi: 10.3390/s24092666.
8
Reading in the city: mobile eye-tracking and evaluation of text in an everyday setting.城市中的阅读:日常环境下的移动眼动追踪与文本评估
Front Psychol. 2023 Oct 18;14:1205913. doi: 10.3389/fpsyg.2023.1205913. eCollection 2023.
Data Brief. 2018 Jan 3;17:71-75. doi: 10.1016/j.dib.2017.12.047. eCollection 2018 Apr.
4
Mobile gaze tracking system for outdoor walking behavioral studies.用于户外行走行为研究的移动视线跟踪系统。
J Vis. 2016;16(3):27. doi: 10.1167/16.3.27.
5
Active vision in passive locomotion: real-world free viewing in infants and adults.被动运动中的主动视觉:婴儿和成人在现实世界中的自由观看
Dev Sci. 2015 Sep;18(5):736-50. doi: 10.1111/desc.12251. Epub 2014 Nov 28.
6
Brain mechanisms for emotional influences on perception and attention: what is magic and what is not.大脑对感知和注意的情绪影响的机制:什么是魔术,什么不是。
Biol Psychol. 2013 Mar;92(3):492-512. doi: 10.1016/j.biopsycho.2012.02.007. Epub 2012 Feb 21.
7
The where, what and when of gaze allocation in the lab and the natural environment.实验室及自然环境中注视分配的地点、内容和时间。
Vision Res. 2011 Sep 1;51(17):1920-31. doi: 10.1016/j.visres.2011.07.002. Epub 2011 Jul 23.
8
Personal space regulation by the human amygdala.人类杏仁核对个人空间的调节。
Nat Neurosci. 2009 Oct;12(10):1226-7. doi: 10.1038/nn.2381. Epub 2009 Aug 30.
9
Eye movements during information processing tasks: individual differences and cultural effects.信息处理任务中的眼动:个体差异与文化影响。
Vision Res. 2007 Sep;47(21):2714-26. doi: 10.1016/j.visres.2007.05.007. Epub 2007 Jul 5.
10
Flanker effects with faces may depend on perceptual as well as emotional differences.面部的侧翼效应可能取决于感知差异以及情感差异。
Emotion. 2006 Feb;6(1):28-39. doi: 10.1037/1528-3542.6.1.28.