• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

使用具有频谱损失的生成对抗网络(GAN)对眼睛注视速度轨迹进行建模以提高保真度。

Modeling eye gaze velocity trajectories using GANs with spectral loss for enhanced fidelity.

作者信息

Bhandari Shailendra, Lencastre Pedro, Mathema Rujeena, Szorkovszky Alexander, Yazidi Anis, Lind Pedro G

机构信息

Department of Computer Science, OsloMet - Oslo Metropolitan University, P.O. Box 4 St. Olavs plass, N-0130, Oslo, Norway.

OsloMet Artificial Intelligence Lab, Pilestredet 52, N-0166, Oslo, Norway.

出版信息

Sci Rep. 2025 Jun 6;15(1):19929. doi: 10.1038/s41598-025-05286-5.

DOI:10.1038/s41598-025-05286-5
PMID:40481082
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12144148/
Abstract

Accurate modeling of eye gaze dynamics is essential for advancement in human-computer interaction, neurological diagnostics, and cognitive research. Traditional generative models like Markov models often fail to capture the complex temporal dependencies and distributional nuance inherent in eye gaze trajectories data. This study introduces a Generative Adversarial Network (GAN) framework employing Long Short-Term Memory (LSTM) and Convolutional Neural Network (CNN) generators and discriminators to generate high-fidelity synthetic eye gaze velocity trajectories. We conducted a comprehensive evaluation of four GAN architectures: CNN-CNN, LSTM-CNN, CNN-LSTM, and LSTM-LSTM-trained under two conditions: using only adversarial loss ([Formula: see text]) and using a weighted combination of adversarial and spectral losses. Our findings reveal that the LSTM-CNN architecture trained with this new loss function exhibits the closest alignment to the real data distribution, effectively capturing both the distribution tails and the intricate temporal dependencies. The inclusion of spectral regularization significantly enhances the GANs' ability to replicate the spectral characteristics of eye gaze movements, leading to a more stable learning process and improved data fidelity. Comparative analysis with a Hidden Markov Model (HMM) optimized to four hidden states further highlights the advantages of the LSTM-CNN GAN. Statistical metrics show that the HMM-generated data significantly diverges from the real data in terms of mean, standard deviation, skewness, and kurtosis. In contrast, the LSTM-CNN model closely matches the real data across these statistics, affirming its capacity to model the complexity of eye gaze dynamics effectively. These results position the spectrally regularized LSTM-CNN GAN as a robust tool for generating synthetic eye gaze velocity data with high fidelity. Its ability to accurately replicate both the distributional and temporal properties of real data holds significant potential for applications in simulation environments, training systems, and the development of advanced eye-tracking technologies, ultimately contributing to more naturalistic and responsive human-computer interactions.

摘要

准确建模眼睛注视动态对于人机交互、神经诊断和认知研究的进展至关重要。像马尔可夫模型这样的传统生成模型往往无法捕捉眼睛注视轨迹数据中固有的复杂时间依赖性和分布细微差别。本研究引入了一种生成对抗网络(GAN)框架,该框架采用长短期记忆(LSTM)和卷积神经网络(CNN)生成器与判别器来生成高保真的合成眼睛注视速度轨迹。我们对四种GAN架构进行了全面评估:CNN-CNN、LSTM-CNN、CNN-LSTM和LSTM-LSTM,在两种条件下进行训练:仅使用对抗损失([公式:见文本])以及使用对抗损失和频谱损失的加权组合。我们的研究结果表明,使用这种新损失函数训练的LSTM-CNN架构与真实数据分布的对齐最紧密,有效地捕捉了分布尾部和复杂的时间依赖性。频谱正则化的纳入显著增强了GAN复制眼睛注视运动频谱特征的能力,导致更稳定的学习过程和更高的数据保真度。与优化为四个隐藏状态的隐马尔可夫模型(HMM)的比较分析进一步突出了LSTM-CNN GAN的优势。统计指标表明,HMM生成的数据在均值、标准差、偏度和峰度方面与真实数据有显著差异。相比之下,LSTM-CNN模型在这些统计数据上与真实数据紧密匹配,证实了其有效建模眼睛注视动态复杂性的能力。这些结果将频谱正则化的LSTM-CNN GAN定位为生成高保真合成眼睛注视速度数据的强大工具。它准确复制真实数据的分布和时间属性的能力在模拟环境、训练系统以及先进眼动追踪技术的开发中具有巨大潜力,最终有助于实现更自然和响应性更强的人机交互。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/27bb1c79794a/41598_2025_5286_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/2b4d0cb02099/41598_2025_5286_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/53f6dfda26fe/41598_2025_5286_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/ce104dd28587/41598_2025_5286_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/2885a05d6963/41598_2025_5286_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/3b7350d28fe7/41598_2025_5286_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/27bb1c79794a/41598_2025_5286_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/2b4d0cb02099/41598_2025_5286_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/53f6dfda26fe/41598_2025_5286_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/ce104dd28587/41598_2025_5286_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/2885a05d6963/41598_2025_5286_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/3b7350d28fe7/41598_2025_5286_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ee95/12144148/27bb1c79794a/41598_2025_5286_Fig6_HTML.jpg

相似文献

1
Modeling eye gaze velocity trajectories using GANs with spectral loss for enhanced fidelity.使用具有频谱损失的生成对抗网络(GAN)对眼睛注视速度轨迹进行建模以提高保真度。
Sci Rep. 2025 Jun 6;15(1):19929. doi: 10.1038/s41598-025-05286-5.
2
EEG-Based Feature Classification Combining 3D-Convolutional Neural Networks with Generative Adversarial Networks for Motor Imagery.基于脑电的特征分类,结合三维卷积神经网络和生成对抗网络,用于运动想象。
J Integr Neurosci. 2024 Aug 20;23(8):153. doi: 10.31083/j.jin2308153.
3
Deep causal speech enhancement and recognition using efficient long-short term memory Recurrent Neural Network.利用高效长短时记忆递归神经网络进行深度因果语音增强和识别。
PLoS One. 2024 Jan 3;19(1):e0291240. doi: 10.1371/journal.pone.0291240. eCollection 2024.
4
Gaze Tracking Based on Concatenating Spatial-Temporal Features.基于时空特征拼接的注视跟踪。
Sensors (Basel). 2022 Jan 11;22(2):545. doi: 10.3390/s22020545.
5
Eye Gaze-based Early Intent Prediction Utilizing CNN-LSTM.利用卷积神经网络-长短期记忆网络的基于眼动注视的早期意图预测
Annu Int Conf IEEE Eng Med Biol Soc. 2019 Jul;2019:1310-1313. doi: 10.1109/EMBC.2019.8857054.
6
Electrocardiogram generation with a bidirectional LSTM-CNN generative adversarial network.使用双向 LSTM-CNN 生成对抗网络生成心电图。
Sci Rep. 2019 May 1;9(1):6734. doi: 10.1038/s41598-019-42516-z.
7
Generative adversarial network based synthetic data training model for lightweight convolutional neural networks.用于轻量级卷积神经网络的基于生成对抗网络的合成数据训练模型。
Multimed Tools Appl. 2023 May 20:1-23. doi: 10.1007/s11042-023-15747-6.
8
Enhancing classification of cells procured from bone marrow aspirate smears using generative adversarial networks and sequential convolutional neural network.利用生成对抗网络和序列卷积神经网络增强骨髓穿刺涂片获取的细胞分类。
Comput Methods Programs Biomed. 2022 Sep;224:107019. doi: 10.1016/j.cmpb.2022.107019. Epub 2022 Jul 10.
9
Use of one-dimensional CNN for input data size reduction in LSTM for improved computational efficiency and accuracy in hourly rainfall-runoff modeling.一维卷积神经网络(CNN)在长短期记忆网络(LSTM)中用于输入数据降维,以提高小时降雨径流建模的计算效率和准确性。
J Environ Manage. 2024 May;359:120931. doi: 10.1016/j.jenvman.2024.120931. Epub 2024 Apr 27.
10
A Robust Framework for Data Generative and Heart Disease Prediction Based on Efficient Deep Learning Models.基于高效深度学习模型的数据生成与心脏病预测的稳健框架
Diagnostics (Basel). 2022 Nov 22;12(12):2899. doi: 10.3390/diagnostics12122899.

本文引用的文献

1
TNT: An Interpretable Tree-Network-Tree Learning Framework using Knowledge Distillation.TNT:一种使用知识蒸馏的可解释树-网络-树学习框架。
Entropy (Basel). 2020 Oct 24;22(11):1203. doi: 10.3390/e22111203.
2
Utilizing Amari-Alpha Divergence to Stabilize the Training of Generative Adversarial Networks.利用阿马里-阿尔法散度来稳定生成对抗网络的训练。
Entropy (Basel). 2020 Apr 4;22(4):410. doi: 10.3390/e22040410.
3
On the Jensen-Shannon Symmetrization of Distances Relying on Abstract Means.关于基于抽象均值的距离的詹森 - 香农对称化
Entropy (Basel). 2019 May 11;21(5):485. doi: 10.3390/e21050485.
4
Estimation and simulation of foraging trips in land-based marine predators.陆基海洋捕食者觅食旅行的估计和模拟。
Ecology. 2017 Jul;98(7):1932-1944. doi: 10.1002/ecy.1880. Epub 2017 Jun 12.
5
Spatial memory in foraging games.觅食游戏中的空间记忆。
Cognition. 2016 Mar;148:85-96. doi: 10.1016/j.cognition.2015.12.015. Epub 2016 Jan 2.
6
Flexible and practical modeling of animal telemetry data: hidden Markov models and extensions.灵活实用的动物遥测数据建模:隐马尔可夫模型及其扩展。
Ecology. 2012 Nov;93(11):2336-42. doi: 10.1890/11-2241.1.
7
Live Speech Driven Head-and-Eye Motion Generators.实时语音驱动的头眼运动生成器。
IEEE Trans Vis Comput Graph. 2012 Nov;18(11):1902-14. doi: 10.1109/TVCG.2012.74. Epub 2012 Mar 6.
8
Eye guidance in natural vision: reinterpreting salience.自然视觉中的眼动引导:重新诠释显著性
J Vis. 2011 May 27;11(5):5. doi: 10.1167/11.5.5.
9
State-space models of individual animal movement.个体动物运动的状态空间模型。
Trends Ecol Evol. 2008 Feb;23(2):87-94. doi: 10.1016/j.tree.2007.10.009. Epub 2008 Jan 11.