• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

眼动引导的多模态融合:迈向使用可解释人工智能的自适应学习框架

Eye-Guided Multimodal Fusion: Toward an Adaptive Learning Framework Using Explainable Artificial Intelligence.

作者信息

Moradizeyveh Sahar, Hanif Ambreen, Liu Sidong, Qi Yuankai, Beheshti Amin, Di Ieva Antonio

机构信息

Computational NeuroSurgery (CNS) Lab, Macquarie Medical School, Faculty of Medicine, Health and Human Sciences, Macquarie University, Sydney 2113, Australia.

Centre for Applied Artificial Intelligence, School of Computing, Faculty of Science and Engineering, Macquarie University, Sydney 2113, Australia.

出版信息

Sensors (Basel). 2025 Jul 24;25(15):4575. doi: 10.3390/s25154575.

DOI:10.3390/s25154575
PMID:40807742
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12349219/
Abstract

Interpreting diagnostic imaging and identifying clinically relevant features remain challenging tasks, particularly for novice radiologists who often lack structured guidance and expert feedback. To bridge this gap, we propose an Eye-Gaze Guided Multimodal Fusion framework that leverages expert eye-tracking data to enhance learning and decision-making in medical image interpretation. By integrating chest X-ray (CXR) images with expert fixation maps, our approach captures radiologists' visual attention patterns and highlights regions of interest (ROIs) critical for accurate diagnosis. The fusion model utilizes a shared backbone architecture to jointly process image and gaze modalities, thereby minimizing the impact of noise in fixation data. We validate the system's interpretability using Gradient-weighted Class Activation Mapping (Grad-CAM) and assess both classification performance and explanation alignment with expert annotations. Comprehensive evaluations, including robustness under gaze noise and expert clinical review, demonstrate the framework's effectiveness in improving model reliability and interpretability. This work offers a promising pathway toward intelligent, human-centered AI systems that support both diagnostic accuracy and medical training.

摘要

解读诊断成像并识别临床相关特征仍然是具有挑战性的任务,特别是对于那些常常缺乏结构化指导和专家反馈的新手放射科医生而言。为了弥补这一差距,我们提出了一种眼动引导的多模态融合框架,该框架利用专家眼动追踪数据来加强医学图像解读中的学习和决策。通过将胸部X光(CXR)图像与专家注视图相结合,我们的方法捕捉放射科医生的视觉注意力模式,并突出显示对准确诊断至关重要的感兴趣区域(ROI)。融合模型利用共享骨干架构来联合处理图像和注视模态,从而将注视数据中的噪声影响降至最低。我们使用梯度加权类激活映射(Grad-CAM)验证系统的可解释性,并评估分类性能以及与专家注释的解释一致性。包括注视噪声下的稳健性和专家临床审查在内的综合评估证明了该框架在提高模型可靠性和可解释性方面的有效性。这项工作为支持诊断准确性和医学培训的智能、以人为本的人工智能系统提供了一条充满希望的途径。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/b886397b40c5/sensors-25-04575-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/78f5b720cd22/sensors-25-04575-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/1826de99c4c0/sensors-25-04575-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/2b1dea418105/sensors-25-04575-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/f6e3fede11f3/sensors-25-04575-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/d2748506e2af/sensors-25-04575-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/9153071f5b02/sensors-25-04575-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/b886397b40c5/sensors-25-04575-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/78f5b720cd22/sensors-25-04575-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/1826de99c4c0/sensors-25-04575-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/2b1dea418105/sensors-25-04575-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/f6e3fede11f3/sensors-25-04575-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/d2748506e2af/sensors-25-04575-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/9153071f5b02/sensors-25-04575-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/af38/12349219/b886397b40c5/sensors-25-04575-g007.jpg

相似文献

1
Eye-Guided Multimodal Fusion: Toward an Adaptive Learning Framework Using Explainable Artificial Intelligence.眼动引导的多模态融合:迈向使用可解释人工智能的自适应学习框架
Sensors (Basel). 2025 Jul 24;25(15):4575. doi: 10.3390/s25154575.
2
CXR-MultiTaskNet a unified deep learning framework for joint disease localization and classification in chest radiographs.CXR-MultiTaskNet:一种用于胸部X光片中疾病联合定位与分类的统一深度学习框架。
Sci Rep. 2025 Aug 31;15(1):32022. doi: 10.1038/s41598-025-16669-z.
3
Shedding light on ai in radiology: A systematic review and taxonomy of eye gaze-driven interpretability in deep learning.解读放射学中的人工智能:深度学习中眼动驱动可解释性的系统综述与分类法
Eur J Radiol. 2024 Mar;172:111341. doi: 10.1016/j.ejrad.2024.111341. Epub 2024 Feb 1.
4
A deep learning approach to direct immunofluorescence pattern recognition in autoimmune bullous diseases.深度学习方法在自身免疫性大疱性疾病中的直接免疫荧光模式识别。
Br J Dermatol. 2024 Jul 16;191(2):261-266. doi: 10.1093/bjd/ljae142.
5
Are Artificial Intelligence Models Listening Like Cardiologists? Bridging the Gap Between Artificial Intelligence and Clinical Reasoning in Heart-Sound Classification Using Explainable Artificial Intelligence.人工智能模型能像心脏病专家一样“聆听”吗?利用可解释人工智能弥合人工智能与心音分类临床推理之间的差距。
Bioengineering (Basel). 2025 May 22;12(6):558. doi: 10.3390/bioengineering12060558.
6
Deep Learning and Image Generator Health Tabular Data (IGHT) for Predicting Overall Survival in Patients With Colorectal Cancer: Retrospective Study.深度学习与图像生成器健康表格数据(IGHT)用于预测结直肠癌患者的总生存期:回顾性研究
JMIR Med Inform. 2025 Aug 19;13:e75022. doi: 10.2196/75022.
7
ItpCtrl-AI: End-to-end interpretable and controllable artificial intelligence by modeling radiologists' intentions.ItpCtrl-AI:通过对放射科医生的意图进行建模实现端到端可解释和可控的人工智能。
Artif Intell Med. 2025 Feb;160:103054. doi: 10.1016/j.artmed.2024.103054. Epub 2024 Dec 12.
8
Novel Artificial Intelligence-Driven Infant Meningitis Screening From High-Resolution Ultrasound Imaging.基于高分辨率超声成像的新型人工智能驱动的婴儿脑膜炎筛查
Ultrasound Med Biol. 2025 Jun 28. doi: 10.1016/j.ultrasmedbio.2025.04.009.
9
Explainable Artificial Intelligence (XAI) in the Era of Large Language Models: Applying an XAI Framework in Pediatric Ophthalmology Diagnosis using the Gemini Model.大语言模型时代的可解释人工智能(XAI):在儿科眼科诊断中使用Gemini模型应用XAI框架
AMIA Jt Summits Transl Sci Proc. 2025 Jun 10;2025:566-575. eCollection 2025.
10
Deep learning-based image classification for AI-assisted integration of pathology and radiology in medical imaging.基于深度学习的图像分类,用于医学成像中病理学与放射学的人工智能辅助整合。
Front Med (Lausanne). 2025 Jun 2;12:1574514. doi: 10.3389/fmed.2025.1574514. eCollection 2025.

本文引用的文献

1
Eye Gaze Guided Cross-Modal Alignment Network for Radiology Report Generation.用于生成放射学报告的眼动引导跨模态对齐网络
IEEE J Biomed Health Inform. 2024 Dec;28(12):7406-7419. doi: 10.1109/JBHI.2024.3422168. Epub 2024 Dec 5.
2
Analyzing Eye Paths Using Fractals.利用分形分析眼径。
Adv Neurobiol. 2024;36:827-848. doi: 10.1007/978-3-031-47606-8_42.
3
Eye-Gaze-Guided Vision Transformer for Rectifying Shortcut Learning.眼动引导视觉Transformer 用于纠正捷径学习。
IEEE Trans Med Imaging. 2023 Nov;42(11):3384-3394. doi: 10.1109/TMI.2023.3287572. Epub 2023 Oct 27.
4
Explainable AI in medical imaging: An overview for clinical practitioners - Beyond saliency-based XAI approaches.医学成像中的可解释人工智能:临床从业者概述——超越基于显著性的可解释人工智能方法
Eur J Radiol. 2023 May;162:110786. doi: 10.1016/j.ejrad.2023.110786. Epub 2023 Mar 20.
5
Skill Characterisation of Sonographer Gaze Patterns during Second Trimester Clinical Fetal Ultrasounds using Time Curves.使用时间曲线对孕中期临床胎儿超声检查中超声医师注视模式的技能特征分析
Proc Eye Track Res Appl Symp. 2022 Jun;2022. doi: 10.1145/3517031.3529637. Epub 2022 Jun 8.
6
Changes in Radiologists' Gaze Patterns Against Lung X-rays with Different Abnormalities: a Randomized Experiment.不同异常肺部 X 光片下放射科医生注视模式的变化:一项随机实验。
J Digit Imaging. 2023 Jun;36(3):767-775. doi: 10.1007/s10278-022-00760-2. Epub 2023 Jan 9.
7
REFLACX, a dataset of reports and eye-tracking data for localization of abnormalities in chest x-rays.REFLACX,一个包含报告和眼动数据的数据集,用于定位胸部 X 光片中的异常。
Sci Data. 2022 Jun 18;9(1):350. doi: 10.1038/s41597-022-01441-z.
8
Explainable artificial intelligence (XAI) in deep learning-based medical image analysis.深度学习在医学影像分析中的可解释人工智能(XAI)。
Med Image Anal. 2022 Jul;79:102470. doi: 10.1016/j.media.2022.102470. Epub 2022 May 4.
9
Recent advances and clinical applications of deep learning in medical image analysis.深度学习在医学图像分析中的最新进展和临床应用。
Med Image Anal. 2022 Jul;79:102444. doi: 10.1016/j.media.2022.102444. Epub 2022 Apr 4.
10
On Smart Gaze Based Annotation of Histopathology Images for Training of Deep Convolutional Neural Networks.基于智能凝视的组织病理学图像标注在深度卷积神经网络训练中的应用
IEEE J Biomed Health Inform. 2022 Jul;26(7):3025-3036. doi: 10.1109/JBHI.2022.3148944. Epub 2022 Jul 1.