• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

MedVH:迈向医学背景下大型视觉语言模型幻觉的系统评估

MedVH: Toward Systematic Evaluation of Hallucination for Large Vision Language Models in the Medical Context.

作者信息

Gu Zishan, Chen Jiayuan, Liu Fenglin, Yin Changchang, Zhang Ping

机构信息

Department of Computer Science and Engineering, The Ohio State University, Columbus, Ohio 43210, USA.

Department of Biomedical Informatics, The Ohio State University, Columbus, Ohio 43210, USA.

出版信息

Adv Intell Syst. 2025 Jul 21. doi: 10.1002/aisy.202500255.

DOI:10.1002/aisy.202500255
PMID:40843006
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12363988/
Abstract

Large vision language models (LVLMs) have achieved superior performance on natural image and text tasks, inspiring extensive fine-tuning research. However, their robustness against hallucination in clinical contexts remains understudied. We propose the Medical Visual Hallucination Test (MedVH), a novel evaluation framework assessing hallucination tendencies in both medical-specific and general-purpose LVLMs. MedVH encompasses six tasks targeting medical hallucinations, including two traditional tasks and four novel tasks formatted as multi-choice visual question answering and long response generation. Our extensive experiments with six evaluation metrics reveal that medical LVLMs, despite promising performance on standard medical tasks, are particularly susceptible to hallucinations-often more so than general models. This raises significant concerns about domain-specific model reliability. For real-world applications, medical LVLMs must accurately integrate medical knowledge while maintaining robust reasoning to prevent hallucination. We explore mitigation methods without model-specific fine-tuning, including prompt engineering and collaboration between general and domain-specific models. Our work provides a foundation for future evaluation studies. The dataset is available at PhysioNet: https://physionet.org/content/medvh.

摘要

大型视觉语言模型(LVLMs)在自然图像和文本任务上取得了卓越的性能,这激发了广泛的微调研究。然而,它们在临床环境中抵御幻觉的能力仍未得到充分研究。我们提出了医学视觉幻觉测试(MedVH),这是一个新颖的评估框架,用于评估医学专用和通用LVLMs中的幻觉倾向。MedVH包含六项针对医学幻觉的任务,包括两项传统任务和四项格式为多项选择视觉问答和长响应生成的新任务。我们使用六种评估指标进行的广泛实验表明,医学LVLMs尽管在标准医学任务上表现出色,但特别容易产生幻觉——通常比通用模型更容易产生幻觉。这引发了对特定领域模型可靠性的重大担忧。对于实际应用,医学LVLMs必须在准确整合医学知识的同时保持强大的推理能力,以防止产生幻觉。我们探索了无需特定模型微调的缓解方法,包括提示工程以及通用模型和特定领域模型之间的协作。我们的工作为未来的评估研究奠定了基础。该数据集可在PhysioNet上获取:https://physionet.org/content/medvh 。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/0b6a5fab91af/nihms-2099209-f0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/37468e734bb7/nihms-2099209-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/a5be84cc0389/nihms-2099209-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/b7f290014750/nihms-2099209-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/412b59ee41bc/nihms-2099209-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/a8c93d653efa/nihms-2099209-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/e504f7929b61/nihms-2099209-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/0b6a5fab91af/nihms-2099209-f0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/37468e734bb7/nihms-2099209-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/a5be84cc0389/nihms-2099209-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/b7f290014750/nihms-2099209-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/412b59ee41bc/nihms-2099209-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/a8c93d653efa/nihms-2099209-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/e504f7929b61/nihms-2099209-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0177/12363988/0b6a5fab91af/nihms-2099209-f0007.jpg

相似文献

1
MedVH: Toward Systematic Evaluation of Hallucination for Large Vision Language Models in the Medical Context.MedVH:迈向医学背景下大型视觉语言模型幻觉的系统评估
Adv Intell Syst. 2025 Jul 21. doi: 10.1002/aisy.202500255.
2
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
3
Short-Term Memory Impairment短期记忆障碍
4
Multi-model assurance analysis showing large language models are highly vulnerable to adversarial hallucination attacks during clinical decision support.多模型保证分析表明,在临床决策支持过程中,大语言模型极易受到对抗性幻觉攻击。
Commun Med (Lond). 2025 Aug 2;5(1):330. doi: 10.1038/s43856-025-01021-3.
5
Sexual Harassment and Prevention Training性骚扰与预防培训
6
Evaluating the effectiveness of biomedical fine-tuning for large language models on clinical tasks.评估生物医学微调对大语言模型在临床任务上的有效性。
J Am Med Inform Assoc. 2025 Jun 1;32(6):1015-1024. doi: 10.1093/jamia/ocaf045.
7
Evaluating the Reasoning Capabilities of Large Language Models for Medical Coding and Hospital Readmission Risk Stratification: Zero-Shot Prompting Approach.评估大型语言模型在医学编码和医院再入院风险分层方面的推理能力:零样本提示方法。
J Med Internet Res. 2025 Jul 30;27:e74142. doi: 10.2196/74142.
8
UpGen: Unleashing Potential of Foundation Models for Training-Free Camouflage Detection via Generative Models.UpGen:通过生成模型释放基础模型在无训练伪装检测方面的潜力。
IEEE Trans Image Process. 2025;34:5400-5413. doi: 10.1109/TIP.2025.3599101.
9
The quantity, quality and findings of network meta-analyses evaluating the effectiveness of GLP-1 RAs for weight loss: a scoping review.评估胰高血糖素样肽-1受体激动剂(GLP-1 RAs)减肥效果的网状Meta分析的数量、质量及结果:一项范围综述
Health Technol Assess. 2025 Jun 25:1-73. doi: 10.3310/SKHT8119.
10
A dataset and benchmark for hospital course summarization with adapted large language models.一个用于医院病程总结的数据集和基准测试,采用了适配的大语言模型。
J Am Med Inform Assoc. 2025 Mar 1;32(3):470-479. doi: 10.1093/jamia/ocae312.

本文引用的文献

1
BioInstruct: instruction tuning of large language models for biomedical natural language processing.BioInstruct:用于生物医学自然语言处理的大型语言模型的指令调整。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1821-1832. doi: 10.1093/jamia/ocae122.
2
PMC-LLaMA: toward building open-source language models for medicine.PMC-LLaMA:为医学构建开源语言模型的努力。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1833-1843. doi: 10.1093/jamia/ocae045.
3
Large language models encode clinical knowledge.大语言模型编码临床知识。
Nature. 2023 Aug;620(7972):172-180. doi: 10.1038/s41586-023-06291-2. Epub 2023 Jul 12.
4
Deep learning for chest X-ray analysis: A survey.深度学习在胸部 X 光分析中的应用:综述。
Med Image Anal. 2021 Aug;72:102125. doi: 10.1016/j.media.2021.102125. Epub 2021 Jun 5.
5
COVID-CheXNet: hybrid deep learning framework for identifying COVID-19 virus in chest X-rays images.COVID-CheXNet:用于在胸部X光图像中识别新冠病毒的混合深度学习框架。
Soft comput. 2023;27(5):2657-2672. doi: 10.1007/s00500-020-05424-3. Epub 2020 Nov 21.
6
A dataset of clinically generated visual questions and answers about radiology images.一个包含临床生成的关于放射影像的视觉问题和答案的数据集。
Sci Data. 2018 Nov 20;5:180251. doi: 10.1038/sdata.2018.251.