文献检索文档翻译深度研究
Suppr Zotero 插件Zotero 插件
邀请有礼套餐&价格历史记录

新学期,新优惠

限时优惠:9月1日-9月22日

30天高级会员仅需29元

1天体验卡首发特惠仅需5.99元

了解详情
不再提醒
插件&应用
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
高级版
套餐订阅购买积分包
AI 工具
文献检索文档翻译深度研究
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2025

使用 ChatGPT-3 检索临床放射学信息的信息和参考文献的准确性。

Accuracy of Information and References Using ChatGPT-3 for Retrieval of Clinical Radiological Information.

机构信息

Department of Diagnostic Imaging, Division of Neuroradiology, The Hospital for Sick Children, Toronto, Canada.

Department of Medical Imaging, University of Toronto, Toronto, Canada.

出版信息

Can Assoc Radiol J. 2024 Feb;75(1):69-73. doi: 10.1177/08465371231171125. Epub 2023 Apr 20.


DOI:10.1177/08465371231171125
PMID:37078489
Abstract

To assess the accuracy of answers provided by ChatGPT-3 when prompted with questions from the daily routine of radiologists and to evaluate the text response when ChatGPT-3 was prompted to provide references for a given answer. ChatGPT-3 (San Francisco, OpenAI) is an artificial intelligence chatbot based on a large language model (LLM) that has been designed to generate human-like text. A total of 88 questions were submitted to ChatGPT-3 using textual prompt. These 88 questions were equally dispersed across 8 subspecialty areas of radiology. The responses provided by ChatGPT-3 were assessed for correctness by cross-checking them with peer-reviewed, PubMed-listed references. In addition, the references provided by ChatGPT-3 were evaluated for authenticity. A total of 59 of 88 responses (67%) to radiological questions were correct, while 29 responses (33%) had errors. Out of 343 references provided, only 124 references (36.2%) were available through internet search, while 219 references (63.8%) appeared to be generated by ChatGPT-3. When examining the 124 identified references, only 47 references (37.9%) were considered to provide enough background to correctly answer 24 questions (37.5%). In this pilot study, ChatGPT-3 provided correct responses to questions from the daily clinical routine of radiologists in only about two thirds, while the remainder of responses contained errors. The majority of provided references were not found and only a minority of the provided references contained the correct information to answer the question. Caution is advised when using ChatGPT-3 to retrieve radiological information.

摘要

评估 ChatGPT-3 在回答放射科医生日常工作中的问题时的准确性,并评估 ChatGPT-3 在提供给定答案的参考文献时的文本回复。ChatGPT-3(旧金山,OpenAI)是一种基于大型语言模型(LLM)的人工智能聊天机器人,旨在生成类人文本。总共向 ChatGPT-3 提交了 88 个问题,使用文本提示。这 88 个问题平均分布在放射科的 8 个亚专科领域。通过与同行评议的 PubMed 列出的参考文献交叉检查,评估 ChatGPT-3 提供的回复的正确性。此外,还评估了 ChatGPT-3 提供的参考文献的真实性。对放射学问题的 88 个回复中的 59 个(67%)是正确的,而 29 个回复(33%)存在错误。在提供的 343 个参考文献中,只有 124 个参考文献(36.2%)可以通过互联网搜索获得,而 219 个参考文献(63.8%)似乎是由 ChatGPT-3 生成的。在检查 124 个已识别的参考文献时,只有 47 个参考文献(37.9%)被认为提供了足够的背景信息,可以正确回答 24 个问题(37.5%)。在这项初步研究中,ChatGPT-3 仅在大约三分之二的情况下正确回答了放射科医生日常临床工作中的问题,而其余的回复则包含错误。提供的参考文献大多数未找到,只有少数提供的参考文献包含正确的信息来回答问题。在使用 ChatGPT-3 检索放射学信息时应谨慎。

相似文献

[1]
Accuracy of Information and References Using ChatGPT-3 for Retrieval of Clinical Radiological Information.

Can Assoc Radiol J. 2024-2

[2]
ChatGPT and oral cancer: a study on informational reliability.

BMC Oral Health. 2025-1-17

[3]
Evaluating ChatGPT Responses on Thyroid Nodules for Patient Education.

Thyroid. 2024-3

[4]
Evaluating the reference accuracy of large language models in radiology: a comparative study across subspecialties.

Diagn Interv Radiol. 2025-5-12

[5]
ChatGPT and Artificial Intelligence in Transplantation Research: Is It Always Correct?

Cureus. 2023-7-19

[6]
Dr. Google vs. Dr. ChatGPT: Exploring the Use of Artificial Intelligence in Ophthalmology by Comparing the Accuracy, Safety, and Readability of Responses to Frequently Asked Patient Questions Regarding Cataracts and Cataract Surgery.

Semin Ophthalmol. 2024-8

[7]
A Review of ChatGPT as a Reliable Source of Scientific Information Regarding Endodontic Local Anesthesia.

J Endod. 2025-5

[8]
How AI Responds to Common Lung Cancer Questions: ChatGPT vs Google Bard.

Radiology. 2023-6

[9]
An Assessment of the Accuracy and Consistency of ChatGPT in the Management of Midshaft Clavicle Fractures.

Cureus. 2025-4-8

[10]
Assessing the accuracy and quality of artificial intelligence (AI) chatbot-generated responses in making patient-specific drug-therapy and healthcare-related decisions.

BMC Med Inform Decis Mak. 2024-12-24

引用本文的文献

[1]
AI Methods Tailored to Influenza, RSV, HIV, and SARS-CoV-2: A Focused Review.

Pathogens. 2025-7-30

[2]
Optimizing patient education for radioactive iodine therapy and the role of ChatGPT incorporating chain-of-thought technique: ChatGPT questionnaire.

Digit Health. 2025-7-7

[3]
Evaluating the Accuracy of Privacy-Preserving Large Language Models in Calculating the Spinal Instability Neoplastic Score (SINS).

Cancers (Basel). 2025-6-20

[4]
ChatGPT for mechanobiology and medicine: A perspective.

Mechanobiol Med. 2023-7-5

[5]
AI-Driven Information for Relatives of Patients with Malignant Middle Cerebral Artery Infarction: A Preliminary Validation Study Using GPT-4o.

Brain Sci. 2025-4-11

[6]
Can deepseek and ChatGPT be used in the diagnosis of oral pathologies?

BMC Oral Health. 2025-4-25

[7]
Large Language Models for Chatbot Health Advice Studies: A Systematic Review.

JAMA Netw Open. 2025-2-3

[8]
ChatGPT and oral cancer: a study on informational reliability.

BMC Oral Health. 2025-1-17

[9]
Assessing the Quality, Readability, and Acceptability of AI-Generated Information in Plastic and Aesthetic Surgery.

Cureus. 2024-11-17

[10]
Comparing the Accuracy of Two Generated Large Language Models in Identifying Health-Related Rumors or Misconceptions and the Applicability in Health Science Popularization: Proof-of-Concept Study.

JMIR Form Res. 2024-12-2

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

推荐工具

医学文档翻译智能文献检索