Suppr超能文献

人工智能聊天机器人是否符合基于证据的癌症筛查建议?

Are AI chatbots concordant with evidence-based cancer screening recommendations?

作者信息

Nickel Brooke, Ayre Julie, Marinovich M Luke, Smith David P, Chiam Karen, Lee Christoph I, Wilt Timothy J, Taba Melody, McCaffery Kirsten, Houssami Nehmat

机构信息

Sydney Health Literacy Lab, Sydney School of Public Health, Faculty of Medicine and Health, The University of Sydney, Australia; Wiser Healthcare, Sydney School of Public Health, Faculty of Medicine and Health, The University of Sydney, Australia.

Sydney Health Literacy Lab, Sydney School of Public Health, Faculty of Medicine and Health, The University of Sydney, Australia.

出版信息

Patient Educ Couns. 2025 May;134:108677. doi: 10.1016/j.pec.2025.108677. Epub 2025 Jan 21.

Abstract

OBJECTIVE

This study aimed to assess whether information from AI chatbots on benefits and harms of breast and prostate cancer screening were concordant with evidence-based cancer screening recommendations.

METHODS

Seven unique prompts (four breast cancer; three prostate cancer) were presented to ChatGPT in March 2024. A total of 60 criteria (30 breast; 30 prostate) were used to assess the concordance of information. Concordance was scored between 0 and 2 against the United States Preventive Services Task Force (USPSTF) breast and prostate cancer screening recommendations independently by international cancer screening experts.

RESULTS

43 of 60 (71.7 %) criteria were completely concordant, 3 (5 %) were moderately concordant and 14 (23.3 %) were not concordant or not present, with most of the non-concordant criteria (9 of 14, 64.3 %) being from prompts for the oldest age groups. ChatGPT hallucinations (i.e., completely made up, non-sensical or irrelevant information) were found in 9 of 60 criteria (15 %).

CONCLUSIONS

ChatGPT provided information mostly concordant with USPSTF breast and prostate cancer screening recommendations, however, important gaps exist. These findings provide insights into the role of AI to communicate cancer screening benefits and harms and hold increased relevance for periods of guideline change.

PRACTICE IMPLICATIONS

AI generated information on cancer screening should be taken in conjunction with official screening recommendations and/or information from clinicians.

摘要

目的

本研究旨在评估人工智能聊天机器人提供的关于乳腺癌和前列腺癌筛查利弊的信息是否与基于证据的癌症筛查建议一致。

方法

2024年3月向ChatGPT提出了七个独特的提示(四个关于乳腺癌;三个关于前列腺癌)。总共使用60条标准(30条关于乳腺癌;30条关于前列腺癌)来评估信息的一致性。国际癌症筛查专家根据美国预防服务工作组(USPSTF)的乳腺癌和前列腺癌筛查建议,独立对一致性进行0至2分的评分。

结果

60条标准中有43条(71.7%)完全一致,3条(5%)中度一致,14条(23.3%)不一致或未提及,大多数不一致的标准(14条中的9条,64.3%)来自针对最年长年龄组的提示。在60条标准中有9条(15%)发现了ChatGPT的幻觉(即完全编造、无意义或不相关的信息)。

结论

ChatGPT提供的信息大多与USPSTF的乳腺癌和前列腺癌筛查建议一致,然而,仍存在重要差距。这些发现为人工智能在传达癌症筛查利弊方面的作用提供了见解,并且在指南变更期间具有更高的相关性。

实践意义

人工智能生成的关于癌症筛查的信息应与官方筛查建议和/或临床医生提供的信息相结合。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验