• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Can Artificial Intelligence Fool Residency Selection Committees? Analysis of Personal Statements by Real Applicants and Generative AI, a Randomized, Single-Blind Multicenter Study.人工智能能骗过住院医师选拔委员会吗?真实申请者与生成式人工智能个人陈述分析,一项随机、单盲多中心研究。
JB JS Open Access. 2024 Oct 24;9(4). doi: 10.2106/JBJS.OA.24.00028. eCollection 2024 Oct-Dec.
2
Assessing the Reproducibility of the Structured Abstracts Generated by ChatGPT and Bard Compared to Human-Written Abstracts in the Field of Spine Surgery: Comparative Analysis.评估 ChatGPT 和 Bard 生成的结构化摘要与脊柱外科领域人类撰写的摘要在可重复性方面的比较:对比分析。
J Med Internet Res. 2024 Jun 26;26:e52001. doi: 10.2196/52001.
3
Distinguishing Authentic Voices in the Age of ChatGPT: Comparing AI-Generated and Applicant-Written Personal Statements for Plastic Surgery Residency Application.ChatGPT 时代的真实声音辨别:比较整形外科学住院医师申请的 AI 生成和申请人撰写的个人陈述。
Ann Plast Surg. 2023 Sep 1;91(3):324-325. doi: 10.1097/SAP.0000000000003653.
4
Residency Application Selection Committee Discriminatory Ability in Identifying Artificial Intelligence-Generated Personal Statements.住院医师申请选拔委员会识别人工智能生成个人陈述的歧视能力。
J Surg Educ. 2024 Jun;81(6):780-785. doi: 10.1016/j.jsurg.2024.02.009. Epub 2024 Apr 27.
5
Detecting Artificial Intelligence-Generated Personal Statements in Professional Physical Therapist Education Program Applications: A Lexical Analysis.在专业物理治疗师教育项目申请中检测人工智能生成的个人陈述:一项词汇分析。
Phys Ther. 2024 Apr 2;104(4). doi: 10.1093/ptj/pzae006.
6
Generative artificial intelligence chatbots may provide appropriate informational responses to common vascular surgery questions by patients.生成式人工智能聊天机器人可能会为患者关于常见血管外科问题提供恰当的信息性回复。
Vascular. 2025 Feb;33(1):229-237. doi: 10.1177/17085381241240550. Epub 2024 Mar 18.
7
Evaluation of the Current Status of Artificial Intelligence for Endourology Patient Education: A Blind Comparison of ChatGPT and Google Bard Against Traditional Information Resources.评估人工智能在泌尿内镜患者教育中的现状:ChatGPT 和 Google Bard 与传统信息资源的盲对比。
J Endourol. 2024 Aug;38(8):843-851. doi: 10.1089/end.2023.0696. Epub 2024 May 17.
8
Evaluation of the Performance of Generative AI Large Language Models ChatGPT, Google Bard, and Microsoft Bing Chat in Supporting Evidence-Based Dentistry: Comparative Mixed Methods Study.评估生成式 AI 大语言模型 ChatGPT、Google Bard 和 Microsoft Bing Chat 在支持循证牙科方面的性能:比较混合方法研究。
J Med Internet Res. 2023 Dec 28;25:e51580. doi: 10.2196/51580.
9
Does using artificial intelligence take the person out of personal statements? We can't tell.使用人工智能是否会使个人陈述失去人情味?我们无法判断。
Surgery. 2024 Dec;176(6):1610-1616. doi: 10.1016/j.surg.2024.08.018. Epub 2024 Sep 19.
10
Digital Ink and Surgical Dreams: Perceptions of Artificial Intelligence-Generated Essays in Residency Applications.数字墨水与手术梦想:对人工智能生成论文在住院医师申请中的看法。
J Surg Res. 2024 Sep;301:504-511. doi: 10.1016/j.jss.2024.06.020. Epub 2024 Jul 22.

引用本文的文献

1
Artificial Intelligence vs Human Authorship in Spine Surgery Fellowship Personal Statements: Can ChatGPT Outperform Applicants?脊柱外科住院医师个人陈述中的人工智能与人类撰写:ChatGPT 能比申请者表现更出色吗?
Global Spine J. 2025 May 20:21925682251344248. doi: 10.1177/21925682251344248.

本文引用的文献

1
Comparing scientific abstracts generated by ChatGPT to real abstracts with detectors and blinded human reviewers.使用检测器和不知情的人类评审员,将ChatGPT生成的科学摘要与真实摘要进行比较。
NPJ Digit Med. 2023 Apr 26;6(1):75. doi: 10.1038/s41746-023-00819-6.
2
Will ChatGPT Match to Your Program?ChatGPT 是否能符合你的项目要求?
Am J Phys Med Rehabil. 2023 Jun 1;102(6):545-547. doi: 10.1097/PHM.0000000000002238. Epub 2023 Mar 13.
3
Perspectives of Orthopedic Surgery Program Directors on the USMLE Step 1 Scoring Change.骨科外科学项目主任对 USMLE 第 1 阶段评分变化的看法。
Orthopedics. 2022 Sep-Oct;45(5):e257-e262. doi: 10.3928/01477447-20220425-03. Epub 2022 Apr 29.
4
Perceptions of USMLE Step 1 Pass/Fail Score Reporting Among Orthopedic Surgery Residency Program Directors.骨科住院医师项目主任对美国医师执照考试(USMLE)第 1 步通过/不通过评分报告的看法。
Orthopedics. 2022 Jan-Feb;45(1):e30-e34. doi: 10.3928/01477447-20211124-08. Epub 2021 Dec 2.
5
Evaluating the Standardized Letter of Recommendation Form in Applicants to Orthopaedic Surgery Residency.评估骨科住院医师申请人标准化推荐信表。
J Am Acad Orthop Surg. 2020 Oct 1;28(19):814-822. doi: 10.5435/JAAOS-D-19-00423.
6
Use of Standardized Letters of Recommendation for Orthopaedic Surgery Residency Applications: A Single-Institution Retrospective Review.使用标准化推荐信申请骨科住院医师培训:单机构回顾性研究。
J Bone Joint Surg Am. 2020 Feb 19;102(4):e14. doi: 10.2106/JBJS.19.00130.
7
Matching in Orthopaedic Surgery.骨科手术中的匹配。
J Am Acad Orthop Surg. 2020 Feb 15;28(4):135-144. doi: 10.5435/JAAOS-D-19-00313.
8
Have personal statements become impersonal? An evaluation of personal statements in anesthesiology residency applications.个人陈述是否变得缺乏个性?对麻醉学住院医师申请中个人陈述的评估。
J Clin Anesth. 2010 Aug;22(5):346-51. doi: 10.1016/j.jclinane.2009.10.007.

人工智能能骗过住院医师选拔委员会吗?真实申请者与生成式人工智能个人陈述分析,一项随机、单盲多中心研究。

Can Artificial Intelligence Fool Residency Selection Committees? Analysis of Personal Statements by Real Applicants and Generative AI, a Randomized, Single-Blind Multicenter Study.

作者信息

Lum Zachary C, Guntupalli Lohitha, Saiz Augustine M, Leshikar Holly, Le Hai V, Meehan John P, Huish Eric G

机构信息

Department of Surgery, Kiran Patel School of Osteopathic and Allopathic Medicine, Nova Southeastern University, Davie, Florida.

Department of Orthopaedic Surgery, School of Medicine, University of California: Davis, Sacramento, California.

出版信息

JB JS Open Access. 2024 Oct 24;9(4). doi: 10.2106/JBJS.OA.24.00028. eCollection 2024 Oct-Dec.

DOI:10.2106/JBJS.OA.24.00028
PMID:39450246
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11498924/
Abstract

INTRODUCTION

The potential capabilities of generative artificial intelligence (AI) tools have been relatively unexplored, particularly in the realm of creating personalized statements for medical students applying to residencies. This study aimed to investigate the ability of generative AI, specifically ChatGPT and Google BARD, to generate personal statements and assess whether faculty on residency selection committees could (1) evaluate differences between real and AI statements and (2) determine differences based on 13 defined and specific metrics of a personal statement.

METHODS

Fifteen real personal statements were used to generate 15 unique and distinct personal statements from ChatGPT and BARD each, resulting in a total of 45 statements. Statements were then randomized, blinded, and presented to a group of faculty reviewers on residency selection committees. Reviewers assessed the statements by 14 metrics including if the personal statement was AI-generated or real. Comparison of all metrics was performed.

RESULTS

Faculty correctly identified 88% (79/90) real statements, 90% (81/90) BARD, and 44% (40/90) ChatGPT statements. Accuracy of identifying real and BARD statements was 89%, but this dropped to 74% when including ChatGPT. In addition, the accuracy did not increase as faculty members reviewed more personal statements (area under the curve [AUC] 0.498, p = 0.966). BARD performed poorer than both real and ChatGPT across all metrics (p < 0.001). Comparing real with ChatGPT, there was no difference in most metrics, except for Personal Interests, Reasons for Choosing Residency, Career Goals, Compelling Nature and Originality, and all favoring the real personal statements (p = 0.001, p = 0.002, p < 0.001, p < 0.001, and p < 0.001, respectively).

CONCLUSION

Faculty members accurately identified real and BARD statements, but ChatGPT deceived them 56% of the time. Although AI can craft convincing statements that are sometimes indistinguishable from real ones, replicating the humanistic experience, personal nuances, and individualistic elements found in real personal statements is difficult. Residency selection committees might want to prioritize these particular metrics while assessing personal statements, given the growing capabilities of AI in this arena.

CLINICAL RELEVANCE

Residency selection committees may want to prioritize certain metrics unique to the human element such as personal interests, reasons for choosing residency, career goals, compelling nature, and originality when evaluating personal statements.

摘要

引言

生成式人工智能(AI)工具的潜在能力尚未得到充分探索,尤其是在为申请住院医师培训项目的医学生撰写个性化陈述方面。本研究旨在调查生成式AI,特别是ChatGPT和谷歌BARD生成个人陈述的能力,并评估住院医师选拔委员会的教员是否能够(1)评估真实陈述与AI生成陈述之间的差异,以及(2)根据个人陈述的13项明确且具体的指标确定差异。

方法

使用15篇真实的个人陈述,分别从ChatGPT和BARD生成15篇独特且不同的个人陈述,共得到45篇陈述。然后将这些陈述随机化、遮蔽,并呈现给住院医师选拔委员会的一组教员评审员。评审员通过14项指标评估这些陈述,包括个人陈述是否由AI生成或真实。对所有指标进行了比较。

结果

教员正确识别出88%(79/90)的真实陈述、90%(81/90)的BARD生成陈述和44%(40/90)的ChatGPT生成陈述。识别真实陈述和BARD生成陈述的准确率为89%,但纳入ChatGPT后降至74%。此外,随着教员评审更多个人陈述,准确率并未提高(曲线下面积[AUC]为0.498,p = 0.966)。在所有指标上,BARD的表现均不如真实陈述和ChatGPT(p < 0.001)。将真实陈述与ChatGPT生成陈述进行比较,除个人兴趣、选择住院医师培训项目的原因、职业目标、吸引力和原创性外多数指标无差异,且这些方面均更有利于真实的个人陈述(分别为p = 0.001、p = 0.002、p < 0.001、p < 0.001和p < 0.001)。

结论

教员能够准确识别真实陈述和BARD生成的陈述,但ChatGPT有56%的时间能骗过他们。尽管AI可以撰写令人信服的陈述,有时与真实陈述难以区分,但复制真实个人陈述中的人文体验、个人细微差别和个性化元素却很困难。鉴于AI在这一领域的能力不断增强,住院医师选拔委员会在评估个人陈述时可能希望优先考虑这些特定指标。

临床意义

住院医师选拔委员会在评估个人陈述时,可能希望优先考虑某些独特的人文元素指标,如个人兴趣、选择住院医师培训项目的原因、职业目标、吸引力和原创性。