Suppr超能文献

急诊医学应用中人工智能生成与人工撰写的个人陈述的比较分析

Comparative Analysis of Artificial Intelligence-Generated and Human-Written Personal Statements in Emergency Medicine Applications.

作者信息

Steele Elizabeth, Steratore Anthony, Dilcher Brian Z, Bandi Kathryn

机构信息

Emergency Medicine, West Virginia University School of Medicine, Morgantown, USA.

Emergency Medicine, University of Pittsburgh Medical Center, Pittsburgh, USA.

出版信息

Cureus. 2025 Jul 26;17(7):e88818. doi: 10.7759/cureus.88818. eCollection 2025 Jul.

Abstract

Introduction Personal statements (PSs) have long been part of the Electronic Residency Application Service (ERAS) application; however, there exist only limited guidelines for their creation and even fewer for their role in the application review process. Applicants invest significant time in writing their PSs, and still, program directors rank PSs among the least important factors in interview and rank order list decisions. The emergence of generative artificial intelligence (AI), particularly large language models (LLMs) like ChatGPT, has introduced questions of ethics and originality across all aspects of education, particularly in the generation of free-form documents such as the PS. This study evaluates whether AI-generated PSs are distinguishable from authentic ones written by applicants and their perceived impact on residency selection. Methods Five AI-generated PSs were created using ChatGPT, incorporating applicant location, hobbies, and career background. Five de-identified, authentic PSs randomly selected from incoming emergency medicine (EM) interns were used for comparison. A Qualtrics survey was distributed electronically to the Council of Residency Directors (CORD) community. Respondents rated the PSs on writing quality, ability to convey personal attributes, and perceived influence on interview decisions. Statistical analyses (ANOVA and Wilcoxon tests) were used to assess differences between AI-generated and authentic statements. Results A total of 66 responses were collected over a two-month period. Of these, eight respondents did not regularly review ERAS applications, and 28 did not complete the survey beyond the initial question, leaving 30 responses for analysis. There were no statistically significant differences between AI-generated and authentic PSs in grammar and writing style (p = 0.5897), expression of personal attributes (p = 0.6827), overall quality (p = 0.2757), or perceived influence on interview decisions (p = 0.5457). Free-text comments reflected skepticism about the value of the PS in the selection process. Conclusion AI-generated PSs performed comparably to authentic ones, potentially further challenging the relevance of PSs in residency applications. These findings suggest an inherent lack of originality in the PS and may support re-evaluating the role of the PS and even exploring more meaningful ways to assess applicant fit in the residency selection process. Novel methods, such as structured interviews, standardized video responses, personality inventories, or situational judgment tests, may be considered to supplement the role intended for the PS.

摘要

引言

个人陈述(PS)长期以来一直是电子住院医师申请服务(ERAS)申请的一部分;然而,关于其撰写的指导方针有限,而关于其在申请审核过程中作用的指导方针则更少。申请人花费大量时间撰写个人陈述,但项目主任仍将个人陈述列为面试和排名决定中最不重要的因素之一。生成式人工智能(AI)的出现,尤其是像ChatGPT这样的大型语言模型(LLM),在教育的各个方面都引发了伦理和原创性问题,特别是在生成诸如个人陈述这样的自由形式文档时。本研究评估了人工智能生成的个人陈述是否能与申请人撰写的真实陈述区分开来,以及它们对住院医师选拔的潜在影响。

方法

使用ChatGPT创建了五份人工智能生成的个人陈述,纳入了申请人的所在地、爱好和职业背景。从即将入学的急诊医学(EM)实习生中随机选取五份匿名的真实个人陈述用于比较。通过电子方式向住院医师主任委员会(CORD)社区分发了一份Qualtrics调查问卷。受访者对个人陈述的写作质量、传达个人特质的能力以及对面试决定的感知影响进行评分。使用统计分析(方差分析和威尔科克森检验)来评估人工智能生成的陈述与真实陈述之间的差异。

结果

在两个月的时间里共收集到66份回复。其中,8名受访者不经常审核ERAS申请,28名受访者在回答初始问题后未完成调查,剩余30份回复用于分析。在语法和写作风格(p = 0.5897)、个人特质的表达(p = 0.6827)、整体质量(p = 0.2757)或对面试决定的感知影响(p = 0.5457)方面,人工智能生成的个人陈述与真实个人陈述之间没有统计学上的显著差异。自由文本评论反映了对个人陈述在选拔过程中的价值的怀疑。

结论

人工智能生成的个人陈述与真实陈述表现相当,这可能进一步挑战了个人陈述在住院医师申请中的相关性。这些发现表明个人陈述本身缺乏原创性,并可能支持重新评估个人陈述的作用,甚至探索更有意义的方式来评估申请人在住院医师选拔过程中的适应性。可以考虑采用新的方法,如结构化面试、标准化视频回复、个性量表或情景判断测试,来补充个人陈述原本 intended 的作用。 (注:原文此处intended表述有误,可能影响理解,推测应为“intended”)

相似文献

本文引用的文献

1
Predicting Surgical Resident Performance With Situational Judgment Tests.用情境判断测验预测外科住院医师的表现。
Acad Med. 2024 Aug 1;99(8):884-888. doi: 10.1097/ACM.0000000000005680. Epub 2024 Feb 27.
6
Will ChatGPT Match to Your Program?ChatGPT 是否能符合你的项目要求?
Am J Phys Med Rehabil. 2023 Jun 1;102(6):545-547. doi: 10.1097/PHM.0000000000002238. Epub 2023 Mar 13.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验