Steele Elizabeth, Steratore Anthony, Dilcher Brian Z, Bandi Kathryn
Emergency Medicine, West Virginia University School of Medicine, Morgantown, USA.
Emergency Medicine, University of Pittsburgh Medical Center, Pittsburgh, USA.
Cureus. 2025 Jul 26;17(7):e88818. doi: 10.7759/cureus.88818. eCollection 2025 Jul.
Introduction Personal statements (PSs) have long been part of the Electronic Residency Application Service (ERAS) application; however, there exist only limited guidelines for their creation and even fewer for their role in the application review process. Applicants invest significant time in writing their PSs, and still, program directors rank PSs among the least important factors in interview and rank order list decisions. The emergence of generative artificial intelligence (AI), particularly large language models (LLMs) like ChatGPT, has introduced questions of ethics and originality across all aspects of education, particularly in the generation of free-form documents such as the PS. This study evaluates whether AI-generated PSs are distinguishable from authentic ones written by applicants and their perceived impact on residency selection. Methods Five AI-generated PSs were created using ChatGPT, incorporating applicant location, hobbies, and career background. Five de-identified, authentic PSs randomly selected from incoming emergency medicine (EM) interns were used for comparison. A Qualtrics survey was distributed electronically to the Council of Residency Directors (CORD) community. Respondents rated the PSs on writing quality, ability to convey personal attributes, and perceived influence on interview decisions. Statistical analyses (ANOVA and Wilcoxon tests) were used to assess differences between AI-generated and authentic statements. Results A total of 66 responses were collected over a two-month period. Of these, eight respondents did not regularly review ERAS applications, and 28 did not complete the survey beyond the initial question, leaving 30 responses for analysis. There were no statistically significant differences between AI-generated and authentic PSs in grammar and writing style (p = 0.5897), expression of personal attributes (p = 0.6827), overall quality (p = 0.2757), or perceived influence on interview decisions (p = 0.5457). Free-text comments reflected skepticism about the value of the PS in the selection process. Conclusion AI-generated PSs performed comparably to authentic ones, potentially further challenging the relevance of PSs in residency applications. These findings suggest an inherent lack of originality in the PS and may support re-evaluating the role of the PS and even exploring more meaningful ways to assess applicant fit in the residency selection process. Novel methods, such as structured interviews, standardized video responses, personality inventories, or situational judgment tests, may be considered to supplement the role intended for the PS.
引言
个人陈述(PS)长期以来一直是电子住院医师申请服务(ERAS)申请的一部分;然而,关于其撰写的指导方针有限,而关于其在申请审核过程中作用的指导方针则更少。申请人花费大量时间撰写个人陈述,但项目主任仍将个人陈述列为面试和排名决定中最不重要的因素之一。生成式人工智能(AI)的出现,尤其是像ChatGPT这样的大型语言模型(LLM),在教育的各个方面都引发了伦理和原创性问题,特别是在生成诸如个人陈述这样的自由形式文档时。本研究评估了人工智能生成的个人陈述是否能与申请人撰写的真实陈述区分开来,以及它们对住院医师选拔的潜在影响。
方法
使用ChatGPT创建了五份人工智能生成的个人陈述,纳入了申请人的所在地、爱好和职业背景。从即将入学的急诊医学(EM)实习生中随机选取五份匿名的真实个人陈述用于比较。通过电子方式向住院医师主任委员会(CORD)社区分发了一份Qualtrics调查问卷。受访者对个人陈述的写作质量、传达个人特质的能力以及对面试决定的感知影响进行评分。使用统计分析(方差分析和威尔科克森检验)来评估人工智能生成的陈述与真实陈述之间的差异。
结果
在两个月的时间里共收集到66份回复。其中,8名受访者不经常审核ERAS申请,28名受访者在回答初始问题后未完成调查,剩余30份回复用于分析。在语法和写作风格(p = 0.5897)、个人特质的表达(p = 0.6827)、整体质量(p = 0.2757)或对面试决定的感知影响(p = 0.5457)方面,人工智能生成的个人陈述与真实个人陈述之间没有统计学上的显著差异。自由文本评论反映了对个人陈述在选拔过程中的价值的怀疑。
结论
人工智能生成的个人陈述与真实陈述表现相当,这可能进一步挑战了个人陈述在住院医师申请中的相关性。这些发现表明个人陈述本身缺乏原创性,并可能支持重新评估个人陈述的作用,甚至探索更有意义的方式来评估申请人在住院医师选拔过程中的适应性。可以考虑采用新的方法,如结构化面试、标准化视频回复、个性量表或情景判断测试,来补充个人陈述原本 intended 的作用。 (注:原文此处intended表述有误,可能影响理解,推测应为“intended”)