Suppr超能文献

ChatGPT能骗过面试吗?整形外科住院医师申请中的人工智能个人陈述:一项比较研究。

Can ChatGPT Fool the Match? Artificial Intelligence Personal Statements for Plastic Surgery Residency Applications: A Comparative Study.

作者信息

Chen Jeffrey, Tao Brendan K, Park Shihyun, Bovill Esta

机构信息

Michael G. DeGroote School of Medicine, Faculty of Health Sciences, McMaster University, Hamilton, Ontario, Canada.

Faculty of Medicine, University of British Columbia, Vancouver, British Columbia, Canada.

出版信息

Plast Surg (Oakv). 2024 Jul 23:22925503241264832. doi: 10.1177/22925503241264832.

Abstract

Personal statements can be decisive in Canadian residency applications. With the rise in AI technology, ethical concerns regarding authenticity and originality become more pressing. This study explores the capability of ChatGPT in producing personal statements for plastic surgery residency that match the quality of statements written by successful applicants. ChatGPT was utilized to generate a cohort of personal statements for CaRMS (Canadian Residency Matching Service) to compare with previously successful Plastic Surgery applications. Each AI-generated and human-written statement was randomized and anonymized prior to assessment. Two retired members of the plastic surgery residency selection committee from the University of British Columbia, evaluated these on a 0 to 10 scale and provided a binary response judging whether each statement was AI or human written. Statistical analysis included Welch 2-sample tests and Cohen's Kappa for agreement. Twenty-two personal statements (11 AI-generated by ChatGPT and 11 human-written) were evaluated. The overall mean scores were 7.48 (SD 0.932) and 7.68 (SD 0.716), respectively, with no significant difference between AI and human groups ( = .4129). The average accuracy in distinguishing between human and AI letters was 65.9%. The Cohen's Kappa value was 0.374. ChatGPT can generate personal statements for plastic surgery residency applications with quality indistinguishable from human-written counterparts, as evidenced by the lack of significant scoring difference and moderate accuracy in discrimination by experienced surgeons. These findings highlight the evolving role of AI and the need for updated evaluative criteria or guidelines in the residency application process.

摘要

个人陈述在加拿大住院医师申请中可能起决定性作用。随着人工智能技术的兴起,关于真实性和原创性的伦理问题变得更加紧迫。本研究探讨了ChatGPT生成与成功申请者所写质量相当的整形外科住院医师申请个人陈述的能力。利用ChatGPT为加拿大住院医师匹配服务(CaRMS)生成一组个人陈述,以便与之前成功的整形外科申请进行比较。在评估之前,将每篇人工智能生成的和人工撰写的陈述进行随机化和匿名化处理。两名来自英属哥伦比亚大学的整形外科住院医师选拔委员会退休成员,以0到10分的尺度对这些陈述进行评估,并给出一个二元反应,判断每篇陈述是由人工智能还是人工撰写的。统计分析包括韦尔奇双样本检验和用于一致性分析的科恩卡方检验。共评估了22篇个人陈述(11篇由ChatGPT生成,11篇由人工撰写)。总体平均得分分别为7.48(标准差0.932)和7.68(标准差0.716),人工智能组和人工组之间无显著差异(P = 0.4129)。区分人工和人工智能信件的平均准确率为65.9%。科恩卡方值为0.374。ChatGPT能够生成质量与人工撰写的难以区分的整形外科住院医师申请个人陈述,经验丰富的外科医生在评分上缺乏显著差异以及辨别准确率一般就证明了这一点。这些发现凸显了人工智能不断演变的作用以及在住院医师申请过程中更新评估标准或指南的必要性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f2df/11561920/498bc3a9f35c/10.1177_22925503241264832-fig1.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验