• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

非英语语言的大语言模型辅助手术同意书:内容分析与可读性评估

Large Language Model-Assisted Surgical Consent Forms in Non-English Language: Content Analysis and Readability Evaluation.

作者信息

Oh Namkee, Kim Jongman, Park Sunghae, An Sunghyo, Lee Eunjin, Do Hayeon, Baik Jiyoung, Gwon Suk Min, Rhu Jinsoo, Choi Gyu-Seong, Park Seonmin, Cho Jai Young, Lee Hae Won, Lee Boram, Jeong Eun Sung, Lee Jeong-Moo, Choi YoungRok, Kwon Jieun, Kim Kyeong Deok, Kim Seok-Hwan, Chun Gwang-Sik

机构信息

Department of Surgery, Samsung Medical Center, 81 Ilwonro, Seoul, Republic of Korea, 82 1093650277.

Department of Surgery, Seoul National University Bundang Hospital, Seongnam-si, Republic of Korea.

出版信息

J Med Internet Res. 2025 Jun 19;27:e73222. doi: 10.2196/73222.

DOI:10.2196/73222
PMID:40537063
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12200805/
Abstract

BACKGROUND

Surgical consent forms convey critical information; yet, their complex language can limit patient comprehension. Large language models (LLMs) can simplify complex information and improve readability, but evidence of the impact of LLM-generated modifications on content preservation in non-English consent forms is lacking.

OBJECTIVE

This study evaluates the impact of LLM-assisted editing on the readability and content quality of surgical consent forms in Korean-particularly consent documents for standardized liver resection-across multiple institutions.

METHODS

Standardized liver resection consent forms were collected from 7 South Korean medical institutions, and these forms were simplified using ChatGPT-4o. Thereafter, readability was assessed using KReaD and Natmal indices, while text structure was evaluated based on character count, word count, sentence count, words per sentence, and difficult word ratio. Content quality was analyzed across 4 domains-risk, benefit, alternative, and overall impression-using evaluations from 7 liver resection specialists. Statistical comparisons were conducted using paired 2-sided t tests, and a linear mixed-effects model was applied to account for institutional and evaluator variability.

RESULTS

Artificial intelligence-assisted editing significantly improved readability, reducing the KReaD score from 1777 (SD 28.47) to 1335.6 (SD 59.95) (P<.001) and the Natmal score from 1452.3 (SD 88.67) to 1245.3 (SD 96.96) (P=.007). Sentence length and difficult word ratio decreased significantly, contributing to increased accessibility (P<.05). However, content quality analysis showed a decline in the risk description scores (before: 2.29, SD 0.47 vs after: 1.92, SD 0.32; P=.06) and overall impression scores (before: 2.21, SD 0.49 vs after: 1.71, SD 0.64; P=.13). The linear mixed-effects model confirmed significant reductions in risk descriptions (β₁=-0.371; P=.01) and overall impression (β₁=-0.500; P=.03), suggesting potential omissions in critical safety information. Despite this, qualitative analysis indicated that evaluators did not find explicit omissions but perceived the text as overly simplified and less professional.

CONCLUSIONS

Although LLM-assisted surgical consent forms significantly enhance readability, they may compromise certain aspects of content completeness, particularly in risk disclosure. These findings highlight the need for a balanced approach that maintains accessibility while ensuring medical and legal accuracy. Future research should include patient-centered evaluations to assess comprehension and informed decision-making as well as broader multilingual validation to determine LLM applicability across diverse health care settings.

摘要

背景

手术同意书传达关键信息;然而,其复杂的语言可能会限制患者的理解。大语言模型(LLMs)可以简化复杂信息并提高可读性,但缺乏关于大语言模型生成的修改对非英语同意书内容保留影响的证据。

目的

本研究评估大语言模型辅助编辑对韩国手术同意书(特别是标准化肝切除术的同意文件)在多个机构中的可读性和内容质量的影响。

方法

从7家韩国医疗机构收集标准化肝切除术同意书,并使用ChatGPT-4o对这些表格进行简化。此后,使用KReaD和Natmal指数评估可读性,同时根据字符数、单词数、句子数、每个句子的单词数和难词比例评估文本结构。通过7位肝切除专家的评估,对4个领域(风险、益处、替代方案和总体印象)的内容质量进行分析。使用配对双侧t检验进行统计比较,并应用线性混合效应模型来考虑机构和评估者的变异性。

结果

人工智能辅助编辑显著提高了可读性,KReaD分数从1777(标准差28.47)降至1335.6(标准差59.95)(P<.001),Natmal分数从1452.3(标准差88.67)降至1245.3(标准差96.96)(P=.007)。句子长度和难词比例显著降低,有助于提高易读性(P<.05)。然而,内容质量分析显示风险描述分数有所下降(之前:2.29,标准差0.47;之后:1.92,标准差0.32;P=.06)和总体印象分数有所下降(之前:2.21,标准差0.49;之后:1.71,标准差0.64;P=.13)。线性混合效应模型证实风险描述(β₁=-0.371;P=.01)和总体印象(β₁=-0.500;P=.03)显著降低,表明关键安全信息可能存在遗漏。尽管如此,定性分析表明评估者未发现明确的遗漏,但认为文本过于简化且专业性不足。

结论

尽管大语言模型辅助的手术同意书显著提高了可读性,但它们可能会损害内容完整性的某些方面,特别是在风险披露方面。这些发现凸显了需要一种平衡的方法,在确保医疗和法律准确性的同时保持易读性。未来的研究应包括以患者为中心的评估,以评估理解和知情决策,以及更广泛的多语言验证,以确定大语言模型在不同医疗环境中的适用性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5bdb/12200805/4ff8c1f19ebd/jmir-v27-e73222-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5bdb/12200805/93165e8302ac/jmir-v27-e73222-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5bdb/12200805/4ff8c1f19ebd/jmir-v27-e73222-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5bdb/12200805/93165e8302ac/jmir-v27-e73222-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5bdb/12200805/4ff8c1f19ebd/jmir-v27-e73222-g002.jpg

相似文献

1
Large Language Model-Assisted Surgical Consent Forms in Non-English Language: Content Analysis and Readability Evaluation.非英语语言的大语言模型辅助手术同意书:内容分析与可读性评估
J Med Internet Res. 2025 Jun 19;27:e73222. doi: 10.2196/73222.
2
Enhancing the Readability of Online Patient Education Materials Using Large Language Models: Cross-Sectional Study.使用大语言模型提高在线患者教育材料的可读性:横断面研究。
J Med Internet Res. 2025 Jun 4;27:e69955. doi: 10.2196/69955.
3
Can artificial intelligence improve the readability of patient education information in gynecology?人工智能能否提高妇科患者教育信息的可读性?
Am J Obstet Gynecol. 2025 Jun 25. doi: 10.1016/j.ajog.2025.06.047.
4
Intravenous magnesium sulphate and sotalol for prevention of atrial fibrillation after coronary artery bypass surgery: a systematic review and economic evaluation.静脉注射硫酸镁和索他洛尔预防冠状动脉搭桥术后房颤:系统评价与经济学评估
Health Technol Assess. 2008 Jun;12(28):iii-iv, ix-95. doi: 10.3310/hta12280.
5
Improving Patient Communication by Simplifying AI-Generated Dental Radiology Reports With ChatGPT: Comparative Study.通过使用ChatGPT简化人工智能生成的牙科放射学报告来改善患者沟通:比较研究
J Med Internet Res. 2025 Jun 9;27:e73337. doi: 10.2196/73337.
6
Artificial Intelligence Shows Limited Success in Improving Readability Levels of Spanish-language Orthopaedic Patient Education Materials.人工智能在提高西班牙语骨科患者教育材料的可读性方面成效有限。
Clin Orthop Relat Res. 2025 Feb 11. doi: 10.1097/CORR.0000000000003413.
7
Clinical Management of Wasp Stings Using Large Language Models: Cross-Sectional Evaluation Study.使用大语言模型对黄蜂蜇伤进行临床管理:横断面评估研究
J Med Internet Res. 2025 Jun 4;27:e67489. doi: 10.2196/67489.
8
Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.在基层医疗机构或医院门诊环境中,如果患者出现以下症状和体征,可判断其是否患有 COVID-19。
Cochrane Database Syst Rev. 2022 May 20;5(5):CD013665. doi: 10.1002/14651858.CD013665.pub3.
9
Using a Large Language Model for Breast Imaging Reporting and Data System Classification and Malignancy Prediction to Enhance Breast Ultrasound Diagnosis: Retrospective Study.使用大语言模型进行乳腺影像报告和数据系统分类及恶性肿瘤预测以增强乳腺超声诊断:回顾性研究
JMIR Med Inform. 2025 Jun 11;13:e70924. doi: 10.2196/70924.
10
Performance of ChatGPT-4o and Four Open-Source Large Language Models in Generating Diagnoses Based on China's Rare Disease Catalog: Comparative Study.ChatGPT-4o与四个开源大语言模型基于中国罕见病目录生成诊断的性能:比较研究
J Med Internet Res. 2025 Jun 18;27:e69929. doi: 10.2196/69929.

本文引用的文献

1
Assessing the Current Limitations of Large Language Models in Advancing Health Care Education.评估大语言模型在推进医疗保健教育方面的当前局限性。
JMIR Form Res. 2025 Jan 16;9:e51319. doi: 10.2196/51319.
2
ChatGPT as a Support Tool for Informed Consent and Preoperative Patient Education Prior to Penile Prosthesis Implantation.ChatGPT作为阴茎假体植入术前知情同意和患者术前教育的辅助工具。
J Clin Med. 2024 Dec 10;13(24):7482. doi: 10.3390/jcm13247482.
3
ChatGPT May Improve Access to Language-Concordant Care for Patients With Non-English Language Preferences.
ChatGPT 可能会改善为有非英语语言偏好的患者提供语言匹配护理的可及性。
JMIR Med Educ. 2024 Dec 10;10:e51435. doi: 10.2196/51435.
4
Health information for all: do large language models bridge or widen the digital divide?全民健康信息:大型语言模型是弥合还是扩大了数字鸿沟?
BMJ. 2024 Oct 11;387:e080208. doi: 10.1136/bmj-2024-080208.
5
Artificial Intelligence-Powered Surgical Consent: Patient Insights.人工智能驱动的手术同意书:患者见解
Cureus. 2024 Aug 29;16(8):e68134. doi: 10.7759/cureus.68134. eCollection 2024 Aug.
6
Assessing artificial intelligence responses to common patient questions regarding inflatable penile prostheses using a publicly available natural language processing tool (ChatGPT).评估人工智能对常见患者问题的反应,这些问题涉及可充气阴茎假体,使用一个公开可用的自然语言处理工具(ChatGPT)。
Can J Urol. 2024 Jun;31(3):11880-11885.
7
Bridging the literacy gap for surgical consents: an AI-human expert collaborative approach.弥合手术同意书的读写能力差距:一种人工智能与人类专家的协作方法。
NPJ Digit Med. 2024 Mar 8;7(1):63. doi: 10.1038/s41746-024-01039-2.
8
Large Language Model-Based Chatbot vs Surgeon-Generated Informed Consent Documentation for Common Procedures.基于大语言模型的聊天机器人与外科医生生成的常见手术知情同意书文档。
JAMA Netw Open. 2023 Oct 2;6(10):e2336997. doi: 10.1001/jamanetworkopen.2023.36997.
9
Informed Consent: Legal Obligation or Cornerstone of the Care Relationship?知情同意:法律义务还是医患关系基石?
Int J Environ Res Public Health. 2023 Jan 24;20(3):2118. doi: 10.3390/ijerph20032118.
10
Allegations of Failure to Obtain Informed Consent in Spinal Surgery Medical Malpractice Claims.脊柱手术医疗事故索赔中未获得知情同意的指控。
JAMA Surg. 2017 Jun 21;152(6):e170544. doi: 10.1001/jamasurg.2017.0544.