Department of Plastic Surgery, University of Pittsburgh Medical Center, 1350 Locust Street, G103, Pittsburgh, PA 15219, United States.
Department of Plastic Surgery, University of Rome "Tor Vergata", Rome, Italy.
J Plast Reconstr Aesthet Surg. 2023 Dec;87:390-402. doi: 10.1016/j.bjps.2023.10.091. Epub 2023 Oct 20.
Appropriate patient education and preparation prior to surgery represent a fundamental step in managing expectations, avoiding unnecessary encounters and eventually achieving optimal outcomes. Thus, the objective of this study is to evaluate ChatGPT's potential as a viable source for patient education by comparing its responses and provided references to frequently asked questions on body contouring, with Google's. A Google search was conducted on July 15th, 2023, using the search term "body contouring surgery". The first 15 questions under the "People also ask" section and answers provided by Google were recorded. The 15 questions were then asked to ChatGPT-3.5. Four plastic surgeons evaluated the answers from 1 to 5 according to the Global Quality Scale. The mean score for responses given by Google was 2.55 ± 1.29, indicating poor quality but some information present, of very limited use to patients. The mean score for responses produced by ChatGPT was 4.38 ± 0.67, suggesting that the content was of good quality, useful to patients, and encompassed the most important topics. The difference was statistically significant (p = 0.001). Deficiencies in providing references represent one of the most evident weaknesses of ChatGPT. However, ChatGPT did not appear to spread misinformation, and the content of the generated responses was deemed of good quality and useful to patients. The integration of AI technology as a source for patient education has the potential to optimize patient queries on body contouring questions.
适当的术前患者教育和准备是管理预期、避免不必要的医患接触并最终实现最佳治疗效果的基础步骤。因此,本研究旨在评估 ChatGPT 是否可作为患者教育的一种可行资源,我们将其回答和提供的身体塑形相关常见问题的参考文献与谷歌进行了比较。2023 年 7 月 15 日,我们使用“body contouring surgery”进行了谷歌搜索,记录了“People also ask”部分的前 15 个问题及谷歌提供的答案。然后,我们向 ChatGPT-3.5 提出了这 15 个问题。四位整形外科医生根据全球质量量表(Global Quality Scale)对问题回答进行了 1-5 分的评估。谷歌回答的平均得分为 2.55±1.29,表明其质量较差但存在一些信息,对患者的帮助非常有限。ChatGPT 回答的平均得分为 4.38±0.67,表明其内容质量较高,对患者有用,涵盖了最重要的主题。二者的差异具有统计学意义(p=0.001)。提供参考文献方面的不足是 ChatGPT 最明显的弱点之一。然而,ChatGPT 似乎没有传播错误信息,生成的回答内容质量高且对患者有用。将人工智能技术整合为患者教育资源具有优化患者对身体塑形问题查询的潜力。