Pandey V K, Munshi A, Mohanti B K, Bansal K, Rastogi K
Radiation Oncology, Manipal Hospital Dwarka, Delhi, India.
Radiation Oncology, Manipal Hospital Dwarka, Delhi, India.
Cancer Radiother. 2024 Jun;28(3):258-264. doi: 10.1016/j.canrad.2023.11.005. Epub 2024 Jun 12.
Commercial vendors have created artificial intelligence (AI) tools for use in all aspects of life and medicine, including radiation oncology. AI innovations will likely disrupt workflows in the field of radiation oncology. However, limited data exist on using AI-based chatbots about the quality of radiation oncology information. This study aims to assess the accuracy of ChatGPT, an AI-based chatbot, in answering patients' questions during their first visit to the radiation oncology outpatient department and test knowledge of ChatGPT in radiation oncology.
Expert opinion was formulated using a set of ten standard questions of patients encountered in outpatient department practice. A blinded expert opinion was taken for the ten questions on common queries of patients in outpatient department visits, and the same questions were evaluated on ChatGPT version 3.5 (ChatGPT 3.5). The answers by expert and ChatGPT were independently evaluated for accuracy by three scientific reviewers. Additionally, a comparison was made for the extent of similarity of answers between ChatGPT and experts by a response scoring for each answer. Word count and Flesch-Kincaid readability score and grade were done for the responses obtained from expert and ChatGPT. A comparison of the answers of ChatGPT and expert was done with a Likert scale. As a second component of the study, we tested the technical knowledge of ChatGPT. Ten multiple choice questions were framed with increasing order of difficulty - basic, intermediate and advanced, and the responses were evaluated on ChatGPT. Statistical testing was done using SPSS version 27.
After expert review, the accuracy of expert opinion was 100%, and ChatGPT's was 80% (8/10) for regular questions encountered in outpatient department visits. A noticeable difference was observed in word count and readability of answers from expert opinion or ChatGPT. Of the ten multiple-choice questions for assessment of radiation oncology database, ChatGPT had an accuracy rate of 90% (9 out of 10). One answer to a basic-level question was incorrect, whereas all answers to intermediate and difficult-level questions were correct.
ChatGPT provides reasonably accurate information about routine questions encountered in the first outpatient department visit of the patient and also demonstrated a sound knowledge of the subject. The result of our study can inform the future development of educational tools in radiation oncology and may have implications in other medical fields. This is the first study that provides essential insight into the potentially positive capabilities of two components of ChatGPT: firstly, ChatGPT's response to common queries of patients at OPD visits, and secondly, the assessment of the radiation oncology knowledge base of ChatGPT.
商业供应商已开发出人工智能(AI)工具,用于生活和医学的各个方面,包括放射肿瘤学。AI创新可能会扰乱放射肿瘤学领域的工作流程。然而,关于使用基于AI的聊天机器人获取放射肿瘤学信息质量的数据有限。本研究旨在评估基于AI的聊天机器人ChatGPT在患者首次就诊于放射肿瘤学门诊时回答患者问题的准确性,并测试ChatGPT在放射肿瘤学方面的知识。
使用门诊实践中遇到的一组十个患者标准问题制定专家意见。对门诊就诊患者常见问题的十个问题采用盲法专家意见,并在ChatGPT 3.5版本上评估相同的问题。由三位科学评审员独立评估专家和ChatGPT的答案的准确性。此外,通过对每个答案进行回应评分,比较ChatGPT和专家答案的相似程度。对从专家和ChatGPT获得的回答进行单词计数、弗莱什-金凯德可读性得分和年级水平评估。使用李克特量表对ChatGPT和专家的答案进行比较。作为研究的第二个部分,我们测试了ChatGPT的专业知识。设计了十个难度递增的多项选择题——基础、中级和高级,并在ChatGPT上评估回答。使用SPSS 27版本进行统计测试。
经过专家评审,门诊就诊中常见问题的专家意见准确性为100%,ChatGPT的准确性为80%(8/10)。在专家意见或ChatGPT答案的单词计数和可读性方面观察到明显差异。在用于评估放射肿瘤学数据库的十个多项选择题中,ChatGPT的准确率为90%(10题中的9题)。一个基础水平问题的答案不正确,而中级和困难水平问题的所有答案都是正确的。
ChatGPT为患者首次门诊就诊时遇到的常规问题提供了合理准确的信息,并且也展示了对该主题的扎实知识。我们的研究结果可为放射肿瘤学教育工具的未来发展提供参考,并且可能对其他医学领域有影响。这是第一项对ChatGPT的两个组成部分的潜在积极能力提供重要见解的研究:第一,ChatGPT对门诊患者常见问题的回答;第二,对ChatGPT放射肿瘤学知识库的评估。