Özcan Fatma, Örücü Atar Merve, Köroğlu Özlem, Yılmaz Bilge
Department of Physical Medicine and Rehabilitation, University of Health Sciences, Gaziler Physical Medicine and Rehabilitation Training and Research Hospital, Ankara, Turkey.
J Spinal Cord Med. 2025 Sep;48(5):852-857. doi: 10.1080/10790268.2024.2361551. Epub 2024 Jun 11.
The use of artificial intelligence chatbots to obtain information about patients' diseases is increasing. This study aimed to determine the reliability and usability of ChatGPT for spinal cord injury-related questions.
Three raters simultaneously evaluated a total of 47 questions on a 7-point Likert scale for reliability and usability, based on the three most frequently searched keywords in Google Trends ('general information', 'complications' and 'treatment').
Inter-rater Cronbach α scores indicated substantial agreement for both reliability and usability scores (α between 0.558 and 0.839, and α between 0.373 and 0.772, respectively). The highest mean reliability score was for 'complications' (mean 5.38). The lowest average was for the 'general information' section (mean 4.20). The 'treatment' had the highest mean scores for the usability (mean 5.87) and the lowest mean value was recorded in the 'general information' section (mean 4.80).
The answers given by ChatGPT to questions related to spinal cord injury were reliable and useful. Nevertheless, it should be kept in mind that ChatGPT may provide incorrect or incomplete information, especially in the 'general information' section, which may mislead patients and their relatives.
使用人工智能聊天机器人获取患者疾病信息的情况日益增多。本研究旨在确定ChatGPT对于脊髓损伤相关问题的可靠性和可用性。
基于谷歌趋势中搜索频率最高的三个关键词(“一般信息”、“并发症”和“治疗”),三位评估者同时以7分李克特量表对总共47个问题的可靠性和可用性进行评估。
评估者间的Cronbach α分数表明,可靠性和可用性分数均具有高度一致性(α分别在0.558至0.839之间以及0.373至0.772之间)。“并发症”的平均可靠性分数最高(平均5.38)。“一般信息”部分的平均分最低(平均4.20)。“治疗”在可用性方面的平均分数最高(平均5.87),而“一般信息”部分的平均值最低(平均4.80)。
ChatGPT对脊髓损伤相关问题给出的答案可靠且有用。然而,应牢记ChatGPT可能会提供不正确或不完整的信息,尤其是在“一般信息”部分,这可能会误导患者及其亲属。