Bahçeci Tuncer, Elmaağaç Burak, Ceyhan Erman
Adnan Menderes University, Faculty of Medicine Department of Urology, Aydın, Turkey.
Kayseri City Hospital, Urology Clinic, Kayseri, Turkey.
Int J Impot Res. 2025 Apr 22. doi: 10.1038/s41443-025-01056-z.
Failure to achieve spontaneous pregnancy within 12 months despite unprotected intercourse is called infertility. The rapid development of digital health data has led more people to search for healthcare-related topics on the Internet. Many infertile individuals and couples use the Internet as their primary source for information on infertility diagnosis and treatment. However, it is important to assess the readability, understandability, and actionability of the information provided by these sources for patients. There is a gap in the literature addressing this aspect. This study aims to compare the readability, understandability, and actionability of responses generated by Microsoft Copilot (MC), an AI chatbot, and Google Search (GS), an internet search engine, for infertility-related queries. Prospectively a Google Trends analysis was conducted to identify the top 20 queries related to infertility in February, 2024. Then these queries were entered into GS and MC in May 2024. Answers from both platforms were recorded for further analysis. Outputs were assessed using automated readability tools, and readability scores were calculated. Understandability and actionability of answers were evaluated using the Patient Education Materials Assessment Tool for Printable Materials (PEMAT-P) tool. GS was found to have significantly higher Automated Readability Index (ARI) and Flesch-Kincaid Grade Level (FKGL) scores than MC (p = 0.044), while no significant differences were observed in the Flesch Reading Ease, Gunning Fog Index, Simplified Measure of Gobbledygook (SMOG), and Coleman-Liau Index scores. Both GS and MC outputs had readability scores above the 8th-grade level, indicating advanced reading levels. According to PEMAT-P, MC outperformed GS in terms of understandability (68.65 ± 11.99 vs. 54.50 ± 15.09, p = 0.001) and actionability (29.85 ± 17.8 vs. 1 ± 4.47, p = 0.000). MC provides more understandable and actionable responses to infertility related queries, that it might have great potential for patient education.
尽管有未采取保护措施的性行为,但在12个月内未能自然受孕被称为不孕症。数字健康数据的快速发展导致越来越多的人在互联网上搜索与医疗保健相关的主题。许多不孕不育的个人和夫妇将互联网作为获取不孕症诊断和治疗信息的主要来源。然而,评估这些来源为患者提供的信息的可读性、可理解性和可操作性非常重要。在解决这方面的文献中存在差距。本研究旨在比较人工智能聊天机器人Microsoft Copilot(MC)和互联网搜索引擎Google Search(GS)针对不孕症相关查询生成的回答的可读性、可理解性和可操作性。前瞻性地进行了谷歌趋势分析,以确定2024年2月与不孕症相关的前20个查询。然后在2024年5月将这些查询输入GS和MC。记录两个平台的答案以供进一步分析。使用自动可读性工具评估输出,并计算可读性分数。使用可打印材料患者教育材料评估工具(PEMAT-P)工具评估答案的可理解性和可操作性。发现GS的自动可读性指数(ARI)和弗莱施-金凯德年级水平(FKGL)分数显著高于MC(p = 0.044),而在弗莱施阅读简易度、冈宁雾度指数、胡言乱语简化度量(SMOG)和科尔曼-廖指数分数方面未观察到显著差异。GS和MC的输出可读性分数均高于八年级水平,表明阅读水平较高。根据PEMAT-P,MC在可理解性(68.65±11.99对54.50±15.09,p = 0.001)和可操作性(29.85±17.8对1±4.47,p = 0.000)方面优于GS。MC对不孕症相关查询提供了更易理解和可行的回答,它可能在患者教育方面具有巨大潜力。