• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

土耳其牙科专业考试中人工智能系统回答口腔修复学问题的比较

Comparison of artificial intelligence systems in answering prosthodontics questions from the dental specialty exam in Turkey.

作者信息

Tosun Busra, Yilmaz Zeynep Sen

机构信息

Department of Prosthodontics, Faculty of Dentistry, Bolu Abant Izzet Baysal University, Bolu, Turkey.

Department of Prosthodontics, Faculty of Dentistry, The University of Atatürk, Erzurum, Turkey.

出版信息

J Dent Sci. 2025 Jul;20(3):1454-1459. doi: 10.1016/j.jds.2025.01.025. Epub 2025 Jan 31.

DOI:10.1016/j.jds.2025.01.025
PMID:40654425
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12254736/
Abstract

UNLABELLED

: Artificial intelligence (AI) is increasingly vital in dentistry, supporting diagnostics, treatment planning, and patient education. However, AI systems face challenges, especially in delivering accurate information within specialized dental fields. This study aimed to evaluate the performance of seven AI-based chatbots (ChatGPT-3.5, ChatGPT-4, Gemini, Gemini Advanced, Claude AI, Microsoft Copilot, and Smodin AI) in correctly answering prosthodontics questions from the Dental Specialty Exam (DUS) in Turkey.

MATERIALS AND METHODS

The dataset for this study consists of 128 multiple-choice prosthodontics questions from the DUS, a national exam administered in Turkey by the Student Selection and Placement Center (ÖSYM) between 2012 and 2021. Chatbot performance was assessed by categorizing the questions into case-based and knowledge-based.

RESULTS

ChatGPT-4 achieved the highest accuracy (75.8 %), while Gemini AI had the lowest (46.1 %). Gemini AI also had more incorrect (69) than correct answers (59). ChatGPT-4 and ChatGPT-3.5 showed significantly higher accuracy in knowledge-based questions compared to case-based ones (p < 0.05). For case-based questions, Gemini and Gemini Advanced had the lowest accuracy (36.4 %), while other chatbots averaged 45.5 %. In knowledge-based questions, ChatGPT-4 performed best (78.6 %) and Gemini AI the worst (47 %).

CONCLUSION

ChatGPT-4 excelled in knowledge-based prosthodontic questions, showing potential to enhance dental education through personalized learning and clinical reasoning support. However, its limitations in case-based scenarios highlight the need for optimization to better address complex clinical situations. These findings suggest that AI models can significantly contribute to dental education and clinical practice.

摘要

未标注

人工智能(AI)在牙科领域的重要性日益凸显,可辅助诊断、治疗计划制定以及患者教育。然而,人工智能系统面临诸多挑战,尤其是在专业牙科领域提供准确信息方面。本研究旨在评估七个基于人工智能的聊天机器人(ChatGPT-3.5、ChatGPT-4、Gemini、Gemini Advanced、Claude AI、Microsoft Copilot和Smodin AI)正确回答土耳其牙科专业考试(DUS)中修复学问题的表现。

材料与方法

本研究的数据集由128道来自DUS的修复学选择题组成,该考试是土耳其学生选拔与安置中心(ÖSYM)在2012年至2021年间组织的全国性考试。通过将问题分为基于病例和基于知识两类来评估聊天机器人的表现。

结果

ChatGPT-4的准确率最高(75.8%),而Gemini AI的准确率最低(46.1%)。Gemini AI答错的题目(69道)比答对的题目(59道)还多。与基于病例的问题相比,ChatGPT-4和ChatGPT-3.5在基于知识的问题上表现出显著更高的准确率(p < 0.05)。对于基于病例的问题,Gemini和Gemini Advanced的准确率最低(36.4%),而其他聊天机器人的平均准确率为45.5%。在基于知识的问题上,ChatGPT-4表现最佳(78.6%),Gemini AI表现最差(47%)。

结论

ChatGPT-4在基于知识的修复学问题上表现出色,显示出通过个性化学习和临床推理支持来加强牙科教育的潜力。然而其在基于病例场景中的局限性凸显了优化以更好应对复杂临床情况的必要性。这些发现表明人工智能模型可为牙科教育和临床实践做出重大贡献。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a936/12254736/d2089f67ef13/gr2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a936/12254736/95dfc9243b5e/gr1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a936/12254736/d2089f67ef13/gr2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a936/12254736/95dfc9243b5e/gr1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a936/12254736/d2089f67ef13/gr2.jpg

相似文献

1
Comparison of artificial intelligence systems in answering prosthodontics questions from the dental specialty exam in Turkey.土耳其牙科专业考试中人工智能系统回答口腔修复学问题的比较
J Dent Sci. 2025 Jul;20(3):1454-1459. doi: 10.1016/j.jds.2025.01.025. Epub 2025 Jan 31.
2
Artificial Intelligence in Peripheral Artery Disease Education: A Battle Between ChatGPT and Google Gemini.外周动脉疾病教育中的人工智能:ChatGPT与谷歌Gemini的较量
Cureus. 2025 Jun 1;17(6):e85174. doi: 10.7759/cureus.85174. eCollection 2025 Jun.
3
Accuracy of ChatGPT-3.5, ChatGPT-4o, Copilot, Gemini, Claude, and Perplexity in advising on lumbosacral radicular pain against clinical practice guidelines: cross-sectional study.ChatGPT-3.5、ChatGPT-4o、Copilot、Gemini、Claude和Perplexity在依据临床实践指南对腰骶神经根性疼痛提供建议方面的准确性:横断面研究
Front Digit Health. 2025 Jun 27;7:1574287. doi: 10.3389/fdgth.2025.1574287. eCollection 2025.
4
Accuracy and Reliability of Artificial Intelligence Chatbots as Public Information Sources in Implant Dentistry.人工智能聊天机器人作为种植牙科公共信息来源的准确性和可靠性
Int J Oral Maxillofac Implants. 2025 Jun 25;0(0):1-23. doi: 10.11607/jomi.11280.
5
Performance of 3 Conversational Generative Artificial Intelligence Models for Computing Maximum Safe Doses of Local Anesthetics: Comparative Analysis.用于计算局部麻醉药最大安全剂量的3种对话式生成人工智能模型的性能:比较分析
JMIR AI. 2025 May 13;4:e66796. doi: 10.2196/66796.
6
Performance of 7 Artificial Intelligence Chatbots on Board-style Endodontic Questions.7款人工智能聊天机器人在根管治疗式问题上的表现
J Endod. 2025 Jun 26. doi: 10.1016/j.joen.2025.06.014.
7
Comparative analysis of LLMs performance in medical embryology: A cross-platform study of ChatGPT, Claude, Gemini, and Copilot.大语言模型在医学胚胎学中的性能比较分析:ChatGPT、Claude、Gemini和Copilot的跨平台研究
Anat Sci Educ. 2025 May 11. doi: 10.1002/ase.70044.
8
Evaluating the validity and consistency of artificial intelligence chatbots in responding to patients' frequently asked questions in prosthodontics.评估人工智能聊天机器人在回答患者口腔修复学常见问题时的有效性和一致性。
J Prosthet Dent. 2025 Apr 7. doi: 10.1016/j.prosdent.2025.03.009.
9
Cognitive Domain Assessment of Artificial Intelligence Chatbots: A Comparative Study Between ChatGPT and Gemini's Understanding of Anatomy Education.人工智能聊天机器人的认知领域评估:ChatGPT与Gemini对解剖学教育理解的比较研究
Med Sci Educ. 2025 Feb 15;35(3):1295-1304. doi: 10.1007/s40670-025-02303-0. eCollection 2025 Jun.
10
Performance of artificial intelligence on Turkish dental specialization exam: can ChatGPT-4.0 and gemini advanced achieve comparable results to humans?人工智能在土耳其牙科专业考试中的表现:ChatGPT-4.0和Gemini Advanced能否取得与人类相当的成绩?
BMC Med Educ. 2025 Feb 10;25(1):214. doi: 10.1186/s12909-024-06389-9.

本文引用的文献

1
Performance of ChatGPT 3.5 and 4 on U.S. dental examinations: the INBDE, ADAT, and DAT.ChatGPT 3.5和4在美国牙科考试中的表现:国际牙科执照考试(INBDE)、高级牙科能力倾向测试(ADAT)和牙科入学考试(DAT)
Imaging Sci Dent. 2024 Sep;54(3):271-275. doi: 10.5624/isd.20240037. Epub 2024 Jul 2.
2
Evaluating the performance of ChatGPT-3.5 and ChatGPT-4 on the Taiwan plastic surgery board examination.评估ChatGPT-3.5和ChatGPT-4在台湾整形外科医师资格考试中的表现。
Heliyon. 2024 Jul 18;10(14):e34851. doi: 10.1016/j.heliyon.2024.e34851. eCollection 2024 Jul 30.
3
Concerns with the Usage of ChatGPT in Academia and Medicine: A Viewpoint.
对ChatGPT在学术界和医学领域应用的担忧:一种观点。
Am J Med Open. 2023 Feb 27;9:100036. doi: 10.1016/j.ajmo.2023.100036. eCollection 2023 Jun.
4
Redefining Healthcare With Artificial Intelligence (AI): The Contributions of ChatGPT, Gemini, and Co-pilot.用人工智能(AI)重新定义医疗保健:ChatGPT、Gemini和Copilot的贡献。
Cureus. 2024 Apr 7;16(4):e57795. doi: 10.7759/cureus.57795. eCollection 2024 Apr.
5
The Quality of AI-Generated Dental Caries Multiple Choice Questions: A Comparative Analysis of ChatGPT and Google Bard Language Models.人工智能生成的龋齿多项选择题的质量:ChatGPT和谷歌巴德语言模型的比较分析
Heliyon. 2024 Mar 19;10(7):e28198. doi: 10.1016/j.heliyon.2024.e28198. eCollection 2024 Apr 15.
6
Conformity of ChatGPT recommendations with the AUA/SUFU guideline on postprostatectomy urinary incontinence.与 AUA/SUFU 后前列腺切除术后尿失禁指南的一致性。
Neurourol Urodyn. 2024 Apr;43(4):935-941. doi: 10.1002/nau.25442. Epub 2024 Mar 7.
7
Artificial intelligence in dental education: ChatGPT's performance on the periodontic in-service examination.人工智能在口腔医学教育中的应用:ChatGPT 在牙周在职考试中的表现。
J Periodontol. 2024 Jul;95(7):682-687. doi: 10.1002/JPER.23-0514. Epub 2024 Jan 10.
8
Personalized Care in Eye Health: Exploring Opportunities, Challenges, and the Road Ahead for Chatbots.眼部健康的个性化护理:探索聊天机器人的机遇、挑战与未来之路。
J Pers Med. 2023 Dec 2;13(12):1679. doi: 10.3390/jpm13121679.
9
ChatGPT Versus Consultants: Blinded Evaluation on Answering Otorhinolaryngology Case-Based Questions.ChatGPT与医学顾问的对比:对耳鼻喉科基于病例问题回答的盲法评估
JMIR Med Educ. 2023 Dec 5;9:e49183. doi: 10.2196/49183.
10
Performance of ChatGPT, Bard, Claude, and Bing on the Peruvian National Licensing Medical Examination: a cross-sectional study.ChatGPT、Bard、Claude 和 Bing 在秘鲁国家医师执照考试中的表现:一项横断面研究。
J Educ Eval Health Prof. 2023;20:30. doi: 10.3352/jeehp.2023.20.30. Epub 2023 Nov 20.