Suppr超能文献

评估ChatGPT 4o在牙支持固定修复体治疗计划中的准确性、可重复性和一致性:简单与复杂临床病例的对比分析

Assessing the accuracy, repeatability, and consistency of ChatGPT 4o in treatment planning for tooth-supported fixed prostheses: a comparative analysis of simple and complex clinical cases.

作者信息

İşisağ Özer, Karakaya Kevser

机构信息

Faculty of Dentistry Department of Prosthodontics, Afyonkarahisar Health Sciences University, Güvenevler Neighborhood, İnönü Boulevard No:4,, Afyonkarahisar, Turkey.

出版信息

Clin Oral Investig. 2025 Sep 2;29(9):433. doi: 10.1007/s00784-025-06521-z.

Abstract

OBJECTIVES

The aim of this study is to evaluate the effectiveness of ChatGPT 4o in planning tooth-supported fixed prostheses by examining the accuracy, repeatability, and consistency of its responses across various prosthodontic cases.

MATERIALS AND METHODS

1,140 responses were generated for 38 questions regarding simple and complex prosthetic cases, using case schemas from Fundamentals of Fixed Prosthodontics. Each question was asked 30 times at different times of day (morning, afternoon, and evening) to assess response consistency.

RESULTS

ChatGPT demonstrated high accuracy (97.08%) for simple fixed prostheses but struggled with complex cases, showing a low accuracy rate of 19.69%. Statistically significant differences were observed in the response accuracy between question categories, indicating that ChatGPT performs better with straightforward cases than with complex ones. Repeatability was high across all question types, though accuracy varied, especially in complex scenarios.

CONCLUSIONS

The findings highlight ChatGPT's potential as an assistive tool in simple prosthodontic cases yet emphasize the need for clinician expertise in more complex treatment planning.

CLINICAL RELEVANCE

While ChatGPT shows promise as a supplementary tool for dental education and practice, reliance on AI alone for intricate cases remains premature.

摘要

目的

本研究旨在通过检查ChatGPT 4o在各种修复病例中的回答准确性、可重复性和一致性,评估其在设计牙支持固定修复体方面的有效性。

材料与方法

使用《固定义齿修复学基础》中的病例模式,针对简单和复杂修复病例的38个问题生成了1140个回答。每个问题在一天中的不同时间(上午、下午和晚上)被询问30次,以评估回答的一致性。

结果

ChatGPT在简单固定修复体方面表现出较高的准确性(97.08%),但在复杂病例中表现不佳,准确率仅为19.69%。不同问题类别之间的回答准确性存在统计学上的显著差异,这表明ChatGPT在简单病例中的表现优于复杂病例。尽管准确性有所不同,尤其是在复杂场景中,但所有问题类型的可重复性都很高。

结论

研究结果凸显了ChatGPT在简单修复病例中作为辅助工具的潜力,但也强调了在更复杂的治疗计划中临床医生专业知识的必要性。

临床意义

虽然ChatGPT有望成为牙科教育和实践的辅助工具,但仅依靠人工智能处理复杂病例仍为时过早。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验