Latario Luke D, Fowler John R
University of Pittsburgh Medical Center Passavant, PA, USA.
University of Pittsburgh, PA, USA.
Hand (N Y). 2024 Nov 16:15589447241295328. doi: 10.1177/15589447241295328.
Artificial intelligence offers opportunities to improve the burden of health care administrative tasks. Application of machine learning to coding and billing for clinic encounters may represent time- and cost-saving benefits with low risk to patient outcomes.
Gemini, a publicly available large language model chatbot, was queried with 139 de-identified patient encounters from a single surgeon and asked to provide the Current Procedural Terminology code based on the criteria for different encounter types. Percent agreement and Cohen's kappa coefficient were calculated.
Gemini demonstrated 68% agreement for all encounter types, with a kappa coefficient of 0.586 corresponding to moderate interrater reliability. Agreement was highest for postoperative encounters (n = 43) with 98% agreement and lowest for new encounters (n = 27) with 48% agreement. Gemini recommended billing levels greater than the surgeon's billing level 31 times and lower billing levels 10 times, with 4 wrong encounter type codes.
A publicly available chatbot without specific programming for health care billing demonstrated moderate interrater reliability with a hand surgeon in billing clinic encounters. Future integration of artificial intelligence tools in physician workflow may improve the accuracy and speed of billing encounters and lower administrative costs.
人工智能为减轻医疗保健管理任务的负担提供了机会。将机器学习应用于临床诊疗的编码和计费可能具有节省时间和成本的优势,且对患者预后风险较低。
使用来自一位外科医生的139份去识别化患者诊疗记录查询公开可用的大型语言模型聊天机器人Gemini,并要求其根据不同诊疗类型的标准提供当前程序术语代码。计算百分比一致性和科恩kappa系数。
Gemini对所有诊疗类型的一致性为68%,kappa系数为0.586,对应中等程度的评分者间信度。术后诊疗记录(n = 43)的一致性最高,为98%,新诊疗记录(n = 27)的一致性最低,为48%。Gemini推荐的计费级别高于外科医生计费级别的有31次,低于计费级别的有10次,出现了4次错误的诊疗类型代码。
一个没有针对医疗保健计费进行特定编程的公开可用聊天机器人在与手外科医生进行临床诊疗计费时表现出中等程度的评分者间信度。未来将人工智能工具整合到医生工作流程中可能会提高计费诊疗的准确性和速度,并降低管理成本。