Sharma Brihat, Gao Yanjun, Miller Timothy, Churpek Matthew M, Afshar Majid, Dligach Dmitriy
University of Wisconsin-Madison.
Boston Children's Hospital and Harvard Medical School.
Proc Conf Assoc Comput Linguist Meet. 2023 Jul;2023(ClinicalNLP):78-85.
Generative artificial intelligence (AI) is a promising direction for augmenting clinical diagnostic decision support and reducing diagnostic errors, a leading contributor to medical errors. To further the development of clinical AI systems, the Diagnostic Reasoning Benchmark (DR.BENCH) was introduced as a comprehensive generative AI framework, comprised of six tasks representing key components in clinical reasoning. We present a comparative analysis of in-domain versus out-of-domain language models as well as multi-task versus single task training with a focus on the problem summarization task in DR.BENCH (Gao et al., 2023). We demonstrate that a multi-task, clinically-trained language model outperforms its general domain counterpart by a large margin, establishing a new state-of-the-art performance, with a ROUGE-L score of 28.55. This research underscores the value of domain-specific training for optimizing clinical diagnostic reasoning tasks.
生成式人工智能(AI)是增强临床诊断决策支持和减少诊断错误的一个有前景的方向,诊断错误是医疗差错的一个主要原因。为了推动临床AI系统的发展,诊断推理基准(DR.BENCH)作为一个全面的生成式AI框架被引入,它由代表临床推理关键组成部分的六个任务组成。我们对领域内与领域外语言模型以及多任务与单任务训练进行了比较分析,重点关注DR.BENCH中的问题总结任务(Gao等人,2023年)。我们证明,经过临床训练的多任务语言模型比其通用领域的对应模型有大幅提升,建立了新的最先进性能,ROUGE-L分数为28.55。这项研究强调了针对特定领域训练对优化临床诊断推理任务的价值。