Kim Chanwoo, Gadgil Soham U, DeGrave Alex J, Cai Zhuo Ran, Daneshjou Roxana, Lee Su-In
Paul G. Allen School of Computer Science and Engineering, University of Washington.
Medical Scientist Training Program, University of Washington.
medRxiv. 2023 Jun 12:2023.06.07.23291119. doi: 10.1101/2023.06.07.23291119.
Building trustworthy and transparent image-based medical AI systems requires the ability to interrogate data and models at all stages of the development pipeline: from training models to post-deployment monitoring. Ideally, the data and associated AI systems could be described using terms already familiar to physicians, but this requires medical datasets densely annotated with semantically meaningful concepts. Here, we present a foundation model approach, named MONET (edical ccept rriever), which learns how to connect medical images with text and generates dense concept annotations to enable tasks in AI transparency from model auditing to model interpretation. Dermatology provides a demanding use case for the versatility of MONET, due to the heterogeneity in diseases, skin tones, and imaging modalities. We trained MONET on the basis of 105,550 dermatological images paired with natural language descriptions from a large collection of medical literature. MONET can accurately annotate concepts across dermatology images as verified by board-certified dermatologists, outperforming supervised models built on previously concept-annotated dermatology datasets. We demonstrate how MONET enables AI transparency across the entire AI development pipeline from dataset auditing to model auditing to building inherently interpretable models.
构建值得信赖且透明的基于图像的医学人工智能系统,需要在开发流程的各个阶段具备审查数据和模型的能力:从训练模型到部署后监测。理想情况下,数据和相关人工智能系统可以使用医生已经熟悉的术语来描述,但这需要用具有语义意义的概念进行密集注释的医学数据集。在此,我们提出一种基础模型方法,名为MONET(医学概念检索器),它学习如何将医学图像与文本联系起来,并生成密集的概念注释,以实现从模型审核到模型解释等人工智能透明度方面的任务。由于疾病、肤色和成像方式的异质性,皮肤病学为MONET的通用性提供了一个具有挑战性的用例。我们基于105,550张皮肤病学图像以及来自大量医学文献的自然语言描述对MONET进行了训练。经皮肤科专科医生验证,MONET可以准确注释皮肤病学图像中的概念,优于基于先前概念注释的皮肤病学数据集构建的监督模型。我们展示了MONET如何在从数据集审核到模型审核再到构建内在可解释模型的整个人工智能开发流程中实现人工智能透明度。