Garbin Christian, Marques Oge
College of Engineering & Computer Science, Florida Atlantic University, 777 Glades Rd, EE441, Boca Raton, FL 33431-0991.
Radiol Artif Intell. 2022 Jan 26;4(2):e210127. doi: 10.1148/ryai.210127. eCollection 2022 Mar.
Artificial intelligence applications for health care have come a long way. Despite the remarkable progress, there are several examples of unfulfilled promises and outright failures. There is still a struggle to translate successful research into successful real-world applications. Machine learning (ML) products diverge from traditional software products in fundamental ways. Particularly, the main component of an ML solution is not a specific piece of code that is written for a specific purpose; rather, it is a generic piece of code, a model, customized by a training process driven by hyperparameters and a dataset. Datasets are usually large, and models are opaque. Therefore, datasets and models cannot be inspected in the same, direct way as traditional software products. Other methods are needed to detect failures in ML products. This report investigates recent advancements that promote auditing, supported by transparency, as a mechanism to detect potential failures in ML products for health care applications. It reviews practices that apply to the early stages of the ML lifecycle, when datasets and models are created; these stages are unique to ML products. Concretely, this report demonstrates how two recently proposed checklists, datasheets for datasets and model cards, can be adopted to increase the transparency of crucial stages of the ML lifecycle, using ChestX-ray8 and CheXNet as examples. The adoption of checklists to document the strengths, limitations, and applications of datasets and models in a structured format leads to increased transparency, allowing early detection of potential problems and opportunities for improvement. Artificial Intelligence, Machine Learning, Lifecycle, Auditing, Transparency, Failures, Datasheets, Datasets, Model Cards © RSNA, 2022.
人工智能在医疗保健领域的应用已经取得了长足的进步。尽管取得了显著进展,但仍有一些未兑现的承诺和彻底失败的例子。将成功的研究转化为成功的实际应用仍然存在困难。机器学习(ML)产品在根本上与传统软件产品不同。特别是,ML解决方案的主要组件不是为特定目的编写的特定代码片段;相反,它是一段通用代码,即一个模型,由超参数和数据集驱动的训练过程进行定制。数据集通常很大,而且模型不透明。因此,无法像检查传统软件产品那样直接检查数据集和模型。需要其他方法来检测ML产品中的故障。本报告调查了最近的进展,这些进展以透明度为支撑,推动审计作为一种检测医疗保健应用中ML产品潜在故障的机制。它回顾了适用于ML生命周期早期阶段(即创建数据集和模型时)的实践;这些阶段是ML产品所特有的。具体而言,本报告以ChestX-ray8和CheXNet为例,展示了如何采用最近提出的两个清单——数据集数据表和模型卡——来提高ML生命周期关键阶段的透明度。采用清单以结构化格式记录数据集和模型的优势、局限性及应用,可提高透明度,从而能够早期发现潜在问题和改进机会。人工智能、机器学习、生命周期、审计、透明度、故障、数据表、数据集、模型卡 © RSNA,2022年