Cabitza Federico, Campagner Andrea, Balsano Clara
Dipartimento di Informatica, Sistemistica e Comunicazione, Università degli Studi di Milano-Bicocca, Milano, Italy.
IRCCS Istituto Ortopedico Galeazzi, Milano, Italy.
Ann Transl Med. 2020 Apr;8(7):501. doi: 10.21037/atm.2020.03.63.
Interest in the application of machine learning (ML) techniques to medicine is growing fast and wide because of their ability to endow decision support systems with so-called artificial intelligence, particularly in those medical disciplines that extensively rely on digital imaging. Nonetheless, achieving a pragmatic and ecological validation of medical AI systems in real-world settings is difficult, even when these systems exhibit very high accuracy in laboratory settings. This difficulty has been called the "last mile of implementation." In this review of the concept, we claim that this metaphorical mile presents two chasms: the hiatus of human trust and the hiatus of machine experience. The former hiatus encompasses all that can hinder the concrete use of AI at the point of care, including availability and usability issues, but also the contradictory phenomena of cognitive ergonomics, such as automation bias (overreliance on technology) and prejudice against the machine (clearly the opposite). The latter hiatus, on the other hand, relates to the production and availability of a sufficient amount of reliable and accurate clinical data that is suitable to be the "experience" with which a machine can be trained. In briefly reviewing the existing literature, we focus on this latter hiatus of the last mile, as it has been largely neglected by both ML developers and doctors. In doing so, we argue that efforts to cross this chasm require data governance practices and a focus on data work, including the practices of data awareness and data hygiene. To address the challenge of bridging the chasms in the last mile of medical AI implementation, we discuss the six main socio-technical challenges that must be overcome in order to build robust bridges and deploy potentially effective AI in real-world clinical settings.
由于机器学习(ML)技术能够赋予决策支持系统所谓的人工智能,尤其是在那些广泛依赖数字成像的医学学科中,因此其在医学领域的应用正迅速广泛地发展。尽管如此,即使这些系统在实验室环境中表现出非常高的准确性,要在现实环境中对医学人工智能系统进行务实且符合实际情况的验证也很困难。这种困难被称为“实施的最后一英里”。在对这一概念的综述中,我们认为这个隐喻的一英里存在两个鸿沟:人类信任的鸿沟和机器经验的鸿沟。前一个鸿沟涵盖了所有可能阻碍人工智能在医疗点具体应用的因素,包括可用性和易用性问题,还有认知人机工程学的矛盾现象,比如自动化偏差(过度依赖技术)和对机器的偏见(显然是相反的情况)。另一方面,后一个鸿沟涉及到生产和提供足够数量的可靠且准确的临床数据,这些数据适合作为机器可以训练的“经验”。在简要回顾现有文献时,我们关注最后一英里的后一个鸿沟,因为它在很大程度上被机器学习开发者和医生忽视了。我们认为,跨越这个鸿沟需要数据治理实践,并专注于数据工作,包括数据意识和数据卫生等实践。为应对在医学人工智能实施的最后一英里中弥合这些鸿沟的挑战,我们讨论了为在现实临床环境中构建稳健桥梁并部署可能有效的人工智能必须克服的六个主要社会技术挑战。