Suppr超能文献

公共医疗保健中机器学习解决方案的可解释性:CRISP-ML方法。

Interpretability of Machine Learning Solutions in Public Healthcare: The CRISP-ML Approach.

作者信息

Kolyshkina Inna, Simoff Simeon

机构信息

Analytikk Consulting, Sydney, NSW, Australia.

School of Computer, Data and Mathematical Sciences, Western Sydney University, Sydney, NSW, Australia.

出版信息

Front Big Data. 2021 May 26;4:660206. doi: 10.3389/fdata.2021.660206. eCollection 2021.

Abstract

Public healthcare has a history of cautious adoption for artificial intelligence (AI) systems. The rapid growth of data collection and linking capabilities combined with the increasing diversity of the data-driven AI techniques, including machine learning (ML), has brought both ubiquitous opportunities for data analytics projects and increased demands for the regulation and accountability of the outcomes of these projects. As a result, the area of interpretability and explainability of ML is gaining significant research momentum. While there has been some progress in the development of ML methods, the methodological side has shown limited progress. This limits the practicality of using ML in the health domain: the issues with explaining the outcomes of ML algorithms to medical practitioners and policy makers in public health has been a recognized obstacle to the broader adoption of data science approaches in this domain. This study builds on the earlier work which introduced CRISP-ML, a methodology that determines the interpretability level required by stakeholders for a successful real-world solution and then helps in achieving it. CRISP-ML was built on the strengths of CRISP-DM, addressing the gaps in handling interpretability. Its application in the Public Healthcare sector follows its successful deployment in a number of recent real-world projects across several industries and fields, including credit risk, insurance, utilities, and sport. This study elaborates on the CRISP-ML methodology on the determination, measurement, and achievement of the necessary level of interpretability of ML solutions in the Public Healthcare sector. It demonstrates how CRISP-ML addressed the problems with data diversity, the unstructured nature of data, and relatively low linkage between diverse data sets in the healthcare domain. The characteristics of the case study, used in the study, are typical for healthcare data, and CRISP-ML managed to deliver on these issues, ensuring the required level of interpretability of the ML solutions discussed in the project. The approach used ensured that interpretability requirements were met, taking into account public healthcare specifics, regulatory requirements, project stakeholders, project objectives, and data characteristics. The study concludes with the three main directions for the development of the presented cross-industry standard process.

摘要

公共医疗保健领域在采用人工智能(AI)系统方面一直持谨慎态度。数据收集和链接能力的迅速增长,再加上包括机器学习(ML)在内的数据驱动型AI技术的日益多样化,既为数据分析项目带来了无处不在的机遇,也增加了对这些项目成果的监管和问责要求。因此,ML的可解释性和可说明性领域正获得显著的研究动力。虽然ML方法的开发取得了一些进展,但在方法学方面进展有限。这限制了ML在医疗领域的实用性:向医疗从业者和公共卫生领域的政策制定者解释ML算法的结果所存在的问题,已成为该领域更广泛采用数据科学方法的公认障碍。本研究基于早期引入CRISP-ML的工作,CRISP-ML是一种方法,它确定利益相关者对于成功的实际解决方案所需的可解释性水平,然后帮助实现这一水平。CRISP-ML建立在CRISP-DM的优势之上,解决了处理可解释性方面的差距。它在公共医疗保健部门的应用是基于其最近在多个行业和领域的一些实际项目中的成功部署,包括信用风险、保险、公用事业和体育。本研究详细阐述了CRISP-ML方法,该方法用于确定、衡量和实现公共医疗保健部门中ML解决方案所需的可解释性水平。它展示了CRISP-ML如何解决医疗保健领域的数据多样性、数据的非结构化性质以及不同数据集之间相对较低的关联性等问题。研究中使用的案例研究的特征对于医疗保健数据来说是典型的,并且CRISP-ML成功解决了这些问题,确保了项目中所讨论的ML解决方案所需的可解释性水平。所采用的方法确保了可解释性要求得到满足,同时考虑到了公共医疗保健的具体情况、监管要求、项目利益相关者、项目目标和数据特征。该研究最后给出了所提出的跨行业标准流程的三个主要发展方向。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2881/8187858/ac477a291949/fdata-04-660206-g0001.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验