Suppr超能文献

基于机器学习的高血压预测模型的可解释性研究。

On the interpretability of machine learning-based model for predicting hypertension.

机构信息

Data Systems Group, Institute of Computer Science, University of Tartu, 2 J. Liivi St., 50409, Tartu, Estonia.

Houston Methodist Center, Tartu, Estonia.

出版信息

BMC Med Inform Decis Mak. 2019 Jul 29;19(1):146. doi: 10.1186/s12911-019-0874-0.

Abstract

BACKGROUND

Although complex machine learning models are commonly outperforming the traditional simple interpretable models, clinicians find it hard to understand and trust these complex models due to the lack of intuition and explanation of their predictions. The aim of this study to demonstrate the utility of various model-agnostic explanation techniques of machine learning models with a case study for analyzing the outcomes of the machine learning random forest model for predicting the individuals at risk of developing hypertension based on cardiorespiratory fitness data.

METHODS

The dataset used in this study contains information of 23,095 patients who underwent clinician-referred exercise treadmill stress testing at Henry Ford Health Systems between 1991 and 2009 and had a complete 10-year follow-up. Five global interpretability techniques (Feature Importance, Partial Dependence Plot, Individual Conditional Expectation, Feature Interaction, Global Surrogate Models) and two local interpretability techniques (Local Surrogate Models, Shapley Value) have been applied to present the role of the interpretability techniques on assisting the clinical staff to get better understanding and more trust of the outcomes of the machine learning-based predictions.

RESULTS

Several experiments have been conducted and reported. The results show that different interpretability techniques can shed light on different insights on the model behavior where global interpretations can enable clinicians to understand the entire conditional distribution modeled by the trained response function. In contrast, local interpretations promote the understanding of small parts of the conditional distribution for specific instances.

CONCLUSIONS

Various interpretability techniques can vary in their explanations for the behavior of the machine learning model. The global interpretability techniques have the advantage that it can generalize over the entire population while local interpretability techniques focus on giving explanations at the level of instances. Both methods can be equally valid depending on the application need. Both methods are effective methods for assisting clinicians on the medical decision process, however, the clinicians will always remain to hold the final say on accepting or rejecting the outcome of the machine learning models and their explanations based on their domain expertise.

摘要

背景

虽然复杂的机器学习模型通常优于传统的简单可解释模型,但由于缺乏对预测结果的直观理解和信任,临床医生发现很难理解和信任这些复杂模型。本研究旨在通过对基于心肺功能数据预测高血压发病风险的机器学习随机森林模型结果进行分析的案例研究,展示各种与模型无关的机器学习模型解释技术的实用性。

方法

本研究使用的数据集包含了 1991 年至 2009 年期间在亨利福特健康系统接受临床医生推荐的运动平板压力测试的 23095 名患者的信息,并且这些患者都有完整的 10 年随访。本研究应用了 5 种全局可解释性技术(特征重要性、部分依赖图、个体条件期望、特征交互、全局替代模型)和 2 种局部可解释性技术(局部替代模型、Shapley 值),以展示这些解释技术如何帮助临床医生更好地理解和信任基于机器学习的预测结果。

结果

进行并报告了多项实验。结果表明,不同的解释技术可以揭示模型行为的不同见解,全局解释可以使临床医生了解训练后的响应函数所建模的整个条件分布,而局部解释则可以促进对特定实例的条件分布的小部分的理解。

结论

各种解释技术可以对机器学习模型的行为有不同的解释。全局可解释性技术的优势在于它可以对整个总体进行概括,而局部可解释性技术则侧重于在实例层面上进行解释。这两种方法都可以根据应用需求同样有效。这两种方法都是辅助临床医生进行医疗决策的有效方法,但是,临床医生始终可以根据自己的专业知识对机器学习模型及其解释的结果做出接受或拒绝的最终决定。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/884c/6664803/74606f8c1771/12911_2019_874_Fig1_HTML.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验