Che Zhengping, Purushotham Sanjay, Khemani Robinder, Liu Yan
University of Southern California, Los Angeles, CA, USA.
Children's Hospital Los Angeles, Los Angeles, CA, USA.
AMIA Annu Symp Proc. 2017 Feb 10;2016:371-380. eCollection 2016.
Exponential surge in health care data, such as longitudinal data from electronic health records (EHR), sensor data from intensive care unit (ICU), etc., is providing new opportunities to discover meaningful data-driven characteristics and patterns ofdiseases. Recently, deep learning models have been employedfor many computational phenotyping and healthcare prediction tasks to achieve state-of-the-art performance. However, deep models lack interpretability which is crucial for wide adoption in medical research and clinical decision-making. In this paper, we introduce a simple yet powerful knowledge-distillation approach called interpretable mimic learning, which uses gradient boosting trees to learn interpretable models and at the same time achieves strong prediction performance as deep learning models. Experiment results on Pediatric ICU dataset for acute lung injury (ALI) show that our proposed method not only outperforms state-of-the-art approaches for morality and ventilator free days prediction tasks but can also provide interpretable models to clinicians.
医疗保健数据呈指数级增长,例如来自电子健康记录(EHR)的纵向数据、重症监护病房(ICU)的传感器数据等,这为发现有意义的数据驱动的疾病特征和模式提供了新机会。最近,深度学习模型已被用于许多计算表型分析和医疗保健预测任务,以实现一流的性能。然而,深度模型缺乏可解释性,而这对于在医学研究和临床决策中广泛应用至关重要。在本文中,我们介绍了一种简单而强大的知识蒸馏方法,称为可解释模仿学习,它使用梯度提升树来学习可解释模型,同时实现与深度学习模型一样强大的预测性能。针对小儿ICU急性肺损伤(ALI)数据集的实验结果表明,我们提出的方法不仅在死亡率和无呼吸机天数预测任务上优于现有方法,还能为临床医生提供可解释模型。