Wang Jiaxuan, Oh Jeeheh, Wang Haozhu, Wiens Jenna
University of Michigan.
KDD. 2018 Aug;2018:2417-2426. doi: 10.1145/3219819.3220070. Epub 2018 Jul 19.
In many settings, it is important that a model be capable of providing reasons for its predictions (., the model must be interpretable). However, the model's reasoning may not conform with well-established knowledge. In such cases, while interpretable, the model lacks . In this work, we formally define credibility in the linear setting and focus on techniques for learning models that are both accurate and credible. In particular, we propose a regularization penalty, expert yielded estimates (EYE), that incorporates expert knowledge about well-known relationships among covariates and the outcome of interest. We give both theoretical and empirical results comparing our proposed method to several other regularization techniques. Across a range of settings, experiments on both synthetic and real data show that models learned using the EYE penalty are significantly more credible than those learned using other penalties. Applied to two large-scale patient risk stratification task, our proposed technique results in a model whose top features overlap significantly with known clinical risk factors, while still achieving good predictive performance.
在许多情况下,一个模型能够为其预测提供理由(即,模型必须是可解释的)是很重要的。然而,模型的推理可能不符合已确立的知识。在这种情况下,虽然模型是可解释的,但它缺乏……在这项工作中,我们在线性设置中正式定义了可信度,并专注于学习既准确又可信的模型的技术。具体来说,我们提出了一种正则化惩罚,即专家给出的估计(EYE),它纳入了关于协变量与感兴趣的结果之间的已知关系的专家知识。我们给出了理论和实证结果,将我们提出的方法与其他几种正则化技术进行了比较。在一系列情况下,对合成数据和真实数据的实验表明,使用EYE惩罚学习到的模型比使用其他惩罚学习到的模型明显更可信。应用于两个大规模患者风险分层任务时,我们提出的技术产生了一个模型,其顶级特征与已知临床风险因素有显著重叠,同时仍实现了良好的预测性能。