Li Jiawei, Li Yiming, Xiang Xingchun, Xia Shu-Tao, Dong Siyi, Cai Yun
Tsinghua Shenzhen International Graduate School, Tsinghua University, Shenzhen 518055, China.
PCL Research Center of Networks and Communications, Peng Cheng Laboratory, Shenzhen 518055, China.
Entropy (Basel). 2020 Oct 24;22(11):1203. doi: 10.3390/e22111203.
Deep Neural Networks (DNNs) usually work in an end-to-end manner. This makes the trained DNNs easy to use, but they remain an ambiguous decision process for every test case. Unfortunately, the interpretability of decisions is crucial in some scenarios, such as medical or financial data mining and decision-making. In this paper, we propose a Tree-Network-Tree (TNT) learning framework for explainable decision-making, where the knowledge is alternately transferred between the tree model and DNNs. Specifically, the proposed TNT learning framework exerts the advantages of different models at different stages: (1) a novel James-Stein Decision Tree (JSDT) is proposed to generate better knowledge representations for DNNs, especially when the input data are in low-frequency or low-quality; (2) the DNNs output high-performing prediction result from the knowledge embedding inputs and behave as a teacher model for the following tree model; and (3) a novel distillable Gradient Boosted Decision Tree (dGBDT) is proposed to learn interpretable trees from the soft labels and make a comparable prediction as DNNs do. Extensive experiments on various machine learning tasks demonstrated the effectiveness of the proposed method.
深度神经网络(DNN)通常以端到端的方式工作。这使得训练好的DNN易于使用,但对于每个测试用例来说,它们仍然是一个模糊的决策过程。不幸的是,决策的可解释性在某些场景中至关重要,例如医学或金融数据挖掘与决策。在本文中,我们提出了一种用于可解释决策的树-网络-树(TNT)学习框架,其中知识在树模型和DNN之间交替传递。具体而言,所提出的TNT学习框架在不同阶段发挥不同模型的优势:(1)提出了一种新颖的詹姆斯-斯坦决策树(JSDT),为DNN生成更好的知识表示,特别是当输入数据处于低频或低质量时;(2)DNN从知识嵌入输入中输出高性能的预测结果,并作为后续树模型的教师模型;(3)提出了一种新颖的可蒸馏梯度提升决策树(dGBDT),从软标签中学习可解释的树,并做出与DNN相当的预测。在各种机器学习任务上进行的大量实验证明了所提方法的有效性。