McTavish Hayden, Zhong Chudi, Achermann Reto, Karimalis Ilias, Chen Jacques, Rudin Cynthia, Seltzer Margo
University of British Columbia.
University of California, San Diego.
Proc AAAI Conf Artif Intell. 2022;36(9):9604-9613. doi: 10.1609/aaai.v36i9.21194. Epub 2022 Jun 28.
Sparse decision tree optimization has been one of the most fundamental problems in AI since its inception and is a challenge at the core of interpretable machine learning. Sparse decision tree optimization is computationally hard, and despite steady effort since the 1960's, breakthroughs have been made on the problem only within the past few years, primarily on the problem of finding optimal sparse decision trees. However, current state-of-the-art algorithms often require impractical amounts of computation time and memory to find optimal or near-optimal trees for some real-world datasets, particularly those having several continuous-valued features. Given that the search spaces of these decision tree optimization problems are massive, can we practically hope to find a sparse decision tree that competes in accuracy with a black box machine learning model? We address this problem via smart guessing strategies that can be applied to any optimal branch-and-bound-based decision tree algorithm. The guesses come from knowledge gleaned from black box models. We show that by using these guesses, we can reduce the run time by multiple orders of magnitude while providing bounds on how far the resulting trees can deviate from the black box's accuracy and expressive power. Our approach enables guesses about how to bin continuous features, the size of the tree, and lower bounds on the error for the optimal decision tree. Our experiments show that in many cases we can rapidly construct sparse decision trees that match the accuracy of black box models. To summarize: , .
自人工智能诞生以来,稀疏决策树优化一直是其最基本的问题之一,也是可解释机器学习核心的一项挑战。稀疏决策树优化在计算上很困难,尽管自20世纪60年代以来一直在持续努力,但直到过去几年才在该问题上取得突破,主要是在寻找最优稀疏决策树的问题上。然而,对于一些现实世界的数据集,特别是那些具有多个连续值特征的数据集,当前最先进的算法通常需要大量不切实际的计算时间和内存来找到最优或接近最优的树。鉴于这些决策树优化问题的搜索空间巨大,我们真的能期望找到一个在准确性上能与黑箱机器学习模型相媲美的稀疏决策树吗?我们通过可应用于任何基于最优分支定界的决策树算法的智能猜测策略来解决这个问题。这些猜测来自从黑箱模型中收集的知识。我们表明,通过使用这些猜测,我们可以将运行时间减少多个数量级,同时给出所得树与黑箱的准确性和表达能力的偏差程度的界限。我们的方法能够对如何对连续特征进行分箱、树的大小以及最优决策树的误差下界进行猜测。我们的实验表明,在许多情况下,我们可以快速构建出与黑箱模型准确性相匹配的稀疏决策树。总结如下: , 。