Lage Isaac, Chen Emily, He Jeffrey, Narayanan Menaka, Kim Been, Gershman Samuel J, Doshi-Velez Finale
Harvard University.
Google.
Proc AAAI Conf Hum Comput Crowdsourc. 2019;7(1):59-67. Epub 2019 Oct 28.
Recent years have seen a boom in interest in interpretable machine learning systems built on models that can be understood, at least to some degree, by domain experts. However, exactly what kinds of models are truly human-interpretable remains poorly understood. This work advances our understanding of precisely which factors make models interpretable in the context of decision sets, a specific class of logic-based model. We conduct carefully controlled human-subject experiments in two domains across three tasks based on human-simulatability through which we identify specific types of complexity that affect performance more heavily than others-trends that are consistent across tasks and domains. These results can inform the choice of regularizers during optimization to learn more interpretable models, and their consistency suggests that there may exist common design principles for interpretable machine learning systems.
近年来,人们对基于模型构建的可解释机器学习系统兴趣大增,这些模型至少在一定程度上能够被领域专家理解。然而,究竟哪些类型的模型才是真正可被人类解释的,目前仍知之甚少。这项工作推进了我们对在决策集(一种特定类型的基于逻辑的模型)背景下使模型可解释的精确因素的理解。我们基于人类可模拟性在两个领域的三项任务中进行了精心控制的人体实验,通过这些实验我们识别出了比其他因素对性能影响更大的特定类型的复杂性——这些趋势在不同任务和领域中是一致的。这些结果可以为优化过程中选择正则化器以学习更可解释的模型提供参考,并且它们的一致性表明,可解释机器学习系统可能存在共同的设计原则。
Proc AAAI Conf Hum Comput Crowdsourc. 2019
Front Artif Intell. 2023-4-24
Proc Natl Acad Sci U S A. 2019-10-16
BMC Bioinformatics. 2021-5-28
IEEE Trans Neural Netw Learn Syst. 2020-2-13
Sensors (Basel). 2022-3-15
Adv Neural Inf Process Syst. 2018-12
Artif Intell Med. 2013-10-18
Entropy (Basel). 2020-10-24
Front Artif Intell. 2023-2-23
Diagnostics (Basel). 2022-11-15
Philos Technol. 2021
Front Big Data. 2021-5-26
Sensors (Basel). 2021-4-3
Nature. 2000-10-5
Cogn Psychol. 1980-1