Suppr超能文献

对事后基于规则的全局解释进行定量评估,模型无关方法。

A Quantitative Evaluation of Global, Rule-Based Explanations of Post-Hoc, Model Agnostic Methods.

作者信息

Vilone Giulia, Longo Luca

机构信息

Artificial Intelligence Cognitive Load Research Lab, Applied Intelligence Research Centre, School of Computer Science, Technological University Dublin, Dublin, Ireland.

出版信息

Front Artif Intell. 2021 Nov 3;4:717899. doi: 10.3389/frai.2021.717899. eCollection 2021.

Abstract

Understanding the inferences of data-driven, machine-learned models can be seen as a process that discloses the relationships between their input and output. These relationships consist and can be represented as a set of inference rules. However, the models usually do not explicit these rules to their end-users who, subsequently, perceive them as black-boxes and might not trust their predictions. Therefore, scholars have proposed several methods for extracting rules from data-driven machine-learned models to explain their logic. However, limited work exists on the evaluation and comparison of these methods. This study proposes a novel comparative approach to evaluate and compare the rulesets produced by five model-agnostic, post-hoc rule extractors by employing eight quantitative metrics. Eventually, the Friedman test was employed to check whether a method consistently performed better than the others, in terms of the selected metrics, and could be considered superior. Findings demonstrate that these metrics do not provide sufficient evidence to identify superior methods over the others. However, when used together, these metrics form a tool, applicable to every rule-extraction method and machine-learned models, that is, suitable to highlight the strengths and weaknesses of the rule-extractors in various applications in an objective and straightforward manner, without any human interventions. Thus, they are capable of successfully modelling distinctively aspects of explainability, providing to researchers and practitioners vital insights on what a model has learned during its training process and how it makes its predictions.

摘要

理解数据驱动的机器学习模型的推理可被视为一个揭示其输入与输出之间关系的过程。这些关系由一组推理规则组成并可表示出来。然而,这些模型通常不会向终端用户明确这些规则,因此终端用户将它们视为黑箱,可能不信任其预测结果。所以,学者们提出了几种从数据驱动的机器学习模型中提取规则以解释其逻辑的方法。然而,关于这些方法的评估和比较的研究工作有限。本研究提出了一种新颖的比较方法,通过使用八个定量指标来评估和比较由五个与模型无关的事后规则提取器生成的规则集。最终,采用弗里德曼检验来检查一种方法在所选指标方面是否始终比其他方法表现更好,是否可被视为更优。研究结果表明,这些指标没有提供足够的证据来确定哪种方法比其他方法更优。然而,当一起使用时,这些指标形成了一种工具,适用于每种规则提取方法和机器学习模型,也就是说,适合以客观直接的方式突出规则提取器在各种应用中的优缺点,无需任何人为干预。因此,它们能够成功地对可解释性的不同方面进行建模,为研究人员和从业者提供关于模型在训练过程中学到了什么以及如何进行预测的重要见解。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9a49/8596373/71aff13fbe54/frai-04-717899-g001.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验