Department of Information Engineering, University of Pisa, Largo Lucio Lazzarino, 1, Pisa, 56126, Italy; Bioengineering & Robotics Research Center E. Piaggio, University of Pisa, Largo Lucio Lazzarino, 1, Pisa, 56126, Italy.
Institute of Neuroscience, Consiglio Nazionale delle Ricerche, Via Raoul Follereau, 3, Vedano al Lambro (MB), 20854, Italy.
Comput Methods Programs Biomed. 2023 Jun;236:107550. doi: 10.1016/j.cmpb.2023.107550. Epub 2023 Apr 16.
Explainable artificial intelligence (XAI) is a technology that can enhance trust in mental state classifications by providing explanations for the reasoning behind artificial intelligence (AI) models outputs, especially for high-dimensional and highly-correlated brain signals. Feature importance and counterfactual explanations are two common approaches to generate these explanations, but both have drawbacks. While feature importance methods, such as shapley additive explanations (SHAP), can be computationally expensive and sensitive to feature correlation, counterfactual explanations only explain a single outcome instead of the entire model.
To overcome these limitations, we propose a new procedure for computing global feature importance that involves aggregating local counterfactual explanations. This approach is specifically tailored to fMRI signals and is based on the hypothesis that instances close to the decision boundary and their counterfactuals mainly differ in the features identified as most important for the downstream classification task. We refer to this proposed feature importance measure as Boundary Crossing Solo Ratio (BoCSoR), since it quantifies the frequency with which a change in each feature in isolation leads to a change in classification outcome, i.e., the crossing of the model's decision boundary.
Experimental results on synthetic data and real publicly available fMRI data from the Human Connect project show that the proposed BoCSoR measure is more robust to feature correlation and less computationally expensive than state-of-the-art methods. Additionally, it is equally effective in providing an explanation for the behavior of any AI model for brain signals. These properties are crucial for medical decision support systems, where many different features are often extracted from the same physiological measures and a gold standard is absent. Consequently, computing feature importance may become computationally expensive, and there may be a high probability of mutual correlation among features, leading to unreliable results from state-of-the-art XAI methods.
可解释人工智能(XAI)是一种技术,它可以通过提供人工智能(AI)模型输出背后的推理解释来增强对心理状态分类的信任,尤其是对于高维且高度相关的脑信号。特征重要性和反事实解释是生成这些解释的两种常见方法,但它们都有缺点。虽然特征重要性方法,如 Shapley 加性解释(SHAP),可能计算成本高且对特征相关性敏感,但反事实解释仅解释单个结果,而不是整个模型。
为了克服这些限制,我们提出了一种新的计算全局特征重要性的程序,该程序涉及聚合局部反事实解释。这种方法是专门针对 fMRI 信号设计的,其基于这样一种假设:接近决策边界的实例及其反事实主要在被识别为对下游分类任务最重要的特征方面有所不同。我们将这种新提出的特征重要性度量方法称为边界穿越独奏比(BoCSoR),因为它量化了每个特征孤立变化导致分类结果变化的频率,即模型决策边界的穿越。
在合成数据和来自 Human Connect 项目的真实可用 fMRI 数据上的实验结果表明,与最新方法相比,所提出的 BoCSoR 度量方法对特征相关性更稳健且计算成本更低。此外,它在为脑信号的任何 AI 模型的行为提供解释方面同样有效。这些特性对于医疗决策支持系统至关重要,在这些系统中,通常从相同的生理测量中提取许多不同的特征,并且不存在黄金标准。因此,计算特征重要性可能变得计算成本高昂,并且特征之间可能存在高度的相关性,从而导致最新的 XAI 方法产生不可靠的结果。