Tan Li
Adobe, San Francisco, CA 94103, USA.
Entropy (Basel). 2025 Jul 29;27(8):814. doi: 10.3390/e27080814.
We propose a novel interpretability framework that integrates instance-wise feature selection with causal reasoning to explain decisions made by black-box image classifiers. Instead of relying on feature importance or mutual information, our method identifies input regions that exert the greatest causal influence on model predictions. Causal influence is formalized using a structural causal model and quantified via a conditional mutual information term. To optimize this objective efficiently, we employ continuous subset sampling and the matrix-based Rényi's α-order entropy functional. The resulting explanations are compact, semantically meaningful, and causally grounded. Experiments across multiple vision datasets demonstrate that our method outperforms existing baselines in terms of predictive fidelity.
我们提出了一种新颖的可解释性框架,该框架将实例级特征选择与因果推理相结合,以解释黑箱图像分类器所做的决策。我们的方法不是依赖于特征重要性或互信息,而是识别对模型预测产生最大因果影响的输入区域。因果影响通过结构因果模型进行形式化,并通过条件互信息项进行量化。为了有效地优化这一目标,我们采用了连续子集采样和基于矩阵的雷尼α阶熵泛函。由此产生的解释简洁、语义有意义且基于因果关系。在多个视觉数据集上的实验表明,我们的方法在预测保真度方面优于现有的基线方法。