College of Information Science and Engineering/College of Artificial Intelligence, China University of Petroleum (Beijing), Beijing, 102249, China; Computational Optimisation and Learning (COL) Lab, School of Computer Science, University of Nottingham, Nottingham, United Kingdom; The Lab for Uncertainty in Data and Decision Making (LUCID), School of Computer Science, University of Nottingham, Nottingham, United Kingdom.
The Lab for Uncertainty in Data and Decision Making (LUCID), School of Computer Science, University of Nottingham, Nottingham, United Kingdom.
Neural Netw. 2024 Aug;176:106345. doi: 10.1016/j.neunet.2024.106345. Epub 2024 Apr 27.
Local Interpretability Model-agnostic Explanations (LIME) is a well-known post-hoc technique for explaining black-box models. While very useful, recent research highlights challenges around the explanations generated. In particular, there is a potential lack of stability, where the explanations provided vary over repeated runs of the algorithm, casting doubt on their reliability. This paper investigates the stability of LIME when applied to multivariate time series classification. We demonstrate that the traditional methods for generating neighbours used in LIME carry a high risk of creating 'fake' neighbours, which are out-of-distribution in respect to the trained model and far away from the input to be explained. This risk is particularly pronounced for time series data because of their substantial temporal dependencies. We discuss how these out-of-distribution neighbours contribute to unstable explanations. Furthermore, LIME weights neighbours based on user-defined hyperparameters which are problem-dependent and hard to tune. We show how unsuitable hyperparameters can impact the stability of explanations. We propose a two-fold approach to address these issues. First, a generative model is employed to approximate the distribution of the training data set, from which within-distribution samples and thus meaningful neighbours can be created for LIME. Second, an adaptive weighting method is designed in which the hyperparameters are easier to tune than those of the traditional method. Experiments on real-world data sets demonstrate the effectiveness of the proposed method in providing more stable explanations using the LIME framework. In addition, in-depth discussions are provided on the reasons behind these results.
局部可解释模型不可知解释(LIME)是一种用于解释黑盒模型的知名事后技术。虽然非常有用,但最近的研究强调了生成的解释所面临的挑战。特别是,存在缺乏稳定性的问题,即算法的重复运行提供的解释会发生变化,这让人对其可靠性产生怀疑。本文研究了 LIME 在多元时间序列分类中的稳定性。我们证明了 LIME 中用于生成邻居的传统方法存在很高的风险,会创建出“假”邻居,这些邻居与训练模型的分布不相关,并且远离要解释的输入。由于时间序列数据具有很强的时间依赖性,因此这种风险尤为明显。我们讨论了这些离群邻居如何导致不稳定的解释。此外,LIME 根据用户定义的超参数对邻居进行加权,这些超参数取决于问题且难以调整。我们展示了不合适的超参数如何影响解释的稳定性。我们提出了一种两方面的方法来解决这些问题。首先,我们使用生成模型来近似训练数据集的分布,从中可以为 LIME 创建分布内的样本,从而创建有意义的邻居。其次,我们设计了一种自适应加权方法,其中超参数比传统方法更容易调整。在真实数据集上的实验证明了我们的方法在使用 LIME 框架提供更稳定解释方面的有效性。此外,我们还对这些结果背后的原因进行了深入讨论。