Yuan Hao, Yu Haiyang, Gui Shurui, Ji Shuiwang
IEEE Trans Pattern Anal Mach Intell. 2023 May;45(5):5782-5799. doi: 10.1109/TPAMI.2022.3204236. Epub 2023 Apr 3.
Deep learning methods are achieving ever-increasing performance on many artificial intelligence tasks. A major limitation of deep models is that they are not amenable to interpretability. This limitation can be circumvented by developing post hoc techniques to explain predictions, giving rise to the area of explainability. Recently, explainability of deep models on images and texts has achieved significant progress. In the area of graph data, graph neural networks (GNNs) and their explainability are experiencing rapid developments. However, there is neither a unified treatment of GNN explainability methods, nor a standard benchmark and testbed for evaluations. In this survey, we provide a unified and taxonomic view of current GNN explainability methods. Our unified and taxonomic treatments of this subject shed lights on the commonalities and differences of existing methods and set the stage for further methodological developments. To facilitate evaluations, we provide a testbed for GNN explainability, including datasets, common algorithms and evaluation metrics. Furthermore, we conduct comprehensive experiments to compare and analyze the performance of many techniques. Altogether, this work provides a unified methodological treatment of GNN explainability and a standardized testbed for evaluations.
深度学习方法在许多人工智能任务上的性能正不断提高。深度模型的一个主要局限性在于它们难以解释。可以通过开发事后解释预测的技术来规避这一局限性,从而催生了可解释性这一领域。最近,深度模型在图像和文本方面的可解释性取得了显著进展。在图数据领域,图神经网络(GNN)及其可解释性正在迅速发展。然而,对于GNN可解释性方法既没有统一的论述,也没有用于评估的标准基准和测试平台。在本次综述中,我们对当前的GNN可解释性方法提供了统一的分类视角。我们对这一主题的统一分类论述揭示了现有方法的异同,为进一步的方法学发展奠定了基础。为便于评估,我们提供了一个GNN可解释性测试平台,包括数据集、常用算法和评估指标。此外,我们进行了全面的实验,以比较和分析许多技术的性能。总之,这项工作为GNN可解释性提供了统一的方法论述和标准化的评估测试平台。