School of Information Science and Engineering, Shandong Normal University, Jinan, China.
Neural Netw. 2024 Apr;172:106095. doi: 10.1016/j.neunet.2024.106095. Epub 2024 Jan 4.
Deep neural networks have demonstrated superior performance in artificial intelligence applications, but the opaqueness of their inner working mechanism is one major drawback in their application. The prevailing unit-based interpretation is a statistical observation of stimulus-response data, which fails to show a detailed internal process of inherent mechanisms of neural networks. In this work, we analyze a convolutional neural network (CNN) trained in the classification task and present an algorithm to extract the diffusion pathways of individual pixels to identify the locations of pixels in an input image associated with object classes. The pathways allow us to test the causal components which are important for classification and the pathway-based representations are clearly distinguishable between categories. We find that the few largest pathways of an individual pixel from an image tend to cross the feature maps in each layer that is important for classification. And the large pathways of images of the same category are more consistent in their trends than those of different categories. We also apply the pathways to understanding adversarial attacks, object completion, and movement perception. Further, the total number of pathways on feature maps in all layers can clearly discriminate the original, deformed, and target samples.
深度神经网络在人工智能应用中表现出了卓越的性能,但它们内部工作机制的不透明性是其应用的一个主要缺点。目前基于单元的解释是对刺激-反应数据的统计观察,无法展示神经网络固有机制的详细内部过程。在这项工作中,我们分析了在分类任务中训练的卷积神经网络(CNN),并提出了一种算法来提取单个像素的扩散路径,以识别与目标类相关的输入图像中像素的位置。这些路径使我们能够测试对分类很重要的因果成分,并且基于路径的表示在类别之间是明显可区分的。我们发现,来自图像的单个像素的少数最大路径往往会穿过对分类很重要的各层的特征图。并且同一类别图像的大路径在趋势上比不同类别图像的更一致。我们还将这些路径应用于理解对抗攻击、物体补全和运动感知。此外,所有层的特征图上的路径总数可以清楚地区分原始、变形和目标样本。