• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

使用注意力知识蒸馏解释神经网络。

Explaining Neural Networks Using Attentive Knowledge Distillation.

机构信息

Division of Computer Science and Engineering, Jeonbuk National University, Jeollabuk-do 54896, Korea.

Research Center for Artificial Intelligence Technology, Jeonbuk National University, Jeollabuk-do 54896, Korea.

出版信息

Sensors (Basel). 2021 Feb 11;21(4):1280. doi: 10.3390/s21041280.

DOI:10.3390/s21041280
PMID:33670125
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7916876/
Abstract

Explaining the prediction of deep neural networks makes the networks more understandable and trusted, leading to their use in various mission critical tasks. Recent progress in the learning capability of networks has primarily been due to the enormous number of model parameters, so that it is usually hard to interpret their operations, as opposed to classical white-box models. For this purpose, generating saliency maps is a popular approach to identify the important input features used for the model prediction. Existing explanation methods typically only use the output of the last convolution layer of the model to generate a saliency map, lacking the information included in intermediate layers. Thus, the corresponding explanations are coarse and result in limited accuracy. Although the accuracy can be improved by iteratively developing a saliency map, this is too time-consuming and is thus impractical. To address these problems, we proposed a novel approach to explain the model prediction by developing an attentive surrogate network using the knowledge distillation. The surrogate network aims to generate a fine-grained saliency map corresponding to the model prediction using meaningful regional information presented over all network layers. Experiments demonstrated that the saliency maps are the result of spatially attentive features learned from the distillation. Thus, they are useful for fine-grained classification tasks. Moreover, the proposed method runs at the rate of 24.3 frames per second, which is much faster than the existing methods by orders of magnitude.

摘要

解释深度神经网络的预测结果可以使网络更加易于理解和信任,从而促进其在各种关键任务中的应用。最近,神经网络的学习能力取得了显著的进展,这主要得益于模型参数的大量增加,使得其操作通常难以解释,这与经典的白盒模型形成了鲜明的对比。为此,生成显著图是一种识别模型预测中使用的重要输入特征的常用方法。现有的解释方法通常仅使用模型最后一个卷积层的输出生成显著图,而缺乏中间层包含的信息。因此,相应的解释比较粗糙,导致准确性有限。虽然通过迭代开发显著图可以提高准确性,但这太耗时,不太实际。为了解决这些问题,我们提出了一种新的方法,通过使用知识蒸馏来开发注意代理网络来解释模型的预测。代理网络旨在使用来自蒸馏的有意义的区域信息,生成与模型预测相对应的细粒度显著图。实验表明,显著图是从蒸馏中学习到的空间注意特征的结果。因此,它们对细粒度分类任务非常有用。此外,该方法的运行速度为每秒 24.3 帧,比现有的方法快几个数量级。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/fbc8994c8c81/sensors-21-01280-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/7ed1edeeb0b9/sensors-21-01280-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/aa4860aaadc7/sensors-21-01280-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/34e63c027cee/sensors-21-01280-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/7eb4ea465782/sensors-21-01280-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/6a4ca58b5f64/sensors-21-01280-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/262f351dc333/sensors-21-01280-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/624ea75d9618/sensors-21-01280-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/86d48632cea6/sensors-21-01280-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/9ecfd841d2e0/sensors-21-01280-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/fbc8994c8c81/sensors-21-01280-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/7ed1edeeb0b9/sensors-21-01280-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/aa4860aaadc7/sensors-21-01280-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/34e63c027cee/sensors-21-01280-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/7eb4ea465782/sensors-21-01280-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/6a4ca58b5f64/sensors-21-01280-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/262f351dc333/sensors-21-01280-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/624ea75d9618/sensors-21-01280-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/86d48632cea6/sensors-21-01280-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/9ecfd841d2e0/sensors-21-01280-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a2e/7916876/fbc8994c8c81/sensors-21-01280-g010.jpg

相似文献

1
Explaining Neural Networks Using Attentive Knowledge Distillation.使用注意力知识蒸馏解释神经网络。
Sensors (Basel). 2021 Feb 11;21(4):1280. doi: 10.3390/s21041280.
2
TSGB: Target-Selective Gradient Backprop for Probing CNN Visual Saliency.TSGB:用于探测 CNN 视觉显著性的目标选择梯度反向传播。
IEEE Trans Image Process. 2022;31:2529-2540. doi: 10.1109/TIP.2022.3157149. Epub 2022 Mar 21.
3
Achieving efficient interpretability of reinforcement learning via policy distillation and selective input gradient regularization.通过策略蒸馏和选择性输入梯度正则化实现强化学习的高效可解释性。
Neural Netw. 2023 Apr;161:228-241. doi: 10.1016/j.neunet.2023.01.025. Epub 2023 Jan 24.
4
Deep Visual Attention Prediction.深度视觉注意力预测。
IEEE Trans Image Process. 2018 May;27(5):2368-2378. doi: 10.1109/TIP.2017.2787612. Epub 2017 Dec 27.
5
DISC: Deep Image Saliency Computing via Progressive Representation Learning.通过渐进式表示学习进行深度图像显著度计算。
IEEE Trans Neural Netw Learn Syst. 2016 Jun;27(6):1135-49. doi: 10.1109/TNNLS.2015.2506664. Epub 2016 Jan 5.
6
Saliency-driven system models for cell analysis with deep learning.基于深度学习的细胞分析的显著驱动系统模型。
Comput Methods Programs Biomed. 2019 Dec;182:105053. doi: 10.1016/j.cmpb.2019.105053. Epub 2019 Aug 26.
7
PiCANet: Pixel-wise Contextual Attention Learning for Accurate Saliency Detection.PiCANet:用于精确显著性检测的逐像素上下文注意力学习
IEEE Trans Image Process. 2020 Apr 23. doi: 10.1109/TIP.2020.2988568.
8
Predicting Human Eye Fixations via an LSTM-based Saliency Attentive Model.基于长短期记忆网络的显著性注意力模型预测人眼注视点
IEEE Trans Image Process. 2018 Jun 29. doi: 10.1109/TIP.2018.2851672.
9
Continual learning with attentive recurrent neural networks for temporal data classification.用于时态数据分类的基于注意力循环神经网络的持续学习
Neural Netw. 2023 Jan;158:171-187. doi: 10.1016/j.neunet.2022.10.031. Epub 2022 Nov 11.
10
Clinical validation of saliency maps for understanding deep neural networks in ophthalmology.用于理解眼科中深度神经网络的显著图的临床验证。
Med Image Anal. 2022 Apr;77:102364. doi: 10.1016/j.media.2022.102364. Epub 2022 Jan 22.

引用本文的文献

1
Deep learning prediction model for central lymph node metastasis in papillary thyroid microcarcinoma based on cytology.基于细胞学的甲状腺微小乳头状癌中央淋巴结转移的深度学习预测模型
Cancer Sci. 2023 Oct;114(10):4114-4124. doi: 10.1111/cas.15930. Epub 2023 Aug 13.
2
Deep Learning-Based Classification and Targeted Gene Alteration Prediction from Pleural Effusion Cell Block Whole-Slide Images.基于深度学习的胸腔积液细胞块全切片图像分类及靶向基因改变预测
Cancers (Basel). 2023 Jan 25;15(3):752. doi: 10.3390/cancers15030752.

本文引用的文献

1
CT prostate segmentation based on synthetic MRI-aided deep attention fully convolution network.基于合成 MRI 辅助深度注意全卷积网络的 CT 前列腺分割。
Med Phys. 2020 Feb;47(2):530-540. doi: 10.1002/mp.13933. Epub 2019 Dec 3.