• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过注意力机制和对抗性逻辑对配对提高对抗鲁棒性

Improving Adversarial Robustness via Attention and Adversarial Logit Pairing.

作者信息

Li Xingjian, Goodman Dou, Liu Ji, Wei Tao, Dou Dejing

机构信息

Big Data Lab, Baidu Research, Beijing, China.

X-Lab, Baidu Inc., Beijing, China.

出版信息

Front Artif Intell. 2022 Jan 27;4:752831. doi: 10.3389/frai.2021.752831. eCollection 2021.

DOI:10.3389/frai.2021.752831
PMID:35156010
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8829878/
Abstract

Though deep neural networks have achieved the state of the art performance in visual classification, recent studies have shown that they are all vulnerable to the attack of adversarial examples. In this paper, we develop improved techniques for defending against adversarial examples. First, we propose an enhanced defense technique denoted , which encourages both attention map and logit for the pairs of examples to be similar. When being applied to clean examples and their adversarial counterparts, improves accuracy on adversarial examples over adversarial training. We show that can effectively increase the average activations of adversarial examples in the key area and demonstrate that it focuses on discriminate features to improve the robustness of the model. Finally, we conduct extensive experiments using a wide range of datasets and the experiment results show that our achieves defense performance. For example, on , under strong 200-iteration Projected Gradient Descent (PGD) gray-box and black-box attacks where prior art has 34 and 39% accuracy, our method achieves and . Compared with previous work, our work is evaluated under highly challenging PGD attack: the maximum perturbation ∈ {0.25, 0.5} i.e. ∈ {0.25, 0.5} with 10-200 attack iterations. To the best of our knowledge, such a strong attack has not been previously explored on a wide range of datasets.

摘要

尽管深度神经网络在视觉分类方面已经取得了最优性能,但最近的研究表明,它们都容易受到对抗样本的攻击。在本文中,我们开发了改进技术来抵御对抗样本。首先,我们提出了一种增强防御技术,记为 ,它鼓励示例对的注意力图和逻辑值相似。当应用于干净示例及其对抗对应物时, 相对于对抗训练提高了对抗样本的准确率。我们表明, 可以有效地增加关键区域对抗样本的平均激活,并证明它专注于区分特征以提高模型的鲁棒性。最后,我们使用广泛的数据集进行了大量实验,实验结果表明我们的 实现了防御性能。例如,在 上,在强大的200次迭代投影梯度下降(PGD)灰盒和黑盒攻击下,现有技术的准确率分别为34%和39%,我们的方法实现了 和 。与之前的工作相比,我们的工作是在极具挑战性的PGD攻击下进行评估的:最大扰动 ∈ {0.25, 0.5},即 ∈ {0.25, 0.5},攻击迭代次数为10 - 200次。据我们所知,之前尚未在广泛的数据集上探索过如此强大的攻击。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/67a1651c6c1f/frai-04-752831-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/1ab1ffa1e5f2/frai-04-752831-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/fb0caf27dacc/frai-04-752831-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/43545511c665/frai-04-752831-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/68387da81254/frai-04-752831-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/97f52afc0219/frai-04-752831-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/67a1651c6c1f/frai-04-752831-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/1ab1ffa1e5f2/frai-04-752831-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/fb0caf27dacc/frai-04-752831-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/43545511c665/frai-04-752831-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/68387da81254/frai-04-752831-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/97f52afc0219/frai-04-752831-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b50a/8829878/67a1651c6c1f/frai-04-752831-g006.jpg

相似文献

1
Improving Adversarial Robustness via Attention and Adversarial Logit Pairing.通过注意力机制和对抗性逻辑对配对提高对抗鲁棒性
Front Artif Intell. 2022 Jan 27;4:752831. doi: 10.3389/frai.2021.752831. eCollection 2021.
2
Uni-image: Universal image construction for robust neural model.Uni-image:用于稳健神经模型的通用图像构建。
Neural Netw. 2020 Aug;128:279-287. doi: 10.1016/j.neunet.2020.05.018. Epub 2020 May 21.
3
Enhancing robustness in video recognition models: Sparse adversarial attacks and beyond.增强视频识别模型的鲁棒性:稀疏对抗攻击及其他。
Neural Netw. 2024 Mar;171:127-143. doi: 10.1016/j.neunet.2023.11.056. Epub 2023 Nov 25.
4
Towards evaluating the robustness of deep diagnostic models by adversarial attack.通过对抗攻击评估深度诊断模型的稳健性。
Med Image Anal. 2021 Apr;69:101977. doi: 10.1016/j.media.2021.101977. Epub 2021 Jan 22.
5
Between-Class Adversarial Training for Improving Adversarial Robustness of Image Classification.基于类间对抗训练提高图像分类对抗鲁棒性。
Sensors (Basel). 2023 Mar 20;23(6):3252. doi: 10.3390/s23063252.
6
Enhancing adversarial defense for medical image analysis systems with pruning and attention mechanism.利用剪枝和注意力机制增强医学图像分析系统的对抗防御能力。
Med Phys. 2021 Oct;48(10):6198-6212. doi: 10.1002/mp.15208. Epub 2021 Sep 14.
7
Adversarial Robustness of Deep Reinforcement Learning Based Dynamic Recommender Systems.基于深度强化学习的动态推荐系统的对抗鲁棒性
Front Big Data. 2022 May 3;5:822783. doi: 10.3389/fdata.2022.822783. eCollection 2022.
8
Interpreting and Improving Adversarial Robustness of Deep Neural Networks With Neuron Sensitivity.基于神经元敏感性的深度神经网络对抗鲁棒性解释与改进。
IEEE Trans Image Process. 2021;30:1291-1304. doi: 10.1109/TIP.2020.3042083. Epub 2020 Dec 23.
9
Towards improving fast adversarial training in multi-exit network.针对多出口网络中快速对抗训练的改进。
Neural Netw. 2022 Jun;150:1-11. doi: 10.1016/j.neunet.2022.02.015. Epub 2022 Feb 25.
10
Improving adversarial robustness of medical imaging systems via adding global attention noise.通过添加全局注意力噪声来提高医学成像系统的对抗鲁棒性。
Comput Biol Med. 2023 Sep;164:107251. doi: 10.1016/j.compbiomed.2023.107251. Epub 2023 Jul 11.