• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于不可察觉和可转移对抗攻击的扩散模型

Diffusion Models for Imperceptible and Transferable Adversarial Attack.

作者信息

Chen Jianqi, Chen Hao, Chen Keyan, Zhang Yilan, Zou Zhengxia, Shi Zhenwei

出版信息

IEEE Trans Pattern Anal Mach Intell. 2025 Feb;47(2):961-977. doi: 10.1109/TPAMI.2024.3480519. Epub 2025 Jan 9.

DOI:10.1109/TPAMI.2024.3480519
PMID:39405140
Abstract

Many existing adversarial attacks generate -norm perturbations on image RGB space. Despite some achievements in transferability and attack success rate, the crafted adversarial examples are easily perceived by human eyes. Towards visual imperceptibility, some recent works explore unrestricted attacks without -norm constraints, yet lacking transferability of attacking black-box models. In this work, we propose a novel imperceptible and transferable attack by leveraging both the generative and discriminative power of diffusion models. Specifically, instead of direct manipulation in pixel space, we craft perturbations in the latent space of diffusion models. Combined with well-designed content-preserving structures, we can generate human-insensitive perturbations embedded with semantic clues. For better transferability, we further "deceive" the diffusion model which can be viewed as an implicit recognition surrogate, by distracting its attention away from the target regions. To our knowledge, our proposed method, DiffAttack, is the first that introduces diffusion models into the adversarial attack field. Extensive experiments conducted across diverse model architectures (CNNs, Transformers, and MLPs), datasets (ImageNet, CUB-200, and Standford Cars), and defense mechanisms underscore the superiority of our attack over existing methods such as iterative attacks, GAN-based attacks, and ensemble attacks. Furthermore, we provide a comprehensive discussion on future research avenues in diffusion-based adversarial attacks, aiming to chart a course for this burgeoning field.

摘要

许多现有的对抗攻击在图像RGB空间上生成 -范数扰动。尽管在可迁移性和攻击成功率方面取得了一些成果,但精心制作的对抗样本很容易被人眼察觉。为了实现视觉上的不可察觉性,一些近期的工作探索了无 -范数约束的无限制攻击,但缺乏攻击黑盒模型的可迁移性。在这项工作中,我们通过利用扩散模型的生成能力和判别能力,提出了一种新颖的不可察觉且可迁移的攻击方法。具体而言,我们不是在像素空间中直接进行操作,而是在扩散模型的潜在空间中制作扰动。结合精心设计的内容保留结构,我们可以生成嵌入语义线索的对人类不敏感的扰动。为了获得更好的可迁移性,我们进一步“欺骗”扩散模型,该模型可被视为一种隐式识别替代模型,方法是将其注意力从目标区域转移开。据我们所知,我们提出的方法DiffAttack是首个将扩散模型引入对抗攻击领域的方法。在各种模型架构(卷积神经网络、Transformer和多层感知器)、数据集(ImageNet、CUB - 200和斯坦福汽车数据集)以及防御机制上进行的广泛实验强调了我们的攻击方法相对于现有方法(如迭代攻击、基于生成对抗网络的攻击和集成攻击)的优越性。此外,我们对基于扩散的对抗攻击的未来研究方向进行了全面讨论,旨在为这个新兴领域指明方向。

相似文献

1
Diffusion Models for Imperceptible and Transferable Adversarial Attack.用于不可察觉和可转移对抗攻击的扩散模型
IEEE Trans Pattern Anal Mach Intell. 2025 Feb;47(2):961-977. doi: 10.1109/TPAMI.2024.3480519. Epub 2025 Jan 9.
2
Towards Transferable Adversarial Attacks on Image and Video Transformers.面向图像和视频Transformer的可迁移对抗攻击
IEEE Trans Image Process. 2023;32:6346-6358. doi: 10.1109/TIP.2023.3331582. Epub 2023 Nov 20.
3
DualFlow: Generating imperceptible adversarial examples by flow field and normalize flow-based model.双流:通过流场和基于归一化流的模型生成不可察觉的对抗样本。
Front Neurorobot. 2023 Feb 9;17:1129720. doi: 10.3389/fnbot.2023.1129720. eCollection 2023.
4
Imperceptible Transfer Attack and Defense on 3D Point Cloud Classification.三维点云分类中的隐形迁移攻击与防御。
IEEE Trans Pattern Anal Mach Intell. 2023 Apr;45(4):4727-4746. doi: 10.1109/TPAMI.2022.3193449. Epub 2023 Mar 7.
5
Remix: Towards the transferability of adversarial examples.对抗样本的可迁移性研究
Neural Netw. 2023 Jun;163:367-378. doi: 10.1016/j.neunet.2023.04.012. Epub 2023 Apr 18.
6
Enhancing robustness in video recognition models: Sparse adversarial attacks and beyond.增强视频识别模型的鲁棒性:稀疏对抗攻击及其他。
Neural Netw. 2024 Mar;171:127-143. doi: 10.1016/j.neunet.2023.11.056. Epub 2023 Nov 25.
7
Adversarial attack vulnerability of medical image analysis systems: Unexplored factors.对抗攻击对医学影像分析系统的漏洞:未知因素。
Med Image Anal. 2021 Oct;73:102141. doi: 10.1016/j.media.2021.102141. Epub 2021 Jun 18.
8
Strengthening transferability of adversarial examples by adaptive inertia and amplitude spectrum dropout.通过自适应惯性和幅度谱丢弃增强对抗样本的可转移性。
Neural Netw. 2023 Aug;165:925-937. doi: 10.1016/j.neunet.2023.06.031. Epub 2023 Jun 30.
9
Adv-BDPM: Adversarial attack based on Boundary Diffusion Probability Model.Adv-BDPM:基于边界扩散概率模型的对抗攻击。
Neural Netw. 2023 Oct;167:730-740. doi: 10.1016/j.neunet.2023.08.048. Epub 2023 Sep 9.
10
Improving the Transferability of Adversarial Examples by Feature Augmentation.通过特征增强提高对抗样本的可迁移性
IEEE Trans Neural Netw Learn Syst. 2025 May 8;PP. doi: 10.1109/TNNLS.2025.3563855.