• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过概率训练实现内在对抗鲁棒性。

Toward Intrinsic Adversarial Robustness Through Probabilistic Training.

出版信息

IEEE Trans Image Process. 2023;32:3862-3872. doi: 10.1109/TIP.2023.3290532. Epub 2023 Jul 14.

DOI:10.1109/TIP.2023.3290532
PMID:37428673
Abstract

Modern deep neural networks have made numerous breakthroughs in real-world applications, yet they remain vulnerable to some imperceptible adversarial perturbations. These tailored perturbations can severely disrupt the inference of current deep learning-based methods and may induce potential security hazards to artificial intelligence applications. So far, adversarial training methods have achieved excellent robustness against various adversarial attacks by involving adversarial examples during the training stage. However, existing methods primarily rely on optimizing injective adversarial examples correspondingly generated from natural examples, ignoring potential adversaries in the adversarial domain. This optimization bias can induce the overfitting of the suboptimal decision boundary, which heavily jeopardizes adversarial robustness. To address this issue, we propose Adversarial Probabilistic Training (APT) to bridge the distribution gap between the natural and adversarial examples via modeling the latent adversarial distribution. Instead of tedious and costly adversary sampling to form the probabilistic domain, we estimate the adversarial distribution parameters in the feature level for efficiency. Moreover, we decouple the distribution alignment based on the adversarial probability model and the original adversarial example. We then devise a novel reweighting mechanism for the distribution alignment by considering the adversarial strength and the domain uncertainty. Extensive experiments demonstrate the superiority of our adversarial probabilistic training method against various types of adversarial attacks in different datasets and scenarios.

摘要

现代深度神经网络在实际应用中取得了许多突破,但它们仍然容易受到一些难以察觉的对抗性扰动的影响。这些定制的扰动可以严重干扰当前基于深度学习的方法的推断,并可能给人工智能应用带来潜在的安全隐患。到目前为止,对抗性训练方法通过在训练阶段涉及对抗性示例,在对抗性攻击方面取得了优异的鲁棒性。然而,现有的方法主要依赖于优化从自然示例相应生成的注入性对抗性示例,而忽略了对抗性领域中的潜在对手。这种优化偏差会导致次优决策边界的过度拟合,这严重危及对抗性的鲁棒性。为了解决这个问题,我们提出了对抗性概率训练 (APT),通过对潜在的对抗性分布进行建模来弥合自然和对抗性示例之间的分布差距。我们没有通过繁琐且昂贵的对抗性抽样来形成概率性域,而是在特征级别估计对抗性分布参数以提高效率。此外,我们根据对抗性概率模型和原始对抗性示例来分离分布对齐。然后,我们通过考虑对抗性强度和域不确定性,为分布对齐设计了一种新颖的重新加权机制。广泛的实验表明,我们的对抗性概率训练方法在不同数据集和场景中的各种对抗性攻击中具有优越性。

相似文献

1
Toward Intrinsic Adversarial Robustness Through Probabilistic Training.通过概率训练实现内在对抗鲁棒性。
IEEE Trans Image Process. 2023;32:3862-3872. doi: 10.1109/TIP.2023.3290532. Epub 2023 Jul 14.
2
Interpreting and Improving Adversarial Robustness of Deep Neural Networks With Neuron Sensitivity.基于神经元敏感性的深度神经网络对抗鲁棒性解释与改进。
IEEE Trans Image Process. 2021;30:1291-1304. doi: 10.1109/TIP.2020.3042083. Epub 2020 Dec 23.
3
A Hamiltonian Monte Carlo Method for Probabilistic Adversarial Attack and Learning.一种用于概率对抗攻击和学习的哈密顿蒙特卡罗方法。
IEEE Trans Pattern Anal Mach Intell. 2022 Apr;44(4):1725-1737. doi: 10.1109/TPAMI.2020.3032061. Epub 2022 Mar 4.
4
Enhancing robustness in video recognition models: Sparse adversarial attacks and beyond.增强视频识别模型的鲁棒性:稀疏对抗攻击及其他。
Neural Netw. 2024 Mar;171:127-143. doi: 10.1016/j.neunet.2023.11.056. Epub 2023 Nov 25.
5
Learning defense transformations for counterattacking adversarial examples.学习防御变换以反击对抗样本。
Neural Netw. 2023 Jul;164:177-185. doi: 10.1016/j.neunet.2023.03.008. Epub 2023 Mar 24.
6
Boosting adversarial robustness via self-paced adversarial training.通过自步对抗训练提高对抗鲁棒性。
Neural Netw. 2023 Oct;167:706-714. doi: 10.1016/j.neunet.2023.08.063. Epub 2023 Sep 9.
7
DEFEAT: Decoupled feature attack across deep neural networks.击败:跨深度神经网络的解耦特征攻击。
Neural Netw. 2022 Dec;156:13-28. doi: 10.1016/j.neunet.2022.09.009. Epub 2022 Sep 20.
8
Adv-BDPM: Adversarial attack based on Boundary Diffusion Probability Model.Adv-BDPM:基于边界扩散概率模型的对抗攻击。
Neural Netw. 2023 Oct;167:730-740. doi: 10.1016/j.neunet.2023.08.048. Epub 2023 Sep 9.
9
Generalizable and Discriminative Representations for Adversarially Robust Few-Shot Learning.用于对抗鲁棒少样本学习的可泛化和判别性表示
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):5480-5493. doi: 10.1109/TNNLS.2024.3379172. Epub 2025 Feb 28.
10
Towards Adversarial Robustness with Early Exit Ensembles.迈向具有早期退出集成的对抗鲁棒性。
Annu Int Conf IEEE Eng Med Biol Soc. 2022 Jul;2022:313-316. doi: 10.1109/EMBC48229.2022.9871347.