• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

隐私保护的黑盒分类器对抗在线对抗攻击。

Privacy Preserving Defense For Black Box Classifiers Against On-Line Adversarial Attacks.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2022 Dec;44(12):9503-9520. doi: 10.1109/TPAMI.2021.3125931. Epub 2022 Nov 7.

DOI:10.1109/TPAMI.2021.3125931
PMID:34748482
Abstract

Deep learning models have been shown to be vulnerable to adversarial attacks. Adversarial attacks are imperceptible perturbations added to an image such that the deep learning model misclassifies the image with a high confidence. Existing adversarial defenses validate their performance using only the classification accuracy. However, classification accuracy by itself is not a reliable metric to determine if the resulting image is "adversarial-free". This is a foundational problem for online image recognition applications where the ground-truth of the incoming image is not known and hence we cannot compute the accuracy of the classifier or validate if the image is "adversarial-free" or not. This paper proposes a novel privacy preserving framework for defending Black box classifiers from adversarial attacks using an ensemble of iterative adversarial image purifiers whose performance is continuously validated in a loop using Bayesian uncertainties. The proposed approach can convert a single-step black box adversarial defense into an iterative defense and proposes three novel privacy preserving Knowledge Distillation (KD) approaches that use prior meta-information from various datasets to mimic the performance of the Black box classifier. Additionally, this paper proves the existence of an optimal distribution for the purified images that can reach a theoretical lower bound, beyond which the image can no longer be purified. Experimental results on six public benchmark datasets namely: 1) Fashion-MNIST, 2) CIFAR-10, 3) GTSRB, 4) MIO-TCD, 5) Tiny-ImageNet, and 6) MS-Celeb show that the proposed approach can consistently detect adversarial examples and purify or reject them against a variety of adversarial attacks.

摘要

深度学习模型已被证明易受对抗攻击的影响。对抗攻击是指在图像中添加不可察觉的扰动,从而使深度学习模型以高置信度错误分类图像。现有的对抗防御仅使用分类准确率来验证其性能。然而,分类准确率本身并不是确定生成的图像是否“无对抗”的可靠指标。这是在线图像识别应用中的一个基本问题,因为传入图像的真实情况未知,因此我们无法计算分类器的准确率,也无法验证图像是否“无对抗”。本文提出了一种新颖的隐私保护框架,使用迭代对抗图像净化器的集合来保护黑盒分类器免受对抗攻击,其性能在循环中使用贝叶斯不确定性不断验证。所提出的方法可以将单步黑盒对抗防御转换为迭代防御,并提出了三种新颖的隐私保护知识蒸馏 (KD) 方法,这些方法使用来自各种数据集的先验元信息来模拟黑盒分类器的性能。此外,本文证明了纯化图像存在最优分布,可以达到理论下限,超过该下限,图像就无法再被纯化。在六个公共基准数据集(即:1)Fashion-MNIST、2)CIFAR-10、3)GTSRB、4)MIO-TCD、5)Tiny-ImageNet 和 6)MS-Celeb 上的实验结果表明,所提出的方法可以一致地检测对抗样本,并对各种对抗攻击进行纯化或拒绝。

相似文献

1
Privacy Preserving Defense For Black Box Classifiers Against On-Line Adversarial Attacks.隐私保护的黑盒分类器对抗在线对抗攻击。
IEEE Trans Pattern Anal Mach Intell. 2022 Dec;44(12):9503-9520. doi: 10.1109/TPAMI.2021.3125931. Epub 2022 Nov 7.
2
Adversarial example defense based on image reconstruction.基于图像重建的对抗样本防御。
PeerJ Comput Sci. 2021 Dec 24;7:e811. doi: 10.7717/peerj-cs.811. eCollection 2021.
3
Beware the Black-Box: On the Robustness of Recent Defenses to Adversarial Examples.谨防黑箱:论近期针对对抗样本防御的稳健性
Entropy (Basel). 2021 Oct 18;23(10):1359. doi: 10.3390/e23101359.
4
ABCAttack: A Gradient-Free Optimization Black-Box Attack for Fooling Deep Image Classifiers.ABC攻击:一种用于欺骗深度图像分类器的无梯度优化黑盒攻击。
Entropy (Basel). 2022 Mar 15;24(3):412. doi: 10.3390/e24030412.
5
Enhancing robustness in video recognition models: Sparse adversarial attacks and beyond.增强视频识别模型的鲁棒性:稀疏对抗攻击及其他。
Neural Netw. 2024 Mar;171:127-143. doi: 10.1016/j.neunet.2023.11.056. Epub 2023 Nov 25.
6
Adversarial Attack and Defense in Deep Ranking.深度排序中的对抗攻击与防御
IEEE Trans Pattern Anal Mach Intell. 2024 Aug;46(8):5306-5324. doi: 10.1109/TPAMI.2024.3365699. Epub 2024 Jul 2.
7
Sinkhorn Adversarial Attack and Defense.Sinkhorn对抗攻击与防御
IEEE Trans Image Process. 2022;31:4039-4049. doi: 10.1109/TIP.2022.3180207. Epub 2022 Jun 14.
8
DualFlow: Generating imperceptible adversarial examples by flow field and normalize flow-based model.双流:通过流场和基于归一化流的模型生成不可察觉的对抗样本。
Front Neurorobot. 2023 Feb 9;17:1129720. doi: 10.3389/fnbot.2023.1129720. eCollection 2023.
9
Image Super-Resolution as a Defense Against Adversarial Attacks.图像超分辨率作为对抗对抗攻击的一种防御手段。
IEEE Trans Image Process. 2019 Sep 19. doi: 10.1109/TIP.2019.2940533.
10
Between-Class Adversarial Training for Improving Adversarial Robustness of Image Classification.基于类间对抗训练提高图像分类对抗鲁棒性。
Sensors (Basel). 2023 Mar 20;23(6):3252. doi: 10.3390/s23063252.