Suppr超能文献

最小对抗样本。

Minimum Adversarial Examples.

作者信息

Du Zhenyu, Liu Fangzheng, Yan Xuehu

机构信息

College of Electronic Engineering, National University of Defense Technology, Hefei 230037, China.

出版信息

Entropy (Basel). 2022 Mar 12;24(3):396. doi: 10.3390/e24030396.

Abstract

Deep neural networks in the area of information security are facing a severe threat from adversarial examples (AEs). Existing methods of AE generation use two optimization models: (1) taking the successful attack as the objective function and limiting perturbations as the constraint; (2) taking the minimum of adversarial perturbations as the target and the successful attack as the constraint. These all involve two fundamental problems of AEs: the minimum boundary of constructing the AEs and whether that boundary is reachable. The reachability means whether the AEs of successful attack models exist equal to that boundary. Previous optimization models have no complete answer to the problems. Therefore, in this paper, for the first problem, we propose the definition of the minimum AEs and give the theoretical lower bound of the amplitude of the minimum AEs. For the second problem, we prove that solving the generation of the minimum AEs is an NPC problem, and then based on its computational inaccessibility, we establish a new third optimization model. This model is general and can adapt to any constraint. To verify the model, we devise two specific methods for generating controllable AEs under the widely used distance evaluation standard of adversarial perturbations, namely Lp constraint and SSIM constraint (structural similarity). This model limits the amplitude of the AEs, reduces the solution space's search cost, and is further improved in efficiency. In theory, those AEs generated by the new model which are closer to the actual minimum adversarial boundary overcome the blindness of the adversarial amplitude setting of the existing methods and further improve the attack success rate. In addition, this model can generate accurate AEs with controllable amplitude under different constraints, which is suitable for different application scenarios. In addition, through extensive experiments, they demonstrate a better attack ability under the same constraints as other baseline attacks. For all the datasets we test in the experiment, compared with other baseline methods, the attack success rate of our method is improved by approximately 10%.

摘要

信息安全领域的深度神经网络正面临对抗样本(AE)的严重威胁。现有的AE生成方法使用两种优化模型:(1)以成功攻击为目标函数,将扰动限制作为约束条件;(2)以对抗扰动的最小值为目标,成功攻击作为约束条件。这些都涉及到AE的两个基本问题:构建AE的最小边界以及该边界是否可达。可达性是指成功攻击模型的AE是否存在等于该边界的情况。先前的优化模型对这些问题没有完整的答案。因此,在本文中,针对第一个问题,我们提出了最小AE的定义,并给出了最小AE幅度的理论下界。针对第二个问题,我们证明了解决最小AE的生成是一个NPC问题,然后基于其计算不可达性,我们建立了一个新的第三种优化模型。该模型具有通用性,可以适应任何约束条件。为了验证该模型,我们在对抗扰动的广泛使用的距离评估标准(即Lp约束和SSIM约束(结构相似性))下设计了两种生成可控AE的具体方法。该模型限制了AE的幅度,降低了解空间的搜索成本,并在效率上得到了进一步提高。从理论上讲,新模型生成的那些更接近实际最小对抗边界的AE克服了现有方法对抗幅度设置的盲目性,并进一步提高了攻击成功率。此外,该模型可以在不同约束条件下生成幅度可控的精确AE,适用于不同的应用场景。此外,通过大量实验,在与其他基线攻击相同的约束条件下,它们展示了更好的攻击能力。对于我们在实验中测试的所有数据集,与其他基线方法相比,我们方法的攻击成功率提高了约10%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/702b/8947511/425906a4edf4/entropy-24-00396-g0A1.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验