Sun Guangling, Hu Haoqi, Su Yuying, Liu Qi, Lu Xiaofeng
Shanghai University, School of Communication and Information Engineering, 99 Shangda Road, Baoshan District, Shanghai, 200444 China.
Multimed Tools Appl. 2023;82(5):7443-7461. doi: 10.1007/s11042-022-13641-1. Epub 2022 Aug 23.
Albeit Deep neural networks (DNNs) are widely used in computer vision, natural language processing and speech recognition, they have been discovered to be fragile to adversarial attacks. Specifically, in computer vision, an attacker can easily deceive DNNs by contaminating an input image with perturbations imperceptible to humans. As one of the important vision tasks, face verification is also subject to adversarial attack. Thus, in this paper, we focus on defending against the adversarial attack for face verification to mitigate the potential risk. We learn a network via an implementation of stacked residual blocks, namely adversarial perturbations alleviation network (ApaNet), to alleviate latent adversarial perturbations hidden in the input facial image. During the supervised learning of ApaNet, only the Labeled Faces in the Wild (LFW) is used as the training set, and the legitimate examples and corresponding adversarial examples produced by projected gradient descent algorithm compose supervision and inputs respectively. By leveraging the middle and high layer's activation of FaceNet, the discrepancy between an image output by ApaNet and the supervision is calculated as the loss function to optimize ApaNet. Empirical experiment results on the LFW, YouTube Faces DB and CASIA-FaceV5 confirm the effectiveness of the proposed defender against some representative white-box and black-box adversarial attacks. Also, experimental results show the superiority performance of the ApaNet as comparing with several currently available techniques.
尽管深度神经网络(DNN)在计算机视觉、自然语言处理和语音识别中得到了广泛应用,但人们发现它们对对抗性攻击很脆弱。具体而言,在计算机视觉中,攻击者可以通过用人类难以察觉的扰动污染输入图像来轻易欺骗DNN。作为重要的视觉任务之一,面部验证也容易受到对抗性攻击。因此,在本文中,我们专注于防御面部验证的对抗性攻击,以减轻潜在风险。我们通过堆叠残差块的实现方式学习一个网络,即对抗性扰动缓解网络(ApaNet),以减轻隐藏在输入面部图像中的潜在对抗性扰动。在ApaNet的监督学习过程中,仅将野生标注人脸(LFW)用作训练集,由投影梯度下降算法生成的合法示例和相应的对抗性示例分别构成监督和输入。通过利用FaceNet中层和高层的激活,将ApaNet输出的图像与监督之间的差异计算为损失函数,以优化ApaNet。在LFW、YouTube Faces DB和CASIA-FaceV5上的实证实验结果证实了所提出的防御方法对一些代表性白盒和黑盒对抗性攻击的有效性。此外,实验结果表明,与目前几种可用技术相比,ApaNet具有优越的性能。