Yang Yang, Idris Norisma Binti, Liu Chang, Wu Hui, Yu Dingguo
Faculty of Computer Science and Information Technology, Universiti Malaya, Kuala Lumpur, Malaysia.
Institute of Intelligent Media Technology, Communication University of Zhejiang, Hangzhou, China.
PeerJ Comput Sci. 2024 Oct 4;10:e2356. doi: 10.7717/peerj-cs.2356. eCollection 2024.
The harm caused by deepfake face images is increasing. To proactively defend against this threat, this paper innovatively proposes a destructive active defense algorithm for deepfake face images (DADFI). This algorithm adds slight perturbations to the original face images to generate adversarial samples. These perturbations are imperceptible to the human eye but cause significant distortions in the outputs of mainstream deepfake models. Firstly, the algorithm generates adversarial samples that maintain high visual fidelity and authenticity. Secondly, in a black-box scenario, the adversarial samples are used to attack deepfake models to enhance their offensive capabilities. Finally, destructive attack experiments were conducted on the mainstream face datasets CASIA-FaceV5 and CelebA. The results demonstrate that the proposed DADFI algorithm not only improves the generation speed of adversarial samples but also increases the success rate of active defense. This achievement can effectively reduce the harm caused by deepfake face images.
深度伪造人脸图像造成的危害日益增加。为了积极抵御这种威胁,本文创新性地提出了一种针对深度伪造人脸图像的破坏性主动防御算法(DADFI)。该算法对原始人脸图像添加轻微扰动以生成对抗样本。这些扰动肉眼难以察觉,但会导致主流深度伪造模型的输出产生显著失真。首先,该算法生成保持高视觉保真度和真实性的对抗样本。其次,在黑盒场景中,利用对抗样本攻击深度伪造模型以增强其攻击能力。最后,在主流人脸数据集CASIA - FaceV5和CelebA上进行了破坏性攻击实验。结果表明,所提出的DADFI算法不仅提高了对抗样本的生成速度,还提高了主动防御的成功率。这一成果能够有效降低深度伪造人脸图像造成的危害。