Sun Jianyuan, Yu Hongchuan, Zhang Jian J, Dong Junyu, Yu Hui, Zhong Guoqiang
Department of Computer Science and Technology, Qingdao University, Qingdao 266071, China; Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford GU2 7XH, UK.
National Centre for Computer Animation, Bournemouth University, Poole BH12 5BB, UK.
Neural Netw. 2022 Oct;154:179-189. doi: 10.1016/j.neunet.2022.07.013. Epub 2022 Jul 16.
Face image-sketch synthesis is widely applied in law enforcement and digital entertainment fields. Despite the extensive progression in face image-sketch synthesis, there are few methods focusing on generating a color face image from a sketch. The existing methods pay less attention to learning the illumination or highlight distribution on the face region. However, the illumination is the key factor that makes the generated color face image looks vivid and realistic. Moreover, existing methods tend to employ some image preprocessing technologies and facial region patching approaches to generate high-quality face images, which results in the high complexity and memory consumption in practice. In this paper, we propose a novel end-to-end generative adversarial fusion model, called GAF, which fuses two U-Net generators and a discriminator by jointly learning the content and adversarial loss functions. In particular, we propose a parametric tanh activation function to learn and control illumination highlight distribution over faces, which is integrated between the two U-Net generators by an illumination distribution layer. Additionally, we fuse the attention mechanism into the second U-Net generator of GAF to keep the identity consistency and refine the generated facial details. The qualitative and quantitative experiments on the public benchmark datasets show that the proposed GAF has better performance than existing image-sketch synthesis methods in synthesized face image quality (FSIM) and face recognition accuracy (NLDA). Meanwhile, the good generalization ability of GAF has also been verified. To further demonstrate the reliability and authenticity of face images generated using GAF, we use the generated face image to attack the well-known face recognition system. The result shows that the face images generated by GAF can maintain identity consistency and well maintain everyone's unique facial characteristics, which can be further used in the benchmark of facial spoofing. Moreover, the experiments are implemented to verify the effectiveness and rationality of the proposed parametric tanh activation function and attention mechanism in GAF.
面部图像-草图合成在执法和数字娱乐领域有着广泛应用。尽管面部图像-草图合成已取得了长足进展,但专注于从草图生成彩色面部图像的方法却很少。现有方法较少关注对面部区域光照或高光分布的学习。然而,光照是使生成的彩色面部图像看起来生动逼真的关键因素。此外,现有方法倾向于采用一些图像预处理技术和面部区域修补方法来生成高质量面部图像,这在实际应用中导致了高复杂度和高内存消耗。在本文中,我们提出了一种新颖的端到端生成对抗融合模型,称为GAF,它通过联合学习内容和对抗损失函数来融合两个U-Net生成器和一个判别器。具体而言,我们提出了一种参数化双曲正切激活函数来学习和控制面部的光照高光分布,并通过一个光照分布层将其集成在两个U-Net生成器之间。此外,我们将注意力机制融合到GAF的第二个U-Net生成器中,以保持身份一致性并细化生成的面部细节。在公共基准数据集上进行的定性和定量实验表明,所提出的GAF在合成面部图像质量(FSIM)和人脸识别准确率(NLDA)方面比现有图像-草图合成方法具有更好的性能。同时,GAF良好的泛化能力也得到了验证。为了进一步证明使用GAF生成的面部图像的可靠性和真实性,我们使用生成的面部图像来攻击著名的人脸识别系统。结果表明,GAF生成的面部图像能够保持身份一致性,并很好地保留每个人独特的面部特征,可进一步用于面部欺骗基准测试。此外,还进行了实验以验证所提出的参数化双曲正切激活函数和注意力机制在GAF中的有效性和合理性。