Computer Vision Institute, College of Computer Science and Software Engineering, Shenzhen University, Shenzhen, China; Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen, China; Guangdong Key Laboratory of Intelligent Information Processing, Shenzhen University, Shenzhen, China.
Computer Vision Institute, College of Computer Science and Software Engineering, Shenzhen University, Shenzhen, China; Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen, China; Guangdong Key Laboratory of Intelligent Information Processing, Shenzhen University, Shenzhen, China.
Neural Netw. 2022 Jan;145:209-220. doi: 10.1016/j.neunet.2021.10.017. Epub 2021 Oct 28.
Although significant progress has been made in synthesizing high-quality and visually realistic face images by unconditional Generative Adversarial Networks (GANs), there is still a lack of control over the generation process in order to achieve semantic face editing. In this paper, we propose a novel learning framework, called GuidedStyle, to achieve semantic face editing on pretrained StyleGAN by guiding the image generation process with a knowledge network. Furthermore, we allow an attention mechanism in StyleGAN generator to adaptively select a single layer for style manipulation. As a result, our method is able to perform disentangled and controllable edits along various attributes, including smiling, eyeglasses, gender, mustache, hair color and attractive. Both qualitative and quantitative results demonstrate the superiority of our method over other competing methods for semantic face editing. Moreover, we show that our model can be also applied to different types of real and artistic face editing, demonstrating strong generalization ability.
尽管无条件生成对抗网络 (GANs) 在合成高质量、逼真的人脸图像方面取得了重大进展,但在实现语义人脸编辑方面,仍然缺乏对生成过程的控制。在本文中,我们提出了一种新颖的学习框架,称为 GuidedStyle,通过使用知识网络引导图像生成过程,从而在预训练的 StyleGAN 上实现语义人脸编辑。此外,我们允许 StyleGAN 生成器中的注意力机制自适应地选择单个层进行样式操作。因此,我们的方法能够沿着各种属性(包括微笑、眼镜、性别、胡须、头发颜色和吸引力)进行解耦和可控的编辑。定性和定量的结果都表明,我们的方法在语义人脸编辑方面优于其他竞争方法。此外,我们还表明,我们的模型也可以应用于不同类型的真实和艺术人脸编辑,展示了强大的泛化能力。