Alhlffee Mahmood H B, Huang Yea-Shuan, Chen Yi-An
College of Computer Science and Electrical Engineering, Chung-Hua University, Hsinchu, Taiwan.
Department of Computer Science and Information Engineering, Chung-Hua University, Hsinchu, Taiwan.
PeerJ Comput Sci. 2022 Feb 16;8:e897. doi: 10.7717/peerj-cs.897. eCollection 2022.
One of the key challenges in facial recognition is multi-view face synthesis from a single face image. The existing generative adversarial network (GAN) deep learning methods have been proven to be effective in performing facial recognition with a set of pre-processing, post-processing and feature representation techniques to bring a frontal view into the same position in-order to achieve high accuracy face identification. However, these methods still perform relatively weak in generating high quality frontal-face image samples under extreme face pose scenarios. The novel framework architecture of the two-pathway generative adversarial network (TP-GAN), has made commendable progress in the face synthesis model, making it possible to perceive global structure and local details in an unsupervised manner. More importantly, the TP-GAN solves the problems of photorealistic frontal view synthesis by relying on texture details of the landmark detection and synthesis functions, which limits its ability to achieve the desired performance in generating high-quality frontal face image samples under extreme pose. We propose, in this paper, a landmark feature-based method (LFM) for robust pose-invariant facial recognition, which aims to improve image resolution quality of the generated frontal faces under a variety of facial poses. We therefore augment the existing TP-GAN generative global pathway with a well-constructed 2D face landmark localization to cooperate with the local pathway structure in a landmark sharing manner to incorporate empirical face pose into the learning process, and improve the encoder-decoder global pathway structure for better representation of facial image features by establishing robust feature extractors that select meaningful features that ease the operational workflow toward achieving a balanced learning strategy, thus significantly improving the photorealistic face image resolution. We verify the effectiveness of our proposed method on both Multi-PIE and FEI datasets. The quantitative and qualitative experimental results show that our proposed method not only generates high quality perceptual images under extreme poses but also significantly improves upon the TP-GAN results.
人脸识别中的关键挑战之一是从单张人脸图像进行多视角人脸合成。现有的生成对抗网络(GAN)深度学习方法已被证明在执行人脸识别时是有效的,通过一系列预处理、后处理和特征表示技术,将正面视图调整到相同位置,以实现高精度的人脸识别。然而,这些方法在极端人脸姿态场景下生成高质量正面人脸图像样本时,表现仍然相对较弱。双路径生成对抗网络(TP-GAN)的新颖框架架构在人脸合成模型方面取得了值得称赞的进展,使其能够以无监督的方式感知全局结构和局部细节。更重要的是,TP-GAN通过依赖地标检测和合成功能的纹理细节解决了逼真正面视图合成的问题,这限制了其在极端姿态下生成高质量正面人脸图像样本时达到理想性能的能力。在本文中,我们提出了一种基于地标特征的方法(LFM)用于鲁棒的姿态不变人脸识别,旨在提高在各种人脸姿态下生成的正面人脸的图像分辨率质量。因此,我们通过精心构建的二维人脸地标定位增强现有的TP-GAN生成全局路径,以地标共享的方式与局部路径结构协作,将经验性人脸姿态纳入学习过程,并改进编码器-解码器全局路径结构,通过建立鲁棒的特征提取器来更好地表示人脸图像特征,该提取器选择有意义的特征,简化操作流程以实现平衡的学习策略,从而显著提高逼真人脸图像分辨率。我们在Multi-PIE和FEI数据集上验证了我们提出方法的有效性。定量和定性实验结果表明,我们提出的方法不仅在极端姿态下生成高质量的感知图像,而且在TP-GAN的结果上有显著改进。