IEEE Trans Image Process. 2023;32:5652-5663. doi: 10.1109/TIP.2023.3322593. Epub 2023 Oct 17.
Face recognition has achieved remarkable success owing to the development of deep learning. However, most of existing face recognition models perform poorly against pose variations. We argue that, it is primarily caused by pose-based long-tailed data - imbalanced distribution of training samples between profile faces and near-frontal faces. Additionally, self-occlusion and nonlinear warping of facial textures caused by large pose variations also increase the difficulty in learning discriminative features of profile faces. In this study, we propose a novel framework called Symmetrical Siamese Network (SSN), which can simultaneously overcome the limitation of pose-based long-tailed data and pose-invariant features learning. Specifically, two sub-modules are proposed in the SSN, i.e., Feature-Consistence Learning sub-Net (FCLN) and Identity-Consistence Learning sub-Net (ICLN). For FCLN, the inputs are all face images on training dataset. Inspired by the contrastive learning, we simulate pose variations of faces and constrain the model to focus on the consistent areas between the original face image and its corresponding virtual pose face images. For ICLN, only profile images are used as inputs, and we propose to adopt Identity Consistence Loss to minimize the intra-class feature variation across different poses. The collaborative learning of two sub-modules guarantees that the parameters of network are updated in a relatively equal probability between near-frontal face images and profile images, so that the pose-based long-tailed problem can be effectively addressed. The proposed SSN shows comparable results over the state-of-the-art methods on several public datasets. In this study, LightCNN is selected as the backbone of SSN, and existing popular networks also can be used into our framework for pose-robust face recognition.
人脸识别技术由于深度学习的发展已经取得了显著的成功。然而,现有的大多数人脸识别模型在面对姿态变化时表现不佳。我们认为,这主要是由于基于姿态的长尾数据造成的——训练样本在侧面人脸和正面人脸之间的分布不平衡。此外,由于姿态变化较大导致的面部纹理的自遮挡和非线性变形也增加了学习侧面人脸判别特征的难度。在这项研究中,我们提出了一种名为对称孪生网络(Symmetrical Siamese Network,SSN)的新框架,它可以同时克服基于姿态的长尾数据和姿态不变特征学习的局限性。具体来说,SSN 中提出了两个子模块,即特征一致性学习子网络(Feature-Consistence Learning sub-Net,FCLN)和身份一致性学习子网络(Identity-Consistence Learning sub-Net,ICLN)。对于 FCLN,输入是训练数据集中的所有人脸图像。受对比学习的启发,我们模拟人脸的姿态变化,并约束模型关注原始人脸图像与其对应虚拟姿态人脸图像之间的一致区域。对于 ICLN,仅使用侧面图像作为输入,我们提出采用身份一致性损失来最小化不同姿态下的类内特征变化。两个子模块的协同学习保证了网络参数在正面人脸图像和侧面人脸图像之间以相对相等的概率进行更新,从而有效地解决了基于姿态的长尾问题。在几个公开数据集上,所提出的 SSN 与最先进的方法相比取得了相当的结果。在本研究中,选择 LightCNN 作为 SSN 的骨干网络,现有的流行网络也可以被应用到我们的框架中,以实现对姿态鲁棒的人脸识别。