Kumar Akhil, Bhattacharjee Swarnava, Kumar Ambrish, Jayakody Dushantha Nalin K
School of Computer Science Engineering and Technology, Bennett University, Greater Noida, India.
Liverpool John Moores University, Liverpool, England.
Sci Rep. 2025 Mar 17;15(1):9102. doi: 10.1038/s41598-025-93096-0.
Facial identity recognition is one of the challenging problems in the domain of computer vision. Facial identity comprises the facial attributes of a person's face ranging from age progression, gender, hairstyle, etc. Manipulating facial attributes such as changing the gender, hairstyle, expressions, and makeup changes the entire facial identity of a person which is often used by law offenders to commit crimes. Leveraging the deep learning-based approaches, this work proposes a one-step solution for facial attribute manipulation and detection leading to facial identity recognition in few-shot and traditional scenarios. As a first step towards performing facial identity recognition, we created the Facial Attribute Manipulation Detection (FAM) Dataset which consists of twenty unique identities with thirty-eight facial attributes generated by the StyleGAN3 inversion. The Facial Attribute Detection (FAM) Dataset has 11,560 images richly annotated in YOLO format. To perform facial attribute and identity detection, we developed the Spatial Transformer Block (STB) and Squeeze-Excite Spatial Pyramid Pooling (SE-SPP)-based Tiny YOLOv7 model and proposed as FIR-Tiny YOLOv7 (Facial Identity Recognition-Tiny YOLOv7) model. The proposed model is an improvised variant of the Tiny YOLOv7 model. For facial identity recognition, the proposed model achieved 10.0% higher mAP in the one-shot scenario, 30.4% higher mAP in the three-shot scenario, 15.3% higher mAP in the five-shot scenario, and 0.1% higher mAP in the traditional 70% - 30% split scenario as compared to the Tiny YOLOv7 model. The results obtained with the proposed model are promising for general facial identity recognition under varying facial attribute manipulation.
面部身份识别是计算机视觉领域中具有挑战性的问题之一。面部身份包括一个人面部的各种属性,如年龄增长、性别、发型等。操纵面部属性,如改变性别、发型、表情和妆容,会改变一个人的整体面部身份,而违法者常常利用这一点来犯罪。利用基于深度学习的方法,这项工作提出了一种一步式解决方案,用于面部属性操纵和检测,从而在少样本和传统场景中实现面部身份识别。作为实现面部身份识别的第一步,我们创建了面部属性操纵检测(FAM)数据集,该数据集由20个独特身份组成,包含通过StyleGAN3反演生成的38种面部属性。面部属性检测(FAM)数据集有11560张以YOLO格式进行了丰富注释的图像。为了进行面部属性和身份检测,我们开发了基于空间变换器模块(STB)和挤压激励空间金字塔池化(SE-SPP)的Tiny YOLOv7模型,并将其作为面部身份识别-Tiny YOLOv7(FIR-Tiny YOLOv7)模型提出。所提出的模型是Tiny YOLOv7模型的改进变体。对于面部身份识别,与Tiny YOLOv7模型相比,所提出的模型在单样本场景下mAP提高了10.0%,在三样本场景下mAP提高了30.4%,在五样本场景下mAP提高了15.3%,在传统的70%-30%分割场景下mAP提高了0.1%。所提出的模型所获得的结果对于在不同面部属性操纵下的一般面部身份识别很有前景。