Guo Yanan, Cao Lin, Du Kangning
Key Laboratory of Information and Communication Systems, Ministry of Information Industry, Beijing Information Science and Technology University, Beijing, China.
Key Laboratory of the Ministry of Education for Optoelectronic Measurement Technology and Instrument, Beijing Information Science and Technology University, Beijing, China.
Front Neurorobot. 2022 Jun 10;16:823484. doi: 10.3389/fnbot.2022.823484. eCollection 2022.
The task of sketch face recognition refers to matching cross-modality facial images from sketch to photo, which is widely applied in the criminal investigation area. Existing works aim to bridge the cross-modality gap by inter-modality feature alignment approaches, however, the small sample problem has received much less attention, resulting in limited performance. In this paper, an effective Cross Task Modality Alignment Network (CTMAN) is proposed for sketch face recognition. To address the small sample problem, a meta learning training episode strategy is first introduced to mimic few-shot tasks. Based on the episode strategy, a two-stream network termed modality alignment embedding learning is used to capture more modality-specific and modality-sharable features, meanwhile, two cross task memory mechanisms are proposed to collect sufficient negative features to further improve the feature learning. Finally, a cross task modality alignment loss is proposed to capture modality-related information of cross task features for more effective training. Extensive experiments are conducted to validate the superiority of the CTMAN, which significantly outperforms state-of-the-art methods on the UoM-SGFSv2 set A, set B, CUFSF, and PRIP-VSGC dataset.
草图人脸识别任务是指将跨模态的面部图像从草图匹配到照片,该任务在刑事侦查领域有着广泛的应用。现有工作旨在通过跨模态特征对齐方法来弥合跨模态差距,然而,小样本问题却很少受到关注,导致性能有限。在本文中,我们提出了一种用于草图人脸识别的有效跨任务模态对齐网络(CTMAN)。为了解决小样本问题,我们首先引入了一种元学习训练情节策略来模拟少样本任务。基于该情节策略,我们使用了一种称为模态对齐嵌入学习的双流网络来捕获更多特定模态和模态共享的特征,同时,提出了两种跨任务记忆机制来收集足够的负特征以进一步改进特征学习。最后,我们提出了一种跨任务模态对齐损失,以捕获跨任务特征的模态相关信息,从而进行更有效的训练。我们进行了大量实验来验证CTMAN的优越性,它在UoM-SGFSv2集A、集B、CUFSF和PRIP-VSGC数据集上显著优于现有方法。