IEEE Trans Cybern. 2021 Apr;51(4):1849-1859. doi: 10.1109/TCYB.2019.2909480. Epub 2021 Mar 17.
Person reidentification (Re-ID) aims to match observations of individuals across multiple nonoverlapping camera views. Recently, metric learning-based methods have played important roles in addressing this task. However, metrics are mostly learned in supervised manners, of which the performance relies heavily on the quantity and quality of manual annotations. Meanwhile, metric learning-based algorithms generally project person features into a common subspace, in which the extracted features are shared by all views. However, it may result in information loss since these algorithms neglect the view-specific features. Besides, they assume person samples of different views are taken from the same distribution. Conversely, these samples are more likely to obey different distributions due to view condition changes. To this end, this paper proposes an unsupervised cross-view metric learning method based on the properties of data distributions. Specifically, person samples in each view are taken from a mixture of two distributions: one models common prosperities among camera views and the other focuses on view-specific properties. Based on this, we introduce a shared mapping to explore the shared features. Meanwhile, we construct view-specific mappings to extract and project view-related features into a common subspace. As a result, samples in the transformed subspace follow the same distribution and are equipped with comprehensive representations. In this paper, these mappings are learned in an unsupervised manner by clustering samples in the projected space. Experimental results on five cross-view datasets validate the effectiveness of the proposed method.
人体重识别(Re-ID)旨在匹配跨多个非重叠摄像机视角的个体观察结果。最近,基于度量学习的方法在解决这个任务方面发挥了重要作用。然而,度量值主要是通过监督学习方式学习的,其性能严重依赖于手动标注的数量和质量。同时,基于度量学习的算法通常将人体特征投影到一个公共子空间中,其中所有视图都共享提取的特征。然而,由于这些算法忽略了特定于视图的特征,因此可能会导致信息丢失。此外,它们假设不同视图的人体样本来自相同的分布。相反,由于视图条件的变化,这些样本更有可能服从不同的分布。为此,本文提出了一种基于数据分布特性的无监督跨视图度量学习方法。具体来说,每个视图中的人体样本取自两个分布的混合:一个分布模型化了相机视图之间的共同特征,另一个分布专注于特定于视图的特征。在此基础上,我们引入了一个共享映射来探索共享特征。同时,我们构建了特定于视图的映射,将视图相关的特征提取并投影到一个公共子空间中。结果,变换后的子空间中的样本遵循相同的分布,并具有全面的表示。在本文中,这些映射是通过在投影空间中对样本进行聚类以无监督的方式学习的。在五个跨视图数据集上的实验结果验证了所提出方法的有效性。