Wang Song, Guo Xin, Tie Yun, Qi Lin, Guan Ling
IEEE Trans Image Process. 2020 Oct 13;PP. doi: 10.1109/TIP.2020.3029424.
Local feature descriptor learning aims to represent distinctive images or patches with the same local features, where their representation is invariant under different types of deformation. Recent studies have demonstrated that descriptor learning based on Convolutional Neural Network (CNN) is able to improve the matching performance significantly. However, they tend to ignore the importance of sample selection during the training process, leading to unstable quality of descriptors and learning efficiency. In this paper, a dual hard batch construction method is proposed to sample the hard matching and non-matching examples for training, improving the performance of the descriptor learning on different tasks. To construct the dual hard training batches, the matching examples with the minimum similarity are selected as the hard positive pairs. For each positive pair, the most similar non-matching example is then sampled from the generated hard positive pairs in the same batch as the corresponding negative. By sampling the hard positive pairs and the corresponding hard negatives, the hard batches are produced to force the CNN model to learn the descriptors with more efforts. In addition, based on the above dual hard batch construction, an ℓ22 triplet loss function is built for optimizing the training model. Specifically, we analyze the superiority of the ℓ22 loss function when dealing with hard examples, and also demonstrate it in the experiments. With the benefits of the proposed sampling strategy and the ℓ22 triplet loss function, our method achieves better performance compared to state-of-the-art on the reference benchmarks for different matching tasks.
局部特征描述符学习旨在用相同的局部特征来表示独特的图像或图像块,其表示在不同类型的变形下具有不变性。最近的研究表明,基于卷积神经网络(CNN)的描述符学习能够显著提高匹配性能。然而,它们往往在训练过程中忽略样本选择的重要性,导致描述符质量不稳定和学习效率低下。本文提出了一种双硬批次构建方法,用于对困难匹配和非匹配示例进行采样以进行训练,提高描述符学习在不同任务上的性能。为了构建双硬训练批次,选择相似度最小的匹配示例作为硬正样本对。对于每一个正样本对,然后从同一批次中生成的硬正样本对中采样最相似的非匹配示例作为相应的负样本。通过对硬正样本对和相应的硬负样本进行采样,生成硬批次以迫使CNN模型更加努力地学习描述符。此外,基于上述双硬批次构建,构建了一个ℓ22三元组损失函数来优化训练模型。具体来说,我们分析了ℓ22损失函数在处理困难示例时的优越性,并在实验中进行了验证。受益于所提出的采样策略和ℓ22三元组损失函数,我们的方法在不同匹配任务的参考基准上比现有技术取得了更好的性能。