Xu Huali, Liu Li, Zhi Shuaifeng, Fu Shaojing, Su Zhuo, Cheng Ming-Ming, Liu Yongxiang
IEEE Trans Image Process. 2024;33:2058-2073. doi: 10.1109/TIP.2024.3374222. Epub 2024 Mar 18.
Existing Cross-Domain Few-Shot Learning (CDFSL) methods require access to source domain data to train a model in the pre-training phase. However, due to increasing concerns about data privacy and the desire to reduce data transmission and training costs, it is necessary to develop a CDFSL solution without accessing source data. For this reason, this paper explores a Source-Free CDFSL (SF-CDFSL) problem, in which CDFSL is addressed through the use of existing pretrained models instead of training a model with source data, avoiding accessing source data. However, due to the lack of source data, we face two key challenges: effectively tackling CDFSL with limited labeled target samples, and the impossibility of addressing domain disparities by aligning source and target domain distributions. This paper proposes an Enhanced Information Maximization with Distance-Aware Contrastive Learning (IM-DCL) method to address these challenges. Firstly, we introduce the transductive mechanism for learning the query set. Secondly, information maximization (IM) is explored to map target samples into both individual certainty and global diversity predictions, helping the source model better fit the target data distribution. However, IM fails to learn the decision boundary of the target task. This motivates us to introduce a novel approach called Distance-Aware Contrastive Learning (DCL), in which we consider the entire feature set as both positive and negative sets, akin to Schrödinger's concept of a dual state. Instead of a rigid separation between positive and negative sets, we employ a weighted distance calculation among features to establish a soft classification of the positive and negative sets for the entire feature set. We explore three types of negative weights to enhance the performance of CDFSL. Furthermore, we address issues related to IM by incorporating contrastive constraints between object features and their corresponding positive and negative sets. Evaluations of the 4 datasets in the BSCD-FSL benchmark indicate that the proposed IM-DCL, without accessing the source domain, demonstrates superiority over existing methods, especially in the distant domain task. Additionally, the ablation study and performance analysis confirmed the ability of IM-DCL to handle SF-CDFSL. The code will be made public at https://github.com/xuhuali-mxj/IM-DCL.
现有的跨域少样本学习(CDFSL)方法在预训练阶段需要访问源域数据来训练模型。然而,由于对数据隐私的担忧日益增加,以及降低数据传输和训练成本的需求,有必要开发一种无需访问源数据的CDFSL解决方案。因此,本文探讨了一种无源CDFSL(SF-CDFSL)问题,其中通过使用现有的预训练模型而不是用源数据训练模型来解决CDFSL问题,从而避免访问源数据。然而,由于缺乏源数据,我们面临两个关键挑战:如何利用有限的有标签目标样本有效地解决CDFSL问题,以及无法通过对齐源域和目标域分布来解决域差异问题。本文提出了一种基于距离感知对比学习的增强信息最大化(IM-DCL)方法来应对这些挑战。首先,我们引入了用于学习查询集的转导机制。其次,探索信息最大化(IM)将目标样本映射到个体确定性和全局多样性预测中,帮助源模型更好地拟合目标数据分布。然而,IM未能学习到目标任务的决策边界。这促使我们引入一种名为距离感知对比学习(DCL)的新方法,在该方法中,我们将整个特征集视为正集和负集,类似于薛定谔的双态概念。我们不是在正集和负集之间进行严格分离,而是在特征之间采用加权距离计算,为整个特征集建立正集和负集的软分类。我们探索了三种类型的负权重以提高CDFSL的性能。此外,我们通过在对象特征与其相应的正集和负集之间纳入对比约束来解决与IM相关的问题。在BSCD-FSL基准测试中的4个数据集上的评估表明,所提出的IM-DCL在不访问源域的情况下,优于现有方法,特别是在远域任务中。此外,消融研究和性能分析证实了IM-DCL处理SF-CDFSL的能力。代码将在https://github.com/xuhuali-mxj/IM-DCL上公开。