School of Computer Science, School of Cyberspace Science, Xiangtan University, Xiangtan, Hunan 411105, China.
School of Mechanical Engineering, Xiangtan University, Xiangtan, Hunan 411105, China.
Neural Netw. 2023 Nov;168:602-614. doi: 10.1016/j.neunet.2023.09.045. Epub 2023 Sep 27.
Unsupervised domain adaptation (UDA) trains models using labeled data from a specific source domain and then transferring the knowledge to certain target domains that have few or no labels. Many prior measurement-based works achieve lots of progress, but their feature distinguishing abilities to classify target samples with similar features are not enough; they do not adequately consider the confusing samples in the target domain that are similar to the source domain; and they don't consider negative transfer of the outlier sample in source domain. We address these issues in our work and propose an UDA method with asymmetrical margin disparity loss and outlier sample extraction, called AMD-Net with OSE. We propose an Asymmetrical Margin Disparity Discrepancy (AMD) method and a training strategy based on sample selection mechanism to make the network have better feature extraction ability and the network gets rid of local optimal. Firstly, in the AMD method, we design a multi-label entropy metric to evaluate the marginal disparity loss of the confusing samples in the target domain. This asymmetric marginal disparity loss designment uses the different entropy measurement algorithms of the two domains to excavate the differences of the two domains as much as possible, so as to find the common features of the two domains. Secondly, A sample selection mechanism is designed to evaluate which part of the sample in target domain is confusable. We define the certainty of the sample in the target domain, adopt a progressive learning scheme, and adopt one-hot marginal disparity loss for most of the samples in the target domain with low uncertainty and easy to distinguish. The multi-label marginal calculation method is used only for the uncertainty samples in the target domain whose certainty is less than the threshold value, so that the network can get rid of the local optimal as much as possible. At last, we further propose an outlier sample extraction algorithm (OSE) based on weighted cosine similarity distance for source domain to reduce the negative migration effect caused by outlier samples in the source domain. Extensive experiments on four datasets Office-31, Office-Home, VisDA-2017 and DomainNet demonstrate that our method works well in various UDA settings and outperforms the state-of-the-art methods.
无监督领域自适应 (UDA) 使用特定源域的带标签数据训练模型,然后将知识转移到带有少量或没有标签的特定目标域。许多基于度量的前期工作取得了很多进展,但它们对具有相似特征的目标样本的分类特征区分能力还不够;它们没有充分考虑到与源域相似的目标域中的混淆样本;并且它们没有考虑源域中异常样本的负迁移。我们在工作中解决了这些问题,并提出了一种带有不对称边缘差异损失和异常样本提取的 UDA 方法,称为 AMD-Net 与 OSE。我们提出了一种不对称边缘差异差异 (AMD) 方法和基于样本选择机制的训练策略,使网络具有更好的特征提取能力,并摆脱局部最优。首先,在 AMD 方法中,我们设计了一个多标签熵度量来评估目标域中混淆样本的边缘差异损失。这种不对称的边缘差异损失设计使用两个域的不同熵度量算法,尽可能多地挖掘两个域之间的差异,从而找到两个域的共同特征。其次,设计了一种样本选择机制来评估目标域中的哪部分样本是混淆的。我们定义了目标域中样本的确定性,采用渐进式学习方案,对于目标域中不确定性低、易于区分的大部分样本,采用单热边缘差异损失。只有当目标域中确定性小于阈值的不确定性样本才采用多标签边缘计算方法,以使网络尽可能摆脱局部最优。最后,我们进一步提出了一种基于加权余弦相似度距离的源域异常样本提取算法 (OSE),以减少源域中异常样本带来的负迁移效应。在四个数据集 Office-31、Office-Home、VisDA-2017 和 DomainNet 上进行的广泛实验表明,我们的方法在各种 UDA 设置下都能很好地工作,并优于最先进的方法。