Li Shuang, Xie Binhui, Lin Qiuxia, Liu Chi Harold, Huang Gao, Wang Guoren
IEEE Trans Pattern Anal Mach Intell. 2022 Aug;44(8):4093-4109. doi: 10.1109/TPAMI.2021.3062644. Epub 2022 Jul 1.
Domain adaptation (DA) attempts to transfer knowledge learned in the labeled source domain to the unlabeled but related target domain without requiring large amounts of target supervision. Recent advances in DA mainly proceed by aligning the source and target distributions. Despite the significant success, the adaptation performance still degrades accordingly when the source and target domains encounter a large distribution discrepancy. We consider this limitation may attribute to the insufficient exploration of domain-specialized features because most studies merely concentrate on domain-general feature learning in task-specific layers and integrate totally-shared convolutional networks (convnets) to generate common features for both domains. In this paper, we relax the completely-shared convnets assumption adopted by previous DA methods and propose Domain Conditioned Adaptation Network (DCAN), which introduces domain conditioned channel attention module with a multi-path structure to separately excite channel activation for each domain. Such a partially-shared convnets module allows domain-specialized features in low-level to be explored appropriately. Further, given the knowledge transferability varying along with convolutional layers, we develop Generalized Domain Conditioned Adaptation Network (GDCAN) to automatically determine whether domain channel activations should be separately modeled in each attention module. Afterward, the critical domain-specialized knowledge could be adaptively extracted according to the domain statistic gaps. As far as we know, this is the first work to explore the domain-wise convolutional channel activations separately for deep DA networks. Additionally, to effectively match high-level feature distributions across domains, we consider deploying feature adaptation blocks after task-specific layers, which can explicitly mitigate the domain discrepancy. Extensive experiments on four cross-domain benchmarks, including DomainNet, Office-Home, Office-31, and ImageCLEF, demonstrate the proposed approaches outperform the existing methods by a large margin, especially on the large-scale challenging dataset. The code and models are available at https://github.com/BIT-DA/GDCAN.
域适应(DA)旨在将在有标签的源域中学习到的知识转移到无标签但相关的目标域,而无需大量的目标监督。DA的最新进展主要通过对齐源域和目标域的分布来实现。尽管取得了显著成功,但当源域和目标域存在较大的分布差异时,适应性能仍会相应下降。我们认为这种局限性可能归因于对域特定特征的探索不足,因为大多数研究仅专注于任务特定层中的域通用特征学习,并集成完全共享的卷积网络(convnet)以为两个域生成通用特征。在本文中,我们放宽了先前DA方法采用的完全共享convnet假设,提出了域条件适应网络(DCAN),该网络引入了具有多路径结构的域条件通道注意力模块,以分别激发每个域的通道激活。这种部分共享的convnet模块允许适当地探索低级别的域特定特征。此外,鉴于知识可迁移性随卷积层而变化,我们开发了广义域条件适应网络(GDCAN),以自动确定是否应在每个注意力模块中分别对域通道激活进行建模。然后,可以根据域统计差距自适应地提取关键的域特定知识。据我们所知,这是第一项针对深度DA网络分别探索按域卷积通道激活的工作。此外,为了有效地跨域匹配高级特征分布,我们考虑在任务特定层之后部署特征适应块,这可以明确减轻域差异。在包括DomainNet、Office-Home、Office-31和ImageCLEF在内的四个跨域基准上进行的大量实验表明,所提出的方法在很大程度上优于现有方法,尤其是在大规模具有挑战性的数据集上。代码和模型可在https://github.com/BIT-DA/GDCAN上获取。