Zhang Xin, Chen Ying-Cong
IEEE Trans Image Process. 2023;32:4247-4258. doi: 10.1109/TIP.2023.3295739. Epub 2023 Jul 26.
Deep neural networks suffer from significant performance deterioration when there exists distribution shift between deployment and training. Domain Generalization (DG) aims to safely transfer a model to unseen target domains by only relying on a set of source domains. Although various DG approaches have been proposed, a recent study named DomainBed (Gulrajani and Lopez-Paz, 2020), reveals that most of them do not beat simple empirical risk minimization (ERM). To this end, we propose a general framework that is orthogonal to existing DG algorithms and could improve their performance consistently. Unlike previous DG works that stake on a static source model to be hopefully a universal one, our proposed AdaODM adaptively modifies the source model at test time for different target domains. Specifically, we create multiple domain-specific classifiers upon a shared domain-generic feature extractor. The feature extractor and classifiers are trained in an adversarial way, where the feature extractor embeds the input samples into a domain-invariant space, and the multiple classifiers capture the distinct decision boundaries that each of them relates to a specific source domain. During testing, distribution differences between target and source domains could be effectively measured by leveraging prediction disagreement among source classifiers. By fine-tuning source models to minimize the disagreement at test time, target-domain features are well aligned to the invariant feature space. We verify AdaODM on two popular DG methods, namely ERM and CORAL, and four DG benchmarks, namely VLCS, PACS, OfficeHome, and TerraIncognita. The results show AdaODM stably improves the generalization capacity on unseen domains and achieves state-of-the-art performance.
当部署和训练之间存在分布偏移时,深度神经网络的性能会显著下降。领域泛化(DG)旨在仅依靠一组源域将模型安全地转移到未见的目标域。尽管已经提出了各种DG方法,但最近一项名为DomainBed(Gulrajani和Lopez-Paz,2020)的研究表明,其中大多数方法并不优于简单的经验风险最小化(ERM)。为此,我们提出了一个与现有DG算法正交的通用框架,该框架可以持续提高它们的性能。与以往依赖静态源模型以期成为通用模型的DG工作不同,我们提出的AdaODM在测试时针对不同的目标域自适应地修改源模型。具体来说,我们在一个共享的领域通用特征提取器之上创建多个特定领域的分类器。特征提取器和分类器以对抗的方式进行训练,其中特征提取器将输入样本嵌入到一个领域不变空间中,多个分类器捕捉不同的决策边界,每个决策边界都与一个特定的源域相关。在测试期间,可以通过利用源分类器之间的预测不一致来有效测量目标域和源域之间的分布差异。通过微调源模型以在测试时最小化不一致,目标域特征可以很好地与不变特征空间对齐。我们在两种流行的DG方法(即ERM和CORAL)以及四个DG基准(即VLCS、PACS、OfficeHome和TerraIncognita)上验证了AdaODM。结果表明,AdaODM稳定地提高了在未见域上的泛化能力,并实现了当前最优的性能。