Xia Haifeng, Jing Taotao, Ding Zhengming
IEEE Trans Image Process. 2023;32:1694-1704. doi: 10.1109/TIP.2023.3251103. Epub 2023 Mar 9.
Domain generalization (DG) aims to learn transferable knowledge from multiple source domains and generalize it to the unseen target domain. To achieve such expectation, the intuitive solution is to seek domain-invariant representations via generative adversarial mechanism or minimization of cross-domain discrepancy. However, the widespread imbalanced data scale problem across source domains and category in real-world applications becomes the key bottleneck of improving generalization ability of model due to its negative effect on learning the robust classification model. Motivated by this observation, we first formulate a practical and challenging imbalance domain generalization (IDG) scenario, and then propose a straightforward but effective novel method generative inference network (GINet), which augments reliable samples for minority domain/category to promote discriminative ability of the learned model. Concretely, GINet utilizes the available cross-domain images from the identical category and estimates their common latent variable, which derives to discover domain-invariant knowledge for unseen target domain. According to these latent variables, our GINet further generates more novel samples with optimal transport constraint and deploys them to enhance the desired model with more robustness and generalization ability. Considerable empirical analysis and ablation studies on three popular benchmarks under normal DG and IDG setups suggests the advantage of our method over other DG methods on elevating model generalization. The source code is available in GitHub https://github.com/HaifengXia/IDG.
域泛化(DG)旨在从多个源域学习可迁移的知识,并将其泛化到未见的目标域。为实现这一期望,直观的解决方案是通过生成对抗机制或最小化跨域差异来寻求域不变表示。然而,在实际应用中,源域和类别之间普遍存在的数据规模不平衡问题成为提高模型泛化能力的关键瓶颈,因为它对学习鲁棒分类模型有负面影响。受此观察启发,我们首先提出了一个实际且具有挑战性的不平衡域泛化(IDG)场景,然后提出了一种简单但有效的新方法——生成推理网络(GINet),该方法为少数域/类别扩充可靠样本,以提升所学模型的判别能力。具体而言,GINet利用来自相同类别的可用跨域图像,并估计它们的公共潜在变量,从而发现未见目标域的域不变知识。根据这些潜在变量,我们的GINet进一步通过最优传输约束生成更多新颖样本,并将它们用于增强所需模型,使其具有更高的鲁棒性和泛化能力。在正常DG和IDG设置下对三个流行基准进行的大量实证分析和消融研究表明,我们的方法在提升模型泛化能力方面优于其他DG方法。源代码可在GitHub上获取:https://github.com/HaifengXia/IDG 。