Bertoin David, Rachelson Emmanuel
IEEE Trans Neural Netw Learn Syst. 2024 May;35(5):6693-6702. doi: 10.1109/TNNLS.2022.3212620. Epub 2024 May 2.
Deep neural networks have demonstrated their ability to automatically extract meaningful features from data. However, in supervised learning, information specific to the dataset used for training, but irrelevant to the task at hand, may remain encoded in the extracted representations. This remaining information introduces a domain-specific bias, weakening the generalization performance. In this work, we propose splitting the information into a task-related representation and its complementary context representation. We propose an original method, combining adversarial feature predictors and cyclic reconstruction, to disentangle these two representations in the single-domain supervised case. We then adapt this method to the unsupervised domain adaptation (UDA) problem, consisting of training a model capable of performing on both a source and a target domain. In particular, our method promotes disentanglement in the target domain, despite the absence of training labels. This enables the isolation of task-specific information from both domains and a projection into a common representation. The task-specific representation allows the efficient transfer of knowledge acquired from the source domain to the target domain. In the single-domain case, we demonstrate the quality of our representations on information retrieval tasks and the generalization benefits induced by sharpened task-specific representations. We then validate the proposed method on several classical domain adaptation (DA) benchmarks and illustrate the benefits of disentanglement for DA.
深度神经网络已展现出从数据中自动提取有意义特征的能力。然而,在监督学习中,特定于用于训练的数据集但与手头任务无关的信息,可能仍会编码在提取的表示中。这种残留信息会引入特定领域的偏差,削弱泛化性能。在这项工作中,我们建议将信息拆分为与任务相关的表示及其互补的上下文表示。我们提出一种原始方法,结合对抗性特征预测器和循环重建,以在单域监督情况下解开这两种表示。然后,我们将此方法应用于无监督域适应(UDA)问题,即训练一个能够在源域和目标域上都能运行的模型。特别是,我们的方法促进了目标域中的解缠,尽管没有训练标签。这使得能够从两个域中分离出特定于任务的信息,并投影到一个共同的表示中。特定于任务的表示允许将从源域获取的知识有效地转移到目标域。在单域情况下,我们在信息检索任务上展示了我们表示的质量以及锐化的特定于任务的表示所带来的泛化优势。然后,我们在几个经典的域适应(DA)基准上验证了所提出的方法,并说明了解缠对域适应的好处。