Wang Chenyang, Jiang Junjun, Hu Xingyu, Liu Xianming, Ji Xiangyang
School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China.
Department of Automation, Tsinghua University, Beijing 100084, China.
Neural Netw. 2025 Apr;184:107053. doi: 10.1016/j.neunet.2024.107053. Epub 2024 Dec 20.
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks, as old data from previous tasks is unavailable when learning a new task. To address this, some methods propose replaying data from previous tasks during new task learning, typically using extra memory to store replay data. However, it is not expected in practice due to memory constraints and data privacy issues. Instead, data-free replay methods invert samples from the classification model. While effective, these methods face inconsistencies between inverted and real training data, overlooked in recent works. To that effect, we propose to measure the data consistency quantitatively by some simplification and assumptions. Using this measurement, we gain insight to develop a novel loss function that reduces inconsistency. Specifically, the loss minimizes the KL divergence between distributions of inverted and real data under a tied multivariate Gaussian assumption, which is simple to implement in continual learning. Additionally, we observe that old class weight norms decrease continually as learning progresses. We analyze the reasons and propose a regularization term to balance class weights, making old class samples more distinguishable. To conclude, we introduce Consistency-enhanced data replay with a Debiased classifier for class incremental learning (CwD). Extensive experiments on CIFAR-100, Tiny-ImageNet, and ImageNet100 show consistently improved performance of CwD compared to previous approaches.
深度学习系统在从一系列任务中学习时容易出现灾难性遗忘,因为在学习新任务时,来自先前任务的旧数据不可用。为了解决这个问题,一些方法建议在新任务学习期间重放先前任务的数据,通常使用额外的内存来存储重放数据。然而,由于内存限制和数据隐私问题,在实践中这是不可行的。相反,无数据重放方法通过分类模型对样本进行反向生成。虽然这些方法有效,但它们面临反向生成数据与真实训练数据之间的不一致问题,而这在最近的研究中被忽视了。为此,我们建议通过一些简化和假设来定量测量数据一致性。利用这种测量方法,我们深入了解如何开发一种新的损失函数来减少不一致性。具体来说,在多元高斯分布的假设下,该损失函数最小化反向生成数据与真实数据分布之间的KL散度,这在持续学习中易于实现。此外,我们观察到随着学习的进行,旧类别的权重范数会持续下降。我们分析了原因并提出了一个正则化项来平衡类别权重,使旧类别的样本更具可区分性。总之,我们引入了用于类增量学习的带有去偏置分类器的一致性增强数据重放方法(CwD)。在CIFAR-100、Tiny-ImageNet和ImageNet100上进行的大量实验表明,与先前的方法相比,CwD的性能持续提高。