Zhang Wei, Gu Xiaodong
Department of Electronic Engineering, Fudan University, Shanghai 200438, China.
Entropy (Basel). 2023 May 10;25(5):776. doi: 10.3390/e25050776.
Few shot class incremental learning (FSCIL) is an extremely challenging but valuable problem in real-world applications. When faced with novel few shot tasks in each incremental stage, it should take into account both catastrophic forgetting of old knowledge and overfitting of new categories with limited training data. In this paper, we propose an efficient prototype replay and calibration (EPRC) method with three stages to improve classification performance. We first perform effective pre-training with rotation and mix-up augmentations in order to obtain a strong backbone. Then a series of pseudo few shot tasks are sampled to perform meta-training, which enhances the generalization ability of both the feature extractor and projection layer and then helps mitigate the over-fitting problem of few shot learning. Furthermore, an even nonlinear transformation function is incorporated into the similarity computation to implicitly calibrate the generated prototypes of different categories and alleviate correlations among them. Finally, we replay the stored prototypes to relieve catastrophic forgetting and rectify prototypes to be more discriminative in the incremental-training stage via an explicit regularization within the loss function. The experimental results on CIFAR-100 and ImageNet demonstrate that our EPRC significantly boosts the classification performance compared with existing mainstream FSCIL methods.
少样本类别增量学习(FSCIL)在实际应用中是一个极具挑战性但很有价值的问题。在每个增量阶段面对新的少样本任务时,它既要考虑旧知识的灾难性遗忘,又要考虑在有限训练数据下新类别的过拟合问题。在本文中,我们提出了一种具有三个阶段的高效原型重放与校准(EPRC)方法来提高分类性能。我们首先通过旋转和混合增强进行有效的预训练,以获得一个强大的主干网络。然后采样一系列伪少样本任务进行元训练,这增强了特征提取器和投影层的泛化能力,进而有助于减轻少样本学习的过拟合问题。此外,在相似度计算中引入一个偶数非线性变换函数,以隐式校准不同类别的生成原型并减轻它们之间的相关性。最后,我们重放存储的原型以缓解灾难性遗忘,并在增量训练阶段通过损失函数内的显式正则化将原型校正得更具判别力。在CIFAR-100和ImageNet上的实验结果表明,与现有的主流FSCIL方法相比,我们的EPRC显著提高了分类性能。