Lu Junxin, Sun Shiliang
IEEE Trans Image Process. 2024;33:3353-3368. doi: 10.1109/TIP.2024.3403053. Epub 2024 May 31.
Continual zero-shot learning (CZSL) aims to develop a model that accumulates historical knowledge to recognize unseen tasks, while eliminating catastrophic forgetting for seen tasks when learning new tasks. However, existing CZSL methods, while mitigating catastrophic forgetting for old tasks, often lead to negative transfer problem for new tasks by over-focusing on accumulating old knowledge and neglecting the plasticity of the model for learning new tasks. To tackle these problems, we propose PAMK, a prototype augmented multi-teacher knowledge transfer network that strikes a trade-off between recognition stability for old tasks and generalization plasticity for new tasks. PAMK consists of a prototype augmented contrastive generation (PACG) module and a multi-teacher knowledge transfer (MKT) module. To reduce the cumulative semantic decay of the class representation embedding and mitigate catastrophic forgetting, we propose a continual prototype augmentation strategy based on relevance scores in PACG. Furthermore, by introducing the prototype augmented semantic-visual contrastive loss, PACG promotes intra-class compactness for all classes across all tasks. MKT effectively accumulates semantic knowledge learned from old tasks to recognize new tasks via the proposed multi-teacher knowledge transfer, eliminating the negative transfer problem. Extensive experiments on various CZSL settings demonstrate the superior performance of PAMK compared to state-of-the-art methods. In particular, in the practical task-free CZSL setting, PAMK achieves impressive gains of 3.28%, 3.09% and 3.71% in mean harmonic accuracy on the CUB, AWA1, and AWA2 datasets, respectively.
持续零样本学习(CZSL)旨在开发一种模型,该模型能够积累历史知识以识别未见任务,同时在学习新任务时消除对已见任务的灾难性遗忘。然而,现有的CZSL方法在减轻旧任务的灾难性遗忘时,往往因过度关注积累旧知识而忽视模型学习新任务的可塑性,从而导致新任务的负迁移问题。为了解决这些问题,我们提出了PAMK,即原型增强多教师知识转移网络,它在旧任务的识别稳定性和新任务的泛化可塑性之间取得了平衡。PAMK由一个原型增强对比生成(PACG)模块和一个多教师知识转移(MKT)模块组成。为了减少类表示嵌入的累积语义衰减并减轻灾难性遗忘,我们在PACG中提出了一种基于相关性分数的持续原型增强策略。此外,通过引入原型增强语义-视觉对比损失,PACG促进了所有任务中所有类别的类内紧凑性。MKT通过所提出的多教师知识转移有效地积累从旧任务中学到的语义知识以识别新任务,消除了负迁移问题。在各种CZSL设置上进行的大量实验表明,与现有方法相比,PAMK具有卓越的性能。特别是,在实际的无任务CZSL设置中,PAMK在CUB、AWA1和AWA2数据集上的平均调和准确率分别取得了3.28%、3.09%和3.71%的显著提升。