• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于持续学习的子空间蒸馏

Subspace distillation for continual learning.

作者信息

Roy Kaushik, Simon Christian, Moghadam Peyman, Harandi Mehrtash

机构信息

Monash University, Melbourne, VIC, Australia; CSIRO, Data61, Brisbane, QLD, Australia.

Monash University, Melbourne, VIC, Australia; Australian national University, Canberra, ACT, Australia; CSIRO, Data61, Brisbane, QLD, Australia.

出版信息

Neural Netw. 2023 Oct;167:65-79. doi: 10.1016/j.neunet.2023.07.047. Epub 2023 Aug 6.

DOI:10.1016/j.neunet.2023.07.047
PMID:37625243
Abstract

An ultimate objective in continual learning is to preserve knowledge learned in preceding tasks while learning new tasks. To mitigate forgetting prior knowledge, we propose a novel knowledge distillation technique that takes into the account the manifold structure of the latent/output space of a neural network in learning novel tasks. To achieve this, we propose to approximate the data manifold up-to its first order, hence benefiting from linear subspaces to model the structure and maintain the knowledge of a neural network while learning novel concepts. We demonstrate that the modeling with subspaces provides several intriguing properties, including robustness to noise and therefore effective for mitigating Catastrophic Forgetting in continual learning. We also discuss and show how our proposed method can be adopted to address both classification and segmentation problems. Empirically, we observe that our proposed method outperforms various continual learning methods on several challenging datasets including Pascal VOC, and Tiny-Imagenet. Furthermore, we show how the proposed method can be seamlessly combined with existing learning approaches to improve their performances. The codes of this article will be available at https://github.com/csiro-robotics/SDCL.

摘要

持续学习的一个最终目标是在学习新任务的同时保留在先前任务中学到的知识。为了减轻对先前知识的遗忘,我们提出了一种新颖的知识蒸馏技术,该技术在学习新任务时考虑了神经网络的潜在/输出空间的流形结构。为了实现这一点,我们建议将数据流形近似到其一阶,从而受益于线性子空间来对结构进行建模,并在学习新概念时保持神经网络的知识。我们证明,用子空间进行建模具有几个有趣的特性,包括对噪声的鲁棒性,因此对于减轻持续学习中的灾难性遗忘是有效的。我们还讨论并展示了如何采用我们提出的方法来解决分类和分割问题。从经验上看,我们观察到我们提出的方法在包括Pascal VOC和Tiny-Imagenet在内的几个具有挑战性的数据集上优于各种持续学习方法。此外,我们展示了所提出的方法如何能够无缝地与现有的学习方法相结合以提高它们的性能。本文的代码将在https://github.com/csiro-robotics/SDCL上提供。

相似文献

1
Subspace distillation for continual learning.用于持续学习的子空间蒸馏
Neural Netw. 2023 Oct;167:65-79. doi: 10.1016/j.neunet.2023.07.047. Epub 2023 Aug 6.
2
Continual learning with attentive recurrent neural networks for temporal data classification.用于时态数据分类的基于注意力循环神经网络的持续学习
Neural Netw. 2023 Jan;158:171-187. doi: 10.1016/j.neunet.2022.10.031. Epub 2022 Nov 11.
3
Uncertainty-Aware Contrastive Distillation for Incremental Semantic Segmentation.用于增量语义分割的不确定性感知对比蒸馏
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):2567-2581. doi: 10.1109/TPAMI.2022.3163806. Epub 2023 Jan 6.
4
Variational Data-Free Knowledge Distillation for Continual Learning.用于持续学习的变分无数据知识蒸馏
IEEE Trans Pattern Anal Mach Intell. 2023 Oct;45(10):12618-12634. doi: 10.1109/TPAMI.2023.3271626. Epub 2023 Sep 5.
5
Self-Net: Lifelong Learning via Continual Self-Modeling.自我网络:通过持续自我建模实现终身学习。
Front Artif Intell. 2020 Apr 9;3:19. doi: 10.3389/frai.2020.00019. eCollection 2020.
6
Improving transparency and representational generalizability through parallel continual learning.通过并行持续学习提高透明度和代表性泛化能力。
Neural Netw. 2023 Apr;161:449-465. doi: 10.1016/j.neunet.2023.02.007. Epub 2023 Feb 10.
7
Continual Learning Objective for Analyzing Complex Knowledge Representations.持续学习目标分析复杂知识表示。
Sensors (Basel). 2022 Feb 21;22(4):1667. doi: 10.3390/s22041667.
8
Continual Learning With Knowledge Distillation: A Survey.基于知识蒸馏的持续学习:一项综述。
IEEE Trans Neural Netw Learn Syst. 2024 Oct 18;PP. doi: 10.1109/TNNLS.2024.3476068.
9
Privacy-Preserving Synthetic Continual Semantic Segmentation for Robotic Surgery.机器人手术中的隐私保护合成持续语义分割。
IEEE Trans Med Imaging. 2024 Jun;43(6):2291-2302. doi: 10.1109/TMI.2024.3364969. Epub 2024 Jun 3.
10
CL3: Generalization of Contrastive Loss for Lifelong Learning.CL3:用于终身学习的对比损失的泛化
J Imaging. 2023 Nov 23;9(12):259. doi: 10.3390/jimaging9120259.

引用本文的文献

1
CL3: Generalization of Contrastive Loss for Lifelong Learning.CL3:用于终身学习的对比损失的泛化
J Imaging. 2023 Nov 23;9(12):259. doi: 10.3390/jimaging9120259.