Mazzia Vittorio, Pedrani Alessandro, Caciolai Andrea, Rottmann Kay, Bernardi Davide
IEEE Trans Neural Netw Learn Syst. 2025 Jul;36(7):11759-11775. doi: 10.1109/TNNLS.2024.3498935.
Deep neural networks are becoming increasingly pervasive in academia and industry, matching and surpassing human performance in a wide variety of fields and related tasks. However, just as humans, even the largest artificial neural networks (ANNs) make mistakes, and once-correct predictions can become invalid as the world progresses in time. Augmenting datasets with samples that account for mistakes or up-to-date information has become a common workaround in practical applications. However, the well-known phenomenon of catastrophic forgetting poses a challenge in achieving precise changes in the implicitly memorized knowledge of neural network parameters, often requiring a full model retraining to achieve desired behaviors. That is expensive, unreliable, and incompatible with the current trend of large self-supervised pretraining, making it necessary to find more efficient and effective methods for adapting neural network models to changing data. To address this need, knowledge editing (KE) is emerging as a novel area of research that aims to enable reliable, data-efficient, and fast changes to a pretrained target model, without affecting model behaviors on previously learned tasks. In this survey, we provide a brief review of this recent artificial intelligence field of research. We first introduce the problem of editing neural networks, formalize it in a common framework and differentiate it from more notorious branches of research such as continuous learning. Next, we provide a review of the most relevant KE approaches and datasets proposed so far, grouping works under four different families: regularization techniques, meta-learning, direct model editing, and architectural strategies. Finally, we outline some intersections with other fields of research and potential directions for future works.
深度神经网络在学术界和工业界正变得越来越普遍,在各种各样的领域和相关任务中达到并超越了人类的表现。然而,与人类一样,即使是最大的人工神经网络(ANN)也会犯错,而且随着时间的推移,曾经正确的预测可能会变得无效。用考虑到错误或最新信息的样本扩充数据集已成为实际应用中的常见解决方法。然而,众所周知的灾难性遗忘现象对精确改变神经网络参数中隐含记忆的知识构成了挑战,通常需要对整个模型进行重新训练才能实现期望的行为。这既昂贵又不可靠,而且与当前大规模自监督预训练的趋势不兼容,因此有必要找到更高效有效的方法,使神经网络模型能够适应不断变化的数据。为了满足这一需求,知识编辑(KE)作为一个新的研究领域正在兴起,其目的是在不影响预训练目标模型在先前学习任务上的行为的情况下,实现对该模型进行可靠、数据高效且快速的改变。在本次综述中,我们对这一最新的人工智能研究领域进行简要概述。我们首先介绍编辑神经网络的问题,在一个通用框架中将其形式化,并将其与诸如持续学习等更著名的研究分支区分开来。接下来,我们对迄今为止提出的最相关的KE方法和数据集进行综述,将相关工作分为四个不同的类别:正则化技术、元学习、直接模型编辑和架构策略。最后,我们概述了与其他研究领域的一些交叉点以及未来工作的潜在方向。