Yousef Qais, Li Pu
Group of Process Optimization, Institute for Automation and Systems Engineering, Technische Universität Ilmenau, P.O. Box 100565, 98684, Ilmenau, Germany.
Sci Rep. 2025 Apr 24;15(1):14330. doi: 10.1038/s41598-025-91635-3.
Regularization is an important tool for the generalization of ANN models. Due to the lack of constraints, it cannot guarantee that the model will work in a real environment with input data distribution changes. Inspired by neuroplasticity, this paper introduces a bounded regularization method that can be safely applied during the deployment phase. First, the reliability of neuron outputs is improved by extending our recent neuronal masking method to generate new supporting neurons. The model is then regularized by incorporating a synaptic connection module containing conenctions of the generated neurons to their previous layer. These connections are optimized online by introducing a synaptic rewiring process triggered by the information about the input distribution. This process is formulated as bilevel mixed-integer nonlinear programming (MINLP) with an objective to minimize the outer risk of the output by identifying the connections that minimize the inner risk of the neuron output. To address this optimization problem, a single-wave scheme is introduced to decompose the problem into smaller, parallel sub-problems that minimize the inner cost function while ensuring the aggregated solution to minimize the outer one. In addition, a storage/recovery memory module is proposed to memorize these connections and their corresponding risks, enabling the model to retrieve previous knowledge when encountering similar situations. Experimental results from classification and regression tasks show around 8% improvement in accuracy over state-of-the-art techniques. As a result, the proposed regularization method enhances the adaptability and robustness of ANN models in a variable environment.
正则化是人工神经网络(ANN)模型泛化的重要工具。由于缺乏约束,它不能保证模型在输入数据分布发生变化的实际环境中有效运行。受神经可塑性的启发,本文介绍了一种有界正则化方法,该方法可在部署阶段安全应用。首先,通过扩展我们最近的神经元掩码方法来生成新的支持神经元,从而提高神经元输出的可靠性。然后,通过合并一个突触连接模块对模型进行正则化,该模块包含生成的神经元与其前一层的连接。通过引入由输入分布信息触发的突触重新布线过程,在线优化这些连接。这个过程被表述为双层混合整数非线性规划(MINLP),其目标是通过识别使神经元输出的内部风险最小化的连接来最小化输出的外部风险。为了解决这个优化问题,引入了一种单波方案,将问题分解为更小的并行子问题,这些子问题在最小化内部成本函数的同时,确保聚合解最小化外部成本函数。此外,还提出了一个存储/恢复记忆模块来记忆这些连接及其相应的风险,使模型在遇到类似情况时能够检索先前的知识。分类和回归任务的实验结果表明,与现有技术相比,准确率提高了约8%。因此,所提出的正则化方法增强了ANN模型在可变环境中的适应性和鲁棒性。