Chen Huanhuan, Yao Xin
The Centre of Excellence for Research in Computational Intelligence and Applications (CERCIA), School of Computer Science, University of Birmingham, Birmingham, UK.
IEEE Trans Neural Netw. 2009 Dec;20(12):1962-79. doi: 10.1109/TNN.2009.2034144. Epub 2009 Nov 17.
Negative correlation learning (NCL) is a neural network ensemble learning algorithm that introduces a correlation penalty term to the cost function of each individual network so that each neural network minimizes its mean square error (MSE) together with the correlation of the ensemble. This paper analyzes NCL and reveals that the training of NCL (when lambda = 1) corresponds to training the entire ensemble as a single learning machine that only minimizes the MSE without regularization. This analysis explains the reason why NCL is prone to overfitting the noise in the training set. This paper also demonstrates that tuning the correlation parameter lambda in NCL by cross validation cannot overcome the overfitting problem. The paper analyzes this problem and proposes the regularized negative correlation learning (RNCL) algorithm which incorporates an additional regularization term for the whole ensemble. RNCL decomposes the ensemble's training objectives, including MSE and regularization, into a set of sub-objectives, and each sub-objective is implemented by an individual neural network. In this paper, we also provide a Bayesian interpretation for RNCL and provide an automatic algorithm to optimize regularization parameters based on Bayesian inference. The RNCL formulation is applicable to any nonlinear estimator minimizing the MSE. The experiments on synthetic as well as real-world data sets demonstrate that RNCL achieves better performance than NCL, especially when the noise level is nontrivial in the data set.
负相关学习(NCL)是一种神经网络集成学习算法,它在每个个体网络的代价函数中引入了一个相关惩罚项,使得每个神经网络在最小化其均方误差(MSE)的同时,还要最小化集成的相关性。本文对NCL进行了分析,并揭示了NCL的训练(当λ = 1时)相当于将整个集成作为一个单一的学习机器进行训练,该学习机器仅最小化MSE而不进行正则化。这一分析解释了NCL为何容易过度拟合训练集中的噪声。本文还表明,通过交叉验证调整NCL中的相关参数λ并不能克服过度拟合问题。本文对该问题进行了分析,并提出了正则化负相关学习(RNCL)算法,该算法为整个集成引入了一个额外的正则化项。RNCL将集成的训练目标,包括MSE和正则化,分解为一组子目标,每个子目标由一个个体神经网络实现。在本文中,我们还为RNCL提供了贝叶斯解释,并基于贝叶斯推理提供了一种自动优化正则化参数的算法。RNCL公式适用于任何最小化MSE的非线性估计器。在合成数据集和真实世界数据集上的实验表明,RNCL比NCL具有更好的性能,特别是当数据集中的噪声水平较大时。