Chen Huanhuan, Jiang Bingbing, Yao Xin
IEEE Trans Neural Netw Learn Syst. 2018 Nov;29(11):5366-5379. doi: 10.1109/TNNLS.2017.2784814. Epub 2018 Mar 1.
Negative correlation learning (NCL) is an ensemble learning algorithm that introduces a correlation penalty term to the cost function of each individual ensemble member. Each ensemble member minimizes its mean square error and its error correlation with the rest of the ensemble. This paper analyzes NCL and reveals that adopting a negative correlation term for unlabeled data is beneficial to improving the model performance in the semisupervised learning (SSL) setting. We then propose a novel SSL algorithm, Semisupervised NCL (SemiNCL) algorithm. The algorithm considers the negative correlation terms for both labeled and unlabeled data for the semisupervised problems. In order to reduce the computational and memory complexity, an accelerated SemiNCL is derived from the distributed least square algorithm. In addition, we have derived a bound for two parameters in SemiNCL based on an analysis of the Hessian matrix of the error function. The new algorithm is evaluated by extensive experiments with various ratios of labeled and unlabeled training data. Comparisons with other state-of-the-art supervised and semisupervised algorithms confirm that SemiNCL achieves the best overall performance.
负相关学习(NCL)是一种集成学习算法,它在每个单独的集成成员的代价函数中引入了一个相关惩罚项。每个集成成员最小化其均方误差以及与集成中其他成员的误差相关性。本文分析了NCL,并揭示了对未标记数据采用负相关项有利于在半监督学习(SSL)设置中提高模型性能。然后,我们提出了一种新颖的SSL算法,即半监督NCL(SemiNCL)算法。该算法针对半监督问题考虑了标记数据和未标记数据的负相关项。为了降低计算和内存复杂度,基于分布式最小二乘算法推导了一种加速的SemiNCL。此外,基于对误差函数的海森矩阵的分析,我们推导了SemiNCL中两个参数的界限。通过使用各种标记和未标记训练数据比例进行的大量实验对新算法进行了评估。与其他最新的监督和半监督算法的比较证实,SemiNCL实现了最佳的整体性能。