IEEE Trans Neural Netw Learn Syst. 2013 Jun;24(6):888-99. doi: 10.1109/TNNLS.2013.2246188.
Traditional learning algorithms applied to complex and highly imbalanced training sets may not give satisfactory results when distinguishing between examples of the classes. The tendency is to yield classification models that are biased towards the overrepresented (majority) class. This paper investigates this class imbalance problem in the context of multilayer perceptron (MLP) neural networks. The consequences of the equal cost (loss) assumption on imbalanced data are formally discussed from a statistical learning theory point of view. A new cost-sensitive algorithm (CSMLP) is presented to improve the discrimination ability of (two-class) MLPs. The CSMLP formulation is based on a joint objective function that uses a single cost parameter to distinguish the importance of class errors. The learning rule extends the Levenberg-Marquadt's rule, ensuring the computational efficiency of the algorithm. In addition, it is theoretically demonstrated that the incorporation of prior information via the cost parameter may lead to balanced decision boundaries in the feature space. Based on the statistical analysis of results on real data, our approach shows a significant improvement of the area under the receiver operating characteristic curve and G-mean measures of regular MLPs.
当对类别示例进行区分时,应用于复杂且高度不平衡的训练集的传统学习算法可能无法给出满意的结果。其趋势是产生偏向于代表性过高(多数)类别的分类模型。本文在多层感知器(MLP)神经网络的上下文中研究了这种类别不平衡问题。从统计学习理论的角度,正式讨论了等代价(损失)假设对不平衡数据的影响。提出了一种新的基于代价敏感的算法(CSMLP)来提高(两类)MLP 的区分能力。CSMLP 公式基于一个联合目标函数,该函数使用单个代价参数来区分类错误的重要性。学习规则扩展了 Levenberg-Marquadt 规则,确保了算法的计算效率。此外,从理论上证明了通过代价参数合并先验信息可能导致特征空间中的平衡决策边界。基于对真实数据结果的统计分析,我们的方法在接收者操作特性曲线和常规 MLP 的 G-均值度量的面积方面显示出显著的改进。