Guo Huaping, Zhi Weimei, Liu Hongbing, Xu Mingliang
School of Computer and Information Technology, Xinyang Normal University, Xinyang 464000, China.
School of Information Engineering, Zhengzhou University, Zhengzhou 450000, China.
Comput Intell Neurosci. 2016;2016:5423204. doi: 10.1155/2016/5423204. Epub 2016 Jan 4.
In recent years, imbalanced learning problem has attracted more and more attentions from both academia and industry, and the problem is concerned with the performance of learning algorithms in the presence of data with severe class distribution skews. In this paper, we apply the well-known statistical model logistic discrimination to this problem and propose a novel method to improve its performance. To fully consider the class imbalance, we design a new cost function which takes into account the accuracies of both positive class and negative class as well as the precision of positive class. Unlike traditional logistic discrimination, the proposed method learns its parameters by maximizing the proposed cost function. Experimental results show that, compared with other state-of-the-art methods, the proposed one shows significantly better performance on measures of recall, g-mean, f-measure, AUC, and accuracy.
近年来,不平衡学习问题越来越受到学术界和工业界的关注,该问题涉及在存在严重类分布偏斜的数据情况下学习算法的性能。在本文中,我们将著名的统计模型逻辑判别应用于这个问题,并提出一种新颖的方法来提高其性能。为了充分考虑类不平衡,我们设计了一个新的成本函数,该函数兼顾了正类和负类的准确率以及正类的精确率。与传统的逻辑判别不同,所提出的方法通过最大化所提出的成本函数来学习其参数。实验结果表明,与其他现有方法相比,所提出的方法在召回率、g均值、F值、AUC和准确率等指标上表现出显著更好的性能。