School of Computer Science & Technology, China University of Mining and Technology, Xuzhou, China.
School of Computer Science & Technology, China University of Mining and Technology, Xuzhou, China.
Neural Netw. 2023 Sep;166:555-565. doi: 10.1016/j.neunet.2023.07.030. Epub 2023 Jul 28.
Complementary-label learning (CLL) is widely used in weakly supervised classification, but it faces a significant challenge in real-world datasets when confronted with class-imbalanced training samples. In such scenarios, the number of samples in one class is considerably lower than in other classes, which consequently leads to a decline in the accuracy of predictions. Unfortunately, existing CLL approaches have not investigate this problem. To alleviate this challenge, we propose a novel problem setting that enables learning from class-imbalanced complementary labels for multi-class classification. To tackle this problem, we propose a novel CLL approach called Weighted Complementary-Label Learning (WCLL). The proposed method models a weighted empirical risk minimization loss by utilizing the class-imbalanced complementary labels, which is also applicable to multi-class imbalanced training samples. Furthermore, we derive an estimation error bound to provide theoretical assurance. To evaluate our approach, we conduct extensive experiments on several widely-used benchmark datasets and a real-world dataset, and compare our method with existing state-of-the-art methods. The proposed approach shows significant improvement in these datasets, even in the case of multiple class-imbalanced scenarios. Notably, the proposed method not only utilizes complementary labels to train a classifier but also solves the problem of class imbalance.
互补标签学习(CLL)在弱监督分类中得到了广泛的应用,但在处理具有不平衡训练样本的真实数据集时,它面临着重大的挑战。在这种情况下,一类样本的数量远远低于其他类,这导致预测的准确性下降。不幸的是,现有的 CLL 方法没有研究这个问题。为了缓解这一挑战,我们提出了一种新的问题设置,使我们能够从多类不平衡的互补标签中学习。为了解决这个问题,我们提出了一种新的 CLL 方法,称为加权互补标签学习(WCLL)。该方法通过利用不平衡的互补标签来建模加权经验风险最小化损失,也适用于多类不平衡训练样本。此外,我们还推导出一个估计误差界,以提供理论保证。为了评估我们的方法,我们在几个广泛使用的基准数据集和一个真实数据集上进行了大量实验,并将我们的方法与现有的最先进的方法进行了比较。该方法在这些数据集上取得了显著的改进,即使在多个不平衡类的情况下也是如此。值得注意的是,该方法不仅利用互补标签来训练分类器,而且还解决了类不平衡的问题。