Yang Zhiyong, Xu Qianqian, Bao Shilong, Cao Xiaochun, Huang Qingming
IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):7747-7763. doi: 10.1109/TPAMI.2021.3101125. Epub 2022 Oct 4.
The Area under the ROC curve (AUC) is a well-known ranking metric for problems such as imbalanced learning and recommender systems. The vast majority of existing AUC-optimization-based machine learning methods only focus on binary-class cases, while leaving the multiclass cases unconsidered. In this paper, we start an early trial to consider the problem of learning multiclass scoring functions via optimizing multiclass AUC metrics. Our foundation is based on the M metric, which is a well-known multiclass extension of AUC. We first pay a revisit to this metric, showing that it could eliminate the imbalance issue from the minority class pairs. Motivated by this, we propose an empirical surrogate risk minimization framework to approximately optimize the M metric. Theoretically, we show that: (i) optimizing most of the popular differentiable surrogate losses suffices to reach the Bayes optimal scoring function asymptotically; (ii) the training framework enjoys an imbalance-aware generalization error bound, which pays more attention to the bottleneck samples of minority classes compared with the traditional O(√{1/N}) result. Practically, to deal with the low scalability of the computational operations, we propose acceleration methods for three popular surrogate loss functions, including the exponential loss, squared loss, and hinge loss, to speed up loss and gradient evaluations. Finally, experimental results on 11 real-world datasets demonstrate the effectiveness of our proposed framework. The code is now available at https://github.com/joshuaas/Learning-with-Multiclass-AUC-Theory-and-Algorithms.
ROC曲线下面积(AUC)是用于诸如不平衡学习和推荐系统等问题的一种知名排序指标。绝大多数现有的基于AUC优化的机器学习方法仅关注二分类情况,而未考虑多分类情况。在本文中,我们开启了一项早期尝试,通过优化多分类AUC指标来考虑学习多分类评分函数的问题。我们的基础是基于M指标,它是AUC的一种知名多分类扩展。我们首先重新审视了这个指标,表明它可以消除少数类对中的不平衡问题。受此启发,我们提出了一个经验替代风险最小化框架来近似优化M指标。从理论上讲,我们证明了:(i)优化大多数流行的可微替代损失足以渐近地达到贝叶斯最优评分函数;(ii)训练框架具有不平衡感知泛化误差界,与传统的O(√{1/N})结果相比,它更关注少数类的瓶颈样本。在实践中,为了处理计算操作的低可扩展性,我们针对三种流行的替代损失函数提出了加速方法,包括指数损失、平方损失和铰链损失,以加快损失和梯度评估。最后,在11个真实世界数据集上的实验结果证明了我们提出的框架的有效性。代码现已在https://github.com/joshuaas/Learning-with-Multiclass-AUC-Theory-and-Algorithms上获取。