Wang Lituan, Zhang Lei, Qi Xiaofeng, Yi Zhang
IEEE Trans Neural Netw Learn Syst. 2022 Aug;33(8):3320-3330. doi: 10.1109/TNNLS.2021.3051721. Epub 2022 Aug 3.
Class imbalance is a common problem in real-world image classification problems, some classes are with abundant data, and the other classes are not. In this case, the representations of classifiers are likely to be biased toward the majority classes and it is challenging to learn proper features, leading to unpromising performance. To eliminate this biased feature representation, many algorithm-level methods learn to pay more attention to the minority classes explicitly according to the prior knowledge of the data distribution. In this article, an attention-based approach called deep attention-based imbalanced image classification (DAIIC) is proposed to automatically pay more attention to the minority classes in a data-driven manner. In the proposed method, an attention network and a novel attention augmented logistic regression function are employed to encapsulate as many features, which belongs to the minority classes, as possible into the discriminative feature learning process by assigning the attention for different classes jointly in both the prediction and feature spaces. With the proposed object function, DAIIC can automatically learn the misclassification costs for different classes. Then, the learned misclassification costs can be used to guide the training process to learn more discriminative features using the designed attention networks. Furthermore, the proposed method is applicable to various types of networks and data sets. Experimental results on both single-label and multilabel imbalanced image classification data sets show that the proposed method has good generalizability and outperforms several state-of-the-art methods for imbalanced image classification.
类别不平衡是现实世界图像分类问题中的一个常见问题,一些类别数据丰富,而其他类别则不然。在这种情况下,分类器的表示可能会偏向多数类,学习合适的特征具有挑战性,导致性能不佳。为了消除这种有偏差的特征表示,许多算法级方法根据数据分布的先验知识,明确地学习更多地关注少数类。在本文中,提出了一种基于注意力的方法,称为深度注意力不平衡图像分类(DAIIC),以数据驱动的方式自动更多地关注少数类。在所提出的方法中,采用了一个注意力网络和一个新颖的注意力增强逻辑回归函数,通过在预测空间和特征空间中联合为不同类别分配注意力,将尽可能多的属于少数类别的特征封装到判别特征学习过程中。利用所提出的目标函数,DAIIC可以自动学习不同类别的误分类成本。然后,所学习的误分类成本可用于指导训练过程,使用设计的注意力网络学习更具判别力的特征。此外,所提出的方法适用于各种类型的网络和数据集。在单标签和多标签不平衡图像分类数据集上的实验结果表明,所提出的方法具有良好的通用性,并且在不平衡图像分类方面优于几种当前的先进方法。