Zhao Peng, Li Yi, Tang Baowei, Liu Huiting, Yao Sheng
Key Laboratory of Intelligent Computing and Signal Processing of Ministry of Education, Anhui University, Hefei 230601, China; School of Computer Science and Technology, Anhui University, Hefei 230601, China.
Key Laboratory of Intelligent Computing and Signal Processing of Ministry of Education, Anhui University, Hefei 230601, China; School of Computer Science and Technology, Anhui University, Hefei 230601, China.
Neural Netw. 2023 Apr;161:306-317. doi: 10.1016/j.neunet.2023.01.050. Epub 2023 Feb 4.
In fine-grained image classification, there are only very subtle differences between classes. It is challenging to learn local discriminative features and remove distractive features in fine-grained image classification. Existing fine-grained image classification methods learn discriminative feature mainly via manual part annotation or attention mechanisms. However, due to the large intraclass variance and interclass similarity, the discriminative information and distractive information still are not distinguished effectively. To address this problem, we propose a feature relocation network (FRe-Net) which takes advantage of the different natures of features learned from different stages of the network. Our network consists of a distractive feature learning module and a relocated high-level feature learning module. In the distractive feature learning module, we propose to exploit the difference between low-level features and high-level features to design a distractive loss L, which guides the attention to locate distractive regions more accurately. In the relocated high-level feature learning module, we enhance the representing capacity of the middle-level feature via the attention module and subtract the distractive feature learned from the distractive feature learning module in order to learn more local discriminative features. In end-to-end model training, the distractive feature learning module and the relocated high-level feature learning module are beneficial to each other via joint optimization. We conducted comprehensive experiments on three benchmark datasets widely used in fine-grained image classification. The experimental results show that FRe-Net achieves state-of-the-art performance, which validates the effectiveness of FRe-Net.
在细粒度图像分类中,类别之间只有非常细微的差异。在细粒度图像分类中学习局部判别特征并去除干扰特征具有挑战性。现有的细粒度图像分类方法主要通过手动部分标注或注意力机制来学习判别特征。然而,由于类内方差大且类间相似度高,判别信息和干扰信息仍然没有得到有效区分。为了解决这个问题,我们提出了一种特征重定位网络(FRe-Net),它利用从网络不同阶段学习到的特征的不同性质。我们的网络由一个干扰特征学习模块和一个重定位的高级特征学习模块组成。在干扰特征学习模块中,我们建议利用低级特征和高级特征之间的差异来设计一个干扰损失L,它引导注意力更准确地定位干扰区域。在重定位的高级特征学习模块中,我们通过注意力模块增强中级特征的表示能力,并减去从干扰特征学习模块中学到的干扰特征,以便学习更多的局部判别特征。在端到端模型训练中,干扰特征学习模块和重定位的高级特征学习模块通过联合优化相互受益。我们在细粒度图像分类中广泛使用的三个基准数据集上进行了全面实验。实验结果表明,FRe-Net实现了当前最优性能,这验证了FRe-Net的有效性。
Neural Netw. 2023-4
Neural Comput. 2020-9
IEEE Trans Pattern Anal Mach Intell. 2022-2
Front Neurorobot. 2024-5-3
IEEE Trans Image Process. 2021