• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于细粒度视觉分类的显著特征抑制与跨特征融合网络

Significant feature suppression and cross-feature fusion networks for fine-grained visual classification.

作者信息

Yang Shengying, Yang Xinqi, Wu Jianfeng, Feng Boyang

机构信息

Zhejiang University of Science and Technology, Hangzhou, 310023, China.

Zhejiang Shuren University, Hangzhou, 310023, China.

出版信息

Sci Rep. 2024 Oct 14;14(1):24051. doi: 10.1038/s41598-024-74654-4.

DOI:10.1038/s41598-024-74654-4
PMID:39402140
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11473661/
Abstract

The technique of extracting different distinguishing features by locating different part regions to achieve fine-grained visual classification (FGVC) has made significant improvements. Utilizing attention mechanisms for feature extraction has become one of the mainstream methods in computer vision, but these methods have certain limitations. They typically focus on the most discriminative regions and directly combine the features of these parts, neglecting other less prominent yet still discriminative regions. Additionally, these methods may not fully explore the intrinsic connections between higher-order and lower-order features to optimize model classification performance. By considering the potential relationships between different higher-order feature representations in the object image, we can enable the integrated higher-order features to contribute more significantly to the model's classification decision-making capabilities. To this end, we propose a saliency feature suppression and cross-feature fusion network model (SFSCF-Net) to explore the interaction learning between different higher-order feature representations. These include (1) an object-level image generator (OIG): the intersection of the output feature maps of the last two convolutional blocks of the backbone network is used as an object mask and mapped to the original image for cropping to obtain an object-level image, which can effectively reduce the interference caused by complex backgrounds. (2) A saliency feature suppression module (SFSM): the most distinguishing part of the object image is obtained by a feature extractor, and the part is masked by a two-dimensional suppression method, which improves the accuracy of feature suppression. (3) A cross-feature fusion method (CFM) based on inter-layer interaction: the output feature maps of different network layers are interactively integrated to obtain high-dimensional features, and then the high-dimensional features are channel compressed to obtain the inter-layer interaction feature representation, which enriches the output feature semantic information. The proposed SFSCF-Net can be trained end-to-end and achieves state-of-the-art or competitive results on four FGVC benchmark datasets.

摘要

通过定位不同的局部区域来提取不同的显著特征以实现细粒度视觉分类(FGVC)的技术已经取得了显著进展。利用注意力机制进行特征提取已成为计算机视觉中的主流方法之一,但这些方法存在一定局限性。它们通常聚焦于最具判别力的区域,并直接组合这些部分的特征,而忽略了其他不太突出但仍具有判别力的区域。此外,这些方法可能无法充分探索高阶特征与低阶特征之间的内在联系以优化模型分类性能。通过考虑目标图像中不同高阶特征表示之间的潜在关系,我们可以使集成的高阶特征对模型的分类决策能力做出更显著的贡献。为此,我们提出了一种显著特征抑制与交叉特征融合网络模型(SFSCF-Net)来探索不同高阶特征表示之间的交互学习。这些包括:(1)一个对象级图像生成器(OIG):将骨干网络最后两个卷积块的输出特征图的交集用作对象掩码,并映射到原始图像进行裁剪以获得对象级图像,这可以有效减少复杂背景造成的干扰。(2)一个显著特征抑制模块(SFSM):通过特征提取器获取目标图像中最具区分性的部分,并采用二维抑制方法对该部分进行掩码处理,提高了特征抑制的准确性。(3)一种基于层间交互的交叉特征融合方法(CFM):对不同网络层的输出特征图进行交互集成以获得高维特征,然后对高维特征进行通道压缩以获得层间交互特征表示,丰富了输出特征的语义信息。所提出的SFSCF-Net可以进行端到端训练,并在四个FGVC基准数据集上取得了领先或具有竞争力的结果。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/b2c2f789d5c5/41598_2024_74654_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/7c597945dd7f/41598_2024_74654_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/a4413c6e4a29/41598_2024_74654_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/25363857b4a6/41598_2024_74654_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/496642184f2f/41598_2024_74654_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/b2c2f789d5c5/41598_2024_74654_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/7c597945dd7f/41598_2024_74654_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/a4413c6e4a29/41598_2024_74654_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/25363857b4a6/41598_2024_74654_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/496642184f2f/41598_2024_74654_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6271/11473661/b2c2f789d5c5/41598_2024_74654_Fig5_HTML.jpg

相似文献

1
Significant feature suppression and cross-feature fusion networks for fine-grained visual classification.用于细粒度视觉分类的显著特征抑制与跨特征融合网络
Sci Rep. 2024 Oct 14;14(1):24051. doi: 10.1038/s41598-024-74654-4.
2
Feature relocation network for fine-grained image classification.用于细粒度图像分类的特征重定位网络。
Neural Netw. 2023 Apr;161:306-317. doi: 10.1016/j.neunet.2023.01.050. Epub 2023 Feb 4.
3
SIM-OFE: Structure Information Mining and Object-Aware Feature Enhancement for Fine-Grained Visual Categorization.SIM-OFE:用于细粒度视觉分类的结构信息挖掘与目标感知特征增强
IEEE Trans Image Process. 2024;33:5312-5326. doi: 10.1109/TIP.2024.3459788. Epub 2024 Sep 27.
4
Cross-Part Learning for Fine-Grained Image Classification.用于细粒度图像分类的跨部分学习
IEEE Trans Image Process. 2022;31:748-758. doi: 10.1109/TIP.2021.3135477. Epub 2021 Dec 28.
5
Centralized contrastive loss with weakly supervised progressive feature extraction for fine-grained common thorax disease retrieval in chest x-ray.基于集中对比损失和弱监督渐进式特征提取的胸部 X 射线细粒度常见胸部疾病检索方法。
Med Phys. 2023 Jun;50(6):3560-3572. doi: 10.1002/mp.16144. Epub 2023 Jan 11.
6
Transformer guided self-adaptive network for multi-scale skin lesion image segmentation.Transformer 引导的自适网络用于多尺度皮肤病变图像分割。
Comput Biol Med. 2024 Feb;169:107846. doi: 10.1016/j.compbiomed.2023.107846. Epub 2023 Dec 23.
7
SR-GNN: Spatial Relation-aware Graph Neural Network for Fine-Grained Image Categorization.SR-GNN:用于细粒度图像分类的空间关系感知图神经网络
IEEE Trans Image Process. 2022 Sep 14;PP. doi: 10.1109/TIP.2022.3205215.
8
Image local structure information learning for fine-grained visual classification.细粒度视觉分类中的图像局部结构信息学习。
Sci Rep. 2022 Nov 10;12(1):19205. doi: 10.1038/s41598-022-23835-0.
9
Fine-grained image classification method based on hybrid attention module.基于混合注意力模块的细粒度图像分类方法。
Front Neurorobot. 2024 May 3;18:1391791. doi: 10.3389/fnbot.2024.1391791. eCollection 2024.
10
P-CNN: Part-Based Convolutional Neural Networks for Fine-Grained Visual Categorization.P-CNN:基于部分的卷积神经网络用于细粒度视觉分类。
IEEE Trans Pattern Anal Mach Intell. 2022 Feb;44(2):579-590. doi: 10.1109/TPAMI.2019.2933510. Epub 2022 Jan 7.

引用本文的文献

1
Enhancing occluded and standard bird object recognition using fuzzy-based ensembled computer vision approach with convolutional neural network.使用基于模糊的集成计算机视觉方法与卷积神经网络增强遮挡和标准鸟类目标识别。
Sci Rep. 2025 Jul 1;15(1):22247. doi: 10.1038/s41598-025-05465-4.

本文引用的文献

1
Video Captioning Using Global-Local Representation.使用全局-局部表示的视频字幕
IEEE Trans Circuits Syst Video Technol. 2022 Oct;32(10):6642-6656. doi: 10.1109/tcsvt.2022.3177320. Epub 2022 May 23.
2
Tripartite Feature Enhanced Pyramid Network for Dense Prediction.用于密集预测的三方特征增强金字塔网络
IEEE Trans Image Process. 2023;32:2678-2692. doi: 10.1109/TIP.2023.3272826. Epub 2023 May 16.
3
Convolutional Fine-Grained Classification With Self-Supervised Target Relation Regularization.
IEEE Trans Image Process. 2022;31:5570-5584. doi: 10.1109/TIP.2022.3197931. Epub 2022 Aug 30.
4
The Devil is in the Channels: Mutual-Channel Loss for Fine-Grained Image Classification.问题出在通道上:用于细粒度图像分类的互通道损失
IEEE Trans Image Process. 2020 Feb 20. doi: 10.1109/TIP.2020.2973812.
5
Fine-tuning Convolutional Neural Networks for Biomedical Image Analysis: Actively and Incrementally.用于生物医学图像分析的卷积神经网络微调:主动式与增量式
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2017 Jul;2017:4761-4772. doi: 10.1109/CVPR.2017.506. Epub 2017 Nov 9.