• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于解耦语义分割网络的稳健红外-可见光融合成像

Robust Infrared-Visible Fusion Imaging with Decoupled Semantic Segmentation Network.

作者信息

Zhang Xuhui, Yin Yunpeng, Wang Zhuowei, Wu Heng, Cheng Lianglun, Yang Aimin, Zhao Genping

机构信息

Guangdong Provincial Key Laboratory of Cyber-Physical System, Guangdong University of Technology, Guangzhou 510006, China.

School of Computer, Guangdong University of Technology, Guangzhou 510006, China.

出版信息

Sensors (Basel). 2025 Apr 22;25(9):2646. doi: 10.3390/s25092646.

DOI:10.3390/s25092646
PMID:40363084
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12074422/
Abstract

The fusion of infrared and visible images provides complementary information from both modalities and has been widely used in surveillance, military, and other fields. However, most of the available fusion methods have only been evaluated with subjective metrics of visual quality of the fused images, which are often independent of the following relevant high-level visual tasks. Moreover, as a useful technique especially used in low-light scenarios, the effect of low-light conditions on the fusion result has not been well-addressed yet. To address these challenges, a decoupled and semantic segmentation-driven infrared and visible image fusion network is proposed in this paper, which connects both image fusion and the downstream task to drive the network to be optimized. Firstly, a cross-modality transformer fusion module is designed to learn rich hierarchical feature representations. Secondly, a semantic-driven fusion module is developed to enhance the key features of prominent targets. Thirdly, a weighted fusion strategy is adopted to automatically adjust the fusion weights of different modality features. This effectively merges the thermal characteristics from infrared images and detailed information from visible images. Additionally, we design a refined loss function that employs the decoupling network to constrain the pixel distributions in the fused images and produce more-natural fusion images. To evaluate the robustness and generalization of the proposed method in practical challenge applications, a Maritime Infrared and Visible (MIV) dataset is created and verified for maritime environmental perception, which will be made available soon. The experimental results from both widely used public datasets and the practically collected MIV dataset highlight the notable strengths of the proposed method with the best-ranking quality metrics among its counterparts. Of more importance, the fusion image achieved with the proposed method has over 96% target detection accuracy and a dominant high mAP@[50:95] value that far surpasses all the competitors.

摘要

红外图像与可见光图像的融合提供了来自两种模态的互补信息,并已广泛应用于监控、军事和其他领域。然而,大多数现有的融合方法仅通过融合图像视觉质量的主观指标进行评估,这些指标通常与后续相关的高级视觉任务无关。此外,作为一种特别用于低光照场景的有用技术,低光照条件对融合结果的影响尚未得到很好的解决。为应对这些挑战,本文提出了一种解耦的、语义分割驱动的红外与可见光图像融合网络,该网络将图像融合与下游任务连接起来,以驱动网络进行优化。首先,设计了一个跨模态变压器融合模块来学习丰富的层次特征表示。其次,开发了一个语义驱动融合模块来增强突出目标的关键特征。第三,采用加权融合策略自动调整不同模态特征的融合权重。这有效地融合了红外图像的热特征和可见光图像的详细信息。此外,我们设计了一种改进的损失函数,该函数采用解耦网络来约束融合图像中的像素分布,并生成更自然的融合图像。为了评估所提方法在实际挑战性应用中的鲁棒性和泛化能力,创建了一个用于海上环境感知的海上红外与可见光(MIV)数据集并进行了验证,该数据集即将公开。来自广泛使用的公共数据集和实际采集的MIV数据集的实验结果突出了所提方法的显著优势,在同类方法中具有最佳的质量指标排名。更重要的是,所提方法获得的融合图像具有超过96%的目标检测准确率和占主导地位的高mAP@[50:95]值,远远超过所有竞争对手。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/7f9156797c00/sensors-25-02646-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/934ba63b4640/sensors-25-02646-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/54f1eaefe25e/sensors-25-02646-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/209cb25da0d8/sensors-25-02646-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/2c3bc0e1098c/sensors-25-02646-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/6fca295d2098/sensors-25-02646-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/4cecd28a3848/sensors-25-02646-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/50478f0f764e/sensors-25-02646-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/35c55b51004c/sensors-25-02646-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/aa8acd016f4e/sensors-25-02646-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/dfac59839805/sensors-25-02646-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/b0833db9aedc/sensors-25-02646-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/7f9156797c00/sensors-25-02646-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/934ba63b4640/sensors-25-02646-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/54f1eaefe25e/sensors-25-02646-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/209cb25da0d8/sensors-25-02646-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/2c3bc0e1098c/sensors-25-02646-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/6fca295d2098/sensors-25-02646-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/4cecd28a3848/sensors-25-02646-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/50478f0f764e/sensors-25-02646-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/35c55b51004c/sensors-25-02646-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/aa8acd016f4e/sensors-25-02646-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/dfac59839805/sensors-25-02646-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/b0833db9aedc/sensors-25-02646-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5122/12074422/7f9156797c00/sensors-25-02646-g012.jpg

相似文献

1
Robust Infrared-Visible Fusion Imaging with Decoupled Semantic Segmentation Network.基于解耦语义分割网络的稳健红外-可见光融合成像
Sensors (Basel). 2025 Apr 22;25(9):2646. doi: 10.3390/s25092646.
2
Infrared and Harsh Light Visible Image Fusion Using an Environmental Light Perception Network.使用环境光感知网络的红外与强光可见图像融合
Entropy (Basel). 2024 Aug 16;26(8):696. doi: 10.3390/e26080696.
3
A Generative Adversarial Network for Infrared and Visible Image Fusion Based on Semantic Segmentation.基于语义分割的红外与可见光图像融合生成对抗网络
Entropy (Basel). 2021 Mar 21;23(3):376. doi: 10.3390/e23030376.
4
SIFusion: Lightweight infrared and visible image fusion based on semantic injection.SIFusion:基于语义注入的轻量级红外与可见光图像融合。
PLoS One. 2024 Nov 6;19(11):e0307236. doi: 10.1371/journal.pone.0307236. eCollection 2024.
5
Brain tumor segmentation and detection in MRI using convolutional neural networks and VGG16.使用卷积神经网络和VGG16在磁共振成像(MRI)中进行脑肿瘤分割与检测
Cancer Biomark. 2025 Mar;42(3):18758592241311184. doi: 10.1177/18758592241311184. Epub 2025 Apr 4.
6
Semantic-Aware Fusion Network Based on Super-Resolution.基于超分辨率的语义感知融合网络
Sensors (Basel). 2024 Jun 5;24(11):3665. doi: 10.3390/s24113665.
7
An Infrared and Visible Image Fusion Network Based on Res2Net and Multiscale Transformer.基于Res2Net和多尺度Transformer的红外与可见光图像融合网络
Sensors (Basel). 2025 Jan 28;25(3):791. doi: 10.3390/s25030791.
8
SCFusion: Infrared and Visible Fusion Based on Salient Compensation.SCFusion:基于显著补偿的红外与可见光融合
Entropy (Basel). 2023 Jun 27;25(7):985. doi: 10.3390/e25070985.
9
Nystromformer based cross-modality transformer for visible-infrared person re-identification.基于Nystromformer的跨模态变压器用于可见光-红外行人重识别
Sci Rep. 2025 May 9;15(1):16224. doi: 10.1038/s41598-025-01226-5.
10
Unsupervised end-to-end infrared and visible image fusion network using learnable fusion strategy.基于可学习融合策略的无监督端到端红外与可见光图像融合网络
J Opt Soc Am A Opt Image Sci Vis. 2022 Dec 1;39(12):2257-2270. doi: 10.1364/JOSAA.473908.

本文引用的文献

1
Fully-Connected Transformer for Multi-Source Image Fusion.用于多源图像融合的全连接变压器
IEEE Trans Pattern Anal Mach Intell. 2025 Mar;47(3):2071-2088. doi: 10.1109/TPAMI.2024.3523364.
2
Dif-Fusion: Toward High Color Fidelity in Infrared and Visible Image Fusion With Diffusion Models.扩散:利用扩散模型实现红外与可见光图像融合中的高色彩保真度
IEEE Trans Image Process. 2023;32:5705-5720. doi: 10.1109/TIP.2023.3322046. Epub 2023 Oct 24.
3
TGFuse: An Infrared and Visible Image Fusion Approach Based on Transformer and Generative Adversarial Network.
TGFuse:一种基于Transformer和生成对抗网络的红外与可见光图像融合方法
IEEE Trans Image Process. 2023 May 10;PP. doi: 10.1109/TIP.2023.3273451.
4
LRRNet: A Novel Representation Learning Guided Fusion Network for Infrared and Visible Images.LRRNet:一种用于红外与可见光图像的新型表示学习引导融合网络。
IEEE Trans Pattern Anal Mach Intell. 2023 Sep;45(9):11040-11052. doi: 10.1109/TPAMI.2023.3268209. Epub 2023 Aug 7.
5
Infrared and Visible Image Fusion Technology and Application: A Review.红外与可见光图像融合技术及应用综述。
Sensors (Basel). 2023 Jan 4;23(2):599. doi: 10.3390/s23020599.
6
Infrared and Visible Image Fusion with Significant Target Enhancement.具有显著目标增强功能的红外与可见光图像融合
Entropy (Basel). 2022 Nov 10;24(11):1633. doi: 10.3390/e24111633.
7
An improved Deeplabv3+ semantic segmentation algorithm with multiple loss constraints.一种具有多重损失约束的改进型 Deeplabv3+ 语义分割算法。
PLoS One. 2022 Jan 19;17(1):e0261582. doi: 10.1371/journal.pone.0261582. eCollection 2022.
8
GMNet: Graded-Feature Multilabel-Learning Network for RGB-Thermal Urban Scene Semantic Segmentation.GMNet:用于RGB-热红外城市场景语义分割的分级特征多标签学习网络
IEEE Trans Image Process. 2021;30:7790-7802. doi: 10.1109/TIP.2021.3109518. Epub 2021 Sep 14.
9
U2Fusion: A Unified Unsupervised Image Fusion Network.U2Fusion:一种统一的无监督图像融合网络。
IEEE Trans Pattern Anal Mach Intell. 2022 Jan;44(1):502-518. doi: 10.1109/TPAMI.2020.3012548. Epub 2021 Dec 8.
10
DenseFuse: A Fusion Approach to Infrared and Visible Images.密集融合:一种红外与可见光图像的融合方法。
IEEE Trans Image Process. 2018 Dec 18. doi: 10.1109/TIP.2018.2887342.