• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

扩散:利用扩散模型实现红外与可见光图像融合中的高色彩保真度

Dif-Fusion: Toward High Color Fidelity in Infrared and Visible Image Fusion With Diffusion Models.

作者信息

Yue Jun, Fang Leyuan, Xia Shaobo, Deng Yue, Ma Jiayi

出版信息

IEEE Trans Image Process. 2023;32:5705-5720. doi: 10.1109/TIP.2023.3322046. Epub 2023 Oct 24.

DOI:10.1109/TIP.2023.3322046
PMID:37843992
Abstract

Color plays an important role in human visual perception, reflecting the spectrum of objects. However, the existing infrared and visible image fusion methods rarely explore how to handle multi-spectral/channel data directly and achieve high color fidelity. This paper addresses the above issue by proposing a novel method with diffusion models, termed as Dif-Fusion, to generate the distribution of the multi-channel input data, which increases the ability of multi-source information aggregation and the fidelity of colors. In specific, instead of converting multi-channel images into single-channel data in existing fusion methods, we create the multi-channel data distribution with a denoising network in a latent space with forward and reverse diffusion process. Then, we use the the denoising network to extract the multi-channel diffusion features with both visible and infrared information. Finally, we feed the multi-channel diffusion features to the multi-channel fusion module to directly generate the three-channel fused image. To retain the texture and intensity information, we propose multi-channel gradient loss and intensity loss. Along with the current evaluation metrics for measuring texture and intensity fidelity, we introduce Delta E as a new evaluation metric to quantify color fidelity. Extensive experiments indicate that our method is more effective than other state-of-the-art image fusion methods, especially in color fidelity. The source code is available at https://github.com/GeoVectorMatrix/Dif-Fusion.

摘要

颜色在人类视觉感知中起着重要作用,反映了物体的光谱。然而,现有的红外与可见光图像融合方法很少探索如何直接处理多光谱/通道数据并实现高色彩保真度。本文通过提出一种使用扩散模型的新方法(称为Dif-Fusion)来解决上述问题,以生成多通道输入数据的分布,这增强了多源信息聚合能力和色彩保真度。具体而言,与现有融合方法将多通道图像转换为单通道数据不同,我们在具有正向和反向扩散过程的潜在空间中使用去噪网络创建多通道数据分布。然后,我们使用去噪网络提取包含可见光和红外信息的多通道扩散特征。最后,我们将多通道扩散特征输入到多通道融合模块中,直接生成三通道融合图像。为了保留纹理和强度信息,我们提出了多通道梯度损失和强度损失。除了当前用于测量纹理和强度保真度的评估指标外,我们引入Delta E作为一种新的评估指标来量化色彩保真度。大量实验表明,我们的方法比其他现有先进图像融合方法更有效,尤其是在色彩保真度方面。源代码可在https://github.com/GeoVectorMatrix/Dif-Fusion获取。

相似文献

1
Dif-Fusion: Toward High Color Fidelity in Infrared and Visible Image Fusion With Diffusion Models.扩散:利用扩散模型实现红外与可见光图像融合中的高色彩保真度
IEEE Trans Image Process. 2023;32:5705-5720. doi: 10.1109/TIP.2023.3322046. Epub 2023 Oct 24.
2
FDNet: An end-to-end fusion decomposition network for infrared and visible images.FDNet:一种用于红外与可见光图像的端到端融合分解网络。
PLoS One. 2023 Sep 18;18(9):e0290231. doi: 10.1371/journal.pone.0290231. eCollection 2023.
3
RADFNet: An infrared and visible image fusion framework based on distributed network.RADFNet:一种基于分布式网络的红外与可见光图像融合框架。
Front Plant Sci. 2023 Jan 24;13:1056711. doi: 10.3389/fpls.2022.1056711. eCollection 2022.
4
SCFusion: Infrared and Visible Fusion Based on Salient Compensation.SCFusion:基于显著补偿的红外与可见光融合
Entropy (Basel). 2023 Jun 27;25(7):985. doi: 10.3390/e25070985.
5
Deep unsupervised endoscopic image enhancement based on multi-image fusion.基于多图像融合的深度无监督内窥镜图像增强。
Comput Methods Programs Biomed. 2022 Jun;221:106800. doi: 10.1016/j.cmpb.2022.106800. Epub 2022 Apr 26.
6
Semantic-Aware Fusion Network Based on Super-Resolution.基于超分辨率的语义感知融合网络
Sensors (Basel). 2024 Jun 5;24(11):3665. doi: 10.3390/s24113665.
7
Different Input Resolutions and Arbitrary Output Resolution: A Meta Learning-Based Deep Framework for Infrared and Visible Image Fusion.不同输入分辨率与任意输出分辨率:一种基于元学习的红外与可见光图像融合深度框架
IEEE Trans Image Process. 2021;30:4070-4083. doi: 10.1109/TIP.2021.3069339. Epub 2021 Apr 7.
8
MJ-GAN: Generative Adversarial Network with Multi-Grained Feature Extraction and Joint Attention Fusion for Infrared and Visible Image Fusion.MJ-GAN:用于红外与可见光图像融合的具有多粒度特征提取和联合注意力融合的生成对抗网络
Sensors (Basel). 2023 Jul 12;23(14):6322. doi: 10.3390/s23146322.
9
Multi-Spectral Fusion and Denoising of Color and Near-Infrared Images Using Multi-Scale Wavelet Analysis.基于多尺度小波分析的彩色与近红外图像的多光谱融合与去噪
Sensors (Basel). 2021 May 22;21(11):3610. doi: 10.3390/s21113610.
10
Fusing Infrared and Visible Images of Different Resolutions via Total Variation Model.基于全变差模型的不同分辨率红外与可见光图像融合。
Sensors (Basel). 2018 Nov 8;18(11):3827. doi: 10.3390/s18113827.

引用本文的文献

1
HC-SPA: Hyperbolic Cosine-Based Symplectic Phase Alignment for Fusion Optimization.HC-SPA:用于融合优化的基于双曲余弦的辛相位对齐
Sensors (Basel). 2025 Aug 13;25(16):5003. doi: 10.3390/s25165003.
2
Robust Infrared-Visible Fusion Imaging with Decoupled Semantic Segmentation Network.基于解耦语义分割网络的稳健红外-可见光融合成像
Sensors (Basel). 2025 Apr 22;25(9):2646. doi: 10.3390/s25092646.
3
A facial structure sampling contrastive learning method for sketch facial synthesis.一种用于草图面部合成的面部结构采样对比学习方法。
Sci Rep. 2025 May 8;15(1):16056. doi: 10.1038/s41598-025-00574-6.
4
Multi-scene image fusion via memory aware synapses.通过记忆感知突触实现多场景图像融合。
Sci Rep. 2025 Apr 24;15(1):14280. doi: 10.1038/s41598-025-88261-4.
5
A dual-stream feature decomposition network with weight transformation for multi-modality image fusion.一种具有权重变换的双流特征分解网络用于多模态图像融合。
Sci Rep. 2025 Mar 3;15(1):7467. doi: 10.1038/s41598-025-92054-0.
6
Diffusion-driven multi-modality medical image fusion.扩散驱动的多模态医学图像融合。
Med Biol Eng Comput. 2025 Feb 11. doi: 10.1007/s11517-025-03300-6.
7
MGFusion: a multimodal large language model-guided information perception for infrared and visible image fusion.MGFusion:一种用于红外与可见光图像融合的多模态大语言模型引导的信息感知方法
Front Neurorobot. 2024 Dec 23;18:1521603. doi: 10.3389/fnbot.2024.1521603. eCollection 2024.
8
Semantic information guided diffusion posterior sampling for remote sensing image fusion.用于遥感图像融合的语义信息引导扩散后验采样
Sci Rep. 2024 Nov 8;14(1):27259. doi: 10.1038/s41598-024-78778-5.
9
MEEAFusion: Multi-Scale Edge Enhancement and Joint Attention Mechanism Based Infrared and Visible Image Fusion.MEEAFusion:基于多尺度边缘增强和联合注意力机制的红外与可见光图像融合
Sensors (Basel). 2024 Sep 9;24(17):5860. doi: 10.3390/s24175860.
10
Semantic-Aware Fusion Network Based on Super-Resolution.基于超分辨率的语义感知融合网络
Sensors (Basel). 2024 Jun 5;24(11):3665. doi: 10.3390/s24113665.