School of Electronics Engineering, Vellore Institute of Technology, Vellore, India.
School of Electronics Engineering, VIT-AP University, Amaravathi, India.
PLoS One. 2024 Oct 23;19(10):e0309651. doi: 10.1371/journal.pone.0309651. eCollection 2024.
Multimodal medical image fusion methods, which combine complementary information from many multi-modality medical images, are among the most important and practical approaches in numerous clinical applications. Various conventional image fusion techniques have been developed for multimodality image fusion. Complex procedures for weight map computing, fixed fusion strategy and lack of contextual understanding remain difficult in conventional and machine learning approaches, usually resulting in artefacts that degrade the image quality. This work proposes an efficient hybrid learning model for medical image fusion using pre-trained and non-pre-trained networks i.e. VGG-19 and SNN with stacking ensemble method. The model leveraging the unique capabilities of each architecture, can effectively preserve the detailed information with high visual quality, for numerous combinations of image modalities in image fusion challenges, notably improved contrast, increased resolution, and lower artefacts. Additionally, this ensemble model can be more robust in the fusion of various combinations of source images that are publicly available from Havard-Medical-Image-Fusion Datasets, GitHub. and Kaggle. Our proposed model performance is superior in terms of visual quality and performance metrics to that of the existing fusion methods in literature like PCA+DTCWT, NSCT, DWT, DTCWT+NSCT, GADCT, CNN and VGG-19.
多模态医学图像融合方法结合了来自多种多模态医学图像的互补信息,是众多临床应用中最重要和实用的方法之一。已经开发了各种常规的图像融合技术用于多模态图像融合。在传统和机器学习方法中,权重图计算的复杂过程、固定的融合策略和缺乏上下文理解仍然很困难,通常会导致图像质量下降的伪影。这项工作提出了一种使用预训练和非预训练网络(即 VGG-19 和 SNN)的混合学习模型,用于医学图像融合,该模型采用堆叠集成方法。该模型利用每个架构的独特功能,可以有效地保留详细信息,具有高质量的视觉效果,适用于图像融合挑战中多种模态图像的组合,特别是对比度提高、分辨率增加和伪影降低。此外,该集成模型在融合来自 Harvard-Medical-Image-Fusion Datasets、GitHub 和 Kaggle 等公共来源的各种组合的源图像时更加稳健。与文献中现有的融合方法(如 PCA+DTCWT、NSCT、DWT、DTCWT+NSCT、GADCT、CNN 和 VGG-19)相比,我们提出的模型在视觉质量和性能指标方面表现更优。