School of Information and Artificial Intelligence, Yunnan University, Kunming, 650504, China.
School of Information and Artificial Intelligence, Yunnan University, Kunming, 650504, China.
Comput Biol Med. 2023 Jun;159:106923. doi: 10.1016/j.compbiomed.2023.106923. Epub 2023 Apr 14.
The main purpose of multimodal medical image fusion is to aggregate the significant information from different modalities and obtain an informative image, which provides comprehensive content and may help to boost other image processing tasks. Many existing methods based on deep learning neglect the extraction and retention of multi-scale features of medical images and the construction of long-distance relationships between depth feature blocks. Therefore, a robust multimodal medical image fusion network via the multi-receptive-field and multi-scale feature (MFNet) is proposed to achieve the purpose of preserving detailed textures and highlighting the structural characteristics. Specifically, the dual-branch dense hybrid dilated convolution blocks (DHDCB) is proposed to extract the depth features from multi-modalities by expanding the receptive field of the convolution kernel as well as reusing features, and establish long-range dependencies. In order to make full use of the semantic features of the source images, the depth features are decomposed into multi-scale domain by combining the 2-D scale function and wavelet function. Subsequently, the down-sampling depth features are fused by the proposed attention-aware fusion strategy and inversed to the feature space with equal size of source images. Ultimately, the fusion result is reconstructed by a deconvolution block. To force the fusion network balancing information preservation, a local standard deviation-driven structural similarity is proposed as the loss function. Extensive experiments prove that the performance of the proposed fusion network outperforms six state-of-the-art methods, which SD, MI, Q and Q are about 12.8%, 4.1%, 8.5% and 9.7% gains, respectively.
多模态医学图像融合的主要目的是聚合来自不同模态的重要信息,并获得具有丰富信息的图像,该图像提供全面的内容,可能有助于提升其他图像处理任务。许多现有的基于深度学习的方法忽略了医学图像的多尺度特征的提取和保留,以及深度特征块之间的长程关系的构建。因此,提出了一种基于多感受野和多尺度特征的稳健多模态医学图像融合网络(MFNet),以达到保留详细纹理和突出结构特征的目的。具体来说,提出了双分支密集混合扩张卷积块(DHDCB),通过扩展卷积核的感受野和重用特征来从多模态中提取深度特征,并建立长距离依赖关系。为了充分利用源图像的语义特征,将深度特征通过结合 2-D 尺度函数和小波函数分解到多尺度域中。然后,通过所提出的注意感知融合策略融合下采样的深度特征,并将其逆变换到与源图像具有相同大小的特征空间。最终,通过反卷积块重构融合结果。为了迫使融合网络平衡信息保留,提出了基于局部标准差的结构相似性作为损失函数。广泛的实验证明,所提出的融合网络的性能优于六种最先进的方法,在 SD、MI、Q 和 Q 方面分别提高了约 12.8%、4.1%、8.5%和 9.7%。