Lu Chenggang, Zhang Jianwei, Zhang Dan, Mou Lei, Yuan Jinli, Xia Kewen, Guo Zhitao, Zhang Jiong
IEEE J Biomed Health Inform. 2025 May 22;PP. doi: 10.1109/JBHI.2025.3572301.
Brain tumors are highly lethal and debilitating pathological changes that require timely diagnosis and treatment. Magnetic resonance imaging (MRI), a non-invasive diagnostic tool, provides complementary multi-modal information crucial for accurate tumor detection and delineation. However, existing methods struggle to effectively fuse multi-modal information from MRI sequences and often fail to perform modality-specific feature extraction, which hinders accurate tumor segmentation. Furthermore, the inherent challenges posed by the blurred boundaries and complex morphological characteristics of tumor structures present additional substantial obstacles to achieving precise segmentation. To address these issues, we propose FiHam, a fine-grained hierarchical progressive modal-aware network that introduces a novel multi-modal fusion strategy and an advanced feature extraction mechanism. Specifically, FiHam employs a progressive fusion strategy that extracts modality-specific features at lower levels and integrates multi-modal features at higher levels to effectively leverage complementary information from tumor images. Additionally, we design a gated cross-attention modal-fusion module that adaptively selects and integrates dual-modal features using cross-attention mechanisms to enhance modality fusion. To further refine segmentation accuracy, we incorporate a tiny U-Net into the encoder to capture boundary features and complex tumor morphology. Extensive experiments on three large-scale, multi-modal brain tumor datasets demonstrate that FiHam achieves state-of-the-art performance, delivering significant improvements in segmentation accuracy and generalizability across diverse MRI modalities.
脑肿瘤是极具致死性和致残性的病理变化,需要及时诊断和治疗。磁共振成像(MRI)作为一种非侵入性诊断工具,提供了对准确肿瘤检测和勾勒至关重要的互补多模态信息。然而,现有方法难以有效地融合来自MRI序列的多模态信息,并且常常无法进行特定模态的特征提取,这阻碍了准确的肿瘤分割。此外,肿瘤结构边界模糊和形态特征复杂所带来的固有挑战,给实现精确分割带来了额外的重大障碍。为了解决这些问题,我们提出了FiHam,这是一种细粒度分层渐进模态感知网络,它引入了一种新颖的多模态融合策略和先进的特征提取机制。具体而言,FiHam采用渐进融合策略,在较低层次提取特定模态特征,在较高层次整合多模态特征,以有效利用肿瘤图像中的互补信息。此外,我们设计了一个门控交叉注意力模态融合模块,利用交叉注意力机制自适应地选择和整合双模态特征,以增强模态融合。为了进一步提高分割精度,我们在编码器中并入一个微小U-Net,以捕捉边界特征和复杂的肿瘤形态。在三个大规模多模态脑肿瘤数据集上进行的大量实验表明,FiHam实现了最优性能,在分割精度和跨不同MRI模态的通用性方面有显著提升。