Yuan Chunbao, Jia Xibin, Wang Luo, Yang Chuanxu
School of Computer Science, Beijing University of Technology, Beijing, China.
Curr Med Imaging. 2025 Jul 30. doi: 10.2174/0115734056361649250717162910.
Magnetic Resonance Imaging (MRI) is a crucial method for clinical diagnosis. Different abdominal MRI sequences provide tissue and structural information from various perspectives, offering reliable evidence for doctors to make accurate diagnoses. In recent years, with the rapid development of intelligent medical imaging, some studies have begun exploring deep learning methods for MRI sequence recognition. However, due to the significant intra-class variations and subtle inter-class differences in MRI sequences, traditional deep learning algorithms still struggle to effectively handle such types of complex distributed data. In addition, the key features for identifying MRI sequence categories often exist in subtle details, while significant discrepancies can be observed among sequences from individual samples. In contrast, current deep learning based MRI sequence classification methods tend to overlook these fine-grained differences across diverse samples.
To overcome the above challenges, this paper proposes a fine-grained prototype network, SequencesNet, for MRI sequence classification. A network combining convolutional neural networks (CNNs) with improved vision transformers is constructed for feature extraction, considering both local and global information. Specifically, a Feature Selection Module (FSM) is added to the visual transformer, and fine-grained features for sequence discrimination are selected based on fused attention weights from multiple layers. Then, a Prototype Classification Module (PCM) is proposed to classify MRI sequences based on fine-grained MRI representations.
Comprehensive experiments are conducted on a public abdominal MRI sequence classification dataset and a private dataset. Our proposed SequencesNet achieved the highest accuracy with 96.73% and 95.98% in two sequence classification datasets, respectively, and outperfom the comparative prototypes and fine-grained models. The visualization results exhibit that our proposed sequencesNet can better capture fine-grained information.
The proposed SequencesNet shows promising performance in MRI sequence classification, excelling in distinguishing subtle inter-class differences and handling large intra-class variability. Specifically, FSM enhances clinical interpretability by focusing on fine-grained features, and PCM improves clustering by optimizing prototype-sample distances. Compared to baselines like 3DResNet18 and TransFG, SequencesNet achieves higher recall and precision, particularly for similar sequences like DCE-LAP and DCE-PVP.
The proposed new MRI sequence classification model, SequencesNet, addresses the problem of subtle inter-class differences and significant intraclass variations existing in medical images. The modular design of SequencesNet can be extended to other medical imaging tasks, including but not limited to multimodal image fusion, lesion detection, and disease staging. Future work can be done to decrease the computational complexity and increase the generalization of the model.
磁共振成像(MRI)是临床诊断的关键方法。不同的腹部MRI序列从不同角度提供组织和结构信息,为医生进行准确诊断提供可靠依据。近年来,随着智能医学成像的快速发展,一些研究开始探索用于MRI序列识别的深度学习方法。然而,由于MRI序列中存在显著的类内变化和细微的类间差异,传统的深度学习算法仍难以有效处理这类复杂的分布式数据。此外,识别MRI序列类别的关键特征往往存在于细微细节中,而各个样本的序列之间可能存在显著差异。相比之下,当前基于深度学习的MRI序列分类方法往往忽略了不同样本之间的这些细粒度差异。
为了克服上述挑战,本文提出了一种用于MRI序列分类的细粒度原型网络SequencesNet。构建了一个将卷积神经网络(CNN)与改进的视觉Transformer相结合的网络用于特征提取,同时考虑局部和全局信息。具体而言,在视觉Transformer中添加了一个特征选择模块(FSM),并基于多层融合注意力权重选择用于序列区分的细粒度特征。然后,提出了一个原型分类模块(PCM),基于细粒度的MRI表示对MRI序列进行分类。
在一个公共腹部MRI序列分类数据集和一个私有数据集上进行了全面实验。我们提出的SequencesNet在两个序列分类数据集中分别达到了96.73%和95.98%的最高准确率,优于对比原型和细粒度模型。可视化结果表明,我们提出的SequencesNet能够更好地捕捉细粒度信息。
所提出的SequencesNet在MRI序列分类中表现出良好的性能,在区分细微的类间差异和处理较大的类内变异性方面表现出色。具体而言,FSM通过关注细粒度特征增强了临床可解释性,而PCM通过优化原型-样本距离改进了聚类。与3DResNet18和TransFG等基线相比,SequencesNet实现了更高的召回率和精确率,特别是对于DCE-LAP和DCE-PVP等相似序列。
所提出的新的MRI序列分类模型SequencesNet解决了医学图像中存在的细微类间差异和显著类内变化的问题。SequencesNet的模块化设计可以扩展到其他医学成像任务,包括但不限于多模态图像融合、病变检测和疾病分期。未来的工作可以致力于降低计算复杂度并提高模型的泛化能力。