Yu Zhengyang, Zhao Tongtong, Xi Zuqiang, Zhang Yaofeng, Zhang Xiaodong, Wang Xiaoying
Department of Radiology, Peking University First Hospital, Beijing, China.
Beijing Smart Tree Medical Technology Co., Ltd., Beijing, China.
Med Phys. 2025 Feb;52(2):951-964. doi: 10.1002/mp.17494. Epub 2024 Nov 4.
Developing deep learning models for segmenting medical images in multiple modalities with less data and annotation is an attractive and challenging task, which was previously discussed as being accomplished by complex external frameworks for bridging the gap between different modalities. Exploring the generalization ability of networks in medical images in different modalities could provide more simple and accessible methods, yet comprehensive testing could still be needed.
To explore the feasibility and robustness of using computed tomography (CT) images to assist the segmentation of magnetic resonance (MR) images via the generalization, in the segmentation of renal parenchyma of renal cell carcinoma (RCC) patients.
Nephrographic CT images and fat-suppressed T2-weighted (fs-T2 W) images were retrospectively collected. The pure CT dataset included 116 CT images. Additionally, 240 MR images were randomly divided into subsets A and B. From subset A, three training datasets were constructed, each containing 40, 80, and 120 images, respectively. Similarly, three datasets were constructed from subset B. Subsequently, datasets with mixed modality were created by combining these pure MR datasets with the 116 CT images. The 3D-UNET models for segmenting the renal parenchyma in two steps were trained using these 13 datasets: segmenting kidneys and then the renal parenchyma. These models were evaluated in internal MR (n = 120), CT (n = 65) validation datasets, and an external validation dataset of CT (n = 79), using the mean of the dice similarity coefficient (DSC). To demonstrate the robustness of generalization ability over different proportions of modalities, we compared the models trained with mixed modality in three different proportions and pure MR, using repeated measures analysis of variance (RM-ANOVA). We developed a renal parenchyma volume quantification tool by the trained models. The mean differences and Pearson correlation coefficients between the model segmentation volume and the ground truth segmentation volume were calculated for its evaluation.
The mean DSCs of models trained with 116 data in CT in the validation of MR were 0.826, 0.842, and 0.953, respectively, for the predictions of kidney segmentation model on whole image, renal parenchymal segmentation model on kidneys with RCC and without RCC. For all models trained with mixed modality, the means of DSC were above 0.9, in all validations of CT and MR. According to the results of the comparison between models trained with mixed modality and pure MR, the means of DSC of the former were significantly greater or equal to the latter, at all three different proportions of modalities. The differences of volumes were all significantly lower than one-third of the volumetric quantification error of a previous method, and the Pearson correlation coefficients of volumes were all above 0.96 on kidneys with and without RCC of three validations.
CT images could be used to assist the segmentation of MR images via the generalization, with or without the supervision of MR data. This ability showed acceptable robustness. A tool for accurately measuring renal parenchymal volume on CT and MR images was established.
利用较少的数据和标注来开发用于分割多种模态医学图像的深度学习模型是一项具有吸引力且具有挑战性的任务,此前讨论的是通过复杂的外部框架来弥合不同模态之间的差距以完成该任务。探索不同模态下医学图像中网络的泛化能力可以提供更简单且易于使用的方法,但仍可能需要全面的测试。
通过在肾细胞癌(RCC)患者肾实质分割中进行泛化,探讨使用计算机断层扫描(CT)图像辅助磁共振(MR)图像分割的可行性和稳健性。
回顾性收集肾实质期CT图像和脂肪抑制T2加权(fs-T2 W)图像。纯CT数据集包含116幅CT图像。此外,240幅MR图像被随机分为子集A和B。从子集A中构建了三个训练数据集,每个分别包含40、80和120幅图像。同样,从子集B中构建了三个数据集。随后,通过将这些纯MR数据集与116幅CT图像相结合创建了混合模态数据集。使用这13个数据集训练用于分两步分割肾实质的3D-UNET模型:先分割肾脏,然后分割肾实质。这些模型在内部MR(n = 120)、CT(n = 65)验证数据集以及外部CT(n = 79)验证数据集中进行评估,使用骰子相似系数(DSC)的均值。为了证明在不同模态比例下泛化能力的稳健性,我们使用重复测量方差分析(RM-ANOVA)比较了以三种不同比例的混合模态和纯MR训练的模型。我们通过训练的模型开发了一种肾实质体积量化工具。计算模型分割体积与真实分割体积之间的平均差异和皮尔逊相关系数以进行评估。
在MR验证中,以CT中的116个数据训练的模型,对于全图像肾脏分割模型、有RCC和无RCC肾脏的肾实质分割模型,其DSC均值分别为0.826、0.842和0.953。对于所有以混合模态训练的模型,在CT和MR的所有验证中,DSC均值均高于0.9。根据混合模态训练模型与纯MR训练模型的比较结果,在所有三种不同模态比例下,前者的DSC均值均显著大于或等于后者。体积差异均显著低于先前一种方法体积量化误差的三分之一,并且在三次验证中有RCC和无RCC的肾脏上,体积的皮尔逊相关系数均高于0.96。
无论有无MR数据监督,CT图像均可通过泛化用于辅助MR图像分割。这种能力显示出可接受的稳健性。建立了一种用于在CT和MR图像上准确测量肾实质体积的工具。