Zhang Huixian, Li Hailong, Ali Redha, Jia Wei, Pan Wen, Reeder Scott B, Harris David, Masch William, Aslam Anum, Shanbhogue Krishna, Parikh Nehal A, Dillman Jonathan R, He Lili
Imaging Research Center, Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, OH, USA.
Department of Computer Science, University of Cincinnati, Cincinnati, OH, USA.
J Imaging Inform Med. 2024 Dec 20. doi: 10.1007/s10278-024-01362-w.
To develop and validate a modality-invariant Swin U-Net Transformer (UNETR) deep learning model for liver and spleen segmentation on abdominal T1-weighted (T1w) or T2-weighted (T2w) MR images from multiple institutions for pediatric and adult patients with known or suspected chronic liver diseases. In this IRB-approved retrospective study, clinical abdominal axial T1w and T2w MR images from pediatric and adult patients were retrieved from four study sites, including Cincinnati Children's Hospital Medical Center (CCHMC), New York University (NYU), University of Wisconsin (UW) and University of Michigan / Michigan Medicine (UM). The whole liver and spleen were manually delineated as the ground truth masks. We developed a modality-invariant 3D Swin UNETR using a modality-invariant training strategy, in which each patient's T1w and T2w MR images were treated as separate training samples. We conducted both internal and external validation experiments. A total of 241 T1w and 339 T2w MR sequences from 304 patients (age [mean standard deviation], 31.8 20.3 years; 132 [43%] female) were included for model development. The Swin UNETR achieved a Dice similarity coefficient (DSC) of 0.95 ± 0.02 on T1w images and 0.93 ± 0.05 on T2w images for liver segmentation. This is significantly better than the U-Net model (0.90 ± 0.05, p < 0.001 and 0.90 ± 0.13, p < 0.001, respectively). The Swin UNETR achieved a DSC of 0.88 ± 0.12 on T1w images and 0.93 ± 0.10 on T2w images for spleen segmentation, and it significantly outperformed a modality-invariant U-Net model (0.80 ± 0.18, p = 0.001 and 0.88 ± 0.12, p = 0.002, respectively). Our study demonstrated that a modality-invariant Swin UNETR model can segment the liver and spleen on routinely collected clinical bi-parametric abdominal MR images from children and adult patients.
开发并验证一种模态不变的Swin U-Net Transformer(UNETR)深度学习模型,用于对来自多个机构的已知或疑似慢性肝病的儿科和成年患者的腹部T1加权(T1w)或T2加权(T2w)磁共振成像(MRI)进行肝脏和脾脏分割。在这项经机构审查委员会批准的回顾性研究中,从四个研究地点检索了儿科和成年患者的临床腹部轴向T1w和T2w MRI图像,包括辛辛那提儿童医院医疗中心(CCHMC)、纽约大学(NYU)、威斯康星大学(UW)和密歇根大学/密歇根医学中心(UM)。将整个肝脏和脾脏手动勾勒为真实掩膜。我们使用模态不变训练策略开发了一种模态不变的3D Swin UNETR,其中将每个患者的T1w和T2w MRI图像视为单独的训练样本。我们进行了内部和外部验证实验。共有来自304例患者(年龄[均值±标准差],31.8±20.3岁;132例[43%]为女性)的241个T1w和339个T2w MR序列纳入模型开发。Swin UNETR在肝脏分割的T1w图像上的Dice相似系数(DSC)为0.95±0.02,在T2w图像上为0.93±0.05。这显著优于U-Net模型(分别为0.90±0.05,p<0.001和0.90±0.13,p<0.001)。Swin UNETR在脾脏分割的T1w图像上的DSC为0.88±0.12,在T2w图像上为0.93±0.10,并且显著优于模态不变的U-Net模型(分别为0.80±0.18,p = 0.001和0.88±0.12,p = 0.002)。我们的研究表明,一种模态不变的Swin UNETR模型可以对儿童和成年患者常规采集的临床双参数腹部MRI图像上的肝脏和脾脏进行分割。