Department of Medical Physics, Memorial Sloan-Kettering Cancer Center, New York, NY 10065, United States of America.
Phys Med Biol. 2020 Oct 7;65(20):205001. doi: 10.1088/1361-6560/ab9fca.
To develop and evaluate a deep learning method to segment parotid glands from MRI using unannotated MRI and unpaired expert-segmented CT datasets. We introduced a new self-derived organ attention deep learning network for combined CT to MRI image-to-image translation (I2I) and MRI segmentation, all trained as an end-to-end network. The expert segmentations available on CT scans were combined with the I2I translated pseudo MR images to train the MRI segmentation network. Once trained, the MRI segmentation network alone is required. We introduced an organ attention discriminator that constrains the CT to MR generator to synthesize pseudo MR images that preserve organ geometry and appearance statistics as in real MRI. The I2I translation network training was regularized using the organ attention discriminator, global image-matching discriminator, and cycle consistency losses. MRI segmentation training was regularized by using cross-entropy loss. Segmentation performance was compared against multiple domain adaptation-based segmentation methods using the Dice similarity coefficient (DSC) and Hausdorff distance at the 95th percentile (HD95). All networks were trained using 85 unlabeled T2-weighted fat suppressed (T2wFS) MRIs and 96 expert-segmented CT scans. Performance upper-limit was based on fully supervised MRI training done using the 85 T2wFS MRI with expert segmentations. Independent evaluation was performed on 77 MRIs never used in training. The proposed approach achieved the highest accuracy (left parotid: DSC 0.82 ± 0.03, HD95 2.98 ± 1.01 mm; right parotid: 0.81 ± 0.05, HD95 3.14 ± 1.17 mm) compared to other methods. This accuracy was close to the reference fully supervised MRI segmentation (DSC of 0.84 ± 0.04, a HD95 of 2.24 ± 0.77 mm for the left parotid, and a DSC of 0.84 ± 0.06 and HD95 of 2.32 ± 1.37 mm for the right parotid glands).
为了开发和评估一种使用未注释的 MRI 和未配对的专家分割 CT 数据集从 MRI 中分割腮腺的深度学习方法。我们引入了一种新的自衍生器官注意深度学习网络,用于 CT 到 MRI 的图像到图像转换 (I2I) 和 MRI 分割,所有这些都作为端到端网络进行训练。在 CT 扫描上获得的专家分割与 I2I 转换的伪 MR 图像相结合,用于训练 MRI 分割网络。一旦训练完成,仅需要 MRI 分割网络。我们引入了一个器官注意鉴别器,该鉴别器约束 CT 到 MR 生成器合成伪 MR 图像,以保留真实 MRI 中的器官几何形状和外观统计信息。I2I 转换网络的训练受到器官注意鉴别器、全局图像匹配鉴别器和循环一致性损失的约束。MRI 分割训练受到交叉熵损失的约束。使用 Dice 相似系数 (DSC) 和第 95 个百分位数的 Hausdorff 距离 (HD95) 比较了与多个基于域适应的分割方法的分割性能。所有网络都使用 85 个未标记的 T2 加权脂肪抑制 (T2wFS) MRI 和 96 个专家分割 CT 扫描进行训练。性能上限基于使用 85 个 T2wFS MRI 和专家分割进行的完全监督 MRI 训练。在从未用于训练的 77 个 MRI 上进行了独立评估。与其他方法相比,所提出的方法达到了最高的准确性(左腮腺:DSC 0.82 ± 0.03,HD95 2.98 ± 1.01mm;右腮腺:0.81 ± 0.05,HD95 3.14 ± 1.17mm)。该准确性接近参考的完全监督 MRI 分割(左腮腺的 DSC 为 0.84 ± 0.04,HD95 为 2.24 ± 0.77mm,右腮腺的 DSC 为 0.84 ± 0.06,HD95 为 2.32 ± 1.37mm)。