Kim Soopil, Park Heejung, Kang Myeongkyun, Jin Kyong Hwan, Adeli Ehsan, Pohl Kilian M, Park Sang Hyun
Department of Robotics and Mechatronics Engineering, Daegu Gyeongbuk Institute of Science and Technology, Republic of Korea; Department of Psychiatry and Behavioral Sciences, Stanford University, CA 94305, USA.
Department of Robotics and Mechatronics Engineering, Daegu Gyeongbuk Institute of Science and Technology, Republic of Korea.
Med Image Anal. 2024 Jul;95:103156. doi: 10.1016/j.media.2024.103156. Epub 2024 Mar 25.
The state-of-the-art multi-organ CT segmentation relies on deep learning models, which only generalize when trained on large samples of carefully curated data. However, it is challenging to train a single model that can segment all organs and types of tumors since most large datasets are partially labeled or are acquired across multiple institutes that may differ in their acquisitions. A possible solution is Federated learning, which is often used to train models on multi-institutional datasets where the data is not shared across sites. However, predictions of federated learning can be unreliable after the model is locally updated at sites due to 'catastrophic forgetting'. Here, we address this issue by using knowledge distillation (KD) so that the local training is regularized with the knowledge of a global model and pre-trained organ-specific segmentation models. We implement the models in a multi-head U-Net architecture that learns a shared embedding space for different organ segmentation, thereby obtaining multi-organ predictions without repeated processes. We evaluate the proposed method using 8 publicly available abdominal CT datasets of 7 different organs. Of those datasets, 889 CTs were used for training, 233 for internal testing, and 30 volumes for external testing. Experimental results verified that our proposed method substantially outperforms other state-of-the-art methods in terms of accuracy, inference time, and the number of parameters.
最先进的多器官CT分割依赖于深度学习模型,这些模型只有在经过精心策划的数据的大样本上训练时才能泛化。然而,训练一个能够分割所有器官和肿瘤类型的单一模型具有挑战性,因为大多数大型数据集只是部分标注的,或者是在多个采集方式可能不同的机构中获取的。一种可能的解决方案是联邦学习,它通常用于在多机构数据集上训练模型,而数据不会在各站点之间共享。然而,由于“灾难性遗忘”,在各站点对模型进行本地更新后,联邦学习的预测可能不可靠。在这里,我们通过使用知识蒸馏(KD)来解决这个问题,以便利用全局模型和预训练的器官特异性分割模型的知识对本地训练进行正则化。我们在多头U-Net架构中实现这些模型,该架构学习不同器官分割的共享嵌入空间,从而无需重复过程即可获得多器官预测。我们使用7种不同器官的8个公开可用的腹部CT数据集对所提出的方法进行评估。在这些数据集中,889个CT用于训练,233个用于内部测试,30个容积用于外部测试。实验结果证实,我们提出的方法在准确性、推理时间和参数数量方面大大优于其他最先进的方法。