Department of Electrical Engineering, City University of Hong Kong, Hong Kong SAR, China.
Department of Physics & Computer Science, Wilfrid Laurier University, Waterloo, Ontario, Canada.
Med Phys. 2024 Oct;51(10):7240-7256. doi: 10.1002/mp.17291. Epub 2024 Jul 15.
Vessel-wall volume and localized three-dimensional ultrasound (3DUS) metrics are sensitive to the change of carotid atherosclerosis in response to medical/dietary interventions. Manual segmentation of the media-adventitia boundary (MAB) and lumen-intima boundary (LIB) required to obtain these metrics is time-consuming and prone to observer variability. Although supervised deep-learning segmentation models have been proposed, training of these models requires a sizeable manually segmented training set, making larger clinical studies prohibitive.
We aim to develop a method to optimize pre-trained segmentation models without requiring manual segmentation to supervise the fine-tuning process.
We developed an adversarial framework called the unsupervised shape-and-texture generative adversarial network (USTGAN) to fine-tune a convolutional neural network (CNN) pre-trained on a source dataset for accurate segmentation of a target dataset. The network integrates a novel texture-based discriminator with a shape-based discriminator, which together provide feedback for the CNN to segment the target images in a similar way as the source images. The texture-based discriminator increases the accuracy of the CNN in locating the artery, thereby lowering the number of failed segmentations. Failed segmentation was further reduced by a self-checking mechanism to flag longitudinal discontinuity of the artery and by self-correction strategies involving surface interpolation followed by a case-specific tuning of the CNN. The U-Net was pre-trained by the source dataset involving 224 3DUS volumes with 136, 44, and 44 volumes in the training, validation and testing sets. The training of USTGAN involved the same training group of 136 volumes in the source dataset and 533 volumes in the target dataset. No segmented boundaries for the target cohort were available for training USTGAN. The validation and testing of USTGAN involved 118 and 104 volumes from the target cohort, respectively. The segmentation accuracy was quantified by Dice Similarity Coefficient (DSC), and incorrect localization rate (ILR). Tukey's Honestly Significant Difference multiple comparison test was employed to quantify the difference of DSCs between models and settings, where was considered statistically significant.
USTGAN attained a DSC of % in LIB and % in MAB, improving from the baseline performance of % in LIB (p ) and % in MAB (p ). Our approach outperformed six state-of-the-art domain-adaptation models (MAB: , LIB: ). The proposed USTGAN also had the lowest ILR among the methods compared (LIB: 2.5%, MAB: 1.7%).
Our framework improves segmentation generalizability, thereby facilitating efficient carotid disease monitoring in multicenter trials and in clinics with less expertise in 3DUS imaging.
血管壁体积和局部三维超声(3DUS)指标对颈动脉粥样硬化在医学/饮食干预下的变化很敏感。获得这些指标所需的血管中膜边界(MAB)和管腔内膜边界(LIB)的手动分割既耗时又容易受到观察者变异性的影响。尽管已经提出了监督深度学习分割模型,但这些模型的训练需要一个相当大的手动分割训练集,使得更大的临床研究变得不可行。
我们旨在开发一种无需手动分割即可优化预训练分割模型的方法,以指导精细调整过程。
我们开发了一种对抗性框架,称为无监督形状和纹理生成对抗网络(USTGAN),用于微调在源数据集上预训练的卷积神经网络(CNN),以准确分割目标数据集。该网络集成了一种新颖的基于纹理的鉴别器和基于形状的鉴别器,它们共同为 CNN 提供反馈,以便以类似于源图像的方式分割目标图像。基于纹理的鉴别器提高了 CNN 定位动脉的准确性,从而减少了分割失败的次数。通过标记动脉的纵向不连续性和涉及表面插值的自我校正策略,进一步减少了分割失败的次数,然后针对特定病例调整 CNN。U-Net 由涉及 224 个 3DUS 体积的源数据集进行预训练,其中训练、验证和测试组分别包含 136、44 和 44 个体积。USTGAN 的训练涉及源数据集的相同训练组(136 个体积)和目标数据集的 533 个体积。目标队列没有分割边界用于训练 USTGAN。对 USTGAN 的验证和测试分别涉及来自目标队列的 118 个和 104 个体积。通过 Dice 相似系数(DSC)和不正确定位率(ILR)来量化分割准确性。采用 Tukey Honestly Significant Difference 多重比较检验来量化模型和设置之间 DSCs 的差异,其中 被认为具有统计学意义。
USTGAN 在 LIB 中达到了 %的 DSC 和 %的 MAB,优于基线性能(LIB:%,MAB:%)。与六种最先进的域自适应模型相比(MAB:%,LIB:%),我们的方法表现更好。与比较的方法相比,所提出的 USTGAN 还具有最低的 ILR(LIB:2.5%,MAB:1.7%)。
我们的框架提高了分割的可泛化性,从而促进了在多中心试验和 3DUS 成像专业知识较少的诊所中进行颈动脉疾病的有效监测。