IEEE Trans Med Imaging. 2023 Sep;42(9):2566-2576. doi: 10.1109/TMI.2023.3260169. Epub 2023 Aug 31.
As a pragmatic data augmentation tool, data synthesis has generally returned dividends in performance for deep learning based medical image analysis. However, generating corresponding segmentation masks for synthetic medical images is laborious and subjective. To obtain paired synthetic medical images and segmentations, conditional generative models that use segmentation masks as synthesis conditions were proposed. However, these segmentation mask-conditioned generative models still relied on large, varied, and labeled training datasets, and they could only provide limited constraints on human anatomical structures, leading to unrealistic image features. Moreover, the invariant pixel-level conditions could reduce the variety of synthetic lesions and thus reduce the efficacy of data augmentation. To address these issues, in this work, we propose a novel strategy for medical image synthesis, namely Unsupervised Mask (UM)-guided synthesis, to obtain both synthetic images and segmentations using limited manual segmentation labels. We first develop a superpixel based algorithm to generate unsupervised structural guidance and then design a conditional generative model to synthesize images and annotations simultaneously from those unsupervised masks in a semi-supervised multi-task setting. In addition, we devise a multi-scale multi-task Fréchet Inception Distance (MM-FID) and multi-scale multi-task standard deviation (MM-STD) to harness both fidelity and variety evaluations of synthetic CT images. With multiple analyses on different scales, we could produce stable image quality measurements with high reproducibility. Compared with the segmentation mask guided synthesis, our UM-guided synthesis provided high-quality synthetic images with significantly higher fidelity, variety, and utility ( by Wilcoxon Signed Ranked test).
作为一种实用的数据增强工具,数据合成在基于深度学习的医学图像分析的性能方面通常都有回报。然而,为合成医学图像生成相应的分割掩模既费力又主观。为了获得成对的合成医学图像和分割,提出了使用分割掩模作为合成条件的条件生成模型。然而,这些分割掩模条件生成模型仍然依赖于大型、多样化和标记的训练数据集,并且只能对人体解剖结构提供有限的约束,导致图像特征不真实。此外,不变的像素级条件会降低合成病变的多样性,从而降低数据增强的效果。为了解决这些问题,在这项工作中,我们提出了一种新的医学图像合成策略,即无监督掩模(UM)引导合成,使用有限的手动分割标签同时获得合成图像和分割。我们首先开发了一种基于超像素的算法来生成无监督的结构指导,然后设计了一个条件生成模型,在半监督的多任务设置中,从这些无监督的掩模中同时合成图像和注释。此外,我们设计了多尺度多任务 Fréchet inception 距离(MM-FID)和多尺度多任务标准偏差(MM-STD)来利用合成 CT 图像的保真度和多样性评估。通过在不同尺度上的多次分析,我们可以生成具有高可重复性的稳定图像质量测量值。与分割掩模引导合成相比,我们的 UM 引导合成提供了高质量的合成图像,具有更高的保真度、多样性和实用性(通过 Wilcoxon 符号秩检验)。