IEEE Trans Med Imaging. 2021 Apr;40(4):1113-1122. doi: 10.1109/TMI.2020.3046444. Epub 2021 Apr 1.
Multi-domain data are widely leveraged in vision applications taking advantage of complementary information from different modalities, e.g., brain tumor segmentation from multi-parametric magnetic resonance imaging (MRI). However, due to possible data corruption and different imaging protocols, the availability of images for each domain could vary amongst multiple data sources in practice, which makes it challenging to build a universal model with a varied set of input data. To tackle this problem, we propose a general approach to complete the random missing domain(s) data in real applications. Specifically, we develop a novel multi-domain image completion method that utilizes a generative adversarial network (GAN) with a representational disentanglement scheme to extract shared content encoding and separate style encoding across multiple domains. We further illustrate that the learned representation in multi-domain image completion could be leveraged for high-level tasks, e.g., segmentation, by introducing a unified framework consisting of image completion and segmentation with a shared content encoder. The experiments demonstrate consistent performance improvement on three datasets for brain tumor segmentation, prostate segmentation, and facial expression image completion respectively.
多领域数据在视觉应用中得到了广泛的应用,利用了来自不同模态的互补信息,例如,从多参数磁共振成像(MRI)中分割脑肿瘤。然而,由于可能的数据损坏和不同的成像协议,在实际中,每个领域的图像可用性可能会因多个数据源而有所不同,这使得使用各种输入数据构建通用模型变得具有挑战性。为了解决这个问题,我们提出了一种通用方法来完成实际应用中随机缺失的域(s)数据。具体来说,我们开发了一种新的多域图像完成方法,该方法利用生成对抗网络(GAN)和表示解缠方案,从多个域中提取共享内容编码和分离的风格编码。我们进一步说明,在多域图像完成中学习到的表示可以通过引入一个包含图像完成和分割的统一框架来利用高级任务,例如分割。实验证明,在脑肿瘤分割、前列腺分割和面部表情图像完成三个数据集上的性能均有一致的提高。