Zhang Zequn, Jiang Yun, Wang Yunnan, Xie Baao, Zhang Wenyao, Li Yuhang, Chen Zhen, Jin Xin, Zeng Wenjun
IEEE Trans Med Imaging. 2025 Apr;44(4):1686-1698. doi: 10.1109/TMI.2024.3525095. Epub 2025 Apr 3.
Unsupervised domain adaptation (UDA) in medical image segmentation aims to improve the generalization of deep models by alleviating domain gaps caused by inconsistency across equipment, imaging protocols, and patient conditions. However, existing UDA works remain insufficiently explored and present great limitations: 1) Exhibit cumbersome designs that prioritize aligning statistical metrics and distributions, which limits the model's flexibility and generalization while also overlooking the potential knowledge embedded in unlabeled data; 2) More applicable in a certain domain, lack the generalization capability to handle diverse shifts encountered in clinical scenarios. To overcome these limitations, we introduce MedCon, a unified framework that leverages general unsupervised contrastive pre-training to establish domain connections, effectively handling diverse domain shifts without tailored adjustments. Specifically, it initially explores a general contrastive pre-training to establish domain connections by leveraging the rich prior knowledge from unlabeled images. Thereafter, the pre-trained backbone is fine-tuned using source-based images to ultimately identify per-pixel semantic categories. To capture both intra- and inter-domain connections of anatomical structures, we construct positive-negative pairs from a hybrid aspect of both local and global scales. In this regard, a shared-weight encoder-decoder is employed to generate pixel-level representations, which are then mapped into hyper-spherical space using a non-learnable projection head to facilitate positive pair matching. Comprehensive experiments on diverse medical image datasets confirm that MedCon outperforms previous methods by effectively managing a wide range of domain shifts and showcasing superior generalization capabilities.
医学图像分割中的无监督域适应(UDA)旨在通过缓解因设备、成像协议和患者状况不一致而导致的域差距,来提高深度模型的泛化能力。然而,现有的UDA工作仍未得到充分探索,存在很大局限性:1)设计繁琐,优先考虑对齐统计指标和分布,这限制了模型的灵活性和泛化能力,同时也忽略了未标记数据中潜在的知识;2)在特定领域更适用,缺乏处理临床场景中遇到的各种变化的泛化能力。为了克服这些局限性,我们引入了MedCon,这是一个统一的框架,利用一般的无监督对比预训练来建立域连接,无需进行定制调整即可有效处理各种域变化。具体来说,它首先通过利用未标记图像中的丰富先验知识,探索一种一般的对比预训练来建立域连接。此后,使用基于源的图像对预训练的主干进行微调,以最终识别每个像素的语义类别。为了捕捉解剖结构的域内和域间连接,我们从局部和全局尺度的混合角度构建正负对。在这方面,采用共享权重的编码器-解码器来生成像素级表示,然后使用不可学习的投影头将其映射到超球面空间,以促进正配对匹配。在各种医学图像数据集上进行的综合实验证实,MedCon通过有效管理广泛的域变化并展示出卓越的泛化能力,优于先前的方法。