Qing Yuan Research Institute, Shanghai Jiao Tong University, Shanghai 200240, China.
Shanghai Artificial Intelligence Laboratory, Shanghai 200232, China.
Med Image Anal. 2024 Oct;97:103289. doi: 10.1016/j.media.2024.103289. Epub 2024 Jul 31.
Large amounts of digitized histopathological data display a promising future for developing pathological foundation models via self-supervised learning methods. Foundation models pretrained with these methods serve as a good basis for downstream tasks. However, the gap between natural and histopathological images hinders the direct application of existing methods. In this work, we present PathoDuet, a series of pretrained models on histopathological images, and a new self-supervised learning framework in histopathology. The framework is featured by a newly-introduced pretext token and later task raisers to explicitly utilize certain relations between images, like multiple magnifications and multiple stains. Based on this, two pretext tasks, cross-scale positioning and cross-stain transferring, are designed to pretrain the model on Hematoxylin and Eosin (H&E) images and transfer the model to immunohistochemistry (IHC) images, respectively. To validate the efficacy of our models, we evaluate the performance over a wide variety of downstream tasks, including patch-level colorectal cancer subtyping and whole slide image (WSI)-level classification in H&E field, together with expression level prediction of IHC marker, tumor identification and slide-level qualitative analysis in IHC field. The experimental results show the superiority of our models over most tasks and the efficacy of proposed pretext tasks. The codes and models are available at https://github.com/openmedlab/PathoDuet.
大量数字化的组织病理学数据为通过自监督学习方法开发病理学基础模型展示了广阔的前景。通过这些方法预先训练的基础模型可以作为下游任务的良好基础。然而,自然图像和组织病理学图像之间的差距阻碍了现有方法的直接应用。在这项工作中,我们提出了 PathoDuet,这是一系列针对组织病理学图像的预先训练模型,以及一种新的组织病理学中的自监督学习框架。该框架的特点是引入了一个新的预设令牌和后期任务提升器,以明确利用图像之间的某些关系,如多种放大倍数和多种染色。基于此,我们设计了两个预设任务,跨尺度定位和跨染色转移,分别用于在苏木精和伊红(H&E)图像上预训练模型,并将模型转移到免疫组织化学(IHC)图像上。为了验证我们模型的有效性,我们评估了其在各种下游任务上的性能,包括在 H&E 领域的斑块级结直肠癌亚型分类和全幻灯片图像(WSI)级分类,以及 IHC 标记物表达水平预测、肿瘤识别和 IHC 领域的幻灯片级定性分析。实验结果表明,我们的模型在大多数任务上都具有优越性,并且所提出的预设任务是有效的。代码和模型可在 https://github.com/openmedlab/PathoDuet 上获取。