He Xingxin, Zhong Zhun, Fang Leyuan, He Min, Sebe Nicu
IEEE Trans Image Process. 2023;32:309-320. doi: 10.1109/TIP.2022.3228163. Epub 2022 Dec 21.
Accurate retinal fluid segmentation on Optical Coherence Tomography (OCT) images plays an important role in diagnosing and treating various eye diseases. The art deep models have shown promising performance on OCT image segmentation given pixel-wise annotated training data. However, the learned model will achieve poor performance on OCT images that are obtained from different devices (domains) due to the domain shift issue. This problem largely limits the real-world application of OCT image segmentation since the types of devices usually are different in each hospital. In this paper, we study the task of cross-domain OCT fluid segmentation, where we are given a labeled dataset of the source device (domain) and an unlabeled dataset of the target device (domain). The goal is to learn a model that can perform well on the target domain. To solve this problem, in this paper, we propose a novel Structure-guided Cross-Attention Network (SCAN), which leverages the retinal layer structure to facilitate domain alignment. Our SCAN is inspired by the fact that the retinal layer structure is robust to domains and can reflect regions that are important to fluid segmentation. In light of this, we build our SCAN in a multi-task manner by jointly learning the retinal structure prediction and fluid segmentation. To exploit the mutual benefit between layer structure and fluid segmentation, we further introduce a cross-attention module to measure the correlation between the layer-specific feature and the fluid-specific feature encouraging the model to concentrate on highly relative regions during domain alignment. Moreover, an adaptation difficulty map is evaluated based on the retinal structure predictions from different domains, which enforces the model focus on hard regions during structure-aware adversarial learning. Extensive experiments on the three domains of the RETOUCH dataset demonstrate the effectiveness of the proposed method and show that our approach produces state-of-the-art performance on cross-domain OCT fluid segmentation.
光学相干断层扫描(OCT)图像上准确的视网膜液分割在各种眼部疾病的诊断和治疗中起着重要作用。在给定像素级标注训练数据的情况下,先进的深度模型在OCT图像分割方面表现出了良好的性能。然而,由于域偏移问题,所学习的模型在从不同设备(域)获取的OCT图像上性能会很差。这个问题在很大程度上限制了OCT图像分割在现实世界中的应用,因为每个医院的设备类型通常是不同的。在本文中,我们研究跨域OCT液分割任务,其中我们有源设备(域)的标注数据集和目标设备(域)的未标注数据集。目标是学习一个能在目标域上表现良好的模型。为了解决这个问题,在本文中,我们提出了一种新颖的结构引导交叉注意力网络(SCAN),它利用视网膜层结构来促进域对齐。我们的SCAN受到视网膜层结构对域具有鲁棒性且能反映对液分割重要区域这一事实的启发。鉴于此,我们通过联合学习视网膜结构预测和液分割以多任务方式构建我们的SCAN。为了利用层结构和液分割之间的互利关系,我们进一步引入一个交叉注意力模块来测量特定层特征和特定液特征之间的相关性,鼓励模型在域对齐期间专注于高度相关的区域。此外,基于来自不同域的视网膜结构预测评估一个适应难度图,这在结构感知对抗学习期间迫使模型专注于困难区域。在RETOUCH数据集的三个域上进行的大量实验证明了所提方法的有效性,并表明我们的方法在跨域OCT液分割方面产生了当前最优的性能。