Liu Cong, Zhang Xiaofei, Si Wen, Ni Xinye
Faculty of Business Information, Shanghai Business School, Shanghai 200235, China.
The Affiliated Changzhou No. 2 People's Hospital of Nanjing Medical University, Changzhou 213003, China.
Evid Based Complement Alternat Med. 2021 Mar 5;2021:8894222. doi: 10.1155/2021/8894222. eCollection 2021.
Radiotherapy has become a common treatment option for head and neck (H&N) cancer, and organs at risk (OARs) need to be delineated to implement a high conformal dose distribution. Manual drawing of OARs is time consuming and inaccurate, so automatic drawing based on deep learning models has been proposed to accurately delineate the OARs. However, state-of-the-art performance usually requires a decent amount of delineation, but collecting pixel-level manual delineations is labor intensive and may not be necessary for representation learning. Encouraged by the recent progress in self-supervised learning, this study proposes and evaluates a novel multiview contrastive representation learning to boost the models from unlabelled data. The proposed learning architecture leverages three views of CTs (coronal, sagittal, and transverse plane) to collect positive and negative training samples. Specifically, a CT in 3D is first projected into three 2D views (coronal, sagittal, and transverse planes), then a convolutional neural network takes 3 views as inputs and outputs three individual representations in latent space, and finally, a contrastive loss is used to pull representation of different views of the same image closer ("positive pairs") and push representations of views from different images ("negative pairs") apart. To evaluate performance, we collected 220 CT images in H&N cancer patients. The experiment demonstrates that our method significantly improves quantitative performance over the state-of-the-art (from 83% to 86% in absolute Dice scores). Thus, our method provides a powerful and principled means to deal with the label-scarce problem.
放射治疗已成为头颈部(H&N)癌症的常见治疗选择,需要勾勒出危及器官(OARs)以实现高适形剂量分布。手动绘制OARs既耗时又不准确,因此有人提出基于深度学习模型的自动绘制方法来准确勾勒OARs。然而,最先进的性能通常需要大量的勾勒,但收集像素级的手动勾勒工作量大,而且对于表征学习可能并非必要。受自监督学习近期进展的鼓舞,本研究提出并评估了一种新颖的多视图对比表征学习方法,以从未标记数据中提升模型性能。所提出的学习架构利用CT的三个视图(冠状面、矢状面和横断面)来收集正样本和负样本。具体而言,首先将三维CT投影到三个二维视图(冠状面、矢状面和横断面)中,然后卷积神经网络将这三个视图作为输入,并在潜在空间中输出三个单独的表征,最后,使用对比损失将同一图像不同视图的表征拉近(“正样本对”),并将不同图像视图的表征推开(“负样本对”)。为了评估性能,我们收集了220名头颈部癌症患者的CT图像。实验表明,我们的方法在定量性能上比最先进的方法有显著提高(绝对骰子分数从83%提高到86%)。因此,我们的方法提供了一种强大且有原则的手段来处理标签稀缺问题。