Ibragimov Bulat, Xing Lei
Department of Radiation Oncology, Stanford University School of Medicine, Stanford, California, 94305, USA.
Med Phys. 2017 Feb;44(2):547-557. doi: 10.1002/mp.12045.
Accurate segmentation of organs-at-risks (OARs) is the key step for efficient planning of radiation therapy for head and neck (HaN) cancer treatment. In the work, we proposed the first deep learning-based algorithm, for segmentation of OARs in HaN CT images, and compared its performance against state-of-the-art automated segmentation algorithms, commercial software, and interobserver variability.
Convolutional neural networks (CNNs)-a concept from the field of deep learning-were used to study consistent intensity patterns of OARs from training CT images and to segment the OAR in a previously unseen test CT image. For CNN training, we extracted a representative number of positive intensity patches around voxels that belong to the OAR of interest in training CT images, and negative intensity patches around voxels that belong to the surrounding structures. These patches then passed through a sequence of CNN layers that captured local image features such as corners, end-points, and edges, and combined them into more complex high-order features that can efficiently describe the OAR. The trained network was applied to classify voxels in a region of interest in the test image where the corresponding OAR is expected to be located. We then smoothed the obtained classification results by using Markov random fields algorithm. We finally extracted the largest connected component of the smoothed voxels classified as the OAR by CNN, performed dilate-erode operations to remove cavities of the component, which resulted in segmentation of the OAR in the test image.
The performance of CNNs was validated on segmentation of spinal cord, mandible, parotid glands, submandibular glands, larynx, pharynx, eye globes, optic nerves, and optic chiasm using 50 CT images. The obtained segmentation results varied from 37.4% Dice coefficient (DSC) for chiasm to 89.5% DSC for mandible. We also analyzed the performance of state-of-the-art algorithms and commercial software reported in the literature, and observed that CNNs demonstrate similar or superior performance on segmentation of spinal cord, mandible, parotid glands, larynx, pharynx, eye globes, and optic nerves, but inferior performance on segmentation of submandibular glands and optic chiasm.
We concluded that convolution neural networks can accurately segment most of OARs using a representative database of 50 HaN CT images. At the same time, inclusion of additional information, for example, MR images, may be beneficial to some OARs with poorly visible boundaries.
对头颈部(HaN)癌进行放射治疗的有效计划,关键步骤是准确分割危及器官(OARs)。在本研究中,我们提出了首个基于深度学习的算法,用于分割HaN CT图像中的OARs,并将其性能与最先进的自动分割算法、商业软件以及观察者间的变异性进行比较。
卷积神经网络(CNNs)——深度学习领域的一个概念——被用于研究训练CT图像中OARs的一致强度模式,并在之前未见过的测试CT图像中分割OARs。对于CNN训练,我们在训练CT图像中属于感兴趣OAR的体素周围提取了代表性数量的正强度补丁,以及在属于周围结构的体素周围提取了负强度补丁。这些补丁随后通过一系列CNN层,这些层捕获局部图像特征,如角点、端点和边缘,并将它们组合成更复杂的高阶特征,从而能够有效地描述OAR。将训练好的网络应用于对测试图像中预期相应OAR所在的感兴趣区域中的体素进行分类。然后,我们使用马尔可夫随机场算法对获得的分类结果进行平滑处理。最后,我们提取被CNN分类为OAR的平滑体素的最大连通分量,执行膨胀-腐蚀操作以去除该分量的空洞,从而得到测试图像中OAR的分割结果。
使用50幅CT图像对脊髓、下颌骨、腮腺、颌下腺、喉、咽、眼球、视神经和视交叉的分割验证了CNNs的性能。获得的分割结果从视交叉的37.4%的骰子系数(DSC)到下颌骨的89.5%的DSC不等。我们还分析了文献中报道的最先进算法和商业软件的性能,观察到CNNs在脊髓、下颌骨、腮腺、喉、咽、眼球和视神经的分割上表现出相似或更好的性能,但在颌下腺和视交叉的分割上表现较差。
我们得出结论,卷积神经网络可以使用50幅HaN CT图像的代表性数据库准确分割大多数OARs。同时,纳入额外信息,例如MR图像,可能对一些边界可见性差的OARs有益。