Medical Physics Resident, Department of Medical Physics, CancerCare Manitoba, 675 McDermot Avenue Winnipeg, MB R3E 0V9, Canada.
Department of Physics and Astronomy, University of Manitoba, Allen Building, Winnipeg, MB R3T 2N2, Canada.
Phys Med Biol. 2022 Jun 22;67(11). doi: 10.1088/1361-6560/ac530e.
The purpose of this study was to utilize a deep learning model with an advanced inception module to automatically contour critical organs on the computed tomography (CT) scans of head and neck cancer patients who underwent radiation therapy treatment and interpret the clinical suitability of the model results through activation mapping.This study included 25 critical organs that were delineated by expert radiation oncologists. Contoured medical images of 964 patients were sourced from a publicly available TCIA database. The proportion of training, validation, and testing samples for deep learning model development was 65%, 25%, and 10% respectively. The CT scans and segmentation masks were augmented with shift, scale, and rotate transformations. Additionally, medical images were pre-processed using contrast limited adaptive histogram equalization to enhance soft tissue contrast while contours were subjected to morphological operations to ensure their structural integrity. The segmentation model was based on the U-Net architecture with embedded Inception-ResNet-v2 blocks and was trained over 100 epochs with a batch size of 32 and an adaptive learning rate optimizer. The loss function combined the Jaccard Index and binary cross entropy. The model performance was evaluated with Dice Score, Jaccard Index, and Hausdorff Distances. The interpretability of the model was analyzed with guided gradient-weighted class activation mapping.The Dice Score, Jaccard Index, and mean Hausdorff Distance averaged over all structures and patients were 0.82 ± 0.10, 0.71 ± 0.10, and 1.51 ± 1.17 mm respectively on the testing data sets. The Dice Scores for 86.4% of compared structures was within range or better than published interobserver variability derived from multi-institutional studies. The average model training time was 8 h per anatomical structure. The full segmentation of head and neck anatomy by the trained network required only 6.8 s per patient.High accuracy obtained on a large, multi-institutional data set, short segmentation time and clinically-realistic prediction reasoning make the model proposed in this work a feasible solution for head and neck CT scan segmentation in a clinical environment.
本研究旨在利用具有高级 inception 模块的深度学习模型,自动描绘接受放射治疗的头颈部癌症患者 CT 扫描中的关键器官,并通过激活映射来解释模型结果的临床适用性。该研究纳入了 25 个由专业放射肿瘤学家勾画的关键器官。964 名患者的勾画医学图像来源于公共 TCIA 数据库。深度学习模型开发的训练、验证和测试样本的比例分别为 65%、25%和 10%。CT 扫描和分割掩模通过平移、缩放和旋转变换进行扩充。此外,医学图像使用对比度受限自适应直方图均衡化进行预处理,以增强软组织对比度,而轮廓则进行形态学操作,以确保其结构完整性。分割模型基于 U-Net 架构,嵌入了 Inception-ResNet-v2 块,在 100 个 epoch 上进行训练,批量大小为 32,自适应学习率优化器。损失函数结合了 Jaccard 指数和二进制交叉熵。使用 Dice 评分、Jaccard 指数和 Hausdorff 距离评估模型性能。使用引导梯度加权类激活映射分析模型的可解释性。在测试数据集上,所有结构和患者的平均 Dice 评分、Jaccard 指数和平均 Hausdorff 距离分别为 0.82±0.10、0.71±0.10 和 1.51±1.17mm。86.4%的比较结构的 Dice 评分在范围或优于多机构研究得出的文献报道的观察者间变异性。平均每个解剖结构的模型训练时间为 8 小时。经过训练的网络对头颈部解剖结构的完全分割,每位患者仅需 6.8 秒。该模型在大型多机构数据集上获得了高精度,分割时间短,预测推理符合临床实际,是临床环境中头颈部 CT 扫描分割的可行解决方案。