Medical Artificial Intelligence and Automation (MAIA) Lab, Department of Radiation Oncology, UT Southwestern Medical Center, Dallas, Texas.
Medical Artificial Intelligence and Automation (MAIA) Lab, Department of Radiation Oncology, UT Southwestern Medical Center, Dallas, Texas.
Int J Radiat Oncol Biol Phys. 2024 Mar 15;118(4):1123-1134. doi: 10.1016/j.ijrobp.2023.10.022. Epub 2023 Nov 7.
A reliable and comprehensive cancer prognosis model for oropharyngeal cancer (OPC) could better assist in personalizing treatment. In this work, we developed a vision transformer-based (ViT-based) multilabel model with multimodal input to learn complementary information from available pretreatment data and predict multiple associated endpoints for radiation therapy for patients with OPC.
A publicly available data set of 512 patients with OPC was used for both model training and evaluation. Planning computed tomography images, primary gross tumor volume masks, and 16 clinical variables representing patient demographics, diagnosis, and treatment were used as inputs. To extract deep image features with global attention, we used a ViT module. Clinical variables were concatenated with the learned image features and fed into fully connected layers to incorporate cross-modality features. To learn the mapping between the features and correlated survival outcomes, including overall survival, local failure-free survival, regional failure-free survival, and distant failure-free survival, we employed 4 multitask logistic regression layers. The proposed model was optimized by combining the multitask logistic regression negative-log likelihood losses of different prediction targets.
We employed the C-index and area under the curve metrics to assess the performance of our model for time-to-event prediction and time-specific binary prediction, respectively. Our proposed model outperformed corresponding single-modality and single-label models on all prediction labels, achieving C-indices of 0.773, 0.765, 0.776, and 0.773 for overall survival, local failure-free survival, regional failure-free survival, and distant failure-free survival, respectively. The area under the curve values ranged between 0.799 and 0.844 for different tasks at different time points. Using the medians of predicted risks as the thresholds to identify high-risk and low-risk patient groups, we performed the log-rank test, the results of which showed significantly larger separations in different event-free survivals.
We developed the first model capable of predicting multiple labels for OPC simultaneously. Our model demonstrated better prognostic ability for all the prediction targets compared with corresponding single-modality models and single-label models.
开发一种用于口咽癌(OPC)的可靠且全面的癌症预后模型,可以更好地辅助治疗个体化。在这项工作中,我们开发了一个基于 Vision Transformer(ViT)的多标签模型,具有多模态输入,以从可用的预处理数据中学习互补信息,并预测接受 OPC 放射治疗的患者的多个相关结局。
使用来自 512 名 OPC 患者的公开可用数据集进行模型训练和评估。计划计算机断层扫描图像、原发大体肿瘤体积掩模以及 16 个代表患者人口统计学、诊断和治疗的临床变量被用作输入。为了提取具有全局注意力的深度图像特征,我们使用了 ViT 模块。将临床变量与学习到的图像特征连接起来,并输入全连接层以合并跨模态特征。为了学习特征与相关生存结局(包括总生存、局部无失败生存、区域无失败生存和远处无失败生存)之间的映射,我们使用了 4 个多任务逻辑回归层。通过组合不同预测目标的多任务逻辑回归负对数似然损失来优化提出的模型。
我们分别使用 C 指数和曲线下面积(AUC)度量来评估我们的模型对时间事件预测和时间特定二元预测的性能。与相应的单模态和单标签模型相比,我们提出的模型在所有预测标签上均表现出色,总生存、局部无失败生存、区域无失败生存和远处无失败生存的 C 指数分别为 0.773、0.765、0.776 和 0.773。不同任务在不同时间点的 AUC 值范围在 0.799 到 0.844 之间。使用预测风险的中位数作为阈值来识别高风险和低风险患者组,我们进行了对数秩检验,结果表明在不同的无事件生存中差异更大。
我们开发了第一个能够同时预测 OPC 多个标签的模型。与相应的单模态模型和单标签模型相比,我们的模型在所有预测目标上都表现出更好的预后能力。