SCARLETRED Holding GmbH, Vienna, Austria.
Department of Therapeutic Radiology and Oncology, Comprehensive Cancer Center, Medical University of Graz, Graz, Austria.
Comput Biol Med. 2021 Dec;139:104952. doi: 10.1016/j.compbiomed.2021.104952. Epub 2021 Oct 27.
Although significant advancements in computer-aided diagnostics using artificial intelligence (AI) have been made, to date, no viable method for radiation-induced skin reaction (RISR) analysis and classification is available. The objective of this single-center study was to develop machine learning and deep learning approaches using deep convolutional neural networks (CNNs) for automatic classification of RISRs according to the Common Terminology Criteria for Adverse Events (CTCAE) grading system. Scarletred Vision, a novel and state-of-the-art digital skin imaging method capable of remote monitoring and objective assessment of acute RISRs was used to convert 2D digital skin images using the CIELAB color space and conduct SEV* measurements. A set of different machine learning and deep convolutional neural network-based algorithms has been explored for the automatic classification of RISRs. A total of 2263 distinct images from 209 patients were analyzed for training and testing the machine learning and CNN algorithms. For a 2-class problem of healthy skin (grade 0) versus erythema (grade ≥ 1), all machine learning models produced an accuracy of above 70%, and the sensitivity and specificity of erythema recognition were 67-72% and 72-83%, respectively. The CNN produced a test accuracy of 74%, sensitivity of 66%, and specificity of 83% for predicting healthy and erythema cases. For the severity grade prediction of a 3-class problem (grade 0 versus 1 versus 2), the overall test accuracy was 60-67%, and the sensitivities were 56-82%, 35-59%, and 65-72%, respectively. For estimating the severity grade of each class, the CNN obtained an accuracy of 73%, 66%, and 82%, respectively. Ensemble learning combines several individual predictions to obtain a better generalization performance. Furthermore, we exploited ensemble learning by deploying a CNN model as a meta-learner. The ensemble CNN based on bagging and majority voting shows an accuracy, sensitivity and specificity of 87%, 90%, and 82% for a 2-class problem, respectively. For a 3-class problem, the ensemble CNN shows an overall accuracy of 66%, while for each grade (0, 1, and 2) accuracies were 76%, 69%, and 87%, sensitivities were 70%, 57%, and 71%, and specificities were 78%, 75%, and 95%, respectively. This study is the first to focus on erythema in radiation-dermatitis and produces benchmark results using machine learning models. The outcome of this study validates that the proposed system can act as a pre-screening and decision support tool for oncologists or patients to provide fast, reliable, and efficient assessment of erythema grading.
尽管在人工智能(AI)辅助诊断方面取得了重大进展,但迄今为止,还没有可行的方法可用于分析和分类放射性皮肤反应(RISR)。本单中心研究的目的是开发机器学习和深度学习方法,使用深度卷积神经网络(CNN)根据常见不良事件术语标准(CTCAE)分级系统自动分类 RISR。Scarletred Vision 是一种新颖的最先进的数字皮肤成像方法,能够进行远程监测和急性 RISR 的客观评估,用于使用 CIELAB 颜色空间转换 2D 数字皮肤图像并进行 SEV*测量。已经探索了一组不同的基于机器学习和深度卷积神经网络的算法,用于自动分类 RISR。使用来自 209 名患者的 2263 张不同图像来训练和测试机器学习和 CNN 算法。对于健康皮肤(等级 0)与红斑(等级≥1)的 2 类问题,所有机器学习模型的准确率均高于 70%,红斑识别的敏感性和特异性分别为 67-72%和 72-83%。CNN 对预测健康和红斑病例的测试准确率为 74%,敏感性为 66%,特异性为 83%。对于 3 类问题(等级 0 与 1 与 2)的严重程度预测,整体测试准确率为 60-67%,敏感性分别为 56-82%、35-59%和 65-72%。对于估计每个等级的严重程度,CNN 的准确率分别为 73%、66%和 82%。集成学习将多个个体预测结合起来以获得更好的泛化性能。此外,我们通过部署 CNN 模型作为元学习者来利用集成学习。基于装袋和多数投票的集成 CNN 分别具有 2 类问题的 87%、90%和 82%的准确率、敏感性和特异性。对于 3 类问题,集成 CNN 的整体准确率为 66%,而对于每个等级(0、1 和 2),准确率分别为 76%、69%和 87%,敏感性分别为 70%、57%和 71%,特异性分别为 78%、75%和 95%。这项研究首次专注于放射性皮炎中的红斑,并使用机器学习模型产生基准结果。该研究的结果验证了所提出的系统可以作为肿瘤科医生或患者的预筛选和决策支持工具,提供快速、可靠和高效的红斑分级评估。