Chou Chu-Kuang, Karmakar Riya, Tsao Yu-Ming, Jie Lim Wei, Mukundan Arvind, Huang Chien-Wei, Chen Tsung-Hsien, Ko Chau-Yuan, Wang Hsiang-Chen
Division of Gastroenterology and Hepatology, Department of Internal Medicine, Ditmanson Medical Foundation Chia-Yi Christian Hospital, Chia-Yi 60002, Taiwan.
Obesity Center, Ditmanson Medical Foundation Chia-Yi Christian Hospital, Chia-Yi 60002, Taiwan.
Diagnostics (Basel). 2024 May 29;14(11):1129. doi: 10.3390/diagnostics14111129.
The early detection of esophageal cancer presents a substantial difficulty, which contributes to its status as a primary cause of cancer-related fatalities. This study used You Only Look Once (YOLO) frameworks, specifically YOLOv5 and YOLOv8, to predict and detect early-stage EC by using a dataset sourced from the Division of Gastroenterology and Hepatology, Ditmanson Medical Foundation, Chia-Yi Christian Hospital. The dataset comprised 2741 white-light images (WLI) and 2741 hyperspectral narrowband images (HSI-NBI). They were divided into 60% training, 20% validation, and 20% test sets to facilitate robust detection. The images were produced using a conversion method called the spectrum-aided vision enhancer (SAVE). This algorithm can transform a WLI into an NBI without requiring a spectrometer or spectral head. The main goal was to identify dysplasia and squamous cell carcinoma (SCC). The model's performance was evaluated using five essential metrics: precision, recall, F1-score, mAP, and the confusion matrix. The experimental results demonstrated that the HSI model exhibited improved learning capabilities for SCC characteristics compared with the original RGB images. Within the YOLO framework, YOLOv5 outperformed YOLOv8, indicating that YOLOv5's design possessed superior feature-learning skills. The YOLOv5 model, when used in conjunction with HSI-NBI, demonstrated the best performance. It achieved a precision rate of 85.1% (CI95: 83.2-87.0%, < 0.01) in diagnosing SCC and an F1-score of 52.5% (CI95: 50.1-54.9%, < 0.01) in detecting dysplasia. The results of these figures were much better than those of YOLOv8. YOLOv8 achieved a precision rate of 81.7% (CI95: 79.6-83.8%, < 0.01) and an F1-score of 49.4% (CI95: 47.0-51.8%, < 0.05). The YOLOv5 model with HSI demonstrated greater performance than other models in multiple scenarios. This difference was statistically significant, suggesting that the YOLOv5 model with HSI significantly improved detection capabilities.
食管癌的早期检测存在很大困难,这也是其成为癌症相关死亡主要原因的一个因素。本研究使用了You Only Look Once(YOLO)框架,即YOLOv5和YOLOv8,通过使用来自嘉义基督教医院迪特曼森医学基金会胃肠病学和肝病学部门的数据集来预测和检测早期食管癌。该数据集包括2741张白光图像(WLI)和2741张高光谱窄带图像(HSI-NBI)。它们被分为60%的训练集、20%的验证集和20%的测试集,以促进可靠的检测。这些图像是使用一种称为光谱辅助视觉增强器(SAVE)的转换方法生成的。该算法无需光谱仪或光谱探头就能将白光图像转换为窄带图像。主要目标是识别发育异常和鳞状细胞癌(SCC)。使用五个关键指标评估模型的性能:精确率、召回率、F1分数、平均精度均值(mAP)和混淆矩阵。实验结果表明,与原始RGB图像相比,高光谱图像(HSI)模型对SCC特征表现出更好的学习能力。在YOLO框架内,YOLOv5的表现优于YOLOv8,表明YOLOv5的设计具有卓越的特征学习能力。当与HSI-NBI结合使用时,YOLOv5模型表现最佳。在诊断SCC时,其精确率达到85.1%(95%置信区间:83.2-87.0%,P<0.01),在检测发育异常时F1分数为52.5%(95%置信区间:50.1-54.9%,P<0.01)。这些数据结果比YOLOv8的结果好得多。YOLOv8的精确率为81.7%(95%置信区间:79.6-83.8%,P<0.01),F1分数为49.4%(95%置信区间:47.0-51.8%,P<0.05)。带有HSI-NBI的YOLOv5模型在多种情况下表现优于其他模型。这种差异具有统计学意义,表明带有HSI-NBI的YOLOv5模型显著提高了检测能力。