Waki Kotaro, Nagaoka Katsuya, Okubo Keishi, Kiyama Masato, Gushima Ryosuke, Ohno Kento, Honda Munenori, Yamasaki Akira, Matsuno Kenshi, Furuta Yoki, Miyamoto Hideaki, Naoe Hideaki, Amagasaki Motoki, Tanaka Yasuhito
Department of Gastroenterology and Hepatology, Faculty of Life Sciences, Kumamoto University, 1-1-1 Honjo, Chuo-ku, Kumamoto, Kumamoto, 860-8556, Japan.
Faculty of Advanced Science and Technology, Kumamoto University, Kumamoto, Japan.
Sci Rep. 2025 Feb 1;15(1):4003. doi: 10.1038/s41598-025-86829-8.
There is a currently an unmet need for non-invasive methods to predict the risk of esophageal squamous cell carcinoma (ESCC). Previously, we found that specific soft palate morphologies are strongly associated with increased ESCC risk. However, there is currently no artificial intelligence (AI) system that utilizes oral images for ESCC risk assessment. Here, we evaluated three AI models and three fine-tuning approaches with regard to their ESCC predictive power. Our dataset contained 539 cases, which were subdivided into 221 high-risk cases (2491 images) and 318 non-high-risk cases (2524 images). We used 480 cases (4295 images) for the training dataset, and the rest for validation. The Bilinear convolutional neural network (CNN) model (especially when pre-trained on fractal images) demonstrated diagnostic precision that was comparable to or better than other models for distinguishing between high-risk and non-high-risk groups. In addition, when tested with a small number of images containing soft palate data, the model showed high precision: the best AUC model had 0.91 (sensitivity 0.86, specificity 0.79). This study presents a significant advance in the development of an AI-based non-invasive screening tool for the identification of high-risk ESCC patients. The approach may be particularly suitable for institutes with limited medical imaging resources.
目前,对于预测食管鳞状细胞癌(ESCC)风险的非侵入性方法存在未满足的需求。此前,我们发现特定的软腭形态与ESCC风险增加密切相关。然而,目前尚无利用口腔图像进行ESCC风险评估的人工智能(AI)系统。在此,我们评估了三种AI模型和三种微调方法在ESCC预测能力方面的表现。我们的数据集包含539例病例,分为221例高风险病例(2491张图像)和318例非高风险病例(2524张图像)。我们将480例病例(4295张图像)用于训练数据集,其余用于验证。双线性卷积神经网络(CNN)模型(特别是在分形图像上进行预训练时)在区分高风险和非高风险组方面表现出与其他模型相当或更好的诊断精度。此外,当用少量包含软腭数据的图像进行测试时,该模型显示出高精度:最佳AUC模型为0.91(敏感性0.86,特异性0.79)。本研究在开发基于AI的非侵入性筛查工具以识别高风险ESCC患者方面取得了重大进展。该方法可能特别适用于医学影像资源有限的机构。