Keshavarz Hajar, Ansari Zohreh, Abootalebian Hossein, Sabet Babak, Momenzadeh Mohammadreza
Department of Artificial Intelligence in Medical Sciences, Smart University of Medical Sciences, Tehran, Iran.
Department of Biomedical Engineering, Engineering Faculty, Meybod University, Meybod, Yazd, Iran.
J Med Signals Sens. 2025 Jun 9;15:17. doi: 10.4103/jmss.jmss_23_24. eCollection 2025.
Deep learning has gained much attention in computer-assisted minimally invasive surgery in recent years. The application of deep-learning algorithms in colonoscopy can be divided into four main categories: surgical image analysis, surgical operations analysis, evaluation of surgical skills, and surgical automation. Analysis of surgical images by deep learning can be one of the main solutions for early detection of gastrointestinal lesions and for taking appropriate actions to treat cancer.
This study investigates a simple and accurate deep-learning model for polyp detection. We address the challenge of limited labeled data through transfer learning and employ multi-task learning to achieve both polyp classification and bounding box detection tasks. Considering the appropriate weight for each task in the total cost function is crucial in achieving the best results. Due to the lack of datasets with nonpolyp images, data collection was carried out. The proposed deep neural network structure was implemented on KVASIR-SEG and CVC-CLINIC datasets as polyp images in addition to the nonpolyp images extracted from the LDPolyp videos dataset.
The proposed model demonstrated high accuracy, achieving 100% in polyp/non-polyp classification and 86% in bounding box detection. It also showed fast processing times (0.01 seconds), making it suitable for real-time clinical applications.
The developed deep-learning model offers an efficient, accurate, and cost-effective solution for real-time polyp detection in colonoscopy. Its performance on benchmark datasets confirms its potential for clinical deployment, aiding in early cancer diagnosis and treatment.
近年来,深度学习在计算机辅助微创手术中备受关注。深度学习算法在结肠镜检查中的应用主要可分为四大类:手术图像分析、手术操作分析、手术技能评估和手术自动化。通过深度学习分析手术图像可能是早期检测胃肠道病变并采取适当措施治疗癌症的主要解决方案之一。
本研究探究了一种用于息肉检测的简单且准确的深度学习模型。我们通过迁移学习应对标记数据有限的挑战,并采用多任务学习来完成息肉分类和边界框检测任务。考虑总成本函数中每个任务的适当权重对于取得最佳结果至关重要。由于缺乏包含非息肉图像的数据集,因此进行了数据收集。除了从LDPolyp视频数据集中提取的非息肉图像外,所提出的深度神经网络结构还在KVASIR-SEG和CVC-CLINIC数据集上作为息肉图像实现。
所提出的模型表现出高准确率,在息肉/非息肉分类中达到100%,在边界框检测中达到86%。它还显示出快速的处理时间(0.01秒),使其适用于实时临床应用。
所开发的深度学习模型为结肠镜检查中的实时息肉检测提供了一种高效、准确且经济高效的解决方案。其在基准数据集上的性能证实了其临床部署的潜力,有助于早期癌症诊断和治疗。