Vulpoi Radu Alexandru, Ciobanu Adrian, Drug Vasile Liviu, Mihai Catalina, Barboi Oana Bogdana, Floria Diana Elena, Coseru Alexandru Ionut, Olteanu Andrei, Rosca Vadim, Luca Mihaela
Institute of Gastroenterology and Hepatology, "Grigore T. Popa" University of Medicine and Pharmacy, 700111 Iasi, Romania.
Institute of Computer Science, Romanian Academy, Iasi Branch, 700481 Iasi, Romania.
J Imaging. 2025 Mar 18;11(3):84. doi: 10.3390/jimaging11030084.
This study aims to objectively evaluate the overall quality of colonoscopies using a specially trained deep learning-based semantic segmentation neural network. This represents a modern and valuable approach for the analysis of colonoscopy frames. We collected thousands of colonoscopy frames extracted from a set of video colonoscopy files. A color-based image processing method was used to extract color features from specific regions of each colonoscopy frame, namely, the intestinal mucosa, residues, artifacts, and lumen. With these features, we automatically annotated all the colonoscopy frames and then selected the best of them to train a semantic segmentation network. This trained network was used to classify the four region types in a different set of test colonoscopy frames and extract pixel statistics that are relevant to quality evaluation. The test colonoscopies were also evaluated by colonoscopy experts using the Boston scale. The deep learning semantic segmentation method obtained good results, in terms of classifying the four key regions in colonoscopy frames, and produced pixel statistics that are efficient in terms of objective quality assessment. The Spearman correlation results were as follows: BBPS vs. pixel scores: 0.69; BBPS vs. mucosa pixel percentage: 0.63; BBPS vs. residue pixel percentage: -0.47; BBPS vs. Artifact Pixel Percentage: -0.65. The agreement analysis using Cohen's Kappa yielded a value of 0.28. The colonoscopy evaluation based on the extracted pixel statistics showed a fair level of compatibility with the experts' evaluations. Our proposed deep learning semantic segmentation approach is shown to be a promising tool for evaluating the overall quality of colonoscopies and goes beyond the Boston Bowel Preparation Scale in terms of assessing colonoscopy quality. In particular, while the Boston scale focuses solely on the amount of residual content, our method can identify and quantify the percentage of colonic mucosa, residues, and artifacts, providing a more comprehensive and objective evaluation.
本研究旨在使用经过专门训练的基于深度学习的语义分割神经网络,客观评估结肠镜检查的整体质量。这代表了一种分析结肠镜检查图像帧的现代且有价值的方法。我们从一组视频结肠镜检查文件中提取了数千帧结肠镜检查图像。采用基于颜色的图像处理方法,从每个结肠镜检查图像帧的特定区域提取颜色特征,即肠黏膜、残留物、伪影和管腔。利用这些特征,我们自动标注了所有结肠镜检查图像帧,然后挑选出最佳的图像帧来训练语义分割网络。这个经过训练的网络用于对另一组测试结肠镜检查图像帧中的四种区域类型进行分类,并提取与质量评估相关的像素统计数据。测试结肠镜检查还由结肠镜检查专家使用波士顿量表进行评估。深度学习语义分割方法在对结肠镜检查图像帧中的四个关键区域进行分类方面取得了良好结果,并产生了在客观质量评估方面有效的像素统计数据。斯皮尔曼相关性结果如下:波士顿肠道准备评分(BBPS)与像素分数:0.69;BBPS与黏膜像素百分比:0.63;BBPS与残留像素百分比:-0.47;BBPS与伪影像素百分比:-0.65。使用科恩kappa系数进行的一致性分析得出的值为0.28。基于提取的像素统计数据的结肠镜检查评估与专家评估显示出一定程度的兼容性。我们提出的深度学习语义分割方法被证明是评估结肠镜检查整体质量的一种有前途的工具,并且在评估结肠镜检查质量方面超越了波士顿肠道准备量表。特别是,虽然波士顿量表仅关注残留内容的量,但我们的方法可以识别并量化结肠黏膜、残留物和伪影的百分比,提供更全面和客观的评估。