Bondarenko Andrey, Jumutc Vilen, Netter Antoine, Duchateau Fanny, Abrão Henrique Mendonca, Noorzadeh Saman, Giacomello Giuseppe, Ferrari Filippo, Bourdel Nicolas, Kirk Ulrik Bak, Bļizņuks Dmitrijs
Institute of Applied Computer Systems, Riga Technical University, LV-1048 Riga, Latvia.
Department of Obstetrics and Gynecology, Marseille Hospital, 13005 Marseille, France.
Diagnostics (Basel). 2025 May 15;15(10):1254. doi: 10.3390/diagnostics15101254.
Laparoscopic surgery for endometriosis presents unique challenges due to the complexity of and variability in lesion appearances within the abdominal cavity. This study investigates the application of deep learning models for object detection in laparoscopic videos, aiming to assist surgeons in accurately identifying and localizing endometriosis lesions and related anatomical structures. A custom dataset was curated, comprising of 199 video sequences and 205,725 frames. Of these, 17,560 frames were meticulously annotated by medical professionals. The dataset includes object detection annotations for 10 object classes relevant to endometriosis, alongside segmentation masks for some classes. To address the object detection task, we evaluated the performance of two deep learning models-FasterRCNN and YOLOv9-under both stratified and non-stratified training scenarios. The experimental results demonstrated that stratified training significantly reduced the risk of data leakage and improved model generalization. The best-performing FasterRCNN object detection model achieved a high average test precision of 0.9811 ± 0.0084, recall of 0.7083 ± 0.0807, and mAP50 (mean average precision at 50% overlap) of 0.8185 ± 0.0562 across all presented classes. Despite these successes, the study also highlights the challenges posed by the weak annotations and class imbalances in the dataset, which impacted overall model performances. In conclusion, this study provides valuable insights into the application of deep learning for enhancing laparoscopic surgical precision in endometriosis treatment. The findings underscore the importance of robust dataset curation and advanced training strategies in developing reliable AI-assisted tools for surgical interventions. The latter could potentially improve the guidance of surgical interventions and prevent blind spots occurring in difficult to reach abdominal regions. Future work will focus on refining the dataset and exploring more sophisticated model architectures to further improve detection accuracy.
由于腹腔内病变外观的复杂性和变异性,子宫内膜异位症的腹腔镜手术面临着独特的挑战。本研究调查了深度学习模型在腹腔镜视频中进行目标检测的应用,旨在帮助外科医生准确识别和定位子宫内膜异位症病变及相关解剖结构。精心策划了一个自定义数据集,包括199个视频序列和205,725帧。其中,17,560帧由医学专业人员进行了细致标注。该数据集包括与子宫内膜异位症相关的10个目标类别的目标检测标注,以及部分类别的分割掩码。为解决目标检测任务,我们在分层和非分层训练场景下评估了两种深度学习模型——FasterRCNN和YOLOv9的性能。实验结果表明,分层训练显著降低了数据泄露的风险并提高了模型的泛化能力。表现最佳的FasterRCNN目标检测模型在所有呈现的类别中,平均测试精度达到了0.9811±0.0084,召回率为0.7083±0.0807,mAP50(重叠率为50%时的平均精度均值)为0.8185±0.0562。尽管取得了这些成功,但该研究也凸显了数据集中弱标注和类别不平衡所带来的挑战,这些因素影响了整体模型性能。总之,本研究为深度学习在提高子宫内膜异位症治疗中腹腔镜手术精度方面的应用提供了有价值的见解。研究结果强调了在开发可靠的手术干预人工智能辅助工具时,稳健的数据集策划和先进训练策略的重要性。后者可能会改善手术干预的指导,并防止在难以到达的腹部区域出现盲点。未来的工作将集中在完善数据集和探索更复杂的模型架构,以进一步提高检测精度。