Suppr超能文献

预训练的YOLO检测器应用于未见过的延时图像进行自动传粉者监测的成功与局限

Successes and limitations of pretrained YOLO detectors applied to unseen time-lapse images for automated pollinator monitoring.

作者信息

Ștefan Valentin, Stark Thomas, Wurm Michael, Taubenböck Hannes, Knight Tiffany M

机构信息

Department of Species Interaction Ecology, Helmholtz Centre for Environmental Research - UFZ, Leipzig, Germany.

German Centre for Integrative Biodiversity Research (iDiv) Halle-Jena-Leipzig, Leipzig, Germany.

出版信息

Sci Rep. 2025 Aug 21;15(1):30671. doi: 10.1038/s41598-025-16140-z.

Abstract

Pollinating insects provide essential ecosystem services, and using time-lapse photography to automate their observation could improve monitoring efficiency. Computer vision models, trained on clear citizen science photos, can detect insects in similar images with high accuracy, but their performance in images taken using time-lapse photography is unknown. We evaluated the generalisation of three lightweight YOLO detectors (YOLOv5-nano, YOLOv5-small, YOLOv7-tiny), previously trained on citizen science images, for detecting ~ 1,300 flower-visiting arthropod individuals in nearly 24,000 time-lapse images captured with a fixed smartphone setup. These field images featured unseen backgrounds and smaller arthropods than the training data. YOLOv5-small, the model with the highest number of trainable parameters, performed best, localising 91.21% of Hymenoptera and 80.69% of Diptera individuals. However, classification recall was lower (80.45% and 66.90%, respectively), partly due to Syrphidae mimicking Hymenoptera and the challenge of detecting smaller, blurrier flower visitors. This study reveals both the potential and limitations of such models for real-world automated monitoring, suggesting they work well for larger and sharply visible pollinators but need improvement for smaller, less sharp cases.

摘要

传粉昆虫提供重要的生态系统服务,使用延时摄影来自动观察它们可以提高监测效率。在清晰的公民科学照片上训练的计算机视觉模型能够高精度地检测类似图像中的昆虫,但其在使用延时摄影拍摄的图像中的性能尚不清楚。我们评估了之前在公民科学图像上训练的三种轻量级YOLO探测器(YOLOv5-纳米、YOLOv5-小型、YOLOv7-微小)在使用固定智能手机设置拍摄的近24000张延时图像中检测约1300只访花节肢动物个体的泛化能力。这些野外图像具有未见过的背景,且节肢动物比训练数据中的更小。可训练参数数量最多的模型YOLOv5-小型表现最佳,定位了91.21%的膜翅目个体和80.69%的双翅目个体。然而,分类召回率较低(分别为80.45%和66.90%),部分原因是食蚜蝇模仿膜翅目,以及检测更小、更模糊的访花者存在挑战。这项研究揭示了此类模型在实际自动监测中的潜力和局限性,表明它们对较大且清晰可见的传粉者效果良好,但对更小、不太清晰的情况需要改进。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验