Department of Food, Agricultural, and Biological Engineering, Ohio State University, 590 Woody Hayes Dr, Columbus, OH 43210, USA.
Google, Kirkland, WA 98033, USA.
Sensors (Basel). 2024 Oct 7;24(19):6467. doi: 10.3390/s24196467.
Plant counting is a critical aspect of crop management, providing farmers with valuable insights into seed germination success and within-field variation in crop population density, both of which are key indicators of crop yield and quality. Recent advancements in Unmanned Aerial System (UAS) technology, coupled with deep learning techniques, have facilitated the development of automated plant counting methods. Various computer vision models based on UAS images are available for detecting and classifying crop plants. However, their accuracy relies largely on the availability of substantial manually labeled training datasets. The objective of this study was to develop a robust corn counting model by developing and integrating an automatic image annotation framework. This study used high-spatial-resolution images collected with a DJI Mavic Pro 2 at the V2-V4 growth stage of corn plants from a field in Wooster, Ohio. The automated image annotation process involved extracting corn rows and applying image enhancement techniques to automatically annotate images as either corn or non-corn, resulting in 80% accuracy in identifying corn plants. The accuracy of corn stand identification was further improved by training four deep learning (DL) models, including InceptionV3, VGG16, VGG19, and Vision Transformer (ViT), with annotated images across various datasets. Notably, VGG16 outperformed the other three models, achieving an F1 score of 0.955. When the corn counts were compared to ground truth data across five test regions, VGG achieved an R of 0.94 and an RMSE of 9.95. The integration of an automated image annotation process into the training of the DL models provided notable benefits in terms of model scaling and consistency. The developed framework can efficiently manage large-scale data generation, streamlining the process for the rapid development and deployment of corn counting DL models.
植物计数是作物管理的一个关键方面,它为农民提供了有关种子发芽成功率和田间作物种群密度变化的有价值的信息,这些都是作物产量和质量的关键指标。最近,无人机系统 (UAS) 技术的进步,加上深度学习技术,促进了自动化植物计数方法的发展。各种基于 UAS 图像的计算机视觉模型可用于检测和分类作物植物。然而,它们的准确性在很大程度上依赖于大量可用的手动标记训练数据集。本研究的目的是通过开发和集成自动图像注释框架来开发稳健的玉米计数模型。本研究使用 DJI Mavic Pro 2 在俄亥俄州伍斯特的一个玉米田收集的高空间分辨率图像,该图像是在玉米植株的 V2-V4 生长阶段拍摄的。自动图像注释过程涉及提取玉米行,并应用图像增强技术自动将图像注释为玉米或非玉米,从而在识别玉米植株方面达到 80%的准确率。通过在不同数据集上使用注释图像对四个深度学习 (DL) 模型(InceptionV3、VGG16、VGG19 和 Vision Transformer (ViT))进行训练,进一步提高了玉米植株识别的准确性。值得注意的是,VGG16 优于其他三个模型,F1 得分为 0.955。当将玉米计数与五个测试区域的地面实况数据进行比较时,VGG 的 R 为 0.94,RMSE 为 9.95。将自动图像注释过程集成到 DL 模型的训练中,在模型扩展和一致性方面提供了显著的优势。所开发的框架可以有效地管理大规模数据生成,简化了快速开发和部署玉米计数 DL 模型的过程。