Wang Lele, Zhao Yingjie, Liu Shengbo, Li Yuanhong, Chen Shengde, Lan Yubin
College of Electronic Engineering, College of Artificial Intelligence, South China Agricultural University, Guangzhou, China.
National Center for International Collaboration Research on Precision Agricultural Aviation Pesticides Spraying Technology, Guangzhou, China.
Front Plant Sci. 2022 Mar 11;13:839269. doi: 10.3389/fpls.2022.839269. eCollection 2022.
The precision detection of dense small targets in orchards is critical for the visual perception of agricultural picking robots. At present, the visual detection algorithms for plums still have a poor recognition effect due to the characteristics of small plum shapes and dense growth. Thus, this paper proposed a lightweight model based on the improved You Only Look Once version 4 (YOLOv4) to detect dense plums in orchards. First, we employed a data augmentation method based on category balance to alleviate the imbalance in the number of plums of different maturity levels and insufficient data quantity. Second, we abandoned Center and Scale Prediction Darknet53 (CSPDarknet53) and chose a lighter MobilenetV3 on selecting backbone feature extraction networks. In the feature fusion stage, we used depthwise separable convolution (DSC) instead of standard convolution to achieve the purpose of reducing model parameters. To solve the insufficient feature extraction problem of dense targets, this model achieved fine-grained detection by introducing a 152 × 152 feature layer. The Focal loss and complete intersection over union (CIOU) loss were joined to balance the contribution of hard-to-classify and easy-to-classify samples to the total loss. Then, the improved model was trained through transfer learning at different stages. Finally, several groups of detection experiments were designed to evaluate the performance of the improved model. The results showed that the improved YOLOv4 model had the best mean average precision (mAP) performance than YOLOv4, YOLOv4-tiny, and MobileNet-Single Shot Multibox Detector (MobileNet-SSD). Compared with some results from the YOLOv4 model, the model size of the improved model is compressed by 77.85%, the parameters are only 17.92% of the original model parameters, and the detection speed is accelerated by 112%. In addition, the influence of the automatic data balance algorithm on the accuracy of the model and the detection effect of the improved model under different illumination angles, different intensity levels, and different types of occlusions were discussed in this paper. It is indicated that the improved detection model has strong robustness and high accuracy under the real natural environment, which can provide data reference for the subsequent orchard yield estimation and engineering applications of robot picking work.
果园中密集小目标的精确检测对于农业采摘机器人的视觉感知至关重要。目前,由于李子形状小且生长密集的特点,针对李子的视觉检测算法识别效果仍然较差。因此,本文提出了一种基于改进的YOLOv4(You Only Look Once版本4)的轻量级模型,用于检测果园中的密集李子。首先,我们采用了一种基于类别平衡的数据增强方法,以缓解不同成熟度李子数量不平衡和数据量不足的问题。其次,在选择骨干特征提取网络时,我们舍弃了Center and Scale Prediction Darknet53(CSPDarknet53),选择了更轻量级的MobilenetV3。在特征融合阶段,我们使用深度可分离卷积(DSC)代替标准卷积,以达到减少模型参数的目的。为了解决密集目标特征提取不足的问题,该模型通过引入一个152×152的特征层实现了细粒度检测。加入Focal损失和完整交并比(CIOU)损失,以平衡难分类和易分类样本对总损失的贡献。然后,通过不同阶段的迁移学习对改进后的模型进行训练。最后,设计了几组检测实验来评估改进后模型的性能。结果表明,改进后的YOLOv4模型比YOLOv4、YOLOv4-tiny和MobileNet-单阶段多框检测器(MobileNet-SSD)具有最佳的平均精度均值(mAP)性能。与YOLOv4模型的一些结果相比,改进后模型的模型大小压缩了77.85%,参数仅为原始模型参数的17.92%,检测速度加快了112%。此外,本文还讨论了自动数据平衡算法对模型精度的影响以及改进后模型在不同光照角度、不同强度水平和不同遮挡类型下的检测效果。结果表明,改进后的检测模型在真实自然环境下具有很强的鲁棒性和高精度,可为后续果园产量估计和机器人采摘工作的工程应用提供数据参考。