Zhaosheng Yao, Tao Liu, Tianle Yang, Chengxin Ju, Chengming Sun
Jiangsu Key Laboratory of Crop Genetics and Physiology/Jiangsu Key Laboratory of Crop Cultivation and Physiology, Agricultural College of Yangzhou University, Yangzhou, China.
Jiangsu Co-Innovation Center for Modern Production Technology of Grain Crops, Yangzhou University, Yangzhou, China.
Front Plant Sci. 2022 Apr 27;13:851245. doi: 10.3389/fpls.2022.851245. eCollection 2022.
Wheat ears in unmanned aerial vehicles (UAV) orthophotos are characterized by occlusion, small targets, dense distribution, and complex backgrounds. Rapid identification of wheat ears in UAV orthophotos in a field environment is critical for wheat yield prediction. Three improvements were achieved based on YOLOX-m: mosaic optimized, using BiFPN structure, and attention mechanism, then ablation experiments were performed to verify the effect of each improvement. Three scene datasets were established: images were acquired during three different growing periods, at three planting densities, and under three scenarios of UAV flight heights. In ablation experiments, three improvements had increased recognition accuracies on the experimental dataset. Compared the accuracy of the standard model with our improved model on three scene datasets. Our improved model during three different periods, at three planting densities, and under three scenarios of the UAV flight height, obtaining 88.03%, 87.59%, and 87.93% accuracies, which were, respectively, 2.54%, 1.89%, and 2.15% better than the original model. The results of this study showed that the improved YOLOX-m model can achieve UAV orthophoto wheat recognition under different practical scenarios in large fields, and that the best combination were obtained images from the wheat milk stage, low planting density, and low flight altitude.
无人机正射影像中的麦穗具有遮挡、目标小、分布密集和背景复杂的特点。在田间环境中快速识别无人机正射影像中的麦穗对于小麦产量预测至关重要。基于YOLOX-m实现了三项改进:马赛克优化、使用BiFPN结构和注意力机制,然后进行消融实验以验证各项改进的效果。建立了三个场景数据集:在三个不同生长时期、三种种植密度以及三种无人机飞行高度场景下采集图像。在消融实验中,三项改进提高了实验数据集上的识别准确率。在三个场景数据集上比较了标准模型和改进模型的准确率。我们的改进模型在三个不同时期、三种种植密度以及三种无人机飞行高度场景下,准确率分别为88.03%、87.59%和87.93%,分别比原始模型高2.54%、1.89%和2.15%。研究结果表明,改进后的YOLOX-m模型能够在大田不同实际场景下实现无人机正射影像小麦识别,且最佳组合是从小麦灌浆期、低密度种植和低飞行高度获取的图像。