Gao Yangmingrui, Li Yinglun, Jiang Ruibo, Zhan Xiaohai, Lu Hao, Guo Wei, Yang Wanneng, Ding Yanfeng, Liu Shouyang
Plant Phenomics Research Centre, Academy for Advanced Interdisciplinary Studies, Jiangsu Collaborative Innovation Center for Modern Crop Production, Nanjing Agricultural University, Nanjing, China.
Key Laboratory of Image Processing and Intelligent Control, School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China.
Plant Phenomics. 2023 Jul 18;5:0064. doi: 10.34133/plantphenomics.0064. eCollection 2023.
The green fraction (GF), which is the fraction of green vegetation in a given viewing direction, is closely related to the light interception ability of the crop canopy. Monitoring the dynamics of GF is therefore of great interest for breeders to identify genotypes with high radiation use efficiency. The accuracy of GF estimation depends heavily on the quality of the segmentation dataset and the accuracy of the image segmentation method. To enhance segmentation accuracy while reducing annotation costs, we developed a self-supervised strategy for deep learning semantic segmentation of rice and wheat field images with very contrasting field backgrounds. First, the Digital Plant Phenotyping Platform was used to generate large, perfectly labeled simulated field images for wheat and rice crops, considering diverse canopy structures and a wide range of environmental conditions (sim dataset). We then used the domain adaptation model cycle-consistent generative adversarial network (CycleGAN) to bridge the reality gap between the simulated and real images (real dataset), producing simulation-to-reality images (sim2real dataset). Finally, 3 different semantic segmentation models (U-Net, DeepLabV3+, and SegFormer) were trained using 3 datasets (real, sim, and sim2real datasets). The performance of the 9 training strategies was assessed using real images captured from various sites. The results showed that SegFormer trained using the sim2real dataset achieved the best segmentation performance for both rice and wheat crops (rice: Accuracy = 0.940, F1-score = 0.937; wheat: Accuracy = 0.952, F1-score = 0.935). Likewise, favorable GF estimation results were obtained using the above strategy (rice: = 0.967, RMSE = 0.048; wheat: = 0.984, RMSE = 0.028). Compared with SegFormer trained using a real dataset, the optimal strategy demonstrated greater superiority for wheat images than for rice images. This discrepancy can be partially attributed to the differences in the backgrounds of the rice and wheat fields. The uncertainty analysis indicated that our strategy could be disrupted by the inhomogeneity of pixel brightness and the presence of senescent elements in the images. In summary, our self-supervised strategy addresses the issues of high cost and uncertain annotation accuracy during dataset creation, ultimately enhancing GF estimation accuracy for rice and wheat field images. The best weights we trained in wheat and rice are available: https://github.com/PheniX-Lab/sim2real-seg.
绿色分量(GF)是指在给定观察方向上绿色植被的比例,它与作物冠层的光截获能力密切相关。因此,监测GF的动态变化对于育种者识别具有高辐射利用效率的基因型具有重要意义。GF估计的准确性在很大程度上取决于分割数据集的质量和图像分割方法的准确性。为了提高分割精度同时降低标注成本,我们针对具有非常不同田间背景的水稻和小麦田图像的深度学习语义分割开发了一种自监督策略。首先,利用数字植物表型平台,考虑不同的冠层结构和广泛的环境条件,生成了用于小麦和水稻作物的大量、标注完美的模拟田间图像(模拟数据集)。然后,我们使用域适应模型循环一致生成对抗网络(CycleGAN)来弥合模拟图像和真实图像(真实数据集)之间的现实差距,生成模拟到现实的图像(模拟到现实数据集)。最后,使用3个数据集(真实、模拟和模拟到现实数据集)训练了3种不同的语义分割模型(U-Net、DeepLabV3+和SegFormer)。使用从不同地点拍摄的真实图像评估了9种训练策略的性能。结果表明,使用模拟到现实数据集训练的SegFormer在水稻和小麦作物上均取得了最佳分割性能(水稻:准确率=0.940,F1分数=0.937;小麦:准确率=0.952,F1分数=0.935)。同样,使用上述策略也获得了良好的GF估计结果(水稻: =0.967,均方根误差=0.048;小麦: =0.984,均方根误差=0.028)。与使用真实数据集训练的SegFormer相比,最优策略在小麦图像上比在水稻图像上表现出更大的优势。这种差异部分可归因于水稻田和小麦田背景的不同。不确定性分析表明,我们的策略可能会受到图像中像素亮度不均匀和衰老元素存在的干扰。总之,我们的自监督策略解决了数据集创建过程中成本高和标注准确性不确定的问题,最终提高了水稻和小麦田图像的GF估计准确性。我们在小麦和水稻中训练的最佳权重可在以下网址获取:https://github.com/PheniX-Lab/sim2real-seg。