Zhou Lingli, Hu Anqi, Cheng Yawen, Zhang Wenxiang, Zhang Bingyuan, Lu Xinyu, Wu Qian, Ren Ni
Institute of Agricultural Information, Jiangsu Academy of Agricultural Sciences, Nanjing, China.
Key Laboratory of Intelligent Agricultural Technology (Changjiang Delta), Ministry of Agriculture and Rural Affairs, Nanjing, China.
Front Plant Sci. 2024 Oct 31;15:1460060. doi: 10.3389/fpls.2024.1460060. eCollection 2024.
With the advancement of computer vision technology, vision-based target perception has emerged as a predominant approach for harvesting robots to identify and locate fruits. However, little attention has been paid to the fact that fruits may be obscured by stems or other objects. In order to improve the vision detection ability of fruit harvesting robot, a fruit target selection and location approach considering obstacle perception was proposed. To enrich the dataset for tomato harvesting, synthetic data were generated by rendering a 3D simulated model of the tomato greenhouse environment, and automatically producing corresponding pixel-level semantic segmentation labels. An attention-based spatial-relationship feature extraction module (SFM) with lower computational complexity was designed to enhance the ability of semantic segmentation network DeepLab v3+ in accurately segmenting linear-structured obstructions such as stems and wires. An adaptive K-means clustering method was developed to distinguish individual instances of fruits. Furthermore, a barrier-free fruit selection algorithm that integrates information of obstacles and fruit instances was proposed to identify the closest and largest non-occluded fruit as the optimal picking target. The improved semantic segmentation network exhibited enhanced performance, achieving an accuracy of 96.75%. Notably, the Intersection-over-Union () of wire and stem classes was improved by 5.0% and 2.3%, respectively. Our target selection method demonstrated accurate identification of obstacle types (96.15%) and effectively excluding fruits obstructed by strongly resistant objects (86.67%). Compared to the fruit detection method without visual obstacle avoidance (Yolo v5), our approach exhibited an 18.9% increase in selection precision and a 1.3% reduction in location error. The improved semantic segmentation algorithm significantly increased the segmentation accuracy of linear-structured obstacles, and the obstacle perception algorithm effectively avoided occluded fruits. The proposed method demonstrated an appreciable ability in precisely selecting and locating barrier-free fruits within non-structural environments, especially avoiding fruits obscured by stems or wires. This approach provides a more reliable and practical solution for fruit selection and localization for harvesting robots, while also being applicable to other fruits and vegetables such as sweet peppers and kiwis.
随着计算机视觉技术的发展,基于视觉的目标感知已成为收获机器人识别和定位水果的主要方法。然而,水果可能会被茎或其他物体遮挡这一事实却很少受到关注。为了提高水果收获机器人的视觉检测能力,提出了一种考虑障碍物感知的水果目标选择和定位方法。为了丰富番茄收获的数据集,通过渲染番茄温室环境的3D模拟模型并自动生成相应的像素级语义分割标签来生成合成数据。设计了一种计算复杂度较低的基于注意力的空间关系特征提取模块(SFM),以增强语义分割网络DeepLab v3+准确分割茎和电线等线性结构障碍物的能力。开发了一种自适应K均值聚类方法来区分单个水果实例。此外,还提出了一种整合障碍物和水果实例信息的无障碍水果选择算法,以识别最近且最大的未被遮挡的水果作为最佳采摘目标。改进后的语义分割网络表现出更高的性能,准确率达到96.75%。值得注意的是,电线和茎类别的交并比(IoU)分别提高了5.0%和2.3%。我们的目标选择方法在障碍物类型识别方面表现准确(96.15%),并能有效排除被强抗性物体遮挡的水果(86.67%)。与没有视觉避障功能的水果检测方法(Yolo v5)相比,我们的方法在选择精度上提高了18.9%,定位误差降低了1.3%。改进后的语义分割算法显著提高了线性结构障碍物的分割精度,障碍物感知算法有效避免了被遮挡的水果。所提出的方法在非结构化环境中精确选择和定位无障碍水果方面表现出可观的能力,尤其能够避免被茎或电线遮挡的水果。这种方法为收获机器人的水果选择和定位提供了更可靠、实用的解决方案,同时也适用于其他水果和蔬菜,如甜椒和猕猴桃。