Cui Jie, Zhang Lilian, Gao Lutao, Bai Chunhui, Yang Linnan
College of Big Data, Yunnan Agricultural University, Kunming, China.
Yunnan Engineering Technology Research Center of Agricultural Big Data, Kunming, China.
Front Plant Sci. 2025 Sep 3;16:1607205. doi: 10.3389/fpls.2025.1607205. eCollection 2025.
Accurate identification of cherry maturity and precise detection of harvestable cherry contours are essential for the development of cherry-picking robots. However, occlusion, lighting variation, and blurriness in natural orchard environments present significant challenges for real-time semantic segmentation.
To address these issues, we propose a machine vision approach based on the PIDNet real-time semantic segmentation framework. Redundant loss functions and residual blocks were removed to improve efficiency, and SwiftFormer-XS was adopted as a lightweight backbone to reduce complexity and accelerate inference. A Swift Rep-parameterized Hybrid (SwiftRep-Hybrid) module was designed to integrate local convolutional features with global Transformer-based context, while a Light Fusion Enhance (LFE) module with bidirectional enhancement and bilinear interpolation was introduced to strengthen feature representation. Additionally, a post-processing module was employed to refine class determination and visualize maturity classification results.
The proposed model achieved a mean Intersection over Union (MIoU) of 72.2% and a pixel accuracy (PA) of 99.82%, surpassing state-of-the-art real-time segmentation models such as PIDNet, DDRNet, and Fast-SCNN. Furthermore, when deployed on an embedded Jetson TX2 platform, the model maintained competitive inference speed and accuracy, confirming its feasibility for real-world robotic harvesting applications.
This study presents a lightweight, accurate, and efficient solution for cherry maturity recognition and contour detection in robotic harvesting. The proposed approach enhances robustness under challenging agricultural conditions and shows strong potential for deployment in intelligent harvesting systems, contributing to the advancement of precision agriculture technologies.
准确识别樱桃成熟度并精确检测可收获樱桃的轮廓对于樱桃采摘机器人的发展至关重要。然而,自然果园环境中的遮挡、光照变化和模糊性给实时语义分割带来了重大挑战。
为了解决这些问题,我们提出了一种基于PIDNet实时语义分割框架的机器视觉方法。去除了冗余损失函数和残差块以提高效率,并采用SwiftFormer-XS作为轻量级主干来降低复杂度并加速推理。设计了一个Swift重参数化混合(SwiftRep-Hybrid)模块,将局部卷积特征与基于全局Transformer的上下文集成,同时引入了一个具有双向增强和双线性插值的轻量级融合增强(LFE)模块来加强特征表示。此外,还采用了一个后处理模块来细化类别判定并可视化成熟度分类结果。
所提出的模型实现了72.2%的平均交并比(MIoU)和99.82%的像素准确率(PA),超过了PIDNet、DDRNet和Fast-SCNN等最先进的实时分割模型。此外,当部署在嵌入式Jetson TX2平台上时,该模型保持了有竞争力的推理速度和准确率,证实了其在实际机器人收获应用中的可行性。
本研究为机器人收获中的樱桃成熟度识别和轮廓检测提出了一种轻量级、准确且高效的解决方案。所提出的方法在具有挑战性的农业条件下增强了鲁棒性,并在智能收获系统中显示出强大的部署潜力,有助于推动精准农业技术的发展。