• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

利用点云补全和多模态融合网络的采摘机器人对草莓重量进行分级估计。

Estimating strawberry weight for grading by picking robot with point cloud completion and multimodal fusion network.

作者信息

Chen Yiming, Wang Wei, Chen Junchao, Deng Jizhou, Xiang Yuanping, Qiao Bo, Zhu Xinghui, Li Changyun

机构信息

Hunan Agricultural University, Changsha, 410000, China.

Tencent Music Entertainment, Shenzhen, 518000, China.

出版信息

Sci Rep. 2025 Apr 2;15(1):11227. doi: 10.1038/s41598-025-92641-1.

DOI:10.1038/s41598-025-92641-1
PMID:40175474
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11965529/
Abstract

Strawberry grading by picking robots can eliminate the manual classification, reducing labor costs and minimizing the damage to the fruit. Strawberry size or weight is a key factor in grading, with accurate weight estimation being crucial for proper classification. In this paper, we collected 1521 sets of strawberry RGB-D images using a depth camera and manually measured the weight and size of the strawberries to construct a training dataset for the strawberry weight regression model. To address the issue of incomplete depth images caused by environmental interference with depth cameras, this study proposes a multimodal point cloud completion method specifically designed for symmetrical objects, leveraging RGB images to guide the completion of depth images in the same scene. The method follows a process of locating strawberry pixel regions, calculating centroid coordinates, determining the symmetry axis via PCA, and completing the depth image. Based on this approach, a multimodal fusion regression model for strawberry weight estimation, named MMF-Net, is developed. The model uses the completed point cloud and RGB image as inputs, and extracts features from the RGB image and point cloud by EfficientNet and PointNet, respectively. These features are then integrated at the feature level through gradient blending, realizing the combination of the strengths of both modalities. Using the Percent Correct Weight (PCW) metric as the evaluation standard, this study compares the performance of four traditional machine learning methods, Support Vector Regression (SVR), Multilayer Perceptron (MLP), Linear Regression, and Random Forest Regression, with four point cloud-based deep learning models, PointNet, PointNet++, PointMLP, and Point Cloud Transformer, as well as an image-based deep learning model, EfficientNet and ResNet, on single-modal datasets. The results indicate that among traditional machine learning methods, the SVR model achieved the best performance with an accuracy of 77.7% (PCW@0.2). Among deep learning methods, the image-based EfficientNet model obtained the highest accuracy, reaching 85% (PCW@0.2), while the PointNet + + model demonstrated the best performance among point cloud-based models, with an accuracy of 54.3% (PCW@0.2). The proposed multimodal fusion model, MMF-Net, achieved an accuracy of 87.66% (PCW@0.2), significantly outperforming both traditional machine learning methods and single-modal deep learning models in terms of precision.

摘要

通过采摘机器人对草莓进行分级可以消除人工分类,降低劳动力成本,并将对果实的损伤降至最低。草莓的大小或重量是分级的关键因素,准确的重量估计对于正确分类至关重要。在本文中,我们使用深度相机收集了1521组草莓RGB-D图像,并手动测量了草莓的重量和大小,以构建草莓重量回归模型的训练数据集。为了解决深度相机受环境干扰导致深度图像不完整的问题,本研究提出了一种专门针对对称物体的多模态点云补全方法,利用RGB图像来指导同一场景中深度图像的补全。该方法遵循定位草莓像素区域、计算质心坐标、通过主成分分析确定对称轴以及补全深度图像的过程。基于此方法,开发了一种用于草莓重量估计的多模态融合回归模型,名为MMF-Net。该模型将补全后的点云和RGB图像作为输入,分别通过EfficientNet和PointNet从RGB图像和点云中提取特征。然后,这些特征通过梯度融合在特征层面进行整合,实现了两种模态优势的结合。以正确重量百分比(PCW)指标作为评估标准,本研究在单模态数据集上比较了四种传统机器学习方法(支持向量回归(SVR)、多层感知器(MLP)、线性回归和随机森林回归)、四种基于点云的深度学习模型(PointNet、PointNet++、PointMLP和点云变换器)以及一种基于图像的深度学习模型(EfficientNet和ResNet)的性能。结果表明,在传统机器学习方法中,SVR模型表现最佳,准确率为77.7%(PCW@0.2)。在深度学习方法中,基于图像的EfficientNet模型准确率最高,达到85%(PCW@0.2),而PointNet++模型在基于点云的模型中表现最佳,准确率为54.3%(PCW@0.2)。所提出的多模态融合模型MMF-Net的准确率为87.66%(PCW@0.2),在精度方面显著优于传统机器学习方法和单模态深度学习模型。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/9dacf19a40a3/41598_2025_92641_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/a8513675b5e5/41598_2025_92641_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/b88dc59aa960/41598_2025_92641_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/3dbe77ddb8e7/41598_2025_92641_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/6ef3b64655b6/41598_2025_92641_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/7c1636e505bf/41598_2025_92641_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/11ef8382372a/41598_2025_92641_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/abcb029fa3cd/41598_2025_92641_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/172f9f50cd14/41598_2025_92641_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/06fdd3d77b20/41598_2025_92641_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/29b786f866d3/41598_2025_92641_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/e874ae857b14/41598_2025_92641_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/27e44e17244f/41598_2025_92641_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/1d5cf4bd79d7/41598_2025_92641_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/8d1083cb07d1/41598_2025_92641_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/9dacf19a40a3/41598_2025_92641_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/a8513675b5e5/41598_2025_92641_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/b88dc59aa960/41598_2025_92641_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/3dbe77ddb8e7/41598_2025_92641_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/6ef3b64655b6/41598_2025_92641_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/7c1636e505bf/41598_2025_92641_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/11ef8382372a/41598_2025_92641_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/abcb029fa3cd/41598_2025_92641_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/172f9f50cd14/41598_2025_92641_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/06fdd3d77b20/41598_2025_92641_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/29b786f866d3/41598_2025_92641_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/e874ae857b14/41598_2025_92641_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/27e44e17244f/41598_2025_92641_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/1d5cf4bd79d7/41598_2025_92641_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/8d1083cb07d1/41598_2025_92641_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/62d4/11965529/9dacf19a40a3/41598_2025_92641_Fig14_HTML.jpg

相似文献

1
Estimating strawberry weight for grading by picking robot with point cloud completion and multimodal fusion network.利用点云补全和多模态融合网络的采摘机器人对草莓重量进行分级估计。
Sci Rep. 2025 Apr 2;15(1):11227. doi: 10.1038/s41598-025-92641-1.
2
Brain tumor segmentation and detection in MRI using convolutional neural networks and VGG16.使用卷积神经网络和VGG16在磁共振成像(MRI)中进行脑肿瘤分割与检测
Cancer Biomark. 2025 Mar;42(3):18758592241311184. doi: 10.1177/18758592241311184. Epub 2025 Apr 4.
3
Self-Attention Mechanism-Based Head Pose Estimation Network with Fusion of Point Cloud and Image Features.基于自注意力机制的融合点云与图像特征的头部姿态估计网络
Sensors (Basel). 2023 Dec 18;23(24):9894. doi: 10.3390/s23249894.
4
Integrating PointNet-Based Model and Machine Learning Algorithms for Classification of Rupture Status of IAs.基于点云网络的模型与机器学习算法相结合用于颅内动脉瘤破裂状态的分类
Bioengineering (Basel). 2024 Jun 28;11(7):660. doi: 10.3390/bioengineering11070660.
5
A medical image classification method based on self-regularized adversarial learning.基于自正则化对抗学习的医学图像分类方法。
Med Phys. 2024 Nov;51(11):8232-8246. doi: 10.1002/mp.17320. Epub 2024 Jul 30.
6
Comparison of Graph Fitting and Sparse Deep Learning Model for Robot Pose Estimation.图形拟合与稀疏深度学习模型在机器人位姿估计中的比较。
Sensors (Basel). 2022 Aug 29;22(17):6518. doi: 10.3390/s22176518.
7
RAE-Net: a multi-modal neural network based on feature fusion and evidential deep learning algorithm in predicting breast cancer subtypes on DCE-MRI.RAE-Net:一种基于特征融合和证据深度学习算法的多模态神经网络,用于在动态对比增强磁共振成像(DCE-MRI)上预测乳腺癌亚型
Biomed Phys Eng Express. 2025 Feb 25;11(2). doi: 10.1088/2057-1976/adb494.
8
Point Cloud Hand-Object Segmentation Using Multimodal Imaging with Thermal and Color Data for Safe Robotic Object Handover.点云手-物体分割使用多模态成像与热和颜色数据进行安全机器人物体交接。
Sensors (Basel). 2021 Aug 23;21(16):5676. doi: 10.3390/s21165676.
9
Towards robust multimodal ultrasound classification for liver tumor diagnosis: A generative approach to modality missingness.迈向用于肝肿瘤诊断的稳健多模态超声分类:一种处理模态缺失的生成方法。
Comput Methods Programs Biomed. 2025 Jun;265:108759. doi: 10.1016/j.cmpb.2025.108759. Epub 2025 Mar 30.
10
Multi-objective RGB-D fusion network for non-destructive strawberry trait assessment.用于无损草莓性状评估的多目标RGB-D融合网络。
Front Plant Sci. 2025 Mar 12;16:1564301. doi: 10.3389/fpls.2025.1564301. eCollection 2025.

本文引用的文献

1
Sensory manipulation as a countermeasure to robot teleoperation delays: system and evidence.感觉操纵作为机器人遥操作延迟的对策:系统和证据。
Sci Rep. 2024 Feb 21;14(1):4333. doi: 10.1038/s41598-024-54734-1.
2
Multimodal Learning With Transformers: A Survey.基于Transformer的多模态学习:一项综述。
IEEE Trans Pattern Anal Mach Intell. 2023 Oct;45(10):12113-12132. doi: 10.1109/TPAMI.2023.3275156. Epub 2023 Sep 5.
3
Deep Learning for 3D Point Clouds: A Survey.用于三维点云的深度学习:综述
IEEE Trans Pattern Anal Mach Intell. 2021 Dec;43(12):4338-4364. doi: 10.1109/TPAMI.2020.3005434. Epub 2021 Nov 3.