• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种基于视觉的自动仓储与检索系统中堆叠货物位置检测方法。

A Vision-Based Method for Detecting the Position of Stacked Goods in Automated Storage and Retrieval Systems.

作者信息

Chen Chuanjun, Liu Junjie, Yin Haonan, Huang Biqing

机构信息

Department of Automation, Tsinghua University, Beijing 100084, China.

BZS (Beijing) Technology Development Co., Ltd., No.1 Jiaochangkou, Deshengmenwai, Beijing 100120, China.

出版信息

Sensors (Basel). 2025 Apr 21;25(8):2623. doi: 10.3390/s25082623.

DOI:10.3390/s25082623
PMID:40285312
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12031210/
Abstract

Automated storage and retrieval systems (AS/RS) play a crucial role in modern logistics, yet effectively monitoring cargo stacking patterns remains challenging. While computer vision and deep learning offer promising solutions, existing methods struggle to balance detection accuracy, computational efficiency, and environmental adaptability. This paper proposes a novel machine vision-based detection algorithm that integrates a pallet surface object detection network (STEGNet) with a box edge detection algorithm. STEGNet's core innovation is the Efficient Gated Pyramid Feature Network (EG-FPN), which integrates a Gated Feature Fusion module and a Lightweight Attention Mechanism to optimize feature extraction and fusion. In addition, we introduce a geometric constraint method for box edge detection and employ a Perspective-n-Point (PnP)-based 2D-to-3D transformation approach for precise pose estimation. Experimental results show that STEGNet achieves 93.49% mAP on our proposed GY Warehouse Box View 4-Dimension (GY-WSBW-4D) dataset and 83.2% mAP on the WSGID-B dataset, surpassing existing benchmarks. The lightweight variant maintains competitive accuracy while reducing the model size by 34% and increasing the inference speed by 68%. In practical applications, the system achieves pose estimation with a Mean Absolute Error within 4 cm and a Rotation Angle Error below 2°, demonstrating robust performance in complex warehouse environments. This research provides a reliable solution for automated cargo stack monitoring in modern logistics systems.

摘要

自动存储和检索系统(AS/RS)在现代物流中起着至关重要的作用,但有效地监测货物堆叠模式仍然具有挑战性。虽然计算机视觉和深度学习提供了有前景的解决方案,但现有方法难以在检测精度、计算效率和环境适应性之间取得平衡。本文提出了一种基于机器视觉的新型检测算法,该算法将托盘表面物体检测网络(STEGNet)与箱边检测算法相结合。STEGNet的核心创新在于高效门控金字塔特征网络(EG-FPN),它集成了门控特征融合模块和轻量级注意力机制,以优化特征提取和融合。此外,我们引入了一种用于箱边检测的几何约束方法,并采用基于透视n点(PnP)的二维到三维变换方法进行精确的姿态估计。实验结果表明,STEGNet在我们提出的GY仓库箱视图四维(GY-WSBW-4D)数据集上实现了93.49%的平均精度均值(mAP),在WSGID-B数据集上实现了83.2%的mAP,超过了现有基准。轻量级变体在保持竞争力的精度的同时,将模型大小减少了34%,推理速度提高了68%。在实际应用中,该系统实现了平均绝对误差在4厘米以内、旋转角度误差低于2°的姿态估计,在复杂的仓库环境中表现出强大的性能。这项研究为现代物流系统中的自动货物堆叠监测提供了可靠的解决方案。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/21c4a03cee78/sensors-25-02623-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/fe5e75ff3f58/sensors-25-02623-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/00bc4fde5215/sensors-25-02623-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/2cf4e2c4f64d/sensors-25-02623-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/07720063b347/sensors-25-02623-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/0dcf4134fbd3/sensors-25-02623-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/6bee63e24caf/sensors-25-02623-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/e52a4c24607a/sensors-25-02623-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/e57be4af7dbb/sensors-25-02623-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/12a551ca4ecb/sensors-25-02623-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/4ab023da5674/sensors-25-02623-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/b989d33b7594/sensors-25-02623-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/8f340e813e92/sensors-25-02623-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/08fed3b9a867/sensors-25-02623-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/1b5d94c5122d/sensors-25-02623-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/e5b3ab76288c/sensors-25-02623-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/8c80adfda00a/sensors-25-02623-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/c31c66b865cb/sensors-25-02623-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/21c4a03cee78/sensors-25-02623-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/fe5e75ff3f58/sensors-25-02623-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/00bc4fde5215/sensors-25-02623-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/2cf4e2c4f64d/sensors-25-02623-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/07720063b347/sensors-25-02623-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/0dcf4134fbd3/sensors-25-02623-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/6bee63e24caf/sensors-25-02623-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/e52a4c24607a/sensors-25-02623-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/e57be4af7dbb/sensors-25-02623-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/12a551ca4ecb/sensors-25-02623-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/4ab023da5674/sensors-25-02623-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/b989d33b7594/sensors-25-02623-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/8f340e813e92/sensors-25-02623-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/08fed3b9a867/sensors-25-02623-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/1b5d94c5122d/sensors-25-02623-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/e5b3ab76288c/sensors-25-02623-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/8c80adfda00a/sensors-25-02623-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/c31c66b865cb/sensors-25-02623-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e676/12031210/21c4a03cee78/sensors-25-02623-g018.jpg

相似文献

1
A Vision-Based Method for Detecting the Position of Stacked Goods in Automated Storage and Retrieval Systems.一种基于视觉的自动仓储与检索系统中堆叠货物位置检测方法。
Sensors (Basel). 2025 Apr 21;25(8):2623. doi: 10.3390/s25082623.
2
Detection, segmentation, and 3D pose estimation of surgical tools using convolutional neural networks and algebraic geometry.使用卷积神经网络和代数几何进行手术工具的检测、分割和三维姿态估计。
Med Image Anal. 2021 May;70:101994. doi: 10.1016/j.media.2021.101994. Epub 2021 Feb 7.
3
Brain tumor segmentation and detection in MRI using convolutional neural networks and VGG16.使用卷积神经网络和VGG16在磁共振成像(MRI)中进行脑肿瘤分割与检测
Cancer Biomark. 2025 Mar;42(3):18758592241311184. doi: 10.1177/18758592241311184. Epub 2025 Apr 4.
4
YOLOv8-seg-CP: a lightweight instance segmentation algorithm for chip pad based on improved YOLOv8-seg model.YOLOv8-seg-CP:一种基于改进的YOLOv8-seg模型的用于芯片焊盘的轻量级实例分割算法。
Sci Rep. 2024 Nov 12;14(1):27716. doi: 10.1038/s41598-024-78578-x.
5
Lightweight medical image segmentation network with multi-scale feature-guided fusion.轻量级医疗图像分割网络,具有多尺度特征引导融合。
Comput Biol Med. 2024 Nov;182:109204. doi: 10.1016/j.compbiomed.2024.109204. Epub 2024 Oct 3.
6
A vision model for automated frozen tuna processing.一种用于自动化冷冻金枪鱼加工的视觉模型。
Sci Rep. 2025 Jan 25;15(1):3216. doi: 10.1038/s41598-025-87339-3.
7
SF-YOLOv5: A Lightweight Small Object Detection Algorithm Based on Improved Feature Fusion Mode.SF-YOLOv5:一种基于改进特征融合模式的轻量级小目标检测算法。
Sensors (Basel). 2022 Aug 4;22(15):5817. doi: 10.3390/s22155817.
8
Machine Vision-Assisted Design of End Effector Pose in Robotic Mixed Depalletizing of Heterogeneous Cargo.异构货物机器人混合卸托盘过程中末端执行器姿态的机器视觉辅助设计
Sensors (Basel). 2025 Feb 13;25(4):1137. doi: 10.3390/s25041137.
9
DAHD-YOLO: A New High Robustness and Real-Time Method for Smoking Detection.DAHD-YOLO:一种用于吸烟检测的新型高鲁棒性实时方法。
Sensors (Basel). 2025 Feb 26;25(5):1433. doi: 10.3390/s25051433.
10
A lightweight and efficient model for grape bunch detection and biophysical anomaly assessment in complex environments based on YOLOv8s.一种基于YOLOv8s的轻量级高效模型,用于复杂环境下葡萄串检测和生物物理异常评估。
Front Plant Sci. 2024 Aug 6;15:1395796. doi: 10.3389/fpls.2024.1395796. eCollection 2024.

本文引用的文献

1
An Improved SAC-IA Algorithm Based on Voxel Nearest Neighbor Search.基于体素最近邻搜索的改进 SAC-IA 算法。
Crit Rev Biomed Eng. 2022;50(1):35-46. doi: 10.1615/CritRevBiomedEng.2022044053.
2
A Survey on Vision Transformer.视觉Transformer综述
IEEE Trans Pattern Anal Mach Intell. 2023 Jan;45(1):87-110. doi: 10.1109/TPAMI.2022.3152247. Epub 2022 Dec 5.
3
Integrating Sensor Models in Deep Learning Boosts Performance: Application to Monocular Depth Estimation in Warehouse Automation.深度学习中整合传感器模型可提高性能:在仓库自动化中单目深度估计中的应用。
Sensors (Basel). 2021 Feb 19;21(4):1437. doi: 10.3390/s21041437.
4
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.
5
SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.SegNet:一种用于图像分割的深度卷积编解码器架构。
IEEE Trans Pattern Anal Mach Intell. 2017 Dec;39(12):2481-2495. doi: 10.1109/TPAMI.2016.2644615. Epub 2017 Jan 2.
6
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
7
A Robust O(n) Solution to the Perspective-n-Point Problem.一种鲁棒的 O(n) 视角 n 点问题解决方案。
IEEE Trans Pattern Anal Mach Intell. 2012 Jul;34(7):1444-50. doi: 10.1109/TPAMI.2012.41. Epub 2012 Jan 31.
8
SIFT: Predicting amino acid changes that affect protein function.SIFT:预测影响蛋白质功能的氨基酸变化。
Nucleic Acids Res. 2003 Jul 1;31(13):3812-4. doi: 10.1093/nar/gkg509.