• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于双目视觉和Mask R-CNN的钢卷卷眼姿态检测

Steel Roll Eye Pose Detection Based on Binocular Vision and Mask R-CNN.

作者信息

Su Xuwu, Wang Jie, Wang Yifan, Zhang Daode

机构信息

School of Mechanical Engineering, Hubei University of Technology, Wuhan 430000, China.

出版信息

Sensors (Basel). 2025 Mar 14;25(6):1805. doi: 10.3390/s25061805.

DOI:10.3390/s25061805
PMID:40292976
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11946423/
Abstract

To achieve automation at the inner corner guard installation station in a steel coil packaging production line and enable automatic docking and installation of the inner corner guard after eye position detection, this paper proposes a binocular vision method based on deep learning for eye position detection of steel coil rolls. The core of the method involves using the Mask R-CNN algorithm within a deep-learning framework to identify the target region and obtain a mask image of the steel coil end face. Subsequently, the binarized image of the steel coil end face was processed using the RGB vector space image segmentation method. The target feature pixel points were then extracted using Sobel edges, and the parameters were fitted by the least-squares method to obtain the deflection angle and the horizontal and vertical coordinates of the center point in the image coordinate system. Through the ellipse parameter extraction experiment, the maximum deviations in the pixel coordinate system for the center point in the u and v directions were 0.49 and 0.47, respectively. The maximum error in the deflection angle was 0.45°. In the steel coil roll eye position detection experiments, the maximum deviations for the pitch angle, deflection angle, and centroid coordinates were 2.17°, 2.24°, 3.53 mm, 4.05 mm, and 4.67 mm, respectively, all of which met the actual installation requirements. The proposed method demonstrates strong operability in practical applications, and the steel coil end face position solving approach significantly enhances work efficiency, reduces labor costs, and ensures adequate detection accuracy.

摘要

为实现钢卷包装生产线内护角安装工位的自动化,以及在检测到钢卷卷眼位置后实现内护角的自动对接与安装,本文提出一种基于深度学习的双目视觉方法用于钢卷卷眼位置检测。该方法的核心是在深度学习框架内使用Mask R-CNN算法识别目标区域并获取钢卷端面的掩膜图像。随后,采用RGB向量空间图像分割方法对钢卷端面的二值化图像进行处理。接着利用Sobel边缘提取目标特征像素点,并通过最小二乘法拟合参数,以获取图像坐标系中的偏转角以及中心点的水平和垂直坐标。通过椭圆参数提取实验,像素坐标系中中心点在u和v方向上的最大偏差分别为0.49和0.47。偏转角的最大误差为0.45°。在钢卷卷眼位置检测实验中,俯仰角、偏转角和质心坐标的最大偏差分别为2.17°、2.24°、3.53毫米、4.05毫米和4.67毫米,均满足实际安装要求。所提方法在实际应用中具有较强的可操作性,钢卷端面位置求解方法显著提高了工作效率、降低了人工成本并确保了足够的检测精度。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/1f2fec04d9f2/sensors-25-01805-g023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/dad96dd6f275/sensors-25-01805-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/79598995dbf6/sensors-25-01805-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/ca9f37c3199d/sensors-25-01805-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/14ae992baa41/sensors-25-01805-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/f9f0842fa8e7/sensors-25-01805-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/0283749d3284/sensors-25-01805-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/e6eb298a2caf/sensors-25-01805-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/fadb1a90d7cc/sensors-25-01805-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/4f95360e32be/sensors-25-01805-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/1e54639422e2/sensors-25-01805-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/fd31b9217248/sensors-25-01805-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/7fcdbf143b4e/sensors-25-01805-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/093f5479b8b5/sensors-25-01805-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/222b7512d1eb/sensors-25-01805-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/68110966dc12/sensors-25-01805-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/5e9efcc979fb/sensors-25-01805-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/2f9d305a9f38/sensors-25-01805-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/7b36bf817f6a/sensors-25-01805-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/b37d4cfb2175/sensors-25-01805-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/694db61cbef2/sensors-25-01805-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/12bf249f7cd3/sensors-25-01805-g021.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/b6f14d7a7558/sensors-25-01805-g022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/1f2fec04d9f2/sensors-25-01805-g023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/dad96dd6f275/sensors-25-01805-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/79598995dbf6/sensors-25-01805-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/ca9f37c3199d/sensors-25-01805-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/14ae992baa41/sensors-25-01805-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/f9f0842fa8e7/sensors-25-01805-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/0283749d3284/sensors-25-01805-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/e6eb298a2caf/sensors-25-01805-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/fadb1a90d7cc/sensors-25-01805-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/4f95360e32be/sensors-25-01805-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/1e54639422e2/sensors-25-01805-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/fd31b9217248/sensors-25-01805-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/7fcdbf143b4e/sensors-25-01805-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/093f5479b8b5/sensors-25-01805-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/222b7512d1eb/sensors-25-01805-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/68110966dc12/sensors-25-01805-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/5e9efcc979fb/sensors-25-01805-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/2f9d305a9f38/sensors-25-01805-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/7b36bf817f6a/sensors-25-01805-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/b37d4cfb2175/sensors-25-01805-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/694db61cbef2/sensors-25-01805-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/12bf249f7cd3/sensors-25-01805-g021.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/b6f14d7a7558/sensors-25-01805-g022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d79/11946423/1f2fec04d9f2/sensors-25-01805-g023.jpg

相似文献

1
Steel Roll Eye Pose Detection Based on Binocular Vision and Mask R-CNN.基于双目视觉和Mask R-CNN的钢卷卷眼姿态检测
Sensors (Basel). 2025 Mar 14;25(6):1805. doi: 10.3390/s25061805.
2
Manipulation Planning for Object Re-Orientation Based on Semantic Segmentation Keypoint Detection.基于语义分割关键点检测的物体重新定向操作规划
Sensors (Basel). 2021 Mar 24;21(7):2280. doi: 10.3390/s21072280.
3
An Efficient Building Extraction Method from High Spatial Resolution Remote Sensing Images Based on Improved Mask R-CNN.基于改进的 Mask R-CNN 的高空间分辨率遥感图像高效建筑物提取方法。
Sensors (Basel). 2020 Mar 6;20(5):1465. doi: 10.3390/s20051465.
4
Detection, segmentation, and 3D pose estimation of surgical tools using convolutional neural networks and algebraic geometry.使用卷积神经网络和代数几何进行手术工具的检测、分割和三维姿态估计。
Med Image Anal. 2021 May;70:101994. doi: 10.1016/j.media.2021.101994. Epub 2021 Feb 7.
5
Vision Measurement of Gear Pitting Under Different Scenes by Deep Mask R-CNN.基于 Deep Mask R-CNN 的不同场景下齿轮点蚀的视觉测量。
Sensors (Basel). 2020 Aug 1;20(15):4298. doi: 10.3390/s20154298.
6
A Fast Binocular Localisation Method for AUV Docking.一种用于自主水下航行器对接的快速双目定位方法。
Sensors (Basel). 2019 Apr 11;19(7):1735. doi: 10.3390/s19071735.
7
Development and evaluation of ultrasound image tracking technology based on Mask R-CNN applied to respiratory motion compensation system.基于Mask R-CNN的超声图像跟踪技术在呼吸运动补偿系统中的开发与评估
Quant Imaging Med Surg. 2023 Oct 1;13(10):6827-6839. doi: 10.21037/qims-23-23. Epub 2023 Sep 5.
8
Fully automatic estimation of pelvic sagittal inclination from anterior-posterior radiography image using deep learning framework.利用深度学习框架从前后位 X 射线图像全自动估计骨盆矢状倾斜度。
Comput Methods Programs Biomed. 2020 Feb;184:105282. doi: 10.1016/j.cmpb.2019.105282. Epub 2019 Dec 23.
9
A Multi-Step CNN-Based Estimation of Aircraft Landing Gear Angles.基于多步卷积神经网络的飞机起落架角度估计
Sensors (Basel). 2021 Dec 17;21(24):8440. doi: 10.3390/s21248440.
10
Binocular vision calibration method for a long-wavelength infrared camera and a visible spectrum camera with different resolutions.用于不同分辨率的长波红外相机和可见光谱相机的双目视觉校准方法。
Opt Express. 2021 Feb 1;29(3):3855-3872. doi: 10.1364/OE.405600.

本文引用的文献

1
Crack Detection and Comparison Study Based on Faster R-CNN and Mask R-CNN.基于 Faster R-CNN 和 Mask R-CNN 的裂纹检测与比较研究。
Sensors (Basel). 2022 Feb 5;22(3):1215. doi: 10.3390/s22031215.
2
Sharp U-Net: Depthwise convolutional network for biomedical image segmentation.Sharp U-Net:用于生物医学图像分割的深度卷积网络。
Comput Biol Med. 2021 Sep;136:104699. doi: 10.1016/j.compbiomed.2021.104699. Epub 2021 Jul 29.
3
Intelligent Perception System of Robot Visual Servo for Complex Industrial Environment.机器人视觉伺服智能感知系统在复杂工业环境中的应用。
Sensors (Basel). 2020 Dec 11;20(24):7121. doi: 10.3390/s20247121.
4
Monocular Vision-Based Pose Determination in Close Proximity for Low Impact Docking.基于单目视觉的近距离低冲击对接姿态确定
Sensors (Basel). 2019 Jul 24;19(15):3261. doi: 10.3390/s19153261.