Zhao Delong, Kong Feifei, Lv Nengbin, Xu Zhangmao, Du Fuzhou
School of Mechanical Engineering and Automation, Beihang University, 37 College Road, Haidian District, Beijing 100191, China.
Sensors (Basel). 2024 Jun 25;24(13):4120. doi: 10.3390/s24134120.
The industrial manufacturing model is undergoing a transformation from a product-centric model to a customer-centric one. Driven by customized requirements, the complexity of products and the requirements for quality have increased, which pose a challenge to the applicability of traditional machine vision technology. Extensive research demonstrates the effectiveness of AI-based learning and image processing on specific objects or tasks, but few publications focus on the composite task of the integrated product, the traceability and improvability of methods, as well as the extraction and communication of knowledge between different scenarios or tasks. To address this problem, this paper proposes a common, knowledge-driven, generic vision inspection framework, targeted for standardizing product inspection into a process of information decoupling and adaptive metrics. Task-related object perception is planned into a multi-granularity and multi-pattern progressive alignment based on industry knowledge and structured tasks. Inspection is abstracted as a reconfigurable process of multi-sub-pattern space combination mapping and difference metric under appropriate high-level strategies and experiences. Finally, strategies for knowledge improvement and accumulation based on historical data are presented. The experiment demonstrates the process of generating a detection pipeline for complex products and continuously improving it through failure tracing and knowledge improvement. Compared to the (1.767°, 69.802 mm) and 0.883 obtained by state-of-the-art deep learning methods, the generated pipeline achieves a pose estimation ranging from (2.771°, 153.584 mm) to (1.034°, 52.308 mm) and a detection rate ranging from 0.462 to 0.927. Through verification of other imaging methods and industrial tasks, we prove that the key to adaptability lies in the mining of inherent commonalities of knowledge, multi-dimensional accumulation, and reapplication.
工业制造模式正在从以产品为中心的模式向以客户为中心的模式转变。在定制需求的驱动下,产品的复杂性和对质量的要求都有所提高,这对传统机器视觉技术的适用性提出了挑战。大量研究表明基于人工智能的学习和图像处理在特定对象或任务上是有效的,但很少有出版物关注集成产品的复合任务、方法的可追溯性和可改进性,以及不同场景或任务之间知识的提取和交流。为了解决这个问题,本文提出了一个通用的、知识驱动的、通用视觉检测框架,旨在将产品检测标准化为一个信息解耦和自适应度量的过程。基于行业知识和结构化任务,将与任务相关的目标感知规划为多粒度、多模式的渐进对齐。在适当的高级策略和经验下,将检测抽象为多子模式空间组合映射和差异度量的可重构过程。最后,提出了基于历史数据的知识改进和积累策略。实验展示了为复杂产品生成检测管道并通过故障追踪和知识改进不断优化它的过程。与最先进的深度学习方法获得的(1.767°, 69.802 mm)和0.883相比,生成的管道实现了从(2.771°, 153.584 mm)到(1.034°, 52.308 mm)的姿态估计和从0.462到0.927的检测率。通过对其他成像方法和工业任务的验证,我们证明了适应性的关键在于挖掘知识的内在共性、多维度积累和重新应用。