Zhu Shenrui, Liao Bin, Hua Yi, Zhang Chunlin, Wan Fangyi, Qing Xinlin
School of Aeronautics, Northwestern Polytechnical University, Xi'an 710072, PR China.
ISA Trans. 2023 Feb;133:1-12. doi: 10.1016/j.isatra.2022.07.016. Epub 2022 Jul 23.
Deep learning has become the prevailing trend of intelligent fault diagnosis for rotating machines. Compared to early-stage methods, deep learning methods use automatic feature extraction instead of manual feature design. However, conventional intelligent diagnosis models are trapped by a dilemma that simple models are unable to tackle difficult cases, while complicated models are likely to over-parameterize. In this paper, a transformer-based model, Periodic Representations for Transformers (PRT) is proposed. PRT uses a dense-overlapping split strategy to enhance the feature learning inside sequence patches. Combined with the inherent capability of capturing long range dependencies of transformer, and the further information extraction of class-attention, PRT has excellent feature extraction abilities and could capture characteristic features directly from raw vibration signals. Moreover, PRT adopts a two-stage positional encoding method to encode position information both among and inside patches, which could adapt to different input lengths. A novel inference method to use larger inference sample sizes is further proposed to improve the performance of PRT. The effectiveness of PRT is verified on two datasets, where it achieves comparable and even better accuracies than the benchmark and state-of-the-art methods. PRT has the least FLOPs among the best performing models and could be further improved by the inference strategy, reaching an accuracy near 100%.
深度学习已成为旋转机械智能故障诊断的主流趋势。与早期方法相比,深度学习方法采用自动特征提取而非手动特征设计。然而,传统的智能诊断模型面临一个两难困境:简单模型无法处理复杂情况,而复杂模型又可能出现过参数化问题。本文提出了一种基于Transformer的模型——Transformer的周期表示(PRT)。PRT采用密集重叠分割策略来增强序列块内的特征学习。结合Transformer固有的捕捉长距离依赖的能力以及类注意力的进一步信息提取,PRT具有出色的特征提取能力,能够直接从原始振动信号中捕捉特征。此外,PRT采用两阶段位置编码方法对块间和块内的位置信息进行编码,可适应不同的输入长度。还进一步提出了一种使用更大推理样本量的新颖推理方法来提高PRT的性能。在两个数据集上验证了PRT的有效性,其准确率与基准方法和当前最先进的方法相当,甚至更高。在性能最佳的模型中,PRT的浮点运算次数最少,并且通过推理策略可进一步提高,准确率接近100%。