Aerospace Information Research Institute, Chinese Academy of Sciences, No. 9 Dengzhuang South Road, Haidian District, Beijing 100094, China.
School of Optoelectronics, University of Chinese of Academy Sciences, No. 19(A) Yuquan Road, Shijingshan District, Beijing 100049, China.
Sensors (Basel). 2022 Sep 1;22(17):6618. doi: 10.3390/s22176618.
To address the problems of convolutional neural networks (CNNs) consuming more hardware resources (such as DSPs and RAMs on FPGAs) and their accuracy, efficiency, and resources being difficult to balance, meaning they cannot meet the requirements of industrial applications, we proposed an innovative low-bit power-of-two quantization method: the global sign-based network quantization (GSNQ). This method involves designing different quantization ranges according to the sign of the weights, which can provide a larger quantization-value range. Combined with the fine-grained and multi-scale global retraining method proposed in this paper, the accuracy loss of low-bit quantization can be effectively reduced. We also proposed a novel convolutional algorithm using shift operations to replace multiplication to help to deploy the GSNQ quantized models on FPGAs. Quantization comparison experiments performed on LeNet-5, AlexNet, VGG-Net, ResNet, and GoogLeNet showed that GSNQ has higher accuracy than most existing methods and achieves "lossless" quantization (i.e., the accuracy of the quantized CNN model is higher than the baseline) at low-bit quantization in most cases. FPGA comparison experiments showed that our convolutional algorithm does not occupy on-chip DSPs, and it also has a low comprehensive occupancy in terms of on-chip LUTs and FFs, which can effectively improve the computational parallelism, and this proves that GSNQ has good hardware-adaptation capability. This study provides theoretical and experimental support for the industrial application of CNNs.
为了解决卷积神经网络(CNN)消耗更多硬件资源(如 FPGA 上的 DSP 和 RAM)的问题,以及其准确性、效率和资源难以平衡的问题,即无法满足工业应用的要求,我们提出了一种创新的低比特 2 的幂次量化方法:基于全局符号的网络量化(GSNQ)。该方法根据权重的符号设计不同的量化范围,可以提供更大的量化值范围。结合本文提出的细粒度和多尺度全局再训练方法,可以有效降低低比特量化的精度损失。我们还提出了一种使用移位操作代替乘法的新卷积算法,以帮助在 FPGA 上部署 GSNQ 量化模型。在 LeNet-5、AlexNet、VGG-Net、ResNet 和 GoogLeNet 上进行的量化比较实验表明,GSNQ 在大多数情况下比大多数现有方法具有更高的精度,并在低比特量化时实现了“无损”量化(即量化 CNN 模型的精度高于基线)。FPGA 比较实验表明,我们的卷积算法不占用片上 DSP,并且在片上 LUT 和 FF 方面的综合占用率也很低,可以有效提高计算并行度,这证明了 GSNQ 具有良好的硬件适应性。本研究为 CNN 的工业应用提供了理论和实验支持。