• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

无梯度比特分配在混合精度神经网络中的应用。

GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks.

机构信息

Department of Computer Science, Ben-Gurion University, Beer Sheva 8410501, Israel.

出版信息

Sensors (Basel). 2022 Dec 13;22(24):9772. doi: 10.3390/s22249772.

DOI:10.3390/s22249772
PMID:36560141
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9787339/
Abstract

Quantized neural networks (QNNs) are among the main approaches for deploying deep neural networks on low-resource edge devices. Training QNNs using different levels of precision throughout the network (mixed-precision quantization) typically achieves superior trade-offs between performance and computational load. However, optimizing the different precision levels of QNNs can be complicated, as the values of the bit allocations are discrete and difficult to differentiate for. Moreover, adequately accounting for the dependencies between the bit allocation of different layers is not straightforward. To meet these challenges, in this work, we propose GradFreeBits: a novel joint optimization scheme for training mixed-precision QNNs, which alternates between gradient-based optimization for the weights and gradient-free optimization for the bit allocation. Our method achieves a better or on par performance with the current state-of-the-art low-precision classification networks on CIFAR10/100 and ImageNet, semantic segmentation networks on Cityscapes, and several graph neural networks benchmarks. Furthermore, our approach can be extended to a variety of other applications involving neural networks used in conjunction with parameters that are difficult to optimize for.

摘要

量化神经网络 (QNN) 是在资源有限的边缘设备上部署深度神经网络的主要方法之一。在整个网络中使用不同精度级别(混合精度量化)训练 QNN 通常可以在性能和计算负载之间实现更好的权衡。然而,优化 QNN 的不同精度级别可能很复杂,因为位分配的值是离散的,并且难以区分。此外,充分考虑不同层之间的位分配之间的依赖关系并不简单。为了应对这些挑战,在这项工作中,我们提出了 GradFreeBits:一种用于训练混合精度 QNN 的新联合优化方案,它交替使用基于梯度的权重优化和基于梯度的位分配优化。我们的方法在 CIFAR10/100 和 ImageNet 上的分类网络、Cityscapes 上的语义分割网络以及几个图神经网络基准测试中实现了与当前最先进的低精度分类网络相当或更好的性能。此外,我们的方法可以扩展到其他各种涉及与难以优化的参数一起使用的神经网络的应用。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/df5b8ddf3b61/sensors-22-09772-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/5eb62913fc65/sensors-22-09772-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/2f017202620b/sensors-22-09772-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/8185b4b8da30/sensors-22-09772-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/76f6ef187bbd/sensors-22-09772-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/b2d41c33ec9a/sensors-22-09772-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/df5b8ddf3b61/sensors-22-09772-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/5eb62913fc65/sensors-22-09772-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/2f017202620b/sensors-22-09772-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/8185b4b8da30/sensors-22-09772-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/76f6ef187bbd/sensors-22-09772-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/b2d41c33ec9a/sensors-22-09772-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da9c/9787339/df5b8ddf3b61/sensors-22-09772-g006.jpg

相似文献

1
GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks.无梯度比特分配在混合精度神经网络中的应用。
Sensors (Basel). 2022 Dec 13;22(24):9772. doi: 10.3390/s22249772.
2
Hessian-based mixed-precision quantization with transition aware training for neural networks.基于黑塞矩阵的神经网络混合精度量化与过渡感知训练
Neural Netw. 2025 Feb;182:106910. doi: 10.1016/j.neunet.2024.106910. Epub 2024 Nov 16.
3
Deployable mixed-precision quantization with co-learning and one-time search.采用协同学习和一次性搜索的可部署混合精度量化
Neural Netw. 2025 Jan;181:106812. doi: 10.1016/j.neunet.2024.106812. Epub 2024 Oct 18.
4
Exploiting Retraining-Based Mixed-Precision Quantization for Low-Cost DNN Accelerator Design.利用基于再训练的混合精度量化进行低成本深度神经网络加速器设计。
IEEE Trans Neural Netw Learn Syst. 2021 Jul;32(7):2925-2938. doi: 10.1109/TNNLS.2020.3008996. Epub 2021 Jul 6.
5
MedQ: Lossless ultra-low-bit neural network quantization for medical image segmentation.MedQ:用于医学图像分割的无损超低比特神经网络量化。
Med Image Anal. 2021 Oct;73:102200. doi: 10.1016/j.media.2021.102200. Epub 2021 Aug 2.
6
Effective Training of Convolutional Neural Networks With Low-Bitwidth Weights and Activations.低比特权重和激活值卷积神经网络的有效训练。
IEEE Trans Pattern Anal Mach Intell. 2022 Oct;44(10):6140-6152. doi: 10.1109/TPAMI.2021.3088904. Epub 2022 Sep 14.
7
Training high-performance and large-scale deep neural networks with full 8-bit integers.用全 8 位整数训练高性能和大规模深度神经网络。
Neural Netw. 2020 May;125:70-82. doi: 10.1016/j.neunet.2019.12.027. Epub 2020 Jan 15.
8
Whether the Support Region of Three-Bit Uniform Quantizer Has a Strong Impact on Post-Training Quantization for MNIST Dataset?三位均匀量化器的支持区域对MNIST数据集的训练后量化有很大影响吗?
Entropy (Basel). 2021 Dec 20;23(12):1699. doi: 10.3390/e23121699.
9
Fast and Effective: A Novel Sequential Single-Path Search for Mixed-Precision-Quantized Networks.快速且高效:一种用于混合精度量化网络的新型顺序单路径搜索方法
IEEE Trans Cybern. 2023 Oct;53(10):6187-6199. doi: 10.1109/TCYB.2022.3164285. Epub 2023 Sep 15.
10
SpQuant-SNN: ultra-low precision membrane potential with sparse activations unlock the potential of on-device spiking neural networks applications.SpQuant-SNN:具有稀疏激活的超低精度膜电位开启了片上脉冲神经网络应用的潜力。
Front Neurosci. 2024 Sep 4;18:1440000. doi: 10.3389/fnins.2024.1440000. eCollection 2024.

本文引用的文献

1
Haar Wavelet Feature Compression for Quantized Graph Convolutional Networks.用于量化图卷积网络的哈尔小波特征压缩
IEEE Trans Neural Netw Learn Syst. 2024 Apr;35(4):4542-4553. doi: 10.1109/TNNLS.2023.3285874. Epub 2024 Apr 4.
2
ReLeQ : A Reinforcement Learning Approach for Automatic Deep Quantization of Neural Networks.ReLeQ:一种用于神经网络自动深度量化的强化学习方法。
IEEE Micro. 2020 Sep-Oct;40(5):37-45. doi: 10.1109/mm.2020.3009475. Epub 2020 Jul 15.
3
Towards Efficient U-Nets: A Coupled and Quantized Approach.迈向高效 U-Nets:一种耦合和量化方法。
IEEE Trans Pattern Anal Mach Intell. 2020 Aug;42(8):2038-2050. doi: 10.1109/TPAMI.2019.2907634. Epub 2019 Mar 26.
4
Deep Neural Network Compression by In-Parallel Pruning-Quantization.通过并行剪枝-量化实现深度神经网络压缩。
IEEE Trans Pattern Anal Mach Intell. 2020 Mar;42(3):568-579. doi: 10.1109/TPAMI.2018.2886192. Epub 2018 Dec 12.
5
Reducing the time complexity of the derandomized evolution strategy with covariance matrix adaptation (CMA-ES).降低带协方差矩阵自适应的去随机化进化策略(CMA-ES)的时间复杂度。
Evol Comput. 2003 Spring;11(1):1-18. doi: 10.1162/106365603321828970.