• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

三位均匀量化器的支持区域对MNIST数据集的训练后量化有很大影响吗?

Whether the Support Region of Three-Bit Uniform Quantizer Has a Strong Impact on Post-Training Quantization for MNIST Dataset?

作者信息

Nikolić Jelena, Perić Zoran, Aleksić Danijela, Tomić Stefan, Jovanović Aleksandra

机构信息

Faculty of Electronic Engineering, University of Nis, Aleksandra Medvedeva 14, 18000 Nis, Serbia.

Department of Mobile Network Nis, Telekom Srbija, Vozdova 11, 18000 Nis, Serbia.

出版信息

Entropy (Basel). 2021 Dec 20;23(12):1699. doi: 10.3390/e23121699.

DOI:10.3390/e23121699
PMID:34946005
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8700806/
Abstract

Driven by the need for the compression of weights in neural networks (NNs), which is especially beneficial for edge devices with a constrained resource, and by the need to utilize the simplest possible quantization model, in this paper, we study the performance of three-bit post-training uniform quantization. The goal is to put various choices of the key parameter of the quantizer in question (support region threshold) in one place and provide a detailed overview of this choice's impact on the performance of post-training quantization for the MNIST dataset. Specifically, we analyze whether it is possible to preserve the accuracy of the two NN models (MLP and CNN) to a great extent with the very simple three-bit uniform quantizer, regardless of the choice of the key parameter. Moreover, our goal is to answer the question of whether it is of the utmost importance in post-training three-bit uniform quantization, as it is in quantization, to determine the optimal support region threshold value of the quantizer to achieve some predefined accuracy of the quantized neural network (QNN). The results show that the choice of the support region threshold value of the three-bit uniform quantizer does not have such a strong impact on the accuracy of the QNNs, which is not the case with two-bit uniform post-training quantization, when applied in MLP for the same classification task. Accordingly, one can anticipate that due to this special property, the post-training quantization model in question can be greatly exploited.

摘要

受神经网络(NNs)中权重压缩需求的驱动,这对资源受限的边缘设备尤为有益,同时也受使用尽可能简单的量化模型的需求驱动,在本文中,我们研究了三位训练后均匀量化的性能。目标是将量化器关键参数(支持区域阈值)的各种选择集中在一起,并详细概述这种选择对MNIST数据集训练后量化性能的影响。具体而言,我们分析了使用非常简单的三位均匀量化器,无论关键参数如何选择,是否有可能在很大程度上保持两个NN模型(多层感知器和卷积神经网络)的准确性。此外,我们的目标是回答在训练后三位均匀量化中,确定量化器的最佳支持区域阈值以实现量化神经网络(QNN)的某些预定义准确性是否与量化中一样至关重要的问题。结果表明,三位均匀量化器的支持区域阈值选择对QNN的准确性没有如此强烈的影响,而在相同分类任务中应用于多层感知器的两位均匀训练后量化则并非如此。因此,可以预期,由于这种特殊性质,所讨论的训练后量化模型可以得到极大利用。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/d7c28b653301/entropy-23-01699-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/1a0675f16952/entropy-23-01699-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/5abb85b1c3f9/entropy-23-01699-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/d52e9c0cfee9/entropy-23-01699-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/51a250fdd41b/entropy-23-01699-g004a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/d375a3b4de69/entropy-23-01699-g005a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/0be02fe241c4/entropy-23-01699-g006a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/6eb4ea51ff9b/entropy-23-01699-g007a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/df5b1adf10f8/entropy-23-01699-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/0f608220bec5/entropy-23-01699-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/3fec4cd3c624/entropy-23-01699-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/9b6e495f2471/entropy-23-01699-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/f0caec8c8b20/entropy-23-01699-g012a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/3e3ee85f891d/entropy-23-01699-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/d7c28b653301/entropy-23-01699-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/1a0675f16952/entropy-23-01699-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/5abb85b1c3f9/entropy-23-01699-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/d52e9c0cfee9/entropy-23-01699-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/51a250fdd41b/entropy-23-01699-g004a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/d375a3b4de69/entropy-23-01699-g005a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/0be02fe241c4/entropy-23-01699-g006a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/6eb4ea51ff9b/entropy-23-01699-g007a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/df5b1adf10f8/entropy-23-01699-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/0f608220bec5/entropy-23-01699-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/3fec4cd3c624/entropy-23-01699-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/9b6e495f2471/entropy-23-01699-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/f0caec8c8b20/entropy-23-01699-g012a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/3e3ee85f891d/entropy-23-01699-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/69ef/8700806/d7c28b653301/entropy-23-01699-g014.jpg

相似文献

1
Whether the Support Region of Three-Bit Uniform Quantizer Has a Strong Impact on Post-Training Quantization for MNIST Dataset?三位均匀量化器的支持区域对MNIST数据集的训练后量化有很大影响吗?
Entropy (Basel). 2021 Dec 20;23(12):1699. doi: 10.3390/e23121699.
2
Design of a 2-Bit Neural Network Quantizer for Laplacian Source.用于拉普拉斯源的2比特神经网络量化器设计
Entropy (Basel). 2021 Jul 22;23(8):933. doi: 10.3390/e23080933.
3
Exploiting Retraining-Based Mixed-Precision Quantization for Low-Cost DNN Accelerator Design.利用基于再训练的混合精度量化进行低成本深度神经网络加速器设计。
IEEE Trans Neural Netw Learn Syst. 2021 Jul;32(7):2925-2938. doi: 10.1109/TNNLS.2020.3008996. Epub 2021 Jul 6.
4
Optimization-Based Post-Training Quantization With Bit-Split and Stitching.基于优化的带位分割与拼接的训练后量化
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):2119-2135. doi: 10.1109/TPAMI.2022.3159369. Epub 2023 Jan 6.
5
Training high-performance and large-scale deep neural networks with full 8-bit integers.用全 8 位整数训练高性能和大规模深度神经网络。
Neural Netw. 2020 May;125:70-82. doi: 10.1016/j.neunet.2019.12.027. Epub 2020 Jan 15.
6
Transform Quantization for CNN Compression.用于卷积神经网络压缩的变换量化
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):5700-5714. doi: 10.1109/TPAMI.2021.3084839. Epub 2022 Aug 4.
7
A Convolutional Neural Network-Based Quantization Method for Block Compressed Sensing of Images.一种基于卷积神经网络的图像块压缩感知量化方法。
Entropy (Basel). 2024 May 29;26(6):468. doi: 10.3390/e26060468.
8
GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks.无梯度比特分配在混合精度神经网络中的应用。
Sensors (Basel). 2022 Dec 13;22(24):9772. doi: 10.3390/s22249772.
9
MedQ: Lossless ultra-low-bit neural network quantization for medical image segmentation.MedQ:用于医学图像分割的无损超低比特神经网络量化。
Med Image Anal. 2021 Oct;73:102200. doi: 10.1016/j.media.2021.102200. Epub 2021 Aug 2.
10
Pattern Classification Using Quantized Neural Networks for FPGA-Based Low-Power IoT Devices.基于量化神经网络的 FPGA 低功耗物联网设备的模式分类。
Sensors (Basel). 2022 Nov 10;22(22):8694. doi: 10.3390/s22228694.

引用本文的文献

1
Speaker Recognition Using Constrained Convolutional Neural Networks in Emotional Speech.在情感语音中使用约束卷积神经网络的说话人识别
Entropy (Basel). 2022 Mar 16;24(3):414. doi: 10.3390/e24030414.

本文引用的文献

1
Design of a 2-Bit Neural Network Quantizer for Laplacian Source.用于拉普拉斯源的2比特神经网络量化器设计
Entropy (Basel). 2021 Jul 22;23(8):933. doi: 10.3390/e23080933.
2
Quantization and Deployment of Deep Neural Networks on Microcontrollers.微控制器上的深度神经网络量化与部署。
Sensors (Basel). 2021 Apr 23;21(9):2984. doi: 10.3390/s21092984.
3
Deep Task-Based Quantization.基于深度任务的量化
Entropy (Basel). 2021 Jan 13;23(1):104. doi: 10.3390/e23010104.
4
Optimal Randomness for Stochastic Configuration Network (SCN) with Heavy-Tailed Distributions.具有重尾分布的随机配置网络(SCN)的最优随机性
Entropy (Basel). 2020 Dec 31;23(1):56. doi: 10.3390/e23010056.
5
A Novel Low-Bit Quantization Strategy for Compressing Deep Neural Networks.一种用于压缩深度神经网络的新型低比特量化策略。
Comput Intell Neurosci. 2020 Feb 18;2020:7839064. doi: 10.1155/2020/7839064. eCollection 2020.
6
Deep Neural Network Compression by In-Parallel Pruning-Quantization.通过并行剪枝-量化实现深度神经网络压缩。
IEEE Trans Pattern Anal Mach Intell. 2020 Mar;42(3):568-579. doi: 10.1109/TPAMI.2018.2886192. Epub 2018 Dec 12.