• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过异常值感知量化实现基于传感器的视觉系统的高效深度学习模型压缩

Efficient Deep Learning Model Compression for Sensor-Based Vision Systems via Outlier-Aware Quantization.

作者信息

Yoo Joonhyuk, Ban Guenwoo

机构信息

College of Information and Communication Engineering, Daegu University, Gyeongsan 38453, Republic of Korea.

NPS Co., Ltd., Cheongju 28371, Republic of Korea.

出版信息

Sensors (Basel). 2025 May 5;25(9):2918. doi: 10.3390/s25092918.

DOI:10.3390/s25092918
PMID:40363355
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12074280/
Abstract

With the rapid growth of sensor technology and computer vision, efficient deep learning models are essential for real-time image feature extraction in resource-constrained environments. However, most existing quantized deep neural networks (DNNs) are highly sensitive to outliers, leading to severe performance degradation in low-precision settings. Our study reveals that outliers extending beyond the nominal weight distribution significantly increase the dynamic range, thereby reducing quantization resolution and affecting sensor-based image analysis tasks. To address this, we propose an outlier-aware quantization (OAQ) method that effectively reshapes weight distributions to enhance quantization accuracy. By analyzing previous outlier-handling techniques using structural similarity (SSIM) measurement results, we demonstrated that OAQ significantly reduced the negative impact of outliers while maintaining computational efficiency. Notably, OAQ was orthogonal to existing quantization schemes, making it compatible with various quantization methods without additional computational overhead. Experimental results on multiple CNN architectures and quantization approaches showed that OAQ effectively mitigated quantization errors. In post-training quantization (PTQ), our 4-bit OAQ ResNet20 model achieved improved accuracy compared with full-precision counterparts, while in quantization-aware training (QAT), OAQ enhanced 2-bit quantization performance by 43.55% over baseline methods. These results confirmed the potential of OAQ for optimizing deep learning models in sensor-based vision applications.

摘要

随着传感器技术和计算机视觉的快速发展,高效的深度学习模型对于资源受限环境下的实时图像特征提取至关重要。然而,大多数现有的量化深度神经网络(DNN)对异常值高度敏感,导致在低精度设置下性能严重下降。我们的研究表明,超出标称权重分布的异常值会显著增加动态范围,从而降低量化分辨率并影响基于传感器的图像分析任务。为了解决这个问题,我们提出了一种异常值感知量化(OAQ)方法,该方法有效地重塑权重分布以提高量化精度。通过使用结构相似性(SSIM)测量结果分析先前的异常值处理技术,我们证明了OAQ在保持计算效率的同时显著降低了异常值的负面影响。值得注意的是,OAQ与现有的量化方案正交,使其能够与各种量化方法兼容,而无需额外的计算开销。在多个CNN架构和量化方法上的实验结果表明,OAQ有效地减轻了量化误差。在训练后量化(PTQ)中,我们的4位OAQ ResNet20模型与全精度模型相比,精度得到了提高,而在量化感知训练(QAT)中,OAQ比基线方法将2位量化性能提高了43.55%。这些结果证实了OAQ在优化基于传感器的视觉应用中的深度学习模型方面的潜力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/e84ed67832d9/sensors-25-02918-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/d664a8e409fa/sensors-25-02918-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/e0cd7d577e95/sensors-25-02918-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/f5dd5a06e2af/sensors-25-02918-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/a18e92f82576/sensors-25-02918-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/63d4802412d4/sensors-25-02918-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/8730c469f731/sensors-25-02918-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/900fb3a3e196/sensors-25-02918-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/98b4f58be07f/sensors-25-02918-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/f35988fe7144/sensors-25-02918-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/e84ed67832d9/sensors-25-02918-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/d664a8e409fa/sensors-25-02918-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/e0cd7d577e95/sensors-25-02918-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/f5dd5a06e2af/sensors-25-02918-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/a18e92f82576/sensors-25-02918-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/63d4802412d4/sensors-25-02918-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/8730c469f731/sensors-25-02918-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/900fb3a3e196/sensors-25-02918-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/98b4f58be07f/sensors-25-02918-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/f35988fe7144/sensors-25-02918-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5e45/12074280/e84ed67832d9/sensors-25-02918-g010.jpg

相似文献

1
Efficient Deep Learning Model Compression for Sensor-Based Vision Systems via Outlier-Aware Quantization.通过异常值感知量化实现基于传感器的视觉系统的高效深度学习模型压缩
Sensors (Basel). 2025 May 5;25(9):2918. doi: 10.3390/s25092918.
2
ADFQ-ViT: Activation-Distribution-Friendly post-training Quantization for Vision Transformers.
Neural Netw. 2025 Jun;186:107289. doi: 10.1016/j.neunet.2025.107289. Epub 2025 Feb 22.
3
Quantization Friendly MobileNet (QF-MobileNet) Architecture for Vision Based Applications on Embedded Platforms.面向嵌入式平台视觉应用的量化友好型 MobileNet(QF-MobileNet)架构。
Neural Netw. 2021 Apr;136:28-39. doi: 10.1016/j.neunet.2020.12.022. Epub 2020 Dec 29.
4
Hessian-based mixed-precision quantization with transition aware training for neural networks.基于黑塞矩阵的神经网络混合精度量化与过渡感知训练
Neural Netw. 2025 Feb;182:106910. doi: 10.1016/j.neunet.2024.106910. Epub 2024 Nov 16.
5
Pushing the Limit of Post-Training Quantization.突破训练后量化的极限
IEEE Trans Pattern Anal Mach Intell. 2025 Jul;47(7):5556-5570. doi: 10.1109/TPAMI.2025.3554523.
6
Training high-performance and large-scale deep neural networks with full 8-bit integers.用全 8 位整数训练高性能和大规模深度神经网络。
Neural Netw. 2020 May;125:70-82. doi: 10.1016/j.neunet.2019.12.027. Epub 2020 Jan 15.
7
Deployable mixed-precision quantization with co-learning and one-time search.采用协同学习和一次性搜索的可部署混合精度量化
Neural Netw. 2025 Jan;181:106812. doi: 10.1016/j.neunet.2024.106812. Epub 2024 Oct 18.
8
MedQ: Lossless ultra-low-bit neural network quantization for medical image segmentation.MedQ:用于医学图像分割的无损超低比特神经网络量化。
Med Image Anal. 2021 Oct;73:102200. doi: 10.1016/j.media.2021.102200. Epub 2021 Aug 2.
9
Exploiting Retraining-Based Mixed-Precision Quantization for Low-Cost DNN Accelerator Design.利用基于再训练的混合精度量化进行低成本深度神经网络加速器设计。
IEEE Trans Neural Netw Learn Syst. 2021 Jul;32(7):2925-2938. doi: 10.1109/TNNLS.2020.3008996. Epub 2021 Jul 6.
10
Optimization-Based Post-Training Quantization With Bit-Split and Stitching.基于优化的带位分割与拼接的训练后量化
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):2119-2135. doi: 10.1109/TPAMI.2022.3159369. Epub 2023 Jan 6.

本文引用的文献

1
From Near-Sensor to In-Sensor: A State-of-the-Art Review of Embedded AI Vision Systems.从近传感器到传感器内:嵌入式人工智能视觉系统的最新综述
Sensors (Basel). 2024 Aug 22;24(16):5446. doi: 10.3390/s24165446.