• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

从算法到硬件:深度神经网络高效安全部署综述

From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks.

作者信息

Geng Xue, Wang Zhe, Chen Chunyun, Xu Qing, Xu Kaixin, Jin Chao, Gupta Manas, Yang Xulei, Chen Zhenghua, Sabry Aly Mohamed M, Lin Jie, Wu Min, Li Xiaoli

出版信息

IEEE Trans Neural Netw Learn Syst. 2025 Apr;36(4):5837-5857. doi: 10.1109/TNNLS.2024.3394494. Epub 2025 Apr 4.

DOI:10.1109/TNNLS.2024.3394494
PMID:38875092
Abstract

Deep neural networks (DNNs) have been widely used in many artificial intelligence (AI) tasks. However, deploying them brings significant challenges due to the huge cost of memory, energy, and computation. To address these challenges, researchers have developed various model compression techniques such as model quantization and model pruning. Recently, there has been a surge in research on compression methods to achieve model efficiency while retaining performance. Furthermore, more and more works focus on customizing the DNN hardware accelerators to better leverage the model compression techniques. In addition to efficiency, preserving security and privacy is critical for deploying DNNs. However, the vast and diverse body of related works can be overwhelming. This inspires us to conduct a comprehensive survey on recent research toward the goal of high-performance, cost-efficient, and safe deployment of DNNs. Our survey first covers the mainstream model compression techniques, such as model quantization, model pruning, knowledge distillation, and optimizations of nonlinear operations. We then introduce recent advances in designing hardware accelerators that can adapt to efficient model compression approaches. In addition, we discuss how homomorphic encryption can be integrated to secure DNN deployment. Finally, we discuss several issues, such as hardware evaluation, generalization, and integration of various compression approaches. Overall, we aim to provide a big picture of efficient DNNs from algorithm to hardware accelerators and security perspectives.

摘要

深度神经网络(DNN)已广泛应用于许多人工智能(AI)任务中。然而,由于内存、能源和计算成本巨大,部署这些网络带来了重大挑战。为应对这些挑战,研究人员开发了各种模型压缩技术,如模型量化和模型剪枝。最近,在实现模型效率同时保留性能的压缩方法研究方面激增。此外,越来越多的工作专注于定制DNN硬件加速器,以更好地利用模型压缩技术。除了效率,保护安全和隐私对于部署DNN至关重要。然而,大量且多样的相关工作可能让人应接不暇。这促使我们针对DNN的高性能、成本高效且安全的部署目标,对近期研究进行全面调查。我们的调查首先涵盖主流模型压缩技术,如模型量化、模型剪枝、知识蒸馏以及非线性操作的优化。然后,我们介绍在设计能够适应高效模型压缩方法的硬件加速器方面的最新进展。此外,我们讨论如何集成同态加密以保障DNN部署安全。最后,我们讨论几个问题,如硬件评估、泛化以及各种压缩方法的集成。总体而言,我们旨在从算法到硬件加速器以及安全角度,提供高效DNN的全景图。

相似文献

1
From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks.从算法到硬件:深度神经网络高效安全部署综述
IEEE Trans Neural Netw Learn Syst. 2025 Apr;36(4):5837-5857. doi: 10.1109/TNNLS.2024.3394494. Epub 2025 Apr 4.
2
Random sketch learning for deep neural networks in edge computing.边缘计算中深度神经网络的随机草图学习
Nat Comput Sci. 2021 Mar;1(3):221-228. doi: 10.1038/s43588-021-00039-6. Epub 2021 Mar 25.
3
Non-Structured DNN Weight Pruning-Is It Beneficial in Any Platform?非结构化深度神经网络权重剪枝——在任何平台上都有益吗?
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4930-4944. doi: 10.1109/TNNLS.2021.3063265. Epub 2022 Aug 31.
4
Resource-constrained FPGA/DNN co-design.资源受限的FPGA/DNN协同设计
Neural Comput Appl. 2021;33(21):14741-14751. doi: 10.1007/s00521-021-06113-4. Epub 2021 May 15.
5
SalvageDNN: salvaging deep neural network accelerators with permanent faults through saliency-driven fault-aware mapping.通过显著驱动的故障感知映射来利用永久性故障挽救深度神经网络加速器。
Philos Trans A Math Phys Eng Sci. 2020 Feb 7;378(2164):20190164. doi: 10.1098/rsta.2019.0164. Epub 2019 Dec 23.
6
Towards Model Compression for Deep Learning Based Speech Enhancement.面向基于深度学习的语音增强的模型压缩
IEEE/ACM Trans Audio Speech Lang Process. 2021;29:1785-1794. doi: 10.1109/taslp.2021.3082282. Epub 2021 May 21.
7
SmartDeal: Remodeling Deep Network Weights for Efficient Inference and Training.智能交易:重塑深度网络权重以实现高效推理与训练
IEEE Trans Neural Netw Learn Syst. 2023 Oct;34(10):7099-7113. doi: 10.1109/TNNLS.2021.3138056. Epub 2023 Oct 5.
8
Three-Stage Global Channel Pruning for Resources-Limited Platform.面向资源受限平台的三阶段全局通道剪枝
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):16153-16166. doi: 10.1109/TNNLS.2023.3292152. Epub 2024 Oct 29.
9
Arch-Net: Model conversion and quantization for architecture agnostic model deployment.Arch-Net:用于与架构无关的模型部署的模型转换和量化。
Neural Netw. 2025 Jul;187:107384. doi: 10.1016/j.neunet.2025.107384. Epub 2025 Mar 18.
10
Random and Adversarial Bit Error Robustness: Energy-Efficient and Secure DNN Accelerators.随机和对抗性比特错误鲁棒性:节能且安全的深度神经网络加速器
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3632-3647. doi: 10.1109/TPAMI.2022.3181972.