• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

L1-范数批归一化在深度神经网络高效训练中的应用。

L1 -Norm Batch Normalization for Efficient Training of Deep Neural Networks.

出版信息

IEEE Trans Neural Netw Learn Syst. 2019 Jul;30(7):2043-2051. doi: 10.1109/TNNLS.2018.2876179. Epub 2018 Nov 9.

DOI:10.1109/TNNLS.2018.2876179
PMID:30418924
Abstract

Batch normalization (BN) has recently become a standard component for accelerating and improving the training of deep neural networks (DNNs). However, BN brings in additional calculations, consumes more memory, and significantly slows down the training iteration. Furthermore, the nonlinear square and sqrt operations in the normalization process impede low bit-width quantization techniques, which draw much attention to the deep learning hardware community. In this paper, we propose an L1 -norm BN (L1BN) with only linear operations in both forward and backward propagations during training. L1BN is approximately equivalent to the conventional L2 -norm BN (L2BN) by multiplying a scaling factor that equals (π/2) . Experiments on various convolutional neural networks and generative adversarial networks reveal that L1BN can maintain the same performance and convergence rate as L2BN but with higher computational efficiency. In real application-specified integrated circuit synthesis with reduced resources, L1BN achieves 25% speedup and 37% energy saving compared to the original L2BN. Our hardware-friendly normalization method not only surpasses L2BN in speed but also simplifies the design of deep learning accelerators. Last but not least, L1BN promises a fully quantized training of DNNs, which empowers future artificial intelligence applications on mobile devices with transfer and continual learning capability.

摘要

批量归一化(BN)最近已成为加速和改进深度神经网络(DNN)训练的标准组件。然而,BN 带来了额外的计算量,消耗了更多的内存,并且大大降低了训练迭代速度。此外,归一化过程中的非线性平方和平方根运算阻碍了低比特宽度量化技术的发展,这引起了深度学习硬件社区的广泛关注。在本文中,我们提出了一种 L1-范数 BN(L1BN),在训练过程中前向和后向传播都只有线性运算。L1BN 通过乘以一个缩放因子(等于 (π/2)),与传统的 L2-范数 BN(L2BN)近似等效。在各种卷积神经网络和生成对抗网络上的实验表明,L1BN 可以保持与 L2BN 相同的性能和收敛速度,但计算效率更高。在使用减少资源的专用集成电路综合中,L1BN 与原始的 L2BN 相比,速度提高了 25%,能耗降低了 37%。我们的硬件友好型归一化方法不仅在速度上超过了 L2BN,而且简化了深度学习加速器的设计。最后但同样重要的是,L1BN 可以实现 DNN 的完全量化训练,这为未来具有迁移和持续学习能力的移动设备上的人工智能应用提供了支持。

相似文献

1
L1 -Norm Batch Normalization for Efficient Training of Deep Neural Networks.L1-范数批归一化在深度神经网络高效训练中的应用。
IEEE Trans Neural Netw Learn Syst. 2019 Jul;30(7):2043-2051. doi: 10.1109/TNNLS.2018.2876179. Epub 2018 Nov 9.
2
Training high-performance and large-scale deep neural networks with full 8-bit integers.用全 8 位整数训练高性能和大规模深度神经网络。
Neural Netw. 2020 May;125:70-82. doi: 10.1016/j.neunet.2019.12.027. Epub 2020 Jan 15.
3
Quantization Friendly MobileNet (QF-MobileNet) Architecture for Vision Based Applications on Embedded Platforms.面向嵌入式平台视觉应用的量化友好型 MobileNet(QF-MobileNet)架构。
Neural Netw. 2021 Apr;136:28-39. doi: 10.1016/j.neunet.2020.12.022. Epub 2020 Dec 29.
4
Training Faster by Separating Modes of Variation in Batch-Normalized Models.通过分离批归一化模型中的变化模式实现更快训练。
IEEE Trans Pattern Anal Mach Intell. 2020 Jun;42(6):1483-1500. doi: 10.1109/TPAMI.2019.2895781. Epub 2019 Jan 28.
5
Tree-CNN: A hierarchical Deep Convolutional Neural Network for incremental learning.树卷积神经网络:一种用于增量学习的层次化深度卷积神经网络。
Neural Netw. 2020 Jan;121:148-160. doi: 10.1016/j.neunet.2019.09.010. Epub 2019 Sep 19.
6
DART: Domain-Adversarial Residual-Transfer networks for unsupervised cross-domain image classification.DART:用于无监督跨域图像分类的域对抗残差转移网络。
Neural Netw. 2020 Jul;127:182-192. doi: 10.1016/j.neunet.2020.03.025. Epub 2020 Apr 24.
7
Adversarial Examples: Opportunities and Challenges.对抗样本:机遇与挑战。
IEEE Trans Neural Netw Learn Syst. 2020 Jul;31(7):2578-2593. doi: 10.1109/TNNLS.2019.2933524. Epub 2019 Sep 17.
8
GXNOR-Net: Training deep neural networks with ternary weights and activations without full-precision memory under a unified discretization framework.GXNOR-Net:在统一的离散化框架下使用三进制权重和激活函数训练深度神经网络,无需全精度存储。
Neural Netw. 2018 Apr;100:49-58. doi: 10.1016/j.neunet.2018.01.010. Epub 2018 Feb 2.
9
A structure-time parallel implementation of spike-based deep learning.基于尖峰的深度学习的结构-时间并行实现。
Neural Netw. 2019 May;113:72-78. doi: 10.1016/j.neunet.2019.01.010. Epub 2019 Feb 4.
10
Accelerating deep learning with memcomputing.利用忆阻器计算加速深度学习。
Neural Netw. 2019 Feb;110:1-7. doi: 10.1016/j.neunet.2018.10.012. Epub 2018 Nov 3.

引用本文的文献

1
Improving prediction accuracy of hospital arrival vital signs using a multi-output machine learning model: a retrospective study of JSAS-registry data.使用多输出机器学习模型提高医院到达时生命体征的预测准确性:对JSAS登记数据的回顾性研究
BMC Emerg Med. 2025 May 13;25(1):78. doi: 10.1186/s12873-025-01233-9.
2
Adaptive temporal compression for reduction of computational complexity in human behavior recognition.自适应时间压缩在人类行为识别中降低计算复杂度的应用。
Sci Rep. 2024 May 8;14(1):10560. doi: 10.1038/s41598-024-61286-x.
3
[A survey on the application of convolutional neural networks in the diagnosis of occupational pneumoconiosis].
关于卷积神经网络在职业性尘肺病诊断中的应用调查
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi. 2024 Apr 25;41(2):413-420. doi: 10.7507/1001-5515.202309079.
4
Calculation of solvation force in molecular dynamics simulation by deep-learning method.基于深度学习方法的分子动力学模拟中溶剂化力的计算。
Biophys J. 2024 Sep 3;123(17):2830-2838. doi: 10.1016/j.bpj.2024.02.029. Epub 2024 Mar 4.
5
A three-dimensional deep learning model for inter-site harmonization of structural MR images of the brain: Extensive validation with a multicenter dataset.一种用于脑结构磁共振图像跨站点协调的三维深度学习模型:基于多中心数据集的广泛验证
Heliyon. 2023 Nov 23;9(12):e22647. doi: 10.1016/j.heliyon.2023.e22647. eCollection 2023 Dec.
6
Constraints on Optimising Encoder-Only Transformers for Modelling Sign Language with Human Pose Estimation Keypoint Data.使用人体姿态估计关键点数据对仅编码器的Transformer进行手语建模时的优化限制
J Imaging. 2023 Nov 2;9(11):238. doi: 10.3390/jimaging9110238.
7
Application of Laser-Induced Breakdown Spectroscopy Coupled With Spectral Matrix and Convolutional Neural Network for Identifying Geographical Origins of Franch.激光诱导击穿光谱结合光谱矩阵和卷积神经网络在识别法国葡萄酒地理来源中的应用
Front Artif Intell. 2021 Dec 10;4:735533. doi: 10.3389/frai.2021.735533. eCollection 2021.
8
ES-ImageNet: A Million Event-Stream Classification Dataset for Spiking Neural Networks.ES-ImageNet:用于脉冲神经网络的百万事件流分类数据集。
Front Neurosci. 2021 Nov 25;15:726582. doi: 10.3389/fnins.2021.726582. eCollection 2021.
9
N6-Methyladenosine RNA modification in cerebrospinal fluid as a novel potential diagnostic biomarker for progressive multiple sclerosis.脑脊液 N6-甲基腺苷 RNA 修饰作为一种新型潜在的多发性硬化进展性诊断生物标志物。
J Transl Med. 2021 Jul 22;19(1):316. doi: 10.1186/s12967-021-02981-5.
10
Auto-Refining Reconstruction Algorithm for Recreation of Limited Angle Humanoid Depth Data.用于有限角度人形深度数据再现的自动精炼重建算法。
Sensors (Basel). 2021 May 26;21(11):3702. doi: 10.3390/s21113702.