• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

优化基于模拟记忆的深度神经网络的权重编程。

Optimised weight programming for analogue memory-based deep neural networks.

机构信息

IBM Research-Almaden, 650 Harry Road, San Jose, CA, USA.

IBM Research-Yorktown Heights, 1101 Kitchawan Road, Yorktown Heights, NY, USA.

出版信息

Nat Commun. 2022 Jun 30;13(1):3765. doi: 10.1038/s41467-022-31405-1.

DOI:10.1038/s41467-022-31405-1
PMID:35773285
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9247051/
Abstract

Analogue memory-based deep neural networks provide energy-efficiency and per-area throughput gains relative to state-of-the-art digital counterparts such as graphics processing units. Recent advances focus largely on hardware-aware algorithmic training and improvements to circuits, architectures, and memory devices. Optimal translation of software-trained weights into analogue hardware weights-given the plethora of complex memory non-idealities-represents an equally important task. We report a generalised computational framework that automates the crafting of complex weight programming strategies to minimise accuracy degradations during inference, particularly over time. The framework is agnostic to network structure and generalises well across recurrent, convolutional, and transformer neural networks. As a highly flexible numerical heuristic, the approach accommodates arbitrary device-level complexity, making it potentially relevant for a variety of analogue memories. By quantifying the limit of achievable inference accuracy, it also enables analogue memory-based deep neural network accelerators to reach their full inference potential.

摘要

基于模拟记忆的深度神经网络相对于最先进的数字对应物(如图形处理单元)提供了节能和每面积吞吐量的优势。最近的进展主要集中在硬件感知算法训练以及对电路、架构和存储设备的改进上。将软件训练的权重转换为模拟硬件权重——考虑到大量复杂的存储非理想情况——是一项同样重要的任务。我们报告了一个通用的计算框架,该框架可以自动制作复杂的权重编程策略,以最大限度地减少推理过程中的精度下降,特别是随着时间的推移。该框架与网络结构无关,并且可以很好地推广到递归、卷积和变压器神经网络。作为一种高度灵活的数值启发式方法,该方法可以适应任意设备级别的复杂性,因此对于各种模拟存储器都可能具有相关性。通过量化可实现的推理精度的极限,它还可以使基于模拟存储器的深度神经网络加速器充分发挥其推理潜力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/58b63c0c0522/41467_2022_31405_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/235abb2cd617/41467_2022_31405_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/1a7afae41559/41467_2022_31405_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/b8607716b7e9/41467_2022_31405_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/f57b3cfe5184/41467_2022_31405_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/58b63c0c0522/41467_2022_31405_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/235abb2cd617/41467_2022_31405_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/1a7afae41559/41467_2022_31405_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/b8607716b7e9/41467_2022_31405_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/f57b3cfe5184/41467_2022_31405_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36ff/9247051/58b63c0c0522/41467_2022_31405_Fig5_HTML.jpg

相似文献

1
Optimised weight programming for analogue memory-based deep neural networks.优化基于模拟记忆的深度神经网络的权重编程。
Nat Commun. 2022 Jun 30;13(1):3765. doi: 10.1038/s41467-022-31405-1.
2
Quantization-aware training for low precision photonic neural networks.低精度光神经网络的量化感知训练。
Neural Netw. 2022 Nov;155:561-573. doi: 10.1016/j.neunet.2022.09.015. Epub 2022 Sep 19.
3
Toward Software-Equivalent Accuracy on Transformer-Based Deep Neural Networks With Analog Memory Devices.借助模拟存储设备实现基于Transformer的深度神经网络的软件等效精度。
Front Comput Neurosci. 2021 Jul 5;15:675741. doi: 10.3389/fncom.2021.675741. eCollection 2021.
4
Lean Neural Networks for Autonomous Radar Waveform Design.用于自主雷达波形设计的精简神经网络。
Sensors (Basel). 2022 Feb 9;22(4):1317. doi: 10.3390/s22041317.
5
Equivalent-accuracy accelerated neural-network training using analogue memory.利用模拟内存实现等效精度的加速神经网络训练。
Nature. 2018 Jun;558(7708):60-67. doi: 10.1038/s41586-018-0180-5. Epub 2018 Jun 6.
6
Toward Full-Stack Acceleration of Deep Convolutional Neural Networks on FPGAs.深度卷积神经网络在 FPGAs 上的全栈加速。
IEEE Trans Neural Netw Learn Syst. 2022 Aug;33(8):3974-3987. doi: 10.1109/TNNLS.2021.3055240. Epub 2022 Aug 3.
7
Energy-efficient Mott activation neuron for full-hardware implementation of neural networks.用于神经网络全硬件实现的节能莫特激活神经元。
Nat Nanotechnol. 2021 Jun;16(6):680-687. doi: 10.1038/s41565-021-00874-8. Epub 2021 Mar 18.
8
Mixed-Precision Deep Learning Based on Computational Memory.基于计算内存的混合精度深度学习
Front Neurosci. 2020 May 12;14:406. doi: 10.3389/fnins.2020.00406. eCollection 2020.
9
Accelerating Inference of Convolutional Neural Networks Using In-memory Computing.利用内存计算加速卷积神经网络的推理
Front Comput Neurosci. 2021 Aug 3;15:674154. doi: 10.3389/fncom.2021.674154. eCollection 2021.
10
Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices.使用电阻式交叉点器件训练深度卷积神经网络。
Front Neurosci. 2017 Oct 10;11:538. doi: 10.3389/fnins.2017.00538. eCollection 2017.

引用本文的文献

1
The Role of Phase-Change Memory in Edge Computing and Analog In-Memory Computing: An Overview of Recent Research Contributions and Future Challenges.相变存储器在边缘计算和模拟内存计算中的作用:近期研究贡献与未来挑战综述
Sensors (Basel). 2025 Jun 9;25(12):3618. doi: 10.3390/s25123618.
2
In situ training of an in-sensor artificial neural network based on ferroelectric photosensors.基于铁电光电传感器的传感器内人工神经网络的原位训练。
Nat Commun. 2025 Jan 7;16(1):421. doi: 10.1038/s41467-024-55508-z.
3
Hardware implementation of backpropagation using progressive gradient descent for in situ training of multilayer neural networks.

本文引用的文献

1
Toward Software-Equivalent Accuracy on Transformer-Based Deep Neural Networks With Analog Memory Devices.借助模拟存储设备实现基于Transformer的深度神经网络的软件等效精度。
Front Comput Neurosci. 2021 Jul 5;15:675741. doi: 10.3389/fncom.2021.675741. eCollection 2021.
2
Robo-writers: the rise and risks of language-generating AI.机器人写作:生成语言的人工智能的兴起与风险。
Nature. 2021 Mar;591(7848):22-25. doi: 10.1038/d41586-021-00530-0.
3
Accurate deep neural network inference using computational phase-change memory.利用计算相变化内存实现精确的深度神经网络推理。
使用渐进梯度下降进行多层神经网络原位训练的反向传播的硬件实现。
Sci Adv. 2024 Jul 12;10(28):eado8999. doi: 10.1126/sciadv.ado8999.
4
Resistive Switching Devices for Neuromorphic Computing: From Foundations to Chip Level Innovations.用于神经形态计算的电阻式开关器件:从基础到芯片级创新
Nanomaterials (Basel). 2024 Mar 15;14(6):527. doi: 10.3390/nano14060527.
5
Bringing uncertainty quantification to the extreme-edge with memristor-based Bayesian neural networks.利用基于忆阻器的贝叶斯神经网络将不确定性量化引入极端边缘。
Nat Commun. 2023 Nov 20;14(1):7530. doi: 10.1038/s41467-023-43317-9.
6
Hardware-aware training for large-scale and diverse deep learning inference workloads using in-memory computing-based accelerators.使用基于内存计算的加速器对大规模多样的深度学习推理工作负载进行硬件感知训练。
Nat Commun. 2023 Aug 30;14(1):5282. doi: 10.1038/s41467-023-40770-4.
7
Thousands of conductance levels in memristors integrated on CMOS.在 CMOS 上集成的数千个电导水平的忆阻器。
Nature. 2023 Mar;615(7954):823-829. doi: 10.1038/s41586-023-05759-5. Epub 2023 Mar 29.
Nat Commun. 2020 May 18;11(1):2473. doi: 10.1038/s41467-020-16108-9.
4
Memory devices and applications for in-memory computing.用于内存计算的存储设备和应用。
Nat Nanotechnol. 2020 Jul;15(7):529-544. doi: 10.1038/s41565-020-0655-z. Epub 2020 Mar 30.
5
Parallel programming of an ionic floating-gate memory array for scalable neuromorphic computing.离子浮栅存储器阵列的并行编程可实现可扩展的神经形态计算。
Science. 2019 May 10;364(6440):570-574. doi: 10.1126/science.aaw5581. Epub 2019 Apr 25.
6
Memristive crossbar arrays for brain-inspired computing.忆阻器交叉阵列用于脑启发计算。
Nat Mater. 2019 Apr;18(4):309-323. doi: 10.1038/s41563-019-0291-x. Epub 2019 Mar 20.
7
High-Performance Mixed-Signal Neurocomputing With Nanoscale Floating-Gate Memory Cell Arrays.采用纳米级浮栅存储单元阵列的高性能混合信号神经计算
IEEE Trans Neural Netw Learn Syst. 2018 Oct;29(10):4782-4790. doi: 10.1109/TNNLS.2017.2778940. Epub 2017 Dec 22.
8
Equivalent-accuracy accelerated neural-network training using analogue memory.利用模拟内存实现等效精度的加速神经网络训练。
Nature. 2018 Jun;558(7708):60-67. doi: 10.1038/s41586-018-0180-5. Epub 2018 Jun 6.
9
Mastering the game of Go with deep neural networks and tree search.用深度神经网络和树搜索掌握围棋游戏。
Nature. 2016 Jan 28;529(7587):484-9. doi: 10.1038/nature16961.
10
Deep learning.深度学习。
Nature. 2015 May 28;521(7553):436-44. doi: 10.1038/nature14539.