• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于计算内存的混合精度深度学习

Mixed-Precision Deep Learning Based on Computational Memory.

作者信息

Nandakumar S R, Le Gallo Manuel, Piveteau Christophe, Joshi Vinay, Mariani Giovanni, Boybat Irem, Karunaratne Geethan, Khaddam-Aljameh Riduan, Egger Urs, Petropoulos Anastasios, Antonakopoulos Theodore, Rajendran Bipin, Sebastian Abu, Eleftheriou Evangelos

机构信息

IBM Research - Zurich, Rüschlikon, Switzerland.

Department of Information Technology and Electrical Engineering, ETH Zurich, Zurich, Switzerland.

出版信息

Front Neurosci. 2020 May 12;14:406. doi: 10.3389/fnins.2020.00406. eCollection 2020.

DOI:10.3389/fnins.2020.00406
PMID:32477047
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7235420/
Abstract

Deep neural networks (DNNs) have revolutionized the field of artificial intelligence and have achieved unprecedented success in cognitive tasks such as image and speech recognition. Training of large DNNs, however, is computationally intensive and this has motivated the search for novel computing architectures targeting this application. A computational memory unit with nanoscale resistive memory devices organized in crossbar arrays could store the synaptic weights in their conductance states and perform the expensive weighted summations in place in a non-von Neumann manner. However, updating the conductance states in a reliable manner during the weight update process is a fundamental challenge that limits the training accuracy of such an implementation. Here, we propose a mixed-precision architecture that combines a computational memory unit performing the weighted summations and imprecise conductance updates with a digital processing unit that accumulates the weight updates in high precision. A combined hardware/software training experiment of a multilayer perceptron based on the proposed architecture using a phase-change memory (PCM) array achieves 97.73% test accuracy on the task of classifying handwritten digits (based on the MNIST dataset), within 0.6% of the software baseline. The architecture is further evaluated using accurate behavioral models of PCM on a wide class of networks, namely convolutional neural networks, long-short-term-memory networks, and generative-adversarial networks. Accuracies comparable to those of floating-point implementations are achieved without being constrained by the non-idealities associated with the PCM devices. A system-level study demonstrates 172 × improvement in energy efficiency of the architecture when used for training a multilayer perceptron compared with a dedicated fully digital 32-bit implementation.

摘要

深度神经网络(DNN)彻底改变了人工智能领域,并在图像和语音识别等认知任务中取得了前所未有的成功。然而,大型DNN的训练计算量很大,这促使人们寻找针对该应用的新型计算架构。一种由交叉阵列中纳米级电阻式存储器件组成的计算存储单元,可以将突触权重存储在其电导状态中,并以非冯·诺依曼方式就地执行昂贵的加权求和。然而,在权重更新过程中以可靠的方式更新电导状态是一个基本挑战,限制了这种实现方式的训练精度。在此,我们提出一种混合精度架构,该架构将执行加权求和和不精确电导更新的计算存储单元与以高精度累积权重更新的数字处理单元相结合。基于所提出架构、使用相变存储器(PCM)阵列的多层感知器的硬件/软件联合训练实验,在手写数字分类任务(基于MNIST数据集)上实现了97.73%的测试准确率,与软件基线的差距在0.6%以内。该架构还使用PCM的精确行为模型在广泛的网络类型上进行了进一步评估,即卷积神经网络、长短时记忆网络和生成对抗网络。在不受PCM器件相关非理想因素限制的情况下,实现了与浮点实现相当的准确率。一项系统级研究表明,与专用的全数字32位实现相比,该架构用于训练多层感知器时能量效率提高了172倍。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2ec1/7235420/4d8d505f630a/fnins-14-00406-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2ec1/7235420/d27c4d9c4ee4/fnins-14-00406-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2ec1/7235420/05909eedde13/fnins-14-00406-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2ec1/7235420/e67849961cfd/fnins-14-00406-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2ec1/7235420/4d8d505f630a/fnins-14-00406-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2ec1/7235420/d27c4d9c4ee4/fnins-14-00406-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2ec1/7235420/05909eedde13/fnins-14-00406-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2ec1/7235420/e67849961cfd/fnins-14-00406-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2ec1/7235420/4d8d505f630a/fnins-14-00406-g0004.jpg

相似文献

1
Mixed-Precision Deep Learning Based on Computational Memory.基于计算内存的混合精度深度学习
Front Neurosci. 2020 May 12;14:406. doi: 10.3389/fnins.2020.00406. eCollection 2020.
2
Fully parallel write/read in resistive synaptic array for accelerating on-chip learning.用于加速片上学习的电阻式突触阵列的全并行读写。
Nanotechnology. 2015 Nov 13;26(45):455204. doi: 10.1088/0957-4484/26/45/455204.
3
Memristors for Neuromorphic Circuits and Artificial Intelligence Applications.用于神经形态电路和人工智能应用的忆阻器
Materials (Basel). 2020 Feb 20;13(4):938. doi: 10.3390/ma13040938.
4
A fully hardware-based memristive multilayer neural network.一种完全基于硬件的忆阻式多层神经网络。
Sci Adv. 2021 Nov 26;7(48):eabj4801. doi: 10.1126/sciadv.abj4801. Epub 2021 Nov 24.
5
Spiking CMOS-NVM mixed-signal neuromorphic ConvNet with circuit- and training-optimized temporal subsampling.具有电路和训练优化时间下采样的尖峰CMOS-NVM混合信号神经形态卷积网络
Front Neurosci. 2023 Jul 18;17:1177592. doi: 10.3389/fnins.2023.1177592. eCollection 2023.
6
An energy-efficient in-memory computing architecture for survival data analysis based on resistive switching memories.一种基于电阻式开关存储器的用于生存数据分析的节能内存计算架构。
Front Neurosci. 2022 Aug 9;16:932270. doi: 10.3389/fnins.2022.932270. eCollection 2022.
7
Synapse-Mimetic Hardware-Implemented Resistive Random-Access Memory for Artificial Neural Network.用于人工神经网络的突触模拟硬件实现的电阻式随机存取存储器。
Sensors (Basel). 2023 Mar 14;23(6):3118. doi: 10.3390/s23063118.
8
Memristive Quantized Neural Networks: A Novel Approach to Accelerate Deep Learning On-Chip.忆阻量化神经网络:一种加速芯片上深度学习的新方法。
IEEE Trans Cybern. 2021 Apr;51(4):1875-1887. doi: 10.1109/TCYB.2019.2912205. Epub 2021 Mar 17.
9
Sign backpropagation: An on-chip learning algorithm for analog RRAM neuromorphic computing systems.符号反向传播:一种用于模拟 RRAM 神经形态计算系统的片上学习算法。
Neural Netw. 2018 Dec;108:217-223. doi: 10.1016/j.neunet.2018.08.012. Epub 2018 Sep 1.
10
Thousands of conductance levels in memristors integrated on CMOS.在 CMOS 上集成的数千个电导水平的忆阻器。
Nature. 2023 Mar;615(7954):823-829. doi: 10.1038/s41586-023-05759-5. Epub 2023 Mar 29.

引用本文的文献

1
A near-threshold memristive computing-in-memory engine for edge intelligence.一种用于边缘智能的近阈值忆阻存内计算引擎。
Nat Commun. 2025 Jul 1;16(1):5897. doi: 10.1038/s41467-025-61025-4.
2
Phase-Change Memory for In-Memory Computing.用于内存计算的相变存储器
Chem Rev. 2025 Jun 11;125(11):5163-5194. doi: 10.1021/acs.chemrev.4c00670. Epub 2025 May 22.
3
Enhancing in-situ updates of quantized memristor neural networks: a Siamese network learning approach.增强量化忆阻器神经网络的原位更新:一种暹罗网络学习方法。

本文引用的文献

1
Accurate deep neural network inference using computational phase-change memory.利用计算相变化内存实现精确的深度神经网络推理。
Nat Commun. 2020 May 18;11(1):2473. doi: 10.1038/s41467-020-16108-9.
2
Memory devices and applications for in-memory computing.用于内存计算的存储设备和应用。
Nat Nanotechnol. 2020 Jul;15(7):529-544. doi: 10.1038/s41565-020-0655-z. Epub 2020 Mar 30.
3
Digital Biologically Plausible Implementation of Binarized Neural Networks With Differential Hafnium Oxide Resistive Memory Arrays.基于氧化铪差分电阻式存储器阵列的二值化神经网络的数字生物合理实现
Cogn Neurodyn. 2024 Aug;18(4):2047-2059. doi: 10.1007/s11571-024-10069-1. Epub 2024 Feb 13.
4
The backpropagation algorithm implemented on spiking neuromorphic hardware.在脉冲神经形态硬件上实现的反向传播算法。
Nat Commun. 2024 Nov 8;15(1):9691. doi: 10.1038/s41467-024-53827-9.
5
Fast and robust analog in-memory deep neural network training.快速且稳健的模拟内存深度神经网络训练
Nat Commun. 2024 Aug 20;15(1):7133. doi: 10.1038/s41467-024-51221-z.
6
Training neural networks using physical equations of motion.使用运动物理方程训练神经网络。
Proc Natl Acad Sci U S A. 2024 Jul 23;121(30):e2411913121. doi: 10.1073/pnas.2411913121. Epub 2024 Jul 15.
7
A comprehensive review of advanced trends: from artificial synapses to neuromorphic systems with consideration of non-ideal effects.对先进趋势的全面综述:从人工突触到考虑非理想效应的神经形态系统。
Front Neurosci. 2024 Apr 10;18:1279708. doi: 10.3389/fnins.2024.1279708. eCollection 2024.
8
Hardware implementation of memristor-based artificial neural networks.基于忆阻器的人工神经网络的硬件实现。
Nat Commun. 2024 Mar 4;15(1):1974. doi: 10.1038/s41467-024-45670-9.
9
Enabling Training of Neural Networks on Noisy Hardware.在有噪声的硬件上实现神经网络训练。
Front Artif Intell. 2021 Sep 9;4:699148. doi: 10.3389/frai.2021.699148. eCollection 2021.
10
Neural Network Training Acceleration With RRAM-Based Hybrid Synapses.基于阻变随机存取存储器(RRAM)的混合突触实现神经网络训练加速
Front Neurosci. 2021 Jun 24;15:690418. doi: 10.3389/fnins.2021.690418. eCollection 2021.
Front Neurosci. 2020 Jan 9;13:1383. doi: 10.3389/fnins.2019.01383. eCollection 2019.
4
Physics-based modeling approaches of resistive switching devices for memory and in-memory computing applications.用于存储器和内存计算应用的电阻式开关器件的基于物理的建模方法。
J Comput Electron. 2017;16(4):1121-1143. doi: 10.1007/s10825-017-1101-9. Epub 2017 Nov 13.
5
Training LSTM Networks With Resistive Cross-Point Devices.使用电阻式交叉点器件训练长短期记忆网络
Front Neurosci. 2018 Oct 24;12:745. doi: 10.3389/fnins.2018.00745. eCollection 2018.
6
High-Performance Mixed-Signal Neurocomputing With Nanoscale Floating-Gate Memory Cell Arrays.采用纳米级浮栅存储单元阵列的高性能混合信号神经计算
IEEE Trans Neural Netw Learn Syst. 2018 Oct;29(10):4782-4790. doi: 10.1109/TNNLS.2017.2778940. Epub 2017 Dec 22.
7
Neuromorphic computing with multi-memristive synapses.基于多忆阻器突触的神经形态计算。
Nat Commun. 2018 Jun 28;9(1):2514. doi: 10.1038/s41467-018-04933-y.
8
Efficient and self-adaptive in-situ learning in multilayer memristor neural networks.多层忆阻器神经网络中的高效自适应原位学习。
Nat Commun. 2018 Jun 19;9(1):2385. doi: 10.1038/s41467-018-04484-2.
9
Equivalent-accuracy accelerated neural-network training using analogue memory.利用模拟内存实现等效精度的加速神经网络训练。
Nature. 2018 Jun;558(7708):60-67. doi: 10.1038/s41586-018-0180-5. Epub 2018 Jun 6.
10
Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices.使用电阻式交叉点器件训练深度卷积神经网络。
Front Neurosci. 2017 Oct 10;11:538. doi: 10.3389/fnins.2017.00538. eCollection 2017.