• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

iNALU:改进型神经算术逻辑单元。

iNALU: Improved Neural Arithmetic Logic Unit.

作者信息

Schlör Daniel, Ring Markus, Hotho Andreas

机构信息

Data Science Chair, Institute of Computer Science, University of Wuerzburg, Würzburg, Germany.

Department of Electrical Engineering and Computer Science, University of Applied Sciences and Arts Coburg, Coburg, Germany.

出版信息

Front Artif Intell. 2020 Sep 29;3:71. doi: 10.3389/frai.2020.00071. eCollection 2020.

DOI:10.3389/frai.2020.00071
PMID:33733188
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7861275/
Abstract

Neural networks have to capture mathematical relationships in order to learn various tasks. They approximate these relations implicitly and therefore often do not generalize well. The recently proposed Neural Arithmetic Logic Unit (NALU) is a novel neural architecture which is able to explicitly represent the mathematical relationships by the units of the network to learn operations such as summation, subtraction or multiplication. Although NALUs have been shown to perform well on various downstream tasks, an in-depth analysis reveals practical shortcomings by design, such as the inability to multiply or divide negative input values or training stability issues for deeper networks. We address these issues and propose an improved model architecture. We evaluate our model empirically in various settings from learning basic arithmetic operations to more complex functions. Our experiments indicate that our model solves stability issues and outperforms the original NALU model in means of arithmetic precision and convergence.

摘要

神经网络必须捕捉数学关系以便学习各种任务。它们隐式地近似这些关系,因此通常泛化能力不佳。最近提出的神经算术逻辑单元(NALU)是一种新颖的神经架构,它能够通过网络单元明确表示数学关系,以学习诸如加法、减法或乘法等运算。尽管NALU已被证明在各种下游任务中表现良好,但深入分析发现其在设计上存在实际缺点,例如无法对负输入值进行乘法或除法运算,或者对于更深层次的网络存在训练稳定性问题。我们解决了这些问题并提出了一种改进的模型架构。我们在从学习基本算术运算到更复杂函数的各种设置中对我们的模型进行了实证评估。我们的实验表明,我们的模型解决了稳定性问题,并且在算术精度和收敛性方面优于原始的NALU模型。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/f08fc53446a7/frai-03-00071-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/96bcf0c564f9/frai-03-00071-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/1528a39a60cd/frai-03-00071-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/57c876b36c73/frai-03-00071-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/a83d0a081176/frai-03-00071-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/ae02f97ee0f1/frai-03-00071-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/f08fc53446a7/frai-03-00071-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/96bcf0c564f9/frai-03-00071-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/1528a39a60cd/frai-03-00071-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/57c876b36c73/frai-03-00071-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/a83d0a081176/frai-03-00071-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/ae02f97ee0f1/frai-03-00071-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b22/7861275/f08fc53446a7/frai-03-00071-g0006.jpg

相似文献

1
iNALU: Improved Neural Arithmetic Logic Unit.iNALU:改进型神经算术逻辑单元。
Front Artif Intell. 2020 Sep 29;3:71. doi: 10.3389/frai.2020.00071. eCollection 2020.
2
Universal Background Subtraction Based on Arithmetic Distribution Neural Network.基于算术分布神经网络的通用背景减法
IEEE Trans Image Process. 2022;31:2934-2949. doi: 10.1109/TIP.2022.3162961. Epub 2022 Apr 8.
3
Fast Approximations of Activation Functions in Deep Neural Networks when using Posit Arithmetic.当使用正算数时,深度神经网络中激活函数的快速逼近。
Sensors (Basel). 2020 Mar 10;20(5):1515. doi: 10.3390/s20051515.
4
Discontinuities in recurrent neural networks.循环神经网络中的不连续性。
Neural Comput. 1999 Apr 1;11(3):715-46. doi: 10.1162/089976699300016638.
5
Secure Outsourced Matrix Computation and Application to Neural Networks.安全外包矩阵计算及其在神经网络中的应用
Conf Comput Commun Secur. 2018 Oct;2018:1209-1222. doi: 10.1145/3243734.3243837.
6
Design of a DNA-based reversible arithmetic and logic unit.基于DNA的可逆算术逻辑单元的设计。
IET Nanobiotechnol. 2015 Aug;9(4):226-38. doi: 10.1049/iet-nbt.2014.0056.
7
An Evaluation of Power Side-Channel Resistance for RNS Secure Logic.基于余数数制(RNS)的安全逻辑的功耗侧信道抗性评估
Sensors (Basel). 2022 Mar 14;22(6):2242. doi: 10.3390/s22062242.
8
Exploring the Feasibility of a DNA Computer: Design of an ALU Using Sticker-Based DNA Model.探索DNA计算机的可行性:基于贴纸的DNA模型设计算术逻辑单元
IEEE Trans Nanobioscience. 2017 Sep;16(6):383-399. doi: 10.1109/TNB.2017.2726682. Epub 2017 Jul 13.
9
Elements for a general memory structure: properties of recurrent neural networks used to form situation models.通用记忆结构的要素:用于形成情境模型的递归神经网络的属性。
Biol Cybern. 2008 May;98(5):371-95. doi: 10.1007/s00422-008-0221-5. Epub 2008 Mar 19.
10
Neuromorphic computation with spiking memristors: habituation, experimental instantiation of logic gates and a novel sequence-sensitive perceptron model.基于尖峰神经元的忆阻器神经形态计算:习惯化、逻辑门的实验实例化以及一种新颖的序列敏感感知器模型。
Faraday Discuss. 2019 Feb 18;213(0):521-551. doi: 10.1039/c8fd00111a.