• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

源自现代霍普菲尔德网络的玻尔兹曼机家族中的注意力机制

Attention in a Family of Boltzmann Machines Emerging From Modern Hopfield Networks.

作者信息

Ota Toshihiro, Karakida Ryo

机构信息

CyberAgent, Shibuya, Tokyo, 150-6121, Japan.

RIKEN, Wako, Saitama, 351-0198, Japan

出版信息

Neural Comput. 2023 Jun 20:1-18. doi: 10.1162/neco_a_01597.

DOI:10.1162/neco_a_01597
PMID:37432863
Abstract

Hopfield networks and Boltzmann machines (BMs) are fundamental energy-based neural network models. Recent studies on modern Hopfield networks have broadened the class of energy functions and led to a unified perspective on general Hopfield networks, including an attention module. In this letter, we consider the BM counterparts of modern Hopfield networks using the associated energy functions and study their salient properties from a trainability perspective. In particular, the energy function corresponding to the attention module naturally introduces a novel BM, which we refer to as the attentional BM (AttnBM). We verify that AttnBM has a tractable likelihood function and gradient for certain special cases and is easy to train. Moreover, we reveal the hidden connections between AttnBM and some single-layer models, namely the gaussian-Bernoulli restricted BM and the denoising autoencoder with softmax units coming from denoising score matching. We also investigate BMs introduced by other energy functions and show that the energy function of dense associative memory models gives BMs belonging to exponential family harmoniums.

摘要

霍普菲尔德网络和玻尔兹曼机(BMs)是基于能量的基本神经网络模型。近期对现代霍普菲尔德网络的研究拓宽了能量函数的类别,并为包括注意力模块在内的通用霍普菲尔德网络带来了统一的视角。在这封信中,我们使用相关的能量函数来考虑现代霍普菲尔德网络的BM对应物,并从可训练性的角度研究它们的显著特性。特别地,与注意力模块对应的能量函数自然地引入了一种新颖的BM,我们将其称为注意力玻尔兹曼机(AttnBM)。我们验证了AttnBM在某些特殊情况下具有易于处理的似然函数和梯度,并且易于训练。此外,我们揭示了AttnBM与一些单层模型之间的隐藏联系,即高斯 - 伯努利受限玻尔兹曼机和来自去噪得分匹配的具有softmax单元的去噪自编码器。我们还研究了由其他能量函数引入的玻尔兹曼机,并表明密集联想记忆模型的能量函数给出了属于指数族和声琴的玻尔兹曼机。

相似文献

1
Attention in a Family of Boltzmann Machines Emerging From Modern Hopfield Networks.源自现代霍普菲尔德网络的玻尔兹曼机家族中的注意力机制
Neural Comput. 2023 Jun 20:1-18. doi: 10.1162/neco_a_01597.
2
On the equivalence of Hopfield networks and Boltzmann Machines.Hopfield 网络与 Boltzmann 机的等价性。
Neural Netw. 2012 Oct;34:1-9. doi: 10.1016/j.neunet.2012.06.003. Epub 2012 Jun 23.
3
Universal Hopfield Networks: A General Framework for Single-Shot Associative Memory Models.通用霍普菲尔德网络:单触发联想记忆模型的通用框架
Proc Mach Learn Res. 2022 Jul;162:15561-15583.
4
Boltzmann Machines as Generalized Hopfield Networks: A Review of Recent Results and Outlooks.作为广义霍普菲尔德网络的玻尔兹曼机:近期成果与展望综述
Entropy (Basel). 2020 Dec 29;23(1):34. doi: 10.3390/e23010034.
5
On the effective initialisation for restricted Boltzmann machines via duality with Hopfield model.通过对偶性与 Hopfield 模型实现受限 Boltzmann 机的有效初始化。
Neural Netw. 2021 Nov;143:314-326. doi: 10.1016/j.neunet.2021.06.017. Epub 2021 Jun 23.
6
On the importance of hidden bias and hidden entropy in representational efficiency of the Gaussian-Bipolar Restricted Boltzmann Machines.在高斯双极受限玻尔兹曼机的表示效率中,隐藏偏差和隐藏熵的重要性。
Neural Netw. 2018 Sep;105:405-418. doi: 10.1016/j.neunet.2018.06.002. Epub 2018 Jun 22.
7
Phase diagram of restricted Boltzmann machines and generalized Hopfield networks with arbitrary priors.受限玻尔兹曼机和广义霍普菲尔德网络的相图,具有任意先验。
Phys Rev E. 2018 Feb;97(2-1):022310. doi: 10.1103/PhysRevE.97.022310.
8
Approximation properties of Gaussian-binary restricted Boltzmann machines and Gaussian-binary deep belief networks.高斯-二项式受限玻尔兹曼机和高斯-二项式深度置信网络的逼近性质。
Neural Netw. 2022 Sep;153:49-63. doi: 10.1016/j.neunet.2022.05.020. Epub 2022 Jun 2.
9
A broad class of discrete-time hypercomplex-valued Hopfield neural networks.一类广义的离散时间超复数值 Hopfield 神经网络。
Neural Netw. 2020 Feb;122:54-67. doi: 10.1016/j.neunet.2019.09.040. Epub 2019 Oct 18.
10
In Search of Dispersed Memories: Generative Diffusion Models Are Associative Memory Networks.寻找分散记忆:生成扩散模型是联想记忆网络。
Entropy (Basel). 2024 Apr 29;26(5):381. doi: 10.3390/e26050381.