• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于深度学习的变分张量神经网络。

Variational tensor neural networks for deep learning.

作者信息

Jahromi Saeed S, Orús Román

机构信息

Department of Physics, Institute for Advanced Studies in Basic Sciences (IASBS), Zanjan, 45137-66731, Iran.

Donostia International Physics Center, Paseo Manuel de Lardizabal 4, 20018, San Sebastián, Spain.

出版信息

Sci Rep. 2024 Aug 16;14(1):19017. doi: 10.1038/s41598-024-69366-8.

DOI:10.1038/s41598-024-69366-8
PMID:39152160
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11329717/
Abstract

Deep neural networks (NNs) encounter scalability limitations when confronted with a vast array of neurons, thereby constraining their achievable network depth. To address this challenge, we propose an integration of tensor networks (TN) into NN frameworks, combined with a variational DMRG-inspired training technique. This in turn, results in a scalable tensor neural network (TNN) architecture capable of efficient training over a large parameter space. Our variational algorithm utilizes a local gradient-descent technique, enabling manual or automatic computation of tensor gradients, facilitating design of hybrid TNN models with combined dense and tensor layers. Our training algorithm further provides insight on the entanglement structure of the tensorized trainable weights and correlation among the model parameters. We validate the accuracy and efficiency of our method by designing TNN models and providing benchmark results for linear and non-linear regressions, data classification and image recognition on MNIST handwritten digits.

摘要

当面对大量神经元时,深度神经网络(NNs)会遇到可扩展性限制,从而限制了它们能够达到的网络深度。为应对这一挑战,我们提出将张量网络(TN)集成到NN框架中,并结合一种受变分密度矩阵重整化群(DMRG)启发的训练技术。这进而产生了一种可扩展的张量神经网络(TNN)架构,能够在大参数空间上进行高效训练。我们的变分算法利用局部梯度下降技术,实现张量梯度的手动或自动计算,便于设计具有密集层和张量层组合的混合TNN模型。我们的训练算法还进一步揭示了张量化可训练权重的纠缠结构以及模型参数之间的相关性。我们通过设计TNN模型并提供线性和非线性回归、数据分类以及MNIST手写数字图像识别的基准结果,验证了我们方法的准确性和效率。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/14b1b55bbb9c/41598_2024_69366_Fig19_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/b823f04aecb3/41598_2024_69366_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/590efd6e62a8/41598_2024_69366_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/a5fc4b09823e/41598_2024_69366_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/6dabe734941b/41598_2024_69366_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/8c4d69b52a6f/41598_2024_69366_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/0239b2fc075b/41598_2024_69366_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/a8f45c10374f/41598_2024_69366_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/5a115d813578/41598_2024_69366_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/e15e0cb3cdc1/41598_2024_69366_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/59912987560c/41598_2024_69366_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/15e2022864f2/41598_2024_69366_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/afcf4d12ac67/41598_2024_69366_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/ee1a7771cd83/41598_2024_69366_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/307c04abcdec/41598_2024_69366_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/081c74cd95a5/41598_2024_69366_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/cb1f42b2d2bf/41598_2024_69366_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/63c7eaafb19b/41598_2024_69366_Fig17_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/b4ecd7a4ef17/41598_2024_69366_Fig18_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/14b1b55bbb9c/41598_2024_69366_Fig19_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/b823f04aecb3/41598_2024_69366_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/590efd6e62a8/41598_2024_69366_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/a5fc4b09823e/41598_2024_69366_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/6dabe734941b/41598_2024_69366_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/8c4d69b52a6f/41598_2024_69366_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/0239b2fc075b/41598_2024_69366_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/a8f45c10374f/41598_2024_69366_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/5a115d813578/41598_2024_69366_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/e15e0cb3cdc1/41598_2024_69366_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/59912987560c/41598_2024_69366_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/15e2022864f2/41598_2024_69366_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/afcf4d12ac67/41598_2024_69366_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/ee1a7771cd83/41598_2024_69366_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/307c04abcdec/41598_2024_69366_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/081c74cd95a5/41598_2024_69366_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/cb1f42b2d2bf/41598_2024_69366_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/63c7eaafb19b/41598_2024_69366_Fig17_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/b4ecd7a4ef17/41598_2024_69366_Fig18_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11329717/14b1b55bbb9c/41598_2024_69366_Fig19_HTML.jpg

相似文献

1
Variational tensor neural networks for deep learning.用于深度学习的变分张量神经网络。
Sci Rep. 2024 Aug 16;14(1):19017. doi: 10.1038/s41598-024-69366-8.
2
General-Purpose Bayesian Tensor Learning With Automatic Rank Determination and Uncertainty Quantification.具有自动秩确定和不确定性量化的通用贝叶斯张量学习
Front Artif Intell. 2022 Jan 7;4:668353. doi: 10.3389/frai.2021.668353. eCollection 2021.
3
Analysis of Diffractive Optical Neural Networks and Their Integration with Electronic Neural Networks.衍射光学神经网络及其与电子神经网络集成的分析
IEEE J Sel Top Quantum Electron. 2020 Jan-Feb;26(1). doi: 10.1109/JSTQE.2019.2921376. Epub 2019 Jun 6.
4
Stable tensor neural networks for efficient deep learning.用于高效深度学习的稳定张量神经网络。
Front Big Data. 2024 May 30;7:1363978. doi: 10.3389/fdata.2024.1363978. eCollection 2024.
5
Deep convolutional neural network and IoT technology for healthcare.用于医疗保健的深度卷积神经网络和物联网技术。
Digit Health. 2024 Jan 17;10:20552076231220123. doi: 10.1177/20552076231220123. eCollection 2024 Jan-Dec.
6
Diffractive deep neural network adjoint assist or (DNA): a fast and efficient nonlinear diffractive neural network implementation.衍射深度神经网络伴随辅助或 (DNA):一种快速高效的非线性衍射神经网络实现方法。
Opt Express. 2022 Feb 28;30(5):7441-7456. doi: 10.1364/OE.449415.
7
The deep arbitrary polynomial chaos neural network or how Deep Artificial Neural Networks could benefit from data-driven homogeneous chaos theory.深度任意多项式混沌神经网络或深度人工神经网络如何从数据驱动的均匀混沌理论中受益。
Neural Netw. 2023 Sep;166:85-104. doi: 10.1016/j.neunet.2023.06.036. Epub 2023 Jul 10.
8
Fast, Simple and Accurate Handwritten Digit Classification by Training Shallow Neural Network Classifiers with the 'Extreme Learning Machine' Algorithm.通过使用“极限学习机”算法训练浅层神经网络分类器实现快速、简单且准确的手写数字分类
PLoS One. 2015 Aug 11;10(8):e0134254. doi: 10.1371/journal.pone.0134254. eCollection 2015.
9
Hierarchical Bayesian Inference and Learning in Spiking Neural Networks.分层贝叶斯推断和尖峰神经网络中的学习。
IEEE Trans Cybern. 2019 Jan;49(1):133-145. doi: 10.1109/TCYB.2017.2768554. Epub 2017 Nov 9.
10
DEBI-NN: Distance-encoding biomorphic-informational neural networks for minimizing the number of trainable parameters.DEBI-NN:用于最小化可训练参数数量的距离编码生物形态信息神经网络。
Neural Netw. 2023 Oct;167:517-532. doi: 10.1016/j.neunet.2023.08.026. Epub 2023 Aug 25.

本文引用的文献

1
Randomness-Enhanced Expressivity of Quantum Neural Networks.量子神经网络的随机性增强表现力
Phys Rev Lett. 2024 Jan 5;132(1):010602. doi: 10.1103/PhysRevLett.132.010602.
2
Machine learning for condensed matter physics.用于凝聚态物理的机器学习。
J Phys Condens Matter. 2020 Nov 5;33(5). doi: 10.1088/1361-648X/abb895.
3
Self-Supervised Visual Feature Learning With Deep Neural Networks: A Survey.基于深度神经网络的自监督视觉特征学习:综述
IEEE Trans Pattern Anal Mach Intell. 2021 Nov;43(11):4037-4058. doi: 10.1109/TPAMI.2020.2992393. Epub 2021 Oct 1.
4
A high-bias, low-variance introduction to Machine Learning for physicists.面向物理学家的机器学习高偏差、低方差入门介绍。
Phys Rep. 2019 May 30;810:1-124. doi: 10.1016/j.physrep.2019.03.001. Epub 2019 Mar 14.
5
A Comprehensive Analysis of Deep Regression.深度回归分析的全面分析。
IEEE Trans Pattern Anal Mach Intell. 2020 Sep;42(9):2065-2081. doi: 10.1109/TPAMI.2019.2910523. Epub 2019 Apr 11.
6
Robust Kronecker Component Analysis.稳健克罗内克成分分析
IEEE Trans Pattern Anal Mach Intell. 2019 Oct;41(10):2365-2379. doi: 10.1109/TPAMI.2018.2881476. Epub 2018 Nov 15.
7
Machine learning quantum phases of matter beyond the fermion sign problem.机器学习超越费米子符号问题的物质量子相。
Sci Rep. 2017 Aug 18;7(1):8823. doi: 10.1038/s41598-017-09098-0.
8
Efficient simulation of one-dimensional quantum many-body systems.一维量子多体系统的高效模拟
Phys Rev Lett. 2004 Jul 23;93(4):040502. doi: 10.1103/PhysRevLett.93.040502. Epub 2004 Jul 19.
9
Efficient classical simulation of slightly entangled quantum computations.轻度纠缠量子计算的高效经典模拟。
Phys Rev Lett. 2003 Oct 3;91(14):147902. doi: 10.1103/PhysRevLett.91.147902. Epub 2003 Oct 1.
10
Reward-dependent learning in neuronal networks for planning and decision making.用于规划和决策的神经网络中基于奖励的学习。
Prog Brain Res. 2000;126:217-29. doi: 10.1016/S0079-6123(00)26016-0.