• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于高效深度学习的稳定张量神经网络。

Stable tensor neural networks for efficient deep learning.

作者信息

Newman Elizabeth, Horesh Lior, Avron Haim, Kilmer Misha E

机构信息

Department of Mathematics, Emory University, Atlanta, GA, United States.

Mathematics and Theoretical Computer Science, IBM TJ Watson Research Center, Yorktown, NY, United States.

出版信息

Front Big Data. 2024 May 30;7:1363978. doi: 10.3389/fdata.2024.1363978. eCollection 2024.

DOI:10.3389/fdata.2024.1363978
PMID:38873283
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11170703/
Abstract

Learning from complex, multidimensional data has become central to computational mathematics, and among the most successful high-dimensional function approximators are deep neural networks (DNNs). Training DNNs is posed as an optimization problem to learn network weights or parameters that well-approximate a mapping from input to target data. Multiway data or tensors arise naturally in myriad ways in deep learning, in particular as input data and as high-dimensional weights and features extracted by the network, with the latter often being a bottleneck in terms of speed and memory. In this work, we leverage tensor representations and processing to efficiently parameterize DNNs when learning from high-dimensional data. We propose tensor neural networks (t-NNs), a natural extension of traditional fully-connected networks, that can be trained efficiently in a reduced, yet more powerful parameter space. Our t-NNs are built upon matrix-mimetic tensor-tensor products, which retain algebraic properties of matrix multiplication while capturing high-dimensional correlations. Mimeticity enables t-NNs to inherit desirable properties of modern DNN architectures. We exemplify this by extending recent work on stable neural networks, which interpret DNNs as discretizations of differential equations, to our multidimensional framework. We provide empirical evidence of the parametric advantages of t-NNs on dimensionality reduction using autoencoders and classification using fully-connected and stable variants on benchmark imaging datasets MNIST and CIFAR-10.

摘要

从复杂的多维数据中学习已成为计算数学的核心,而深度神经网络(DNN)是最成功的高维函数逼近器之一。训练DNN被视为一个优化问题,即学习能很好地逼近从输入数据到目标数据映射的网络权重或参数。多路数据或张量在深度学习中以多种方式自然出现,特别是作为输入数据以及作为网络提取的高维权重和特征,而后者在速度和内存方面往往是一个瓶颈。在这项工作中,我们在从高维数据学习时利用张量表示和处理来有效地参数化DNN。我们提出了张量神经网络(t-NN),它是传统全连接网络的自然扩展,可以在一个简化但更强大的参数空间中进行高效训练。我们的t-NN基于模仿矩阵的张量-张量积构建,它在捕获高维相关性的同时保留了矩阵乘法的代数性质。模仿性使t-NN能够继承现代DNN架构的理想特性。我们通过将最近关于稳定神经网络的工作扩展到我们的多维框架来举例说明这一点,该工作将DNN解释为微分方程的离散化。我们在基准成像数据集MNIST和CIFAR-10上,使用自动编码器进行降维和使用全连接及稳定变体进行分类,提供了t-NN参数优势的实证证据。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/e90e3752b259/fdata-07-1363978-g0013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/306374090df5/fdata-07-1363978-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/60f89602a2b7/fdata-07-1363978-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/f023e1a16f1e/fdata-07-1363978-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/dda192add991/fdata-07-1363978-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/f4704d3b57c7/fdata-07-1363978-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/39e3c2eee291/fdata-07-1363978-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/aeefdc9b02c9/fdata-07-1363978-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/58c98d3bf113/fdata-07-1363978-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/8af59bbb6e2f/fdata-07-1363978-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/801775f5060d/fdata-07-1363978-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/d2da3a6fcefd/fdata-07-1363978-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/43ce89028e88/fdata-07-1363978-g0012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/e90e3752b259/fdata-07-1363978-g0013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/306374090df5/fdata-07-1363978-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/60f89602a2b7/fdata-07-1363978-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/f023e1a16f1e/fdata-07-1363978-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/dda192add991/fdata-07-1363978-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/f4704d3b57c7/fdata-07-1363978-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/39e3c2eee291/fdata-07-1363978-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/aeefdc9b02c9/fdata-07-1363978-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/58c98d3bf113/fdata-07-1363978-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/8af59bbb6e2f/fdata-07-1363978-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/801775f5060d/fdata-07-1363978-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/d2da3a6fcefd/fdata-07-1363978-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/43ce89028e88/fdata-07-1363978-g0012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cfe4/11170703/e90e3752b259/fdata-07-1363978-g0013.jpg

相似文献

1
Stable tensor neural networks for efficient deep learning.用于高效深度学习的稳定张量神经网络。
Front Big Data. 2024 May 30;7:1363978. doi: 10.3389/fdata.2024.1363978. eCollection 2024.
2
Nonlinear tensor train format for deep neural network compression.非线性张量火车格式用于深度神经网络压缩。
Neural Netw. 2021 Dec;144:320-333. doi: 10.1016/j.neunet.2021.08.028. Epub 2021 Sep 8.
3
Compact Neural Architecture Designs by Tensor Representations.基于张量表示的紧凑型神经架构设计
Front Artif Intell. 2022 Mar 8;5:728761. doi: 10.3389/frai.2022.728761. eCollection 2022.
4
Block-term tensor neural networks.块张量神经网络。
Neural Netw. 2020 Oct;130:11-21. doi: 10.1016/j.neunet.2020.05.034. Epub 2020 Jun 7.
5
Synergistic Integration of Deep Neural Networks and Finite Element Method with Applications of Nonlinear Large Deformation Biomechanics.深度神经网络与有限元方法在非线性大变形生物力学中的协同集成及应用
Comput Methods Appl Mech Eng. 2023 Nov 1;416. doi: 10.1016/j.cma.2023.116347. Epub 2023 Aug 22.
6
Variational tensor neural networks for deep learning.用于深度学习的变分张量神经网络。
Sci Rep. 2024 Aug 16;14(1):19017. doi: 10.1038/s41598-024-69366-8.
7
Symbolic Deep Networks: A Psychologically Inspired Lightweight and Efficient Approach to Deep Learning.符号深度学习网络:一种受心理学启发的轻量级高效深度学习方法。
Top Cogn Sci. 2022 Oct;14(4):702-717. doi: 10.1111/tops.12571. Epub 2021 Oct 5.
8
Compression of Deep Neural Networks based on quantized tensor decomposition to implement on reconfigurable hardware platforms.基于张量分解量化的深度神经网络压缩及其在可重构硬件平台上的实现。
Neural Netw. 2022 Jun;150:350-363. doi: 10.1016/j.neunet.2022.02.024. Epub 2022 Mar 8.
9
A Gradient-Guided Evolutionary Approach to Training Deep Neural Networks.基于梯度的深度神经网络进化训练方法。
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4861-4875. doi: 10.1109/TNNLS.2021.3061630. Epub 2022 Aug 31.
10
Representational Distance Learning for Deep Neural Networks.深度神经网络的代表性距离学习
Front Comput Neurosci. 2016 Dec 27;10:131. doi: 10.3389/fncom.2016.00131. eCollection 2016.

引用本文的文献

1
Nanoscale Titanium Oxide Memristive Structures for Neuromorphic Applications: Atomic Force Anodization Techniques, Modeling, Chemical Composition, and Resistive Switching Properties.用于神经形态应用的纳米级氧化钛忆阻结构:原子力阳极氧化技术、建模、化学成分及电阻开关特性
Nanomaterials (Basel). 2025 Jan 6;15(1):75. doi: 10.3390/nano15010075.
2
A Tensor Space for Multi-View and Multitask Learning Based on Einstein and Hadamard Products: A Case Study on Vehicle Traffic Surveillance Systems.基于爱因斯坦积和哈达玛积的用于多视图和多任务学习的张量空间:以车辆交通监测系统为例
Sensors (Basel). 2024 Nov 22;24(23):7463. doi: 10.3390/s24237463.

本文引用的文献

1
Dimensionality reduction of longitudinal 'omics data using modern tensor factorizations.利用现代张量分解进行纵向 'omics 数据的降维。
PLoS Comput Biol. 2022 Jul 15;18(7):e1010212. doi: 10.1371/journal.pcbi.1010212. eCollection 2022 Jul.
2
Tensor-tensor algebra for optimal representation and compression of multiway data.张量张量代数在多向数据的最优表示和压缩中的应用。
Proc Natl Acad Sci U S A. 2021 Jul 13;118(28). doi: 10.1073/pnas.2015851118.
3
Predicting clustered weather patterns: A test case for applications of convolutional neural networks to spatio-temporal climate data.
预测聚类天气模式:卷积神经网络在时空气候数据中应用的一个测试案例。
Sci Rep. 2020 Jan 28;10(1):1317. doi: 10.1038/s41598-020-57897-9.
4
Tensor-Factorized Neural Networks.张量分解神经网络。
IEEE Trans Neural Netw Learn Syst. 2018 May;29(5):1998-2011. doi: 10.1109/TNNLS.2017.2690379. Epub 2017 Apr 17.
5
Learning long-term dependencies with gradient descent is difficult.使用梯度下降法学习长期依赖关系是困难的。
IEEE Trans Neural Netw. 1994;5(2):157-66. doi: 10.1109/72.279181.
6
A tensor higher-order singular value decomposition for integrative analysis of DNA microarray data from different studies.一种用于不同研究的DNA微阵列数据综合分析的张量高阶奇异值分解
Proc Natl Acad Sci U S A. 2007 Nov 20;104(47):18371-6. doi: 10.1073/pnas.0709146104. Epub 2007 Nov 14.
7
Some mathematical notes on three-mode factor analysis.关于三模式因子分析的一些数学注释。
Psychometrika. 1966 Sep;31(3):279-311. doi: 10.1007/BF02289464.