Suppr超能文献

块张量神经网络。

Block-term tensor neural networks.

机构信息

SMILE Lab, School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 610031, China.

Center for Quantum Software and Information, University of Technology Sydney, NSW 2007, Australia.

出版信息

Neural Netw. 2020 Oct;130:11-21. doi: 10.1016/j.neunet.2020.05.034. Epub 2020 Jun 7.

Abstract

Deep neural networks (DNNs) have achieved outstanding performance in a wide range of applications, e.g., image classification, natural language processing, etc. Despite the good performance, the huge number of parameters in DNNs brings challenges to efficient training of DNNs and also their deployment in low-end devices with limited computing resources. In this paper, we explore the correlations in the weight matrices, and approximate the weight matrices with the low-rank block-term tensors. We name the new corresponding structure as block-term tensor layers (BT-layers), which can be easily adapted to neural network models, such as CNNs and RNNs. In particular, the inputs and the outputs in BT-layers are reshaped into low-dimensional high-order tensors with a similar or improved representation power. Sufficient experiments have demonstrated that BT-layers in CNNs and RNNs can achieve a very large compression ratio on the number of parameters while preserving or improving the representation power of the original DNNs.

摘要

深度神经网络(DNN)在图像分类、自然语言处理等广泛的应用中取得了优异的性能。尽管性能良好,但 DNN 中的大量参数给 DNN 的有效训练以及在计算资源有限的低端设备中的部署带来了挑战。在本文中,我们探索了权重矩阵之间的相关性,并使用低秩分块张量来近似权重矩阵。我们将新的对应结构命名为分块张量层(BT 层),它可以很容易地适应于神经网络模型,如卷积神经网络(CNN)和循环神经网络(RNN)。特别地,BT 层中的输入和输出被重塑为具有相似或改进表示能力的低维高阶张量。充分的实验表明,CNN 和 RNN 中的 BT 层可以在保持或提高原始 DNN 的表示能力的同时,对参数数量实现非常大的压缩比。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验