SMILE Lab, School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 610031, China.
Center for Quantum Software and Information, University of Technology Sydney, NSW 2007, Australia.
Neural Netw. 2020 Oct;130:11-21. doi: 10.1016/j.neunet.2020.05.034. Epub 2020 Jun 7.
Deep neural networks (DNNs) have achieved outstanding performance in a wide range of applications, e.g., image classification, natural language processing, etc. Despite the good performance, the huge number of parameters in DNNs brings challenges to efficient training of DNNs and also their deployment in low-end devices with limited computing resources. In this paper, we explore the correlations in the weight matrices, and approximate the weight matrices with the low-rank block-term tensors. We name the new corresponding structure as block-term tensor layers (BT-layers), which can be easily adapted to neural network models, such as CNNs and RNNs. In particular, the inputs and the outputs in BT-layers are reshaped into low-dimensional high-order tensors with a similar or improved representation power. Sufficient experiments have demonstrated that BT-layers in CNNs and RNNs can achieve a very large compression ratio on the number of parameters while preserving or improving the representation power of the original DNNs.
深度神经网络(DNN)在图像分类、自然语言处理等广泛的应用中取得了优异的性能。尽管性能良好,但 DNN 中的大量参数给 DNN 的有效训练以及在计算资源有限的低端设备中的部署带来了挑战。在本文中,我们探索了权重矩阵之间的相关性,并使用低秩分块张量来近似权重矩阵。我们将新的对应结构命名为分块张量层(BT 层),它可以很容易地适应于神经网络模型,如卷积神经网络(CNN)和循环神经网络(RNN)。特别地,BT 层中的输入和输出被重塑为具有相似或改进表示能力的低维高阶张量。充分的实验表明,CNN 和 RNN 中的 BT 层可以在保持或提高原始 DNN 的表示能力的同时,对参数数量实现非常大的压缩比。