• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

QTTNet:用于3D物体和视频识别的量化张量列车神经网络

QTTNet: Quantized tensor train neural networks for 3D object and video recognition.

作者信息

Lee Donghyun, Wang Dingheng, Yang Yukuan, Deng Lei, Zhao Guangshe, Li Guoqi

机构信息

Department of Precision Instrumentation, Center for Brain Inspired Computing Research and Beijing Innovation Center for Future Chip, Tsinghua University, Beijing 100084, China.

School of Automation Science and Engineering, Faculty of Electronic and Information Engineering, Xi'an Jiaotong University, Xi'an, Shaanxi 710049, China.

出版信息

Neural Netw. 2021 Sep;141:420-432. doi: 10.1016/j.neunet.2021.05.034. Epub 2021 Jun 5.

DOI:10.1016/j.neunet.2021.05.034
PMID:34146969
Abstract

Relying on the rapidly increasing capacity of computing clusters and hardware, convolutional neural networks (CNNs) have been successfully applied in various fields and achieved state-of-the-art results. Despite these exciting developments, the huge memory cost is still involved in training and inferring a large-scale CNN model and makes it hard to be widely used in resource-limited portable devices. To address this problem, we establish a training framework for three-dimensional convolutional neural networks (3DCNNs) named QTTNet that combines tensor train (TT) decomposition and data quantization together for further shrinking the model size and decreasing the memory and time cost. Through this framework, we can fully explore the superiority of TT in reducing the number of trainable parameters and the advantage of quantization in decreasing the bit-width of data, particularly compressing 3DCNN model greatly with little accuracy degradation. In addition, due to the low bit quantization to all parameters during the inference process including TT-cores, activations, and batch normalizations, the proposed method naturally takes advantage in memory and time cost. Experimental results of compressing 3DCNNs for 3D object and video recognition on ModelNet40, UCF11, and UCF50 datasets verify the effectiveness of the proposed method. The best compression ratio we have obtained is up to nearly 180× with competitive performance compared with other state-of-the-art researches. Moreover, the total bytes of our QTTNet models on ModelNet40 and UCF11 datasets can be 1000× lower than some typical practices such as MVCNN.

摘要

依靠计算集群和硬件能力的快速增长,卷积神经网络(CNN)已成功应用于各个领域并取得了领先成果。尽管有这些令人兴奋的进展,但在训练和推断大规模CNN模型时仍涉及巨大的内存成本,这使得它难以在资源有限的便携式设备中广泛应用。为了解决这个问题,我们为三维卷积神经网络(3DCNN)建立了一个名为QTTNet的训练框架,该框架将张量列车(TT)分解和数据量化结合在一起,以进一步缩小模型大小并降低内存和时间成本。通过这个框架,我们可以充分探索TT在减少可训练参数数量方面的优势以及量化在降低数据位宽方面的优势,特别是在几乎不降低精度的情况下大幅压缩3DCNN模型。此外,由于在推理过程中对包括TT核、激活值和批量归一化在内的所有参数进行低比特量化,该方法在内存和时间成本方面具有天然优势。在ModelNet40、UCF11和UCF50数据集上对3DCNN进行3D对象和视频识别的压缩实验结果验证了该方法的有效性。我们获得的最佳压缩比高达近180倍,与其他最新研究相比具有竞争力。此外,我们的QTTNet模型在ModelNet40和UCF11数据集上的总字节数可能比一些典型方法(如MVCNN)低1000倍。

相似文献

1
QTTNet: Quantized tensor train neural networks for 3D object and video recognition.QTTNet:用于3D物体和视频识别的量化张量列车神经网络
Neural Netw. 2021 Sep;141:420-432. doi: 10.1016/j.neunet.2021.05.034. Epub 2021 Jun 5.
2
Compressing 3DCNNs based on tensor train decomposition.基于张量树分解的 3DCNN 压缩。
Neural Netw. 2020 Nov;131:215-230. doi: 10.1016/j.neunet.2020.07.028. Epub 2020 Aug 7.
3
Nonlinear tensor train format for deep neural network compression.非线性张量火车格式用于深度神经网络压缩。
Neural Netw. 2021 Dec;144:320-333. doi: 10.1016/j.neunet.2021.08.028. Epub 2021 Sep 8.
4
Hybrid tensor decomposition in neural network compression.神经网络压缩中的混合张量分解。
Neural Netw. 2020 Dec;132:309-320. doi: 10.1016/j.neunet.2020.09.006. Epub 2020 Sep 19.
5
MedQ: Lossless ultra-low-bit neural network quantization for medical image segmentation.MedQ:用于医学图像分割的无损超低比特神经网络量化。
Med Image Anal. 2021 Oct;73:102200. doi: 10.1016/j.media.2021.102200. Epub 2021 Aug 2.
6
FPGA-Based Hybrid-Type Implementation of Quantized Neural Networks for Remote Sensing Applications.基于 FPGA 的量化神经网络混合式实现及其在遥感中的应用。
Sensors (Basel). 2019 Feb 22;19(4):924. doi: 10.3390/s19040924.
7
A Hardware-Friendly Low-Bit Power-of-Two Quantization Method for CNNs and Its FPGA Implementation.一种面向硬件的 CNN 低比特数 2 的幂量化方法及其 FPGA 实现。
Sensors (Basel). 2022 Sep 1;22(17):6618. doi: 10.3390/s22176618.
8
Training high-performance and large-scale deep neural networks with full 8-bit integers.用全 8 位整数训练高性能和大规模深度神经网络。
Neural Netw. 2020 May;125:70-82. doi: 10.1016/j.neunet.2019.12.027. Epub 2020 Jan 15.
9
A Novel Low-Bit Quantization Strategy for Compressing Deep Neural Networks.一种用于压缩深度神经网络的新型低比特量化策略。
Comput Intell Neurosci. 2020 Feb 18;2020:7839064. doi: 10.1155/2020/7839064. eCollection 2020.
10
Pre-Computing Batch Normalisation Parameters for Edge Devices on a Binarized Neural Network.在二值化神经网络上为边缘设备预计算批量归一化参数。
Sensors (Basel). 2023 Jun 14;23(12):5556. doi: 10.3390/s23125556.

引用本文的文献

1
Compact Neural Architecture Designs by Tensor Representations.基于张量表示的紧凑型神经架构设计
Front Artif Intell. 2022 Mar 8;5:728761. doi: 10.3389/frai.2022.728761. eCollection 2022.