• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

学习内容加权深度图像压缩

Learning Content-Weighted Deep Image Compression.

作者信息

Li Mu, Zuo Wangmeng, Gu Shuhang, You Jane, Zhang David

出版信息

IEEE Trans Pattern Anal Mach Intell. 2021 Oct;43(10):3446-3461. doi: 10.1109/TPAMI.2020.2983926. Epub 2021 Sep 2.

DOI:10.1109/TPAMI.2020.2983926
PMID:32248094
Abstract

Learning-based lossy image compression usually involves the joint optimization of rate-distortion performance, and requires to cope with the spatial variation of image content and contextual dependence among learned codes. Traditional entropy models can spatially adapt the local bit rate based on the image content, but usually are limited in exploiting context in code space. On the other hand, most deep context models are computationally very expensive and cannot efficiently perform decoding over the symbols in parallel. In this paper, we present a content-weighted encoder-decoder model, where the channel-wise multi-valued quantization is deployed for the discretization of the encoder features, and an importance map subnet is introduced to generate the importance masks for spatially varying code pruning. Consequently, the summation of importance masks can serve as an upper bound of the length of bitstream. Furthermore, the quantized representations of the learned code and importance map are still spatially dependent, which can be losslessly compressed using arithmetic coding. To compress the codes effectively and efficiently, we propose an upper-triangular masked convolutional network (triuMCN) for large context modeling. Experiments show that the proposed method can produce visually much better results, and performs favorably against deep and traditional lossy image compression approaches.

摘要

基于学习的有损图像压缩通常涉及率失真性能的联合优化,并且需要应对图像内容的空间变化以及学习到的代码之间的上下文依赖性。传统熵模型可以根据图像内容在空间上自适应局部比特率,但通常在利用代码空间中的上下文方面存在局限性。另一方面,大多数深度上下文模型计算成本非常高,并且不能有效地对符号进行并行解码。在本文中,我们提出了一种内容加权的编码器 - 解码器模型,其中针对编码器特征的离散化部署了通道级多值量化,并引入了一个重要性映射子网来生成用于空间变化代码剪枝的重要性掩码。因此,重要性掩码的总和可以用作比特流长度的上限。此外,学习到的代码和重要性映射的量化表示在空间上仍然相关,可以使用算术编码进行无损压缩。为了有效且高效地压缩代码,我们提出了一种用于大上下文建模的上三角掩码卷积网络(triuMCN)。实验表明,所提出的方法可以产生视觉上更好的结果,并且在与深度和传统有损图像压缩方法的比较中表现出色。

相似文献

1
Learning Content-Weighted Deep Image Compression.学习内容加权深度图像压缩
IEEE Trans Pattern Anal Mach Intell. 2021 Oct;43(10):3446-3461. doi: 10.1109/TPAMI.2020.2983926. Epub 2021 Sep 2.
2
Learning Context-Based Nonlocal Entropy Modeling for Image Compression.基于学习上下文的非局部熵图像压缩建模
IEEE Trans Neural Netw Learn Syst. 2023 Mar;34(3):1132-1145. doi: 10.1109/TNNLS.2021.3104974. Epub 2023 Feb 28.
3
Hierarchical Recurrent Neural Hashing for Image Retrieval With Hierarchical Convolutional Features.基于层次卷积特征的层次递归神经网络哈希图像检索
IEEE Trans Image Process. 2018;27(1):106-120. doi: 10.1109/TIP.2017.2755766.
4
Efficient and Effective Context-Based Convolutional Entropy Modeling for Image Compression.用于图像压缩的高效且有效的基于上下文的卷积熵建模
IEEE Trans Image Process. 2020 Apr 14. doi: 10.1109/TIP.2020.2985225.
5
End-to-end Optimized ROI Image Compression.端到端优化的感兴趣区域图像压缩
IEEE Trans Image Process. 2019 Dec 25. doi: 10.1109/TIP.2019.2960869.
6
End-to-End Optimized Versatile Image Compression With Wavelet-Like Transform.基于类小波变换的端到端优化通用图像压缩
IEEE Trans Pattern Anal Mach Intell. 2022 Mar;44(3):1247-1263. doi: 10.1109/TPAMI.2020.3026003. Epub 2022 Feb 3.
7
End-to-End Learnt Image Compression via Non-Local Attention Optimization and Improved Context Modeling.通过非局部注意力优化和改进的上下文建模实现端到端学习的图像压缩
IEEE Trans Image Process. 2021;30:3179-3191. doi: 10.1109/TIP.2021.3058615. Epub 2021 Feb 25.
8
Image Compression Based on Hybrid Domain Attention and Postprocessing Enhancement.基于混合域注意力和后处理增强的图像压缩。
Comput Intell Neurosci. 2022 Mar 17;2022:4926124. doi: 10.1155/2022/4926124. eCollection 2022.
9
QARV: Quantization-Aware ResNet VAE for Lossy Image Compression.QARV:用于有损图像压缩的量化感知残差网络变分自编码器
IEEE Trans Pattern Anal Mach Intell. 2024 Jan;46(1):436-450. doi: 10.1109/TPAMI.2023.3322904. Epub 2023 Dec 6.
10
Joint optimization of run-length coding, Huffman coding, and quantization table with complete baseline JPEG decoder compatibility.运行长度编码、哈夫曼编码和量化表的联合优化,与完整基线JPEG解码器完全兼容。
IEEE Trans Image Process. 2009 Jan;18(1):63-74. doi: 10.1109/TIP.2008.2007609.

引用本文的文献

1
Memristor-based storage system with convolutional autoencoder-based image compression network.基于忆阻器的存储系统与基于卷积自动编码器的图像压缩网络。
Nat Commun. 2024 Feb 7;15(1):1132. doi: 10.1038/s41467-024-45312-0.