Suppr超能文献

基于学习的感知视频编码的刚可察觉量化失真建模。

Learning-Based Just-Noticeable-Quantization- Distortion Modeling for Perceptual Video Coding.

出版信息

IEEE Trans Image Process. 2018 Jul;27(7):3178-3193. doi: 10.1109/TIP.2018.2818439.

Abstract

Conventional predictive video coding-based approaches are reaching the limit of their potential coding efficiency improvements, because of severely increasing computation complexity. As an alternative approach, perceptual video coding (PVC) has attempted to achieve high coding efficiency by eliminating perceptual redundancy, using just-noticeable-distortion (JND) directed PVC. The previous JNDs were modeled by adding white Gaussian noise or specific signal patterns into the original images, which were not appropriate in finding JND thresholds due to distortion with energy reduction. In this paper, we present a novel discrete cosine transform-based energy-reduced JND model, called ERJND, that is more suitable for JND-based PVC schemes. Then, the proposed ERJND model is extended to two learning-based just-noticeable-quantization-distortion (JNQD) models as preprocessing that can be applied for perceptual video coding. The two JNQD models can automatically adjust JND levels based on given quantization step sizes. One of the two JNQD models, called LR-JNQD, is based on linear regression and determines the model parameter for JNQD based on extracted handcraft features. The other JNQD model is based on a convolution neural network (CNN), called CNN-JNQD. To our best knowledge, our paper is the first approach to automatically adjust JND levels according to quantization step sizes for preprocessing the input to video encoders. In experiments, both the LR-JNQD and CNN-JNQD models were applied to high efficiency video coding (HEVC) and yielded maximum (average) bitrate reductions of 38.51% (10.38%) and 67.88% (24.91%), respectively, with little subjective video quality degradation, compared with the input without preprocessing applied.

摘要

基于传统预测视频编码的方法由于计算复杂度的急剧增加,已经达到了潜在编码效率提高的极限。作为一种替代方法,感知视频编码(PVC)试图通过消除感知冗余,使用仅可察觉失真(JND)指导的 PVC 来实现高效率的编码。之前的 JND 是通过在原始图像中添加白噪声或特定的信号模式来建模的,由于能量减少导致失真,因此不适合寻找 JND 阈值。在本文中,我们提出了一种新的基于离散余弦变换的能量减少 JND 模型,称为 ERJND,它更适合基于 JND 的 PVC 方案。然后,将所提出的 ERJND 模型扩展到两个基于学习的仅可察觉量化失真(JNQD)模型,作为预处理,可以应用于感知视频编码。这两个 JNQD 模型可以根据给定的量化步长自动调整 JND 水平。这两个 JNQD 模型中的一个称为 LR-JNQD,它基于线性回归,并根据提取的手工特征确定 JNQD 的模型参数。另一个 JNQD 模型基于卷积神经网络(CNN),称为 CNN-JNQD。据我们所知,我们的论文是第一个根据量化步长自动调整 JND 水平的方法,用于预处理视频编码器的输入。在实验中,LR-JNQD 和 CNN-JNQD 模型都应用于高效视频编码(HEVC),与未应用预处理的输入相比,分别获得了最大(平均)比特率降低 38.51%(10.38%)和 67.88%(24.91%),而主观视频质量几乎没有下降。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验