Nakahara Yuta, Matsushima Toshiyasu
Center for Data Science, Waseda University, 1-6-1 Nisniwaseda, Shinjuku-ku, Tokyo 169-8050, Japan.
Department of Pure and Applied Mathematics, Waseda University, 3-4-1 Okubo, Shinjuku-ku, Tokyo 169-8555, Japan.
Entropy (Basel). 2021 Jul 30;23(8):991. doi: 10.3390/e23080991.
In information theory, lossless compression of general data is based on an explicit assumption of a stochastic generative model on target data. However, in lossless image compression, researchers have mainly focused on the coding procedure that outputs the coded sequence from the input image, and the assumption of the stochastic generative model is implicit. In these studies, there is a difficulty in discussing the difference between the expected code length and the entropy of the stochastic generative model. We solve this difficulty for a class of images, in which they have non-stationarity among segments. In this paper, we propose a novel stochastic generative model of images by redefining the implicit stochastic generative model in a previous coding procedure. Our model is based on the quadtree so that it effectively represents the variable block size segmentation of images. Then, we construct the Bayes code optimal for the proposed stochastic generative model. It requires the summation of all possible quadtrees weighted by their posterior. In general, its computational cost increases exponentially for the image size. However, we introduce an efficient algorithm to calculate it in the polynomial order of the image size without loss of optimality. As a result, the derived algorithm has a better average coding rate than that of JBIG.
在信息论中,通用数据的无损压缩基于对目标数据的随机生成模型的明确假设。然而,在无损图像压缩中,研究人员主要关注从输入图像输出编码序列的编码过程,随机生成模型的假设是隐含的。在这些研究中,讨论随机生成模型的期望码长与熵之间的差异存在困难。我们针对一类在各段之间具有非平稳性的图像解决了这一困难。在本文中,我们通过重新定义先前编码过程中的隐含随机生成模型,提出了一种新颖的图像随机生成模型。我们的模型基于四叉树,以便有效地表示图像的可变块大小分割。然后,我们为所提出的随机生成模型构建最优的贝叶斯码。这需要对所有可能的四叉树按其后验概率加权求和。一般来说,其计算成本随图像大小呈指数增长。然而,我们引入了一种高效算法,以图像大小的多项式阶数来计算它,而不会损失最优性。结果,所推导的算法具有比JBIG更好的平均编码率。