Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, 518055, China.
School of Mathematics, Sun Yat-sen University, Guangzhou, 510275, China; Guangdong Province Key Laboratory of Computational Science, Guangzhou, 510275, China.
Neural Netw. 2022 Sep;153:49-63. doi: 10.1016/j.neunet.2022.05.020. Epub 2022 Jun 2.
Despite the successful use of Gaussian-binary restricted Boltzmann machines (GB-RBMs) and Gaussian-binary deep belief networks (GB-DBNs), little is known about their theoretical approximation capabilities to represent distributions of continuous random variables. In this paper, we address the expressive properties of GB-RBMs and GB-DBNs, contributing theoretical insights to the optimal number of hidden variables. We first treat the GB-RBM's unnormalized log-likelihood as a sum of a special two-layer feedforward neural network and a negative quadratic term. Then, a series of simulation results are established, which can be used to relate GB-RBMs to general two-layer feedforward neural networks whose expressive properties are much better understood. On this basis, we show that a two-layer ReLU network with all weights in the second layer being 1, along with a negative quadratic term, can approximate all continuous functions. In addition, we provide qualified lower bounds for the number of hidden variables of GB-RBMs required to approximate distributions whose log-likelihood are given by some classes of smooth functions. Moreover, we further study the universal approximation of GB-DBNs with two hidden layers by providing a sufficient number of hidden variables O(ɛ) that are guaranteed to approximate any given strictly positive continuous distribution within a given error ɛ. Finally, numerical experiments are carried out to verify some of the proposed theoretical results.
尽管高斯二值受限玻尔兹曼机(GB-RBM)和高斯二值深度置信网络(GB-DBN)的应用取得了成功,但对于它们在表示连续随机变量分布方面的理论近似能力,人们知之甚少。在本文中,我们研究了 GB-RBM 和 GB-DBN 的表达能力,为最优隐藏变量数量提供了理论见解。我们首先将 GB-RBM 的非归一化对数似然看作是一个特殊的两层前馈神经网络和一个负二次项的和。然后,建立了一系列模拟结果,可以将 GB-RBM 与一般的两层前馈神经网络联系起来,而后者的表达能力更容易理解。在此基础上,我们证明了具有全一权重的两层 ReLU 网络加上负二次项可以逼近所有连续函数。此外,我们为需要逼近某些光滑函数类对数似然分布的 GB-RBM 所需的隐藏变量数量提供了有条件的下界。此外,我们通过提供足够数量的隐藏变量 O(ɛ)来进一步研究具有两个隐藏层的 GB-DBN 的通用逼近,这些隐藏变量保证可以在给定的误差 ɛ内逼近任何给定的严格正连续分布。最后,进行了数值实验来验证一些提出的理论结果。