Kowalczyk Adam
Telstra Research Laboratories, Clayton, Australia
Neural Netw. 1997 Nov;10(8):1417-1433. doi: 10.1016/s0893-6080(97)00009-9.
We estimate the storage capacity of multilayer perceptron with n inputs, h(1) threshold logic units in the first hidden layer and 1 output. We show that if the network can memorize 50% of all dichotomies of a randomly selected N-tuple of points of R(n) with probability 1, then N</=2(nh(1)+1), while at 100% memorization N</=nh(1)+1. Furthermore, if the bounds are reached, then the first hidden layer must be fully connected to the input. It is shown that such a network has memory capacity (in the sense of Cover) between nh(1)+1 and 2(nh(1)+1) input patterns and for the most efficient networks in this class between 1 and 2 input patterns per connection. Comparing these results with the recent estimates of VC-dimension we find that in contrast to a single neuron case, the VC-dimension exceeds the capacity for a sufficiently large n and h(1). The results are based on the derivation of an explicit expression for the number of dichotomies which can be implemented by such a network for a special class of N-tuples of input patterns which has a positive probability of being randomly chosen.
我们估计具有(n)个输入、第一隐藏层中有(h(1))个阈值逻辑单元和(1)个输出的多层感知器的存储容量。我们表明,如果网络能够以概率(1)记住(\mathbb{R}(n))中随机选择的(N)元组点的所有二分法的(50%),那么(N\leq2(nh(1)+1)),而在(100%)记忆时(N\leq nh(1)+1)。此外,如果达到这些界限,那么第一隐藏层必须与输入完全连接。结果表明,这样的网络具有介于(nh(1)+1)和(2(nh(1)+1))个输入模式之间的记忆容量(在Cover意义下),并且对于该类中最有效的网络,每个连接有(1)到(2)个输入模式。将这些结果与最近对VC维的估计进行比较,我们发现与单个神经元情况不同,对于足够大的(n)和(h(1)),VC维超过了容量。这些结果基于对一类特殊的输入模式(N)元组的二分法数量的显式表达式的推导,该类(N)元组被随机选择的概率为正。