Erdal Mehmet, Schwenker Friedhelm
Institute of Neural Information Processing, Ulm University, 89081 Ulm, Germany.
Entropy (Basel). 2022 Aug 13;24(8):1117. doi: 10.3390/e24081117.
In this paper, we study the learnability of the Boolean inner product by a systematic simulation study. The family of the Boolean inner product function is known to be representable by neural networks of threshold neurons of depth 3 with only 2n+1 units ( the input dimension)-whereas an exact representation by a depth 2 network cannot possibly be of polynomial size. This result can be seen as a strong argument for deep neural network architectures. In our study, we found that this depth 3 architecture of the Boolean inner product is difficult to train, much harder than the depth 2 network, at least for the small input size scenarios n≤16. Nonetheless, the accuracy of the deep architecture increased with the dimension of the input space to 94% on average, which means that multiple restarts are needed to find the compact depth 3 architecture. Replacing the fully connected first layer by a partially connected layer (a kind of convolutional layer sparsely connected with weight sharing) can significantly improve the learning performance up to 99% accuracy in simulations. Another way to improve the learnability of the compact depth 3 representation of the inner product could be achieved by adding just a few additional units into the first hidden layer.
在本文中,我们通过系统的模拟研究来探讨布尔内积的可学习性。已知布尔内积函数族可以由深度为3且仅具有2n + 1个单元(输入维度)的阈值神经元神经网络表示,而深度为2的网络的精确表示不可能具有多项式规模。这一结果可被视为支持深度神经网络架构的有力论据。在我们的研究中,我们发现布尔内积的这种深度为3的架构很难训练,比深度为2的网络困难得多,至少在小输入规模场景n≤16时是这样。尽管如此,深度架构的准确率会随着输入空间维度的增加而提高,平均达到94%,这意味着需要多次重新启动才能找到紧凑的深度为3的架构。在模拟中,用部分连接层(一种具有权重共享的稀疏连接的卷积层)替换全连接的第一层可以显著提高学习性能,准确率高达99%。提高内积紧凑深度为3表示的可学习性的另一种方法可以是在第一个隐藏层中仅添加几个额外的单元来实现。