IEEE Trans Neural Netw Learn Syst. 2022 Nov;33(11):6373-6387. doi: 10.1109/TNNLS.2021.3079113. Epub 2022 Oct 27.
The adaptive hinging hyperplane (AHH) model is a popular piecewise linear representation with a generalized tree structure and has been successfully applied in dynamic system identification. In this article, we aim to construct the deep AHH (DAHH) model to extend and generalize the networking of AHH model for high-dimensional problems. The network structure of DAHH is determined through a forward growth, in which the activity ratio is introduced to select effective neurons and no connecting weights are involved between the layers. Then, all neurons in the DAHH network can be flexibly connected to the output in a skip-layer format, and only the corresponding weights are the parameters to optimize. With such a network framework, the backpropagation algorithm can be implemented in DAHH to efficiently tackle large-scale problems and the gradient vanishing problem is not encountered in the training of DAHH. In fact, the optimization problem of DAHH can maintain convexity with convex loss in the output layer, which brings natural advantages in optimization. Different from the existing neural networks, DAHH is easier to interpret, where neurons are connected sparsely and analysis of variance (ANOVA) decomposition can be applied, facilitating to revealing the interactions between variables. A theoretical analysis toward universal approximation ability and explicit domain partitions are also derived. Numerical experiments verify the effectiveness of the proposed DAHH.
自适应铰链超平面 (AHH) 模型是一种流行的分段线性表示方法,具有广义树状结构,已成功应用于动态系统辨识。本文旨在构建深度 AHH (DAHH) 模型,以扩展和推广 AHH 模型在高维问题中的网络结构。DAHH 的网络结构通过前向生长确定,其中引入活动比来选择有效神经元,并且各层之间没有连接权重。然后,DAHH 网络中的所有神经元都可以以跳过层的格式灵活地连接到输出,并且只有相应的权重是需要优化的参数。通过这样的网络框架,可以在 DAHH 中实现反向传播算法,以有效地处理大规模问题,并且在 DAHH 的训练中不会遇到梯度消失问题。实际上,DAHH 的优化问题可以在输出层保持凸性和凸损失,这在优化中带来了自然的优势。与现有的神经网络不同,DAHH 更容易解释,其中神经元稀疏连接,可以应用方差分析 (ANOVA) 分解,有助于揭示变量之间的相互作用。还推导出了对通用逼近能力和显式域分区的理论分析。数值实验验证了所提出的 DAHH 的有效性。