IEEE Trans Neural Netw Learn Syst. 2012 Oct;23(10):1649-58. doi: 10.1109/TNNLS.2012.2210242.
Simple recurrent error backpropagation networks have been widely used to learn temporal sequence data, including regular and context-free languages. However, the production of relatively large and opaque weight matrices during learning has inspired substantial research on how to extract symbolic human-readable interpretations from trained networks. Unlike feedforward networks, where research has focused mainly on rule extraction, most past work with recurrent networks has viewed them as dynamical systems that can be approximated symbolically by finite-state machine (FSMs). With this approach, the network's hidden layer activation space is typically divided into a finite number of regions. Past research has mainly focused on better techniques for dividing up this activation space. In contrast, very little work has tried to influence the network training process to produce a better representation in hidden layer activation space, and that which has been done has had only limited success. Here we propose a powerful general technique to bias the error backpropagation training process so that it learns an activation space representation from which it is easier to extract FSMs. Using four publicly available data sets that are based on regular and context-free languages, we show via computational experiments that the modified learning method helps to extract FSMs with substantially fewer states and less variance than unmodified backpropagation learning, without decreasing the neural networks' accuracy. We conclude that modifying error backpropagation so that it more effectively separates learned pattern encodings in the hidden layer is an effective way to improve contemporary FSM extraction methods.
简单的递归错误反向传播网络已被广泛用于学习时间序列数据,包括正则和无上下文语言。然而,在学习过程中产生的相对较大且不透明的权重矩阵激发了大量关于如何从训练的网络中提取符号可读解释的研究。与主要关注规则提取的前馈网络不同,过去关于递归网络的大多数工作都将它们视为可以通过有限状态机 (FSM) 进行符号近似的动力系统。使用这种方法,网络的隐藏层激活空间通常被划分为有限数量的区域。过去的研究主要集中在更好的划分这个激活空间的技术上。相比之下,很少有工作试图影响网络训练过程,以在隐藏层激活空间中产生更好的表示,而且已经完成的工作只取得了有限的成功。在这里,我们提出了一种强大的通用技术,使误差反向传播训练过程产生偏差,以便从更容易提取 FSM 的激活空间中学习。使用基于正则和无上下文语言的四个公开可用的数据集,我们通过计算实验表明,与未经修改的反向传播学习相比,修改后的学习方法有助于提取具有更少状态和更少方差的 FSM,而不会降低神经网络的准确性。我们的结论是,修改误差反向传播以便更有效地在隐藏层中分离学习的模式编码是改进当代 FSM 提取方法的有效途径。