Department of Information Engineering, Electronics and Telecommunications (DIET), "Sapienza" University of Rome, Via Eudossiana 18, 00184 Rome, Italy.
Department of Computer Science and Information Technology, La Trobe University, Melbourne, VIC 3086, Australia.
Neural Netw. 2016 Jun;78:65-74. doi: 10.1016/j.neunet.2015.07.006. Epub 2015 Aug 18.
The current big data deluge requires innovative solutions for performing efficient inference on large, heterogeneous amounts of information. Apart from the known challenges deriving from high volume and velocity, real-world big data applications may impose additional technological constraints, including the need for a fully decentralized training architecture. While several alternatives exist for training feed-forward neural networks in such a distributed setting, less attention has been devoted to the case of decentralized training of recurrent neural networks (RNNs). In this paper, we propose such an algorithm for a class of RNNs known as Echo State Networks. The algorithm is based on the well-known Alternating Direction Method of Multipliers optimization procedure. It is formulated only in terms of local exchanges between neighboring agents, without reliance on a coordinating node. Additionally, it does not require the communication of training patterns, which is a crucial component in realistic big data implementations. Experimental results on large scale artificial datasets show that it compares favorably with a fully centralized implementation, in terms of speed, efficiency and generalization accuracy.
当前的大数据洪流需要创新的解决方案,以便对大量异构信息进行高效推断。除了已知的高容量和高速率带来的挑战外,实际的大数据应用还可能施加额外的技术限制,包括对完全去中心化训练架构的需求。虽然在这种分布式设置中存在几种用于训练前馈神经网络的替代方案,但对于递归神经网络 (RNN) 的去中心化训练关注较少。在本文中,我们为一类称为回声状态网络 (Echo State Network) 的 RNN 提出了这样的算法。该算法基于著名的交替方向乘子法 (ADMM) 优化过程。它仅根据相邻代理之间的局部交换进行构建,不依赖于协调节点。此外,它不需要训练模式的通信,这在现实大数据实现中是一个关键组件。在大规模人工数据集上的实验结果表明,它在速度、效率和泛化准确性方面都优于完全集中式实现。