Tchernev Elko B, Mulvaney Rory G, Phatak Dhananjay S
Computer Science and Electrical Engineering Department, University of Maryland Baltimore County, Baltimore, MD 21250, USA.
Neural Comput. 2005 Jul;17(7):1646-64. doi: 10.1162/0899766053723096.
Particular levels of partial fault tolerance (PFT) in feedforward artificial neural networks of a given size can be obtained by redundancy (replicating a smaller normally trained network), by design (training specifically to increase PFT), and by a combination of the two (replicating a smaller PFT-trained network). This letter investigates the method of achieving the highest PFT per network size (total number of units and connections) for classification problems. It concludes that for non-toy problems, there exists a normally trained network of optimal size that produces the smallest fully fault-tolerant network when replicated. In addition, it shows that for particular network sizes, the best level of PFT is achieved by training a network of that size for fault tolerance. The results and discussion demonstrate how the outcome depends on the levels of saturation of the network nodes when classifying data points. With simple training tasks, where the complexity of the problem and the size of the network are well within the ability of the training method, the hidden-layer nodes operate close to their saturation points, and classification is clean. Under such circumstances, replicating the smallest normally trained correct network yields the highest PFT for any given network size. For hard training tasks (difficult classification problems or network sizes close to the minimum), normal training obtains networks that do not operate close to their saturation points, and outputs are not as close to their targets. In this case, training a larger network for fault tolerance yields better PFT than replicating a smaller, normally trained network. However, since fault-tolerant training on its own produces networks that operate closer to their linear areas than normal training, replicating normally trained networks ultimately leads to better PFT than replicating fault-tolerant networks of the same initial size.
给定规模的前馈人工神经网络中特定级别的部分容错能力(PFT)可以通过冗余(复制一个规模较小的正常训练网络)、通过设计(专门训练以提高PFT)以及通过两者结合(复制一个规模较小的经过PFT训练的网络)来实现。本文研究了针对分类问题在每个网络规模(单元和连接的总数)下实现最高PFT的方法。研究得出结论,对于非简单问题,存在一个最优规模的正常训练网络,当进行复制时会产生最小的完全容错网络。此外,研究表明对于特定的网络规模,通过训练该规模的网络以实现容错能力可达到最佳的PFT水平。结果和讨论展示了在对数据点进行分类时,结果如何取决于网络节点的饱和程度。对于简单的训练任务,问题的复杂度和网络规模完全在训练方法的能力范围内,隐藏层节点接近其饱和点运行,分类清晰。在这种情况下,对于任何给定的网络规模,复制最小的正常训练正确网络可产生最高的PFT。对于困难的训练任务(困难的分类问题或接近最小规模的网络),正常训练得到的网络并非接近其饱和点运行,输出也与目标值不太接近。在这种情况下,训练一个更大的网络以实现容错能力比复制一个较小的、正常训练的网络能产生更好的PFT。然而,由于单独的容错训练所产生的网络比正常训练的网络更接近其线性区域运行,最终复制正常训练的网络比复制相同初始规模的容错网络能带来更好的PFT。