Gori Marco, Sperduti Alessandro
Dipartimento di Ingegneria dell'Informazione, Università di Siena, Via Roma 56, Siena 53100, Italy.
Neural Netw. 2005 Oct;18(8):1064-79. doi: 10.1016/j.neunet.2005.07.006. Epub 2005 Sep 29.
The present work deals with one of the major and not yet completely understood topics of supervised connectionist models. Namely, it investigates the relationships between the difficulty of a given learning task and the chosen neural network architecture. These relationships have been investigated and nicely established for some interesting problems in the case of neural networks used for processing vectors and sequences, but only a few studies have dealt with loading problems involving graphical inputs. In this paper, we present sufficient conditions which guarantee the absence of local minima of the error function in the case of learning directed acyclic graphs with recursive neural networks. We introduce topological indices which can be directly calculated from the given training set and that allows us to design the neural architecture with local minima free error function. In particular, we conceive a reduction algorithm that involves both the information attached to the nodes and the topology, which enlarges significantly the class of the problems with unimodal error function previously proposed in the literature.
目前的工作涉及监督式连接主义模型中一个主要但尚未完全理解的主题。具体而言,它研究了给定学习任务的难度与所选神经网络架构之间的关系。对于用于处理向量和序列的神经网络,在一些有趣的问题中已经对这些关系进行了研究并很好地确立了,但只有少数研究涉及涉及图形输入的加载问题。在本文中,我们给出了充分条件,这些条件保证在使用递归神经网络学习有向无环图的情况下误差函数不存在局部最小值。我们引入了可以直接从给定训练集计算得出的拓扑指标,这使我们能够设计出具有无局部最小值误差函数的神经架构。特别是,我们构想了一种既涉及节点附加信息又涉及拓扑的约简算法,该算法显著扩大了文献中先前提出的具有单峰误差函数的问题类别。