Cancelliere Rossella, Gai Mario, Gallinari Patrick, Rubini Luca
University of Turin, Department of Computer Sciences, C.so Svizzera 185, 10149 Torino, Italy.
National Institute of Astrophysics, Astrophys. Observ. of Torino, Pino T.se (TO), Italy.
Neural Netw. 2015 Nov;71:76-87. doi: 10.1016/j.neunet.2015.07.015. Epub 2015 Aug 12.
In this paper we consider the training of single hidden layer neural networks by pseudoinversion, which, in spite of its popularity, is sometimes affected by numerical instability issues. Regularization is known to be effective in such cases, so that we introduce, in the framework of Tikhonov regularization, a matricial reformulation of the problem which allows us to use the condition number as a diagnostic tool for identification of instability. By imposing well-conditioning requirements on the relevant matrices, our theoretical analysis allows the identification of an optimal value for the regularization parameter from the standpoint of stability. We compare with the value derived by cross-validation for overfitting control and optimization of the generalization performance. We test our method for both regression and classification tasks. The proposed method is quite effective in terms of predictivity, often with some improvement on performance with respect to the reference cases considered. This approach, due to analytical determination of the regularization parameter, dramatically reduces the computational load required by many other techniques.
在本文中,我们考虑通过伪逆来训练单隐藏层神经网络,尽管它很受欢迎,但有时会受到数值不稳定问题的影响。已知正则化在这种情况下是有效的,因此我们在蒂霍诺夫正则化框架内引入了问题的矩阵重新表述,这使我们能够将条件数用作识别不稳定性的诊断工具。通过对相关矩阵施加良态条件要求,我们的理论分析能够从稳定性的角度确定正则化参数的最优值。我们将其与通过交叉验证得出的值进行比较,以进行过拟合控制和泛化性能优化。我们针对回归和分类任务测试了我们的方法。所提出的方法在预测性方面相当有效,相对于所考虑的参考案例,性能通常会有所提高。由于正则化参数的解析确定,这种方法极大地减少了许多其他技术所需的计算量。