Shoemaker P A
US Naval Ocean Syst. Center, San Diego, CA.
IEEE Trans Neural Netw. 1991;2(1):158-60. doi: 10.1109/72.80304.
Neural network models are considered as mathematical classifiers whose inputs comprise random variables generated according to arbitrary stationary class distributions, and the implication of learning based on minimization of sum-square classification error over a training set of these observations for which class assignments are absolutely determined is addressed. Expectations for network outputs in such cases are weighted least-squares approximations to a posteriori probabilities for the classes, which justifies interpretation of network outputs as indicating degree of confidence in class membership. The author demonstrates this with a straightforward proof in which class probability densities are regarded as primitives and which for simplicity does not rely on probability theory or statistics. The author cites more detailed results giving conditions for consistency of the estimators and discusses some issues relating to the suitability of neural network models and back-propagation training for approximation of conditional probabilities in classification tasks.
神经网络模型被视为数学分类器,其输入由根据任意平稳类别分布生成的随机变量组成,并且探讨了基于这些观测值的训练集上的平方和分类误差最小化进行学习的意义,其中类别分配是绝对确定的。在这种情况下,网络输出的期望是对类别的后验概率的加权最小二乘逼近,这证明了将网络输出解释为指示类别成员资格的置信度是合理的。作者通过一个直接的证明展示了这一点,其中类别概率密度被视为原语,并且为了简单起见,不依赖于概率论或统计学。作者引用了给出估计量一致性条件的更详细结果,并讨论了一些与神经网络模型的适用性以及反向传播训练在分类任务中逼近条件概率的适用性相关的问题。