Suppr超能文献

在解释临床数据的神经网络中过度训练。

Overtraining in neural networks that interpret clinical data.

作者信息

Astion M L, Wener M H, Thomas R G, Hunder G G, Bloch D A

机构信息

Department of Laboratory Medicine, University of Washington, Seattle 98195.

出版信息

Clin Chem. 1993 Sep;39(9):1998-2004.

PMID:8375090
Abstract

Backpropagation neural networks are a computer-based pattern-recognition method that has been applied to the interpretation of clinical data. Unlike rule-based pattern recognition, backpropagation networks learn by being repetitively trained with examples of the patterns to be differentiated. We describe and analyze the phenomenon of overtraining in backpropagation networks. Overtraining refers to the reduction in generalization ability that can occur as networks are trained. The clinical application we used was the differentiation of giant cell arteritis (GCA) from other forms of vasculitis (OTH) based on results for 807 patients (593 OTH, 214 GCA) and eight clinical predictor variables. The 807 cases were randomly assigned to either a training set with 404 cases or to a cross-validation set with the remaining 403 cases. The cross-validation set was used to monitor generalization during training. Results were obtained for eight networks, each derived from a different random assignment of the 807 cases. Training error monotonically decreased during training. In contrast, the cross-validation error usually reached a minimum early in training while the training error was still decreasing. Training beyond the minimum cross-validation error was associated with an increased cross-validation error. The shape of the cross-validation error curve and the point during training corresponding to the minimum cross-validation error varied with the composition of the data sets and the training conditions. The study indicates that training error is not a reliable indicator of a network's ability to generalize. To find the point during training when a network generalizes best, one must monitor cross-validation error separately.

摘要

反向传播神经网络是一种基于计算机的模式识别方法,已应用于临床数据的解读。与基于规则的模式识别不同,反向传播网络通过对要区分的模式示例进行反复训练来学习。我们描述并分析了反向传播网络中的过训练现象。过训练是指随着网络训练可能出现的泛化能力下降。我们使用的临床应用是根据807例患者(593例其他血管炎,214例巨细胞动脉炎)的结果和八个临床预测变量,将巨细胞动脉炎(GCA)与其他形式的血管炎(OTH)区分开来。807例病例被随机分配到一个包含404例病例的训练集或一个包含其余403例病例的交叉验证集。交叉验证集用于在训练期间监测泛化情况。对八个网络进行了结果分析,每个网络都来自807例病例的不同随机分配。训练误差在训练期间单调下降。相比之下,交叉验证误差通常在训练早期达到最小值,而此时训练误差仍在下降。在最小交叉验证误差之后继续训练会导致交叉验证误差增加。交叉验证误差曲线的形状以及训练期间对应于最小交叉验证误差的点会因数据集的组成和训练条件而有所不同。该研究表明,训练误差不是网络泛化能力的可靠指标。要找到网络泛化最佳的训练点,必须单独监测交叉验证误差。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验