Vergouwe Yvonne, Nieboer Daan, Oostenbrink Rianne, Debray Thomas P A, Murray Gordon D, Kattan Michael W, Koffijberg Hendrik, Moons Karel G M, Steyerberg Ewout W
Center for Medical Decision Sciences, Department of Public Health, Erasmus MC, Rotterdam, the Netherlands.
Department of General Paediatrics, Erasmus MC-Sophia Children's Hospital, Rotterdam, the Netherlands.
Stat Med. 2017 Dec 10;36(28):4529-4539. doi: 10.1002/sim.7179. Epub 2016 Nov 28.
Prediction models fitted with logistic regression often show poor performance when applied in populations other than the development population. Model updating may improve predictions. Previously suggested methods vary in their extensiveness of updating the model. We aim to define a strategy in selecting an appropriate update method that considers the balance between the amount of evidence for updating in the new patient sample and the danger of overfitting. We consider recalibration in the large (re-estimation of model intercept); recalibration (re-estimation of intercept and slope) and model revision (re-estimation of all coefficients) as update methods. We propose a closed testing procedure that allows the extensiveness of the updating to increase progressively from a minimum (the original model) to a maximum (a completely revised model). The procedure involves multiple testing with maintaining approximately the chosen type I error rate. We illustrate this approach with three clinical examples: patients with prostate cancer, traumatic brain injury and children presenting with fever. The need for updating the prostate cancer model was completely driven by a different model intercept in the update sample (adjustment: 2.58). Separate testing of model revision against the original model showed statistically significant results, but led to overfitting (calibration slope at internal validation = 0.86). The closed testing procedure selected recalibration in the large as update method, without overfitting. The advantage of the closed testing procedure was confirmed by the other two examples. We conclude that the proposed closed testing procedure may be useful in selecting appropriate update methods for previously developed prediction models. Copyright © 2016 John Wiley & Sons, Ltd.
当应用于开发人群以外的其他人群时,采用逻辑回归拟合的预测模型往往表现不佳。模型更新可能会改善预测效果。先前提出的方法在模型更新的广度方面各不相同。我们旨在定义一种策略,以选择一种合适的更新方法,该方法要考虑新患者样本中用于更新的证据量与过度拟合风险之间的平衡。我们将大规模重新校准(重新估计模型截距)、重新校准(重新估计截距和斜率)以及模型修订(重新估计所有系数)视为更新方法。我们提出了一种封闭检验程序,该程序允许更新的广度从最小值(原始模型)逐步增加到最大值(完全修订的模型)。该程序涉及多次检验,同时保持近似选定的I型错误率。我们用三个临床实例来说明这种方法:前列腺癌患者、创伤性脑损伤患者以及发热儿童。前列腺癌模型的更新需求完全由更新样本中不同的模型截距驱动(调整值:2.58)。针对原始模型单独检验模型修订显示出具有统计学意义的结果,但导致了过度拟合(内部验证时的校准斜率 = 0.86)。封闭检验程序选择大规模重新校准作为更新方法,未出现过度拟合。另外两个实例证实了封闭检验程序的优势。我们得出结论,所提出的封闭检验程序可能有助于为先前开发的预测模型选择合适的更新方法。版权所有© 2016约翰威立父子有限公司。