Saidi Pouria, Dasarathy Gautam, Berisha Visar
School of Electrical, Computer and Energy Engineering, Arizona State University, Tempe, AZ 85281, USA.
College of Health Solutions, Arizona State University, Tempe, AZ 85281, USA.
Patterns (N Y). 2025 Feb 25;6(4):101185. doi: 10.1016/j.patter.2025.101185. eCollection 2025 Apr 11.
Machine learning (ML) is increasingly used across many disciplines with impressive reported results. However, recent studies suggest that the published performances of ML models are often overoptimistic. Validity concerns are underscored by findings of an inverse relationship between sample size and reported accuracy in published ML models, contrasting with the theory of learning curves where accuracy should improve or remain stable with increasing sample size. This paper investigates factors contributing to overoptimism in ML-driven science, focusing on overfitting and publication bias. We introduce a stochastic model for observed accuracy, integrating parametric learning curves and the aforementioned biases. We construct an estimator that corrects for these biases in observed data. Theoretical and empirical results show that our framework can estimate the underlying learning curve, providing realistic performance assessments from published results. By applying the model to meta-analyses of classifications of neurological conditions, we estimate the inherent limits of ML-driven prediction in each domain.
机器学习(ML)在许多学科中越来越多地被使用,报告的结果令人印象深刻。然而,最近的研究表明,ML模型已发表的性能往往过于乐观。样本量与已发表的ML模型中报告的准确率之间存在反比关系,这一发现凸显了有效性问题,这与学习曲线理论形成对比,在学习曲线理论中,准确率应随着样本量的增加而提高或保持稳定。本文研究了导致ML驱动的科学中过度乐观的因素,重点关注过拟合和发表偏倚。我们引入了一个用于观察到的准确率的随机模型,整合了参数学习曲线和上述偏差。我们构建了一个估计器,用于校正观察数据中的这些偏差。理论和实证结果表明,我们的框架可以估计潜在的学习曲线,根据已发表的结果提供现实的性能评估。通过将该模型应用于神经疾病分类的荟萃分析,我们估计了每个领域中ML驱动预测的固有局限性。