Suppr超能文献

具有相对高维数据的推断模型的变量选择:稳健选择的辅助方法——方法异质性和协变量稳定性。

Variable selection for inferential models with relatively high-dimensional data: Between method heterogeneity and covariate stability as adjuncts to robust selection.

机构信息

School of Veterinary Medicine and Science, University of Nottingham, Sutton Bonington Campus, Leicestershire, LE12 5RD, United Kingdom.

OIE, World Organisation for Animal Health 12, rue de Prony, 75017, Paris, France.

出版信息

Sci Rep. 2020 May 14;10(1):8002. doi: 10.1038/s41598-020-64829-0.

Abstract

Variable selection in inferential modelling is problematic when the number of variables is large relative to the number of data points, especially when multicollinearity is present. A variety of techniques have been described to identify 'important' subsets of variables from within a large parameter space but these may produce different results which creates difficulties with inference and reproducibility. Our aim was evaluate the extent to which variable selection would change depending on statistical approach and whether triangulation across methods could enhance data interpretation. A real dataset containing 408 subjects, 337 explanatory variables and a normally distributed outcome was used. We show that with model hyperparameters optimised to minimise cross validation error, ten methods of automated variable selection produced markedly different results; different variables were selected and model sparsity varied greatly. Comparison between multiple methods provided valuable additional insights. Two variables that were consistently selected and stable across all methods accounted for the majority of the explainable variability; these were the most plausible important candidate variables. Further variables of importance were identified from evaluating selection stability across all methods. In conclusion, triangulation of results across methods, including use of covariate stability, can greatly enhance data interpretation and confidence in variable selection.

摘要

当变量的数量相对于数据点的数量较大时,推理建模中的变量选择是有问题的,尤其是存在多重共线性时。已经描述了多种技术来从大参数空间中识别“重要”变量子集,但这些技术可能会产生不同的结果,从而给推理和可重复性带来困难。我们的目的是评估变量选择会在多大程度上因统计方法而异,以及跨方法的三角测量是否可以增强数据解释。使用包含 408 个主题、337 个解释变量和正态分布结果的真实数据集。我们表明,通过优化模型超参数以最小化交叉验证误差,十种自动化变量选择方法产生了明显不同的结果;选择了不同的变量,模型稀疏度差异很大。对多种方法进行比较提供了有价值的附加见解。两个在所有方法中都被一致选择且稳定的变量占可解释变异性的大部分;这些是最合理的重要候选变量。从评估所有方法的选择稳定性中确定了其他重要变量。总之,跨方法结果的三角测量,包括使用协变量稳定性,可以极大地增强数据解释和对变量选择的信心。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c6e8/7224285/28ddd5eb14ca/41598_2020_64829_Fig1_HTML.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验