Dai Xiaowu, Lyu Xiang, Li Lexin
University of California, Berkeley.
J Am Stat Assoc. 2023;118(543):2158-2170. doi: 10.1080/01621459.2022.2039671. Epub 2022 Mar 14.
Thanks to its fine balance between model flexibility and interpretability, the nonparametric additive model has been widely used, and variable selection for this type of model has been frequently studied. However, none of the existing solutions can control the false discovery rate (FDR) unless the sample size tends to infinity. The knockoff framework is a recent proposal that can address this issue, but few knockoff solutions are directly applicable to nonparametric models. In this article, we propose a novel kernel knockoffs selection procedure for the nonparametric additive model. We integrate three key components: the knockoffs, the subsampling for stability, and the random feature mapping for nonparametric function approximation. We show that the proposed method is guaranteed to control the FDR for any sample size, and achieves a power that approaches one as the sample size tends to infinity. We demonstrate the efficacy of our method through intensive simulations and comparisons with the alternative solutions. our proposal thus makes useful contributions to the methodology of nonparametric variable selection, FDR-based inference, as well as knockoffs.
由于非参数加法模型在模型灵活性和可解释性之间取得了良好平衡,它已被广泛使用,并且针对此类模型的变量选择也得到了频繁研究。然而,现有的解决方案都无法控制错误发现率(FDR),除非样本量趋于无穷大。仿冒框架是最近提出的一种可以解决此问题的方法,但很少有仿冒解决方案能直接应用于非参数模型。在本文中,我们为非参数加法模型提出了一种新颖的核仿冒选择程序。我们整合了三个关键组件:仿冒、用于稳定性的子采样以及用于非参数函数逼近的随机特征映射。我们表明,所提出的方法能够保证在任何样本量下都控制FDR,并且当样本量趋于无穷大时,其检验功效趋近于1。我们通过大量模拟以及与替代解决方案的比较来证明我们方法的有效性。因此,我们的提议为非参数变量选择、基于FDR的推断以及仿冒方法做出了有益贡献。