Xiao Wei, Wu Yichao, Zhou Hua
Department of Statistics, North Carolina State University, Raleigh, NC 27695.
J Comput Graph Stat. 2015 Jul 1;24(3):603-626. doi: 10.1080/10618600.2014.962700. Epub 2015 Sep 16.
The least angle regression (LAR) was proposed by Efron, Hastie, Johnstone and Tibshirani (2004) for continuous model selection in linear regression. It is motivated by a geometric argument and tracks a path along which the predictors enter successively and the active predictors always maintain the same absolute correlation (angle) with the residual vector. Although it gains popularity quickly, its extensions seem rare compared to the penalty methods. In this expository article, we show that the powerful geometric idea of LAR can be generalized in a fruitful way. We propose a ConvexLAR algorithm that works for any convex loss function and naturally extends to group selection and data adaptive variable selection. After simple modification it also yields new exact path algorithms for certain penalty methods such as a convex loss function with lasso or group lasso penalty. Variable selection in recurrent event and panel count data analysis, Ada-Boost, and Gaussian graphical model is reconsidered from the ConvexLAR angle.
最小角回归(LAR)由埃弗龙(Efron)、哈斯蒂(Hastie)、约翰斯通(Johnstone)和蒂布希拉尼(Tibshirani)于2004年提出,用于线性回归中的连续模型选择。它由一个几何论证推动,并追踪一条路径,沿着该路径预测变量依次进入,且活跃预测变量与残差向量始终保持相同的绝对相关性(角度)。尽管它迅速受到欢迎,但与惩罚方法相比,其扩展似乎很少。在这篇说明性文章中,我们表明LAR强大的几何思想可以以富有成效的方式进行推广。我们提出了一种适用于任何凸损失函数的凸LAR算法,并且自然地扩展到组选择和数据自适应变量选择。经过简单修改后,它还为某些惩罚方法(如带有套索或组套索惩罚的凸损失函数)产生了新的精确路径算法。从凸LAR的角度重新考虑了复发事件和面板计数数据分析、Ada - Boost以及高斯图形模型中的变量选择。