Zhao Wanqing, Beach Thomas H, Rezgui Yacine
Cardiff School of Engineering, Cardiff University , Cardiff CF24 3AA, UK.
Proc Math Phys Eng Sci. 2017 Feb;473(2198):20160775. doi: 10.1098/rspa.2016.0775.
Least angle regression, as a promising model selection method, differentiates itself from conventional stepwise and stagewise methods, in that it is neither too greedy nor too slow. It is closely related to norm optimization, which has the advantage of low prediction variance through sacrificing part of model bias property in order to enhance model generalization capability. In this paper, we propose an efficient least angle regression algorithm for model selection for a large class of linear-in-the-parameters models with the purpose of accelerating the model selection process. The entire algorithm works completely in a recursive manner, where the correlations between model terms and residuals, the evolving directions and other pertinent variables are derived explicitly and updated successively at every subset selection step. The model coefficients are only computed when the algorithm finishes. The direct involvement of matrix inversions is thereby relieved. A detailed computational complexity analysis indicates that the proposed algorithm possesses significant computational efficiency, compared with the original approach where the well-known efficient Cholesky decomposition is involved in solving least angle regression. Three artificial and real-world examples are employed to demonstrate the effectiveness, efficiency and numerical stability of the proposed algorithm.
最小角回归作为一种很有前景的模型选择方法,有别于传统的逐步回归和阶段wise回归方法,因为它既不过于贪婪也不过于缓慢。它与范数优化密切相关,范数优化的优点是通过牺牲部分模型偏差特性来降低预测方差,从而提高模型的泛化能力。在本文中,我们提出了一种高效的最小角回归算法,用于一大类参数线性模型的模型选择,目的是加速模型选择过程。整个算法完全以递归方式运行,在每个子集选择步骤中,显式推导并依次更新模型项与残差之间的相关性、演化方向和其他相关变量。仅在算法结束时才计算模型系数。从而避免了直接进行矩阵求逆运算。详细的计算复杂度分析表明,与原始方法相比,所提出的算法具有显著的计算效率,原始方法在求解最小角回归时涉及著名的高效Cholesky分解。使用三个人工和实际例子来证明所提出算法的有效性、效率和数值稳定性。