Seghouane Abd-Krim, Amari Shun-Ichi
Systems Engineering and Complex Systems Research Program, National ICT Australia, Canberra Research Laboratory, Canberra, ACT 2601, Australia.
IEEE Trans Neural Netw. 2007 Jan;18(1):97-106. doi: 10.1109/TNN.2006.882813.
The Akaike information criterion (AIC) is a widely used tool for model selection. AIC is derived as an asymptotically unbiased estimator of a function used for ranking candidate models which is a variant of the Kullback-Leibler divergence between the true model and the approximating candidate model. Despite the Kullback-Leibler's computational and theoretical advantages, what can become inconvenient in model selection applications is their lack of symmetry. Simple examples can show that reversing the role of the arguments in the Kullback-Leibler divergence can yield substantially different results. In this paper, three new functions for ranking candidate models are proposed. These functions are constructed by symmetrizing the Kullback-Leibler divergence between the true model and the approximating candidate model. The operations used for symmetrizing are the average, geometric, and harmonic means. It is found that the original AIC criterion is an asymptotically unbiased estimator of these three different functions. Using one of these proposed ranking functions, an example of new bias correction to AIC is derived for univariate linear regression models. A simulation study based on polynomial regression is provided to compare the different proposed ranking functions with AIC and the new derived correction with AICc.
赤池信息准则(AIC)是一种广泛用于模型选择的工具。AIC是作为用于对候选模型进行排序的函数的渐近无偏估计量推导出来的,该函数是真实模型与近似候选模型之间的库尔贝克-莱布勒散度的一种变体。尽管库尔贝克-莱布勒散度在计算和理论上具有优势,但在模型选择应用中可能不方便的是它们缺乏对称性。简单的例子可以表明,在库尔贝克-莱布勒散度中颠倒参数的作用会产生截然不同的结果。本文提出了三种用于对候选模型进行排序的新函数。这些函数是通过对真实模型与近似候选模型之间的库尔贝克-莱布勒散度进行对称化构建的。用于对称化的运算为算术平均、几何平均和调和平均。研究发现,原始的AIC准则是这三种不同函数的渐近无偏估计量。使用这些提出的排序函数之一,为单变量线性回归模型推导了对AIC进行新的偏差校正的一个例子。提供了一项基于多项式回归的模拟研究,以比较所提出的不同排序函数与AIC以及新推导的校正与AICc。