The AIC criterion and symmetrizing the Kullback-Leibler divergence
The Akaike information criterion (AIC) is a widely used tool for model selection. AIC is derived as an asymptotically unbiased estimator of a function used for ranking candidate models which is a variant of the Kullback-Leibler divergence between the true model and the approximating candidate model. Despite the Kullback-Leibler's computational and theoretical advantages, what can become inconvenient in model selection applications is their lack of symmetry. Simple examples can show that...[Show more]
|Collections||ANU Research Publications|
|Source:||IEEE Transactions on Neural Networks|
|01_Seghouane_The_AIC_criterion_and_2007.pdf||643.52 kB||Adobe PDF||Request a copy|
Items in Open Research are protected by copyright, with all rights reserved, unless otherwise indicated.