Hannan–Quinn information criterion

In statistics, the Hannan–Quinn information criterion (HQC) is a criterion for model selection. It is an alternative to Akaike information criterion (AIC) and Bayesian information criterion (BIC). It is given as

 \mathrm{HQC} = n \log \left( {\mathrm{RSS} \over n } \right) + 2 k \log \log n, \

where k is the number of parameters, n is the number of observations, and RSS is the residual sum of squares that results from linear regression or other statistical model.

Burnham & Anderson (2002, p. 287) say that HQC, "while often cited, seems to have seen little use in practice". They also note that HQC, like BIC, but unlike AIC, is not an estimator of Kullback–Leibler divergence. Claeskens & Hjort (2008, ch. 4) note that HQC, like BIC, but unlike AIC, is not asymptotically efficient, and further point out that whatever method is being used for fine-tuning the criterion will be more important in practice than the term log log n, since this latter number is small even for very large n.

See also

References