Akaike’s Information Criterion

Information Criterions

We define two types of information criterion: the Akaike Information Criterion (AIC) and the Schwarz’s Bayesian Information Criterion (BIC).

The Akaike information criterion is a measure of the relative goodness of fit of a statistical model.

Akaike Information Criterion

  • Akaike’s information criterionis a measure of the goodness of fit of an estimated statistical model. The AIC was developed by Hirotsugu Akaike under the name of an information criterion in 1971.

  • The AIC is a model selection tool i.e. a method of comparing two or more candidate regression models. The AIC methodology attempts to find the model that best explains the data with a minimum of parameters. (i.e. in keeping with the law of parsimony)

  • The AIC is calculated using the likelihood function and the number of parameters . The likelihood value is generally given in code output, as a complement to the AIC. (Likelihood function is not on our course)

  • Given a data set, several competing models may be ranked according to their AIC, with the one having the lowest AIC being the best. (Although, a difference in AIC values of less than two is considered negligible).

\[ AIC = 2p - 2 \ln(L)\]


  • \(p\) is the number of predictor variables in the model.
  • \(L\) is the value of the Likelihood function for the model in question.
  • For AIC to be optimal, \(n\) must be large compared to \(p\).

An alternative to the AIC is the Schwarz BIC, which additionally takes into account the sample size n.

\[BIC = p \ln(n) - 2 \ln(L)\]


  • When using the AIC (or BIC) for selecting the optimal model, we choose the model for which the AIC (or BIC) value is lowest.

  • In a multiple linear regression model, adjusted R square measures the proportion of the variation in the dependent variable accounted for by the independent variables.

  • Adjusted R square is generally considered to be a more accurate goodness-of-fit measure than R square.