Model fit statistics give us an indication of how well our statistical model fits our data. There are lots of different fit statistics depending on the type of model, but here are a few:
Regression
represents the proportion of the variation in the dependent variable that is explained by the model. Since it’s a proportion, it ranges from 0 to 1, with higher values indicating better fit. The formula for is:
which is 1 - (residual sum of squares) / (total sum of squares)
RMSE
The root mean squared error (RMSE) represents the average amount of error per observation. Since it represents the amount of error, lower values are better. The formula is:
AIC and BIC
Akaike Information Criteroin (AIC) and Bayesian Information Criterion (BIC) are fit statistics that take model complexity into account. Essentially, they add in a penalty for more complex models — similar to how regularized regression penalizes model coefficients.
It’s helpful to penalize complexity if we’re going to be assessing our models without using some sort of train/test split or cross-validation. This is because adding complexity will always improve model fit/accuracy/whatever on the training data, although this added complexity does not necessarily mean the model will perform better on unseen data.
AIC
The AIC is defined for models fit by maximum likelihood and is given by:
where is the residual sum of squares, is the number of predictors in the model, and is the error variance.
BIC
BIC is given by:
The main difference here is that instead of using like AIC does in its penalty, BIC uses . In practice, this will more-heavily-penalize models with more predictors. In other words, it places a higher priority on parsimony.
Adjusted
Adjusted adjusts the statistic — which will always increase when more variables are added — to penalize model complexity. The formula is:
Where is the number of observations and is the number of predictors in the model.
Accuracy
TODO
Structural Equation Modeling
TODO