Residuals are:
Predictors are not too multicollinear
The relationship between the predictors and outcomes is linear
$$ Residual = Y - \hat{Y} $$
$$\hat{Y} = b_0 + \sum_{p=1}^{P}{B_pX_p}$$
A lot of what we think of as assumptions about the outcome are assumptions about residuals.
We assume residuals follow a multivariate normal distribution. While this is exactly what it sounds like, it also has ramifications that are somewhat wider reaching.
Specifically, if outcome, conditional on the predictors, is not able to be approximated using a normal distribution, our residuals will be non-normal. This means that ordinal or binary outcomes will often violate this assumption.
Specifically, if outcome, conditional on the predictors, is not able to be approximated using a normal distribution, our residuals will be non-normal. This means that ordinal or binary outcomes will often violate this assumption.
This means that errors are equally distributed across levels of the outcome and the predictors.
Suppose the true model would be estimated as:
$$ Y = b_0 + b_1X_1 + b_2X_2 + Error $$
But we omit X_2 from our model, so we estimate:
$$ Y = b_0 + b_1X_1 + Error $$
In the model we actually fit, our error term contains variance from \( X_2 \). If \( X_2 \) is correlated with \( X_1 \), we have violated this assumption.
$$ VIF_p = \frac{1}{1 - R_p^2} $$
where \( R_p^2 \) is the multiple correlation from a model predicting \( X_p \) with every variable in the model except \( X_p \).
$$ Y = b_0 + b_1X_1 + b_2X_1^2 + b_3X_1^3 + Error $$