In multiple linear regression, hypothesis tests answer questions like:
βj equal to zero (or some other value)?We assume the usual regression model:
Here, k is the number of predictors, so there are p = k + 1 parameters (including the intercept), and n observations.
The test for significance of regression checks whether the model with predictors does better than a model with only an intercept.
We partition the total variability:
SST: total sum of squaresSSR: regression (model) sum of squaresSSE: error (residual) sum of squaresk (one for each slope)n − p = n − k − 1F₀ > Fα, k, n−p or if the p-value is small.| Source | SS | df | MS | F |
|---|---|---|---|---|
| Regression | SSR | k | MSR = SSR / k | F₀ = MSR / MSE |
| Error | SSE | n − p | MSE = SSE / (n − p) | – |
| Total | SST | n − 1 | – | – |
Interpretation: If H₀ is rejected, the model with predictors explains a significant portion of variability in Y (at least one slope is nonzero).
The global F-test is often paired with R² and adjusted R² to summarize overall fit.
R² measures the proportion of total variability in Y explained by the regression. It always increases when you add more predictors.
To “penalize” unnecessary predictors, we use adjusted R²:
R²adj increases only if the new predictor reduces MSE.Once the overall model is significant, we often ask whether individual predictors are helpful given that the others are already in the model.
n − p|t₀| > tα/2, n−p (two-sided test) or based on the p-value.To test whether predictor xj contributes to the model:
If we fail to reject H₀, xj does not provide a significant partial contribution, given that the other predictors remain in the model.
Sometimes we want to test a subset of predictors at once (for example, a block of interaction or quadratic terms).
Extra sum of squares due to the subset:
q = number of coefficients tested (size of the subset).q, dfden = n − p for the full model.F₀ > Fα, q, n−p or based on the p-value.Special case: When q = 1 (testing a single coefficient), the partial F-test is exactly equivalent to the t-test:
Y the model explains and help compare models.