📘 Section 12.1.2 – Properties of Least Squares Estimators

🎯 Why This Matters

Understanding the properties of least squares estimators is crucial for interpreting regression results. It allows students to assess the reliability and accuracy of predictive models for outcomes like GPA, job offers, or salaries.

🔢 Model Setup

The general multiple linear regression model in matrix form is:

Y = Xβ + ε

📐 Least Squares Estimator

The least squares estimator minimizes the sum of squared residuals:

β̂ = (XᵀX)−1XᵀY

🧠 Key Properties

📊 Estimating σ² (Error Variance)

The error variance is estimated by:

σ̂² = SSE / (n − k − 1)

Where SSE = ∑(Yᵢ − Ŷᵢ)². This is also called the Mean Squared Error (MSE).

📈 Covariance Matrix of β̂

The covariance matrix for the estimated coefficients is:

Cov(β̂) = σ̂² (XᵀX)−1

This helps assess the variability and potential correlation between coefficient estimates.

🎓 Student-Based Example

Suppose we want to model a student's GPA based on:

The model is:

GPA = β₀ + β₁x₁ + β₂x₂ + β₃x₃ + ε

🧮 Least Squares Estimation Example

Step 1: Least Squares Function

Given n = 5 observations, the least squares objective is:

L = ∑i=15 [yi − (β₀ + β₁xi1 + β₂xi2 + β₃xi3)]²

Step 2: Matrix Form

X = [ [1, 15, 7.0, 40],
[1, 20, 8.0, 42],
[1, 10, 6.0, 35],
[1, 12, 6.5, 38],
[1, 18, 7.5, 41]
]
Y = [3.5, 3.8, 2.9, 3.2, 3.7]T

Step 3: Estimate Coefficients

Apply the formula:

β̂ = (XTX)−1XTY

Steps involved:

  1. Compute XTX (4×4 matrix)
  2. Compute XTY (4×1 vector)
  3. Multiply the inverse of XTX with XTY to get β̂

📊 Visualization