Confidence Intervals for Regression Coefficients

1. Introduction

In regression analysis, confidence intervals provide a range of plausible values for the true population parameters, given the observed data. We'll focus on the 95% confidence interval for the OLS estimator.

2. Formula for 95% Confidence Interval

For a given coefficient \(\beta_j\), the 95% confidence interval is given by:

\[ \hat{\beta}_j \pm t_{n-k,0.975} \cdot SE(\hat{\beta}_j) \]

Where:

3. Calculating the Standard Error

The standard error for \(\hat{\beta}_j\) is given by:

\[ SE(\hat{\beta}_j) = \sqrt{\hat{\sigma}^2 \cdot [(X'X)^{-1}]_{jj}} \]

Where:

4. Interpretation

A 95% confidence interval means that if we were to repeat the sampling process many times and calculate the confidence interval each time, about 95% of these intervals would contain the true population parameter.

For a given sample:

5. Assumptions

The validity of these confidence intervals relies on several assumptions:

  1. Linearity: The relationship between X and Y is linear.
  2. Independence: The observations are independent of each other.
  3. Homoscedasticity: The variance of the residuals is constant.
  4. Normality: The residuals are normally distributed.
  5. No perfect multicollinearity among the predictors.

Violation of these assumptions can lead to incorrect confidence intervals.

6. Example Calculation

Suppose we have the following results for a coefficient \(\beta_1\):

The 95% confidence interval would be:

\[ \begin{aligned} 2.5 \pm 1.98 \cdot 0.5 &= 2.5 \pm 0.99 \\ &= [1.51, 3.49] \end{aligned} \]

Interpretation: We are 95% confident that the true population value of \(\beta_1\) lies between 1.51 and 3.49.

7. Considerations