Weighted Least Squares is a generalization of Ordinary Least Squares (OLS) that can be used when the standard assumptions of constant variance (homoscedasticity) in the errors do not hold.
In WLS, we give each observation a weight that is inversely proportional to its variance. Observations with higher variance (less reliable) are given lower weight, while observations with lower variance (more reliable) are given higher weight.
Consider the linear model:
where \(\text{Var}(\boldsymbol{\varepsilon}) = \boldsymbol{\Omega} = \text{diag}(\sigma_1^2, \sigma_2^2, \ldots, \sigma_n^2)\)
The WLS estimator is given by:
where \(\mathbf{W} = \boldsymbol{\Omega}^{-1} = \text{diag}(1/\sigma_1^2, 1/\sigma_2^2, \ldots, 1/\sigma_n^2)\)
The main challenge in applying WLS is knowing the correct weights. In practice, these often need to be estimated, which can introduce additional uncertainty.
White's standard errors, also known as heteroscedasticity-consistent (HC) standard errors, were introduced by Halbert White in 1980. They provide a way to obtain consistent estimates of standard errors when heteroscedasticity is present, without having to specify the form of the heteroscedasticity.
Instead of trying to model the heteroscedasticity explicitly (as in WLS), White's approach uses the observed residuals to estimate the variance of the OLS estimator under heteroscedasticity.
The variance-covariance matrix of the OLS estimator under heteroscedasticity is:
White's estimator replaces \(\boldsymbol{\Omega}\) with a diagonal matrix of squared OLS residuals:
where \(\hat{\varepsilon}_i\) are the OLS residuals.
This proof demonstrates that a Weighted Least Squares problem can be transformed into an Ordinary Least Squares problem through a simple transformation of variables.
Consider the linear model:
where \(\text{Var}(\boldsymbol{\varepsilon}) = \boldsymbol{\Omega} = \text{diag}(\sigma_1^2, \sigma_2^2, \ldots, \sigma_n^2)\)
Define the weight matrix \(\mathbf{W}\) as the inverse of the variance matrix:
Define a transformation matrix \(\mathbf{P}\) such that \(\mathbf{P}'\mathbf{P} = \mathbf{W}\). We can choose \(\mathbf{P}\) as:
Transform the original model by pre-multiplying both sides by \(\mathbf{P}\):
Define new variables:
The transformed model is now:
Check the variance of the transformed error term:
The transformed model now has homoscedastic errors (constant variance).
Apply OLS to the transformed model:
This final expression is identical to the WLS estimator.
We have shown that by applying an appropriate transformation, a WLS problem can be converted into an OLS problem. The OLS estimator of the transformed model is identical to the WLS estimator of the original model. This proof demonstrates why WLS is the Best Linear Unbiased Estimator (BLUE) when the weights are correctly specified, as it reduces to OLS in a transformed space where the Gauss-Markov assumptions hold.