Given a linear regression model with normally distributed errors and assuming a flat (improper) prior on the parameters, the Maximum A Posteriori (MAP) estimate is equivalent to the Ordinary Least Squares (OLS) estimate.
Consider the linear regression model:
where \(\mathbf{y}\) is an \(n \times 1\) vector of observations, \(\mathbf{X}\) is an \(n \times p\) matrix of predictors, \(\boldsymbol{\beta}\) is a \(p \times 1\) vector of coefficients, and \(\boldsymbol{\varepsilon}\) is an \(n \times 1\) vector of errors.
Assume \(\boldsymbol{\varepsilon} \sim \mathcal{N}(\mathbf{0}, \sigma^2\mathbf{I})\).
1. The likelihood function is:
2. Assuming a flat prior \(P(\boldsymbol{\beta}) = \text{constant}\), the posterior is proportional to the likelihood:
3. The MAP estimate maximizes the posterior, which is equivalent to maximizing the likelihood. Taking the log:
4. Maximizing this is equivalent to minimizing:
5. This is exactly the OLS objective function. To find the minimum, we differentiate with respect to \(\boldsymbol{\beta}\) and set to zero:
6. Solving this equation:
7. The solution is the OLS estimator:
Conclusion: The MAP estimate with a flat prior on \(\boldsymbol{\beta}\) and normally distributed errors is identical to the OLS estimate. This demonstrates that under these conditions, the Bayesian approach (MAP) and the frequentist approach (OLS) yield the same point estimate for the regression coefficients.