Let's define our matrices and vectors:
We want to prove:
In matrix notation, this is equivalent to:
Key properties:
We have proven that Σ(yᵢ - ŷᵢ)(ŷᵢ - ȳ) = 0 using matrix notation. This proof demonstrates the orthogonality of residuals to centered predicted values, which is a fundamental property in regression analysis. It underlies many important results in linear regression theory and provides insights into the nature of least squares estimation.
Let's visualize regression in n-dimensional space:
To show that Σ(yᵢ - ŷᵢ)(ŷᵢ - ȳ) = 0, or equivalently, e'(ŷ - ȳ1) = 0:
We can write ŷ as the sum of its mean component and the centered component:
Since e is orthogonal to both components of ŷ, it must be orthogonal to ŷ - ȳ1.
This geometric approach provides an intuitive understanding of why Σ(yᵢ - ŷᵢ)(ŷᵢ - ȳ) = 0. It's a direct consequence of how linear regression finds the best fit by projecting the observed values onto the subspace spanned by the predictors. The orthogonality of residuals to centered predicted values is a fundamental property that underpins many key results in regression analysis.