Frisch–Waugh–Lovell theorem

Theorem in statistics and econometrics From Wikipedia, the free encyclopedia

In econometrics, the Frisch–Waugh–Lovell[a] (FWL) theorem is named after the econometricians Ragnar Frisch, Frederick V. Waugh, and Michael C. Lovell.[1][2][3]

Background

The Frisch-Waugh-Lovell theorem is an algebraic result for regressions estimated by least squares, the most commonly used estimator in applied econometrics.[4] Least squares is a method of estimating coefficients in models which are linear in parameters. That is, the outcome variable is modeled as a linear combination of the input variables plus some error term. The least squares solution is that which sets the input variables' coefficients to minimize the sum of squared errors. Under a certain set of assumptions, the Gauss–Markov theorem, least squares estimation is the best linear unbiased estimator.

Let be any outcome variable and a set of predictor variables, such that , and suppose observations of are sampled. If is modeled as a linear function of , it can be written as . The least squares estimator sets the coefficients to minimize the sum of squared errors . With observations this involves minimizing across equations, and is typically written in matrix form as , where and are -dimensional column vectors and is an -by--dimensional matrix. Then, the least squares solution is .[5]

In regressions estimated by least squares, it is common to refer to a coefficient as the effect of that variable "holding constant" the other input variables.[6] For example, if wage is modeled as a function of education and work experience, the coefficient for education is interpreted as the difference in the expectation of wage for a unit difference in education, "holding constant" work experience. Econometrician Arthur Goldberger frames the Frisch-Waugh-Lovell theorem as "giving content to th[is] language".[7]

Definition and interpretation

The Frisch-Waugh-Lovell theorem states that in a least squares-estimated regression of the form

any coefficient can be estimated by the two-step process of:

  1. Regress on the set of other right-hand-side variables, obtaining residuals
  2. Regress on , obtaining

This two-step process is referred to as the residual regression or equivalently the regression anatomy theorem.[8][9][10]

The theorem shows that coefficients in a multiple regression reflect the relationship between the associated variable and the outcome variable after removing the part linearly explained by the other predictor variables.[11][9] This is the basis for understanding the contribution of each single variable to a multivariate regression (see, for instance, Ch. 13 in [12]).

Double residual regression

The double residual regression is the three-step process:

  1. Regress on the set of other right-hand-side variables, obtaining residuals
  2. Regress on the set of right-hand-side variables excluding , obtaining residuals
  3. Regress on , estimating and

This yields an identical coefficient to the two-step process.[7][13] It includes the additional feature that the residuals from the regression in step 3 equal the residuals in the full regression.[11]

Multivariate definition

Consider the regression , where and are -dimensional column vectors, is an -by- matrix, and is an -by- matrix. Then, the Frisch-Waugh-Lovell theorem states that

where , the residuals from the regression of on , and , the residuals from the regression of on . The first expression of is the residual regression, and the second the double residual regression.[7]

History

The origin of the theorem is uncertain, but it was well-established in the realm of linear regression before the Frisch and Waugh paper. George Udny Yule's comprehensive analysis of partial regressions, published in 1907, included the theorem in section 9 on page 184.[14]

Yule emphasized the theorem's importance for understanding multiple and partial regression and correlation coefficients, as mentioned in section 10 of the same paper.[14]

Yule 1907 also introduced the partial regression notation which is still in use today.

In 1962, Richard Stone generalized the theorem to apply to an arbitrary number of variables which may be chosen for special analysis in the same way that time was distinguished in Frisch's and Waugh's original formulation.[15]

In 1963, Lovell published a proof considered more straightforward and intuitive.[2] In recognition, people generally add his name to the theorem name.

Proof

Consider the linear regression and annihilator matrix . Premultiplying both sides of the regression equation by the annihilator matrix removes from and the component linearly explained by :

Then, by the least squares result, and . This concludes the proof.[16][7]

Extensions

Standard errors

The Frisch-Waugh-Lovell theorem applies to the standard errors of the partial and full regressions, where they differ (in the homoskedastic case) only by a degrees of freedom adjustment.[17]

See also

Notes

References

Related Articles

Wikiwand AI