Published online by Cambridge University Press: 05 March 2004
Correcting for heteroskedasticity of unspecified form.
Consider the basic linear regression model with heteroskedasticity, yi = xi′ β + ei, with i = 1,…, n, where ei is a zero-mean random variable distributed independently from all xi and all ej with i ≠ j. The variance of ei is σi2, where all σi2, i = 1,…, n may be different. The “true” generalized least squares estimator (GLSE)
of β is obtained as the ordinary least squares estimator (OLSE) from the regression of the yi /σi on the xi /σi. When, as usual, the σi are unknown, we might instead be tempted to use the absolute value of the residual
, with b the OLSE from the regression of the yi on the xi, and obtain a “feasible” GLSE
, say, by regressing the
on the
. Intuitively, this is not a sensible approach, but one is hard put to find an argument in the literature. What are the (finite or asymptotic) distributional properties of
? (One might conjecture that
is consistent and asymptotically normal.) Is there any sense in which
constitutes an improvement over b?