Heteroskedasticity in Linear Regression Models

LM-Stat (for testing join significance of independent variables)

Heteroskedasticity

Other large sample tests: The Lagrange Multiplier Statistics to test join significance of independent variables

Consider the model:

RcPTQi+FYduv4VBIRbVYXFhERIosnzhbQOtEuIWH

Explain the procedure of LM-test to test the null hypothesis that QRqDZCADs=  and  XX9uSGaIiIBbM4iIZgECAwECAwECAwUqILAFBgAo  have no effect on DZorMIjEQgA7  once the other factors have been controlled for.

The null hypothesis: ol5CWEOXnVOqpVWtl02QAGaBfZ4SYwASuaK9S8Ci .

Estimate the restricted model: AkWvkbR5ggAAOw== . Get the residuals XWaIiIBbM4iIZgECAwECAwECAwECAwECAwUmIAAI .

Regress XWaIiIBbM4iIZgECAwECAwECAwECAwECAwUmIAAI  on XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC . Get the XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC .

Compute QqLTomFqv2NGQEugSMYGC0eTFSrvdRLFjVBsFRdb .  Reject the null  if  QKPIhrVqTFmIlwCVeAgWjqHstAxBHQdE0OJihlGL  at predetermined significance level.

Heteroskedasticity

Remember:

  • Assumption MLR 5 (Homoskedasticity):  mygazcIADs=  or  AA0M88cYEHCqCUAeyE1LMHQEWxIeCQAUgB4AIAEE
  • Sampling variances of the OLS slope estimators:

Under MLR 1 ~ MLR 5 , conditional on the sample values of XWaIiAECAwECAwECAwECAwECAwECAwECAwECAwEC ,

(5.1)           KQZXEd0Zd5fqSFIxF9SKJFBIOnkpsQdgdh1lXxt9

for XWaIiIBbM4iIZgECAwECAwECAwECAwECAwECAwEC , where +ALSOOXwF8N4MuGtMOClF8l4M2vG43ez58+gNSVx  is the total variation in 2xy9wqwyAM+LPEhA5AkcdMikCKHEKJ+QgSoEADs=  and XX9uSG6AbmaIiIBbM4iIZgECAwECAwECAwECAwEC  is the R-squared from regressing 2xy9wqwyAM+LPEhA5AkcdMikCKHEKJ+QgSoEADs=  on all other independent variables (including an intercept).

  • Equation (5.1) still have an unknown parameter 100FYIADs= . The unbiased estimator of 0jw2QU4ueU+4gpFDtigVXgNS7fUjKV5B0eJI6xlk  is

(5.2)           jGpnkaAWQ0TVF7BS5ygAqLVpvaFamovn5aR6qmst

The term GEHQZGEyv2CNj2Cggs0IHQBIIgBmY4QOAYIPfxK1  is the degrees of freedom (df) for the general OLS problem with n observations and k independent variables (there are XW6AbmaIiIBbM4iIZgECAwECAwECAwVBIAAEhWie  parameters in a regression model with k independent variables and an intercept).

  • For conducting confidence interval and tests, we need to estimate the standard deviation of XW6AbmaIiIBbM4iIZgECAwECAwECAwECAwVDICCO , which is just the square root of the Qm9VCEUEQqWnRqlMnkMaRa5DEGxDAoZImUSzzk3E :

(5.3)           8kOVTSjtfZz6pN8y3heAXPXtJ9yrKXACcHkRIVx8

Since f2xGHX9uSGaIiIBbM4iIZgQZEEgUaiUSiJzy9EDh  is unknown, we replace it with its estimator f2xGHX9uSGaIiIBbM4iIZgECAwECAwECAwECAwEC . This gives the standard error of XW6AbmaIiIBbM4iIZgECAwECAwECAwECAwVDICCO

(5.4)           oZicTfcvTSkR3uuvGa03h2esIICOsxH+TDiAsRCH

  1. The standard error formula in (5.4) is not a valid estimator of XX9uWW6AbmaIiIBbM4iIZgECAwECAwECAwECAwEC  if the errors exhibit heteroskedasticity.
    • The presence of  heteroskedasticity does not cause bias in the XW6AbmaIiIBbM4iIZgECAwECAwECAwECAwVFICCO  but it does lead to bias in the usual formula for Qm9VCEUEQqWnRqlMnkMaRa5DEGxDAoZImUSzzk3E , which then invalidates the standard errors.
    • If we suspect heteroskedasticity, then the “usual” OLS standard errors are invalid and some corrective action should be taken.

Heteroskedasticity in OLS

  • Homoskedasticity assumption for multiple regression: the variance of the unobservable error XX9uSG5uRGaIiIBbMwECAwECAwECAwECAwECAwEC , conditional on the explanatory variables, is the same (= constant) for all combinations of outcomes of the explanatory variables.
  • Homoskedasticity fails whenever the variance of the unobservables changes across different segment of the population. The segments are determined by the diffferent values of the explanatory variables.
  • Example:
  • Saving equation 

TDnuSv2JSwkygPPz0IADs=

Homoskedasticity is not constant if the variance of the unobserved factors affecting saving increases with income.

  1. bMLkxnFPdKBMeQG40sYo0ukSBSBsHnoEnVhkY4yi

Homoskedasticty requires that the variance of the error XX9uSG5uRGaIiIBbMwECAwECAwECAwECAwECAwEC  does not depend on the levels of education, experience, or tenure. That is

dEnrUxwG5Q1PzteKGEB7WlEoAnEWQmUoDNKEMLL1

  • Homoskedasticity is needed to justify the usual t tests, F tests, and confidence intervals for OLS estimation of the linear regression model. In the presence of heteroskedasticity:
  • The usual  OLS t-statistics do not have t distributions in the presence of heteroskedasticity and the problem is not resolved by increasing sample size.
  • Similarly, F statistics are no longer F distributed
  • Also LM statistics no longer has an asymptotic chi square distribution.
  • In summary, the statistics we used to test hypothesis under the Gauss-Markov assumptions are not valid in the presence of heteroskedasticity.
  • If XX9uSG5uRGpqamaIiIBbM4iIZgECAwECAwECAwEC  is not constant, OLS in longer BLUE. We will see it is possible to find estimators that are more efficient than OLS in the presence of heteroskedasticity (but it requires knowing the form of the heteroskedasticity).

Heteroskedasticity-robust procedure

  • Heteroskedasticity-robust procedure: methods that remain valid (at least in large samples) whether or not the errors have constant variance.
  • In the presence of heteroskedasticity, OLS estimates are still useful but the standard errors, t, F, LM statistics should be modified so that they are valid, at least asymptotically.
  • Consider the model with a single independent variable

(5.5)           EWRY8iHJ+EXgQo9OFBktUykIADs=

The first four Gauss-Markov assumptions hold but not MLR 5; the errors contain heteroskedasticity:

(5.6)           QIBwSCwaj8ikcslsGh8Bh3NKrVqvgAhWudg6Md4l

The OLS estimator for XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC  is still the same: 

(5.7)           TG3k35GaLfOEBxscU6AnoEGgD1jYLjPUCI9wVIRQ

and can be written as function of population parameter XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC

(5.8)           L00kh4tGTZjElkuIxyUvyCShxDcONaXTgeYdEzYG .

Under assumption MLR 1 through MLR 4 (without the homoskedasticity assumption) we can show that

(5.9)           H0JGg9uIxqLvuBmEUmlJ1uloiKdaTtJIDH6kh3yv

where 9AbC+Lew4cOIEysWwnex48eL6xKZALnyEMqWC7cV  is the total sum of squares of the XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC .

Proof of (5.9):

XWaIiIBbM4iIZgECAwECAwECAwECAwECAwECAwEC

BbZ7DYNXPy7RDQPEc7p98VTAMa2wB+MJRhbfGlol , then

XWaIiIBbM4iIZgECAwECAwECAwECAwECAwECAwEC

EFjsKheKxZawWQQAAOw==

                                    poPtGEAA7    

                                    IyiIvDTD1F8WIVCMKwxUhmCTmEqpgw4odS7YsWCl

ltguAc+7GC0AQADs= .

  • When Owxwp9AVhrcjaRFoNgmAibOIdDZHzpFg5ECuBNTZ , XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwUu , this formula reduces to the usual form, WX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC . Equation (5.9) explicitly shows that, for the simple regression case, the variance of OLS estimator formula derived under homoskedasticity is no longer valid when heteroskedasticity is present.
  • White (1980) gives a valid estimator of XWaIiIBbM4iIZgECAwECAwECAwECAwECAwECAwEC , for heteroskedasticity of any form:

(5.10)    XE2BQiebFJACQohEJ0d9qkuLqQAjRyEBA7JFfWCM    gcAouy6kkmb5MkodslhV290Xxg4NVLEAA7

where XX9uSG5uRGaIiIBbM4iIZgECAwECAwECAwUvICBa  denote the OLS residuals from the initial regression of DZorMIjEQgA7  on XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC .

  • For the general multiple regression model IveWcBozRJsQBQAydGEoxCGAC0Y4mBqCkZokV0Qs , a valid estimator for zekQvYRgOACwQTXmsSBAA7  under assumption MLR 1 through MLR 4 is

(5.11)    XE2BQiebFJACQohEJ0d9qkuLqQAjRyEBA7JFfWCM    CxsAGVWAgmYyAQNNgE5fVo1diGAKV5NghwZLIE0o

where XX9uSG6AbmaIiIBbM4iIZgECAwECAwECAwECAwEC  denotes the i th residual from regressing 2xy9wqwyAM+LPEhA5AkcdMikCKHEKJ+QgSoEADs=  on all other independent variables and WX9uSG6AbmaIiIBbM4iIZgVwIAAogWiO5Rmsa3G+  is the sum of squared residuals from this regression (Note: recall the partialling out representation of the OLS estimates).

  • The square root of the quantity in (5.11) is called the heteroskedasticity-robust standard error for XW6AbmaIiIBbM4iIZgECAwECAwECAwECAwVDICCO . In econometrics, these standard errors are usually attributed to White (1980).
  • Once heteroskedasticity-robust standard errors are obtained, a heteroskedasticity-robust t-statistic can be computed using the standard formula

(5.12)         GJGsX8k4BgEAOw== .

  1. If the heteroskedasticity robust se are valid more often than the usual OLS se, why bother with the OLS se at all?
  2. One reason: if the homoskedasticity assumption holds and the errors are normally distributed, then the OLS t-statistic have exact t distributions regardless of n. The robust se and robust t-statistic are justified only as n becomes larger.
  3. With small sample sizes, the robust t-statistic can have distribution that are not very close to the t distribution so that we can not make inference.
  • The usual F-statistic:

(5.13)         6SKsRJ1UHen5MkXw2F6hGD8Azw8mHxHFEoHqjECN

where AAgADMDAekt8QQA7  is the sum of squared residuals from the unrestricted model, WX9uSG6AbmaIiIBbM4iIZgVrIAAkgWiO5Rmsa3G+  is the sum of squared residuals from the restricted model, XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC  is number of restrictions (i.e., number of explanatory variables dropped). Note that gAiiPCbogGd9Ys9Qvwj6CTRYX26ROyqhIYiQD2Fc .

  1. Special case: most regression packages include a special F statistic for testing overall significance of the regression. In this case, XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC  is that none of the explanatory variables has an effect on y. This F-statistic is

(5.14)         OIkcTWOIpBLDLRRfRVQjpslGK2coliiiT0q2jWSE

Note that (5.14) is the special form of (5.13), i.e. (5.13) is more general.

  • Equation (5.13) and (5.14) are no longer valid when heteroskedasticity presents (but the robust version has no simple form and will not be presented here).
  • To test of multiple exclusion restrictions that is robust to heteroskedasticity, use heteroskedasticity-robust LM statistic.
  • Consider the model

+AE4CIDZ9VSgFnQAiN8R+rFCFoTBBVgEKgMK8Xqg .

Suppose we would like to test

SPnGMVFYpdGTZhKIUElgaIiiiFYgSoRSsI3FNvW6

  1. The usual LM statistic can computed as follow (it only requires OLS regression):

(1) First estimate the restricted model (i.e., the model without XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwUs  and XX9uSGaIiIBbM4iIZgECAwECAwECAwUtILAFBwAs ), obtain the residuals XWaIiIBbM4iIZgECAwECAwECAwECAwECAwUmIAAI ;

(2) Regress XWaIiIBbM4iIZgECAwECAwECAwECAwECAwUmIAAI  on all of the independent variables (including XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwUs , XX9uSGaIiIBbM4iIZgECAwECAwECAwUtILAFBwAs  and an intercept);

(3)  Compute 80ABxoxooiggin5EnIEIkmAEEAADs=  where XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC  is the R-squared from the regression of XWaIiIBbM4iIZgECAwECAwECAwECAwECAwUmIAAI  on x.

  1. The robust version of LM statistic involves some extra work:

(1)   Obtain the residuals XWaIiIBbM4iIZgECAwECAwECAwECAwECAwUmIAAI  form the restricted model.

(2)   Obtain residuals XX9uSGaIiIBbM4iIZgECAwECAwECAwUvIAAEgSAC  from the regression of XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwUs  on wpwqAM83IkfgIAoYREAdUslUBUQhADs= , XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC , XX9uSGaIiIBbM4BuboiIZgECAwECAwECAwUtIAAE .

(3)   Obtain residuals XX9uSGaIiIBbM4iIZgECAwECAwECAwUuIAAEgSAC  from the regression of XX9uSGaIiIBbM4iIZgECAwECAwECAwUtILAFBwAs  on wpwqAM83IkfgIAoYREAdUslUBUQhADs= , XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC , XX9uSGaIiIBbM4BuboiIZgECAwECAwECAwUtIAAE .

(Thus we regress each of the independent variables excluded under the null on all the included independent variables, and keep the residuals each time).

(4)   Find the products between each XX9uSGaIiIBbM4iIZgUxIAAEgSAC1SmmYkOqcCzP  and XWaIiIBbM4iIZgECAwECAwECAwECAwECAwUmIAAI .

(5)   Run the regression of 1 on XX9uSG5uRGaIiIBbM4iIZgECAwECAwECAwECAwEC , XX9uSG5uRGaIiIBbM4iIZgECAwECAwECAwECAwEC  without an intercept.

(6)   Compute aj0SBvDoNADKCzE2clHa5kQCwA8GCqIRAxFIZUqK  where WX9uSG6AbmaIiIBbM4iIZgECAwECAwECAwECAwEC  is the sum of squared residuals from regression in (5). Under XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC , LM is distributed approximately as XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC .

Testing for heteroskedasticity

  • Focus on more modern tests: testing the assumption that the variance of the error does not depend on the independent variables.

The BP test for heteroskedasticity

  • Consider the model

(5.15)         qQ7gI2gSl83Yl7D5CnutbEJAgA7

where assumptions MLR 1 through MLR 4 are maintained. In particular, we assume rU8FUyXAsUrDVIWen4dFpQCSQiVzW4hDqkUHXtjE , so that OLS is unbiased and consistent.

  • The null hypothesis relates to assumption MLR 5 to be true:

(5.16)         GVgRRJhg2r6wxcLQC00F4LMG0YrcZvDNtKHA+A3l

If we can not reject the XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC  at a sufficiently small significance level, we conclude that heteroskedasticity is not a problem.

  • Since we assume a zero conditional expectation, then CQp5CocEBBgAeRXVUeAGz00VlUkLPRt8AJuHjRfC  and so the null of hypothesis of homoskedasticity is equivalent to

(5.17)         xW4RkEqdEHi8DsCyzGSqgDS8cdMrEoDchqTorIlj

This shows that, in order to test for violation of the homoskedasticity assumption, we want to test whether XX9uSG5uRGaIiIBbMwECAwECAwECAwUtICCOJBIc  is related (in expected value) to one or more of the explanatory variables. If XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC  is false, the expected value of XX9uSG5uRGaIiIBbMwECAwECAwECAwUtICCOJBIc , given the independent variables, can be virtually any function of the XX9uSG6AbmaIiIBbM4iIZgUsIAAEhaiUYpo2wKK+ .

  1. A simple approach is to assume a linear function:

(5.18)         5jbD4kBIAMjhH83fLwrFJqlvJtIkJd9PTAYQkgnX

where sSAPMigSvrZAgAOw==  is an error term with mean zero given the 2xy9wqwyAM+LPEhA5AkcdMikCKHEKJ+QgSoEADs= .

  1. The null hypothesis of homoskedasticity is

(5.19)         CbG5I4Owz0dtOGSfkzMUxGendz6OSXG4kGQhJEaR

  1. Then either the F or LM statistics for the overall (join) significance of the independent variables in explaining XX9uSG5uRGaIiIBbMwECAwECAwECAwUtICCOJBIc  can be used to test (5.19).
  2. Since we do not know actual errors in the population model, we should use the OLS residuals CgUpp6di56QSwWAExgyMCRZrqXTkQYAW1IQMuWEB  and estimate the regression

(5.20)         JKFcfhYwSN4ze9k1gdisRn4UO+Cofzqt3iDYQ+St

  1. The F statistic and LM statistics both depend on the R-squared from regression (5.20) (which we call C23XiAIwNEIoHQ3RKEAAZRGZDf1BNAGtgAYN5ShG ).
    • The F statistic is

(5.21)               7yniTdTIHhNHJKRE1UAg0uU9IQh9p0jhDVEGJdBE

which is distributed as r04BslsOrEpBADs=  under the null hypothesis of homoskedasticity.

  • The LM statistic is

(5.22)               MrIkUkWFR1DJmknsiQAhkKCAAA7

which is distributed as XX9uSGpqamaIiIBbM4iIZgECAwECAwECAwECAwEC .

The LM version is called the Breusch-Pagan test for heteroskedasticity (BP test). This form is suggested by Koenker (1983).

  • The BP tests for heteroskedasticity:
  • Estimate the model by OLS. Obtain the squared OLS residuals.
  • Run the auxiliary regression by OLS: squared OLS residuals on the independent variables

JKFcfhYwSN4ze9k1gdisRn4UO+Cofzqt3iDYQ+St

Keep the R-squared from this regression C23XiAIwNEIoHQ3RKEAAZRGZDf1BNAGtgAYN5ShG

  1. Form either the F-stat or the LM stat and compute the p-value. The distribution is A7ZcMBiy5QQ1nZ0NO50WBNKbk4PfFcDRU4bABEBd  for the former and XX9uSG5uWWaIiIBbM4iIZgECAwECAwECAwECAwEC  for the latter.
  2. If the p-value is sufficiently small, we reject the null hypothesis of homoskedasticity.

The White test for heteroskedasticity

An alternative test for heteroskedasticity is to assume that under the null hypothesis the squared error is uncorrelated with all the independent variables XX9uSG6AbmaIiIBbM4iIZgECAwECAwECAwECAwEC , the squares of the independent variables XX9uSG6AbmaIiIBbM4iIZgECAwECAwECAwECAwEC , and all the cross-products nrkAv1AITt5HykcNVdIAJgCmSBBFlEbGe9YAucZQ . This is motivated by White (1980).

  • When the model contains XX9uSGaIiIBbM4BuboiIZgECAwECAwECAwVPIAAw  independent variables, the White test is based on an estimation of

AApKBFQbXbwyRDxSSRBoZshGBX+jDfqx71pBAAA7

  • Compared with the Breusch Pagan test, this equation has six more regressors. The White test for heteroskedasticity is the LM statistic or F statistic (both tests have asymptotic justification) for testing that all of the coefficient delta in the equation are zero, except for the intercept.
  • The White test uses many degrees of freedom for models with just a moderate number of independent variables. (E.g. with 6 independent variables in the model, the White regression would involve 27 regressors). Wooldridge proposed an alternative regression equivalent to the White regression but is more conserving on degrees of freedom.
  • The (special case of) White tests for heteroskedasticity:
  • Estimate the model by OLS. Obtain the OLS residuals XX9uSG5uRGaIiIBbM4iIZgECAwECAwECAwUjICCK and the fitted value XX9uSG5uRG6AbmaIiIBbM4iIZgECAwECAwECAwUw . Compute the squared residuals XX9uSG5uRGaIiIBbM4BuboiIZgECAwECAwECAwEC  and the squared fitted value XX9uSG6AbmaIiIBbM4iIZgECAwECAwECAwECAwEC .
  • Run the auxiliary regression by OLS: squared OLS residuals on the independent variables

JGy6NMwEFDNlELrx5FLJjxv2FI2uPcypBAAA7

Note that XX9uSG5uRG6AbmaIiIBbM4iIZgECAwECAwECAwUw  is a function of all f2xGHW5GM39ZSH9uSG6AbmaIiIBbM4iIZgECAwEC , while XX9uSG6AbmaIiIBbM4iIZgECAwECAwECAwECAwEC  is a function for all the squares and cross products of f2xGHW5GM39ZSH9uSG6AbmaIiIBbM4iIZgECAwEC .

Keep the R-squared from this regression C23XiAIwNEIoHQ3RKEAAZRGZDf1BNAGtgAYN5ShG

  1. Form either the F-stat or the LM stat for the null hypothesis QhuZ66U9Iu1qfihCfTb5eTtRHUYxhfKOw4IuCCNW  and compute the p-value. The distribution is XX9uSGaIiIBbM4BuboiIZgECAwECAwECAwECAwEC  for the former and XX9uSGaIiIBbM4iIZgVGICCOJKkESamO2uqWzyvP  for the latter.
  2. If the p-value is sufficiently small, we reject the null hypothesis of homoskedasticity.

If heteroskedasticity is a problem, some corrective measure should be taken. One possibility is to just use the heteroskedasticity-robust standard errors and test statistics discussed previously. Another possibility is to use Weighted Least Squares Estimation.

Weighted least squares estimation

  • When heteroskedasticity is present, OLS is still unbiased but no longer the most efficient. A more efficient estimator than OLS exists, and it produces t and F statistics that have t and F distributions.
  • The caveat is that we must be very specific about the nature of the heteroskedasticity.
  • Assume that

(5.23)         gPx0rfkHCqcvb4uYEj4Z6FPIgJQAzAGA88LdkIpG

where XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC  is some function of the explanatory variables that determines the heteroskedasticity. Assume for a moment that the function XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC  is known (but 0jw2QU4ueU+4gpFDtigVXgNS7fUjKV5B0eJI6xlk  is still unknown). For a random drawing, we can write

(5.24)         kt7d+EcEfwL2dwp5BSYxkxBDJbhFRw46AWGEaWiz .

Example: consider the simple saving function

(5.25)         yVqymNC01BAAA7

(5.26)         QrUYEFQZy4UJycYb4hLYlVpM0A5Rq6tX2zESG1Xh .

Here rpgCKyjPWx7XwfOQV1XJ5fV5dkWX9qg0cPckuFiF ; the variance of the error is proportional to the level of income. This means that as income increases, the variability in savings increases. The standard deviation of XX9uSG5uRGaIiIBbM4iIZgECAwECAwECAwECAwUo  conditional on XX9uSGaIiIBbM4iIZgECAwECAwECAwVGICAOYmme  is zvqAJaBAagUZSZFWEyToBBGBBEAGhSNNiRPmNJiM .

Given a heteroskedasticity in (5.26), note that if we divide XX9uSG5uRGaIiIBbM4iIZgECAwECAwECAwECAwUo  by XUgZBOWB1ocQqWU1QO08F7DaJeCAAA7  we will have a new error with constant variance. Thus we can transformed (5.26) into

(5.27)         ccWxxC77zJm1sTYnXXUzcGtiFfHxEEADs=  or in more general term

(5.28)         ecSpiopQzxoxF2PepIrlrYp6cQpIWGYhYsHuqoEZ  

Rd2bJLBpA3yY4ZzCTGMcUUwT2eqPNLWCKwIqNxfC

  1. In this transformed equation, XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC  has a zero mean and a constant variance 0jw2QU4ueU+4gpFDtigVXgNS7fUjKV5B0eJI6xlk , conditional on gIDyk0JgByvpAh2GyCFAAA7 .

hShH8axgNZ8B8IBxQaACmRGvy00tRQtcRRAAADs= .

  1. This means that if the original equation (5.25) satisfies the first four Gauss-Markov assumptions, then the transformed equation (5.28) satisfies all five Gauss-Markov assumptions.
  2. Thus, equation (5.28) can be estimated with OLS and we can simply use the resulted statistics for inference.
  3. The resulted estimates are examples of generalized least squares (GLS) estimators. The GLS estimators for correcting heteroskedasticity are called weighted least squares (WLS) estimators. In our example, the weight is XWaIiIBbM4iIZgECAwECAwECAwECAwECAwECAwEC .
  4. The idea of WLS is to give less weight to observations with a higher error variance. As comparison, OLS gives each observation the same weight. Mathematically, the WLS estimator are the values of f0huf2xGHX9ZSH9uSG6AbmaIiIBbM4iIZgECAwEC  that minimize

(5.29)         0KhBhY7I6eKaOyEFVSRglePnn1ict6ckZwJq6KGI

  1. Bringing the square root of XWaIiIBbM4iIZgECAwECAwECAwECAwECAwECAwEC inside the squared residuals, the weighted sum of squared residuals is identical to the sum of squared residuals in the transformed variables

(5.30)         nUyYSnFKBPCqhaAoOlPCjVKCkkYo9qUayUpk756i

The WLS estimators that minimizes (5.29) are simply the OLS estimators from (5.30).

  1. Most regression packages have a feature for computing WLS. We just specify, as usual, the dependent and independent variables, and the weighting function XWaIiIBbM4iIZgECAwECAwECAwECAwECAwECAwEC . That is, we specify weights proportional to the inverse of the variance, not proportional to the standard deviation.

What are the properties of WLS if our choice for XX9uSGaIiIBbM4iIZgECAwECAwECAwECAwECAwEC is incorrect?

  • WLS continues to be unbiased and consistent for estimating the XX9uSG6AbmaIiIBbM4iIZgECAwECAwECAwU7ICCO .
  • However, the reported standard errors, t-stat and F-stat are not valid if we incorrectly specify the form of heteroskedasticity.
  • WLS is only guaranteed to be more efficient than OLS if we have correctly chosen the form of heteroskedasticity.
  • When we are not certain about the form of heteroskedasticity:
    1. Use OLS and compute robust standard errors and test statistics