chapter 5

80
Chapter 5 Heteroskedasticity

Upload: esben

Post on 25-Feb-2016

25 views

Category:

Documents


0 download

DESCRIPTION

Chapter 5. Heteroskedasticity. What is in this Chapter?. How do we detect this problem What are the consequences of this problem? What are the solutions?. What is in this Chapter?. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Chapter 5

Chapter 5

Heteroskedasticity

Page 2: Chapter 5

What is in this Chapter?• How do we detect this problem• What are the consequences of this proble

m?• What are the solutions?

Page 3: Chapter 5

What is in this Chapter?

• First, We discuss tests based on OLS residuals, likelihood ratio test, G-Q test and the B-P test. The last one is an LM test.

• Regarding consequences, we show that the OLS estimators are unbiased but inefficient and the standard errors are also biased, thus invalidating tests of significance

Page 4: Chapter 5

What is in this Chapter?

• Regarding solutions, we discuss solutions depending on particular assumptions about the error variance and general solutions.

• We also discuss transformation of variables to logs and the problems associated with deflators, both of which are commonly used as solutions to the heteroskedasticity problem.

Page 5: Chapter 5

5.1 Introduction

• The homoskedasticity=variance of the error terms is constant

• The heteroskedasticity=variance of the error terms is non-constant

• Illustrative Example – Table 5.1 presents consumption expenditures

(y) and income (x) for 20 families. Suppose that we estimate the equation by ordinary least squares. We get (figures in parentheses are standard errors)

Page 6: Chapter 5

5.1 Introduction

Page 7: Chapter 5

5.1 Introduction

Page 8: Chapter 5

5.1 Introduction

Page 9: Chapter 5

5.1 Introduction

Page 10: Chapter 5

5.1 Introduction

Page 11: Chapter 5

5.1 Introduction

• The residuals from this equation are presented in Table 5.3

• In this situation there is no perceptible increase in the magnitudes of the residuals as the value of x increases

• Thus there does not appear to be a heteroskedasticity problem.

Page 12: Chapter 5

5.2 Detection of Heteroskedasticity

Page 13: Chapter 5

5.2 Detection of Heteroskedasticity

Page 14: Chapter 5

5.2 Detection of Heteroskedasticity

Page 15: Chapter 5

5.2 Detection of Heteroskedasticity

Page 16: Chapter 5

5.2 Detection of Heteroskedasticity

Page 17: Chapter 5

5.2 Detection of Heteroskedasticity

Page 18: Chapter 5

5.2 Detection of Heteroskedasticity

• Some Other Tests– Likelihood Ratio Test – Goldfeld and Quandt Test – Breusch-Pagan Test

Page 19: Chapter 5

5.2 Detection of Heteroskedasticity

• Likelihood Ratio Test

Page 20: Chapter 5

5.2 Detection of Heteroskedasticity

• Goldfeld and Quandt Test– If we do not have large samples, we can use the Gold

feld and Quandt test.– In this test we split the observations into two groups

— one corresponding to large values of x and the other corresponding to small values of x —

– Fit separate regressions for each and then apply an F-test to test the equality of error variances.

– Goldfeld and Quandt suggest omitting some observations in the middle to increase our ability to discriminate between the two error variances.

Page 21: Chapter 5

5.2 Detection of Heteroskedasticity

• Breusch-Pagan Test

Page 22: Chapter 5

5.2 Detection of Heteroskedasticity

Page 23: Chapter 5

5.2 Detection of Heteroskedasticity

• Illustrative Example

Page 24: Chapter 5

5.2 Detection of Heteroskedasticity

Page 25: Chapter 5

5.2 Detection of Heteroskedasticity

Page 26: Chapter 5

5.2 Detection of Heteroskedasticity

Page 27: Chapter 5

5.2 Detection of Heteroskedasticity

Page 28: Chapter 5

5.2 Detection of Heteroskedasticity

Page 29: Chapter 5

5.3 Consequences of Heteroskedasticity

Page 30: Chapter 5

5.3 Consequences of Heteroskedasticity

Page 31: Chapter 5

5.3 Consequences of Heteroskedasticity

Page 32: Chapter 5

5.3 Consequences of Heteroskedasticity

Page 33: Chapter 5

5.3 Consequences of Heteroskedasticity

Page 34: Chapter 5

5.3 Consequences of Heteroskedasticity

Page 35: Chapter 5

5.3 Consequences of Heteroskedasticity

Page 36: Chapter 5

5.3 Consequences of Heteroskedasticity

Page 37: Chapter 5

5.3 Consequences of Heteroskedasticity

Page 38: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

• There are two types of solutions that have been suggested in the literature for the problem of heteroskedasticity: – Solutions dependent on particular assumptions about σi.

– General solutions.• We first discuss category 1. Here we have two m

ethods of estimation: weighted least squares (WLS) and maximum likelihood (ML).

Page 39: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

• WLS

Page 40: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

Thus the constant term in this equation is the slope coefficient in the original equation.

Page 41: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

Page 42: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

Page 43: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

• If we make some specific assumptions about the errors, say that they are normal

• We can use the maximum likelihood method, which is more efficient than the WLS if errors are normal

Page 44: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

Page 45: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

Page 46: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

• Illustrative Example

Page 47: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

Page 48: Chapter 5

5.4 Solutions to the Heteroskedasticity Problem

Page 49: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• There are two remedies often suggested and used for solving the heteroskedasticity problem: – Transforming the data to logs– Deflating the variables by some measure of

"size."

Page 50: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

Page 51: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

Page 52: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• One important thing to note is that the purpose in all these procedures of deflation is to get more efficient estimates of the parameters

• But once those estimates have been obtained, one should make all inferences—calculation of the residuals, prediction of future values, calculation of elasticities at the means, etc., from the original equation—not the equation in the deflated variables.

Page 53: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• Another point to note is that since the purpose of deflation is to get more efficient estimates, it is tempting to argue about the merits of the different procedures by looking at the standard errors of the coefficients.

• However, this is not correct, because in the presence of heteroskedasticity the standard errors themselves are biased, as we showed earlier

Page 54: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• For instance, in the five equations presented above, the second and third are comparable and so are the fourth and fifth.

• In both cases if we look at the standard errors of the coefficient of X, the coefficient in the undeflated equation has a smaller standard error than the corresponding coefficient in the deflated equation

• However, if the standard errors are biased, we have to be careful in making too much of these differences

Page 55: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• In the preceding example we have considered miles M as a deflator and also as an explanatory variable

• In this context we should mention some discussion in the literature on "spurious correlation" between ratios.

• The argument simply is that even if we have two variables X and Y that are uncorrelated, if we deflate both the variables by another variable Z, there could be a strong correlation between X/Z and Y/Z because of the common denominator Z

• It is wrong to infer from this correlation that there exists a close relationship between X and Y

Page 56: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• Of course, if our interest is in fact the relationship between X/Z and Y/Z, there is no reason why this correlation need be called "spurious."

• As Kuh and Meyer point out, "The question of spurious correlation quite obviously does not arise when the hypothesis to be tested has initially been formulated in terms of ratios, for instance, in problems involving relative prices.

Page 57: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• Similarly, when a series such as money value of output is divided by a price index to obtain a 'constant dollar' estimate of output, no question of spurious correlation need arise.

• Thus, spurious correlation can only exist when a hypothesis pertains to undeflated variables and the data have been divided through by another series for reasons extraneous to but not in conflict with the hypothesis framed an exact, i.e., nonstochastic relation."

Page 58: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• In summary, often in econometric work deflated or ratio variables are used to solve the heteroskedasticity problem

• Deflation can sometimes be justified on pure economic grounds, as in the case of the use of "real" quantities and relative prices

• In this case all the inferences from the estimated equation will be based on the equation in the deflated variables.

Page 59: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• However, if deflation is used to solve the heteroskedasticity problem, any inferences we make have to be based on the original equation, not the equation in the deflated variables

• In any case, deflation may increase or decrease the resulting correlations, but this is beside the point. Since the correlations are not comparable anyway, one should not draw any inferences from them.

Page 60: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

• Illustrative Example

Page 61: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

Page 62: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

Page 63: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

Page 64: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

Page 65: Chapter 5

5.5 Heteroskedasticity and the Use of Deflators

Page 66: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

Page 67: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

• When comparing the linear with the log-linear forms, we cannot compare the R2 because R2 is the ratio of explained variance to the total variance and the variances of y and log y are different

• Comparing R2's in this case is like comparing two individuals A and B, where A eats 65% of a carrot cake and B eats 70% of a strawberry cake

• The comparison does not make sense because there are two different cakes.

Page 68: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

• The Box-Cox Test– One solution to this problem is to consider a more

general model of which both the linear and log-linear forms are special cases. Box and Cox consider the transformation

Page 69: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

Page 70: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

Page 71: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

Page 72: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

Page 73: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

Page 74: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

Page 75: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

Page 76: Chapter 5

5.6 Testing the Linear Versus Log-Linear Functional Form

Page 77: Chapter 5

Summary• 1. If the error variance is not constant for all the observati

ons, this is known as the heteroskedasticity problem. The problem is informally illustrated with an example in Section 5.1.

• 2. First, we would like to know whether the problem exists. For this purpose some tests have been suggested. We have discussed the following tests:– (a) Ramsey's test.– (b) Glejser's tests.– (c) Breusch and Pagan's test.– (d) White's test.– (e) Goldfeld and Quandt's test.– (f) Likelihood ratio test.

Page 78: Chapter 5

Summary

• 3. The consequences of the heteroskedasticity problem are:– (a) The least squares estimators are unbiased but inef

ficient.– (b) The estimated variances are themselves biased.– If the heteroskedasticity problem is detected, we can t

ry to solve it by the use of weighted least squares. – Otherwise, we can at least try to correct the error vari

ances

Page 79: Chapter 5

Summary

• 4. There are three solutions commonly suggested for the heteroskedasticity problem– (a) Use of weighted least squares.– (b) Deflating the data by some measure of "size.“– (c) Transforming the data to the logarithmic form.– In weighted least squares, the particular weighting sc

heme used will depend on the nature of heteroskedasticity.

Page 80: Chapter 5

Summary• 5. The use of deflators is similar to the weighted least

squared method, although it is done in a more ad hoc fashion. Some problems with the use of deflators are discussed in Section 5.5.

• 6. The question of estimation in linear versus logarithmic form has received considerable attention during recent years. Several statistical tests have been suggested for testing the linear versus logarithmic form. In Section 5.6 we discuss three of these tests: the Box-Cox test, the BM test, and the PE test. All are easy to implement with standard regression packages. We have not illustrated the use of these tests.