Title: Heteroskedasticity
1Chapter 5
2What is in this Chapter?
- How do we detect this problem
- What are the consequences of this problem?
- What are the solutions?
3What is in this Chapter?
- First, We discuss tests based on OLS residuals,
likelihood ratio test, G-Q test and the B-P test.
The last one is an LM test. - Regarding consequences, we show that the OLS
estimators are unbiased but inefficient and the
standard errors are also biased, thus
invalidating tests of significance
4What is in this Chapter?
- Regarding solutions, we discuss solutions
depending on particular assumptions about the
error variance and general solutions. - We also discuss transformation of variables to
logs and the problems associated with deflators,
both of which are commonly used as solutions to
the heteroskedasticity problem.
55.1 Introduction
- The homoskedasticityvariance of the error terms
is constant - The heteroskedasticityvariance of the error
terms is non-constant - Illustrative Example
- Table 5.1 presents consumption expenditures (y)
and income (x) for 20 families. Suppose that we
estimate the equation by ordinary least squares.
We get (figures in parentheses are standard
errors)
65.1 Introduction
75.1 Introduction
85.1 Introduction
95.1 Introduction
105.1 Introduction
115.1 Introduction
- The residuals from this equation are presented in
Table 5.3 - In this situation there is no perceptible
increase in the magnitudes of the residuals as
the value of x increases - Thus there does not appear to be a
heteroskedasticity problem.
125.2 Detection of Heteroskedasticity
135.2 Detection of Heteroskedasticity
145.2 Detection of Heteroskedasticity
155.2 Detection of Heteroskedasticity
165.2 Detection of Heteroskedasticity
175.2 Detection of Heteroskedasticity
185.2 Detection of Heteroskedasticity
- Some Other Tests
- Likelihood Ratio Test
- Goldfeld and Quandt Test
- Breusch-Pagan Test
195.2 Detection of Heteroskedasticity
205.2 Detection of Heteroskedasticity
- Goldfeld and Quandt Test
- If we do not have large samples, we can use the
Goldfeld and Quandt test. - In this test we split the observations into two
groups one corresponding to large values of x
and the other corresponding to small values of x
- Fit separate regressions for each and then apply
an F-test to test the equality of error
variances. - Goldfeld and Quandt suggest omitting some
observations in the middle to increase our
ability to discriminate between the two error
variances.
215.2 Detection of Heteroskedasticity
225.2 Detection of Heteroskedasticity
235.2 Detection of Heteroskedasticity
245.2 Detection of Heteroskedasticity
255.2 Detection of Heteroskedasticity
265.2 Detection of Heteroskedasticity
275.2 Detection of Heteroskedasticity
285.2 Detection of Heteroskedasticity
295.3 Consequences of Heteroskedasticity
305.3 Consequences of Heteroskedasticity
315.3 Consequences of Heteroskedasticity
325.3 Consequences of Heteroskedasticity
335.3 Consequences of Heteroskedasticity
345.3 Consequences of Heteroskedasticity
355.3 Consequences of Heteroskedasticity
365.3 Consequences of Heteroskedasticity
375.3 Consequences of Heteroskedasticity
385.4 Solutions to the Heteroskedasticity Problem
- There are two types of solutions that have been
suggested in the literature for the problem of
heteroskedasticity - Solutions dependent on particular assumptions
about si. - General solutions.
- We first discuss category 1. Here we have two
methods of estimation weighted least squares
(WLS) and maximum likelihood (ML).
395.4 Solutions to the Heteroskedasticity Problem
405.4 Solutions to the Heteroskedasticity Problem
Thus the constant term in this equation is the
slope coefficient in the original equation.
415.4 Solutions to the Heteroskedasticity Problem
425.4 Solutions to the Heteroskedasticity Problem
435.4 Solutions to the Heteroskedasticity Problem
- If we make some specific assumptions about the
errors, say that they are normal - We can use the maximum likelihood method, which
is more efficient than the WLS if errors are
normal
445.4 Solutions to the Heteroskedasticity Problem
455.4 Solutions to the Heteroskedasticity Problem
465.4 Solutions to the Heteroskedasticity Problem
475.4 Solutions to the Heteroskedasticity Problem
485.4 Solutions to the Heteroskedasticity Problem
495.5 Heteroskedasticity and the Use of Deflators
- There are two remedies often suggested and used
for solving the heteroskedasticity problem - Transforming the data to logs
- Deflating the variables by some measure of
"size."
505.5 Heteroskedasticity and the Use of Deflators
515.5 Heteroskedasticity and the Use of Deflators
525.5 Heteroskedasticity and the Use of Deflators
- One important thing to note is that the purpose
in all these procedures of deflation is to get
more efficient estimates of the parameters - But once those estimates have been obtained, one
should make all inferencescalculation of the
residuals, prediction of future values,
calculation of elasticities at the means, etc.,
from the original equationnot the equation in
the deflated variables.
535.5 Heteroskedasticity and the Use of Deflators
- Another point to note is that since the purpose
of deflation is to get more efficient estimates,
it is tempting to argue about the merits of the
different procedures by looking at the standard
errors of the coefficients. - However, this is not correct, because in the
presence of heteroskedasticity the standard
errors themselves are biased, as we showed earlier
545.5 Heteroskedasticity and the Use of Deflators
- For instance, in the five equations presented
above, the second and third are comparable and so
are the fourth and fifth. - In both cases if we look at the standard errors
of the coefficient of X, the coefficient in the
undeflated equation has a smaller standard error
than the corresponding coefficient in the
deflated equation - However, if the standard errors are biased, we
have to be careful in making too much of these
differences
555.5 Heteroskedasticity and the Use of Deflators
- In the preceding example we have considered miles
M as a deflator and also as an explanatory
variable - In this context we should mention some discussion
in the literature on "spurious correlation"
between ratios. - The argument simply is that even if we have two
variables X and Y that are uncorrelated, if we
deflate both the variables by another variable Z,
there could be a strong correlation between X/Z
and Y/Z because of the common denominator Z - It is wrong to infer from this correlation that
there exists a close relationship between X and Y
565.5 Heteroskedasticity and the Use of Deflators
- Of course, if our interest is in fact the
relationship between X/Z and Y/Z, there is no
reason why this correlation need be called
"spurious." - As Kuh and Meyer point out, "The question of
spurious correlation quite obviously does not
arise when the hypothesis to be tested has
initially been formulated in terms of ratios, for
instance, in problems involving relative prices.
575.5 Heteroskedasticity and the Use of Deflators
- Similarly, when a series such as money value of
output is divided by a price index to obtain a
'constant dollar' estimate of output, no question
of spurious correlation need arise. - Thus, spurious correlation can only exist when a
hypothesis pertains to undeflated variables and
the data have been divided through by another
series for reasons extraneous to but not in
conflict with the hypothesis framed an exact,
i.e., nonstochastic relation."
585.5 Heteroskedasticity and the Use of Deflators
- In summary, often in econometric work deflated or
ratio variables are used to solve the
heteroskedasticity problem - Deflation can sometimes be justified on pure
economic grounds, as in the case of the use of
"real" quantities and relative prices - In this case all the inferences from the
estimated equation will be based on the equation
in the deflated variables.
595.5 Heteroskedasticity and the Use of Deflators
- However, if deflation is used to solve the
heteroskedasticity problem, any inferences we
make have to be based on the original equation,
not the equation in the deflated variables - In any case, deflation may increase or decrease
the resulting correlations, but this is beside
the point. Since the correlations are not
comparable anyway, one should not draw any
inferences from them.
605.5 Heteroskedasticity and the Use of Deflators
615.5 Heteroskedasticity and the Use of Deflators
625.5 Heteroskedasticity and the Use of Deflators
635.5 Heteroskedasticity and the Use of Deflators
645.5 Heteroskedasticity and the Use of Deflators
655.5 Heteroskedasticity and the Use of Deflators
665.6 Testing the Linear Versus Log-Linear
Functional Form
675.6 Testing the Linear Versus Log-Linear
Functional Form
- When comparing the linear with the log-linear
forms, we cannot compare the R2 because R2 is the
ratio of explained variance to the total variance
and the variances of y and log y are different - Comparing R2's in this case is like comparing two
individuals A and B, where A eats 65 of a carrot
cake and B eats 70 of a strawberry cake - The comparison does not make sense because there
are two different cakes.
685.6 Testing the Linear Versus Log-Linear
Functional Form
- The Box-Cox Test
- One solution to this problem is to consider a
more general model of which both the linear and
log-linear forms are special cases. Box and Cox
consider the transformation
695.6 Testing the Linear Versus Log-Linear
Functional Form
705.6 Testing the Linear Versus Log-Linear
Functional Form
715.6 Testing the Linear Versus Log-Linear
Functional Form
725.6 Testing the Linear Versus Log-Linear
Functional Form
735.6 Testing the Linear Versus Log-Linear
Functional Form
745.6 Testing the Linear Versus Log-Linear
Functional Form
755.6 Testing the Linear Versus Log-Linear
Functional Form
765.6 Testing the Linear Versus Log-Linear
Functional Form
77Summary
- 1. If the error variance is not constant for all
the observations, this is known as the
heteroskedasticity problem. The problem is
informally illustrated with an example in Section
5.1. - 2. First, we would like to know whether the
problem exists. For this purpose some tests have
been suggested. We have discussed the following
tests - (a) Ramsey's test.
- (b) Glejser's tests.
- (c) Breusch and Pagan's test.
- (d) White's test.
- (e) Goldfeld and Quandt's test.
- (f) Likelihood ratio test.
78Summary
- 3. The consequences of the heteroskedasticity
problem are - (a) The least squares estimators are unbiased but
inefficient. - (b) The estimated variances are themselves
biased. - If the heteroskedasticity problem is detected, we
can try to solve it by the use of weighted least
squares. - Otherwise, we can at least try to correct the
error variances
79Summary
- 4. There are three solutions commonly suggested
for the heteroskedasticity problem - (a) Use of weighted least squares.
- (b) Deflating the data by some measure of "size.
- (c) Transforming the data to the logarithmic
form. - In weighted least squares, the particular
weighting scheme used will depend on the nature
of heteroskedasticity.
80Summary
- 5. The use of deflators is similar to the
weighted least squared method, although it is
done in a more ad hoc fashion. Some problems with
the use of deflators are discussed in Section
5.5. - 6. The question of estimation in linear versus
logarithmic form has received considerable
attention during recent years. Several
statistical tests have been suggested for testing
the linear versus logarithmic form. In Section
5.6 we discuss three of these tests the Box-Cox
test, the BM test, and the PE test. All are easy
to implement with standard regression packages.
We have not illustrated the use of these tests.