Stat 136 Chapter 10 Nonnormality and Heteroskedasticity
Stat 136 Chapter 10 Nonnormality and Heteroskedasticity
UP School of Statistics
1 Reasons of Nonnormality
2 Implications of Nonnormality
6 Implications of Heteroskedasticity
1 Reasons of Nonnormality
2 Implications of Nonnormality
6 Implications of Heteroskedasticity
However, take note that the β̂ being the BLUE of β does not
require the assumption of normality of error terms.
Note also of the fact that because of the central limit
theorem, a lot of statistics used for hypothesis testing and
confidence interval estimation approximately follows the
normal distribution.
Because of these aforementioned realities, it was stated by a
lot of statisticians that the assumption of normality is the
least bothersome of the assumptions of regression analysis.
However it is still imperative to check for gross nonnormality!
:D
1 Reasons of Nonnormality
2 Implications of Nonnormality
6 Implications of Heteroskedasticity
1 Reasons of Nonnormality
2 Implications of Nonnormality
6 Implications of Heteroskedasticity
The following are some of the tests that may be used for validating
the assumption of normality of the error terms.
χ2 Goodness-of-fit test
Kolmogorov-Smirnov One-Sample Test
Wilk-Shapiro Test
Anderson-Darling Test
Cramer-von-Mises criterion
Jarque-Bera test
All of these tests above test the following hypotheses:
Ho: The error terms are normally distributed.
Ha: The error terms are not normally distributed.
χ2 Goodness-of-fit test
Description: The interest is in the number of subjects or objects
that fall in various categories. In the χ2 goodness-of-fit test, we
wish to test whether a significant difference exists between an
observed number of objects in a category and an expected number
of observations based on the null hypothesis. (This is discussed in
full in Stat 132).
Strengths: A very intuitive test.
Weaknesses: The test is very sensitive to the binning process.
There is also a requirement for the expected frequencies per bin
(≥ 5).
Shapiro-Wilk Test
Description: A test that essentially looks at the correlation
between the ordered data and the theoretical values from the
normal distribution.
Strengths: This is one of the most powerful normality-tests. It is
also one of the most popular tests too. One can use it for small
sample sizes.
Weaknesses: It being a very powerful test is a problem in itself,
especially if the sample size is too large. If the sample size is
sufficiently large this test may detect even trivial departures from
the null hypothesis (for example: outlying residuals). The test is
also very sensitive to many identical residual values.
where
x(i) with parentheses enclosing the subscript index i) is the i th
order statistic, i.e., the i th -smallest number in the sample;
x̄ is the sample mean.
m T V −1
the constants ai are given by (a1 , . . . , an ) = p
m T V −1 V −1 m
where m = (m1 , . . . , mn )T and m1 , . . . , mn are the expected
values of the order statistics of an iid sample from the
standard normal distribution, and V is the covariance matrix
of those order statistics.
UP School of Statistics Statistics 136 Chapter 10
Anderson-Darling Test
Anderson-Darling Test
Description: This is a test that compares the quantile values of
the empirical distribution of the data with the theoretical
distribution of interest. It is based upon the concept that when
given a hypothesized underlying distribution, the data can be
transformed to a uniform distribution.
Strengths: The Anderson-Darling test is one of the most powerful
statistics for detecting most departures from normality. It may be
used with small sample sizes, n ≤ 25.
Weaknesses: Very large sample sizes may reject the assumption
of normality with only slight imperfections, but industrial data with
sample sizes of 200 and more have passed the Anderson-Darling
test. The test is also very sensitive to many identical residual
values.
Cramer-von-Mises Criterion
Description: In statistics the Cramer-von-Mises criterion for
judging the goodness-of-fit of a probability distribution F ∗ (x)
compared to a given distribution F is given by
Z∞
W = 2
[F (x) − F ∗ (x)]2 dF (x)
−∞
Jarque-Bera Test
Description: In statistics, the Jarque-Bera test is a goodness-of-fit
measure of departure from normality, based on the sample kurtosis
and skewness.
Strengths: This is one of the most popular tests, mostly used in
econometrics (it still is).
Weaknesses: The test has low power for distributions with short
tails, especially for bimodal distributions. It is also an asymptotic
test and is very sensitive when the sample size is too low.
1 Reasons of Nonnormality
2 Implications of Nonnormality
6 Implications of Heteroskedasticity
1 Reasons of Nonnormality
2 Implications of Nonnormality
6 Implications of Heteroskedasticity
1 Reasons of Nonnormality
2 Implications of Nonnormality
6 Implications of Heteroskedasticity
The following are some of the tests that may be used for validating
the assumption of normality of the error terms.
Two sample test
Goldfeld-Quandt test
White’s heteroscedasticity test
Breusch-Pagan Test
All of these tests above test the following hypotheses:
Ho: The error variance is constant.
Ha: The error variance is not constant.
Two-Sample Test
Description: If the residual plot gives the impression that the
variance increases or decreases in a systematic manner related to
an independent variable X or to Y , a simple test is to fit separate
regressions to each half of the observations arranged by the level of
X , then compare their MST’s.
Ho: σ12 = σ22 vs. Ha: σ12 ̸= σ22
Strengths: A very intuitive test.
Weaknesses: Less powerful than most tests. Has to assume that
the population is normally distributed.
Test statistic:
(n1 − 1)S12
(n1 − 1)σ12 S2
Fc = 2
= 12 ∼ F(n1 −1,n2 −1)
(n2 − 1)S2 S2
(n2 − 1)σ22
Goldfeld-Quandt Test
Description: The test involves the calculation of two least squares
regression lines, one using the data thought to be associated with
low variance errors and the other using data thought to be
associated with high variance errors.
If the residual variances associated with each regression line are
approximately the same, the homoskedasticity assumption cannot
be rejected. Otherwise, there is a problem of heteroskedasticity.
Strengths: It has a nonparametric test which does not assume
normality of error terms.
Weaknesses: Error variance must be a monotonic function of the
specified explanatory variable. For example, when faced with a
quadratic function mapping the explanatory variable to error
variance the Goldfeld–Quandt test may improperly accept the null
hypothesis of homoskedastic errors.
Breusch-Pagan Test
Description: The Breusch-Pagan test is a test that assumes that
the error terms are normally distributed, with E (ϵi ) = 0 and
Var (ϵi ) = σi2 (i.e., nonconstant variance).
Strengths: A much more focused test on pure heteroskedasticity
than the White test.
Weaknesses: Normal error terms is a requirement to use the said
method.
1 Reasons of Nonnormality
2 Implications of Nonnormality
6 Implications of Heteroskedasticity
then
1
σ1 0 ... 0
1
0 ... 0
W = σ2
. .. .. ..
.. . . .
1
0 0 ...
σn
The larger σi is, the smaller the weight of that observation in
the estimation of β.
UP School of Statistics Statistics 136 Chapter 10
Remedial Measures for Heteroskedasticity