A Generalized Normal Distribution
A Generalized Normal Distribution
Saralees Nadarajah
To cite this article: Saralees Nadarajah (2005) A generalized normal distribution, Journal of
Applied Statistics, 32:7, 685-694, DOI: 10.1080/02664760500079464
SARALEES NADARAJAH
Department of Statistics, University of Nebraska, Lincoln, USA
ABSTRACT Undoubtedly, the normal distribution is the most popular distribution in statistics. In
this paper, we introduce a natural generalization of the normal distribution and provide a
comprehensive treatment of its mathematical properties. We derive expressions for the nth
moment, the nth central moment, variance, skewness, kurtosis, mean deviation about the mean,
mean deviation about the median, Rényi entropy, Shannon entropy, and the asymptotic
distribution of the extreme order statistics. We also discuss estimation by the methods of moments
and maximum likelihood and provide an expression for the Fisher information matrix.
KEY WORDS : Estimation, entropy, generalized normal distribution, moments, normal distribution,
order statistics
Introduction
A random variable X is said to have the normal distribution with parameters m and s if its
p.d.f. is given by
1 (x m)2
f (x) ¼ pffiffiffiffiffiffi exp (1)
2ps 2 s2
In addition to containing equation (1) as the particular case for s ¼ 2, Equation (2) also
contains the Laplace distribution as the particular case for s ¼ 1. Like the normal distri-
bution, equation (2) is bell-shaped and unimodal with mode at x ¼ m. Furthermore,
f (m) ¼ K. Figure 1 illustrates some possible shapes of equation (2).
Figure 2. The generalized normal hazard rate function (equation (7)) for m ¼ 0 and s ¼ 1
Moments
If X has the p.d.f. (equation (2)) then the standardized variable Z ¼ (X m)=s will have
the p.d.f.
s exp {jzjs }
f (z) ¼
2G(1=s)
688 S. Nadarajah
E(X) ¼ m (8)
s2 G(3=s)
E(X 2 ) ¼ m2 þ
G(1=s)
3ms2 G(3=s)
E(X 3 ) ¼ m3 þ
G(1=s)
and
In particular, the first three central moments, skewness and the kurtosis of X are given by
s2 G(3=s)
Var(X) ¼ (9)
G(1=s)
E½{X E(X)}3 ¼ 0
s4 G(5=s)
E½{X E(X)}4 ¼
G(1=s)
Skewness(X) ¼ 0
A Generalized Normal Distribution 689
and
G(1=s) G(5=s)
Kurtosis(X) ¼ (10)
G2 (3=s)
respectively. Note that the expectation is still m and the skewness is still zero. But both the
variance and the kurtosis measures depend on the shape parameter s. Figure 3 shows that
these two measures decrease with respect to s.
Mean Deviations
The amount of scatter in a population is evidently measured to some extent by the totality
of deviations from the mean and median. These are known as the mean deviation about the
mean and the mean deviation about the median – defined by
ð1
d1 (X) ¼ jx E(X)j f (x)dx
1
and
ð1
d2 (X) ¼ jx Median(X)j f (x)dx
1
Figure 3. Variation of variance and kurtosis (given by equations (9) and (10), respectively) versus
s ¼ 0:1; 0:2; . . . ; 10 and s ¼ 1. The y-axis is in log scale
690 S. Nadarajah
respectively. For the generalized normal p.d.f. (equation (2)), both E(X) and Median(X)
are equal to m. Thus, the required measure is EjX mj. Note that one can write
sI
EjX mj ¼ (11)
sG(1=s)
where I denotes the integral
ð1 n x ms o
I¼ (x m) exp dx: (12)
m s
Shannon entropy defined by E½ log f (X) is the particular case of equation (14) for
g " 1. Limiting g " 1 in equation (15) and using L’Hospital’s rule, one obtains
1 s
E½ log f (X) ¼ log
s 2sG(1=s)
Song (2001) observed that the gradient of the Rényi entropy J R0 (g) ¼ (d=dg)JR (g)
is related to the loglikelihood by JR0 (1) ¼ (1=2)Var½log (f (X)). This equality and
the fact that the quantity JR0 (1) remains invariant under location and scale transform-
A Generalized Normal Distribution 691
ations motivated Song to propose 2JR0 (1) as a measure of the shape of a distribution.
From equation (15), the first derivative is
1 1 log g
JR0 (g) ¼
s g(g 1) (g 1)2
Asymptotics
If X1 ; . . . ; Xn is a random sample from equation (2) and if X ¼ (Xp1 ffiffiþ
ffi þ Xn )=npdenotes
ffiffiffiffiffiffiffiffiffiffiffiffiffiffi
the sample mean then by the usual central limit theorem n(X E(X))= Var(X)
approaches the standard normal distribution as n ! 1. Sometimes one would be inter-
ested in the asymptotics of the extreme values Mn ¼ max (X1 ; . . . ; Xn ) and
mn ¼ min (X1 ; . . . ; Xn ). Note from equations (5) and (6) that
1 t m1s n t ms o
1 F(t) exp
2G(1=s) s s
and
1 t þ m1s t þ ms
F( t) exp
2G(1=s) s s
and
F(t xg(t)) t þ xg(t) þ m 1s t þ ms t þ xg(t) þ m s
lim ¼ lim exp
t!1 F(t) t!1 tþm s s
(17)
as t ! 1. If one chooses g(t) (ss =s)tð1sÞ as t ! 1 then both the limits in equations
(16) and (17) would reduce to exp (x) as t ! 1. Hence, it follows from Theorem 1.6.2
in Leadbetter et al. (1983) that there must be norming constants an . 0, bn , cn . 0 and
dn such that
Pr {an (Mn bn ) x} ! exp {exp (x)}
692 S. Nadarajah
and
as n ! 1. The form of the norming constants can also be determined. For instance, using
Corollary 1.6.3 in Leadbetter et al. (1983), one can determine that an ¼ (s=s)( log n)1ð1=sÞ
and bn ¼ s( log n)1=s .
Estimation Issues
We consider estimation by two methods: the method of moments and the method of
maximum likelihood. Let x1 ; . . . ; xn be a random sample from equation (2). Under the
method of moments, equating E(X), Var(X) and Kurtosis(X) in equations (8), (9), (10),
respectively, with the corresponding sample estimates
1X n
s1 ¼ xi ;
n i¼1
1X n
s2 ¼ (xi s1 )2
n i¼1
and
1 X n
s3 ¼ (xi s1 )4
ns22 i¼1
@ log L n s X n
¼ þ sþ1 jxi mjs
@s s s i¼1
A Generalized Normal Distribution 693
and
X n
@ log L n 1 1 xi m xi m
¼ C þ1 log
@s s s s i¼1
s s
@2 log L s(s 1) ðx mÞs2
¼
@m2 s2 s
@2 log L s 2 ðx mÞs1
¼ sign(m x)
@m@s s s
@2 log L 1 ðx mÞs ðx mÞ
¼ 1 þ s log
@m@s ðx mÞ s s
@2 log L 1 ðx mÞs
¼ 2 1 s(s þ 1)
@s 2 s s
@2 log L 1 ðx mÞs ðx mÞ
¼ 1 þ s log
@s@s s s s
and
2
@2 log L ðx mÞs
¼
1
1 þ
2
C
1
þ
1 0 1
C logðx mÞ
@s 2 s 2 s s s 2 s s s
ð1
ðx mÞm s
E ¼ zm exp (zs )dz
s G(1=s) 0
ð1
1
¼ yðmþ1Þ=ðs1Þ exp (y)dy
G(1=s) 0
G(m þ ð1=sÞ)
¼
G(1=s)
m
ð1
ðx mÞ
E logðx mÞ ¼ s zm log z exp (zs )dz
s s G(1=s) 0
ð1
1
¼ yðmþ1Þ=ðs1Þ log y exp (y)dy
sG(1=s) 0
G(ðm þ 1Þ=s)C(ðm þ 1Þ=s)
¼
sG(1=s)
694 S. Nadarajah
and
" # ð1
ðx mÞm ðx mÞ 2 s
E
log ¼ zm ( log z)2 exp (zs )dz
s s G(1=s) 0
ð1
1
¼ 2 yðmþ1Þ=ðs1Þ ( log y)2 exp (y)dy
s G(1=s) 0
G(ðm þ 1Þ=s){C2 (ðm þ 1Þ=s) C0 (ðm þ 1Þ=s)}
¼
s2 G(1=s)
one can show that the elements of the Fisher information matrix are
2
@ log L s(s 1)G(1 ð1=sÞ)
E ¼
@m2 s2 G(1=s)
2
@ log L
E ¼0
@m@s
2
@ log L
E ¼0
@m@s
2
@ log L s
E 2
¼ 2
@s s
2
@ log L 1 1
E ¼ 1þC 1þ
@s@s ss s
and
2
@ log L 1 2 1 1 0 1
E ¼ 1 þ C þ C
@s2 s2 s s s2 s
1 1 1 1
þ G 1þ C2 1 þ C0 1 þ
s s s s
Note that like in the normal distribution, the estimates of m and s are independent.
References
Gradshteyn, I. S. & Ryzhik, I. M. (2000) Table of Integrals, Series, and Products 6th edn (San Diego: Academic
Press).
Leadbetter, M. R., Lindgren, G. & Rootzén, H. (1987) Extremes and Related Properties of Random Sequences
and Processes (New York: Springer-Verlag).
Prudnikov, A. P., Brychkov, Y. A. & Marichev, O. I. (1990) Integrals and Series Vols 1, 2 and 3 (Amsterdam:
Gordon and Breach Science Publishers).
Rényi, A. (1961) On measures of entropy and information, in: Proceedings of the 4th Berkeley Symposium on
Mathematical Statistics and Probability, Vol. I, pp. 547 –561 (Berkeley: University of California Press).
Song, K.-S. (2001) Rényi information, loglikelihood and an intrinsic distribution measure. Journal of Statistical
Planning and Inference, 93, pp. 51–69.