Stochastic Processes by Jyotiprasad Medhi PDF
Stochastic Processes by Jyotiprasad Medhi PDF
Oisknown
as negative binomial distribution with index r, and mean rqp. In particular, when r
is a positive integer, it is known as Pascal distribution.
(See Exercise 1.9 for distribution of the sum of two independent geometric variables).
Example (i). [f X, ¥ are independent r.v.'s. with generating functions A (s), B (s)
respectively, then the generating function D (s) of W = X ~ ¥ is given by
D(s) 5 E{s'} = E{s (57) S EMS EL(15)'},
because of independence of X, ¥; thus Pr {X — ¥ =A} is the coefficient of s', k= 0,
41,42... in the expansion of
Ds) = ABU).Probability Distributions 11
1.1.4 Sum of a Random Number of Discrete Random Variables
In Section 1.1.3 we considered the sum S, =X, + --- +X, of a fixed number of
mutually independent random variables. Sometimes we come across situations when
we have to consider the sum of arandom number N of random variables. For example,
if X, denotes the number of persons involved in the ith accident (in a day in a certain
city) and if the number N of accidents happening on a day is a random variable, then
the sum Sy =X, +++ +Xj,denotes the total number of persons involved in accidents
ona day.
‘We have the following theorem for the sum Sy.
.-, be identically and independently distributed random
and p.gf.
P(s) = L.p,s* for =1,2,... (uy
Sy aX, +...4X yn (1.12)
where N is a random variable independent of the X;’s. Let the distribution of W be
given by Pr {N =n} =, and the p.gf. of Nbe
G(s)=Eg,5". (1.13)
Then the p.g.f. H (5) of Sy is given by the compound function G (P (3)), ie.
H(s)= E Pr {Sy = j}s'=GP(s)). (14)
Proof: Since N is a random variable that can assume values 0, 1, 2, 3,..., the event
Sy =jcan happen in the following mutually exclusive ways: N =n and S,=X,+...
+X, =j fora = 1,2, 3,... To meet the situation N = 0 let us define X_ = Sy
that
XgtX,+..4Xy=Sy, 220.
Wehave hy=Pr(Sy=i) = 5 PrN =aandS,=/}.
Since N is independent of X;"s and therefore of S,,
E Priv =n} PriS,=/) (1.15)12 Stochastic Processes
(1.16)
(The inclusion of the value 0 for m can be justified since Pr (S, = 0) = 1 and
Pr (S,>0}=0.)
‘The p.g.f. of the sum Sy =X, +--- +X, is thus given by
H(s)= E hs! =D Pe {Sy = hs! (1.17)
pheat
=E[[Eerc.=al]e a8)
= x[zrs, =is| &
=ZalPo)"=GP),
since the expression {P (s)}" is the p.g.f. of the sum S, of a fixed number 1 of iid.
random variables (Theorem 1.2). a
As P (s) and G (3) are p.g.f.'s, so also is G (P (s)); G (P (s)) = 1 fors = 1.
“Since G (P (s)) is a compound fuuction, the corresponding distribution is known as
compound distribution. It is atso known as random sum distribution.
Mean and Variance of Sy
We have
Ey) = HS) |= (PGP M2
= PG (P(1) =P()G")
so that
E{Sy} = E{X}E{N}. (1.19)
Again,
HS) =P(S)G (Ps) +G"PE PCF
Thus
E(SR)-E(Sy) =H) Les
= [E(K}) - EXE)
HEW)-EN MEXProbability Distributions 13
Using (1.19) we get
EXSp) = E(X)E(N) — E(N) (EX)?
FEN YER
= E(N) var (X,) + EN?) EX)?
Finally,
var (Sy) = E(S)-[E(Sy)]?
= E(N var (X)+ EK) var (N). (1.20)
Note: The result (1.20) can also be obtained by using the relation
var (X) =E [var (X | Y)] + var (E(X | ¥)) (a)
and taking X = Sy and Y = N, (See Sec. 1.1.5 for conditional expectation).
We have, E(Sy|N=n)=n£(X,) with probability Pr(V =n) =,
$0 that
E(SpIN)= Ene Xe,
= E(X)E()
and
[EGyINI? = LHEHNs,
HEX SEW").
Thus
var (Sy 1) = [EX var (W). @)
Again
var (Sy [M = 2) = nvar(X,) with probability g,14 Stochastic Processes
so that :
Elvar (Sy |N)] = Z nvar(X,)g, = [var X%)] E(N). (c)
Using (b) and (c) in (a) we get (1.20).
Example 1(j). Compound Poisson distribution: If N has a Poisson distribution with
mean A, then G (s) = Eg,s"= exp (A (s— 1)} and hence the sum X, +... +Xy has
pet.
H(s)= GP(s)) =a, [POs " =exp(AP (s)— II}.
‘The distribution having a generating function of the form exp {A [P (s)— 1]}, where
P (s) is itself a generating function, is called compound Poisson distribution.
The mean is given by
E{Sy} = EX} EIN} =EX,).
Here g, = e*.'/a!. Similarly, taking g, = (1 —p) p", one gets compound geometric
distribution, with H (s)=p P (s)/[1-q P(s)].
1.1.5 Generating Function of Bivariate Distribution
Bivariate Distribution: Suppose that X, Y is a pair of integral-valued random
variables with joint probability distribution given by
Pr{X=j,¥ sk} spy, j,k =0,1,2,..., Epa=l. (1.21)
‘The marginal distributions are given by
Pr(X=A=Z perf, J20,1,2,... (122)
Priv=k)=E pees, £2012... (1.23)
B
If (, b) is any point at which f, #0 (i.e. fj > 0), then the conditional distribution of Y
given X =/is given by
Pr{Y
Pr{Y =k 1X =j} = (1.24)Probability Distributions 15
Conditional Expectation
‘The conditional expectation E {¥ 1X =} is given by
EY |X = J) = LkPr{Y =k IX =/}
EkPp
5 + £20,122. (125)
ForX given but unspecified, we note that E {¥ 1X} isa random variable that assumes
the value (Ek palf] with Pr (X =} =f)>0.
Hence the expectation of the random variable E {¥ 1X} is given by
ELEY 1X4) = EE(V IX =f} PrX = J}
Lkpp
{Eh
LG
=LEkPy
ee
=Ek {ze} = LAP =K}= EY). (1.26)
7
In the same way, we can prove that
E(E{Y? 1X} = ElY"]s (1.27)
more generally, for any function (Y) whose expectation exists,
E(E{9(Y) |X}) = Elo (Y)). (1.28)
Note (1). The results (1.26—-1.28) which are given here for discrete random variables
X, ¥ will also hold, mutatis mutandis, for continuous random variables.
Note (2). The result (1.26) holds for all r.v.'s provide E (¥) exists.
Enis (Biometrika (1973) 432) cites an example where E [E (¥ 1X)] =0 but E (¥) does
not exist and consequently (1.26) does not hold.
Bivariate Probability Generating Function
Definition. The probability generating function (bivariate p.g.f.) of a pair of random
variables X, Y with joint distribution given by (1.21) is defined by
POs. si)= ¥ pysish, 1.29)16 Stochastic Processes
Siy 8 being dummy positive real variables chosen so as to make the double series
convergent. We have the following results,
Theorem 1.4. (a) The p.g.f.A (s) of the marginal distribution
of X is given by A (s) = P (s, 1).
(b) The pg-f. B (s) of ¥ is given by
B(s) =P (1,5).
(6) The pact of (% + ¥) is given by P (s,)-
Proof: Since the convergent double series (1.29) consists of positive real terms, the
change of order of summation is justified.
(a) We have from (1.29)
Ps, I= z Pas! = Edea
-¥ psi=, z Pr{X = j}s/, from (1.22)
=A (s) (the pg fofX).
(b) It can be proved in the same way.
(©) We have
<= jt = F J
POsss)= E pas = & & Pens
as can be easily verified.
Now
Pr{X +¥ =m =i (X=,¥em-h =Zn..
Hence.
Pos)= % (PAX +¥ =m) Js",
which shows that P (s, s) is the p.gf. of X +¥. a
Remarks: (1) If X, ¥ are independent,
Dy =PHAX =j,¥ =k} =Pr{X =} Pri =k}
and then
P55) =A(S)BS) =P, PLS)Probability Distributions 17
and conversely, from this relation follows the independence of X and Y.
(2) The probabilities py can be (uniquely) determined from P (s,, 53) as follows:
a re]
Pa TW, elas,
Example 1(k). Consider a series of Bemoulli trials with probability of success p.
Suppose that X denotes the number of failures preceding the first success and ¥ the
number of failures following the first success and preceding the second success. The
sum (X + ¥) gives the number of failures preceding the second success.
The joint distribution of X, Y is given by
Dg =Pr{X =),Y sk} =qi'tp?, j,k $0,1,2,.005
and the bivariate generating function is given by
PO5,)= E. pysisl = Eq! **prsist
tee it
-1 fe Stu
?
“U=sa-59)
‘The p.g-f. of X is given by
p
AG)=PS =r ag) T-9s .
‘The p.g.f. of X + Y is given by
2
rens=(-25) ’
(See Example 1¢h)).
1.2 LAPLACE TRANSFORMS
1.2.1. Introduction
Laplace transform isa generalization of generating function. Laplace transforms
serve as very powerful tools in many situations. They provide an effective means for
the solution of many problems arising in our study. For example, the transforms are
very effective for solving linear differential equations. The Laplace transformation.18 Stochastic Processes
reduces a linear differential equation to an algebraic equation, In the study of some
probability distributions, the method could be used with great advantage, for it
happens quite often that it is easier to find the Laplace transform of a probability
distribution rather than the distribution itself.
Definition. Let f (f) be a function of a positive real variable ¢. Then the Laplace
transform (LT) of f (0) is defined by
Far= femesnytnd Qa)
4
for the range of values of s for which the integral exists.
‘We shall write f (s) = L (f (0) to denote the Laplace transform of f (1).
Example 2(a).
@_TEf(e) = c(const,), Tee) [ exptstbedt= cts (>0)
2
Hpin=t, Fis= frexpeande= us? 6 >0)
:
Hepa, Fisd= [tempest Tint tis" (o> 0)
4
Though f (0 >» >— 1) is infinite at ¢ = 0, the result holds for n > — 1. In particular,
itfo=r", then, Fos) -r(3) Ie =F :
GD) Letf()=e%,then — fis)= ferpcmespanse
°
=Ms-a) (s>a)
Gi) Let f(¢) = sine, then fs) = fexpcansin tat
°
=1(s'+1) (s>0)
(i) Letfe)=e"F,then—7ts)= | exptcstexpl-ayat
3Probability Distributions 19
= fexpi-ne +1} Adar
= Ta+))
G+iyt!
(a >-1).
Example 2(b). Dirac-delta (or Impulse) function located at @ is defined as
&t-a)=1, t=a
=0, r¥a
The L.T. of 8 (t~ a) equals
feree -a)dt=e™.
Unit step function (at a) is defined as
The L-T. of u, (t) ise™/s.
1.2.2 ‘Some Important Properties of Laplace Transforms : see Appendix AL
1.2.3 Inverse Laplace Transform
Definition. If (s) is the L-T. of (0) ic. L (f(@) =F (s), thenf (i)iscalled the inverse
Laplace transform f(s). For example, the inverse Laplace transform of f (3) = 1/5" is
(0 51. There is an inversion formula which gives f (0 in terms of f (s); if f (s) exists,
then f (0) can be uniquely determined subject to certain conditions satisfied by f().
In particular, if two continuous functions have the same transform, they are identical.
Extensive tables of Laplace transforms are also available and reference may be
made to them, whenever necessary, to obtain either the L.TT. of a function f (f) or the
inverse L.T. of a function f (s). Techniques for numerical inversion of Laplace
transforms are discussed in Bellman et al. (1966).
13 LAPLACE (STIELTJES) TRANSFORM OF A PROBABILITY DIS-
‘TRIBUTION OR OF A RANDOM VARIABLE
13.1 Definition
Let X be a non-negative random variable with distribution function
F(x)=Pr{X Sx}.20 ‘Stochastic Processes
The Laplace (Laplace-Stieltjes) transform F* (3) of this distribution is defined, for
5&0, by
F*s)= fewesnaren G.I)
°
We shall say that (3.1) also gives the ‘Laplace (Laplace-Stieltjes) transform of
the random variable X””,
We have
Fs) = Efexp(-sX)} Gla)
and
F*Q)=1. (3.2)
Suppose that X is a continuous variate having density f (x) = F’ (x). Then form (3.1),
Fats)= f expsriftelde 63)
4
(this is the “ordinary” Laplace transform L {f (x)} = 7(s) of the density function
f@).
By a Laplace transform of a r.v.X, we shall mean the L.S.T. of the distribution
function F (-) of X; this is equal to the ordinary L.T. of the density function of X when
this exists. We have
Fe (s)= f(s).
Incase X is an integral-valued random variable with distribution p, = Pr (X=),
k=0,1,2,... and pig. P (s) =Eas', we can stretch the language and define the
LT. of X by
F*(s) =E {exp(-sX_ I = Pf{exp(-s)}. G.4)
‘Thus in case of a discrete random variable assuming non-negative values 0, 1.2,
3,....the L.-T. of the variable differs from its p.g.f. only by a change of variable:
exp (— 5) in the former replaces s in the latter. Thus there is a close analogy between
the properties of Laplace transforms and those of generating functions ors-transforms.
Note. When it exists, the functionProbability Distributions 21
is called the moment generating function of the rv. X (it is the generating function
of the sequence j2,/n!, where 1, = £ {X"} is the nth central moment of X),
‘The function
$0) = Efe}
defined for every r.v. X is called the characteristic function of X.
When X is non-negative and integral valued, then its'p.g.f. is considered; so is
its L.-T. when X is non-negative and continuous as these are easier to handle. The
characteristic function, defined for all r.v.'s, is a more universal tool.
1.3.2. The Laplace Transform of the Distribution Function in Terms of that of
the Density Function.
Let X be a continuous (and non-negative) r.v. having density function f (x) and
distribution function
Pr{X $x} =F(x)= [nae
3
‘The (ordinary) Laplace transform of the distribution function F (x) is
LF} = f exp-snF eee
‘We have from A.6 (Appendix)
Fos) =L{F eo) = fexpcs) {fneha
: t
= ffs} =fisys.
‘The relation can also be obtained by integrating by parts the relation (3.1). Thus
we get
Fs) =f(s)=sF(s). G5)
1.3.3 Mean and Variance in Terms of (Derivatives of) L.T.
‘We note here that differentiation under the integral sign is valid for the L.T. given
by (3.1), since the integrand is bounded and continuous.
Differentiating (3.1) with respect to s, we get22 Stochastic Processes
a -
rms) = f xexp(-sr)dF tr)
# rary acy fa?
war's) =(-ly fe exp(—sx dF (x)
and, in general, for n= 1,2,...
Sr) en" jr exp(-sx)dF (x). G6)
The differentiation under the integral is valid since the new integrands are con-
tinuous and bounded.
We can use the above relation to find £ (X"), the ath moment of X when it exists;
we have
C1]
F “9 = ftexpcso}s'arey fors =0
eo 4
= frdray=£0%, n=I,2,...
°
when the Lh. s, exists, i.c. the rv, X possesses a finite a th momentiflim < SF SIiao
exists.
We have
eare-[Lrre)] G7)
ds 0 "
EQ)= -[Sr «| G8)
sno
d 2
and var (X) = [fro] [fereo} I G9)
=o 100.
1.3.4 Some Important Distributions
1.3.4.1 A special kind of discrete distribution:
Suppose that X is a random variable whose whole mass is concentrated in one
single-point, say, point a. This implies that the variable is ‘almost always’ equal to
a, ie. Pr (X= @) = 1 and Pr (X #a) = 0, Its distribution function is given byProbability Distributions 23
F(x)=Pr{X $x} =0 forx0, OSx . A simple
version of CLT for equal components is as follows.
IfX,....,X,,...arei.id. random variables, with F(X) =u, var 0) = 7 (both
finite) and S, =X, +... +X, then asin — e,
m{ Heal 010
where © (:)is the d.f. of N (0, 1).Probability Distributions 37
13.5 Three Important Theorems
We state, without proof, three theorem conceming the L.T. of probability dis-
tributions
Theorem 1.5.. Uniqueness Theorem: Distinct probability distributions on [0,°) have
distinct Laplace transforms.
Ie has the important implication that a probability distribution is recognizable
by its transform, If the L,T, of a random variable X is known, then by identifying this
with the form of L.T. of a distribution, one can conclude that X has the corresponding
distribution; if this form does not resemble with the form of a L.T. of any standard
distribution one can proceed to find the inverse of the L.T. (for which numerical
methods have also been developed) to get the distribution. Even without finding the
inverse, i.e. the form of the distribution, one can compute the moments of the dis-
tribution. Because of this theorem, perhaps, L.T. has the role it now plays in the study
of probability distribmions.
Theorem 1.6. Continuity Theorem: Let (X,),n= 1,2, ... bea sequence of random
variables with distribution functions {F,} and L.T."s (F,* (s)}.1fas n 0, F, tends
to a distribution function F having transform F* (s), then as m —> 09, F,* (s) > F* (s)
for s > 0, and conversely.
This theorem can be used to obtain the limit distribution of a sequence of random
variables.
Theorem 1.7. Convolution Thearem: The Laplace transform of the convolution of
two independent random variables X, Y is the product of their transforms.
‘The integral
fs0-vvord (3.222)
(denoted by f* g) is called the convolution of the two functions f and g. The con-
volution U of F and G is given by
vere [eu-ydFo) 226)
3
and L{UG)} =L{GOHLIF OQ). (3.22¢)
In case of discrete random variables the result is essentially the same as that statec
in (1.10) for generating function of the convolution of two random variables. In case
of continuous random variables we have now the analogous result.
The above result is equivalent to the assertion that if X, Y are two independent
random variables, then38 Stochastic Processes
Efe"™ "P= E fexp(-sX exp(-s¥)}
= E{exp(-sX)} Efexp(-s¥)). 23)
(Note that the converse is not true).
‘The result can also be stated as:
Theorem 1.8. The L.-T. of the sum of two independent variables is the product of
L.'s of the variables.
Wee can use the result to find the distribution of the sum of two or more inde-
pendent random variables, when the variables are continuous ordiscrete. The method
of generating functions is applicable only when the variables are. non-negative
integral-valued.
The following result immediately follows:
Theorem 1.9. The sum S, = X, +. . . + X, of m (a fixed number) identically and
independently distributed random variables X; has the L.T. equal to [F* (s)]", F* (s)
being the L-T. of X.
As an application we consider the following result:
Theorem 1,10. The sum of k identical and independent negative exponential dis-
tributions with parameters 4 follows gamma distribution with parameters A, k.
Let Xj,G=1,2,...1,2,...4, have densities
Six) = Lexp(-Ax),x 20 forall i.
From Example 3(c), the L.T. of X; is
Fy4(s)=M(s +A)
and so the L.T. of S=X, +Xz+...+X,isQ/s+Ayj*.
But this is the L.T. of the gamma distribution with density f,, (x) given in (3.15)
(Example 3(¢)). Hence the theorem.
‘Wecan easily obtain the mean and variance of the gamma distributionas follows:
E(S) = E(SX,) = ZE(XK)=kA
var (5) = Zvar (X,) =k/A! (see (3.17) and (3.18).
Limiting form of Erlang—k distribution: As k — ©, Erlang-k distribution tends to a
discrete distribution concentrated at 1/A. Let {£,) be a sequence of r.v."s, E, having
Erlang—& distribution with density
k)tx*"Sexp(-kAx)}
TK) &=1,2,3,...
Fa s0=Probability Distributions 39
Is L.T. is given by F,*(s) (245) (i +
As k 900, F,* (5) exp(-s/A). (3.24)
Butexp (—s/A) is the L.T. of the discrete distribution concentrated ata= 1/4 (Example
3(a)). Therefore, from the continuity theorem it follows that, as ke, the distribution
of E, tends to that variable whose whole mass is concentrated at 1/A.
This can also be seen from the fact that, as k =9 ©», the mode (k = 1)/KA of E,
moves to the right to 1/A and that as var (E,) = 1/kA? —> 0, the whole mass of E, tends
to concentrate at the single point 1/A.
‘We thus have exponential distribution for k = 1 and the degenerate distribution
(deterministic case) fork —9 2. A suitable value of k (1 0,1, >Oand0
You might also like