Brownian Motion & Stochastic Integrals
Brownian Motion & Stochastic Integrals
N.J. Nielsen
Introduction
These notes cover a series of lectures given at the University of Kiel in May 2011 in connection
with an Erasmus project and is based on a regular course in stochastic differential equations
given by me at the University of Southern Denmark. Some additional material which there was
no time to cover in the lectures is included at the end. The notes follow the lectures quite closely
since the source file is a slight modification of the file used for the lectures.
The construction of Brownian motion using the Haar system was originally carried out by P.
Lévy in 1948 [2] and Z. Ciesielski in 1961 [1].
General results from functional analysis and probability theory used in the notes can be found in
standard textbooks in these areas of mathematics.
1 Brownian motion
In the following we let (Ω, F, P ) be a fixed probability space.
We start with the following definition:
Definition 1.1 A stochastic process in continuous time is a family (Xt )t≥0 of real random vari-
ables defined on (Ω, F, P ).
Given a stochastic process (Xt )t≥0 we often only consider Xt for t in an interval [0, R].
We shall also need the following definitions:
Definition 1.2 Let(Ft )t≥0 be a family of sub-σ-algebras of F so that Fs ⊆ Ft for all s ≤ t. A
stochastic process (Xt )t≥0 is called adapted if Xt er Ft -measurable for every t ≥ 0.
The following definition is important.
1
Definition 1.3 Let (Ft ) be as in Definition 1.2 and let (Xt ) ⊆ L1 (P ) be an (Ft )-adapted pro-
cess. (Xt ) is called a submartingale if
Xs ≤ E(Xt |Fs ) for all s < t. (1.1)
If there for all s < t is equality in (1.1), then (Xt ) is called a martingale. (Xt ) is said to be a
supermartingale if (−Xt ) is a submartingale.
A process (Xt ) on (Ω, F, P ) is called continuous if the function t → Xt (ω) is continuous for a.a
ω ∈ Ω.
A process (Yt ) is said to have a continuous version if there exists a continuous process (Xt )
so that P (Xt = Yt ) = 1 for all t ≥ 0. If (Xt ) is a process on (Ω, F, P ), then the functions
t → Xt (ω), ω ∈ Ω are called the paths of the process.
Now it is the time to define the Brownian motion.
Definition 1.4 A real stochastic process (Bt ) is called a Brownian motion starting at 0 with
mean value ξ and variance σ 2 if the following conditions are satisfied:
(i) P (B0 = 0) = 1
(ii) Bt − Bs is normally distributed N ((t − s)ξ, (t − s)σ 2 ) for all 0 ≤ s < t.
(iii) Bt1 , Bt2 − Bt1 , . . . Btn − Btn−1 are (stochastically) independent for alle 0 ≤ t1 < t2 <
t3 < · · · tn .
(Bt ) is called a normalized Brownian motion if ξ = 0 and σ 2 = 1.
The essential task of this section is of course to prove the existence of the Brownian motion, i.e.
we have to show that there exists a probability space (Ω, F, P ) and a process (Bt ) on that space
so that the conditions in Definition 1.4 are satisfied. It is of course enough to show the existence
of a normalized Brownian motion (Bt ) for then (ξt + σBt ) is a Brownian motion with mean
value ξ and variance σ 2 . We shall actually show a stronger result, namely that the Brownian
motion has a continuous version. When we in the following talk about a Brownian motion we
will always mean a normalized Brownian motion unless otherwise stated.
We will use Hilbert space theory for the construction so let us recall some of its basic facts.
In the following (·, ·), respectively k · k will denote the inner product, respectively the norm in
an arbitrary Hilbert space H. If we consider several different Hilbert spaces at the same time it
is of course a slight misuse of notation to use the same symbols for the inner products and norms
in these spaces but it is customary and eases the notation.
Let us recall the following elementary theorem from Hilbert space theory:
Theorem 1.5 Let H1 be a Hilbert space with an orthonormal basis (en ) and let (fn ) be an
orthonormal sequence in a Hilbert space H2 . Then the map T : H1 → H2 defined by
X∞
Tx = (x, en )fn for all x ∈ H1 (1.2)
n=1
2
is an isometry of H1 into H2 .
In the following we let (gn ) be a sequence of independent standard Gaussian variables on a
probability space (Ω, F, P ). Actually we can put (Ω, F, P ) = ([0, 1], B, m) where m denotes
the Lebesgue measure on [0, 1].
For the matter of convenience we shall in the sequel consider a constant k as normally distributed
with mean k and variance 0.
We can now prove:
Theorem 1.6 Put H = span(gn ) ⊆ L2 (P ). If T : L2 (0, ∞) → H is an arbitrary isometry and
we put
Bt = T (1[0,t] ) for all t ∈ [0, ∞[, (1.3)
then (Bt ) is a Brownian motion.
Proof: Note that such isometries exist. Indeed, since L2 (0, ∞) is a separable Hilbert space, it
has an orthonormal basis (fn ) and we can e.g. define T by T fn = gn for all n ∈ N.
Let us define (Bt ) by (1.3). Since 0 = T (0) = B0 , it is clear that (i) holds. Next let 0 ≤ s < t.
Since Bt − Bs ∈ H, it is normally distributed with mean value 0 and furthermore we have:
Z
(Bt − Bs )2 dP = kBt − Bs k22 = kT (1]s,t] )k22 = k1]s,t] k22 = (t − s), (1.4)
Ω
The following corollary put our results so far together but gives also new information.
Corollary 1.7 If (fn ) is an arbitrary orthonormal basis for L2 (0, ∞), then the series
∞ Z
X t
Bt = fn (s)ds gn t≥0 (1.5)
n=1 0
converges in L2 (P ) and almost surely for all t ≥ 0. (Bt ) is a Brownian motion on (Ω, F, P ).
Proof: If we define the isometry T : L2 (0, ∞) → L2 (P ) by T fn = gn , it is given by
∞ Z
X ∞
Tf = f (s)fn (s)ds gn , (1.6)
n=1 0
3
where the series converges in L2 (P ). It follows from the above that Bt = T (1[0,t] ) is a Brownian
motion and equation (1.6) gives that
∞ Z
X t
Bt = T (1[0,t] ) = fn (s)ds gn for alle t ≥ 0. (1.7)
n=1 0
Since the terms in this sum are independent, have mean value 0, and
∞
X Z t
E( fn (s)ds gn )2 = kBt k22 = t < ∞, (1.8)
n=1 0
it follows from classical results in probability theory that the series (1.7) converges almost surely
for every t ≥ 0. 2
We shall now prove that there is a continuous version of the Brownian motion and then we do
not as so far have a free choice of the orthonormal basis (fn ) for L2 (0, ∞). We construct an
orthonormal basis (fn ) with the property that there is an A ∈ F with P (A) = 1 so that if ω ∈ A,
then the series in (1.5) converges to Bt (ω) uniformly in t on every compact subinterval of [0, ∞[.
Since every term of the series is continuous in t, this will give that t → Bt (ω) is continuous
for all ω ∈ A. The construction of (fn ) is based on the Haar system (an orthonormal basis for
L2 (0, 1) explained below) with the aid of the Borel-Cantelli lemma.
In the following we let (h̃m ) denote the (non-normalized) be the Haar system, defined as follows
(make a picture!!):
h̃1 (t) = 1 for alle t ∈ [0, 1]. (1.9)
For all k = 0, 1, 2, . . . og ` = 1, 2, . . . , 2k we put
1 if t ∈ [(2` − 2)2−k−1 , (2` − 1)2−k−1 [
h̃2k +` (t) = −1 if t ∈ [(2` − 1)2−k−1 , 2` · 2−k−1 [
0 else.
By direct computation we check that it is an orthonormal system and since it is easy to see that
every indicator function of a dyadic interval belongs to span(hm ), it follows that span(hm ) is
dense in L2 (0, 1). Therefore (hm ) is an orthonormal basis for L2 (0, 1). It follows from Theorem
1.7 that ∞ Z t
X
Bt = hm (s)ds gm 0≤t≤1 (1.11)
m=1 0
is a Brownian motion for t ∈ [0, 1]. The series converges in L2 (P ) and almost surely and the
same is the case if we permute the terms. We should however note that the set with measure
4
1 on which the series converges pointwise depends on the permutation. In order not to get
into difficulties with zero sets we shall fix the order of the terms in the sum. We define for all
0≤t≤1
∞ k+1
2X
Z t Z t X Z t
def
X
Bt = h1 (s)ds g1 + hm (s)ds gm = ∗ hm (s)ds gm (1.12)
0 k=0 m=2k +1 0 m 0
There is a subset Ω̃ ⊆ Ω with P (Ω̃) = 1 so that there to every ω ∈ Ω̃ exists a k(ω) with the
property that Gk (ω) ≤ k for all k ≥ k(ω).
Proof: For every x > 0 we find
r Z ∞ r Z ∞ r
2 −u2 /2 2 u −u2 /2 2 −1 −x2 /2
P (|gm | > x} = e du ≤ e du = x e , (1.14)
π x π x x π
which gives:
k +1
2[ r
2 1 k −k2 /2
P (Gk > k) = P ( (|gm | > k) ≤ 2k P (|g1 | > k) ≤ ·2 e . (1.15)
πk
m=2k +1
Since ∞ r ∞
X 2 X −1 k −k2 /2
P (Gk > k) ≤ k 2 e < ∞,
k=1
π k=1
it follows from the Borel-Cantelli lemma that P (Gk ≤ k from a certain step) = 1. Choosing Ω̃
as this set the statement follows. 2
5
Proof of Theorem 1.8: Let Ω̃ be as in Lemma 1.10 and let ω ∈ Ω̃. Then there exists a k(ω) ≥ 1
so that Gk (ω) ≤ k for alle k ≥ k(ω). If k ≥ k(ω) is now fixed, we find
k+1
2X k+1
2X
Z t Z t
| hm (s)ds · gm (ω)| ≤ hm (s)ds · Gk (ω) ≤ k 2−k/2−1 . (1.16)
0 0
m=2k +1 m=2k +1
for all 0 ≤ t ≤ 1.
Since ∞ −k/2−1
P
k=1 k 2 < ∞, it follows from Weierstrass’ M-test that the series
P∞ P2k+1 Rt
k=k(ω) m=2k +1 0 hm (s)dsgm (ω) converges uniformly for t ∈ [0, 1]. This gives that the series
X Z t
Bt (ω) = ∗ hm (s)dsgm (ω) (1.17)
m 0
In order to find a continuous Brownian motion on [0, ∞[ we define the functions hnm ∈ L2 (0+, ∞)
by
hm (t − n) for t ∈ [n − 1, n]
hnm (t) = n ∈ N, m ∈ N (1.18)
0 else
and note that (hnm )∞
m=1 is an orthonormal basis for L2 (n − 1, n) for all n ∈ N which implies that
(hnm ) is an orthonormal basis for L2 (0, ∞).
The following theorem easily follows from the above:
Theorem 1.11 Let (Ω, F, P ) be a probability space on which there exists a sequence of N (0, 1)-
distributed random variables and let (gnm ) be such a sequence. Define:
∞ X Z t
X
Bt = ∗ hnm (s)ds gnm for allt ≥ 0. (1.19)
n=1 m 0
6
2 The Ito integral
In this section we let (Ω, F, P ) denote a probability space on which we have a Brownian motion
(Bt ) and we shall always assume that it is continuous. Further, B denotes the set of Borel subsets
of R, mn denotes the Lebesgue measure on Rn (m1 = m) and (Ft ) denotes the family of σ–
algebras defined above. Also we let 0 < T < ∞ be a fixed real number. We R T wish to determine a
subspace of functions f of L2 ([0, T ] × Ω.m ⊗ P ) so that we can define 0 f dB as a stochastic
variable. Since it can be proved that for a.a. ω ∈ Ω the function ω → Bt (ω) is not of finite
variation, the Riemann–Stiltjes construction will not work. However, since (Bt ) is a martingale,
we have other means which we are now going to explore.
For every n ∈ N we define the sequence (tnk ) by
−n
n k2 if 0 ≤ k2−n ≤ T
tk =
T if k2−n > T
If n is fixed we shall often write tk instead of tnk .
We let E ⊆ L2 ([0, T ] × Ω, m ⊗ P ) consist of all functions φ of the form
X
φ(t, ω) = ek (ω)1[tnk ,tnk+1 [ (t)
k
where n ∈ N and every ek ∈ L2 (P ) and is Ftnk –measurable. The elements of E are called
elementary functions.
If φ ∈ E is of the form above we define the Ito integral by:
Z T X
φdB = ek (Btk+1 − Btk )
0 k
RT
It is straightforward that the map φ → 0
φdB is linear.
The following theorem is called the Ito isometry for elementary functions.
Theorem 2.1 If φ ∈ E, then
Z T Z T
2
E( φdB) = E( φ2 dm).
0 0
Proof: Let φ be written as above. If j < k, then ej ek (Btj+1 − Btj ) is Ftk –measurable and
therefore independent of (Btk+1 − Btk ). Hence
E(ej ek (Btj+1 − Btj )(Btk+1 − Btk )) = E(ej ek (Btj+1 − Btj )E(Btk+1 − Btk ) = 0.
7
This clearly gives that Z T Z T
2
E( φdB) = E( φ2 dm)
0 0
2
RT
This means that the map φ → 0 φdB is a linear isometry from E → L2 (P ) and can therefore
RT
be extended to a linear isometry from E to L2 (P ). Hence we can define 0 φdB for all f ∈ E
RT
and it is clear that E( 0 φdB) = 0.
RT
Theorem 2.2 If f ∈ E, then ( 0 f dB)0≤t≤T is a martingale.
Proof: Let first φ ∈ E be written as above. Then
Z T X
φdB = ek (Btk+1 − Btk )
0 k
If k < m
E(ek (Btk+1 − Btk ) | Ft ) = ek (Btk+1 − Btk ).
Finally we get:
E(em (Btm+1 − Btm ) | Ft ) = em (Bt − Btm ).
RT Rt
Summing up we get that E( 0 φdB | Ft ) = 0 φdB. Since E(· | Ft ) is an orthogonal projection
RT Rt
on L2 (P ), it follows that E( 0 f dB | Ft ) = 0 f dB for all f ∈ E. 2
Using Doob’s martingale inequality and the Borel–Cantelli lemma, the following can be proved:
Theorem 2.3 The Ito integral
R t has a continuous version, meaning that we can achieve that for
every f ∈ E that map t → 0 f dB is continuous almost surely.
Our next task is to determine E. A B ⊗ F–measurable function f is called progressively measur-
able if for all 0 ≤ t ≤ T f : [0, t] × Ω → R is B ⊗ Ft –measurable. We let P2 (0, T ) denote the
closed subspace of L2 ([0, T ] × Ω, m ⊗ P ) which are progressively measurable.
We can now prove:
Theorem 2.4 E = P2 (0, T ).
8
Proof: Let first g ∈ P2 (0, T ) be bounded so that g(·, ω) is continuous for almost all ω ∈ Ω and
define φn by X
φn (t, ω) = g(tnk , ω)1[tnk ,tnk+1 [ (t)
k
Clearly φn ∈ E for all n ∈ N. Let now ε > 0 and let ω ∈ Ω be fixed. By uniform continuity we
can find a δ > 0 so that |t − s| < δ ⇒ |g(t, ω) − g(s, ω)| < ε. Determine n0 so that 2−n0 < δ
and let n ≥ n0 . Then |g(t, ω) − g(tk , ω)| < ε for all tk ≤ t ≤ tk+1 and therefore
Z T
(g(t, ω) − φn (t, ω))2 dt < ε2 T
0
RT
so that 0 (g(t, ω) − φn (t, ω))2 dt → 0. Since g is bounded, majorized convergence gives that
RT
also E( 0 (g − φn )2 dm) → 0 as well. Hence g ∈ E.
The next step is the tricky one where progressive measurability is used. Let h ∈ P2 (0, T ) be a
bounded function, say |h| ≤ M a.s. We wish to show that there is a sequence (gn ) ⊆ P2 (0, T )
so that for evey n and a.a. ω the function t → gn (t, ω) is continuous and so that gn → h in
L2 (m ⊗ P ). Together with the above this will give that h ∈ E. Let for each n ψn be the non–
Rnegative continuous function which is zero on the intervals ] − ∞, − n1 ] and ]0, ∞[ and so that
∞
−∞ n
ψ (x)dx = 1. We can e.q choose ψn so that its graph is a triangle. (gn ) is now defined by:
Z t
gn (t, ω) = ψn (s − t)h(s, ω)ds for all ω and all 0 ≤ t ≤ T .
0
The properties of the sequence (ψn ) readily give that each gn is continuous in t and |gn | ≤ M
a.s. For fixed t the function (s, u, ω) → ψn (s − u)h(s, ω) is integrable over [0, t] × [0, t] × Ω and
since h ∈ P2 (0, T ), it is B ⊗ B ⊗ Ft –measurable. An application of Fubini’s theorem now gives
that gn is progressively measurable for every n.
Since (ψn ) constitutes an approximative identity with respect to convolution, it follows that
RT
0
(h − gn )2 dm → 0 and an application of majorized convergence gives gn → h in L2 (m ⊗ P ).
Let now f ∈ P2 (0, T ) be arbitrary. For every n ∈ N we define
−n if f (t, ω) < −n
hn (t, ω) = f (t, ω) if −n ≤ f (t, ω) ≤ n
n if f (t, ω) > n
Rt
It is worthwhile to note that if h ∈ L2 ([0, T ]), then 0
hdB is normally distributed with mean 0
RT
and variance 0 h2 dm.
We say that a measurable function f ; [0, T × Ω → R is adapted to the filtration(Ft ), if for fixed
0 ≤ t ≤ T the function ω → f (t, ω) is Ft –measurable. A lengthy and quite demanding proof
gives the following result.
9
Theorem 2.5 (Meyer) If f is adapted to a filtration (Ft ), then it has a progressively measurable
modification g, that is for every 0 ≤ t ≤ T f (t, ·) = g(t, ·) a.s.
We let A2 (0, T ) consist of those functions in L2 (m ⊗ P ) which are adapted to (Ft ). Using
Meyer’s theorem we can now define the Ito integral of an Rf ∈ A2 . WeRchoose namly a progres-
T T
sively measurable modification g of f and simply define 0 f dB = 0 gdB. We shall not go
into details.
The Ito integral can be defined for a larger class of integrants. If f ∈ L2 ([0, T ] × Ω) so that there
is an increasing family (Ht ) of σ–algebras so that
(i) Ft ⊆ Ht for all 0 ≤ t ≤ T .
(ii) For all 0 ≤ s < t ≤ T Bt − Bs is independent of Hs .
(iii) f is (Ht )–adapted.
The arguments are similar to the ones Rgiven above. Note that (ii) implies that (Bt ) is a martingale
t
with respect to (Ht ). It also follows ( 0 f dB) is a martingale.
Let f : [0, T ] × Ω → R be a function satisfying (i)–(iii) and so that
Z T
P ({ω ∈ Ω | f (t, ω)2 dt < ∞}) = 1. (2.1)
0
In that case it can be proved that there is a sequence (fn ) of elementary functions so that
RT RT
0
(f − fn )2 dm → 0 in probability. It turns out that then also ( 0 fn dB) will converge in
probability and we can therefore define
Z T Z T
f dB = lim fn dB in probability.
0 n 0
Note however that since conditional expectations are not continuous in probability, this extended
Ito integral will in general not be a martingale.
Let n ∈ N and let (Ω, F, P ) be a probability space on which we can find n independent Brownian
motions, B1 (t), B2 (t), · · · , Bn (t)). We can then put B(t) = (Bj (t) to get an n–dimensional
Brownian motion. As before we let for every t ≥ 0 Ft denote the σ–algebra generated by
{B(s) | 0 ≤ s ≤ t} and the zero sets N .
If A(t, ω) is an m × n stochastic matrix which is (Ft )–adapted and so that all entries satisfy the
RT
equation (2.1) above, we can define the m–dimensional Ito integral 0 AdB by writing the dB
RT
as a column “vector” and perform matrix multiplication, e.g. the kth coordinate of 0 AdB will
be n Z T
X
Akj dBj .
j=1 0
It follows that if each entry of A is square integrable in both variables, this Ito integral will be an
m–dimensional martingale.
10
3 Ito’s formula
We consider the one dimensional case and let (Bt ) be an one dimensional Brownian motion.
Definition 3.1 An Ito process is a stochastic process of the form
Z t Z t
X t = X0 + u(s, ω)dt + v(s, ω)dBt (ω) t ≥ 0,
0 0
where u and v are so that the integrals make sense for all t ≥ 0.
If X is an Ito process of the form above, we shall often write
Theorem 3.2 Let dXt = udt + vdBt be an Ito process and let g ∈ C 2 ([0, ∞[×R). Then
Yt = g(t, Xt ) is again an Ito process and
∂g ∂g 1 ∂ 2g
dYt = (t, Xt )dt + (t, Xt )dXt + 2
(t, Xt )(dXt )2 .
∂t ∂x 2 ∂x
We shall not prove it here. It is based on Taylor expansions and the difficult part is to show that
that the remainer tends to zero in L2 .
There is also an Ito formula in higher dimensions.
Rt
As an example of the use of Ito’s formular let us compute 0
Bs dBs .
Ito’s formular used with the function x2 gives
so that Z t
1
Bs dBs = (Bt2 − t)
0 2
(Bt2
In particular it follows that − t) is a martingale, a fact we shall prove directly later. The next
theorem is called the Ito representation theorem
Theorem 3.3 Let 0 < T < ∞ and let F ∈ L2 (Ω, FT , P ). Then there is a unique f ∈ A2 (0, T )
so that Z T
F = E(F ) + f dB.
0
We shall only prove it in the one–dimensional case. We need however two lemmas.
11
Lemma 3.4 The set
{φ(Bt1 , Bt2 , · · · , Btn ) | n ∈ N, φ ∈ C0∞ (Rn ), (ti ) ⊆ [0, T ]}
is dense in L2 (Ω, FT , P )
Proof: Let (ti ) be a dense sequence in [0, T ] and let for each n Hn be the σ–algebra generated
by {Bti | 0 ≤ i ≤ n} and the zero sets. Clearly FT is the smallest σ–algebra containing all of
the Hn ’s. Let now g ∈ L2 (Ω, FT , P ) be arbitrary. By the martingale convergence theorem we
get that g = E(g | FT ) = limn E(g | Hn ), where the limit is in L2 and a.s.
A result of Doob and Dynkin gives the existence of a Borel function gn : Rn → R so that for
every n:
E(g | Hn ) = gn (Bt1 , Bt2 , · · · , Btn ).
Let µ denote the distribution Borel measure on Rn of (Bt1 , Bt2 , · · · , Btn ), i.e µ = (Bt1 , Bt2 , · · · , Btn )(P ).
Note that µ has a normal density which implies that C0∞ (Rn ) is dense in L2 (µ).
From the above we get:
Z Z Z
gn2 dµ = 2
gn (Bt1 , Bt2 , · · · , Btn ) dP ≤ g 2 dP
Rn Ω Ω
so that gn ∈ L2 (µ). Hence gn can be approximated well in L2 (µ) by a function φn ∈ C0∞ (Rn )
and hence gn (Bt1 , Bt2 , · · · , Btn can be approximated well in L2 (P ) by φn (Bt1 , Bt2 , · · · , Btn ),
Combining this with the above we get the result. 2
RT RT
Lemma 3.5 Put M = {exp( 0 hdB − 12 0 h(s)2 ds) | h ∈ L2 (0, T )}. Then span(M) is dense
in L2 (Ω, F T , P ).
RT RT
Proof: Note that 0 hdB is normally distributed and therefore exp( 0 hdB) ∈ L2 (P ). Note
RT
also that the term 21 0 h(s)2 ds is actually not needed since we take the span.
Let now g ∈ L2 (Ω, FT , P ) be orthogonal to M. We have to prove that g = 0 a.s.
Let n ∈ N and let {tj | 1 ≤ j ≤ n} For all λ = (λ1 , λ2 , · · · , λn ) ∈ Rn we get that
Z Xn
exp( λj Btj )gdP = 0
Ω j=1
12
We wish to show that g is orthogonal to the set from the previous lemma, so let φ ∈ C0∞ (Rn ).
By the inverse Fourier transform theorem we have
Z
−n
φ(x) = (2π) 2 φ̂(y) exp(i < x, y >)dmn
Rn
In particular Z T
F =1+ Ys h(s)dBs
0
Clearly the function (t, ω) → Yt (ω)h(t) is (Ft )–adapted so we need to verify that it belongs to
L2 (m ⊗ P ) .
Rt Rt
We note that for fixed t 0 hdB is normally distributed with mean 0 and variance σt2 = 0 h(s)2 ds
and hence
R∞ x2
E(Yt2 ) = σ √1 2π −∞ exp(2x − σt2 − 2σ 2 )dx =
t t
R∞ (x−2σ 2 )2
exp(σt2 ) σ √1 2π −∞ exp(− 2σ2t )dx = exp(σt2 ) .
t t
13
Therefore Z T Z T Z t
2
h(t) E(Yt2 )dt = h(t) exp( h(s)2 ds)dt < ∞.
2
0 0 0
Proof: We shall only prove it for n = 1. Let 0 ≤ t < ∞. The representation theorem give us a
unique f t ∈ A2 (0, t) so that Z t
Mt = E(M0 ) + f t dB.
0
If 0 ≤ t1 < t2 , then Z t1
Mt1 = E(Mt2 | Ft1 ) = E(M0 ) + ft2 dB.
0
But Z t1
Mt1 = E(M0 ) + f t1 dB
0
t1 t2
so by uniqueness f (t, ω) = f (t, ω) for almost all (t, ω) ∈ [0, t1 ] × Ω. If we now put
f (t, ω) = f N (t, ω) for almost all 0 ≤ t ≤ N and almost all ω, then f is well–defined and
is clearly the one we need. 2
or in differential form
dXt = b(t, Xt )dt + σ(t, Xt )dBt
where b and σ are so that the integrals make sense.
14
As an example we can consider the equation
Let n, m ∈ N and let Mnm denote the space of all n × m–matrices. Further let b : [0, T ] × Rn →
Rn and σ : [0, T ] × Rn → Mnm be measurable functions so that there exists constants C and D
with
• kb(t, x)k + kσ(t, x)k ≤ C(1 + kxk)
• kb(t, x) − b(t, y)k + kσ(t, x) − σ(t, y)k ≤ Dkx − yk
for all x, y ∈ Rn and all t ∈ [0, T ]. Here k · k denotes the norm in the Euclidian space (we
identify here Mnm with Rnm . Further we let B be an m–dimensional Brownian motion.
We have the following existence and uniqueness theorem:
15
Theorem 4.1 Let Z ∈ L2 (P ) so that Z is independent of {Ft | 0 ≤ t ≤ T . The equation
has a unique solution with X ∈ L2 (m ⊗ P ) so that X is adapted to the σ–algebra FtZ generated
by Z and Ft .
We shall not prove the theorem here. The uniqueness is based on the assumptions above and Ito
isometry. The existence is based on Picard iteration.
(0) (k)
In fact, we put Yt = Z and define Yt inductively by
Z t Z t
(k+1) (k)
Yt =Z+ B(s, Ys )ds + σ(s, Ys(k) )dBs .
0 0
We then use our assumptions to prove that the sequence (Y (k) ) has a limit in L2 (m ⊗ P ). This
limit is our solution. The uniqueness involves Ito isometry.
Definition 4.2 A time homogeneous Ito diffusion (Xt ) is a process that satisfies an equation of
the form
dXt = b(Xt )dt + σ(Xt )dBt 0 ≤ s ≤ t Xs = x ∈ Rn
where b : Rn → Rn and σ : Rn → Mnm satisfy
If n ∈ N, we let h·, ·i denote the canonical inner product on Rn . Hence for all x = (x1 , x2 , . . . , xn ) ∈
Rn og alle y = (y1 , y2 , . . . , yn ) ∈ Rn we have
n
X
hx, yi = xj yj . (5.2)
j=1
Let (Ft )t≥0 be an increasing family of sub-σ-algebras so that Ft contains all sets of measure 0
for all t ≥ 0 (it need not be generated by any Brownian motion). We start with the following
easy result.
Theorem 5.1 Let (Bt ) be a one–dimensional normalized Brownian motion, adapted to (F) and
so that Bt − Bs is independent of Fs for all 0 ≤ s < t (this ensures that (Bt ) is a martingale
with respect to (Ft )). Then (Bt2 − t) is a martingale with respect to (Ft ).
16
Proof: If 0 ≤ s < t, then Bt2 = (Bt − Bs )2 + Bs2 + 2Bs (Bt − Bs ) and hence
where we have used that Bt − Bs and hence also (Bt − Bs )2 are independent of Fs . 2
The main result of this section is to prove that the converse is also true for continuous processes,
namely:
Theorem 5.2 Let (Xt ) be a continous process adapted to (Ft ) so that X0 = 0 and
(i) (Xt ) is a martingale with respect to (Ft ).
(ii) (Xt2 − t) is a martingale with respect to (Ft ).
Then (Xt ) is a (normalized) Brownian motion.
Before we can prove it, we need yet another theorem which is a bit like Ito’s formula and a
lemma.
Theorem 5.3 Let (Xt ) be as in Theorem 5.2 and let f ∈ C 2 (R) so that f , f 0 and f 00 are bounded.
For all 0 ≤ s < t we have
1 t
Z
E(f (Xt ) | Fs ) = Xs + E(f 00 (Xu ) | Fs )du. (5.3)
2 s
Proof: Let Π = (tk )nk=0 be a partition of the interval [s, t] so that s = t0 , t1 < t2 < · · · , < tn = t.
By Taylor’s formula we get
n
X
f (Xt ) = f (Xs ) + (f (Xtk ) − f (Xtk−1 )) (5.4)
k=1
n n
X
0 1 X 00
= f (Xs ) + f (Xtk−1 )(Xtk − Xtk−1 ) + f (Xtk−1 )(Xtk − Xtk−1 )2 + RΠ
k=1
2 k=1
17
Using the continuity of the (Xt ) it can be shown that RΠ → 0 in L2 (P ), when the length
|Π| of Π tends to 0. Hence also E(RΠ | Fs ) → 0 in L2 (P )as |Π| → 0. Since the function
u → E(f 00 (Xu ) | Fs )) is continuous a.s., we get that
n
X Z t
00
E(f (Xtk−1 ) | Fs )(tk − tk−1 ) → E(f 00 (Xu ) | Fs )du a.s. (5.6)
k=1 s
when |Π| → 0 and since f 00 is bounded, the bounded convergence theorem gives that the conver-
gence in (5.6) is also in L2 (P ). Combining the above we get formula (5.3). 2
Proof of Theorem 5.2: The main part of the proof will be to prove that for all 0 ≤ s ≤ t we
have the formula
1
E(exp(iu(Xt − Xs )) | Fs ) = exp(− u2 (t − s)) for all u ∈ R. (5.8)
2
To prove (5.8) fix an s with 0 ≤ s < ∞, a u ∈ R, and apply Theorem 5.3 to the function
f (x) = exp(iux) for all x ∈ R. For all s ≤ t we then obtain:
1 2 t
Z
E(exp(iuXt ) | Fs ) = exp(iuXs ) − u E(exp(iuXv ) | Fs )dv
2 s
or Z t
1
E(exp(iu(Xt − Xs )) | Fs ) = 1 − u2 E(exp(iu(Xv − Xs )) | Fs )dv. (5.9)
2 s
18
Since the integrand on the right side of (5.9) is continuous in v, the left hand side is differentiable
with respect to t and
d 1
E(exp(iu(Xt − Xs )) | Fs ) = − u2 E(exp(iu(Xt − Xs ) | Fs ).
dt 2
This shows that on [s, ∞[ E(exp(iu(Xt − Xs )) | Fs ) is the solution to the differential equation
1
g 0 (t) = − u2 g(t)
2
with the initial condition g(s) = 1. Hence
1
E(exp(iu(Xt − Xs )) | Fs ) = exp(− u2 (t − s)) for all 0 ≤ s ≤ t
2
and equation (5.8) is established.
Let now 0 ≤ s < t. By (5.8) the characteristic function of Xt − Xs is given by:
1
E(exp(iu(Xt − Xs )) = E(E(exp(iu(Xt − Xs )) | Fs )) = exp(− u2 (t − s))
2
and hence Xt − Xs is normally distributed with mean 0 and variance t − s.
Let now 0 = t0 < t1 < t2 < · · · < tn < ∞ and put Y = (Xt1 , Xt2 − Xt1 , . . . , Xtn − Xtn−1 ). If
φY denotes the characteristic function of Y , then we get for all u = (u1 , u2 , . . . , un ) ∈ R:
n
Y
φY (u) = exp(i < u, Y >) = E( exp(iuk (Xtk − Xtk−1 )) =
k=1
n
Y
E(E( exp(iuk (Xtk − Xtk−1 )) | Ftn−1 ) =
k=1
n−1
1 Y
exp(− u2n (tn − tn−1 ))E( exp(iuk (Xtk − Xtk−1 ))
2 k=1
which together with Lemma 5.5 shows that Xt1 , Xt2 − Xt1 , · · · , Xtn − Xtn−1 are independent.
Thus we have proved that (Xt ) is a normalized Brownian motion. 2
In many cases where Theorem 5.2 is used, Ft is for each t the σ–algebra generated by {Xs | 0 ≤
s ≤ t} and the sets of measure 0. However, the theorem is often applied to cases where the Ft ’s
are bigger.
We end this note by showing that the continuity assumption in Theorem 5.2 can not be omitted.
Let us give the following definition:
19
Definition 5.6 An (Ft )–adapted process (Nt ) is called a Poisson process with intensity 1 if N0 =
0 a.s. and for 0 ≤ s < t, Nt − Ns is independent of Fs and Poisson distributed with parameter
t − s.
Hence if (Nt ) is a Poisson process with intensity 1, then Nt − Ns takes values in N ∪ {0} for all
0 ≤ s < t and
(t − s)k
P (Nt − Ns = k) = exp(−(t − s)) for all k ∈ N ∪ {0}
k!
It can be proved that such processes exist.
Easy calculations show that E(Nt − Ns ) = t − s = V (Nt − Ns ). The process (Mt ), where
Mt = Nt − t for all t ∈ [0, ∞[, is called the compensated Poisson process with intensity 1. Note
that (Mt ) is not continuous. We have however:
Theorem 5.7 If (Mt ) is a compensated Poisson process with intensity 1, then it satisfies the
conditions (i) and (ii) in Theorem 5.2.
Proof: Let 0 ≤ s < t. Since Mt − Ms is independent of Fs , we get
E(Mt | Fs ) = Ms + E(Mt − Ms ) = Ms .
6 Girsanov’s theorem
In this section we let again (Bt ) denote a one–dimensional Brownian motion, let 0 < T < ∞,
and let (Ft ) be defined as before. Before we can formulate the main theorem of this section we
need a little preparation. Let us recall that if Q is another probability measure on (Ω, F), then Q is
said to be absolutely continuous with respect to P , written Q << P , if P (A) = 0 ⇒ Q(A) = 0
for all A ∈ F. A famous result of Radon and Nikodym says that in that case there is a unique
h ∈ L1 (P ) so that Z
Q(A) = hdP for all A ∈ F.
A
We often write this as dQ = hdP . In this situation we have:
Theorem 6.1 If f ∈ L1 (Q) and H is a sub–σ–algebra of F, then
EQ (f | H)EP (h | H) = EP (f h | H)
20
Proof: Let A ∈ H be arbitrary. On one hand we have:
Z Z Z
EQ (f | H)hdP = EQ (f | H)dQ = f dQ =
A ZA Z A
f hdP = EP (f h | H)dP .
A A
EQ (f | H)EP (h | H)dP.
A
and put
t
1 t
Z Z
Mt = exp(− adB − a(s, ·)2 ds) for all 0 ≤ t ≤ T . (6.2)
0 2 0
Assume that (Mt )0≤t≤T is a martingale. If we define the measure Q on FT by dQ = MT dP , then
Q is a probability measure and (Yt ) i a Brownian motion with respect to Q.
21
Before we can prove the theorem in a special case, we will investigate when expressions like the
ones in (6.2) form a martingale. Hence let us look on
Z t
1 t
Z
Mt = exp( adB − a(s, ω)2 ds) for all 0 ≤ t. (6.3)
0 2 0
(we write a instead of −a, since the sign does not matter for our investigation)
Ito’s formula gives that
1 1
dMt = Mt (a(t, ·)dBt − a(t, ·)2 dt) + Mt a(t, ·)2 dt =
2 2
Mt a(t, ·)dBt ,
so that Z t
Mt = 1 + aM dB for alle t ≥ 0. (6.4)
0
Our first result states:
Theorem 6.3 (i) (Mt ) is a supermartingale with EMt ≤ 1 for all t ≥ 0.
(ii) (Mt ) is a martingale if and only if EMt = 1 for all t ≥ 0.
Proof: For every n ∈ N we put
Z t
τn = inf{t > 0 | Ms2 a(s, ·)2 ≥ n}
0
22
Since τn is a stopping time, 1[0,τn ] aM ∈ A2 (0, t) and hence (Mt∧τn ) is a martingale with
EMt∧τn = 1 for all n ∈ N. Since Mt∧τn ≥ 0 the Fatou lemma gives that
EMt ≤ lim inf EMt∧τn = 1 for all t ≥ 0.
If we apply Fatou’s lemma for conditional expectations we get for all 0 ≤ s < t that
E(Mt | Fs ) ≤ lim inf E(Mt∧τn | Fs ) = lim Ms∧τn = Ms ,
then we need to show that P (A) = 0. The assumption P (A) > 0 gives that
Z
1 = EMt = E(Mt | Fs )dP =
Z Z ZΩ Z
E(Mt | Fs ) + E(Mt | Fs )dP < Ms dP + Ms dP =
A Ω\A A Ω\A
EMs = 1
which is a contradiction. Hence P (A) = 0 and (Mt ) is a martingale. 2
In connection with applications of the Girsanov theorem it is of course important to find suffi-
cient conditions for (Mt ) being a martingale, often only in the interval [0, T ]. One of the most
important sufficient conditions is the Novikov condition:
1 T
Z
E exp( a(s, ·)2 ds) < ∞ where 0 < T < ∞. (6.6)
2 0
If (6.6) holds for a fixed T , then {Mt | 0 ≤ t ≤ T } is a martingale and if (6.6) holds for all
0 ≤ T < ∞, then {Mt | 0 ≤ t} is a martingale. It lies outside the scope of these lectures to show
this and we shall therefore do something simpler which covers most cases that appear in practice:
Since aM is adapted, it follows from (6.4) that if aM ∈ L2 ([0, t] × Ω) for every 0 ≤ t < ∞
(respectively for every 0 ≤ t ≤ T < ∞), then {Mt | 0 ≤ t} is a martingale (respectively
{Mt | 0 ≤ t ≤ T } is a martingale). The next theorem gives a sufficient condition for this:
Theorem 6.4 Let f : [0, ∞[→ [0, ∞[ be a measurable function and 0 < T < ∞. If
f ∈ L2 [0, T ] (6.7)
and
|a(t, w)| ≤ f (t) for all 0 ≤ t ≤ T and a.a. s ∈ Ω, (6.8)
then:
23
(i) For all 1 ≤ p < ∞ and all 0 ≤ t ≤ T we have Mt ∈ Lp (P ) with
p2 − p t
Z
p
f (s)2 ds .
EMt ≤ exp (6.9)
2 0
Note that in particular Theorem 6.4 is applicable in the important case where a is bounded.
Before we go on, we wish to make a small detour and apply the above to geometric Brownian
motions. Hence let (Xt ) be a geometric Brownian motion starting in a point x ∈ R, say
1 1
Xt = x exp (r − α2 )t + αBt = x exp(rt) exp(αBt − α2 t) for all t ≥ 0,
2 2
where r, α ∈ R. By the above (exp(αBt − 12 α2 t)) is a martingale and therefore E(Xt ) =
x exp(rt) for all t ≥ 0. This can of course also be obtained using that Bt is normally distributed.
We also need:
24
Lemma 6.5 Let 0 ≤ T < ∞ and assume that {Mt | 0 ≤ t ≤ T } is a martingale.Let Q be
the measure on FT defined by dQ = MT dP . If (Xt ) ⊆ L1 (Q) is (Ft )–adapted, then (Xt ) is a
Q–martingale if and only if (Xt Mt ) er en P –martingale.
Proof:
This is an immediate consequence of the following formula which follows directly from Theorem
6.1 together with our assumptions. For all 0 ≤ s < t we have:
EP (Xt Mt | Fs ) = EQ (Xt | Fs ))EP (Mt Fs ) = EQ (Xt | Fs )Ms .
2
Hence we can finish by proving that the integrand belongs to L2 ([0, T ] × Ω). We note that
Mt |1 − Yt a(t)| ≤ Mt + Mt |Yt |f (t)
and since E(Mt ) = 1 for all 0 ≤ t ≤ T , M ∈ L2 ([0, T ] × Ω). Further
Z t
|Yt | ≤ f (s)ds + |Bt |
0
so that Z T
Mt |Yt |f (t) ≤ f (t)Mt f (s)ds + f (t)Mt |Bt | (6.11)
0
For p = 2 Theorem 6.4 gives:
Z T
f (t)2 E(Mt2 )dt ≤
0
Z T Z t
f (t) exp( f (s)2 ds)dt ≤
2
0 0
Z T Z T
2
exp( f (s) ds) f (t)2 dt < ∞
0 0
25
which takes care of the first term in (6.11).
To take care of the second term we use Theorem 6.4 and the Cauchy–Schwartz’s inequality to
get:
2 2 4 12 4 21
√ Z T
E(Mt Bt ) ≤ E(Mt ) E(Bt ) ≤ 3T exp(3 f (s)2 ds)
0
Proof: It follows from Theorem 6.2 that Q is a probability measure on FT and that B̃ is a
Q–Brownian motion. Further we get:
Theorem 6.2 and its corollary can be generalized to higher dimensions. In that case the a in
Theorem 6.2 will take values in Rn and if we interpret a2 as kak2 , then (Mt ) and Q are defined
as before and the result carries over using a multi–dimensional form of Lévy’s result. In the
26
corollary (Bt ) will be an m–dimensional Brownian motion, u will take values in Rn and v will
take values in the space of n×m matrices. The requirement is then that the matrix equation va =
u has a solution satisfying the requirements of the corollary. If our process (Xt ) there represents
a financial market, then the mathematical conditions reflex to some extend the behaviour in
practice of the financial market. In other words, Theorem 6.2 and Corollary 6.6 has a lot of
applications in practice.
At the end we will discuss under which conditions Theorem 6.2 can extended to the case where
T = ∞. Hence we let a : [0, ∞[×Ω → R satisfy (6.1) and define Mt as in (6.3), that is
Z t
1 t 2
Z
Mt = exp( adB − a dm) t ≥ 0.
0 2 0
For convenience we shall assume that F is equal to the σ–algebra generated by {Ft | 0 ≤ t}.
If (Mt ) is a martingale, we can for every t ≥ 0 define a probability measure Qt on Ft by
dQt = Mt dP and the question is now whether there is a probability measure Q on F so that
Q|Ft = Qt for all 0 ≤ t < ∞. The next theorem gives a necessary and sufficient condition for
this to happen.
Theorem 6.7 Assume that {Mt | 0 ≤ t} is a martingale. Then M∞ = limt→∞ Mt exists a.s.
The following statements are equivalent:
(i) There exists a probability measure Q on F with Q << P og Q | Ft = Qt for all t ≥ 0.
(ii) (Mt ) is uniformly integrable.
If (i) (or equivalently (ii)) holds, then dQ = M∞ dP .
Proof: Since EMt = 1 for alle 0 ≤ t, the martingale convergence theorem gives us the existence
of M∞ a.e.
Assume first that (i) and determine f ∈ L1 (P ) so that dQ = f dP . Since Q | Ft = Qt , it clearly
follows that E(f | Ft ) = Mt for alle 0 ≤ t. Let us show that this implies that {Mt | t ≥ 0} is
uniformly integrable. Since (Mt (ω)) is convergent for a.a ω, supt≥0 Mt (ω) < ∞ for a.a ω. If
0 ≤ t < ∞ and x > 0, then
Z Z
Mt dP = E(f | Ft )dP =
(Mt >x) (Mt >x)
Z Z
f dP ≤ f dP,
(Mt >x) (sup Ms >x)
27
which shows that (Mt ) is uniformly integrable.
Assume next that (ii) holds. Then Mt → M∞ in L1 (P ) which implies that EM∞ = lim EMt =
1, and that E(M∞ | Ft ) = lims E(Ms | Ft ) = Mt for all t ≥ 0. If we put dQ = M∞ dP , then Q
is a probability measure and if t ≥ 0 and A ∈ Ft , then:
Z Z
Q(A) = M∞ dP = E(M∞ | Ft )dP =
A
Z A
Mt dP = Qt (A),
A
If we combine Theorem 6.3 with Theorem 6.7 we get the following corollary.
Corollary 6.8 Let f : [0, ∞[→ [0, ∞[ be a measurable function so that
which shows that (Mt ) is bounded in L2 (P ) and therefore uniformly integrable. This proves the
corollary. 2
Girsanov’s Theorem 6.2 holds on the interval [0, ∞[, if we assume that the (Mt ) there is a uni-
formly integrable martingale. If a satisfies the conditions in Corollary 6.8 small modifications of
our proof of Theorem 6.2 will give a proof of this.
Let us end this section with the following example-
Example Lad a være konstant, a 6= 0. (6.7)) and (6.8) are clearly satisfied so that (Mt ) is a
martingale. In fact, Mt = exp(aBt − 21 a2 t) for all t ≥ 0. The martingale convergence theorem
shows that M∞ = limt→∞ Mt exists a.e. Since however M∞ = 0 a.e. (see below) and EMt = 1
for all 0 ≤ t, (Mt ) cannot be uniformly integrable.
28
That M∞ = 0 can be seen as follows: It is sufficient to show that Mt → 0 in probability because
then there is a subsequence (tn ) with Mtn → 0 a.s.
Hence let ε > 0 and determine t0 so that 12 a2 t0 + log ε > 0 and put bt = a−1 ( 12 a2 t + log ε). If
a > 0, then for all t ≥ t0 we get:
Z ∞
1 1
P (Mt ≥ ε) = P (Bt ≥ bt ) = √ exp(− x2 )dx ≤
2tπ bt 2t
1 1
Z ∞
1 √ 1 1 ∞
√ x exp( x2 )dx = t√ − exp(− x2 ) x=bt =
2tπ bt bt 2t 2π 2t
√ 1 1
t √ exp(− b2t ) → 0 for t → ∞
2π 2t
References
[1] Z. Ciesielski, Hölder condition for realizations of Gaussian processes, Trans. Amer. Math.
Soc. 99 (1961), 401–413.
[2] P. Lévy, Processus stochastiques et movement Brownien, Gauthier–Villars, Paris.
[3] I. Karatzas and S.E. Shreve, Brownian motion and stochastic calculus, Springer Verlag
1999.
[4] N.J. Nielsen, Brownian motion, Lecture notes, University of Southern Denmark.
[5] B. Øksendal, Stochastic differential equations, Universitext, 6.edition, Springer Verlag
2005.
29