Ito Shit
Ito Shit
Ito Shit
Aims
This course develops the theory and practice of financial derivatives pricing in continuous time.
The first part of the course provides an introduction to stochastic processes in continuous time. These
processes are essential for modelling asset price processes in continuous time. The main topics which
we will cover are
Theory of martingales in continuous time;
Brownian motion: definitions and properties;
Brownian motion as the limit of a binomial random-walk process;
Introduction to stochastic integration, stochastic differential equations and Itos formula;
Geometric Brownian motion and the Ornstein-Uhlenbeck process;
Introduction to Girsanovs theorem and the martingale representation theorem.
In the second part of the course we will cover some of the core topics in continuous-time financial
derivative pricing. The main topics we will cover are
The Black-Scholes model;
Derivatives pricing using the Black-Scholes model using the martingale and PDE approaches
to pricing;
Extensions to foreign currencies and dividend-paying stocks;
Portfolio risk management using the Greeks;
Introduction to interest rate models;
Introduction to credit risk models.
Web Page
Course information will be posted on Vision, the Universitys online learning environment. All
handouts will also be available on Vision.
Office Hours
Office hours will be announced on Vision.
Feedback
Verbal feedback on tutorial solutions will be given during tutorial hours. For example, if you have
a tutorial solution which differs from the official solution you could ask whether your solution is
correct. Students also have the opportunity to get feedback during office hours. I am also happy to
give feedback on your attempts at past exams.
Reading
Some additional reading is suggested to get a better understanding of the course material.
For the first part of the course, the recommended textbooks are Williams, and Durrett and/or
ksendal. Williams covers introductory probability, integration (i.e. expectation), and the theory
of discrete-time martingales. Durrett and ksendal cover the continuous-time setting, including
Brownian motion, continuous-time martingales, Girsanovs theorem, the martingale representation
theorem and, of course, the theory of stochastic differential equations.
For the second part of the course, the recommended textbooks are Baxter and Rennie, Bj
ork
and Hull. Baxter and Rennie give a good description of the Black-Scholes model. Bjork provides
a nice, simple description of much of the stochastic analysis needed for the course, as well as much
continuous-time theory. Hull has a broad coverage of the practical side of modelling derivatives
markets, and is one of the best known books in the field.
There are many other useful books which can be found in the library, containing similar or related
material.
Bibliography
1. Williams, D. (1991), Probability with martingales. Cambridge University Press.
2. Durrett, R. (1996), Stochastic Calculus, A practical Introduction. CRC Press.
3. ksendal, B. (2003), Stochastic differential equations, 6th ed. Springer.
4. Baxter, M. and Rennie, A. (1996), Financial calculus. Cambridge University Press.
5. Bj
ork (2009), Arbitrage theory in continuous time, 3rd ed. Oxford University Press.
6. Hull, J. (2008) Options, futures and other derivative securities, 7th ed. Prentice Hall.
7. Jacod, J. and Protter, P. (2004) Probability Essentials, 2nd ed. Springer.
(NB: earlier or later editions of these books should be fine.)
Stochastic Processes
1.1
Preliminaries
Probability Space (, F , P)
is a sample space, the set of all possible outcomes, , of an experiment. For example = {H, T }
for tossing a coin.
An event A is a collection of outcomes (paths).
F is a family of events to which we will assign a probability. F must be a -algebra:
(i) , F ;
(ii) If A F then Ac F ;
(iii) If A1 , A2 , . . . is a countably infinite sequence of events in F then
i=1 Ai F .
c c
[
i=1 Ai = (i=1 Ai ) F ]
i=1
P(Ai ).
a R.
Lemma 1.1 If X, Y are measurable w.r.t. G , and f is a continuous function then f (X, Y ) is
measurable w.r.t. G .
Proof. For 1-dim case. This would need the results that a function is continuous if and only if the
preimage of closed sets is closed, and a function is measurable if and only if {X B} G for all
closed subsets B R. Then its easy:
{f (X) a} = X f 1 (, a) G .
3
In this case
Z
EP [X] =
X()dP :=
n
X
ai P(Ai ).
i=1
Typically, if we want to prove a result about EP [X], we first take an approximating sequence of
simple random variables with Xn X a.s., prove the result for EP [Xn ] (hopefully easier), then pass
to the limit.
An example of this is the following result.
Proposition 1.5 If F (x) = P(X x) is the distribution function of X then
Z
xdF (x).
EP [X] =
Proof (sketch, for X 0). Because of monotone convergence theorem (m.c.t.), it is sufficient to
prove this for simple X. Any simple r.v. can be written in the form
X=
n
X
a i 1A i
i=0
where Ai s are disjoint and i Ai = [Just reorder ai in increasing order and add A0 := (i Ai )c ].
P
By definition, EP [X] = ni=1 ai P(Ai ).
Clearly, F (a0 ) = P(X a0 ) = P(A0 ) and F (a1 ) = P(X a1 ) = P(A0 ) + P(A1 ) ... etc
So,
+ + an (F (an ) F (an1 ))
= a1 P(A1 ) + a2 P(A2 ) + + an P(An )
= EP [X] .
X
X
EP
Xn =
EP [Xn ] .
n=1
Proof. Define Yn =
Pn
i=1
n=1
i=1
Xi as
n . By the M.C.T.
EP
"
X
"
Xi = EP
lim
n
X
i=1
h
i
Xi = EP lim Yn
n
i=1
"
M.C.T.
= lim
n
X
EP [Xi ] =
i=1
n
X
#
Xi
i=1
EP [Xi ]
i=1
0c1
x R.
Way to remember:
As y varies, the value of EP [X|Y = y] will vary i.e. the conditional expectation is a function of y.
Hence EP [X|Y ] is a function of a random variable Y and is therefore itself a random variable.
Example: Two fair coins = {HH, HT, T H, T T }
(
1 if ith toss H
Xi =
; i = 1, 2
0 otherwise
Consider EP [X1 + X2 |X1 ].
Elementary calculation:
EP [X1 + X2 |X1 = 0] = 0 + EP [X2 ] =
1
2
3
2
HT
TH
TT
3
2
1
2
1
2
Important observations
Recall: F1 = (X1 ) = {, {T H, T T } , {HH, HT } , }
8
(i)
3
: EP [X1 + X2 |X1 ] =
= {HH, HT }
2
= {X1 = 1}
F1 .
1
: EP [X1 + X2 |X1 ] =
2
= {T H, T T }
= {X1 = 0}
F1 .
(
1 A
1A () :=
0
6 A
Conclusion: A F1 holds:
EP [1A EP [X1 + X2 |X1 ]] = EP [1A (X1 + X2 )] .
This example suggests the following definition:
Given (, F , P), let X be an F -meas. r.v., and let G F be a smaller -algebra (in general, X is
not G -measurable).
Definition 1.8 The conditional expectation EP [X|G ] is defined to be the a.s. unique r.v. satisfying
(i) EP [X|G ] is G -measurable
(ii) EP [1A EP [X|G ]] = EP [1A X] A G .
The definition is equivalent to saying that EP [X|G ] is the unique G -meas. r.v. s.t.
EP [Y EP [X|G ]] = EP [Y X]
G -meas. r.v. Y.
(*)
So, to prove a statement along the lines of EP [X|G ] = Z for some G -meas. r.v. Z, we could show
that
EP [Y Z] = EP [Y X]
G -meas. r.v. Y.
(**)
Properties of conditional expectations
(i) If X 0 a.s. then EP [X|G ] 0 a.s. for any G F
(ii) Linearity EP [a1 X1 + a2 X2 |G ] = a1 EP [X1 |G ] + a2 EP [X2 |G ].
(iii) Taking out what is known If Z is G -meas. r.v., then EP [XZ|G ] = ZEP [X|G ].
Proof. Let Y be a G -meas. r.v. Then since Y Z is G -meas., (*) implies that
EP [Y ZEP [X|G ]] = EP [Y ZX] .
By (**) this means that EP [ZX|G ] = ZEP [X|G ].
10
(vii) Independence and conditioning X is independent of G (i.e. P({X a}A) = P(X a)P(A),
a R, A G ) iff EP [X|G ] = EP [X].
Proof ( = only). Suppose X is independent of G . We need to show that
EP [Y EP [X]] = EP [Y X]
G -meas Y.
LHS = EP [X] EP [Y ]
because EP [X] is a constant.
RHS = EP [Y X] = EP [Y ] EP [X]
by independence of X and Y (Independence of X and Y is an easy exercise!).
Note: The conditional expectation EP [X|G ] is the G -measurable rv Y that minimizes EP [(Y X)2 ].
It is used in least squares estimation (LSE): Think of G as some data available, in which case
EP [X|G ] is the LSE of X.
The monotone convergence theorem and Jensens inequality apply to conditional expectations.
e.g. Monotone convergence theorem
If Xn 0 s.t. 0 X1 X2 . . . and Xn X a.s. as n then
h
i
lim EP [Xn |G ] = EP lim Xn |G = EP [X|G ] a.s.
n
Jensens inequality
If f is a convex function,
EP [f (X)|G ] f (EP [X|G ]) a.s.
Relationship to elementary notion of conditional expectation:
For a fixed value y we know that EP [X|Y = y] is a number which represents a particular value of the
random variable EP [X|(Y )]. Similarly, the r.v. EP [X|G ] can take on values of the form EP [X|A],
for some A G .
For example, in the previous coin tossing example, X = X1 + X2 , G = (X1 ).
We calculated that:
3
EP [X|X1 = 1] = .
2
1
EP [X|X1 = 0] = ,
2
11
1.2
Stochastic processes
Let T [0, ) be a time index. The two usual cases are T = {0, 1, 2, . . . } and T = [0, )
Definition 1.9 A real-valued stochastic process is a family of real-valued random variables X =
{Xt }tT defined on (, F , P).
If T = {0, 1, 2, . . . } then X is a sequence of random variables, and we say that X is a discrete-time
stochastic process (or a time series). If T = [0, ) then we say that X is a continuous-time stochastic
process.
We shall sometimes write X(t) instead of Xt , but these mean the same. For a particular value we
write X(t, ) = Xt (). When we write X() we mean a particular paths of X.
If you are observing a share price over time then information about the process will be gradually
revealed over time. We model this with a filtration.
Definition 1.10 A filtration of (, F , P) is an increasing family of -algebras {Ft }tT such that
Ft F . We shall say that (, F , (Ft )tT , P) is a filtered probability space.
Increasing means that if s t, then Fs Ft . Think of a filtration as the history of the process
X.
Definition 1.11 Let X be a stochastic process defined on (, F , P). We say that X is adapted to
the filtration (Ft )tT if Xt is Ft -measurable for each t T.
Any process X is automatically adapted to the filtration FtX := {Xs : s t} generated by itself.
Example 1.12 Suppose that the first coin is tossed at time 1, and the second at time 2. The
winnings from both bets can be represented by a process Y = {Yt }tT with T = {0, 1, 2}:
Y0 := 0,
Y1 := X1 ,
Y2 := X1 + X2 .
1.3
Martingales
Definition 1.13 A stochastic process {Mt }tT is said to be a P-martingale with respect to a filtration
{Ft }tT if
(i) EP [|Mt |] < for all t T,
(ii) {Mt }tT is {Ft }tT -adapted,
(iii) EP [Mt |Fs ] = Ms if s t.
12
s t.
Note however, if Mt satisfies EP [Mt ] = EP [Ms ] for all s t then Mt does not have to be a martingale.
A process Mt is a {Ft }-supermartingale if (i)-(ii) hold and (iii) EP [Mt |Fs ] Ms a.s. So EP [Mt ]
EP [Ms ] for s t.
A process Mt is a {Ft }-submartingale if (i)-(ii) hold and (iii) EP [Mt |Fs ] Ms a.s. So EP [Mt ]
EP [Ms ] for s t.
Note: A process Mt is a martingale iff it is a super- and sub-martingale.
Proposition 1.14 A discrete-time stochastic process Mn is an Fn -martingale iff
EP [Mn+1 |Fn ] = Mn
n N.
= Mn if EP [Xn+1 ] = 0
Mn if EP [Xn+1 ] 0
Mn if EP [Xn+1 ] 0.
The symmetric random walk is a special case of the last example: Let X1 = 1 with equal probability,
so EP [X1 ] = 0.
Example 1.16 (Product of i.i.d. r.v.s) Let Z1 , Z2 , . . . be i.i.d. with EP [|Z1 |] < .
Q
Let Mn = m0 nk=1 Zk , m0 R. Suppose EP [Z1 ] = 1 and let Fn = {Z1 , Z2 , . . . , Zn }. Then Mn is
a {Fn }-martingale.
"
EP [|Mn |] = EP |m0 |
n
Y
#
|Zk | = |m0 |
k=1
n
Y
EP [|Zk |]
k=1
(
u
Z1 =
d
with probability p
with probability 1 p.
e.g.
14
2
S0 u3
*
S0 u
HH
p
*
H
jp
H
1
S0 u
S0 u2 d
H
p
p
*
*
HH
jp
H
1
S0 ud
S0
HH
HH
p
*
H
H
jp
H
jp
H
1
1
S0 ud2
S0 d
HH
p
*
H
jp
H
1
2
S0 d
HH
H
jp
H
1
S0 d3
e d
ud
then EP [e Z1 ] = 1, so
Mn := (e Z1 )(e Z2 ) . . . (e Zn ) = en Sn
is a martingale.
Example 1.17 (Doobs martingale) Let X be a r.v. on (, F , {Ft }tT , P) such that EP [|X|] <
, and let Mt := EP [X|Ft ]. Then Mt is a martingale.
(i) By Jensens inequality
EP [|Mt |] = EP [|EP [X|Ft ] |]
EP [EP [|X||Ft ]] = EP [|X|] <
(ii) Mt is Ft -measurable, so the process is adapted
(iii) By the tower property, EP [Mt |Fs ] = EP [EP [X|Ft ] |Fs ] = EP [X|Fs ] = Ms .
Proposition 1.18 (Jensens inequality) Suppose Mt is a martingale, and let f be a convex function. Then f (Mt ) is a submartingale (assuming EP [f (Mt )] < ).
Proof. EP [f (Mt )|Fs ] f (EP [Mt |Fs ]) = f (Ms ).
15
1.4
0m
HH
HH
H
HH
H
HH
j m
-1
t=0
t=1
E [M1 ] = 0,
VarP (M1 ) = E M12 E [M1 ]2 = 1.
*
i
H
*
j
H
e
H
*
*
j e
H
m
HH
* HH
j e
j
HH
*
j i
HH
j
m
*
e
HH
j e
*
e
HH
j i
*
e
HH
j e
*
e
HH
j m
(n)
(n)
and
Xt :=
t w.p. 1/2.
Note that E
(n)
Xt
= 0 and VarP
(n)
M0
= 0,
(n)
Xt
(n)
Mt
:=
k
X
(n)
Xi.t ,
t = k.t T(n) .
i=1
(0)
Note that Mt
We know from
Example
1.15 that each process M (n) is a martingale. Furthermore, for any t = kt
h
i
(n)
we have E Mt
= 0, and
VarP
(n)
Mt
k
X
VarP (Xi.t ) =
i=1
k
X
t = k.t = t,
i=1
(n)
e.g. VarP M1
= 1.
Rewrite
(n)
Mt
= t
!
Pk
X
kE
[X
]
i.t
t
i=1
p
.
k VarP (Xt )
Keeping t fixed, the Central Limit Theorem states that the term in brackets has a standard normal
distribution in the limit as n . Therefore,
()
Wt :=Mt
N (0, t).
Properties of Wt
(i) W0 = 0:
(n)
W0 = limn M0
= 0.
As n gets larger, the paths which the random walk Mt takes become continuous. In the
limit, you could draw the trajectories without lifting pen from paper.
(iii) Wt Ws N (0, t s):
Take s = j.t < t = k.t. Then
(n)
Mt
Ms(n)
k
X
Xit
i=j+1
= ts
Pk
!
X
(k
j)E
[X
]
i.t
t
i=j+1
p
.
(k j) VarP (Xt )
Wt Ws = Mt
17
1.5
Lets return to the binomial tree (Example 1.16), but with a time-step of t = 2n . That is,
(n)
St
where
(n)
Zt
(
u
=
d
= S0 Zt Z2t . . . Zt
with probability pu
with probability 1 pu
are i.i.d. r.vs modeling the returns. Suppose we choose pu = 1/2, and we set
1
u = e( 2
1
d = e( 2
2 )t+
2 )t
We have carefully chosen u, d so that the mean and variance of the return over each time interval
are approximately et and 2 t:
Using the expansion ex = 1 + x + x2 /2 + . . . ,
E [Zt ] = pu + (1 p)d
1 2
1 2
1
1
= e( 2 )t+ t + e( 2 )t t
2
2
1
1 2
1
= 1 + ( )t + t + 2 t + . . .
2
2
2
1 2
1
+1 + ( )t t + 2 t + . . .
2
2
= 1 + t + . . .
et .
VarP (Zt ) E (Zt et )2
1
1
= (u et )2 + (d et )2
2
2
1 2t 1 2 t+t
= e (e 2
1)2
2
1 2
1
+ e2t (e 2 t t 1)2
2
1 2t 1 2
1
= e ( t + t + 2 t + . . . )2
2
2
2
1 2t 1 2
1
+ e ( t t + 2 t + . . . )2
2
2
2
2t 2
2
e t = (1 + 2t + . . . ) t
2 t.
Define
(n)
(n)
Xit
ln Zit ( 21 2 )t
:=
=
18
(
t
w.p. 1/2
t w.p. 1/2.
Then
(n)
ln St
= ln S0 +
= ln S0 +
k
X
(n)
ln Zit
(t = k.t)
i=1
k
X
1
(n)
(( 2 )t) + Xit
2
i=1
k
X (n)
1
= ln S0 + k( 2 )t +
Xit
2
i=1
1
(n)
= ln S0 + ( 2 )t + Mt
2
Taking the limit as n ,
1
ln St = ln S0 + ( 2 )t + Wt .
2
Thus
St = S0 e( 2
2 )t+W
t
This process is called geometric Brownian motion, and is the model that we will use for a share price.
19
1.6
The most important example of a continuous-time martingale is a Brownian motion. This process is
named after Robert Brown, a Scottish botanist (born in Montrose, Angus) who was studying pollen
grains in suspension in the early nineteenth century. He observed the pollen was performing a very
random movement and thought this was because the pollen grains were alive. We now know this
rapid movement is due to collisions at the molecular level.
Two of the many people who are responsible for giving a mathematical description of Brownian
motion are Louis Bachelier (in his PhD thesis Theorie de la Speculation in 1900), and Albert
Einstein (in a 1905 paper).
Definition 1.19 A process X is said to have continuous sample paths if the map t 7 Xt () is
continuous for almost all .
Definition 1.20 A stochastic process W = {Wt }t0 is called a standard Brownian motion if
(i) W0 = 0;
(ii) W has continuous sample paths;
(iii) Wt Ws N (0, t s) for s t;
(iv) Wt Ws is independent of FsW := {Wu : u s}.
FtW t0 is the filtration generated by the Brownian motion. Hence (iv) means that the increments
are independent of everything that happened up to time s.
20
(iii) Given s t
E [Wt |Fs ] = E [Wt Ws |Fs ] + E [Ws |Fs ]
= E [Wt Ws ] + Ws = Ws .
In the tutorials, well show that both Mt := Wt2 t and Zt := exp(Wt (2 /2)t) are {Ft }t0 martingales.
Multidimensional Brownian motion
(Wt1 , Wt2 , . . . , Wtn ) where Wti are independent 1-dimensional Brownian motions.
21
1.7
Quadratic variation
Want to measure the amount of oscillation (or volatility) in a process. Take an interval [0, t], and
partition into n intervals of equal length t = t/n.
n1
X
i=0
n1
X
(X(si+1 ) X(si ))2 .
i=0
n1
X
i=0
22
1.8
Riemann integration
Let h be a function on [0, t]. Then we define
Z
n1
X
h(s)ds := lim
h(si )(si+1 si ),
i=0
n1
X
i=0
n1
X
i=0
Stochastic integration
Let Ht be a stochastic process. We will need to define the stochastic integrals
Z t
Z t
Hs dWs .
Hs ds
and
Jt :=
It :=
0
The first integral can be defined pathwise i.e. fix , set h(s) := Hs () and do normal
Riemann integration.
Rt
Note that the variation of It is 0 |Hs |ds, and the quadratic variation [I]t is identically zero.
The second integral is more difficult. It looks like a Riemann-Stieltjes integral, but because Wt has
infinite variation, the sum
X
Hsi [Wsi+1 Wsi ]
i
Pn1
(ii)
Pn1
i=0
i=0
23
In (i),
E
" n1
X
n1
X
i=0
i=0
n1
X
E Wsi (Wsi+1 Wsi )
E [Wsi ] E Wsi+1 Wsi
i=0
= 0.
In (ii),
E
" n1
X
#
Wsi+1 (Wsi+1 Wsi )
i=0
n1
X
=
=
=
=
i=0
n1
X
i=0
n1
X
i=0
n1
X
E Wsi+1 (Wsi+1 Wsi ) 0
E Wsi+1 (Wsi+1
n1
X
W si )
E Wsi (Wsi+1 Wsi )
i=0
E (Wsi+1 Wsi )2
(si+1 si ) = t.
i=0
The approximation (i) leads to the Ito integral, while (ii) leads to the Stratanovich integral. We will
concentrate on the It
o integral.
Construction of the It
o integral
A process Kt is called simple if it has the following form:
Kt = H0 1[0,s1 ] (t) +
n1
X
i=1
Step 1: Define
Z
Ks dWs :=
0
Hi (Wsi+1 Wsi ).
24
Step 2:
t
Z
"
Ks dWs = E
#
X
Hi (Wsi+1 Wsi )
E [Hi ] E Wsi+1 Wsi = 0.
Also
"Z
2 #
Ks dWs
0
!2
X
= E
Hi (Wsi+1 Wsi )
"
#
X
=E
i,j
E Hi2 (Wsi+1 Wsi )2
E Hi (Wsi+1 Wsi )Hj (Wsj+1 Wsj )
i6=j
E Hi2 E (Wsi+1 Wsi )2
E Hi2 (si+1 si )
"
=E
#
X
Hi2 (si+1 si )
i
t
Z
Ks2 ds
=E
.
Step 3:
Any continuous, adapted process Ht with E
hR
t
0
i
Hs2 ds < for all t can be approximated by a
as n .
Rt
0
t
Kn (s)dWs J(t)
i.e. E
as n .
Hs dWs := J(t).
0
Properties of It
o integrals
(i) Jt is adapted to {Ft }t0 , where Ft := {Wu : u t}.
Rt
Ru
Rt
(ii) 0 Hs dWs = 0 Hs dWs + u Hs dWs for u < t.
Rt
Rt
Rt
(iii) 0 (aHs + bHs0 )dWs = a 0 Hs dWs + b 0 Hs0 dWs for any a, b R.
i
hR
t
2
(iv) If E 0 Hs ds < for all t > 0 then the results of Step 2 are preserved:
Z
Hs dWs = 0
E
0
and
"Z
E
2 #
Hs dWs
Z
=E
Hs2 ds
.
hR
t
Hs2 ds
0
Rt
0
< t then Jt :=
Rt
0
Rt
0
Hs2 ds.
Hs dWs is a martingale.
Rt
Hs2 ds
0
1.9
It
o processes and It
os formula
(*)
Then
[X]t = [X0 + I + J]t = [I + J, I + J]t
= [I]t + 2[I, J]t + [J]t
Z t
s2 ds
=0+0+
0
In shorthand,
d[X]t = t2 dt.
Its easy to define integrals of an Ito process: Multiply both sides of (*) by another process t0 ,
t0 dXt = t0 t dt + t0 t dWt .
Then
s0 dXs
Z
:=
s0 s ds
Z
+
s0 s dWs .
It
os formula
Recall the fundamental theorem of calculus:
t
f 0 (s)ds.
f (t) f (0) =
0
f 0 (xs )dxs .
The stochastic version of the fundamental theorem of calculus is the Ito formula.
Proposition 1.25 Let f (t, x) be a smooth function. Let f(t, x) :=
2
f 00 (t, x) = xf2 (t, x). Then f (t, Wt ) is an It
o process satisfying
f
(t, x),
t
f 0 (t, x) :=
f (t, Wt ) f (0, W0 )
Z
Z t
Z t
1 t 00
0
=
f (s, Ws )ds +
f (s, Ws )dWs +
f (s, Ws )ds.
2 0
0
0
Proof (Sketch). Write down Taylor series expansion of f (around 0)
X f
X f
f (t, Wt ) = f (0, W0 ) +
si +
Wi
s
x
i
i
X f
1 X 2f
2
+
(s
)
+
si Wi
i
2 i s2
sx
i
1 X 2f
+
(Wi )2 + o((si )3 , (Wi )3 )
2 i x2
27
f
(t, x)
x
and
where si+1 := si+1 si , Wi+1 := Wsi+1 Wsi for some partition [si , si+1 ] of [0, t]. Let si ,
Wi 0 in above.
Itos formula is usually abbreviated to
1
df (t, Wt ) = (f(t, Wt ) + f 00 (t, Wt ))dt + f 0 (t, Wt )dWt .
2
If f (t, x) = f (x) this becomes
1
df (Wt ) = f 0 (Wt )dWt + f 00 (Wt )dt.
2
Examples 1.26
(ii) Calculate
Rt
0
1
Wt2 ...
2
Thus,
Wt2
= X t X0 =
dXs = 2
0
so in fact,
Rt
0
Ws dWs = 12 Wt2 21 t.
28
Ws dWs + t,
0
Multidimensional It
os formula for general It
o processes
Proposition 1.27 Let X = (Xt )t0 and Y = (Yt )t0 be two Ito processes, and let f be a smooth
function of two real variables. Then the stochastic process Zt := f (Xt , Yt ) is also an Ito process,
satisfying
dZt = df (Xt , Yt )
f
f
(Xt , Yt )dXt +
(Xt , Yt )dYt
=
x
y
1 2f
2f
+
(Xt , Yt )d[X, Y ]t
(X
,
Y
)d[X]
+
2
t
t
t
2 x2
xy
2f
+ 2 (Xt , Yt )d[Y ]t .
y
We can calculate the quadratic variation and covariation processes above as follows. Suppose that
we can write dXt and dYt in the form
dXt = t dt + t dWt ,
dYt = 0t dt + t0 dWt .
Then we can use the multiplication table
dWt
dt
dWt
dt
0
dt
0
0
to obtain
d[X]t = (dXt )2 = (t dt + t dWt )2 = t2 dt,
d[X, Y ]t = dXt dYt = (t dt + t dWt )(0t dt + t0 dWt ) = t t0 dt,
d[Y ]t = (dYt )2 = (0t dt + t0 dWt )2 = (t0 )2 dt.
Itos formula is just a Taylor expansion up to degree 2.
Example 1.28 Product rule: What is d(Xt Yt )?
Let f (x, y) = xy. Then
d(Xt Yt ) = df (Xt , Yt )
f
f
(Xt , Yt )dXt +
(Xt , Yt )dYt
=
x
y
2f
1 2f
2
+
(X
,
Y
)(dX
)
+
(Xt , Yt )dXt dYt
t
t
t
2 x2
xy
1 2f
+
(Xt , Yt )(dYt )2
2 y 2
= Yt dXt + Xt dYt + dXt dYt .
29
1.10
The movement of a drifting particle X (could be an asset price...), which is affected by some noise,
is frequently modeled by writing
Xt+t = Xt
+ (t, Xt )t + (t, Xt )Wt
= old position + drift
+ noise,
where Wt := Wt+t Wt , and t is very small.
As t 0, we write the limit of the above equation as a stochastic differential equation (SDE)
dXt = (t, Xt )dt + (t, Xt )dWt ,
X0 = x0 ,
meaning
Z
X t = x0 +
Z
(t, Xt )dt +
(t, Xt )dWt .
0
The coefficient is called the diffusion coefficient, while the coeff. is called the drift coeff.
[If = 0 this is just an ordinary differential equation (ODE):
dXt = (t, Xt )dt
or
dXt
= (t, Xt ).]
dt
, R.
dSt
dt
= St is St = S0 et .
Define Zt := ln St . In preparation for Itos formula, lets write Zt = f (St ), where f (s) := ln s,
f 0 (s) = 1/s, f 00 (s) = 1/s2 . Then
1
dZt = f 0 (St )dSt + f 00 (St )d[S]t
2
1
1 1
= (St dt + St dWt )
(St dt + St dWt )2
St
2 St2
1
= dt + dWt 2 dt
2
1 2
= ( )dt + dWt .
2
To solve this SDE, just integrate:
Z t
Z t
1 2
1
Zt Z0 =
( )ds +
dWs = ( 2 )t + Wt .
2
2
0
300
Therefore
1
ln St = ln S0 + ( 2 )t + Wt
2
Hence
St = S0 e( 2
2 )t+W
t
1.11
Rt
0
2
W
Theorem 1.30 Let Wt be a BM, and let Mt be a continuous
R t Ft -martingale such
R tthat2 EP [Mt ] < .
Then there exists an adapted process Ht such that Mt = 0 Hs dWs , and [M ]t = 0 Hs ds < .
The brief version of this is dMt = Ht dWt and d[M ]t = (dMt )2 = Ht2 (dWt )2 = Ht2 dt.
Rt
Integrals of the form 0 Xs dMs are just Ito integrals w.r.t. B.M.
Z
Z
Xs dMs =
Xs Hs dWs
In particular, if Xt =
Rt
0
Example1.31 Consider
the martingale
Mt = exp Wt 12 2 t (see tutorial). How to represent this as a stochastic integral?
Apply Itos formula to f (t, x) = exp x 21 2 t
dMt = df (t, Wt )
1 2
1 2
1 2
1
1
= eWt 2 t dWt 2 eWt 2 t dt + 2 eWt 2 t dt
2
2
= Mt dWt .
and XT = X.
31
Using the martingale representation theorem, we see that there exists an adapted process ht such
that
Z t
Xt = X 0 +
hs dWs .
0
Now letting t = T ,
Z
X = EP [X] +
hs dWs .
0
Rt
Later on, h represents a trading strategy, and 0 hs dWs is the gains from trading the asset W , so
any financial payoff X can be synthesized by a trading strategy (the market is complete). This is
not quite right however, because our asset does not follow a Brownian motion!
1.12
Change of Measure
EP [X1Ai ] =
i=1
Q(Ai ).
i=1
dQ
dP
Pn
i=1
n
X
dQ
EQ [Y ] :=
ai Q(Ai ) =
ai E P
1A
dP i
i=1
i=1
"
#
n
dQ X
dQ
= EP
ai 1Ai = EP
Y
dP i=1
dP
The M.C.T. does the rest: If Yn are simple r.v.s such that Yn Y then
dQ
dQ
EQ [Y ] = lim EQ [Yn ] = lim EP
Yn = EP
Y .
n
n
dP
dP
33
Now
Xt is a Q-martingale
EQ [Xt |Fs ] = Xs t > s
EQ [Xt 1A ] = EQ [Xs 1A ]
A Fs
EP [Zt Xt 1A ] = EP [Zs Xs 1A ]
A Fs
EP [Zt Xt |Fs ] = Zs Xs t > s
Zt Xt is a P-martingale.
Let Wt be a P-B.M., let Ft be the filtration generated by Wt , fix a time T (0, ). Suppose that
RT
t is an Ft -adapted process such that 0 s2 ds < a.s.
R
t = dWt + t dt). If t is non-zero then W
t
t := Wt + t s ds (equivalently, dW
Define the process W
0
has a non-zero drift.
Theorem 1.37 (Girsanov Theorem) If
Z t
Z
1 t 2
Zt := exp
s dWs
ds
2 0 s
0
t is a
is a P-martingale then there is a probability measure Q P defined on FT , such that W
dQ
Q-Brownian motion, and dP = ZT .
Proof. We only give the proof for = constant, in which case Zt is automatically a martingale. Since
= ZT . From
Z0 = 1, we have EP [ZT ] = 1. Define Q by Q(A) := EP [1A ZT ] (for A F ), so dQ
dP
Lemma 1.36,
1
2
t t is a Q-martingale
exp W
2
1
2
t t is a P-martingale
Zt exp W
2
1 2
1 2
exp Wt t exp Wt + t t
2
2
is a P-martingale
1
2
exp ( )Wt ( ) t
2
is a P-martingale.
t is a Q-BM.
Since Wt is a P-BM, the above statements are true, and hence (see Tutorials) W
Remark 1.38 A sufficient condition for Zt to be a P-martingale is Novikovs condition
Z T
1
2
EP exp
ds
< .
2 0 s
Theorem 1.39 (Girsanov converse) If WRt is a P-Brownian motion, and Q P, then there exists
t := Wt + t s ds is a Q-Brownian motion. Moreover,
an adapted process t such that W
0
Z T
Z
dQ
1 T 2
= exp
s dWs
ds .
dP
2 0 s
0
34
Proof. The process Zt := EP dQ
|Ft is a positive martingale with Z0 = 1. Using the MRT, there
dP
exists an adapted process Ht such that dZt = Ht dWt . Define t = Ht /Zt , so that dZt = t Zt dWt .
Define Yt = ln(Zt ), so Y0 = 0. By Itos formula,
1
1 1
1
dZt
(dZt )2 = t dWt t2 dt.
2
Zt
2 Zt
2
Rt
Rt
Integrating this gives Yt = 0 s dWs 12 0 s2 ds. Hence
dYt =
Zt = eYt = e
Note that ZT =
dQ
,
dP
Rt
0
s dWs 21
Rt
0
s2 ds
t is a Q-B.M.
so Q is the measure Girsanovs Theorem, and W
35