Professional Documents
Culture Documents
19 24
2001 Elsevier Science B.V. All rights reserved.
Diffusion Processes
S. R. S. V a r a d h a n
1. Brownian motion
j=l 2([j -- [j l ) J
853
854 S. R. S. Varadhan
1. If we denote by H~,c,rl,r2 the set of functions x(.) that satisfy a local H61der
condition of the form
Ix(~) - x(t)l _< c l t - ~1~
for T~ < s < t < T2 then for every ~ > 1/2, T1 < T2 and C < oo
P0[H~,c,~,,~2] = 0 .
2. On the other hand for every ~ < 1/2, and T1 < T2 < oc,
lim P0[H~cr~ r2] = 1 .
C._+o o ~ ~
P0 [kim
--'~1 < ~ J~t]( x ( J2k
- - ' ) t-ix- 1 ) ~) - 2= t for a l l t _ > 0 ] = 1 (1.2)
It is clear from these properties that a typical Brownian path, being nowhere
differentiable, is not of bounded variation in any finite interval. However it has a
definite 'quadratic variation' in any interval that is almost surely equal to the
length of the interval.
One of the ways of understanding the behavior of a Brownian path is to start
with a random walk {S, : n _> 1} which, for each n, is the sum
&=xl+&+...+x.
of the first n terms of a sequence {X/} of independent identically distributed
random variables that take the value :kl with probability 1/2 each. While the
central limit theorem asserts that for large n, the distribution S,,/v~ is asymp-
totically the standard normal, Brownian motion is to be thought of as the limit in
distribution of S[,t]/x/~ as a function of t.
The Wiener measure P0 or the Brownian motion x(.) is not only the prime
example of a diffusion process but it is also the building block out of which other
diffusion processes are constructed. A discussion of Brownian motion and its
properties can be found, among other places, in the following books (Stroock,
1993; Revuz and Yor, 1999). Usually one refers to the measure P0 as the Wiener
measure and a random path x(.) that is 'distributed' according to the Wiener
measure as Brownian motion.
Brownian motion derives its name from the botanist Robert Brown who
observed in 1828, that, in water, pollen from plants dispersed in an irregular
swarming motion. From that time many scientists have examined the Brownian
phenomenon including Einstein. Norbert Wiener was the first one who essentially
constructed the measure corresponding to Brownian motion, and proved the
Diffusion processes 855
H61der continuity of it paths in Wiener (1923) and Wiener (1924). It was the first
rigorous example of integration in an infinite dimensional function space.
Just as any multivariate normal random vector can be generated by a linear
transformation y = Tx from one with covariance equal to the Identity matrix, one
can start with any orthonormal basis {fj(.)} in L2[0, oc] and represent the
Brownian motion as
x(t) = zj (t)
J
where {Zj} are independent standard normal random variables and
t(
Fj(t) :
f s)ds .
While formally the expansion is always valid its convergence is far more delicate.
Such expansions have been considered for Brownian motion and more generally
for Gaussian processes in It6 and Nisio (1968).
p(t,X) = --~7
1 exp [-x2
~1 (2.2)
= fA q(S,X(S), t,y)dy
able. Examples of stopping times are the first time some thing happens, for
instance
r(co) = inf[t : x(t) >_ a]
In order to determine if r(co) _< t, we only need to observe the path up to time t.
On the other hand, the last time some thing happens like
are not stopping times. An important fact is that for Brownian motion as well as
for most other diffusion processes, the Markov property extends to stopping
times and this is called the strong Markov property and was first considered by
Hunt (1956).
Associated with any stopping time r there is a natural a-field ffz defined by
Roughly speaking if we observe the process only up to time z and then stop, then
~ , represents the information that is available. The strong Markov property for
Brownian motion, first established by G. A. Hunt [H], states that with respect to
any Px0, the future increments y~(t) = x(~ + t) - x(z) is again a Brownian motion,
i.e. distributed according to P0 and more over is stochastically independent of the
events in Y~. This can be restated in the form
yield
for all f and s, t > 0. Such semigroups are characterized by their infinitesimal
generators
A dTt
dt t=0
For our Brownian motion semigroup it is an easy calculation to derive that for
any smooth function f
1 d2f(x)
( A f ) ( x ) = lim -1 [ [ f ( x + y) - f ( x ) ] p ( t , y ) d y - 2 dx 2
t-*O t J
and consequently
,Zu - ~xx - - f
are solved by
4. Stochastic integrals
Since x(t) is almost surely a continuous function of t we can define
Y = fo f(t)x(t)dt for smooth functions f with compact support and will give us a
Gaussian random variable with mean 0 and variance fff(s)f(t)s/~ t ds dt. In
fact, i f f is a smooth function we can even define ff(t)dx(t) as -ff'(t)x(t)dt after
an integration by part and calculate
E
f/ f(t)dx(t) = f'(s)f'(t)s A t ds dt =
/ If(t)] 2 dt
By completion in L2[P], we can now define ff(t)dx(t) for functions f in L2[0, cc]
and will give us a mean zero normal random variable with variance [[f[]~. This
was carried out already by Wiener and yields the following result known as the
Cameron Martin formula [see Cameron and Martin (1953)].
Let F(t) be of the form F(t) = fof(s)ds for some f in L2[0, oc]. Let x(.) be
Brownian Motion distributed according to the Wiener measure P0. Then the
distribution PF of y(.) = x(.) + F(.) is absolutely continuous with respect to P0
and has the Radon-Nikodym derivative
S = Zf(tj)(x(tj+i ) -- x(tj))
J
if we let f(t) = f(t, co) depend on co, but insist that for every t, f(t, co) be mea-
surable with respect to g t then the increment x(tj+l)- x(tj) is independent of
f(tj, co). One still gets
Diffusion processes 859
F__ 7
/-- /
leading to It6's definition of the stochastic integral for the class of 'progressively
measurable' functions f(t, co) satisfying
t, co :dt <oc
Since we do not necessarily have to integrate all the way to ec, it suffices that f is
progressively measurable and satisfies
is well defined under these conditions and has the following properties.
(1) The process y(t) is almost surely continuous in t and is a progressively
measurable martingale relative to fit.
(2) If ~ is any stopping time, y(~ A t) is also a martingale and
y(~ A t) =
/0 {l~>s(co)f(s, co)]dx(s) =
/0 f~(s, co)dx(s) .
[
exp kY(t) - 5
is again a martingale.
5. It6 calculus
t
The stochastic integral y(t) = fof(s, co)dx(s) can be thought of as dy = f d x and
the first step in developing any calculus is the chain rule. What is d~b(y(t))? The
problem exists already if f = 1 i.e. what is d~b(x(t))? If we take ~b(x) = x 2 one
might think that dxZ(t) = 2x(t)dx(t) or
This is clearly wrong because the left hand side has expectation t while the right
hand side has expectation 0. A better guess is
This can indeed be seen to be correct by using (1.2). More generally It6's formula
asserts that
f(x(t)) - f(x(O)) =
i0'f'(x(s))dx(s) + ~ fn(x(s))ds
Since x(-) is not of bounded variation but has finite quadratic variation, formally
(dx) 2 = dt and (dx)i(dt) j = 0 if i + 2j _> 3. We have to take an additional term in
the Taylor expansion. We can now compute
1 9
dO(t,y(t)) = Otdt + ~ydy + 1~byy(dy)- = [qSt+ 0yyf2] dt + ~yf dx
or
,b(t,y(t)) - +(O,y(O))
t 1 2
+
/0'
In particular if f = 1
t 1
O(t,x(t)) - q$(O,x(O))= io [~(s,x(s)) +~dPxx(S,X(S)))ds
+
/o'
One can use It6's formula to establish a connection with the infinitesimal
generator 1/2D~. For example if u(t,x) solves
Ht z D~u
with u(O,x) = f ( x ) , It6's formula for v(t,x) = u(T - t,x) yields
dv(t,x(t)) = // v~(s,x(s))dx(s)
ut -[- ~Dxul
2 -L g(x, t)u = 0 in [0, T] x [a, b]
1 [ (y-x-#(t-s))21
q~(s,x, t,y) - v/2~z( t _ s) exp - ~2~--_s i
Symbolically dy = dx + pdt and there are analogous results. But the basic
difference is that the infinitesimal generator 1/2D 2 has now to be replaced by
1/2D 2 + #Dx.
Often, in considering stochastic models, we come across situations where the
drift is a restoring force that is proportional to the displacement. Formally
dy = - # y dt + dx . (6.1)
Such an example will not be a process with independent increments, but will still
be Markov. We can solve (6.1) explicitly with initial condition y(0) = y, to obtain
From the representation (6.2), we see that given y(O) = y, y(t) is again normally
distributed, but now with mean e-~ty and variance 1/2#(1 -e-2~t). We there-
fore have a time homogeneous Markov process with transition probability
density
As t ~ oc this has a limit which is a stationary density for the process and is in
fact the normal distribution with mean 0 and variance 1/2#. For this Markov
process, the generator is 1/2D 2 -#yDy. The way to compute the generator
quickly is by applying It6's formula
/ 1 ,
du(y(t)) = u'dy(t) + u"(dy(t)) 2 = u'[-#ydtJ + u dx + 5u dt
=[~D2u-ktyDyu]dt+Dyudx
The generator is given by the coefficient of the dt term. More generally we can
have models of the form dy = b(y)dt + dx, written as
where 0,co is the new Brownian path x(~r) - x(s) for a > s. This guarantees that
y(t) is still a Markov process and It6's formula computes the generator as
1/2D 2 + b(y)Dy. The transition probability is no longer explicit. It is of course
given by
t 1 t ds 1
e x p l f 0 b(x(s))dx(s)-~ fo [b(x(s))]2
A
7. One-dimensional diffusions
y(t) = x +
ft b(v,y(v))dv + a(v,y(v))dx(v)
u 1 ~2u Ou
~s ~-~a(s,x) bT~2 + b(~, x) = 0 (7.2)
with the boundary condition u(t,x)=f(x). The relation between a,b and
p(.,-,., .) can also expressed by
Diffusion processes 865
lira
t--+s t 1 s f ( y -- x)4p(s, x, t, dy) 0
In fact this was the way Kolmogorov [see Kolmogorov (1931)], formulated the
problem of constructing the transition probabilities from given coefficients.
the existence uniqueness o f p can be established. See for instance Friedman (1964)
for the details. From p one constructs the Markov process in a canonical manner
via finite dimensional distributions. The almost sure continuity of the paths
follows from estimates on p that are obtained along the way.
If we drop the lower bound a > c > 0 but strengthen the smoothness
assumptions by demanding additional regularity on a and b as functions its
variables of s and x, one can prove the existence of smooth solutions u for
the Equation (7.2), provided the boundary data f is smooth. The value of the
solution u(s,x) can be shown for each fixed t to be a nonnegative bounded linear
functional o f f and a suitable version of Riesz representation theorem will give
one should, that if more than one method works in a given situation they lead to
the same measures Qs#.
c. Martingale formulation
No matter how the measures are constructed it is always true that for any
smooth f
Here ~s is the space of continuous maps co = x(-) from Is, oo] into R. ~t~s is the
a-field generated by {x(v) :s < v < t}. And
f0 ~ V(x(s)--~
1 ds = t
and use the stopping times to change the time scale of the process through the
following map ~ : f2 -~ ~2
a (co)(t) = x( t(co), co)
If Q is a solution to the martingale problem for [a(x), b(x)] then Q#7 ~ solves the
martingale problem for [V(x)a(x), V(x)b(x)]. One can therefore transfer existence
and uniqueness results from one to the other. For the matingale formulation
Stroock and Varadhan (1997) is a good source.
In the one dimensional case there are some special results. For instance in the
Stochastic Differential Equations formulation the Lipschitz condition can be
weakened to a H t l d e r condition in x of exponent 1/2. Using a combination of
random time change and Girsanov's formula one can reduce the case of any
Diffusion processes 867
bounded [a, b] with a positive lower bound for a to the case o f a - 1, b - 0, which
is the case of Brownian motion and enjoys both existence and uniqueness. This
gives a way of directly constructing such processes from Brownian motion. These
results can be found in Dynkin (1959) or Stroock and Varadhan (1997) in full
detail.
d. Limit theorems
Suppose that we have a Markov chain with transition probability rch(x, dy) and
this represents a transition in a time step of duration h. We can construct a
piecewise constant stochastic process Xh(t) by defining first, a Markov chain ~hn
with transition probability ~Zh and then making a stochastic process out of it by
defining,
and
lim
h~O - 1f lY - xl2+6zch(x, dy) = 0
hold locally uniformly in x for some bounded continuous functions [a(x), b(x)]
and that there is unique family Qx that is a solution to the martingale problem for
these coefficients that start at time 0 from the point x, then it follows that
lim Qh,x = Qx
h-+O
in the sense of weak convergence. The basic idea in this approach is that with ~zhf
defined as f f (y)~zh(x, dy),
n-1
f(Xh(nh)) -- f(Xh(O)) -- ~(TChf -- f)(Xh((n -- 1)h))
j--O
is a martingale with respect to (Q, ~nh, Qh,x) and under our assumptions
1
h~olim~ (~zhf - f ) = Lf
868 S. R. S. Varadhan
To start with, we note that one can define Brownian motion in d dimensions by
just taking d independent Brownian motions and making them the d components
of a vector valued process {xj(.)}. We get a Markov process (actually independent
increments) with transition probability density given by
and the transition probability density itself is the fundamental solution of the heat
equation
at 2 ;=1 ~x2
For our Brownian motion we have taken the identity J as the covariance
matrix. We can instead take any positive definite matrix A for the covariance and
if A = o-a* with representing the adjoint operation, we can represent the new
process as the linear transform y(t) --ax(t). The new transition probabilities
will be
1 1 [ ((y-x),A-l(y-x)}.l
p(s,x, t,y) = [2~(t - s)y/2 [det A] 1/2 exp - 2(t - s)
i ij~l 82
L = ~..= aij axi---~xj
We can add a constant drift vector vector b and then the corresponding
transition probability density and generator will be given respectively by
Diffusion processes 869
1 1
p(s,x,t,y) =
[2n(t - s)]d/2 Idet A] 1/2
exp[-{(Y- x - b(t- s))'A-l(Y--2(t
- s) x - b(t- s))).1
1~--~ ~2 b ~
L =-~ i~7..lai,j~
,
+ Z,] J~xj
1 d ~2
L = ~. "= a i , j ( t , x ) ~ + Zj bj(t,x) Oxj
and can be constructed by any of the methods outlined in the one dimensional
case. The essential intuition is that we have a vector valued process x(t) such that
the conditional distribution of x(t+h)-x(t), given the past history
{ x ( s ) : 0 < s < t}, is approximately normally distributed in R d with mean
hb(t,x(t)) and covariance ha(t,x(t)). Here b(t,x)= {bj(t,x)} is an R d valued
function on [0, eel x R d while a(t, x) = aij(t, x) is a symmetric positive semidefinite
d x d matrix for each (t, x) E [0, oe) x R d.
Just as in the one dimensional case if we assume that a(t,x) can be written as
a(t,x)a(t, x)* and that both a(t,x) and b(t,x) satisfy a Lipschitz condition of the
form
with a constant C that does not depend on t, then we can solve the stochastic
differential equation
bj(., .) are uniformly bounded by some constant C, and that a(-,.) and b(.,.)
satisfy a H61der condition
for some exponent ~ and constant C, then just as in the one dimensional case we
can get a fundamental solution p(s, x, t,y) that can serve as the transition proba-
bility density for the diffusion. The heat operator is of course replaced by
alas + L~ where
By the continuity of paths one can see that Ix(z)[ = 1 and it is not difficult to see
that Px[CO : r(co) < oo] = 1 for x ~ B. Then for any continuous data f on Ixl = 1
.(x) = E
8x? = 0
J J
in B and for b E ~B,
If we try to construct the one dimensional Brownian motion on the half line
rather than the full line we run into trouble. The Brownian path eventually
reaches 0 and wants to get out to the negative side. We must do some thing to
keep it nonnegative. We may decide to give it enough of a kick to keep it non-
negative. This is easier to see in the discrete setting of a symmetric random walk
or a gambler's ruin situation. Every time the gambler loses his entire fortune (i.e.
reaches 0) we provide him with enough to make one more bet (i.e. move him from
0 to 1). He may lose it again in which case we make another contribution. Every
time the random walk comes to zero, at the next step it is moved to 1.
The continuous analog is the set of relations y(t) = x(t) + F(t), F(t) is non-
decreasing and continuous in t, y(t) >_ 0 and x(t) is a given Brownian path. We are
interested in the minimal F that achieves this. Alternately F is allowed to increase
only when y(t) = 0. It turns out that such a pair of function y(t), F(t) exist and is
unique for every continuous path x(t) and in fact are given explicitly by
u(t,y) = EV(y(t))ly(0) = y]
u 1 O2U
a t - 2 a y 2 on[O, oo) x[O, oc)
References
Cameron, R. H. and W. T. Martin (1953). The transformation of Wiener integrals by non linear
transformations. Trans. Amer. Math. Soc. 75, 552-575.
Dynkin, E. B. (1959). One dimensional continuous strong Markov processes. Theor. Prob. Appl.
4, 3-54.
Ethier, S. N, and T. G. Kurtz (1986). Markovprocesses. Characterization and convergence. Wiley Series
in Probability and Mathematical Statistics: John Wiley & Sons, Inc.
Feller, W. (1957). Generalized second order differential operators and their lateral conditions. Illinois
J. Math. 1, 495-504.
Friedman, A. (1964). A Partial Differential Equations of Parabolic Type. Prentice Hall, New Jersey.
Hunt, G. A. (1956). Some theorems concerning Brownian motion. Trans. Amer. Math. Soc. 81,
294-319.
It6, K. (1942). Differential equations determining a Markoffprocess (original Japanese title: Zenkoku
Sizyo Sugaku Danwakai-si). J. Pan-Japan Math. Coll. 1077.
It6, K. and M. Nisio (1968). On the Convergence of Sums of Independent Banach Space Valued
Random Variables, Osaka. J. Math. 5, 35-48.
Kolmogorov, A. N. (1931). Uber die Analytischen Methoden in der Wahrschienlichtkeitsrechnung.
Math. Ann. 104, 415-458.
McKean, H. P. (1969). Stochastic Integrals. Academic Press.
Revuz, D. and M. Yor (1999). Continuous Martingale andBrownian Motion. 3rd edn., Springer-Verlag.
Stroock, D. W. (1993). Probability Theory. An Analytic View. Cambridge University Press.
Stroock, D. W. and S. R. S. Varadhan (1997). Multidimensional Diffusion Processes. 2nd edn.,
Springer-Verlag.
Stroock, D. W. and S. R. S. Varadhan (1969). Diffusion Processes with continuous coefficients, I, II.
Comm. Pure Appl. Math. XXII, 345 400, 479 530.
Stroock, D. W. and S. R. S. Varadhan (1971). Diffusion Processes with boundary conditions. Comm.
Pure Appl. Math. XXIV, 147225.
Wentzel, A. D. (1959). On the boundary conditions for multidimensional diffusion processes. Theor.
Prob. Appl. 4, 164-177.
Wiener, N. (1923). N. Differential-Space. J. Math. Phys. 2, 131-174.
Wiener, N. (1924). The average value of a functional. Proc. Lon. Math. Soc. 22, 454-467.