Autocorrelation function and the Wiener

I.
BASIC CONCEPTS AND DEFINITIONS
Autocorrelation function and the Wiener-Khinchin theorem
Consider a time series x(t) (signal). Assuming that this signal is known over an infinitely
long interval [−T, T ], with T → ∞, we can build the following function
1ZT
dtx(t)x(t + τ ),
T →∞ T 0
G(τ ) = lim
(1)
known as the autocorrelation function of the signal x(t) (ACF).
ACF is an even function
Proof:
n
o
1ZT
G(−τ ) = lim
dtx(t)x(t − τ ) = t − τ = y|T−τ−τ , dt = dy
(2)
T →∞ T 0
"
#
Z T
Z T
1 Z0
1 Z T −τ
= lim
dyx(τ + y)x(y) = lim
(. . .) +
(. . .) −
(. . .) = G(τ )
T →∞ T
T →∞ T −τ
−τ
0
T −τ
The last equality holds in the limit T → ∞.
The wiener-Khinchin theorem
This theorem plays a central role in the stochastic series analysis, since it relates the
Fourier transform of x(t) to the ACF.
Introduce a new function S(ω) according to
1
| x̂(ω) |2 ,
T →∞ 2πT
S(ω) = lim
(3)
where the forward Fourier transform of x(t) is given by
Z
T
x̂(ω) =
dte−iωt x(t).
(4)
0
Note that if for real signal x(t), the Fourier transform obeys the following symmetry
x̂(−ω) = ω̂ ∗
(5)
The function S(ω) is called the power spectral density of x(t) (psd).
The following lines relate S(ω) to G(τ ) (see Fig.1 for details)
Z
T
1
dte−iωt x(t)
T →∞ 2πT 0
= {(t, t0 ) → (t0 , τ = t − t0 )}
Z
S(ω) = lim
T
0
dt0 eiωt x∗ (t0 )
0
(6)
1
τ
t
000000000000
000000000000
111111111111
T111111111111
000000000000
111111111111
000000000000
111111111111
T
0
1111111111
0000000000
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
0000000000
1111111111
111111111111
000000000000
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
000000000000
111111111111
T
0
t’
T
t’
−T
FIG. 1:
Transformation of the integration domain under the change of variables
{(t, t0 ) → (t0 , τ = t − t0 )}.
"
Z T −τ
Z 0
Z T
1 ZT
−iωτ
0
0 ∗ 0
−iωτ
= lim
dτ e
dt x(t )x (t + τ ) +
dτ e
dt0 x(t0 )x∗ (t0 + τ )
T →∞ 2πT
0
0
−T
−τ
"
#
Z T
1 Z0
−iωτ
−iωτ
=
dτ e
(T G(τ ) + O(1/T )) +
dτ e
(T G(τ ) + O(1/T ))
2πT −T
0
Z T
1
=
lim
dτ e−iωτ G(τ ).
2π T →∞ −T
#
Because the ACF is an even function, we can write
1 Z∞
1 Z∞
1Z∞
±iωτ
S(ω) =
G(τ )e
dτ =
G(τ ) cos ωτ dτ =
G(τ ) cos ωτ dτ
2π −∞
2π −∞
π 0
(7)
Consequently, S(−ω) = S(ω) and using the following representation of the delta-function
δ(x) =
1 Z ∞ ±iωx
e
dω,
2π −∞
(8)
we conclude that
Z
G(τ ) =
∞
−∞
eiωτ S(ω)dω.
(9)
What happens in case of a deterministic signal?
Let x(t) = A cos Ωt, then the Fourier transform
Z
T
x̂(ω) =
0
T
dte−iωt A cos Ωt = A δΩ,ω + O(1).
2
2
(10)
1.5
1.5
(a)
1
1
S(ω)
0.5
G(τ)
(b)
0
deterministic
stochastic
0.5
-0.5
-1
0
-1.5
-10
-5
0
τ
5
10
0
2
ω
4
FIG. 2: (a) ACF G(τ ) of a deterministic (dashed line) and a stochastic (solid line) signals. (b)
The corresponding psd S(ω).
The psd
1
A2 T 2
x̂(ω)x̂∗ (ω) = lim
δΩ,ω + O(1) = O(T )δΩ,ω .
T →∞ 2πT
T →∞ 8π
S(ω) = lim
(11)
Consequently, S(ω = Ω) diverges as T → ∞.
The ACF is found as
A2 Z T
dt cos Ωt cos Ω(t + τ )
(12)
G(τ ) = lim
T →∞ T
0
A2 Z T
A2 cos Ωτ
= lim
+ O(1/T )
dt cos2 Ωt cos Ωτ − sin Ωt cos Ωt sin Ωτ =
T →∞ T
2
0
Stochastic vs deterministic.
The ACF and the psd functions can be used to distinguish between stochastic and deterministic signals. Thus, the ACF of a deterministic signal is periodic, whereas the ACF
of a stochastic signal typically decays exponentially with τ (see Fig. 2(a)). The power spectral function S(ω) of a deterministic signal consists of delta peaks, whereas the psd of a
stochastic signal typically shows a power law decay (Fig. 2(b)).
Stochastic processes
Loosely speaking, any stochastic process is a probabilistic time series. Consider a system,
whose dynamics is described by a certain time-dependent random variable x(t).
Introduce the joint probability to observe the values of x1 , x2 , . . . at the respective times
t1 > t2 > . . .
p(x1 , t1 ; x2 , t2 ; . . .)
3
(13)
For any two moments of time t1 > t2 we can define the conditional probability
p(x1 , t1 |x2 , t2 ) =
p(x1 , t1 ; x2 , t2 )
p(x2 , t2 )
(14)
Note that
Z
p(x1 , t1 ) =
Ω
dx2 p(x1 , t1 ; x2 , t2 ),
(15)
where we integrate over the entire probability space Ω.
Ensemble average
The integration over x, as in the above equation, is associated with the averaging over the
ensemble of different realizations of the stochastic process x(t). We will denote the ensemble
average by h. . .i. Thus, the conditional time-dependent average of x is given by
hx(t)|x0 , t0 i =
Z
dx x p(x, t|x0 , t0 ).
(16)
Similarly, the conditional (non-stationary) ACF can be computed as
hx(t)x(t0 )|x0 , t0 i =
Z
dx dx0 x x0 p(x, t; x0 , t0 |x0 , t0 ) =
Z
dx dx0 x x0 p(x, t|x0 , t0 )p(x0 , t0 |x0 , t0 ).
(17)
The last equality only holds for Markovian processes (see below).
Markov processes:
For any t1 > t2 > t3 > t4 , the probability at times t1 and t2 only conditionally depends
on the state at time t3 , i.e.
p(x1 , t1 ; x2 , t2 |x3 , t3 ; x4 , t4 ) = p(x1 , t1 ; x2 , t2 |x3 , t3 ).
(18)
As a consequence of this property, we have
p(x1 , t1 ; x2 , t2 |x3 , t3 ) = p(x1 , t1 |x2 , t2 )p(x2 , t2 |x3 , t3 ).
(19)
Indeed
p(x1 , t1 |x2 , t2 )p(x2 , t2 |x3 , t3 ) = p(x1 , t1 |x2 , t2 ; x3 , t3 )p(x2 , t2 |x3 , t3 )
p(x1 , t1 ; x2 , t2 ; x3 , t3 ) p(x2 , t2 ; x3 , t3 )
=
p(x2 , t2 ; x3 , t3 )
p(x3 , t3 )
= p(x1 , t1 ; x2 , t2 |x3 , t3 ).
This proves the earlier result for non-stationary ACF Eq. (17).
4
(20)
The Chapman-Kolmogorov equation
From Eq. (19), one easily obtains
Z
p(x1 , t1 ; x2 , t2 ; x3 , t3 ) Z
p(x1 , t1 |x3 , t3 ) = dx2
= dx2 p(x1 , t1 |x2 , t2 )p(x2 , t2 |x3 , t3 ). (21)
p(x3 , t3 )
Ω
Ω
the relation between the transitional probabilities (1|3) and (1|2), (2|3) is known as the
Chapman-Kolmogorov equation.
Stationary processes
Process x(t) is called stationary if for any , x(t + ) has the same statistics as x(t).
Important properties of a stationary process:
hx(t)i = const
hx(t)x(t0 )i = f (t − t0 ).
(22)
Stationary ACF can be obtained from Eq. (17) by assuming that the initial moment of
time t0 is in the remote infinity
G(τ ) =
=
lim hx(t)x(t0 )|x0 , t0 i
t0 →−∞
lim
t0 →−∞
Z
=
Z
(23)
xx0 dx dx0 p(x, t; x0 , t0 |x0 , t0 ) = lim
Z
t0 →−∞
xx0 dx dx0 p(x, t|x0 , t0 )p(x0 , t0 |x0 , t0 )
xx0 dx dx0 p(x, t|x0 , t0 )ps (x0 )
(24)
Example: The random telegraph process.
Ergodic processes
For an ergodic process, the averaging over time is equivalent to the averaging over the
ensemble. Note that ergodicity is stronger than stationarity. Example:
x(t) = A, A is uniformly distributed in [0; 1]
(25)
Then, any realization is a straight line x(t) = Ai , but Ai are different for different realizations,
implying that hx(t)i = 0.5, whereas the average over time for every single realization is given
by
R
dt x(t) = Ai .
Computation of Power spectral density
For stationary processes, the function S(ω) can be computed by taking the ensemble
average of the square of the Fourier transform of x(t). Thus we have
hx̂(ω)x̂∗ (ω 0 )i =
Z
0 0
dt dt0 e−iωt eiω t hx(t)x(t0 )i,
5
(26)
where the integrals are taken in the limit of infinitely large intervals T → ∞. Because for a
stationary process
hx(t)x(t0 )i = G(t − t0 ),
(27)
consequently
hx̂(ω)x̂∗ (ω 0 )i =
=
Z
Z
0 0
dt dt0 e−iωt eiω t G(t − t0 ) = {(t, t0 ) → (t0 , τ = t − t0 )}
0
0
dt0 ei(ω −ω)t
Z
= 2π
0
Z
(28)
0
dτ e−iω τ G(τ )
0
dt0 ei(ω −ω)t S(ω 0 ) = (2π)2 S(ω 0 )δ(ω − ω 0 ).
White noise and stochastic differential equations
The above result allows us to compute S(ω) for systems, described by linear stochastic
differential equations. Any such equation can be written in the form
ẋ = f (x) + Dξ(t),
(29)
where x is a vector variable, which characterizes the state of a system, f (x) is a linear
function, D is the noise strength (intensity), and, ξ(t) is the so called white noise. It can be
introduced as a completely uncorrelated time series with the ACF given
hξ(t)ξ(t0 )i = δ(t − t0 ).
(30)
Now it is easy to compute S(ω) for x(t) in Eq. (29). To this end, one needs to take the
Fourier transform of both sides in Eq. (29) and use Eq. (28).
Examples
We will compute S(ω) for the following standard processes:
(1): The one-dimensional Wiener process is described by the equation
ẋ = ξ(t).
(31)
ˆ
iωx̂(ω) = ξ(ω),
(32)
Thus, we have in the Fourier space
Consequently
hx̂(ω)x̂∗ (ω 0 )i =
6
ˆ ξˆ∗ (ω 0 )i
hξ(ω)
(iω)(−iω 0 )
(33)
But since by the definition of the white noise
ˆ ξˆ∗ (ω 0 )i = (2π)2
hξ(ω)
1 Z
Gξ (τ )dτ e−iωτ δ(ω − ω 0 ) = (2π)δ(ω − ω 0 ),
2π
(34)
we obtain
2πδ(ω − ω 0 )
hx̂(ω)x̂ (ω )i =
= (2π)2 S(ω)δ(ω − ω 0 )
2
ω
1
1
S(ω) =
.
2π ω 2
∗
0
(35)
If we formally try to compute the stationary ACF G(τ ) using the Wiener-Khinchin theorem,
we obtain a contradictory result
Z
G(τ ) =
∞
1 iωτ
e dω ∼ τ.
2πω 2
−∞
(36)
The contradiction comes to surface if we now try to compute S(ω) using the G(τ ). Clearly,
the function f (x) = x is not a square-integrable on the interval (−∞, ∞), implying that
its Fourier transform does not exist. Physically, it means that the Wiener process is not a
stationary process, as we will see in more details later.
(2): Overdamped particle, the Ornstein-Uhlenbeck process The equation of
motion is
ẋ = −αx + Dξ(t).
(37)
Proceeding as before, we obtain
hx̂(ω)x̂∗ (ω 0 )i =
D2 (2π)δ(ω − ω 0 )
,
α2 + ω 2
(38)
Consequently
S(ω) =
D2
.
2π(α2 + ω 2 )
(39)
This type of the power spectrum is known as the Lorentzian. It appears in many applications,
such as: chemical reactions, bistable systems, random telegraph process (details later). The
stationary ACF does exist and is given by
Z
G(τ ) =
∞
−∞
D2 −ατ
D2
iωτ
e
dω
=
e .
2π(α2 + ω 2 )
2α
The last result shows that hx2 i = G(0) = D2 /2α.
7
(40)
1
3
10
G(t)
0.5
2
10
SVx ( ω/ωc )
1
10
0
-0.5
0
0
0.1
10
-1
10
10
20
t
30
40
50
0.01
-2
10
0.001
-3
10
0
1
2
3
ω/ωc
4
5
FIG. 3: Psd of the 2D plasma for different γ as in the legend and kT /m = ωc = 1. Inset: the ACF
function for γ = 0.1.
(3): 2D plasma in magnetic field
Consider a charged particle, confined to move on a plane in a constant magnetic field B,
which is perpendicular to the plane of motion. The equation of motion of a single particle is
ṙ = v
s
kT
q
v × B − γv + 2γ
ξ(t),
v̇ =
m
m
(41)
where γ is the damping coefficient and T is the absolute temperature. We are interested in
the power spectrum of the x-component of the velocity vx . On the ground of the symmetry
reasons, it is clear that Svx (ω) = Svy (ω). The scalar form of the equations of motion is
qB
v˙x =
vy − γvx +
m
s
qB
v˙y = − vx − γvy +
m
kT
ξx (t),
m
2γ
s
kT
ξy (t).
m
2γ
(42)
In what follows we assume that the sources of the white noise ξx and ξy are uncorrelated,
implying that hξx (t)ξy (t0 )i = 0. Taking the Fourier transform, we obtain
qB
iωv̂x =
v̂y − γv̂x +
m
s
qB
iω ŷ = − v̂x − γv̂y +
m
8
2γ
s
kT ˆ
ξx ,
m
2γ
kT ˆ
ξy .
m
(43)
Finally, making use of hξx (ω)ξy (ω 0 )i = 0, we get
1 2kT γ
ω 2 + γ 2 + ωc2
2π m (ωc2 + γ 2 − ω 2 )2 + 4ω 2 γ 2
1 2kT
γ(ω 2 + γ 2 )(ω 2 + γ 2 + ωc2 )
=
,
2π m γ 2 (ω 2 + γ 2 + ωc2 )2 + ω 2 (ω 2 + γ 2 − ωc2 )2
Svx (ω) = Svy (ω) =
(44)
where ωc = qB/m is the cyclotron frequency.
The stationary ACF is found as
G(τ ) =
kT −γ|τ |
e
cos ωc τ .
m
(45)
The psd and the ACF are shown in Fig. 3 for different values of γ as in the legend.
II.
SOLVING STOCHASTIC ODES
The Wiener Process.
Central role in stochastic calculus plays the Wiener process, which has been shortly
discussed above. Consider a continuous stochastic process W (t), with the conditional probability p(w, t|w0 , t0 ), satisfying the following (diffusion) equation
1
∂t p(w, t|w0 , t0 ) = ∂w2 p(w, t|w0 , t0 ).
2
(46)
It is well known that the solution of this equation with the initial condition given by the
delta function p(w, t|w0 , t0 ) = δ(w − w0 ), at t = t0 , is the Gaussian
!
(w − w0 )2
.
p(w, t|w0 , t0 ) = q
exp −
2(t − t0 )
2π(t − t0 )
1
(47)
Using this result, we conclude that
hW (t)i =
Z
∞
−∞
dw w p(w, t|w0 , t0 ) = w0 ,
h(W (t) − w0 )2 i = t − t0 .
(48)
Important is to observe that random variables ∆Wi = W (ti ) − W (ti−1 ), with any ti > ti−1
are independent. Indeed, using the Markov property, we can write the joint distribution as
p(wn , tn ; wn−1 , tn−1 ; . . . ; w0 , t0 )
(49)
= [p(wn , tn |wn−1 , tn−1 )p(wn−1 , tn−1 |wn−2 , tn−2 ) . . . p(w1 , t1 |w0 , t0 )] p(w0 , t0 ).
9
(50)
Now

n 
Y

!
(wi − wi−1 )2 
q
p(wn , tn ; wn−1 , tn−1 ; . . . ; w0 , t0 ) =
exp −
p(w0 , t0 )

2(ti − ti−1 ) 
2π(ti − ti−1 )
i=0
1
= p(∆wn ; ∆wn−1 . . . ; w0 ).
(51)
Using the last equation, we obtain
h∆Wi2 i = ti − ti−1
(52)
hW (t)W (s)|[w0 , t0 ]i = h(W (t) − W (s))W (s)i + hW (s)2 i = min(t − t0 , s − s0 ) + w02 .
Basics of the Ito and Stratonovich stochastic calculus
The starting point for the discussion is the relation between the Wiener process and the
white noise
∆Wi = W (ti ) − W (ti−1 ) = ξ(ti−1 )∆t.
(53)
Then the solution of a general stochastic equation
ẋ = f (x) + g(x)ξ(t),
(54)
is represented via the integral
Z
x(T ) = x(t0 ) +
Z
T
ẋ dt =
t0
Z
T
T
[f (x(t))dt + g(x(t))ξ(t)dt] =
t0
[f (x(t))dt + g(x(t))dW (t)]
t0
(55)
This result raises the question of the interpretation of a stochastic integral of the form
Z
T
g(x(t))dW (t).
(56)
t0
This integral is understood in the mean-square limit sense. More precisely, a sequence of
random variables Xn (ω) is said to converge to X(ω) in the sense of the mean-square limit,
if
Z
lim
n→∞
dω p(ω)[Xn (ω) − X(ω)]2 = 0
(57)
Ito stochastic integral
Consider
Z
T
W (t)dW (t).
t0
10
(58)
Partitioning the interval [t0 , T ] into the subintervals by ti , i = 1, . . . , n, the Ito integral is
defined as the mean-square limit of the sum
n
X
Wi−1 (Wi − Wi−1 ) =
i=1
n
X
Wi−1 ∆Wi =
i=1
=
n h
i
1X
(Wi−1 + ∆Wi )2 − (Wi−1 )2 − (∆Wi )2
2 i=1
n
1
1X
[W (T )2 − W (t0 )2 ] −
(∆Wi )2 .
2
2 i=1
(59)
Now compute the mean-square limit of the sum in the last equation. Using Eq. (52), we get
n
X
h
(∆Wi )2 i =
i=1
n
X
(ti − ti−1 ) = T − t0 .
(60)
i=1
Additionally, we have
"
h
n
X
#2
(∆Wi ) − (T − t0 ) i
2
i=1
X
= h
∆Wi4 + 2
i
X
(61)
∆Wi2 ∆Wj2 − 2(t − t0 )
i<j
X
∆Wi2 + (T − t0 )2 i.
i
But because ∆Wi are independent Gaussian variables, it holds
h∆Wi2 ∆Wj2 i = (ti − ti−1 )(tj − tj−1 ),
h∆Wi4 i = 3(ti − ti−1 )2 .
(62)
Finally
"
h
n
X
#2
(∆Wi ) − (T − t0 ) i
2
(63)
i=1
= 2
X
(ti − ti−1 )2 +
i
X
[(ti − ti−1 −
all i,j
(T − t0 )
(T − t0 )
(T − t0 )2
)(tj − tj−1 −
)] = 2
→ 0.
n
n
n
This completes the proof and we obtain
Z
T
W (t)dW (t) =
t0
i
1h
W (T )2 − W (t0 )2 − (T − t0 ) .
2
(64)
Stratonovich interpretation
The integral Eq. (58) can also be approximated by taking the mid point in every subinterval [ti−1 , ti ]
n
X
1
n
1X
(Wi−1 + Wi ) (Wi − Wi−1 ) =
(Wi−1 ∆Wi + Wi ∆Wi )
2 i=1
i=1 2
1
=
2
1
=
2
(
(
(65)
)
X
1X
1
(W (T )2 − W (t0 )2 −
∆Wi2 ) −
((Wi − ∆Wi2 )2 − Wi2 − ∆Wi2 )
2
2
i
i
)
X
X
1
1
1
(W (T )2 − W (t0 )2 −
∆Wi2 ) + (W (T )2 − W (t0 )2 +
∆Wi2 ) = (W (T )2 − W (t0 )2 ).
2
2
2
i
i
11
Change of variables: the Ito formula
Recalling that according to Eq. (52),
h∆Wi2 i = ti − ti−1 = ∆t,
(66)
we will shortly write in what follows dW 2 = dt.
Then it is easy to see how the differentiation chain rule for an arbitrary function F (x)
changes in case when x(t) satisfies the stochastic differential equation Eq. (54)
dF [x(t)] = F [x(t) + dx(t)] − F [x(t)]
1
= F 0 [x(t)]dx(t) + F 00 [x(t)]dx(t)2 + . . .
2
1
= F 0 [x(t)] {f (x)dt + g(x)dW (t)} + F 00 [x(t)] {f (x)dt + g(x)dW (t)}2 . (67)
2
√
Retaining only terms of the order of dW (t) ∼ dt and dt, we obtain
1
f (x)F [x] + g(x)2 F 00 [x] dt + g(x)F 0 [x]dW (t).
2
dF [x(t)] =
0
(68)
Equivalence of an Ito sde to a Stratonovich sde
Consider a Stratonovich sde
dx(t) = ẋdt = α(x)dt + β(x)dW (t).
(69)
The solution of this equation is represented as a sum of a regular and a stochastic
Stratonovich integrals
Z
Z
T
x(T ) =
T
α(x(t))dt + (S)
t0
Z T
=
α(x(t))dt +
X
t0
β(x(t))dW (t)
t0
β
i
(70)
Z T
X
1
1
(xi + xi−1 ) ∆Wi =
α(x(t))dt +
β xi−1 + ∆xi ∆Wi ,
2
2
t0
i
where ∆xi = xi − xi−1 . But
1
β xi−1 + ∆xi
2
= β(xi−1 ) +
∂β 1
1 ∂ 2β
∆xi +
∂x 2
2 ∂x2
1
∆xi
2
2
+ ...
(71)
Our aim is to express the r.h.s. of Eq. (70) using the Ito interpretation. To this end, we set
∆xi = a(xi−1 )∆t + b(xi−1 )∆Wi ,
(72)
with a(x) and b(x) different from α(x) and β(x). Combining Eq. (72) and Eq. (71), we obtain
1
b(xi−1 )∂x β(xi−1 )∆Wi .β xi−1 + ∆xi
2
1
1
= β(xi−1 ) + a(xi−1 )∂x β(xi−1 ) + b2 (xi−1 )∂x2 β(xi−1 ) ∆t
4
2
1
+ b(xi−1 )∂x β(xi−1 )∆Wi .
(73)
2
12
Therefore, neglecting the terms of the order of dt dW and dt2 , and keeping in mind that
∆Wi2 = ∆t, we conclude
Z
Z
T
(S)
T
β(x(t))dW (x(t)) = (I)
t0
t0
Z T
1
β(x(t))dW (t) + (I)
b(x(t))∂x β(x(t))dt
2
t0
(74)
This shows that a Stratonovich sde
dx = α(x)dt + β(x)dW (t)
(75)
dx = a(x)dt + b(x)dW (t),
(76)
1
a(x) = α(x) + β(x)∂x β(x), b(x) = β(x).
2
(77)
is equivalent to an Ito sde
where
Connection between Fokker-Planck equation and a stochastic sde
For an arbitrary function f (x), where x(t) satisfies the Ito sde Eq. (76), we obtain using
the Ito formula
1
hdf (x(t))i = (a(x)f 0 (x) + b2 (x)f 00 (x)) dt,
2
where we have used the fact that hdW (t)i
=
(78)
0 and further assumed that
hb(x(t))f 0 (x(t))dW (t)i = 0. Then
hdf (x(t))i
df (x(t))
d
= h
i = hf (x(t))i
(79)
dt
dt
dt
Z
Z
1
= dxf (x)∂t p(x, t|x0 , t0 ) = dx a(x)∂x f + b(x)2 ∂x2 f p(x, t|x0 , t0 ),
2
where p(x, t|x0 , t0 ) is the conditional probability. Using integration by parts and natural
boundary conditions at ±∞, we get
Z
Z
dxf (x)∂t p =
1
dxf (x) −∂x (a(x)p) + ∂x2 (b(x)2 p) .
2
(80)
Because f (x) is arbitrary, we are left with the equation for p(x, t|x0 , t0 ), which is called the
Fokker-Planck equation, corresponding to the Ito sde Eq. (76)
∂t p = −∂x
1
a(x)p − ∂x (b(x)2 p) = −∂x J(x),
2
13
(81)
where J(x) is the probability current.
Using our previous results, it is easy to show that if Eq. (76) i treated in the Stratonovich
interpretation, then the corresponding Fokker-Planck equation becomes
1
∂t p = −∂x a(x)p − b(x)∂x (b(x)p) = −∂x J(x).
2
(82)
More generally, if a system is described by a set of stochastic equations
ẋi = fi (x1 , x2 , . . . , xn ) +
n
X
gij (x1 , x2 , . . . , xn )ξj (t), (i = 1, 2, . . . , n),
(83)
j=1
where ξk (t) represent sources of uncorrelated white Gaussian noise. Than in the Stratonovich
interpretation, we have
1
∂t p = −∂i (fi p) + ∂i (gim ∂k gkm p) .
2
(84)
In the last equation we have used Einstein’s summation convention over a pair of repeated
indexes.
Example: active Brownian particle
The rotation of the direction vector p is described by the equation
ṗ = η(t) × p,
(85)
where η is a 3-dimensional vector with components given by uncorrelated white Gaussian
noise, i.e hηi (t)ηk (t0 )i = 2Dδ(t − t0 ), and the indexes (i, k) = (1, 2, 3), corresponding to
(x, y, z).
We want to show that if Eq. (85) is treated in the Stratonovich interpretation, than it
gives rise to the following Smoluchowski equation for the probability density ρ(p)
∂t ρ = DR2 ρ,
(86)
where R = p × ∇p .
Recall that the cross product between any two vectors a and b can be written with the
help of the antisymmetric unity tensor Eijk , which is equal either to ±1 for all different (ijk)
(= +1 for (ijk) = (123)), or it is equal to 0 if any two indexes (ijk) are identical.
a × b = Eijk aj bk ei ,
where ei denotes the unity vector, pointing along the i-axes.
14
(87)
Eq. (86) can be rewritten in terms of Eijk
∂t ρ = D (Eijk pj ∂pk Eims pm ∂ps ) ρ,
(88)
Than in the Stratonovich interpretation, we have
1
∂t ρ = ∂i (σim ∂k σkm ρ)
2
(89)
By comparing Eq. (85) with Eq. (83), we see that
σik = Eikm pm .
(90)
Now we can rewrite Eq. (89) for the case of Eq. (85)
1
∂t ρ = 2D ∂pi (Eims ps ∂pk Ekmj pj ρ)
2
(91)
Finally, after noticing that ∂pi pk = δik and after the appropriate permutation of indexes in
Eq. (91), we conclude that Eq. (91) and Eq. (88) are identical.
How to use the Fokker-Planck equation to compute the evolution equation
for the moments hxm (t)i.
Example: diffusion coefficient of a classical Brownian particle.
Consider one-dimensional motion of a classical particle m in contact with a fluctuating
environment
p
m
√
ṗ = −αp + 2Dξ(t),
ẋ =
(92)
where α is the damping coefficient and D characterizes the strength of fluctuations. The
corresponding Fokker-Planck equation is given by
∂t ρ(x, p, t) = −∂x
p
ρ + ∂p [αpρ + D∂p ρ].
m
(93)
The stationary solution is x-independent
r
ρs =
!
αp2
α
exp −
.
2πD
2D
(94)
In order to recover the Maxwell’s velocity distribution at temperature T
r
ρs (v) =
!
m
mv 2
exp −
,
2πkT
2kT
15
(95)
one needs to impose the Einstein’s condition
D = αmkT.
(96)
We are interested in the diffusion coefficient of the particle
h(x(t) − x(0))2 i
.
t→∞
2t
D∞ = lim
(97)
In order to find hx(t)i, we use the Fokker-Planck equation, integration by parts and natural
BCs at ±∞
∂t hx(t)i =
Z
Z
dxdp x∂t ρ(x, t) =
dxdp x ∂x −
p
hp(t)i
ρ + ∂p [αpρ + D∂p ρ] =
.(98)
m
m
Similarly, we find the equations for hp(t)i, hx(t)p(t)i, hp(t)2 i and hx(t)2 i
Z
Z
p
∂t hp(t)i =
dxdp p∂t ρ(x, t) = dxdp p ∂x − ρ + ∂p [αpρ + D∂p ρ] = −αhp(t)i,
m
Z
p
hp(t)2 i
∂t hx(t)p(t)i =
dxdp px ∂x − ρ + ∂p [αpρ + D∂p ρ] =
− αhx(t)p(t)i
m
m
Z
2hx(t)p(t)i
p
2
2
∂t hx(t) i =
dxdp x ∂x − ρ + ∂p [αpρ + D∂p ρ] =
m m
Z
p
∂t hp(t)2 i =
dxdp p2 ∂x − ρ + ∂p [αpρ + D∂p ρ] = −2αhp(t)2 i + 2D.
(99)
m
Eqs. (98,99) represent a closed system of linear equations for the first five moments, which
can be solved analytically
hp(t)i = p0 e−α(t−t0 )
(100)
D
1 − e−2α(t−t0 ) + p20 e−2α(t−t0 )
hp(t)2 i =
α
p0 hx(t)i = x0 +
1 − e−α(t−t0 )
αm
!
!
D
p20
D
2D
p20
−2α(t−t0 )
−
e
+ (xp)0 − 2 +
e−α(t−t0 )
hx(t)p(t)i = 2 −
α m
αm α2 m
α m αm
2Dt
+ C1 e−α(t−t0 ) + C2 e−2α(t−t0 ) .
(101)
hx(t)2 i =
(αm)2
The terms proportional to C1 and C2 vanish in the limit t → ∞ in Eq. (97) and we finally
obtain
D∞ =
D
kT
=
.
2
(αm)
αm
16
(102)
Numerical solution of stochastic equations
In practice, stochastic equations can be solved using methods with much lower accuracy
then the solution of the corresponding deterministic equations would require.
√
Thus, the dt-accurate Euler scheme for equation Eq. (76) in the Ito interpretation is
given by
x(t + dt) = x(t) + a(x(t))dt + b(x(t))dW (t),
√
dW (t) = N (0, 1) dt,
(103)
where N (0, 1) is a normally distributed random number with variance one and zero mean.
The corresponding numerical scheme for the Stratonovich interpretation is
y = x(t) + a(x(t))dt + b(x(t))dW (t),
1
1
x(t + dt) = x(t) + (x(t) + x(y)) dt + (b(x(t)) + b(y)) dW (t),
2√
2
dW (t) = N (0, 1) dt,
(104)
where dW (t) is the same in the first and in the second equations.
The advantage of this method is the fact that it is explicit. We now show that this
explicit algorithm is indeed equivalent to the implicit algorithm, which corresponds to the
definition of the Stratonovich interpretation.
Thus, according to the definition of the Stratonovich integration, we should have had the
following implicit scheme
x(t + dt) + x(t)
,
2
x(t + dt) = x(t) + a(x̄)dt + b(x̄)dW (t),
√
dW (t) = N (0, 1) dt.
x̄ =
(105)
For simplicity set a(x) = 0 and use the Newton’s method to solve the algebraic equation
!
x(t) + x(t + dt)
dW (t).
0 = x(t + dt) − x(t) − b
2
(106)
The iteration scheme of the Newton’s method for an algebraic equation g(y) = 0 is
y (n+1) = y (n) −
17
g(y (n) )
.
g 0 (y (n) )
(107)
By setting y (n) = x(t) and y (n+1) = x(t + dt), we obtain in our case
−b(x(t))dW (t)
1 − 21 b0 (x(t))dW (t)
1
≈ x(t) + b(x(t))dW (t) + b(x(t))b0 (x(t))dt + O(dt dW (t)).
2
x(t + dt) = x(t) −
(108)
This result is equivalent to Eq. (104). Indeed
1
(b(x(t)) + b(x(t) + b(x(t))dW (t))) dW (t)
2
1
= x(t) + b(x(t))dW (t) + b(x(t))b0 (x(t))dt + O(dt dW (t)).
2
x(t + dt) = x(t) +
18
(109)