Stationary Random Process
The concept of stationarity plays an important role in solving
practical problems
involving random processes. Just like time-invariance is an important characteristics of
many deterministic systems, stationarity describes certain time-invariant property of a
random process. Stationarity also leads to frequency-domain description of a random
process.
Strict-sense Stationary Process
A random process X (t ) is called strict-sense stationary (SSS) if its probability
structure is invariant with time. In terms of the joint distribution function, X (t ) is
called SSS if
FX (t1 ), X (t2 )..., X ( tn ) ( x1 , x2 ..., xn ) FX ( t1 t0 ), X ( t2 t0 )..., X ( tn t0 ) ( x1 , x2 ..., xn )
n N , t0 and for all choices of sample points t1 , t2 ,...tn .
Thus the joint distribution functions of any set of random variables X ( t1 ), X ( t 2 ), ..., X (t n )
does not depend on the placement of the origin of the time axis. This requirement is a
very strict. Less strict form of stationarity may be defined.
Particularly,
if
FX (t1 ), X (t2 )..... X (tn ) ( x1 , x2 .....xn ) FX (t1 t0 ), X (t2 t0 )..... X ( tn t0 ) ( x1 , x2 .....xn ) for n 1, 2,.., k ,
X (t ) is called kth order stationary.
If X (t ) is stationary up to order 1
FX (t1 ) ( x1 ) FX (t1 t0 ) ( x1 ) t 0 T
Let us assume t0 t1 . Then
FX (t1 ) ( x1 ) FX (0) ( x1 ) which is independent of time.
As a consequence
EX (t1 ) EX (0) X (0) constant
If X (t ) is stationary up to order 2
then
FX (t1 ), X (t2 ) ( x1 , x 2 .) FX (t1 t0 ), X (t2 t0 ) ( x1 , x 2 )
Put t 0 t 2
FX (t1 ), X (t2 ) ( x1 , x2 ) FX (t1 t2 ), X (0) ( x1 , x2 )
This implies that the second-order distribution depends only on the time-lag t1 t2 .
As a consequence, for such a process
RX (t1 , t2 ) E ( X (t1 ) X (t2 ))
xx
1 2
f X (0) X ( t1 t2 ) ( x1 , x2 ) dx1 dx2
= RX (t1 t2 )
Similarly,
CX (t1 , t2 )= C X (t1 t2 )
Therefore, the autocorrelation function of a SSS process depends only on the time lag
t1 t2 .
We can also define the joint stationarity of two random processes. Two processes
X (t ) and Y (t ) are called jointly strict-sense stationary
if their joint probability
distributions of any order is invariant under the translation of time. A complex process
Z (t ) X (t ) jY (t ) is called SSS if X (t ) and Y (t ) are jointly SSS.
Example An iid process is SSS. This is because, n,
FX (t1 ), X (t2 )..., X (tn ) ( x1 , x2 ..., xn ) FX (t1 ) ( x1 ) FX (t2 ) ( x2 )...FX (tn ) ( xn )
FX ( x1 ) FX ( x2 )...FX ( xn )
FX (t1 t0 ), X (t2 t0 )..., X (tn t0 ) ( x1 , x2 ..., xn )
Example The Poisson process is {N (t ), t 0} not stationary, because
EN (t ) t
which varies with time.
Wide-sense stationary process
It is very difficult to test whether a process is SSS or not. A subclass of the SSS process
called the wide sense stationary process is extremely important from practical point of
view.
A random process { X (t )} is called wide sense stationary process (WSS) if
EX (t ) X constant
and
EX (t1 ) X (t2 ) RX (t1 t2 ) is a function of time lag t1 t2 .
(Equivalently, Cov( X (t1 ) X (t2 )) C X (t1 t2 ) is a function of time lag t1 t2 )
Remark
(1) For a WSS process { X (t )},
EX 2 (t ) RX (0) constant
var( X (t )=EX 2 (t ) ( EX (t )) 2 constant
C X (t1 , t2 ) EX (t1 ) X (t2 ) EX (t1 ) EX (t2 )
RX (t2 t1 ) X2
C X (t1 , t2 ) is a function of lag (t2 t1 ).
(2) An SSS process is always WSS, but the converse is not always true.
Example: Sinusoid with random phase
Consider the random process X (t ) given by
X (t ) A cos(w0 t ) where A and w0 are constants and is uniformly distributed
between 0 and 2 .
This is the model of the carrier wave (sinusoid of fixed frequency) used to
analyse the noise performance of many receivers.
Note that
1
0 2
f ( ) 2
0 otherwise
By applying the rule for the transformation of a random variable, we get
1
-A x A
f X (t ) ( x) A2 x 2
0 otherwise
which is independent of t. Hence
Note that
X (t ) is first-order stationary.
EX (t ) EA cos( w0 t )
2
1
A cos( w t ) 2 d
0
0
0 which is a constant
and
RX (t1 , t2 ) EX (t1 ) X (t2 )
EA cos( w0 t1 ) A cos( w0 t2 )
2
A
E[c os( w0 t1 w0 t2 ) c os( w0 t1 w0 t2 )]
2
A2
E[c os( w0 (t1 t2 ) 2 ) c os( w0 (t1 t2 )]
2
A2
c os( w0 (t1 t2 ) which is a function of the lag t1 t2 .
2
Hence
X (t ) is wide-sense stationary.
Example: Sinusoid with random amplitude
Consider the random process X (t ) given by
X (t ) A cos(w0 t ) where and w0 are constants and A is a random variable. Here,
EX (t ) EA cos( w0t )
which is independent of time only if EA 0.
RX (t1 , t2 ) EX (t1 ) X (t2 )
EA cos( w0t1 ) A cos(w0t2 )
EA2 cos( w0t1 ) cos(w0t2 )
1
EA2 [c os( w0 (t1 t2 ) 2 ) c os( w0 (t1 t2 )]
2
which will not be function of (t1 t2 ) only.
Example: Random binary wave
Consider a binary random process X (t ) consisting of a sequence of random pulses of
duration T with the following features:
1
Pulse amplitude AK is a random variable with two values p AK (1) and
2
1
p AK ( 1)
2
Pulse amplitudes at different pulse durations are independent of each other.
The start time of the pulse sequence can be any value between 0 to T. Thus the
random start time D (Delay) is uniformly distributed between 0 and T.
A realization of the random binary wave is shown in Fig. above. Such waveforms are
used in binary munication- a pulse of amplitude 1is used to transmit ‘1’ and a pulse of
amplitude -1 is used to transmit ‘0’.
The random process X (t) can be written as,
X (t )
A rect
n
n
(t nT D)
T
For any t,
1
1
(1)( ) 0
2
2
1
1
EX 2 (t ) 12 (1) 2 1
2
2
EX (t ) 1
Thus mean and variance of the process are constants.
To find the autocorrelation function RX (t1 , t2 ) let us consider the case 0 t1 t1 T .
Depending on the delay D, the points t1 and t2 may lie on one or two pulse intervals.
Case 1:
X (t1 ) X (t2 ) 1
Case 2:
X (t1 ) X (t1 ) (1)(1) 1
Case 3:
X (t1 ) X (t2 ) 1
Thus,
0 D t1
1
X (t1 ) X (t2 )
1
or t2 D T
t1 D t2
RX (t1 , t2 ) EEX (t1 ) X (t2 ) | D
E ( X (t1 ) X (t2 ) | P(0 D t1 or t2 D T )) E ( X (t1 ) X (t 2 )) | t1 D t 2 ).P (t1 D t 2 )
1
1 t t
t t
1 1 2 1 (1 1 ) 2 1
T
2
2 T
t t
1 2 1
T
We also have, RX (t2 , t1 ) EX (t2 ) X (t1 ) EX (t1 ) X (t2 ) RX (t1 , t2 )
So that RX (t1 , t2 ) 1
t2 t1
T
t2 t1 T
For t2 t1 T , t1 and t2 are at different pulse intervals.
EX (t1 ) X (t2 ) EX (t1 ) EX (t2 ) 0
Thus the autocorrelation function for the random binary waveform depends on
t2 t1 , and we can write
RX ( ) 1
1
T
T
The plot of RX ( ) is shown below.
RX ( )
-T1
T1
Example Gaussian Random Process
Consider the Gaussian process
X (t ) discussed earlier. For any positive integer
n, X (t1 ), X (t 2 ),..... X (t n ) is jointly Gaussian with the joint density function given by
f X (t1 ), X (t2 )... X (tn ) ( x1 , x2 ,...xn )
1
X'CX1X
e 2
2
n
det(CX )
where CX E ( X μ X )( X μ X ) '
and μ X E ( X) E ( X1 ), E ( X 2 )......E ( X n ) '.
If X (t ) is WSS, then
EX (t1 ) X
EX (t2 ) X
..
μ X ..
..
..
EX (t )
n
X
X (t )
X X (t1 ) X
1
X (t2 ) X X (t2 ) X
..
CX E ..
..
..
X (tn ) X X (tn ) X
C X (t2 t1 )... C X (tn t1 )
C X (0)
C (t t )
C X (0)... C X (t2 tn )
X 2 1
C X (0)
C X (tn t1 ) C X (tn t1 )...
We see that f X (t1 ), X (t2 )... X (tn ) ( x1 , x2 ,...xn ) depends on the time-lags. Thus, for a Gaussian
random process, WSS implies strict sense stationarity, because this process is completely
described by the mean and the autocorrelation functions.
© Copyright 2026 Paperzz