Functions of Random Variables

8
Functions of Random
Variables
Learning Objectives

The purpose of this section is to:



Describe the problem of finding the distribution of a function of
random variables
This modeling is required for propagation of uncertainties in risk
and reliability analysis
Describe the basic analytical and approximate methods for
estimating the mean and variance of a function of random
variables
–
They can be used to approximate the distribution of the function
UN 0701 Engineering Risk and Reliability
201.
Section 8. Functions of Random Variables
(this page left blank intentionally)
202.
UN 0701 Engineering Risk and Reliability
Section 8. Functions of Random Variables
8.1 Introduction
Fundamental Problem




Many engineering problems involve a functional relationship
between a system response variable and one or more basic
(independent) variables
If any of the basic variables are random, the response variable
will also be random
The probability distribution and the moments of the dependent
variable will also be functionally related to the basic random
variables
How do we estimate the probability distribution of the
dependent variable?

How can we estimate mean and standard deviation of the
function
UN 0701 Engineering Risk and Reliability
203.
Section 8. Functions of Random Variables
8.2 Distributions of Functions of RVs
Functions of RVs

Types of functions encountered in risk and reliability problems
can be classified as



Scalar Function – representing a single response quantity
Vector Function – representing multiple responses
Scala Functions

Univariate function: depends on a single RV,
–

Multivariate function: depends on multiple RVs,
–

(1) Linear, or (2) Nonlinear functions
(1) Linear, or (2) Nonlinear functions
Vector Functions

More complex and not covered in this course
Analysis of Functions of RVs

The analysis of a “random” function means



Probability distribution of a function




Determine the complete probability distribution
Determine the moments (mean and SD) only
Transformation method based on calculus concepts
Easy to apply to univariate (scalar) and linear functions
For multivariate functions, only simple cases can be solved
Moments of a function

Since distribution of a function is analytically complex, it is easier
to derive moments of the function
–
–

204.
Exact solutions are possible for simple functions (linear functions)
For complex functions, approximations are based on the Taylor series
Simulation Method: A general numerical approach to solve
these problems
UN 0701 Engineering Risk and Reliability
Section 8. Functions of Random Variables
Functions of Single RVs



The probability distributions of functions of random variables
can be derived through transformation
Consider the function of a single random variable
where X is a random variable and g(x) is a monotonically
increasing function of x
Generally we can solve y = g(x) for x
y
to find the inverse function
e.g. y = ex
where g-1 is the inverse function of g
x
Distribution of the Function

Under these conditions, we can solve directly for the CDF of
the dependent variable Y as
or


In general when finding the distributions of functionally related
random variables, we must work with their CDF’s
To obtain the PDF of Y, we must find the CDF and then
differentiate it, therefore, given
UN 0701 Engineering Risk and Reliability
205.
Section 8. Functions of Random Variables
PDF of Y

The result is

Replacing g-1(y) with x we obtain
or
for functions where y increases with x
Graphical Interpretation
y
y = g(x)
dy
dx
x


206.
The likelihood that Y takes on a value in an interval of width dy
centered on the value of y is equal to the likelihood that X takes on
a value in an interval centered on the corresponding value
x = g-1(y), but with a width dx = dg-1(y)
The interval widths are generally not equal because of the slope
of the function g(x) at y
UN 0701 Engineering Risk and Reliability
Section 8. Functions of Random Variables
Monotonically Decreasing Functions

For monotonically decreasing functions, we have
which results in

However, because the first term on the right hand side is
negative we get
or
where |..| denotes the absolute value
Example 8.1
Consider a linear relationship between X and Y as Y = aX + b.
Find the distribution of Y given that X N(X,X).

Solution: The inverse function is


Recall the PDF for the Normal distribution, we get


Therefore, Y follows the Normal distribution with Y = aX + b
and Y = aX
UN 0701 Engineering Risk and Reliability
207.
Section 8. Functions of Random Variables
Example 8.2
If X has a LOG-NORMAL distribution with parameters l and z,
what is the distribution of Y = ln(X)?

Solution: The inverse function is


Recall the PDF for the Log-Normal distribution, we get


Therefore, Y is a Normal variate with the mean value l and
standard deviation z. That is, E[ln(X)] = l and Var[ln(X)] = z2
Functions of Two RVs

Consider, for example, the following functional relationship

We want to map f(x,y) into g(z,y) or g(x,z)
The joint probability distribution function (CDF) of X and Y is

and the joint PDF is

208.
The CDF of Z is given as
UN 0701 Engineering Risk and Reliability
Section 8. Functions of Random Variables
Transformation



Consider the mapping of f(x,y) to g(z,y)
The inverse function is equal to x = z - y
The transformation is
where the Jacobian J(z,y) is equal to

Therefore,
Joint PDF of Z

The joint PDF is obtained by taking the derivative of the above
distribution function as

For independent X and Y

Therefore, the joint PDF becomes
UN 0701 Engineering Risk and Reliability
209.
Section 8. Functions of Random Variables
Joint PDF of Z (cont’d)

It can also be shown that for the mapping of f(x,y) to g(x,z)
and for independent X and Y



The above equations are convolution integrals
In many problems, performing the integration may be
challenging, especially since many practical PDF’s are defined
by different functions (e.g. Weibull, Gumbel, etc.)
Need to use numerical methods such as Monte Carlo
Simulation, or approximation methods such as FORM and
SORM
Example 8.3
Consider the special case where X and Y are statistically
independent Normal random variables with means and standard
deviations X, X and Y, Y, respectively. Find the probability
density for Z = X – Y.

210.
Solution: The probability density is
UN 0701 Engineering Risk and Reliability
Section 8. Functions of Random Variables
Solution (cont’d)

Simplifying the power terms

Substituting back
where
and

After integration and algebraic reduction..., the final result for
the density function of Z becomes
which is a Normal density function with mean and standard
deviation of
and

In conclusion, any linear function of Normal random variables
is also a Normal random variable
UN 0701 Engineering Risk and Reliability
211.
Section 8. Functions of Random Variables
8.3 Moments of Functions of RVs


In many cases, the moments (e.g. mean and variance) of the
function may be obtained more easily than the full distribution
of the function
The moments may be sufficient for practical purposes even if
the correct probability distribution cannot be determined




With knowledge of mean and standard deviation, the COV can be
estimated, which is a relative measure of variability
This is meant as “propagation of uncertainty”
A function’s distribution can be approximated by a Normal
distribution, if its mean and SD are known
Similar to the probability distributions, the moments of functions
of random variables are related to the moments of the basic
random variables
Mean and Variance of Linear Functions



212.
Consider a univariate linear function Y = aX + b where a and b
are constants
The mean value of Y is given by the mathematical expectation
of the function as
The variance of Y is given as
UN 0701 Engineering Risk and Reliability
Section 8. Functions of Random Variables
Linear Functions (cont’d)

Consider a bivariate linear function

where a1 and a2 are constants
The mean value of Y is then

Therefore
The expected value of a sum is the
sum of the expected values.

The corresponding variance is given as

where Cov[X1,X2] is the covariance between X1 and X2
If X1 and X2 were statistically independent, then Cov[X1,X2]
would be equal to zero and the variance would reduce to
UN 0701 Engineering Risk and Reliability
213.
Section 8. Functions of Random Variables
Linear Functions (cont’d)

For a linear multivariate function
where ai are constants the mean is equal to
and the variance would be
where rXi,Xj is the correlation coefficient between Xi and Xj
Mean and Variance of a Product

Consider the product of n independent random variables as
(product is a non-linear function)

The mean value is


Therefore,
It can also be shown that

Therefore, the variance of the independent random variables is
214.
UN 0701 Engineering Risk and Reliability
Section 8. Functions of Random Variables
Function of a Single RV

Consider a general function (can be non-linear) of a single
random variable X

The exact moments of Y are given by the mathematical
expectation of g(X)
The mean would be

and the variance

Note that the density function fX(x) may not always be known,
and even in cases when it is known, the integrations may be
difficult to perform
Taylor Series Expansion


To avoid integration, we can approximate the function using
Taylor series expansion
The Taylor series expansion of the function g(x) about a point
x = a is
where g(n)(a) is the nth derivative of g with respect to x at point a
UN 0701 Engineering Risk and Reliability
215.
Section 8. Functions of Random Variables
Approximation of Mean

Substituting the first 3 terms of the Taylor series expanded at
the mean point for the single variable case

Assuming the second-order term is negligible, the first-order
approximation of the mean is
Approximation of Variance

Similarly for the variance, substituting the first 3 terms of the
Taylor series expanded at the mean point

Ignoring the higher order moments
216.
UN 0701 Engineering Risk and Reliability
Section 8. Functions of Random Variables
General Function of Multiple RVs

Consider a general function of n random variables

The mean of Y is given by the mathematical expectation

The variance is

Similar to the single variable case, the mean and variance can
be approximated using Taylor series expansion
Approximation of Mean

Expanding the general function about the mean values of Xn

Ignoring the higher order terms gives the first-order
approximation of the mean
UN 0701 Engineering Risk and Reliability
217.
Section 8. Functions of Random Variables
Approximation of Variance

The first-order approximation of the variance is

If the random variables are uncorrelated
Summary

Functions are often used to describe the overall response of a
system that may depend on other random variables


The uncertainty associated with a function can be described by





218.
e.g. PRA for core damage frequency analysis
Full probability distribution
Mean, SD and COV
This problem can be solved for analytically simple form of
functions (e.g., linear or product function)
In case of complex functions, approximations are required
Simulation is a versatile, numerical method that is universally
used to solve this class of problems
UN 0701 Engineering Risk and Reliability