IDEAS THAT WE WILL NEED FROM CHAPTER 5
Chapter 5 is about the simultaneous behavior of two or more random variables. For example, we might
be concerned with the probability that a randomly selected person will have a body mass index at most
35 AND a systolic blood pressure at most 140. Body mass index (Y1) is one random variable, and
systolic blood pressure (Y2) is another. The values 35 and 140 are arbitrary. As we slide these values
around, we can create a function called the joint (cumulative) distribution function:
πΉπ1, π2 (y1,y2) = P(Y1 ο£ y1 AND Y2 ο£ y2)
Much of chapter 5 is devoted to methods for calculating and manipulating these functions.
Fortunately for us, we will only need to be able to calculate the means and standard deviations for linear
combinations of multiple variables. That is, we will frequently be concerned with functions of the form
W ο½ a1Y1 ο« a2Y2 ο« ... ο« akYk ο« b
Suppose that each Yi has expected value denoted by οi . What can we say about E (W ) ?
RESULT #1: E (W ) ο½ a1ο1 ο« a2 ο2 ο« ... ο« ak οk ο« b ο½ a1E (Y1 ) ο« a2 E (Y2 ) ο« ... ο« ak E (Yk ) ο« b
While I am deliberately avoiding developing the machinery for a proof of this important result, it is not
surprising that expectations distribute over linear operations on multiple variables in the same way that
they distribute over linear functions of single variables.
Example 1. You are scoring college applicants on the basis of their High School GPA (Y1) and their
total SAT score (Y2). The scoring function you use is S ο½ 10Y1 ο« Y2 / 40 .
If applicants have a mean High School GPA of 3.1 and a mean total SAT of 1100, what is the mean for
the values of S?
The mean is 10*3.1 + 1100/40 = 58.5
Example 2. The mean time to service a rental car when it is returned is 25 minutes. 10 cars are
returned during the day. What is the expected total time to service all of them?
Intuitively, you know that this must be 25*10 = 250 minutes. Formally, let Y1 be the time to
service the first car, Y2 the time to service the second, etc.
total ο½ T ο½ Y1 ο« Y2 ο« ... ο« Y10
Then E(T) = 25 + 25 + β¦.+ 25 = 10*25 = 250
In addition to getting the mean, we also need to know the standard deviation. This works out reasonably
easily IF we can assume the variables are independent.
Definition: Let the random variable π1 have c.d.f. πΉπ1 , and let the random variable π2 have c.d.f. πΉπ2 .
Let the two random variables together have the joint c.d.f. πΉπ1 ,π2 . The two random variables are said to
be independent if, for all π¦1 πππ π¦2 , we have πΉπ1 ,π2 (π¦1 , π¦2 ) = πΉπ1 (π¦1 )πΉπ2 (π¦2 ). If π1 and π2 are not
independent, they are said to be dependent.
Note: If the two random variables are continuous, then independence also implies that the joint p.d.f. is
factorable into the product of the individual (marginal) p.d.f.βs.
If the two random variables are discrete, then independence implies that the joint p.m.f. is factorable into
the product of the individual (marginal) p.m.f.βs.
1
Example: Consider the joint p.d.f. for two continuous r.v.βs π1 and π2 given by
1
1
ππ1 ,π2 (π¦1 , π¦2 ) = 2ππ2 ππ₯π {β 2π2 [(π¦1 β π1 )2 + (π¦2 β π2 )2 ]} , for ββ < π¦1 < +β, and
ββ < π¦2 < +β. This joint p.d.f. may be factored into a product of two normal p.d.f.βs with different
means but the same standard deviation. Thus the two r.v.βs π1 and π2 are independent.
Example: p. 253, 5.52
Example: Consider the joint p.m.f. for two discrete r.v.βs π1 and π2 given by
1
ππ1 ,π2 (π¦1 , π¦2 ) = 36, for (π¦1 , π¦2 ) β {1, 2, 3, 4, 5, 6}, or ππ1 ,π2 (π¦1 , π¦2 ) = 0, otherwise. This joint p.m.f.
may be factored into the product of
1
ππ1 (π¦1 ) = 6, for (π¦1 ) β {1, 2, 3, 4, 5, 6}, or ππ1 , (π¦1 ) = 0, otherwise; and
1
ππ2 (π¦2 ) = 6, for (π¦2 ) β {1, 2, 3, 4, 5, 6}, or ππ2 , (π¦2 ) = 0, otherwise. Thus the two r.v.βs π1 and π2 are
independent.
Example: p. 251, 5.45
The most important consequence of independence is that the variances add in a sensible way. If the
variables were not independent, then the variance of the linear combination would also include terms
involving covariances of pairs of the Yβs.
RESULT #2: If the variables are independent, V (W ) ο½ a12V (Y1 ) ο« a22V (Y2 ) ο« .... ο« ak2V (Yk ) .
Notice that the additive constant, the βbβ term, disappears, just as it did for the single variable case in
chapters 3 and 4.
Important Note: Random sampling implies that the measurements taken on separate members of the
random sample will be independent random variables.
Whether or not variables are independent can be an interesting question. For example, we are fairly sure
that body mass index is NOT independent of blood pressure β those people with extremely high BMI are
more likely to have high blood pressures. Similarly, High School GPA is not independent of total SAT.
Notice that this does not mean that those with high HS GPA must have high total SAT, but only that
they are more likely to have high total SAT. On the other hand, the HS GPAβs of two randomly selected
applicants are independent because the results of the first applicant will not give any information about
the results of the second applicant.
Example 2, continued. Suppose that the standard deviation for a single carβs service time is 6 minutes.
Also, assume it is reasonable that the times for different cars are independent.
a. What is the standard deviation for T = total time to service the 10 cars?
Solution:
Each car has a time with variance ο³ 2 ο½ 62 ο½ 36 . Hence, the variance for T is
V (T ) ο½ V (Y1 ) ο« V (Y2 ) ο« ... ο« V (Y10 ) ο½ 10 ο΄ 36 ο½ 360
and the standard deviation is
360 ο½ 18.97
2
b. Suppose that T ο½ T /10 is the mean time in the sample of 10 cars. What is the expected value and
standard deviation for T ?
Using the basic results from chapters 3 and 4, we know E (T ) ο½ E (T ) /10 ο½ 25
And V (T ) ο½ V (T ) /102 ο½ 3.6
The standard deviation for T is
3.6 ο½ 1.897
It is extremely important to notice that the standard deviation for T was much smaller than the
standard deviation for any single car. This fundamental property of averages is the principle that
makes statistical inference possible!
Problem A. Suppose that the time between service calls on a copy machine is exponentially distributed
with a mean ο’ . Let T be the sample mean for records of 36 independent times.
a. Express E (T ) and V (T ) as expressions using ο’ .
b. Suppose the sample mean is 20.8 days. Provide a common sense estimate of ο’ . Assuming that the
distribution of T is nearly bell-shaped, is it likely that the true value of ο’ differs from your estimate by
more than 10 days?
The behavior of the the sample mean ( T ) is of great concern to people doing practical statistics. Does it
provide a stable estimate of the population mean ο? Results #1 and #2 yield the following:
RESULT #3. Suppose that Y1 , Y2 ,..., Yn are independent random variables all having the same mean ο
and variance ο³ 2 . Then Y has mean ο and variance ο³ 2 / n .
proof: Using result #1, E (Y ) ο½ E (Y1 / n ο« Y2 / n ο« ... ο« Yn / n) ο½ ο / n ο« ο / n ο« ... ο« ο / n ο½ ο
Using result #2 and the assumption of independence,
V (Y ) ο½ V (Y1 / n ο« Y2 / n ο« ... ο« Yn / n) ο½ ο³ 2 / n2 ο« ο³ 2 / n2 ο« .... ο« ο³ 2 / n2 ο½ ο³ 2 / n
Notice that we only needed independence for the variance calculation.
MOMENT GENERATING FUNCTIONS (This is actually covered in Chapter 6.)
Another useful property of independent random variables is that the expectation distributes over
PRODUCTS. Notice that expectations always distribute over +/-, but it takes the special case of
independent to allow them to distribute over multiplication!
RESULT #4: If the variables are independent, then
E ( g1 (Y1 ) ο΄ g2 (Y2 ) ο΄ ... ο΄ gk (Yk )) ο½ E( g1 (Y1 )) ο΄ E( g2 (Y2 )) ο΄ ... ο΄ E( g k (Yk ))
There are many useful consequences of this, and we will mostly be concerned with the implications for
the moment generating function of the linear combination W.
E(etW ) ο½ E(exp ο»ο₯ a jYj ο« bο½) ο½ E(eta1Y1 ο΄ eta2Y2 ο΄ ... ο΄ etakYk ο΄ etB ) ο½ etb ο E (e
ta jY j
) ο½ etb ο m j (a jt )
3
That is, the moment generating functions will multiply!
Example 3. Suppose that Y1 , Y2 ,..., Yn are independent random variables having the exponential
distribution with mean ο’ .
a. What is the distribution of T ο½ Y1 ο« Y2 ο« ... ο« Yn ?
b. What is the distribution of T ο½ T / n ?
Example 4. Suppose that Y1 is a binomial variable with parameters n1 , p and Y2 is an independent
binomial variable with parameters n2 , p . Note that the values of p are the same, but not necessarily the
values of n . What is the distribution of T ο½ Y1 ο« Y2 ? Would the same result hold if the values of p were
allowed to differ?
Problem B. Suppose that Y1 is a normally distributed variable with parameters ο1 , ο³12 and Y2 is an
independent normally distributed variable with parameters ο2 , ο³ 22 . What is the distribution of
T ο½ Y1 ο« Y2 ?
4
© Copyright 2026 Paperzz