Solution Series 7 - D-MATH

Probability and statistics
April 26, 2016
Lecturer: Prof. Dr. Mete SONER
Coordinator: Yilin WANG
Solution Series 7
Q1. Suppose two random variables X1 and X2 have a continuous joint distribution for which the
joint p.d.f. is as follows:
(
4x1 x2 for 0 < x1 , x2 < 1,
f (x1 , x2 ) =
0
otherwise.
Calculate Cov(X1 , X2 ). Determine the joint p.d.f. of two new random variables Y1 and Y2 ,
which are defined by the relations
X1
and Y2 = X1 X2 .
X2
Y1 =
Solution:
We apply the formula
Cov(X1 , X2 ) = E(X1 X2 ) − E(X1 )E(X2 ).
Since the marginal density for X1 is
1
Z
f1 (x1 ) =
4x1 x2 dx2 = 2x1 ,
0
1
Z
E(X1 ) =
x1 f1 (x1 )dx1 = 2/3 = E(X2 ).
0
We have also
Z
1
Z
E(X1 X2 ) =
1
x1 x2 4x1 x2 dx1 dx2 = 4/9.
0
0
Thus Cov(X1 , X2 ) = 0. Actually, one can see from f1 (x1 )f2 (x2 ) = f (x1 , x2 ) that X1 and X2
are independent.
Joint p.d.f. of (Y1 , Y2 ) = (r1 (X1 , X2 ), r2 (X1 , X2 )): where r1 (x1 , x2 ) = x1 /x2 and r2 (x1 , x2 ) =
x1 x2 . We have that
p
√
x1 = r1 r2 and x2 = r2 /r1 .
p
p
3
1
1
∂x1 /∂r1 ∂x2 /∂r1
r
/r
−
r
/r
2
1
2
1
p
Jac(r1 , r2 ) = det
= det p
=
.
∂x1 /∂r2 ∂x2 /∂r2
4
2r1
r1 /r2
1/r1 r2
1
Probability and statistics
April 26, 2016
By the transformation formula of the density,
p
√
fY1 ,Y2 (r1 , r2 ) = fX1 ,X2 ( r1 r2 , r2 /r1 )Jac(r1 , r2 )
= 4r2 /2r1 10<√r1 r2 ,√r2 /r1 <1
= 10<√r1 r2 ,√r2 /r1 <1 2r2 /r1 .
Let α and β be positive numbers. A random variable X has the gamma distribution with parameters
α and β if X has a continuous distribution for which the p.d.f. is
( α
β
xα−1 e−βx for x > 0,
Γ(α)
f (x|α, β) =
0
for x ≤ 0.
Where Γ is the function defined as: for α > 0,
Z
Γ(α) =
∞
xα−1 e−x dx.
0
Q2. Let X have the gamma distribution with parameters α and β.
(a) For k = 1, 2, . . . , show that the k-th moment of X is
E(X k ) =
Γ(α + k)
α(α + 1) · · · (α + k − 1)
=
.
k
β Γ(α)
βk
What are E(X) and V ar(X)?
(b) What is the moment generating function of X?
(c) If the random variables X1 , · · · , Xk are independent, and if Xi (for i = 1, . . . , k) has the
gamma distribution with parameters αi and β, show that the sum X1 + · · · + Xk has
the gamma distribution with parameters α1 + · · · + αk and β.
Solution:
(a) For k = 1, 2, · · ·
k
E[X ] =
=
=
=
Z ∞
βα
xk xα−1 e−βx dx
Γ(α) 0
Z ∞
βα
xα+k−1 e−βx dx
Γ(α) 0
Z ∞
βα
β −α−k y α+k−1 e−y dy
Γ(α) 0
Γ(α + k)
,
β k Γ(α)
where the change of variable y = βx is used. It can be easily seen that for α > 0,
Γ(α + 1) = αΓ(α).
2
Probability and statistics
April 26, 2016
We then deduce the second equality.
E(X) = α/β
V ar(X) = E(X 2 ) − E(X)2 = α/β 2 .
(b) By definition of moment generating function,
ψX (t) = E(etX )
Z ∞
βα
=
etx xα−1 e−βx dx
Γ(α) 0
Z ∞
βα
xα−1 e−(β−t)x dx
=
Γ(α) 0
α
β
=
β−t
The above computation is only valid for t < β.
(c) If ψi denotes the m.g.f of Xi , then it follows from the last question that for i = 1, · · · , k,
αi
β
.
ψi (t) =
β−t
The m.g.f. ψ of X1 + · · · + Xk is, by independence,
ψ(t) =
k
Y
ψi (t) =
i=1
β
β−t
α1 +···+αk
for t < β.
It coincides with the m.g.f. of a Gamma random variable with parameter (α1 + · · · +
αk , β), in an open interval of 0, thus the sum X1 + · · · + Xk has the Gamma distribution.
Q3. Service Times in a Queue. For i = 1, · · · , n, suppose that customer i in a queue must
wait time Xi for service once reaching the head of the queue. Let Z be the rate at which
the average customer is served. A typical probability model for this situation is to say that,
conditional on Z = z, X1 , . . . , Xn are i.i.d. with a distribution having the conditional p.d.f.
g1 (xi |z) = z exp(−zxi ) for xi > 0. Suppose that Z is also unknown and has the p.d.f.
f2 (z) = 2 exp(−2z) for z > 0.
(a) What is the joint p.d.f. of X1 , . . . , Xn , Z.
(b) What is the marginal joint distribution of X1 , . . . , Xn .
(c) What is the conditionalP
p.d.f. g2 (z|x1 , . . . , xn ) of Z given X1 = xi , . . . , Xn = xn ? For
this we can set y = 2 + ni=1 xi .
(d) What is the expected average service rate given the observations X1 = x1 , · · · , Xn = xn ?
Solution:
3
Probability and statistics
April 26, 2016
(a) The joint p.d.f. of X1 , · · · , Xn , Z is
f (x1 , · · · , xn , z) =
n
Y
g1 (xi |z)f2 (z)
i=1
n
= 2z exp(−z[2 + x1 + · · · + xn ]),
if z, x1 , · · · , xn > 0 and 0 otherwise.
(b) The marginal joint distribution of X1 , · · · , Xn is obtained by integrating z out of the
joint p.d.f. above.
Z ∞
2(n!)
2Γ(n + 1)
=
,
f (x1 , · · · , xn , z)dz =
n+1
(2 + x1 + · · · + xn )
(2 + x1 + · · · + xn )n+1
0
for all xi > 0 and 0 otherwise.
P
(c) We set y = 2 + ni=1 xi , for z > 0
g2 (z|x1 , · · · , xn ) = f (x1 , · · · , xn , z)
y n+1
2(n!)
z n exp(−zy)y n+1
n!
n+1
y
=
z n+1−1 e−yz ,
Γ(n + 1)
=
we recognize the conditional distribution of Z given X1 = x1 , · · · Xn = xn is Gamma
distribution with parameter α = n + 1, β = y.
(d) The conditional expected value of Z given X1 = x1 , · · · Xn = xn is the expected value
of Gamma distribution with parameter α = n + 1, β = y, which by Q2.a, equals to
E(Z|X1 = x1 , · · · Xn = xn ) =
n+1
P
.
2 + ni=1 xi
Q4. Least-squares line.
(a) Let (x1 , y1 ), . . . , (xn , yn ) be a set of n points of R2 and xi s are not all the same. Show
that the straight line defined by the equation y(x) = β̂0 + β̂1 x that minimizes the sum
of the squares of the vertical deviations of all the points from the line has the following
slope and intercept, i.e. (β̂0 , β̂1 ) minimizes
I(β0 , β1 ) :=
n
X
(β0 + β1 xi − yi )2
i=1
over all choices of (β0 , β1 ) ∈ R2 :
Pn
(y − ȳ)(xi − x̄)
Pn i
β̂1 = i=1
,
2
i=1 (xi − x̄)
β̂0 = ȳ − β̂1 x̄,
4
Probability and statistics
April 26, 2016
Table 1: Data for Q4.(b)
i
xi
yi
1
2
3
4
5
6
7
8
0.5
1.0
1.5
2.0
2.5
3.0
3.5
4.0
40
41
43
42
44
42
43
42
P
P
where x̄ = n1 ni=1 xi and ȳ = n1 ni=1 yi .
The minimizing line is called the least-squares line. Remark that the least-squares line
passes through the point (x̄, ȳ).
(b) Fit a straight line of the form y = β0 + β1 x to these values by the method of least
squares (with your calculator or Excel).
Solution:
(a) Using the fact that I(β0 , β1 ) → +∞ as k(β0 , β1 )k → ∞ (which is true since the xi s are
not all the same), the infimum of I is in approximated in some compact set. Since I is
continuous, the infimum of I is a minimum. We can look for critical points (β̂0 , β̂1 ):
∂β0 I(β̂0 , β̂1 ) = 2
∂β1 I(β̂0 , β̂1 ) = 2
n
X
i=1
n
X
β̂0 + β̂1 xi − yi = 0
xi (β̂0 + β̂1 xi − yi ) = 0.
i=1
We solve the above system:
nβ̂0 + nx̄β̂1 = nȳ
5
⇒
β̂0 = ȳ − x̄β̂1 ,
Probability and statistics
April 26, 2016
and
nx̄β̂0 +
n
X
!
x2i
β̂1 =
i=1
n
X
xi y i
i=1
n
X
⇒ nx̄(ȳ − x̄β̂1 ) +
!
x2i
β̂1 =
i=1
⇒
⇒
n
X
!
x2i
− nx̄
i=1
n
X
2
β̂1 =
!
2
(xi − x̄)
i=1
β̂1 =
n
X
xi y i
i=1
n
X
xi yi − nx̄ȳ
i=1
n
X
(xi − x̄)(yi − ȳ)
i=1
Pn
(x − x̄)(yi − ȳ)
Pn i
.
⇒ β̂1 = i=1
2
i=1 (xi − x̄)
(b) We apply the above formula to find β̂1 = 40.89 and β̂0 = 0.55.
6