Lecture 6 Notes

Jim Lambers
MAT 280
Spring Semester 2009-10
Lecture 6 Notes
These notes correspond to Section 11.5 in Stewart and Sections 2.5 and 3.5 in Marsden and Tromba.
The Chain Rule
Recall from single-variable calculus that if a function 𝑔(π‘₯) is differentiable at π‘₯0 , and 𝑓 (π‘₯) is
differentiable at 𝑔(π‘₯0 ), then the derivative of the composition (𝑓 ∘ 𝑔)(π‘₯) = 𝑓 (𝑔(π‘₯)) is given by the
Chain Rule
(𝑓 ∘ 𝑔)β€² (π‘₯0 ) = 𝑓 β€² (𝑔(π‘₯0 ))𝑔 β€² (π‘₯0 ).
We now generalize the Chain Rule to functions of several variables. Let f : 𝐷 βŠ† ℝ𝑛 β†’ β„π‘š , and let
g : π‘ˆ βŠ† ℝ𝑝 β†’ 𝐷. That is, the range of g is the domain of f .
Assume that g is differentiable at a point p0 ∈ π‘ˆ , and that f is differentiable at the point
q0 = g(p0 ). Then, f has a Jacobian matrix 𝐽f (q0 ), and g has a Jacobian matrix 𝐽g (p0 ). These
matrices contain the first partial derivatives of f and g evaluated at q0 and p0 , respectively.
Then, the Chain Rule states that the derivative of the composition (f ∘ g) : π‘ˆ β†’ β„π‘š , defined
by (f ∘ g)(x) = f (g(x)), at p0 , is given by the Jacobian matrix
𝐽f ∘g (p0 ) = 𝐽f (g(p0 ))𝐽g (p0 ).
That is, the derivative of f ∘ g at p0 is the product, in the sense of matrix multiplication, of the
derivative of f at g(p0 ) and the derivative of g at p0 . This is entirely analogous to the Chain Rule
from single-variable calculus, in which the derivative of 𝑓 ∘ 𝑔 at π‘₯0 is the product of the derivative
of 𝑓 at 𝑔(π‘₯0 ) and the derivative of 𝑔 at π‘₯0 .
It follows from the rules of matrix multiplication that the partial derivative of the 𝑖th component
function of f ∘ g with respect to the variable π‘₯𝑗 , an independent variable of g, is given by the dot
product of the gradient of the 𝑖th component function of f with the vector that contains the partial
derivatives of the component functions of g with respect to π‘₯𝑗 . We now illustrate the application
of this general Chain Rule with some examples.
Example Let 𝑓 : ℝ3 β†’ ℝ be defined by
𝑓 (π‘₯, 𝑦, 𝑧) = 𝑒𝑧 cos 2π‘₯ sin 3𝑦,
and let g : ℝ β†’ ℝ3 be a vector-valued function of one variable defined by
g(𝑑) = ⟨π‘₯(𝑑), 𝑦(𝑑), 𝑧(𝑑)⟩ = ⟨2𝑑, 𝑑2 , 𝑑3 ⟩.
1
Then, 𝑓 ∘ g is a scalar-valued function of 𝑑,
3
(𝑓 ∘ g)(𝑑) = 𝑒𝑧(𝑑) cos 2π‘₯(𝑑) sin 3𝑦(𝑑) = 𝑒𝑑 cos 4𝑑 sin 3𝑑2 .
To compute its derivative with respect to 𝑑, we first compute
] [
]
[
βˆ‡π‘“ = 𝑓π‘₯ 𝑓𝑦 𝑓𝑧 = βˆ’2𝑒𝑧 sin 2π‘₯ sin 3𝑦 3𝑒𝑧 cos 2π‘₯ cos 3𝑦 𝑒𝑧 cos 2π‘₯ sin 3𝑦 ,
and
gβ€² (𝑑) = ⟨π‘₯β€² (𝑑), 𝑦 β€² (𝑑), 𝑧 β€² (𝑑)⟩ = ⟨2, 2𝑑, 3𝑑2 ⟩,
and then apply the Chain Rule to obtain
𝑑𝑓
𝑑𝑑
= βˆ‡π‘“ (π‘₯(𝑑), 𝑦(𝑑), 𝑧(𝑑)) β‹… gβ€² (𝑑)
⎑
=
[
𝑓π‘₯ (π‘₯(𝑑), 𝑦(𝑑), 𝑧(𝑑)) 𝑓𝑦 (π‘₯(𝑑), 𝑦(𝑑), 𝑧(𝑑)) 𝑓𝑧 (π‘₯(𝑑), 𝑦(𝑑), 𝑧(𝑑))
]
⎣
𝑑π‘₯
𝑑𝑑
𝑑𝑦
𝑑𝑑
𝑑𝑧
𝑑𝑑
⎀
⎦
𝑑𝑦
𝑓𝑧
𝑑π‘₯
+ 𝑓𝑦 (π‘₯(𝑑), 𝑦(𝑑), 𝑧(𝑑)) + 𝑓𝑧 (π‘₯(𝑑), 𝑦(𝑑), 𝑧(𝑑))
𝑑𝑑
𝑑𝑑
𝑑𝑑
= (βˆ’2𝑒𝑧(𝑑) sin 2π‘₯(𝑑) sin 3𝑦(𝑑))(2) + (3𝑒𝑧(𝑑) cos 2π‘₯(𝑑) cos 3𝑦(𝑑))(2𝑑) + (𝑒𝑧(𝑑) cos 2π‘₯(𝑑) sin 3𝑦(𝑑))(3𝑑2 )
= 𝑓π‘₯ (π‘₯(𝑑), 𝑦(𝑑), 𝑧(𝑑))
3
3
3
= βˆ’4𝑒𝑑 sin 4𝑑 sin 3𝑑2 + 6𝑑𝑒𝑑 cos 4𝑑 cos 3𝑑2 + 3𝑑2 𝑒𝑑 cos 4𝑑 sin 3𝑑2 .
β–‘
Example Let 𝑓 : ℝ2 β†’ ℝ be defined by
𝑓 (π‘₯, 𝑦) = π‘₯2 𝑦 + π‘₯𝑦 2 ,
and let g : ℝ2 β†’ ℝ2 be defined by
[
g(𝑠, 𝑑) =
π‘₯(𝑠, 𝑑)
𝑦(𝑠, 𝑑)
]
[
=
2𝑠 + 𝑑
𝑠 βˆ’ 2𝑑
]
.
Then, 𝑓 ∘ g is a scalar-valued function of 𝑠 and 𝑑,
(𝑓 ∘ g)(𝑠, 𝑑) = π‘₯(𝑠, 𝑑)2 𝑦(𝑠, 𝑑) + π‘₯(𝑠, 𝑑)𝑦(𝑠, 𝑑)2 = (2𝑠 + 𝑑)2 (𝑠 βˆ’ 2𝑑) + (2𝑠 + 𝑑)(𝑠 βˆ’ 2𝑑)2 .
To compute its gradient, which includes its partial derivatives with respect to 𝑠 and 𝑑, we first
compute
[
] [
]
βˆ‡π‘“ = 𝑓π‘₯ 𝑓𝑦 = 2π‘₯𝑦 + 𝑦 2 π‘₯2 + 2π‘₯𝑦 ,
and
[
𝐽g (𝑠, 𝑑) =
π‘₯𝑠 π‘₯𝑑
𝑦𝑠 𝑦𝑑
2
]
[
=
2 1
1 βˆ’2
]
,
and then apply the Chain Rule to obtain
βˆ‡(𝑓 ∘ g)(𝑠, 𝑑) = βˆ‡π‘“ (π‘₯(𝑠, 𝑑), 𝑦(𝑠, 𝑑))𝐽g (𝑠, 𝑑)
[
]
[
] π‘₯𝑠 π‘₯𝑑
𝑓π‘₯ (π‘₯(𝑠, 𝑑), 𝑦(𝑠, 𝑑)) 𝑓𝑦 (π‘₯(𝑠, 𝑑), 𝑦(𝑠, 𝑑))
=
𝑦𝑠 𝑦𝑑
[
𝑓π‘₯ (π‘₯(𝑠, 𝑑), 𝑦(𝑠, 𝑑))π‘₯𝑠 + 𝑓𝑦 (π‘₯(𝑠, 𝑑), 𝑦(𝑠, 𝑑))𝑦𝑠
=
]
𝑓π‘₯ (π‘₯(𝑠, 𝑑), 𝑦(𝑠, 𝑑))π‘₯𝑑 + 𝑓𝑦 (π‘₯(𝑠, 𝑑), 𝑦(𝑠, 𝑑))𝑦𝑑
[
[2π‘₯(𝑑)𝑦(𝑑) + 𝑦(𝑑)2 ](2) + [π‘₯(𝑑)2 + 2π‘₯(𝑑)𝑦(𝑑)](1)
=
]
[2π‘₯(𝑑)𝑦(𝑑) + 𝑦(𝑑)2 ](1) + [π‘₯(𝑑)2 + 2π‘₯(𝑑)𝑦(𝑑)](βˆ’2)
[
4(2𝑠 + 𝑑)(𝑠 βˆ’ 2𝑑) + 2(𝑠 βˆ’ 2𝑑)2 + (2𝑠 + 𝑑)2 + 2(2𝑠 + 𝑑)(𝑠 βˆ’ 2𝑑)
=
2(2𝑠 + 𝑑)(𝑠 βˆ’ 2𝑑) + (𝑠 βˆ’ 2𝑑)2 βˆ’ 2(2𝑠 + 𝑑)2 βˆ’ 4(2𝑠 + 𝑑)(𝑠 βˆ’ 2𝑑)
]
.
β–‘
Example Let 𝑓 : ℝ β†’ ℝ be defined by
𝑓 (π‘₯) = π‘₯3 + 2π‘₯2 ,
and let 𝑔 : ℝ2 β†’ ℝ be defined by
𝑔(𝑒, 𝑣) = sin 𝑒 cos 𝑣.
Then 𝑓 ∘ 𝑔 is a scalar-valued function of 𝑒 and 𝑣,
(𝑓 ∘ 𝑔)(𝑒, 𝑣) = (sin 𝑒 cos 𝑣)3 + 2(sin 𝑒 cos 𝑣)2 .
To compute its gradient, which includes partial derivatives with respect to 𝑒 and 𝑣, we first compute
𝑓 β€² (π‘₯) = 3π‘₯2 + 4π‘₯,
and
βˆ‡π‘” =
[
𝑔𝑒 𝑔𝑣
]
=
[
cos 𝑒 cos 𝑣 βˆ’ sin 𝑒 sin 𝑣
]
,
and then use the Chain Rule to obtain
βˆ‡(𝑓 ∘ g)(𝑒, 𝑣) = 𝑓 β€² (𝑔(𝑒, 𝑣))βˆ‡π‘”(𝑒, 𝑣)
= [3(𝑔(𝑒, 𝑣))2 + 4𝑔(𝑒, 𝑣)]
[
]
cos 𝑒 cos 𝑣 βˆ’ sin 𝑒 sin 𝑣
[
]
= [3 sin2 𝑒 cos2 𝑣 + 4 sin 𝑒 cos 𝑣] cos 𝑒 cos 𝑣 βˆ’ sin 𝑒 sin 𝑣
[
]
(3 sin2 𝑒 cos2 𝑣 + 4 sin 𝑒 cos 𝑣) cos 𝑒 cos 𝑣 βˆ’(3 sin2 𝑒 cos2 𝑣 + 4 sin 𝑒 cos 𝑣) sin 𝑒 sin 𝑣 .
=
β–‘
3
Example Let f : ℝ2 β†’ ℝ2 be defined by
[
f (π‘₯, 𝑦) =
𝑓1 (π‘₯, 𝑦)
𝑓2 (π‘₯, 𝑦)
]
[
=
π‘₯2 𝑦
π‘₯𝑦 2
]
,
and let g : ℝ β†’ ℝ2 be defined by
g(𝑑) = ⟨π‘₯(𝑑), 𝑦(𝑑)⟩ = ⟨cos 𝑑, sin π‘‘βŸ©.
Then f ∘ g is a vector-valued function of 𝑑,
f (𝑑) = ⟨cos2 𝑑 sin 𝑑, cos 𝑑 sin2 π‘‘βŸ©.
To compute its derivative with respect to 𝑑, we first compute
[
]
] [
(𝑓1 )π‘₯ (𝑓1 )𝑦
2π‘₯𝑦 π‘₯2
𝐽f (π‘₯, 𝑦) =
,
=
(𝑓2 )π‘₯ (𝑓2 )𝑦
𝑦 2 2π‘₯𝑦
and gβ€² (𝑑) = βŸ¨βˆ’ sin 𝑑, cos π‘‘βŸ©, and then use the Chain Rule to obtain
]
][ β€²
[
π‘₯ (𝑑)
(𝑓1 )π‘₯ (π‘₯(𝑑), 𝑦(𝑑)) (𝑓1 )𝑦 (π‘₯(𝑑), 𝑦(𝑑))
β€²
β€²
(f ∘ g) (𝑑) = 𝐽f (π‘₯(𝑑), 𝑦(𝑑))g (𝑑) =
𝑦 β€² (𝑑)
(𝑓2 )π‘₯ (π‘₯(𝑑), 𝑦(𝑑)) (𝑓2 )𝑦 (π‘₯(𝑑), 𝑦(𝑑))
][
]
[
βˆ’ sin 𝑑
2π‘₯(𝑑)𝑦(𝑑)
π‘₯(𝑑)2
=
𝑦(𝑑)2
2π‘₯(𝑑)𝑦(𝑑)
cos 𝑑
= ⟨2 cos 𝑑 sin 𝑑(βˆ’ sin 𝑑) + cos2 𝑑(cos 𝑑), sin2 𝑑(βˆ’ sin 𝑑) + 2 cos 𝑑 sin 𝑑(cos 𝑑)⟩
= βŸ¨βˆ’2 cos 𝑑 sin2 𝑑 + cos3 𝑑, βˆ’ sin3 𝑑 + 2 cos2 𝑑 sin π‘‘βŸ©.
β–‘
The Implicit Function Theorem
The Chain Rule can also be used to compute partial derivatives of implicitly defined functions in
a more convenient way than is provided by implicit differentiation. Let the equation 𝐹 (π‘₯, 𝑦) = 0
implicitly define 𝑦 as a differentiable function of π‘₯. That is, 𝑦 = 𝑓 (π‘₯) where 𝐹 (π‘₯, 𝑓 (π‘₯)) = 0 for π‘₯ in
the domain of 𝑓 . If 𝐹 is differentiable, then, by the Chain Rule, we can differentiate the equation
𝐹 (π‘₯, 𝑦(π‘₯)) = 0 with respect to π‘₯ and obtain
𝐹π‘₯ + 𝐹𝑦
𝑑𝑦
= 0,
𝑑π‘₯
which yields
𝑑𝑦
𝐹π‘₯
=βˆ’ .
𝑑π‘₯
𝐹𝑦
4
By the Implicit Function Theorem, the equation 𝐹 (π‘₯, 𝑦) = 0 defines 𝑦 implicitly as a function of π‘₯
near (π‘₯0 , 𝑦0 ), where 𝐹 (π‘₯0 , 𝑦0 ) = 0, provided that 𝐹𝑦 (π‘₯0 , 𝑦0 ) βˆ•= 0 and 𝐹π‘₯ and 𝐹𝑦 are continuous near
(π‘₯0 , 𝑦0 ). Under these conditions, we see that 𝑑𝑦/𝑑π‘₯ is defined at (π‘₯0 , 𝑦0 ) as well.
Example Let 𝐹 : ℝ2 β†’ ℝ2 be defined by
𝐹 (π‘₯, 𝑦) = π‘₯2 + 𝑦 2 βˆ’ 4.
The equation 𝐹 (π‘₯, 𝑦) = 0 defines 𝑦 implicitly as a function of π‘₯, provided that 𝐹 satisfies the
conditions of the Implicit Function Theorem.
We have
𝐹π‘₯ = 2π‘₯, 𝐹𝑦 = 2𝑦.
Since both of these partial derivatives are polynomials, and therefore are continuous on all of ℝ2 ,
it follows that if 𝐹𝑦 βˆ•= 0, then 𝑦 can be implicitly defined as a function of π‘₯ at a point where
𝐹 (π‘₯, 𝑦, 𝑧) = 0, and
𝑑𝑦
𝐹π‘₯
π‘₯
=βˆ’
=βˆ’ .
𝑑π‘₯
𝐹𝑦
𝑦
For example, at the point (π‘₯, 𝑦) = (0, 2), 𝐹 (π‘₯, 𝑦) = 0, and 𝐹𝑦 = 4. Therefore, 𝑦 can be implicitly
defined as a function of π‘₯ near this point, and at π‘₯ = 0, we have 𝑑𝑦/𝑑π‘₯ = 0. β–‘
(0)
(0)
(0)
More generally, let 𝐹 : 𝐷 βŠ† ℝ𝑛+1 β†’ ℝ, and let p0 = (π‘₯1 , π‘₯2 , . . . , π‘₯𝑛 , 𝑦 (0) ) ∈ 𝐷 be such that
(0) (0)
(0)
𝐹 (π‘₯1 , π‘₯2 , . . . , π‘₯𝑛 , 𝑦 (0) ) = 0. In this case, the Implicit Function Theorem states that if 𝐹𝑦 βˆ•= 0
near p0 , and all first partial derivatives of 𝐹 are continuous near p0 , then this equation defines 𝑦
as a function of π‘₯1 , π‘₯2 , . . . , π‘₯𝑛 , and
βˆ‚π‘¦
𝐹π‘₯
= βˆ’ 𝑖,
βˆ‚π‘₯𝑖
𝐹𝑦
(0)
𝑖 = 1, 2, . . . , 𝑛.
(0)
(0)
To see this, we differentiate the equation 𝐹 (π‘₯1 , π‘₯2 , . . . , π‘₯𝑛 , 𝑦 (0) ) = 0 with respect to π‘₯𝑖 to obtain
the equation
βˆ‚π‘¦
𝐹π‘₯𝑖 + 𝐹𝑦
= 0,
βˆ‚π‘₯𝑖
where all partial derivatives are evaluated at p0 , and solve for βˆ‚π‘¦/βˆ‚π‘₯𝑖 at p0 .
Example Let 𝐹 : ℝ3 β†’ ℝ be defined by
𝐹 (π‘₯, 𝑦, 𝑧) = π‘₯2 𝑧 + 𝑧 2 𝑦 βˆ’ 2π‘₯𝑦𝑧 + 1.
The equation 𝐹 (π‘₯, 𝑦, 𝑧) = 0 defines 𝑧 implicitly as a function of π‘₯ and 𝑦, provided that 𝐹 satisfies
the conditions of the Implicit Function Theorem.
We have
𝐹π‘₯ = 2π‘₯𝑧 βˆ’ 2𝑦𝑧, 𝐹𝑦 = 2𝑦𝑧 βˆ’ 2π‘₯𝑧, 𝐹𝑧 = π‘₯2 + 2𝑦𝑧 βˆ’ 2π‘₯𝑦.
5
Since all of these partial derivatives are polynomials, and therefore are continuous on all of ℝ3 , it
follows that if 𝐹𝑧 βˆ•= 0, then 𝑧 can be implicitly defined as a function of π‘₯ and 𝑦 at a point where
𝐹 (π‘₯, 𝑦, 𝑧) = 0, and
𝑧π‘₯ = βˆ’
𝐹π‘₯
2𝑦𝑧 βˆ’ 2π‘₯𝑧
= 2
,
𝐹𝑧
π‘₯ + 2𝑦𝑧 βˆ’ 2π‘₯𝑦
𝑧𝑦 = βˆ’
𝐹𝑦
2π‘₯𝑧 βˆ’ 2𝑦𝑧
= 2
.
𝐹𝑧
π‘₯ + 2𝑦𝑧 βˆ’ 2π‘₯𝑦
For example, at the point (π‘₯, 𝑦, 𝑧) = (1, 0, βˆ’1), 𝐹 (π‘₯, 𝑦, 𝑧) = 0, and 𝐹𝑧 = 1. Therefore, 𝑧 can be
implicitly defined as a function of π‘₯ and 𝑦 near this point, and at (π‘₯, 𝑦) = (1, 0), we have 𝑧π‘₯ = 2
and 𝑧𝑦 = βˆ’2. β–‘
We now consider the most general case: let F : 𝐷 βŠ† ℝ𝑛+π‘š β†’ β„π‘š , and let
(0)
(0)
(0)
(0)
(0)
p0 = (π‘₯1 , π‘₯2 , . . . , π‘₯(0)
𝑛 , 𝑦1 , 𝑦2 , . . . , π‘¦π‘š ) ∈ 𝐷
be such that
(0)
(0)
(0)
(0)
(0)
F(π‘₯1 , π‘₯2 , . . . , π‘₯(0)
𝑛 , 𝑦1 , 𝑦2 , . . . , π‘¦π‘š ) = 0.
If we differentiate this system of equations with respect to π‘₯𝑖 , we obtain the systems of linear
equations
βˆ‚π‘¦π‘š
βˆ‚π‘¦1
βˆ‚π‘¦2
F π‘₯𝑖 + F 𝑦1
+ F 𝑦2
+ β‹… β‹… β‹… + F π‘¦π‘š
= 0, 𝑖 = 1, 2, . . . , 𝑛,
βˆ‚π‘₯𝑖
βˆ‚π‘₯𝑖
βˆ‚π‘₯𝑖
where all partial derivatives are evaluated at p0 .
(0) (0)
(0)
To examine the solvability of these systems of equations, we first define x0 = (π‘₯1 , π‘₯2 , . . . , π‘₯𝑛 ),
and denote the component functions of the vector-valued function F by F = ⟨𝐹1 , 𝐹2 , . . . , πΉπ‘š ⟩. We
then define the Jacobian matrices
⎀
⎑
βˆ‚πΉ1
βˆ‚πΉ1
βˆ‚πΉ1
(p
)
(p
)
(p
)
0
0
0
βˆ‚π‘₯1
βˆ‚π‘₯2
βˆ‚π‘₯𝑛
βŽ₯
⎒ βˆ‚πΉ
βˆ‚πΉ2
2
(p
)
⎒ βˆ‚π‘₯12 (p0 ) βˆ‚πΉ
0
βˆ‚π‘₯
βˆ‚π‘₯𝑛 (p0 ) βŽ₯
2
𝐽x,F (p0 ) = ⎒
βŽ₯,
⎦
⎣
βˆ‚πΉπ‘š
βˆ‚πΉπ‘š
βˆ‚πΉπ‘š
βˆ‚π‘₯1 (p0 )
βˆ‚π‘₯2 (p0 )
βˆ‚π‘₯𝑛 (p0 )
⎑
⎒
⎒
𝐽y,F (p0 ) = ⎒
⎣
and
⎑
⎒
⎒
𝐽y (x0 ) = ⎒
⎣
βˆ‚π‘¦1
βˆ‚π‘¦1 (p0 )
βˆ‚π‘¦2
βˆ‚π‘¦1 (p0 )
βˆ‚πΉ1
βˆ‚π‘¦2 (p0 )
βˆ‚πΉ2
βˆ‚π‘¦2 (p0 )
βˆ‚πΉ1
βˆ‚π‘¦π‘š (p0 )
βˆ‚πΉ2
βˆ‚π‘¦π‘š (p0 )
βˆ‚π‘¦π‘š
βˆ‚π‘¦1 (p0 )
βˆ‚πΉπ‘š
βˆ‚π‘¦2 (p0 )
βˆ‚πΉπ‘š
βˆ‚π‘¦π‘š (p0 )
βˆ‚π‘¦1
βˆ‚π‘₯1 (x0 )
βˆ‚π‘¦2
βˆ‚π‘₯1 (x0 )
βˆ‚π‘¦1
βˆ‚π‘₯2 (x0 )
βˆ‚π‘¦2
βˆ‚π‘₯2 (x0 )
βˆ‚π‘¦1
βˆ‚π‘₯𝑛 (x0 )
βˆ‚π‘¦1
βˆ‚π‘₯𝑛 (x0 )
βˆ‚π‘¦π‘š
βˆ‚π‘₯1 (x0 )
βˆ‚π‘¦π‘š
βˆ‚π‘₯2 (x0 )
βˆ‚π‘¦1
βˆ‚π‘₯𝑛 (x0 )
6
⎀
βŽ₯
βŽ₯
βŽ₯,
⎦
⎀
βŽ₯
βŽ₯
βŽ₯.
⎦
Then, from our previous differentiation with respect to π‘₯𝑖 , for each 𝑖 = 1, 2, . . . , 𝑛, we can concisely
express our systems of equations as a single system
𝐽x,F (p0 ) + 𝐽y,F (p0 )𝐽y (x0 ) = 0.
If the matrix 𝐽y,F (p0 ) is invertible (also nonsingular), which is the case if and only if its determinant is nonzero, and if all first partial derivatives of F are continuous near p0 , then the equation
F(p) = 0 implicitly defines 𝑦1 , 𝑦2 , . . . , π‘¦π‘š as a function of π‘₯1 , π‘₯2 , . . . , π‘₯𝑛 , and
𝐽y (x0 ) = βˆ’[𝐽y,F (p0 )]βˆ’1 𝐽x,F (p0 ),
where [𝐽y,F (p0 )]βˆ’1 is the inverse of the matrix 𝐽y,F (p0 ).
Example Let F : ℝ4 β†’ ℝ2 by defined by
[
] [
]
𝐹1 (π‘₯, 𝑦, 𝑒, 𝑣)
π‘₯𝑒 + 𝑦 2 𝑣
F(π‘₯, 𝑦, 𝑠, 𝑑) =
=
.
𝐹2 (π‘₯, 𝑦, 𝑒, 𝑣)
π‘₯2 𝑣 + 𝑦𝑒 + 1
Then the vector equation F(π‘₯, 𝑦, 𝑒, 𝑣) = 0 implicitly defines (𝑒, 𝑣) as a function of (π‘₯, 𝑦), provided
that F satisifes the conditions of the Implicit Function Theorem. We will compute the partial
derivatives of 𝑒 and 𝑣 with respect to π‘₯ and 𝑦, at a point that satisfies this equation.
We have
[
] [
]
βˆ‚πΉ1
βˆ‚πΉ1
𝑒 2𝑦𝑣
βˆ‚π‘₯
βˆ‚π‘¦
𝐽(π‘₯,𝑦),F (π‘₯, 𝑦, 𝑒, 𝑣) = βˆ‚πΉ2 βˆ‚πΉ2 =
,
2π‘₯𝑣 𝑒
βˆ‚π‘₯
βˆ‚π‘¦
[ βˆ‚πΉ1 βˆ‚πΉ1 ] [
]
π‘₯ 𝑦2
βˆ‚π‘’
βˆ‚π‘£
𝐽(𝑒,𝑣),F (π‘₯, 𝑦, 𝑒, 𝑣) = βˆ‚πΉ2 βˆ‚πΉ2 =
.
𝑦 π‘₯2
βˆ‚π‘’
βˆ‚π‘£
From the formula for the inverse of a 2 × 2 matrix,
]βˆ’1
[
]
[
1
π‘Ž 𝑏
𝑑 βˆ’π‘
=
,
𝑐 𝑑
π‘Žπ‘‘ βˆ’ 𝑏𝑐 βˆ’π‘ π‘Ž
we obtain
[
𝐽(𝑒,𝑣) (π‘₯, 𝑦) =
𝑒π‘₯ 𝑒𝑦
𝑣π‘₯ 𝑣𝑦
]
= βˆ’[𝐽(𝑒,𝑣),F (π‘₯, 𝑦, 𝑒, 𝑣)]βˆ’1 𝐽(π‘₯,𝑦),F (π‘₯, 𝑦, 𝑒, 𝑣)
[ 2
][
]
1
π‘₯ βˆ’π‘¦ 2
𝑒 2𝑦𝑣
= βˆ’ 3
π‘₯
2π‘₯𝑣 𝑒
π‘₯ βˆ’ 𝑦 3 βˆ’π‘¦
[ 2
]
1
π‘₯ 𝑒 βˆ’ 2π‘₯𝑦 2 𝑣 2π‘₯2 𝑦𝑣 βˆ’ 𝑦 2 𝑒
=
.
2π‘₯2 𝑣 βˆ’ 𝑦𝑒
π‘₯𝑒 βˆ’ 2𝑦 2 𝑣
𝑦 3 βˆ’ π‘₯3
These partial derivatives can then be evaluated at any point (π‘₯, 𝑦, 𝑒, 𝑣) such that F(π‘₯, 𝑦, 𝑒, 𝑣) = 0,
such as (π‘₯, 𝑦, 𝑒, 𝑣) = (0, 1, 0, βˆ’1). Note that the matrix 𝐽(𝑒,𝑣),F (π‘₯, 𝑦, 𝑒, 𝑣) is not invertible (that is,
singular) if its determinant π‘₯3 βˆ’ 𝑦 3 = 0; that is, if π‘₯ = 𝑦. When this is the case, (𝑒, 𝑣) can not be
implicitly defined as a function of (π‘₯, 𝑦). β–‘
7
Practice Problems
1. For each of the following pairs of functions f and g, use the Chain Rule to compute all of the
first partial derivatives of f ∘ g with respect to the independent variables of g.
(a)
𝑓 (π‘₯, 𝑦, 𝑧) =
√
π‘₯2 + 𝑦 2 + 𝑧 2 ,
g(𝑑) = βŸ¨π‘’π‘‘ cos 𝑑, 𝑒𝑑 sin 𝑑, 𝑑𝑒𝑑 ⟩
(b)
𝑓 (π‘₯, 𝑦) = ln
√
π‘₯2 + 𝑦 2 ,
g(𝑒, 𝑣) = βŸ¨π‘’ cos 𝑣, 𝑒 sin π‘£βŸ©
(c)
f (π‘₯, 𝑦) = ⟨π‘₯3 + 3𝑦 2 , 𝑦 3 βˆ’ 3π‘₯2 ⟩,
g(𝑑) = ⟨cos 𝑑, sin π‘‘βŸ©
(d)
f (𝑒, 𝑣) =
〈(
1+
(
𝑣
𝑒)
𝑣
𝑒)
𝑉
𝑒
cos
cos 𝑒, 1 + cos
sin 𝑒, sin
2
2
2
2
2
2
βŒͺ
,
g(π‘₯, 𝑦) = ⟨π‘₯2 𝑦, π‘₯𝑦 2 ⟩
2. Compute the indicated derivatives or partial derivatives of the functions that are implicitly
defined by the following equations. Then obtain the equation of the tangent line (or tangent
plane) to the curve (surface) at the indicated point.
√
(a) 𝐹 (π‘₯, 𝑦) = (π‘₯2 + 𝑦 2 )2 βˆ’ 2(π‘₯2 βˆ’ 𝑦 2 ) = 0, 𝑑𝑦/𝑑π‘₯, (π‘₯0 , 𝑦0 ) = ( 3/2, 1/2)
√
(b) 𝐹 (π‘₯, 𝑦, 𝑧) = π‘₯2 + 𝑦 2 + 𝑧 2 βˆ’ 8 π‘₯2 + 𝑦 2 + 15 = 0, βˆ‚π‘§/βˆ‚π‘₯, βˆ‚π‘§/βˆ‚π‘¦, (π‘₯0 , 𝑦0 , 𝑧0 ) = (4, 0, 1)
3. (Just for fun) Describe the graphs of the curves or surfaces defined by the equations in
Problem 2.
Additional Practice Problems
Additional practice problems from the recommended textbooks are:
βˆ™ Stewart: Section 11.5, Exercises 1-11 odd, 17, 19 odd
βˆ™ Marsden/Tromba: Section 2.5, Exercises 5, 9, 11; Section 3.5, Exercises 3, 5
8