506 APPENDIX A. MATHEMATICS BACKGROUND Figure A.9: A

506
APPENDIX A. MATHEMATICS BACKGROUND
Figure A.9: A strictly concave-contoured (strictly quasiconcave) function
There are functions for which the contours look like those of a concave
function but which are not themselves concave. An example here would
be ' (f (x)) where f is a concave function and is an arbitrary monotonic
transformation.
These remarks lead us to the de…nition:
De…nition A.23 A function f is (strictly) concave-contoured if all the sets
B(y0 ) in (A.31) are (strictly) convex.
A synonym for (strictly) concave-contoured is (strictly) quasiconcave. Try
not to let this (unfortunately necessary) jargon confuse you. Take, for example,
a “conventional” looking utility function such as
U (x) = x1 x2 :
(A.32)
According to de…nition A.23 this function is strictly quasiconcave: if you draw
the set of points B( ) := f(x1 ; x2 ) : x1 x2
g you will get a strictly convex
set. Furthermore, although U in (A.32) is not a concave function, it is a simple
transformation of the strictly concave function
^ (x) = log x1 + log x2 ;
U
(A.33)
^ . But when we draw those contours
and has the same shape of contour map as U
on a diagram with the usual axes we would colloquially describe their shape
A.7. MAXIMISATION
507
as being “convex to the origin”! There is nothing seriously wrong here: the
de…nition, the terminology and our intuitive view are all correct; it is just a
matter of the way in which we visualise the function. Finally, the following
complementary property is sometimes useful:
De…nition A.24 A function f is (strictly) quasiconvex if
siconcave.
A.6.6
f is (strictly) qua-
The Hessian property
Rn to R. Let fij (x) denote
Consider a twice-di¤erentiable function f from D
@ 2 f (x)
@xi @xj . The symmetric matrix
2
f11 (x) f12 (x)
6 f21 (x) f22 (x)
6
4 :::
:::
fn1 (x) fn2 (x)
is known as the Hessian matrix of f .
3
::: f1n (x)
::: f2n (x) 7
7
5
:::
:::
::: fnn (x)
De…nition A.25 The Hessian matrix of f at x is negative semide…nite if, for
any vector w 2 Rn , it is true that
n X
n
X
wi wj fij (x)
0:
i=1 j=1
A twice-di¤erentiable function f from D to R is concave if and only if f is
negative semi-de…nite for all x 2 D.
De…nition A.26 The Hessian matrix of f at x is negative de…nite if, for any
vector w 2 Rn ,w 6= 0, it is true that
n X
n
X
wi wj fij (x) < 0:
i=1 j=1
A twice-di¤erentiable function f from D to R is strictly concave if f is
negative de…nite for all x 2 D; but the reverse is not true – a strictly concave
function f may have a negative semi-de…nite Hessian.
If the Hessian of f is negative de…nite for all x 2 D we will say that f has
the Hessian property.
A.7
Maximisation
Because a lot of economics is concerned with optimisation we will brie‡y overview
the main techniques and results. However this only touches the edge of a very
big subject: you should consult the references in section A.9 for more details.
508
APPENDIX A. MATHEMATICS BACKGROUND
A.7.1
The basic technique
The problem of maximising a function of n variables
max f (x)
x2X
(A.34)
X Rn is straightforward if the function f is di¤erentiable and the domain X
is unbounded. We adopt the usual …rst-order condition (FOC)
@f (x)
= 0; i = 1; 2; :::; n
@xi
(A.35)
and then solve for the values of (x1 ; x2 ; :::; xn ) that satisfy (A.35). However the
FOC is, at best, a necessary condition for a maximum of f . The problem is
that the FOC is essentially a simple hill-climbing rule: “if I’m really at the top
of the hill then the ground must be ‡at just where I’m standing.” There are a
number of di¢ culties with this:
The rule only picks out “stationary points” of the function f . As Figure
A.10 illustrates, this condition is satis…ed by a minimum (point C) as well
as a maximum (point A), or by a point of in‡ection (E). To eliminate
points such as C and E we may look at the second-order conditions which
essentially require that at the top of the hill (a point such as A) the slope
must be (locally) decreasing in every direction.
Even if we eliminate minima and points of in‡ection the FOC may pick
out multiple “local” maxima. In Figure A.10 points A and D are each
local maxima, but obviously A is the point that we really want. we may
be able to eliminate. This problem may be sidestepped by introducing
a priori restrictions on the nature of the function f that eliminate the
possibility of multiple stationary points –for example by requiring that f
be strictly concave.
If we have been careless in specifying the problem then the hill-climbing
rule may be completely misleading. We have assumed that each x-component
can range freely from 1 to +1. But suppose – as if often in the case
in economics – that the de…nition of the variable is such that only nonnegative values make sense. Then it is clear from Figure A.10 that A is
an irrelevant point and the maximum is at B. In climbing the hill we have
reached a logical “wall” and we can climb no higher.
Likewise if we have overlooked the requirement that the function f be
everywhere di¤erentiable the hill-climbing rule represented by the FOC
may be misleading. If we draw the function
8
x 1
< x
f (x) =
:
2 x
x>1
A.7. MAXIMISATION
509
Figure A.10: Di¤erent types of stationary point
it is clear that it is continuous and has a maximum at x = 1. But the
FOC as stated in (A.35) is useless because the di¤erential of f is unde…ned
exactly at x = 1.
If we can sweep these di¢ culties aside then we can use the solution to the
system of equations provided by the FOC in a powerful way. To see what is
usually done, slightly rewrite the maximisation problem (A.34) as
max f (x; p)
x2Rn
(A.36)
where p represents a vector of parameters, a set of numbers that are …xed for the
particular maximisation problem in hand but which can be used to characterise
the di¤erent members of a whole class of maximisation problems and their
solutions. For example p might represent prices (outside the control of a small
…rm and therefore taken as given) and might x represent the list of quantities
of inputs and outputs that the …rm chooses in its production process; pro…ts
depend on both the parameters and the choice variables.
We can then treat the FOC (A.35) as a system of n equations in n unknowns
(the components of x).Without further regularity conditions such a system is
not guaranteed to have a solution nor, if it has a solution, will it necessarily
be unique. However, if it does then we can write it as a function of the given
510
APPENDIX A. MATHEMATICS BACKGROUND
parameters p:
x1 = x1 (p)
x2 = x2 (p)
:::
xn = xn (p)
9
>
>
=
(A.37)
>
>
;
We may refer to the functions x1 ( ) in (A.37) as the response functions in that
they indicate how the optimal values of the choice variables (x ) would change
in response to changes in values of the given parameters p.
A.7.2
Constrained maximisation
By itself the basic technique in section A.7.1 is of limited value in economics:
optimisation is usually subject to some side constraints which have not yet been
introduced. We now move on to a simple case of constrained optimisation that,
although restricted in its immediate applicability to economic problems, forms
the basis of other useful techniques. We consider the problem of maximising a
di¤erentiable function of n variables
max f (x; p)
(A.38)
x2Rn
subject to the m equality constraints
G1 (x; p) = 0
G2 (x; p) = 0
:::
Gm (x; p) = 0
9
>
>
=
(A.39)
>
>
;
There is a standard technique for solving this kind of problem: this is to incorporate the constraint in a new maximand. To do this introduce the Lagrange
multipliers 1 ; :::; m , a set of non-negative variables, one for each constraint.
The constrained maximisation problem in the n variables x1 ; :::; xn , is equivalent
to the following (unconstrained) maximisation problem in the n + m variables
x1 ; :::; xn ; 1 ; :::; m ,
L(x; ; p) := f (x; p)
m
X
jG
j
(x; p)
(A.40)
j=1
, where L is the Lagrangean function. By introducing the Lagrange multipliers
we have transformed the constrained optimisation problem into one that is of
the same format as in section A.7.1, namely
max L(x; ; p)
x;
(A.41)
A.7. MAXIMISATION
511
The FOC for solving (A.41) are found by di¤erentiating (A.40) with respect
to each of the n + m variables and setting each to zero.
@L(x ; ; p)
@xi
@L(x ; ; p)
@ j
=
0; i = 1; :::; n
(A.42)
=
0; j = 1; :::; m
(A.43)
where the “ ” means that the di¤erential is being evaluated at a solution point
(x ; ). So the FOC consist of the n equations
m
@f (x ; p) X
=
@xi
j=1
@Gj (x ; p)
; i = 1; :::; n
j
@xi
(A.44)
plus the m constraint equations (A.39) evaluated at x . We therefore have a
system of n + m equations (A.44,A.39) in n + m variables.
As in section A.7.1, if the system of equations does have a unique solution
(x ; ), then this can be written as a function of the parameters p:
9
x1 = x1 (p) >
>
=
x2 = x2 (p)
(A.45)
:::
>
>
;
xn = xn (p)
1
2
m
= 1 (p)
= 2 (p)
:::
= m (p)
9
>
>
=
>
>
;
(A.46)
Once again the functions x1 ( ) in (A.45) are the response functions and have the
same interpretation. The Lagrange multipliers in (A.46) also have an interesting
interpretation which is handled in A.7.4 below.
If the equations (A.44,A.39) yield more than one solution, but f in (A.38)
is quasiconcave and the set of x satisfying (A.39) is convex then we can appeal
to the commonsense result in Theorem A.12.
A.7.3
More on constrained maximisation
Now modify the problem in section A.7.2 in two ways that are especially relevant
to economic problems
Instead of allowing each component xi to range freely from 1 to +1.we
restrict to some interval of the real line. So we will now write the domain
restriction x 2 X where we will take X to be the non-negative orthant of
Rn . The results below can be adapted to other speci…cations of X.
512
APPENDIX A. MATHEMATICS BACKGROUND
We replace the equality constraints in
equality constraints
G1 (x; p)
G2 (x; p)
:::
Gm (x; p)
(A.39) by the corresponding in9
0 >
>
=
0
(A.47)
>
>
;
0
This is reasonable in economic applications of optimisation. For example
the appropriate way of stating a budget constraint is “expenditure must
not exceed income” rather than “...must equal...”.
So the problem is now
max f (x; p)
x2X
subject to (A.47). The solution to this modi…ed problem is similar to that
for the standard Lagrangean – see Intriligator (1971), pages 49-60. Again we
transform the problem by forming a Lagrangean (as in A.40):
max
x2X; >0
L(x; ; p)
(A.48)
However, instead of (A.42, A.43)we now have the following FOCs:
@L(x ; ; p)
@xi
@L(x ; ; p)
xi
@xi
=
0; i = 1; :::; n
(A.49)
0; i = 1; :::; n
(A.50)
0; j = 1; :::; m
(A.51)
0; j = 1; :::; m
(A.52)
and
@L(x ; ; p)
@ j
@L(x ; ; p)
j
@ j
=
This set of equations and inequalities is conventionally known as the KuhnTucker conditions. They have important implications relating the values of the
variables and the Lagrange multipliers at the optimum.
Applying this result we …nd
@f (x ; p)
@xi
m
X
j=1
@Gj (x ; p)
; i = 1; :::; n
j
@xi
(A.53)
with (A.44) if xi > 0. Note that if, for some i, xi = 0 we could have strict
inequality in (A.53). Figure A.11 illustrates this possibility for a case where the
objective function is strictly concave: note that the conventional condition of
“slope=0” (A.42) (which would appear to be satis…ed at point A) is irrelevant
here since a point such as A would violate the constraint xi 0; at the optimum
(point B) the Lagrangean has a strictly decreasing slope. Similar interpretations
will apply to the Lagrange multipliers:
A.7. MAXIMISATION
513
Figure A.11: A case where xi = 0 at the optimum
1. If the Lagrange multiplier associated with constraint j is strictly positive
at the optimum ( j > 0), then it must be binding (Gj (x ; p) = 0).
2. Conversely one could have an optimum where one or more Lagrange multiplier ( j = 0) is zero in which case the constraint may be slack – i.e.
not binding –(Gj (x ; p) < 0).
So, for each j at the optimum, there is at most one inequality condition: if
there is a strict inequality on the Lagrange multiplier then the corresponding
constraint must be satis…ed with equality (case 1); if there is a strict inequality
on the constraint then the corresponding Lagrange multiplier must be equal
to zero (case 2). These facts are conventionally known as the complementary
slackness condition. However, note that one can have cases where both the
Lagrange multiplier ( j = 0) and the constraint is binding (Gj (x ; p) = 0).
Again if the system (A.53,A.47) yields a unique solution it can be written as
a function of the parameters p which in turn determines the response functions;
but if it yields more than one solution, but f in (A.38) is quasiconcave and the
set of x satisfying (A.47) is convex then we can use the following.
Theorem A.12 If f : Rn 7! R is quasiconcave and A
set of values x that solve the problem
max f (x) subject to x 2 A
is convex.
Rn is convex then the