Econ 20740: Analysis of Collective Decision-Making

Econ 20740: Analysis of Collective Decision-Making
Richard Van Weelden∗
May 22, 2017
1
Aggregation of Preferences
In this course we will be interested in environments in which decisions need to be made
by a group. We begin with the topic of social choice theory. When looking at decisions
made by groups, there are two principal components: the aggregation of preferences and the
aggregation of information. Social choice theory largely deals with the first of these, and
we’ll consider informational issues later in the course. In the first part of the class we ask the
questions: If we knew everyone’s preferences, how would we determine which outcome(s) are
preferable to others. For this analysis we will follow Chapters 1 and 7 of Taylor and Pacelli
(2008).
What are some of the issues inherent in aggregating preferences? Recall that in previous
economics classes (Econ 200, Econ 201) we represented individual’s preferences by a utility
function. One alternative is preferred to another if it gives the individual a higher utility.
What happens when there are multiple individuals and they disagree on which alternative
is preferred? If one alternative pareto dominates another the comparison is easy: if all
individuals would prefer A to B then clearly A should be preferred to B by the group.
However we have seen in Econ 200/201 that there are generally many pareto dominant
allocations, so we won’t always be able to rely on a pareto criterion when comparing different
allocations. Allocations that are pareto dominated should be rejected, but how do we choose
among different pareto optimal choices?
Part of the complexity is that it is difficult to make inter-personal comparisons. Recall
that a utility function is meant to represent preferences. This allows us to analyze the
consumer’s problem by maximizing an objective function using calculus. Recall that in
order to represent an individual’s preferences by a utility function we need the individual’s
preferences to be complete and transitive:
1. Complete: For any two alternatives A and B either A B or B A.
2. Transitive: If A B and B C then A C.
∗
Assistant Professor of Economics, University of Chicago. Email: [email protected].
1
Complete means simply that the individual is always capable of comparing the two alternatives and determining which alternative she prefers. Transitive means that, if A is
preferred to B, and B is preferred to C, then A must be preferred to C. This is necessary since, if C were preferred to A then the individual would not be able to choose when
A, B and C are all available. If an individual has transitive preferences then we can assign a higher utility level to more preferred alternatives and compare alternatives by their
associated utility levels.
In this class we will consider the case where there are n > 1 individuals/voters, and each
individual i = 1, . . . , n has preferences i over the some set of alternatives X. We will be
interested in finding a social choice function f which maps the individual’s preferences and
the the available alternatives into the outcome that is chosen or preferred. That is, a social
choice function is a mapping,
f : X × (1 , . . . , n ) → X
If there is not a unique alternative preferred then this mapping would not be a function but
rather a correspondence with a set of preferred alternatives. For this reason we will often
talk about a social choice rule.
Notice that one way to choose an alternative is to take a utilitarian approach. This
approach is often associated with the philosopher Jeremy Bentham. If individual preferences
i are represented by a utility function ui we can define a social welfare function
U (x) = Σni=1 ui (x)
and select x ∈ X to maximize U (x). However, while we can (and sometimes do) use a
utilitarian approach it has some drawbacks. As we have seen in Econ 200/201, if preferences
can be represented by a utility function u the same preferences can also be represented
by any monotonic transformation of u. So which utility function to use? Generally with
heterogenous preferences we need different utility functions to represent the preferences of
different individuals, and how to choose which of the utility functions to use? Depending
on which utility function we choose would give a different utilitarian social welfare function
which would lead to a different choice. We will explore these issues in problem set 1.
2
Collective Choice with Two Alternatives
We begin, by supposing there are only two alternatives, so X = {A, B}. As mentioned above
one approach is to construct a utilitarian social welfare function, and select the alternative
that maximizes the sum of all voters utilities. Another approach is to consider which alternative is preferred by a majority of the individuals. We could then define the social choice
function that always selects the alternative that a majority of the voters prefer. This is
perhaps the most common way of making decisions at least in democratic societies: vote
by majority rule. Assume there are n individuals/voters and assume that n is odd. Also
assume that no individual is indifferent between A and B. By assuming an odd number of
individuals and no assuming there is no indifference we ensure that there will be no ties.
Definition 1. A social choice function satisfies majority rule if alternative A is preferred to
B if and only if A i B for a majority of individuals.
2
While majority rule is not perfect—it does not take into account the intensity of preferences in any way—it has several advantages. Moreover, if the individuals have binary
preferences, so either they prefer A or they prefer B, it is not clear what intensity of preference means. To say that person 1 has a stronger preference for one alternative than person
2 requires some other dimension that the individuals care about so that we can consider
the differing willingness to sacrifice in the other dimension of the sake of this dimension.
For example, person 1 might have a greater willingness to pay money for their preferred
alternative or to spend time showing up to vote or attending a meeting to get their preferred
alternative, etc. We will discuss such environments later in the class, but for now consider
only a setting in which there are only two alternatives and no other relevant dimensions with
which to compare intensity of preferences.
The first advantage of majority rule is that it always allows us to make a choice: if there
are an odd number of individuals and no individual is indifferent then one alternative is
preferred to the other.1 Second, it treats all voters equally: if the preferences of two individuals are reversed the same alternative would be preferred. Third, it treats the alternatives
equally, and is not biased in favor of one alternative or the other. Finally, if one alternative
becomes more popular (more people prefer it than before), it can only become more likely
to be chosen. While there are many other decision making rules we could consider (appoint
a dictator, weighted voting, super-majority requirements, etc), May’s theorem shows that
majority rule is the unique decision making mechanism to satisfy all of these criteria.
Theorem 1. (May 1952) Suppose there are n (odd) individuals and two alternatives, X =
{A, B}. Then majority rule is the unique social choice function to satisfy the following
properties:
• Anonymity: For any i, j and preferences i , j , 0i , 0j such that A i B, A ≺j B, A ≺0i
B, A 0j B, then f ({A, B}, (1 , . . . , n )) = f ({A, B}, (1 , . . . , 0i , . . . , 0j , . . . , n )).
• Neutrality: For each x ∈ {A, B}, if f ({A, B}, (1 , . . . , n )) = x and, B 0i A if and
only if A i B for all i = 1, . . . , n, then f ({A, B}, (01 , . . . , 0n )) 6= x.
• Monotonicity: Consider two preference profiles i , 0i for some individual i such that
A ≺i B, and A 0i B. Then if f ({A, B}, (1 , . . . , n )) = A then
f ({A, B}, (1 , . . . , i−1 , 0i , i+1 , . . . , n )) = A. Similarly, if
f ({A, B}, (1 , . . . , i−1 , 0i , i+1 , . . . , n )) = B then f ({A, B}, (1 , . . . , n )) = B.
Anonymity says that it only matters the number of people who prefer A to B, not who
those individuals are, in order to determine whether A is preferred to B. This is usually
considered desirable since it treats everyone equally and gives them an equal weight. However
it is violated in many rules used in practice, from systems that give only some members voting
rights to the electoral college used to elect U.S. Presidents. Neutrality reflects that neither
alternative is advantaged over the other, so if all preferences are reversed this would reverse
which alternative is chosen. Neutrality is natural if we don’t have an ex-ante reason to
think one alternative is preferred to another; it is violated, for example, with super-majority
1
And if we had an even number of voters split evenly over which alternative they preferred it would be
natural to be indifferent over those alternatives. Similarly if some voters are indifferent.
3
requirements or when unanimity is required (e.g. jury deliberations). Finally, monotonicity
says that increasing the number of individuals who prefer one alternative can never cause
that alternative not to be chosen.
So we see that one way to make decisions between two alternatives is by majority rule.
And if we want to avoid making interpersonal utility comparisons May’s theorem tells us that,
in environments in which neutrality, anonymity, and monotonicity are desirable properties,
majority rule is the “correct” way to make decisions. As we will see when there are more
than two alternatives things get more complicated.
3
The Condorcet Paradox
Now we consider collective choice with more than two alternatives. For simplicity, we assume
none of the voters are indifferent over any of the alternatives. In general, having three
alternatives is sufficient to illustrate the difficulties. When there are three alternatives there
will not always exist one alternative that is the first choice of a majority of voters. One way
to compare alternatives is to compare each two alternatives in X by majority rule, and say
that one alternative x ∈ X is preferred to another, x0 , if a majority of voters prefer x to x0 .2
This was suggested by the Marquis de Condorcet, and any alternative that is not defeated
by any other alternative is referred to as a Condorcet winner.
Definition 2. An alternative x ∈ X is a Condorcet winner if, for all x0 ∈ X, x i x0 for at
least n/2 individuals.
If a Condorcet winner exists it should (arguably) be preferred to any alternative that
is not a Condorcet winner: a non-Condorcet winner means that more than half of the
individuals could be made better off. However, a Condorcet winner may not exist.
Let X = {A, B, C}, and assume that n = 3. So we have three voters and three alternatives. Consider the following preferences:
• Voter 1: A 1 B 1 C
• Voter 2: B 2 C 2 A
• Voter 3: C 3 A 3 B
Notice that, voter 1 and voter 2 both prefer B to C. So, since a majority prefers B
to C we should prefer B to C. And since a majority (1 and 3) prefer A to B we should
prefer A to B. So, for preferences to be transitive we must have that A is preferred to C.
However, voters 2 and 3 both prefer C to A, so a majority prefers C to A. This means that
aggregating preferences by Condorcet’s method violates transitivity and a Condorcet winner
may not exist.
This is the famous voting paradox or Condorcet paradox, discovered by the Marquis de
Condorcet in the late eighteenth century. It shows that, even if each voter has transitive
preferences, aggregating preferences by majority rule can produce non-transitive preferences.
2
Another reason we might apply this method is that there could be three alternatives but only two of
them would be available at each time.
4
And we know from Econ 200/201 that having non-transitive preferences means that we
can’t use Condorcet’s method to represent the preferences with something akin to a utility
function.
4
Aggregating Preferences with 3 (or more) Alternatives
The Condorcet Paradox hints at the difficulties associated with constructing a social choice
function with more than two alternatives. We can use the Condorcet method and compare
each tuple. However, as the above example demonstrates a Condorcet winner need not exist.
This means that we wont always be able to make a choice. However, it will exist in many
circumstances.
Example 1. Let X = {A, B, C}, and assume that n = 5. Consider the following preferences:
• Voter 1: A 1 B 1 C
• Voter 2: B 2 C 2 A
• Voter 3: C 3 A 3 B
• Voter 4: A 4 C 4 B
• Voter 5: B 5 A 5 C
Notice that since A is preferred to B by voters 1, 3, and 4, and preferred to C by voters 1,
4, and 5, that A is the Condorcet winner here.
In the above example a unique Condorcet winner exists. This is not always the case:
there could exist multiple Condorcet winners. If we add a sixth voter with the following
preferences there are two Condorcet winners.
• Voter 6: B 6 A 6 C
Notice that since A is preferred to B by three voters (1, 3, 4) and to C by four voters (1, 4,
5, 6), A is not defeated by either B or C and so remains a Condorcet winner. B is preferred
to A by three voters (2, 5, 6), and to C by four voters (1, 2, 5, 6) and B is also a Condorcet
winner.
Several other methods have been suggested. Perhaps the most common (e.g. in most
elections in the United States and Westminster Democracies) is to choose by plurality rule.
Under plurality rule, the alternative that is the first choice of the most voters is the one
selected (if tied we could declare multiple winners then decide by some other criteria such a
flipping a coin). Note however that plurality rule does not guarantee that a Condorcet winner (if it exists) will be selected. This was arguably the case in 2000 when George W. Bush
defeated Al Gore but, given that a majority of the Ralph Nader voters probably preferred
5
Gore to Bush, it is possible that Al Gore was the Condorcet winner.3 In the literature this
often described as saying that plurality rule violates “Independence of Irrelevant Alternatives”. Since Ralph Nader was clearly not going to be elected, individuals’ ranking of Nader
is in some sense irrelevant relative to their ranking of Bush and Gore. Arguably, whether
Bush or Gore would be selected should depend only each voter’s comparison of Bush and
Gore. However, plurality rule does not guarantee this, so one alternative could be selected
even if a majority prefers a different alternative.
Example 2. Suppose 49% of individuals have preferences B G N , 48% of individuals
have preferences G B N , and 3% of individuals have preferences N G B. Then B
is chosen by plurality rule, but if N was removed from the choice set G would defeat B with
51% of the individuals preferring G. Note that here G is the Condorcet winner.
While Bush and Gore had pretty even support and Nader received few votes similar issues
can arise with three “serious” candidates. Former professional wrestler Jesse Ventura was
elected Governor of Minnesota in 1999 as a third party candidate with 37% of the popular
vote, and exit polls indicated that he would have lost a head to head election with either
of the major party candidates (Norm Coleman and Hubert Humphrey III). Similarly, in the
2016 Republican primary, it is possible (though far from certain) that Donald Trump would
have been defeated in many of the early states if he were running in a two candidate election:
during the early states many different candidates divided the anti-Trump vote. Any choice
rule that satisfies independence of irrelevant alternatives will select a Condorcet winner—if
a Condorcet winner exists—but plurality rule does not satisfy independence of irrelevant
alternatives.
Another well-know approach to aggregating preferences is the Borda rule. In the Borda
rule, if there are k alternatives, each individual ranks the alternatives in order of desirability.
Each alternative is then given k points if it is the first choice of an individual, k − 1 points
if it is second, down to 1 point for the individual’s least preferred alternative. The points
are then added up across the individuals, and the alternatives are compared based on the
sum of their points. While its use in political settings is rare, some variation of this is used
to rank NCAA college sports teams, determine baseball MVPs, and in the Eurovision song
contest. Obviously the Borda rule does not satisfy independence of irrelevant alternatives
and will not guarantee that a Condorcet winner, if it exists, is selected. The Borda rule
embraces that it doesn’t satisfy independence of irrelevant alternatives since it regards the
ranking as informative. That is, knowing whether a voter considers a candidate the second
best alternative or the worst alternative provides information about the voter’s estimation
of this alternative. But the concern is that changing the choice set can have an unnatural
effects on the outcome.
Consider the case with 5 voters from Example 1 above. In that example, under the
Borda rule, A receives 11 points, B receives 10 points, and C receives 9 points. So A would
be selected under Borda rule. Now suppose we add a new alternative B 0 in which we get
outcome B but everyone pays a penny. This is a pareto dominated alternative since all
3
US presidential elections, because of the electoral college, is not exactly a plurality rule election, and
Gore received more votes than Bush but still lost the election. However, each state is determined by plurality
rule and Gore may well have been the Condorcet winner in Florida.
6
individuals would prefer B to B 0 . Since a penny is a small amount, we would expect the
preferences to be
• Voter 1: A 1 B 1 B 0 1 C
• Voter 2: B 2 B 0 2 C 2 A
• Voter 3: C 3 A 3 B 3 B 0
• Voter 4: A 4 C 4 B 4 B 0
• Voter 5: B 5 B 0 5 A 5 C
Adding this alternative doesn’t reveal any information about preferences (since a penny is
small) and the alternative B 0 would never be chosen. Still adding B 0 changes the outcome
under Borda rule: now B receives 15 points, A receives 14 and C would receive 11. So, after
changing the choice set to allow for alternative B 0 , B would be chosen instead. Notice that
A remains the Condorcet winner.
Another approach, more common than the Borda rule in practice, is the Hare system.
This is often referred to as single-transferrable vote, and a variation of the Hare system is
used to run elections in countries such as Ireland and Australia and for Party leadership
elections in some countries (e.g. the New Democratic Party in Canada). Under the Hare
system, the alternative(s) that is/are the top choice of the fewest number of individuals
is removed.4 This is continued until all remaining alternatives are the first choice of the
same number of individuals. The Hare system, and the closely related approach of run-off
elections, is a popular proposal among those who favor electoral reform. However the Hare
system is not a panacea. Not only does it often fail to select a Condorcet winner (if it
exists), but it can also fail to satisfy monotonicity: if individuals become more favorable
to one alternative this can cause it to no longer be selected. This is demonstrated in the
following example.
Example 3. Suppose there are 17 individuals and the preferences are as follows:
• Voters 1-7: A B C
• Voters 8-12: B C A
• Voters 13-16: C B A
• Voter 17: C B A
As there are seven with A their most preferred, and five each with B and C most preferred,
the Hare procedure removes both B and C in the first round meaning that A is selected [Note:
though A is selected, B is the Condorcet winner]. Now suppose we change the preferences of
Voter 17 so that she likes alternative A more than C and B.
4
A related approach of having a run-off between the top two vote getters is used in other systems such
as electing the French president and in Louisiana Senate races. It has also recently been introduced in other
states, such as California, as an alternative to party primaries.
7
• Voter 17: A C B
Now there are five for whom B is most preferred and four for whom C is. So C is removed at
the first stage. Now we have a head-to-head comparison on A and B: as eight voters prefer
A and nine voters who prefer B, B is chosen over A. So making A more desirable to voter
17 causes it to no longer be chosen.
Another approach, that is sometimes used, known as sequential pairwise voting, could
violate an even more fundamental property, as it may not produce a pareto optimal outcome.
Under sequential pairwise voting, there is a fixed order with which the alternatives are
compared to each other by majority rule, with the winner compared to the next alternative
in sequence. For example, if X = {A, B, C, D}, we could compare A and B, then the winner
with C, and finally compare the winner of that comparison with D. As the following example
shows, this approach may not lead to an outcome that is pareto optimal.
Example 4. Suppose preferences are as follows:
• Voter 1: A 1 B 1 D 1 C
• Voter 2: B 2 D 2 C 2 A
• Voter 3: C 3 A 3 B 3 D
Notice that, with these preferences, B pareto dominates D. If we use the sequential pairwise
procedure then A would defeat B in the first round (voters 1 and 3 prefer A to B), then C
would defeat A (voters 2 and 3 prefer C to A) subsequently, and finally D would defeat C
(voters 1 and 2 prefer D to C). So alternative D would be chosen even though it is pareto
dominated by alternative B.
Finally, one alternative that has been used in many cases historically (e.g. absolute
monarchy) is dictatorship. This approach works in terms of ensuring complete transitive
preferences, and the choice is independent of irrelevant alternatives. Choosing the most
preferred alternative of one agent is a simple problem that we have experience solving.
However it has the obvious drawback that only one individual’s preferences matter. It
is then an very strong rejection of the anonymity property we discussed regarding May’s
theorem.
It should be noted that when considering each of these choice rules we are assuming
that individuals’ true preferences are known. If the preferences are not known, individuals
must reveal their preferences through voting or in some other manner. Different social
choice rules are more prone to manipulation by individuals mis-reporting their preferences—
independence of irrelevant alternatives and monotonicity will be important for ensuring
that individuals have an incentive to report truthfully. We will discuss the incentives for
individuals to truthfully reveal their preferences later in the course.
5
Desirable Properties for Social Choice
As the above analysis shows, while there are many different ways to aggregate preferences,
there are criticisms of each of the ways we have suggested so far. It is then worth asking
8
whether a social choice rule that satisfies all the properties we have suggested as desirable
exists or not. If so then such a rule would be a candidate for the “correct” way to aggregate
preferences. If not then there is no “correct” way to aggregate preferences, and whichever
approach we take will involve compromises. While we have so far been looking for a choice
rule that maps the preferences of the individuals into an alternative that is selected, we now
look for a way to rank all the alternatives. We now define a social choice ordering (so as
not to conflict with the terminology used above). A social choice ordering is a function from
the set of alternatives, and the preferences of the individual voters, into a social preference
relation,
f (X, (1 , . . . , n )) =
The social preference relation ranks each of the alternatives in X in order of their desirability (with the possibility of ties). This allows us to order all the alternatives, which will
be useful for precisely defining the conditions we want the social choice ordering to satisfy,
and will simplify the proof of our main result. Note, of course, that to determine which
alternative would be selected is the same as identifying the most preferred alternative.
Before proceeding let us state a list of properties that we often want a social choice
ordering to satisfy precisely. The first thing we want to do is ensure that it is always possible
to make a choice: while we may identify multiple alternatives as equally good, and so we can
be indifferent between certain alternatives, we can’t reject all alternatives or find a pair that
we can’t compare. That is, for any preference profile, we want the resulting social preference
relation to be complete and transitive. Second, at a minimum we want to make sure that the
alternative chosen is pareto optimal—we know that if something is pareto dominated that
there exists a different alternative that is preferred by all voters. We insist on a very weak
notion of pareto optimality: that it is not possible to make all individuals strictly better
off.5 We also want to make sure that the social choice rule is monotonic: if one alternative
is made more desirable to at least one voter then this shouldn’t cause it to be ranked lower
from a social perspective. Perhaps more controversially, we want the social choice rule to
satisfy independence of irrelevant alternatives. Finally, we have a preference for treating
different individuals equally and so are averse to dictatorial choice rules. Formally we define
these conditions as follows.
Definition 3. (Unrestricted Domain) A social choice order satisfies unrestricted domain
if, for all preference rankings (1 , . . . , n ), f (X, (1 , . . . , n )) =, where is a complete,
transitive preference relation over X.
Definition 4. (Pareto Optimality) A social choice ordering satisfies pareto optimality if
x y for all x, y ∈ X for which x i y for all i = 1, . . . , n.
Definition 5. (Monotonicity) A social choice ordering satisfies monotonicity if, for all
(1 , . . . , n ) such that x i y for some i, if the preferences of i changes so that now y 0i x,
then if y x where = f (X, (1 , . . . , n )) this implies that y 0 x where 0 = f (X, (1
, . . . , i−1 , 0i , i+1 , . . . , n )).
5
This is a very weak notion of pareto optimality. We don’t even require that if one alternative makes
everyone weakly better off, and one person strictly better off, that that alternative is preferred. Since we are
leading to an impossibility result it makes sense to ask for the weakest conditions possible.
9
Definition 6. (Independence of Irrelevant Alternatives) A social choice ordering satisfies
independence of irrelevant alternatives if, for all preference profiles (1 , . . . , n ) and
(01 , . . . , 0n ), if x, y ∈ X such that x i y if and only if x 0i y and x ≺i y if and only
if x ≺0i y, then x y if and only if x 0 y and x ≺ y if and only if x ≺0 y, where
= f (X, (1 , . . . , n )) and 0 = f (X, (01 , . . . , 0n )).
Definition 7. (Non-Dictatorial) A social choice ordering is dictatorial if there exists an j
such for all x, y ∈ X, x j y implies that x y. A social choice function is non-dictatorial
if it is not dictatorial.
The most complicated, and controversial condition is Independence of Irrelevant Alternatives. The simplest way to think about what it means is to imagine a group choosing
between two alternatives {A, B}. Suppose A is preferred to B. Independence of Irrelevant
Alternatives requires that if another option C was available then the presence of C should
not make B preferred to A. This is considered desirable since why should the ranking of A
and B depend on an “irrelevant” factor such as whether or not C is available or what what
features C has? Also, if the ranking of A and B could be affected by whether or not C is
available, it could create the incentive to provide fewer alternatives to voters.6
The search for a choice ordering that is guaranteed to satisfy the above conditions leads
to one of the seminal results in economics, the Arrow Impossibility Theorem. This theorem
resulted in Kenneth Arrow being awarded the Nobel prize in 1972. The Arrow Impossibility
Theorem states that there does not exist any social choice ordering that satisfies these five
conditions. It is not simply that no such procedure for aggregating preferences has been
found, but rather it will never be possible to find such a choice rule. This does not mean,
of course, that we cannot talk about aggregating preferences. It does mean, however, that
there is no “correct” or perfect way to aggregate preferences and the choice of mechanism
will involve trade-offs. This is a common theme in economics: it typically isn’t possible to
have everything we want so there will be tradeoffs. While we will omit the proof of many of
the results in the class, given the importance of the Impossibility theorem in economics we
will go over its proof.
6
Arrow’s Impossibility Theorem
Let N = {1, . . . , n} be the set of voters, and X be the set of alternatives. We now state and
prove the Arrow Impossibility Theorem.
Theorem 2. (Arrow 1950) Suppose that X at least three elements. Then there does not exist
a social choice ordering satisfying unrestricted domain, pareto optimality, monotonicity, and
independence of irrelevant alternatives that is non-dictatorial.
Before proceeding with the proof we introduce the concept of a decisive set of voters for
each pair of alternatives. A set Vyx ⊆ N is a decisive set for a pair of alternatives (x, y)
6
For example, in the 2016 election it was believed that Michael Bloomberg decided not to run, at least
in part, because plurality rule elections violate Independence of Irrelevant Alternatives and he thought that
running would make it more likely Trump would defeat Clinton. Bloomberg preferred Clinton and ended up
endorsing her instead.
10
if x i y for all i ∈ Vyx implies that x y regardless of the preferences of those not in
Vyx . What constitutes a decisive set depends on the social choice ordering (e.g. if applying
Condorcet’s method a decisive set is any set containing a majority of the voters, under
dictatorship a set is decisive if and only if the dictator is included). It is immediate that a
decisive set exists for each (x, y) for any social choice ordering that satisfies pareto optimality
since N = {1, . . . , n} is a decisive set for any pair: if x is preferred to y by all individuals
it must be socially preferred. We then say that a set V is a minimally decisive set if it is
a decisive set for some pair of alternatives, but no subset of V is a decisive set for any pair
0
of alternatives. That is, there does not exist any (x0 , y 0 ) such that Vyx0 is a decisive set for
0
(x0 , y 0 ) and Vyx0 ⊂ V . A minimally decisive set must exist since the empty set is not a decisive
set for any pair of alternatives—again, by the Pareto criterion.
Proving Arrow’s theorem will then consist of showing that for any choice rule that satisfies unrestricted domain, pareto optimality, monotonicity, and independence of irrelevant
alternatives, there exists some individual j ∈ N such that {j} is a decisive set for all pairs of
alternatives in X. This will then imply that j is a dictator and so no choice ordering that satisfies unrestricted domain, pareto optimality, monotonicity, and independence of irrelevant
alternatives can also be non-dictatorial.
We now proceed with the proof of Arrow’s Impossibility Theorem. The version of the
proof we present comes from Geanakoplos (2005).
Proof. Let f be a social choice ordering that satisfies unrestricted domain, monotonicity,
independence of irrelevant alternatives and pareto optimality. Let V be a minimally decisive
set, which, due to pareto optimality must exist. Then there exist two alternatives x, y ∈ X
such that V is a decisive set for the alternative pair (x, y). Let z be any element of X that
does not equal either x or y. Since X has at least three elements, such a z must exist.
Let j be any individual in V . Since we have assumed unrestricted domain we must be
able to account for all possible preferences. Consider the following preference profile:
• Voter j: x j y j z
• Any voter i 6= j in V : z i x i y
• Any voter i ∈
/ V : y i z i x
By independence of irrelevant alternatives and the definition of V we know that x y.
We must consider the comparison of y and z. Note that, by independence of irrelevant
alternatives and monotonicity, if z y then V \{j} is a decisive set for (z, y). However, since
V is minimally decisive, V \{j} cannot be a decisive set for any alternative pair. Hence,
y z and, by transitivity, x z.
Since, in the specified preferences, voter j is the only individual who prefers x to z
and all other individuals strictly prefer z, by monotonicity and independence of irrelevant
alternatives {j} is a decisive set for (x, z). Moreover, since {j} ⊆ V is a decisive set, and V
is minimally decisive, V = {j}, and so {j} is a decisive set for (x, y) as well.
Since we allowed z to be any element of X not equal to x or y, we can then conclude that
{j} is a decisive set for (x, z) for all z ∈ X. This is true for a fixed x. It remains to show
that, for all w 6= x, {j} is a decisive set for (w, x) and a decisive set for (w, z), where z 6= x..
11
We first consider the case in which z 6= x. Consider the following preference profile:
• Voter j: w j x j z
• Voters i 6= j: z i w i x
Since {j} is decisive for (x, z), x z and, by pareto optimality, w x. So by transitivity
w z. Since in this preference profile w z even though only voter j prefers w to z and
all other voters prefer z to w, by independence of irrelevant alternatives and monotonicity
we can conclude that {j} is a decisive set for (w, z). Since (w, z) are arbitrary we can then
conclude that {j} is a decisive set for all (w, z) where z 6= x.
Finally we conclude by showing that, {j} is a decisive set for (w, x) for any w. To do so
take any z not equal to w or x and consider the preference profile
• Voter j: w j z j x
• Voters i 6= j: z i x i w
Since {j} is a decisive set for (w, z), w is preferred to z. By pareto, z is preferred to x, so by
transitivity w x. As j is the only individual who prefers w to x, independence of irrelevant
alternatives and monotonicity then imply that {j} is a decisive set for (w, x).
We have now established that for any social choice ordering that satisfies unrestricted
domain, monotonicity, independence of irrelevant alternatives, and pareto optimality, there
must exist one individual whose preferences are decisive for the ranking of every pair of alternatives. This individual is a dictator, and so the choice ordering must be dictatorial. This
completes the proof that there cannot exist a social choice ordering that satisfies unrestricted
domain, monotonicity, independence of irrelevant alternatives, and pareto optimality that is
non-dictatorial.
This shows that, regardless of the choice rule used, there always exists some preference
profile for which at least one of the above conditions will be violated. Some choice rules
are guaranteed to satisfy some of the properties, but no choice rule is guaranteed to satisfy
all of them. Of course, the above properties are only violated for some preference profiles.
The decision of which choice rule to use then depends largely on: (1) which violations are
most troubling in the environment considered, (2): how likely are the individuals to have
preferences for which the above conditions are violated. In the next section we consider what
properties we might expect the preferences of the individuals to satisfy.
7
Single Peaked Preferences
One of the demands of the Arrow’s Impossibility Theorem is unrestricted domain: that no
matter what the preferences of the voters are it is possible to aggregate preferences. This is
important because we do not know, ex-ante, when designing a system what the preferences
of the individuals who will use the system to make decisions might be. However we might
expect certain preference profiles to be more common than others. This is the case when
preferences can be ordered in some way—for example, in terms of their ideology.
12
Think back to the 2000 election, with Bush, Gore, and Nader. We can clearly rank these
three candidates in terms of ideology: Bush was the most conservative and Nader was the
most liberal. Now consider a voter whose first choice is Nader: her second choice could
either be Bush or Gore. Which is more likely? Since a Nader voter is probably on the
left of the political spectrum, and Gore is ideologically to the left of Bush, it is likely that
they would prefer Gore to Bush. This is the reason why we think Gore may have been a
Condorcet winner in the 2000 election: since we think Gore would have been preferred to
Bush by a large majority of Nader voters. This suggests that certain preferences would be
more common than others: in particular N B G and B N G should be very rare.
This suggests that a natural assumption about individual preferences is that they are
single peaked. Suppose we can order the alternatives in X and let x1 be the smallest alternative up to xk as the largest alternative: x1 < x2 < . . . < xk . In this case we are associating
each alternative with a real number (e.g., ideological position on a left-right scale, what is
the tax rate, how much public good to provide, etc.).7 Since the magnitudes are arbitrary
assume that X ⊆ [0, 1]. We can then define single-peaked preferences as follows:
Definition 8. Suppose individual i’s most preferred alternative is xk̂ ∈ X = {x1 , . . . , xk },
when the alternatives have been ordered so that x1 < x2 < . . . < xk . Individual i’s preferences
are single-peaked if, for all k 0 > k̂, xk0 i xk0 +1 and, for all k 0 < k̂, xk0 i xk0 −1 .
As an example of this definition, consider the Bush-Gore-Nader example in which X =
{B, G, N }. In this case, since N is on the left and B on the right, then x1 = N , x2 = G,
and x3 = B. To be single peaked any individual i for whom Bush (x3 ) is the most preferred
alternative must prefer Gore to Nader (i.e. if x3 is the first choice then x2 i x1 ), and any
individual with Nader (x1 ) as their first choice must prefer Gore to Bush (i.e. if x1 is the
first choice then x2 i x3 ). The assumption of single peaked preferences would rule the
preferences N B G and B N G.
A common example of preferences that are single-peaked are when each individual i has
a bliss point ti and her utility from any alternative xk is
u(ti , xk ) = −(ti − xk )2 .
More preferred alternatives are then clearly associated with higher utility levels. The interpretation of these preferences is that each person has a ideal policy in ideological space—an
ideal set of policy positions a candidate could have, her ideal policies that she would implement if she were a dictator—and her utility is maximized the closer the policy is to her
ideal point. We will see that focusing attention on single-peaked preferences can help in
aggregating preferences.
Suppose that no two voters share the same bliss point and that there are 2n+1 voters (an
odd number of voters). We can re-index the voters {1, . . . , 2n + 1} so that t1 < . . . < t2n+1 .
In that case voter n + 1 is the median voter. Then we can compare each pair of alternatives
in X by Condorcet’s method. Now take any two alternatives x, y ∈ X, and without loss of
generality assume that x < y. Notice that the median voter prefers x to y if and only if
−(tn+1 − x)2 > −(tn+1 − y)2
7
For example, political scientists place individuals on scale from 1 to 7 in terms of how conservative or
liberal they are.
13
or equivalently
tn+1 − x < y − tn+1
which is equivalent to
x+y
.
2
In this case ti reflects the individual’s most preferred policy (where the locate on the political
spectrum, their ideal tax rate, the amount of public good provision they think is optimal,
etc.) Notice that, since for all i < n + 1, ti < tn+1 , this implies that all voters with ideal
point lower than the median would also prefer x to y. So if the median voter prefers x
to y then x is preferred to y under Condorcet’s method. Similarly, if tn+1 > x+y
then
2
the median voter, and all voters with ideal point higher than the median, would prefer y
to x. So if the median voter prefers y to x then y is preferred to x under Condorcet’s
method. Hence the alternative that is preferred by the median voter is the alternative that
is socially preferred under Condorcet’s method.8 This means that there is no possibility of
non-transitive preferences when preferences are single peaked and the voters’ preferences are
transitive. We can then use the Condorcet method to aggregate preferences when all voters
have single-peaked preferences.
tn+1 <
Theorem 3. (Black 1958) Suppose that there are an odd number of voters and that the
preferences of the all voters are single-peaked. Then the social preference ordering induced
by Condorcet’s method is a transitive preference ordering over X that satisfies independence of
irrelevant alternatives, pareto optimality, monotonicity, and non-dictatorship. In particular,
the most preferred alternative of the median voter is the Condorcet winner in X and so is
the most preferred alternative according to the induced social ordering.
This shows that if we can relax the requirement of unrestricted domain, and instead focus
attention on preference profiles that are single peaked, it is possible to aggregate preferences
with Condorcet’s method. Single-peaked preferences are not necessary for a Condorcet
winner to exist, but are sufficient, and when preferences are not single-peaked it is not
guaranteed that a Condorcet winner will exist. While we know that no method will work
for all preferences, Dasgupta and Maskin (2008) show that, among all approaches that could
be used, Condorcet’s method satisfies the other axioms for the largest set of preferences.
However, unlike other methods, it doesn’t always tell us how to make a choice.
The assumption of single-peaked preferences can more easily be applied in some settings
than others, however. If the alternatives were to build a bridge, not build a bridge, or to
build half a bridge, individuals would probably not have single-peaked preferences. Similarly
when voting on host cities for the Olympics it may not be possible to order the alternatives
and Condorcet cycles are possible. However, in electoral competition, given that ideological
considerations are of central importance, we can usually order the alternatives (candidates)
and the voters in terms of ideology and the voters are likely to have preferences that are (or
at least are close to) single-peaked. However there are important caveats that we will return
to later. That the median voter is decisive under Condorcet’s method will be very important
when discussing electoral competition, which we turn to in the next section. A detailed
treatment of electoral competition is available in Roemer (2009) on the Chalk website.
8
This does not mean that the median voter is a dictator. If other individuals’ preferences change the
median voter’s preferred alternative would no longer be socially preferred.
14
8
Electoral Competition
So far we have not discussed strategic considerations: we have assumed that individuals’
preferences are known and that the alternatives are fixed and exogenous. We will consider
the incentives for individuals to reveal their preferences later in the course, and we turn
to determining the set of alternatives now. Often the alternatives that individuals vote
over are strategically determined: for example, in an election the candidates for office run
campaigns in which they offer different platforms to the voters, and the voters then decide
which candidate to vote for depending (at least in part) on what the candidates have promised
to do. Given that the set of alternatives is endogenous, we want to get a sense of which
alternatives will be offered to voters in elections. Moreover, Black’s theorem shows that there
is a natural policy alternative to select when preferences are single-peaked, so we want to
get a sense of whether competitive elections are likely to select this alternative. This leads
us to the model of electoral competition due to Downs (1957). In the Downsian model two
candidates compete for office, and we solve for the Nash equilibrium platforms proposed by
each candidate. Recall from Econ 201 that a Nash equilibrium is when both candidates are
optimizing given the strategy of the other candidate.
Suppose there are two candidates for office, 1 and 2. The two candidates i ∈ {1, 2}
compete by choosing any policy xi ∈ X = [0, 1] to implement if elected. This policy is
observed by the voters and the candidates commit to the policy before the election. Both
candidates 1 and 2 are purely office motivated and receive a payoff of 1 if elected and a
payoff of 0 if not elected. There are 2n + 1 voters with single-peaked preferences over the
policy implemented with ideal points t1 < . . . < t2n+1 . The candidates compete by majority
rule. Assume that if a voter is indifferent between the two candidates she flips a balanced
coin and votes for each candidate with probability 1/2. This game has a unique equilibrium:
both candidates propose the bliss point of the median voter.
Theorem 4. (Downs 1957) In the unique Nash equilibrium x1 = x2 = tn+1 .
This is frequently called the Median Voter Theorem, that electoral competition creates
an incentive for both parties to locate at the ideal point of the median voter. You may notice
the similarity between the Downsian model and Hotelling’s linear city which you may have
seen in Econ 201 or an IO class. In order to attract the greatest number of customers, the
firms located at the median location of the customers. The same logic leads candidates to
locate at the median of the electorate.
Notice how important the assumption that preferences are single-peaked is for the median
voter theorem. While the voters will only choose between two alternatives, the candidates
can choose any two alternatives in X to offer them. When a Condorcet winner exists, both
candidates must propose a Condorcet winner: otherwise the other candidate could propose
a Condorcet winner and defeat them for sure. However, if a Condorcet winner does not
exist, then there cannot exist a pure strategy equilibrium. If a Condorcet winner does not
exist then whatever policy 1 offers 2 could find a policy to defeat it, but then 1 would
want to change the policy it implemented to something that defeats 2 and so on. If we had
considered electoral competition with X = {A, B, C}, three voters, and the preferences from
the Condorcet paradox in section 3, then there cannot not exist a pure strategy equilibrium;
15
only a mixed strategy equilibrium could exist.9 In this class we will focus on pure strategy
equilibria.
This also leads to an important caveat about single-peakedness and the median voter
theorem. Suppose that instead of having an ideal point in one dimensional space the voters
each have an ideal point ti = (t1i , t2i ) in two dimensional space, and the alternatives in X also
consist of two dimensions. In this case a median voter does not exist, ordering the set of
alternatives is more difficult, and a Condorcet winner will not always exist.10 To guarantee
that a Condorcet winner exists requires strong symmetry assumptions about the distribution
of voter ideal points. So, while the Downsian logic is compelling, we also need to be aware of
the caveats about the theoretical prediction. In the next section we evaluate the prediction
empirically.
9
The Empirical Relevance of the Median Voter Theorem
The Median Voter Theorem is the starting point for modeling political competition. However
its prediction that both candidates converge to median voter is often criticized, especially
in recent years, as counter-factual. We now consider some of the empirical work testing the
predictions of the median voter theorem and look at the evidence that the median voter
theorem may not hold. In the next section we will consider some theories of why candidates’
platforms may diverge.
Arguably the Downsian equilibrium was a close approximation of reality when it was written in 1957. A 1950 report of the American Political Science Association titled “Towards A
More Responsible Two-Party System” called on political parties to offer clear, differentiated
alternatives to the voters. Democrat John F. Kennedy criticized the Republican Eisenhower
administration for being weak on defense in during the election of 1960 and proposed substantial tax cuts soon after taking office. Moreover, in 1964 presidential candidate Barry
Goldwater promised that unlike the candidates in previous elections he would offer the voters “a choice, not an echo” before being defeated in one of the most lopsided elections in
American history.
In recent elections, however, few would argue that the parties are interchangeable. The
concern now is that the parties have become extremely polarized. For instance, most people
perceived a large difference between Obama and Romney in 2012 and Clinton and Trump
in 2016. Similarly, the parties were considered very polarized in the second half of the twentieth century. We want to discuss the evidence regarding whether or not parties platforms
converge. A series of papers by Nolan McCarty, Howard Rosenthal, and Keith Poole address
this by estimating the ideology of legislators—known as NOMINATE scores—from different
parties. This literature is summarized in their book, McCarty et al. (2008) and the relevant
chapter is available under Library Reserves on the Chalk website.
9
In the mixed strategy equilibrium each candidate randomizes with equal probability over each policy in
X.
10
If there are two dimensions but only one of them involves disagreements (e.g. one dimension is ideology and second dimension is quality of “valence” of the candidate) then a Condorcet winner still exists.
Alternatively if candidates have fixed characteristics in the second dimension, existence is again guaranteed.
16
McCarty, Poole and Rosenthal estimate the policy positions of members of the legislature.
The approach is to assume that the legislator (in the House or Senate) is voting as if they were
maximizing a utility function with an ideal point xi —this could be their personal ideal policy
or it could be the policy they have promised to the voters—then estimate this objective point
based on the voting record of the legislator. Namely, those who vote similarly are considered
to have ideologically close positions. We are interested in how much variation there is across
legislators and how much the voting records vary within and across parties. How much of
ideology appears to be driven by constituency preferences? Do Senators from the same state
vote the same way?
Consider three politicians who are currently serving in the U.S. Senate: Bernie Sanders
(I-VT), Ted Cruz (R-TX), and Susan Collins (R-ME). Cruz is considered to be very conservative, Sanders very liberal, and Collins as a Republican in a Democratic state is viewed
as a moderate. For example, she was one of two Republican senators to vote against the
confirmation of Betsey DeVos and was one of the few to meet with Obama’s Supreme Court
Nominee, Merrick Garland. How would we see this relationship in the data? Suppose we
were to see votes that looked something like this:
Bill 1: Cruz and Collins YEA, Sanders NAY
Bill 2: Cruz NAY, Collins and SandersYEA
Looking that this data we would see that Collins voted with Cruz once and Sanders once
on bills that Cruz and Collins disagreed on. We could interpret this as a sign that Collins
was ideologically between Cruz and Sanders—provided of course that this pattern was borne
out in the larger sample of bills (recent congresses have had 500 to 1200 votes a year). We
could also conclude that yes on bill 1 fits with the Cruz side and yes on bill 2 fits on the
Sanders side. Nothing in the data would tell us which side was conservative and which side
was liberal, though we can make that judgment based on who was estimated to be on which
side of the spectrum. Let’s represent the conservative side with higher numbers since the
real line is ordered from left to right.
Of course we would need to look at the whole range of votes, in which case not all votes
would fit so neatly. We would likely see some bills with Cruz and Sanders voting against
Collins. However we would still estimate that Collins is between Cruz and Sanders if such
an outcome is less common than Collins voting with either Sanders or Cruz. In this case we
assume that the legislators make “errors” —that is, there are elements outside the model
such as local conditions, interest groups, or a personal connection to some issue lead them to
cast votes different from the model predictions. The approach is then to estimate legislative
ideal points based on minimizing the “errors” so that the the model fits the data as well as
possible. This approach is known as maximum likelihood estimation.
Assume there are K legislators, and M bills that the legislators will vote on. Assume to
begin with that we know which side represents the conservative side of each bill and which
represents the liberal side. We must estimate the ideology of each of the K legislators and
the M bills, so that is K + M parameters to estimate. But we have KM data points since
each legislator votes on each bill (excepting votes that are missed). As we have many more
data points than parameters to estimate we have enough data to estimate to estimate the
parameters we are interested in. We assume that on each bill m legislator k voters for the
conservative side if and only if
xk + εk,m > ym
17
where εkm are independent and identically distributed draws from the distribution (normally
distributed for example). We can then calculate the probability that the legislator k would
cast the vote they did, if their ideology is xk and the cutoff on bill m is ym . This probability
of the observed vote is then
P r(εk,m > ym − xk ) vote on conservative side,
p(xk , ym ) =
P r(εk,m < ym − xk ) otherwise.
But then, of course, we observed KM different votes, and the probability of all KM votes
happening, given that each ε is independent, is just the product of the different probabilities.
We then define the Likelihood Function
L(x1 , . . . , xK , y1 , . . . , yM ) =
K Y
M
Y
p(xk , ym ).
k=1 m=1
That is, the likelihood function is the probability that, if the actual ideologies of the legislators were (x1 , . . . xK ) and the actual vote cutoffs were (y1 , . . . , yM ) what the probability of
observing the profile of votes is. We then find the legislator ideologies and vote cutoffs for
which the actual vote profile we observed would be as likely as possible, by maximizing
L(x1 , . . . , xK , y1 , . . . , yM )
over (x1 , . . . xK ) and (y1 , . . . , yM ). This provides an estimate for the ideal point of each
legislator—conditional on correctly identifying which bills are conservative and which are
liberal. Of course, we may not know ex-ante which side is conservative and which side is
liberal, and the vote profile will only match the data well if we have correctly identified
which side is which. This means we have an additional M things to estimate: whether a
yes vote is conservative or liberal on each bill. In principal, we could then compare over the
2M different combinations to see which combination of bills being conservative matches the
data the best by giving us the highest value of the Likelihood function. This gives us an
estimate of which vote on each bill represents the “conservative” position, as well as how
far to the left or right each vote is. In practice, since 2M is an extremely large number, this
a very computationally hard problem, so computational techniques are necessary to avoid
calculating the likelihood function for every possible permutation. For example, on some
bills it is clear what the conservative position is from who is voting on which side—so we
may not need to check every permutation. Using this approach we get what we are really
interested in: an estimate of the ideal point each legislator is maximizing. We can then plot
how the ideal points vary within, and across, the parties based on how they vote.
We can also compare polarization over time by comparing legislators who are in different
Congresses. Also, while it becomes more complicated, we can add additional dimensions and
see how much the explanatory power of the model is improved. Determining whether the
additional dimensions adds enough to make sense including is known as Factor analysis.
This estimation leads to the following conclusion:
1. The voting patterns of legislators elected from different parties is very different.
2. This polarization has increased since the 1960s: there used to be substantial overlap
between the legislators from each party, but such Representatives and Senators have
largely disappeared.
18
3. We can extend the methodology to allow for multiple dimensions, but, excepting the
civil rights era, more dimensions do not add additional explanatory power. The voting
records of legislators is largely captured by a one dimensional Liberal-Conservative
axis.
This approach establishes (subject to all the caveats about the data and all the assumptions underlying the estimation) that legislators from different parties have very different
voting records. This does not in and of itself prove that the policies taken by the winning
candidate diverge from those of the median member of the district (at least in House districts
where only one representative is elected). It is possible that the winning candidate in each
district locates at the median in their district and that districts that elect Republicans are
simply more conservative than those who elect Democrats. While it is certainly true that
Republicans are elected in more conservative districts than Democrats, some evidence that
this is not the whole story is provided by Lee et al. (2004).11 This paper is posted on Chalk.
They take a regression discontinuity approach, and look at elections in which the share
of the two-party vote is almost exactly 50-50. The idea is that when an election is that close,
whether a Republican or Democrat is elected in these districts is more or less random. This
can, at least partially, be tested by looking at how districts that barely elected a Republican
compare to those who barely elected a Democrat. They then look to see whether there
is discontinuity—a discrete jump in the voting record of the of the legislator—at exactly
50% of the two-party vote. Figure VI (on page 840) provides evidence of such a jump at
50%. In fact, it indicates that the vote share has little effect on the voting record of the
legislator, except in determining which party is elected. This indicates that a Republican
and a Democrats would vote very differently, even if they were elected to represent the same
constituency. This, provides evidence against the median voter theorem.
10
10.1
Models of Divergent Platforms
Policy Motivation and Uncertainty
As there is strong empirical evidence refuting the Downsian prediction that platforms converge, it is important to ask what is missing from Downs’s model. Many different proposals
have been put forward to explain why candidates may diverge in equilibrium. One natural
element of the Downs model to question is the assumption that parties are purely office motivated. What happens if instead we assume that parties/candidates are policy motivated?
This leads to the Calvert-Wittman model.
Consider the Downs model, but instead of assuming that parties care only about holding
office assume that parties care about policy. Assume that party 1 has utility function −x2 ,
and party 2 has utility function −(1−x)2 , where x is the policy of the candidate who is elected
by the voters. In this case x = 0 is the bliss point of Party 1 (the Democrats) and x = 1 is
the bliss point of party 2 (the Republicans). Here the parties don’t care about who holds
office at all—they only care about the policy outcome—though, of course, we could all a mix
11
Another piece of evidence that candidates do not converge to the median in the district comes from
Senate. In the Senate two candidates represent the same state at each time, but typically have different
voting records (particularly when they come from different parties).
19
of both. We maintain the assumption from the Downsian model that parties can perfectly
commit to the policies they propose. Unlike in the case with office motivated candidates,
this assumption is not innocuous: the parties may not propose their most preferred policy
but we assume they will follow through on their promises. What is the equilibrium of this
game?
First note that xL = xR = tn+1 is still an equilibrium. If either party deviates then it
has no effect on the implemented policy (since the deviating party is defeated), so neither
party has an incentive to deviate. It is, in fact, the only equilibrium. To see this, note that,
although party 1 would prefer to implement a policy to the left of the median, and party
2 would prefer a policy to the right of the median, in order for the party’s platform to be
implemented it must win the election. If we were to have an equilibrium in which one party
always wins with a non-median policy, the other party would “undercut” them by locating
closer to the median voter’s ideal point but in the direction of that party’s ideal policy. If
the two candidates were to tie by locating an equal distance from the median, then one party
could move slightly closer to the median voter and win for sure. This is similar to Bertrand
competition from Econ 201. While the parties would like to pull the policy towards their
ideal point (and firms would like to set price above cost) since the either party can take the
election by locating slightly closer to the median voter’s ideal point (either firm could win
the entire market by undercutting its rivals price slightly), competition forces both parties
to locate at the median. This means that more than just policy preferences are necessary to
explain divergent platforms.
The simplest model of policy divergence comes from a combination of policy preferences
and uncertainty by the parties about the ideal point of the voter. Suppose that there are
2n + 1 voters and two policy motivated parties with ideal points 0 and 1. Suppose now
that the parties don’t know the ideal point of the median voter, but believe that the ideal
point of the median voter is tM = 1/2 + ε, where ε is Normally distributed with mean 0
and variance σ 2 . Now note that we cannot have an equilibrium with xL = xR = 1/2: either
party could deviate and choose a policy closer to their ideal point. This has no effect on
the policy if they aren’t elected, but moves the policy closer to their ideal point if they are.
Since the location of the median voter’s ideal point is unknown, the election is random, and
the deviating party wins with positive probability. In equilibrium the parties must locate
away from the median.
Theorem 5. There exists a unique pure strategy symmetric equilibrium. In this equilibrium
xL = 1/2 − d, xR = 1/2 + d where d ∈ (0, 1/2). The degree of divergence d is increasing in
σ with
lim d = 0,
σ→0
lim d = 1/2.
σ→∞
This indicates that a combination of policy motivation and uncertainty can generate the
prediction of divergent platforms. If the parties were office motivated then both parties
would locate at the expected median even if there is uncertainty. This uncertainty could
be about the median voter’s ideal point or anything else (i.e. whether the voters will like
the candidate of the party personally, uncertainty about which voters will turn out to vote,
etc.) However, the pressure to move to the center is still apparent, and parties only move a
20
non-trivial distance from the expected median if the uncertainty is not too small. For more
details on this model see the first two chapters of Roemer (2009).
10.2
Electoral Accountability
The Downs and Calvert-Wittman models both assume that candidates can make binding
commitments to specific policies and will always follow through on them. This assumption
seems unrealistic: there is no formal mechanism underlying this commitment and candidates
may make promises that are impossible to fulfill. The question of which promises can be
committed to is considered in Alesina (1988), available on Chalk.
Suppose there are two candidates, L and R, who are concerned only with policy. The
ideal point of the L is 0 and the ideal point of R is 1, and they have quadratic loss as before.
Let’s assume there is no uncertainty about the median voter’s ideal point: it’s tn+1 = 1/2.
We assume that these candidates (parties) are long-lived and care about the policy in
each period. Future outcomes are usually discounted because of people are impatient or
because it is uncertain whether the candidates will be running again. Let β ∈ (0, 1) reflect
the weight placed on future periods.12 Suppose the game is infinitely repeated so the is no
known last period; we can interpret this as the players not being certain when the game will
end. If the policy chosen by the elected candidate at each period s = 0, 1, 2, ... is x(s) the
payoff to the left candidate is
∞
X
β s x(s)2 ,
−
s=0
and to the right candidate it’s
−
∞
X
β s (1 − x(s))2 .
s=0
Note that in this formulation, the payoff in period 0 is not discounted, one period in the
future is weighted β, two periods in the future are weighted at β 2 , and so on.
We look for a symmetric equilibrium in which, if elected, candidate L chooses xL = 1/2−d
and candidate R chooses xR = 1/2 + d in every period if they are elected. That is, we look
for an equilibrium in candidate behavior is stationary: that is, the policy they choose if
elected is the same in every period. If d = 0 then we have that the candidates converge to
the median, and d > 0 then they don’t. One equilibrium is for each candidate to choose their
ideal policy in every period (d = 1/2). However, voters can do better if the median voter
conditions her vote on the incumbent’s behavior in office: Suppose the median voter votes
to re-elect the incumbent if, and only if, the policy they chose was no further than d from
the median voter’s ideal point. We now ask what the most moderate (smallest d) outcomes
that would not give the candidates an incentive not to deviate. To do this we have to make
sure the candidate wouldn’t be willing to deviate to their ideal policy and lose election.
12
For example, it could be that in every period, with probability 1 − β the candidate is prevented from
running due to health reasons; assume that if the candidate is prevented from running the game ends and a
new game that is unaffected by the play in this game starts.
21
As everything is symmetric, we only need to look how the R candidate would behave if
elected. For these strategies to be an equilibrium, so we must have
∞
X
s
2
−β (1 − (1/2 + d)) ≥ 0 +
∞
X
s=0
−β s (1 − (1/2 − d))2 .
s=1
The left hand side reflects the payoff from choosing xR = 1/2 + d every period and being
re-elected forever. The right hand since reflects the payoff from choosing xR = 1 today then
losing office and having the rival win office forever with xL = 1/2−d. To have an equilibrium
the left hand side must be at least is large as the right.
The above condition gives an infinite sum which is, in general, difficult to calculate.
However since the payoff is the same in each period, it becomes easy. Notice that
!
∞
∞
∞
∞
X
X
X
X
βs = 1 +
βs = 1 + β
β s−1 = 1 + β
βs ,
s=0
s=1
which implies that
s=1
∞
X
βs =
s=0
and
∞
X
s=1
s
β =β
∞
X
s=0
s=0
1
,
1−β
βs =
β
.
1−β
Using this we can calculate the infinite sums, and the condition to be an equilibrium reduces
to
−β(1 − (1/2 − d))2
− (1 − (1/2 + d))2
≥
.
1−β
1−β
To characterize the minimum level of divergence we must have this hold with equality, so we
look for d that solves
2
2
β
1
1
1
−d =
+d .
1−β 2
1−β 2
This implies that the most moderate policy that can be supported solves
1/2 − d p
= β,
1/2 + d
and so
√
1− β
1
√ ∈ 0,
d=
.
2
2(1 + β)
This leads to the following result.
Theorem 6. Let d =
rium:
√
1− √β
2(1+ β)
∈ 0, 21 . The following strategies constitute a Nash Equilib-
• Candidate L chooses 1/2 − d if elected, and candidate R chooses 1/2 + d.
22
• All voters to the left of the median vote for L, and all to the right of the median vote
for R. The median voter re-elects the incumbent if and only if the policy they chose is
in [1/2 − d, 1/2 + d].
Moreover, this is the lowest level of polarization that can be supported in any stationary
equilibrium.
The above result says that, although elections can incentivize compromise, politicians
cannot commit to the median policy unless they become extremely patient. Extreme patience
would only emerge if elections are extremely frequent: if we think of elections as happening
every four years it is likely the future is significantly discounted. This is one reason why the
median voter theorem breaks down: it simply isn’t possible for candidates to make binding
commitments in campaigns and so we have to look at the policies they will have an incentive
to choose if elected. In this way the identity not just the promises of candidates matter.
However, even though promises may not be binding it is possible to create incentives for
compromise by the candidates.
While we do not have time to to go into it here, we can make the model richer in many
different ways. If we include other forms of “shirking” by elected officials (lack of effort or
forms of rent-seeking) the same basic logic applies—we consider such a model in Problem Set
2. In addition we can incorporate other features such as uncertainty about the median voter’s
preferences or imperfect observability of the elected official’s behavior. These possibilities,
aside from ensuring the incumbent will not be re-elected forever, will generally increase the
amount of shirking by the official. The reason for this is that “good” behavior may not be
observed and so will not necessarily be rewarded with re-election and “bad” behavior will
not necessarily result in losing office. This is general problem of “moral hazard” and it is
typically the case that individuals will shirk on their responsibilities more when they are less
easily monitored. Some empirical evidence of such shirking by politicians, focused more on
politician effort, is available in the Snyder and Stromberg (2010) paper on Chalk.
10.3
Citizen Candidates
One question, particularly if we think polarization is driven (at least partly) by the preferences of the candidates, is where those preferences come from. The citizen candidate model
views candidates for office as citizens with policy preferences of their own who decided to
run for office. They assume the opposite of what the papers in the Downs/Calvert/Wittman
framework assume. Namely that candidates always implement their ideal policy if they are
elected.13 The candidates are strategic, however, in their decision of whether or not to run
for office. There are two papers that introduced this literature—one of them, Osborne and
Slivinski (1996) is available on Chalk. This gives a model of the ideology of those who run
for office.
There are a continuum of citizens with ideal points t ∈ [0, 1], and they all have singlepeaked preferences over the implemented policy x. Assume that the ideal points of the
citizens are uniformly distributed from [0, 1]. All citizens care about the policy implemented,
13
It is possible to consider a repeated citizen candidate model but we will not do that in this class.
23
and can also become candidates by paying a cost c > 0. If the citizen becomes a candidate
and wins election they receive a utility benefit of b > 0. Hence the utility of each citizen is

elected,
 b−c
−(x − t)2 − c ran and lost and candidate with ideal point x won,

−(x − t)2
didn’t run for office, and candidate with ideal point x elected.
The election takes place by plurality rule: the candidate with the largest share of the vote
is elected. We assume that all citizens who do not become candidates vote sincerely in that
they vote for the candidate that would provide them with the highest utility. Voters who are
indifferent randomize with equal probability across the candidates they are indifferent over.
First note that, if b > 2c, there cannot exist an equilibrium with only one candidate: in
that case, another candidate with the same preferences could also run for office and win with
probability 1/2 and earn strictly higher payoff.
Is there an equilibrium with two candidates? If so, are both candidates located at the
median? To have an equilibrium requires that none of the candidates for office who ran for
office would receive a higher (expected) payoff from not running, and that no citizen could
receive a higher expected utility from running for office. If there are two candidates, both
with the median policy, then each candidate receives half the vote and wins with probability
1/2.
Now consider a candidate of type x that is slightly higher than 1/2. Note that, all citizens
of type t > x would prefer this candidate to the median candidate, so this candidate would
receive at least 1 − x of the vote. As the other voters are indifferent between the median
candidates their votes split evenly: hence each candidate receives less than x/2 votes. So if
x is close enough to 1/2, they would win the election and have an incentive to run for office.
This shows that we cannot have an equilibrium with two candidates and converging ideal
points. So any two candidate equilibrium must involve non-converging platforms. Since a
candidate who would lose office for sure could avoid the cost of running for office could save
the cost of running for office by not declaring as a candidate, for both candidates to run for
office, the candidates must tie.
So any two-candidate equilibrium must involve two candidates, x1 = 1/2 − d and x2 =
1/2 + d running for office. As the median voter is indifferent between them each candidate is
elected with probability 1/2. To determine when this constitutes an equilibrium we need to
check whether (1) neither candidate who is running for office could do better by not running
(2) no citizen who did not declare themselves a candidate could increase their utility by
running for office. The first part is immediate: since each candidate wins with probability
1/2 and b > 2c, and the expected policy is better for each candidate than from not running,
there is no incentive for either candidate not to run. Next note that there is no incentive for
any citizen of type t < x1 or t > x2 to run for office: they could never win office and would
only siphon off votes from the candidate on their side, leading to a worse policy outcome.
So we need to check whether a candidate of type t ∈ 12 − d, 12 + d would have an incentive
to run. The most successful such candidate would be one with t = 1/2.
So for which values of d would a median
citizen have no incentive to run for office. Note
1+d
that all voters with type t ∈ 1−d
,
would vote for candidate 1/2, all voters of type
2
2
1+d
t > 2 would vote for candidate x2 and all voters of type t < 1−d
would vote for candidate
2
x1 . Hence the median candidate would receive d share of the votes, and candidates x1 and
24
x2 would each receive 1/2 − d. So if d < 1/3, a median citizen would not have an incentive
to become a candidate.
Theorem 7. If b > 2c then there exist multiple equilibria with two candidates. For any
d ∈ (0, 1/3), there exists an equilibrium in which two candidates run for office x1 = 1/2 − d
and x2 = 1/2 + d and both candidates win with probability 1/2.
There are clearly many elements of electoral competition missing from the citizen candidate model—in particular, candidates always pursue their most preferred policy if elected—
but there are several interesting results. In particular, notice that the candidates can be as
polarized as 1/6 and 5/6. This is a long way from the median voter theorem. A median
candidate, even though they are the Condorcet winner, would not run for office, because the
non-centrist candidates crowd out the votes for the centrist candidates. As we saw at the
beginning of the course, plurality rule elections may not select the Condorcet winner when
there are three candidates. Arguably, the Republican and Democratic party do this in U.S.
elections, making it difficult for centrist parties to form. Note however that, as long as the
costs aren’t too high, there would also exist an equilibrium with three candidates for office,
x = 1/6, x = 1/2, x = 5/6. Each candidate would receive 1/3 of the votes and each win
office with 1/3 of the probability. So it is difficult to make tight predictions from a citizen
candidate framework.
11
Strategic Voters
Central to the strategic consideration of whether to run for office or not is how voters decide
who to vote for if there are more than two candidates. In Osborne and Slivinski (1996) it
is assumed that voters vote for their first choice. This may not aways be true—think of the
voters who worried that a vote for Nader was a “wasted vote” in the 2000 election or during
the 2016 Republican primary the “never Trump” faction rallying around Ted Cruz even if
they preferred another candidate such as John Kasich.14 The set of equilibria can change
a lot if voters are strategic about which candidate to support, and it will often complicate
the analysis further. Strategic considerations about which candidate to vote for, or which
alternative to support, takes on even greater significance in a small committee with a few
voters. We first illustrate the effect of strategic behavior by voters in a plurality rule election,
and then return to our considerations from the first part of the class about the incentive for
voters to report their preferences truthfully under different decision making mechanisms.
Consider the citizen candidate model, but allowing for strategic voters. This is the
approach taken in Besley and Coate (1997). We now assume that voters seek to affect the
election, which requires, of course, that one vote can influence the outcome. We assume now
that there are 2n + 1 citizen types, that the median type’s ideal point is at tn+1 = 1/2 and
the furthest left and right citizens are t1 = 0 and t2n+1 = 1. Suppose also that there are
several citizens with each of the ideal points (so there are many voters with the same ideal
point as the median voter) and that there are an equal number of citizens of each type. If
voters care only about the outcome and not who they vote for intrinsically, then who each
14
For example, South Carolina Senator Lindsey Graham supported Ted Cruz saying he was his sixteenth
choice out of the seventeen candidates, but he was the most likely to candidate to defeat his last choice.
25
vote only matters if pivotal. We can then construct several equilibria based on the voter’s
expectations on when the election is likely to be pivotal.
First note that, unlike in the case in which voters are sincere, we can have an equilibrium
in which there are two candidates for office, x1 , x2 both who have the median ideal point
(assume the costs are such that a third candidate wouldn’t run). That is x1 = x2 = tn+1 .
Such an equilibrium can be supported with the following voting strategies. If there are two
candidates and the voter is indifferent between the candidates, flip a coin. If there are two
median candidates and a third candidate enters to the left or right of the median, abandon
candidate 2 and each voter votes for candidate 1 or 3 based on which candidate they prefer.
Note that then there would be no incentive for a non-median citizen to run—they would
surely lose. In this equilibrium, the voters on the left respond to the threat of a candidate
on the right by rallying around one of the two candidates at the center, and voters on the
right respond to a threat from the left in the same way. (Note: this coordination is very
difficult for a large number of voters to achieve, but campaigns, endorsements, and polls can
facilitate this coordination by telling the voters which candidate to support).
While the median outcome is an equilibrium with strategic voters, but not with sincere
ones, when the voters behaved sincerely this places an upper bound on how much polarization
is possible. This is not true when voters are strategic. We can have an equilibrium in which
two candidates x1 = t1 = 0 and x2 = t2n+1 = 1 run for office. Voters to the right of the
median vote for candidate 2 regardless of who else runs for office, and those on the left vote
for candidate 1 regardless of who runs for office, and median voters flip a coin between parties
1 and 2. In this case, both candidates win with probability 1/2 and each voter’s vote could
affect the outcome. Since the voters care only about who wins the election, and since nobody
else is voting for any other candidate their vote can never cause a different candidate (for
example a citizen with median policy preferences who ran for office) to be elected, if another
candidate were to enter the race, no voter would have an incentive to deviate and vote for
them. Basically, this is saying that, if voters (similarly activists, volunteers, fundraisers)
are skeptical that a new candidate is likely to win they will not support it. Arguably this
coordination contributes to stability of the two-party structure in the United States. For a
third party to be able to win office it needs to not only be desirable to the voters, but also
be able to convince them that it can win.
The coordination on which candidate to vote for is part of the large class of coordination
games. Other examples include figuring out where to meet for lunch (e.g. we would like to
go to the same place so as to meet each other) and which side of the road to drive on (it
works best of we all drive on the same side of the road, but whether that’s the left or right
may not matter as much). We look for an equilibrium of a game because they are stable—we
if we get to this point, there is no incentive for anyone to change what they are doing—and
in coordination games there are often many equilibria. When there are many equilibria it is
difficult to know which equilibrium is more likely to be played. We often think that some
equilibria are “focal” and so more likely to be played (Thomas Schelling articulated this
concept and won a Nobel Prize in 2005 for his work). For example, we might think that it
is focal to meet in a location we have met at in the past, drive on the side of the road the
street signs say, or vote for the parties that were competitive in the last election. One form
of behavior that is often thought to be focal is to tell the truth, so if it is an equilibrium
for everyone to sincerely report their preferences we expect this equilibrium to be played.
26
If reporting your preferences sincerely isn’t an equilibrium, then it becomes less clear which
equilibrium to expect to be played.
This has shown that plurality rule can cause voters to, for strategic reasons, vote for
alternatives other than their first choice. This raises many issues: (1) We can’t necessarily
interpret vote totals in multi-candidate elections as reflecting the true preferences (2) The
set of equilibria are different (often larger) when voters behave strategically so it is difficult
to make tight predictions about what will happen. If sincere behavior constitutes an equilibrium, even one that isn’t unique, this is likely to a “focal” equilibrium. (3) It is possible
that the most preferred alternative may not be selected because of voters’ expectation that
it won’t win. In fact, it is possible for the election to come down to any two candidates.
This means that, due to the strategic coordination issues, it is difficult to predict the outcome when the voters are behaving strategically. We expect voters to abandon third place
candidates, but can’t be sure ex-ante which candidate(s) will be abandoned. So if we design
an electoral system with considerable room for strategic manipulation we can’t make tight
predictions about what the outcome will be with more than two candidates. When there
are only two candidates there is no reason for individuals (at least if they know which alternative they prefer) to behave strategically. We now consider how the incentive to behave
strategically varies with the electoral rules, and study whether there exists a mechanism that
is immune to strategic manipulation.
12
The Gibbard-Satterthwaite Theorem
It has been well known for a long time that individual voters may have an incentive to
strategically misrepresent their preferences. When Borda proposed the Borda method it
was quickly pointed out to him that individuals would have an incentive to mis-report their
preferences. He famously dismissed the concern, saying that his method was meant for
“honorable men”. To an economist Borda’s response is unsatisfying. First, it is not clear
that it is dishonorable to mis-report ones preferences in an effort to get an outcome the
individual thinks is more desirable. Second, even if we accept that doing so is dishonorable,
if it is also advantageous to misreport ones preferences this would put the “honorable” people
at a disadvantage. So such a procedure would require that everyone who might use this
procedure would be “honorable”. We then want to discuss whether it is possible to design
a system for aggregating preferences in which no voter finds it beneficial to mis-report their
preferences. This leads to the Gibbard-Satterthwaite theorem.
When considering the Arrow’s Impossibility Theorem we assumed that individual preferences are known. But it is rare that we would know the exact preferences of the individuals.
Some of the conditions of Arrow’s theorem—namely independence of irrelevant alternatives
and monotonicity—play a central role in ensuring that individuals do not benefit from misreporting their preferences. Independence of irrelevant alternatives is probably the most
controversial requirement of Arrow’s Impossibility Theorem and some would argue that it
is not always desirable—this is the rationale for the Borda method, for example, which embraces its violation of IIA when calculating its ranking. However, independence of irrelevant
alternatives and monotonicity both play a key role in ensuring that individuals cannot profit
from misreporting their preferences. It is the violation of independence of irrelevant alter27
natives that creates the scope for strategic behavior in plurality rule elections, for example.
Similarly, the following example shows the scope for strategic manipulation in Borda.
Example 5. Suppose there are three voters with preferences
• Voter 1: A 1 B 1 C 1 D
• Voter 2: A 2 B 2 C 2 D
• Voter 3: B 3 A 3 C 3 D
Under the Borda procedure A receives 11 points and B receives 10 points, so A is chosen.
Note, however, that if Voter 3 misreports her preferences as
• Voter 3: B 03 C 03 D 03 A
then A receives 9 points and B still receives 10 points and so B is chosen.
Monotonicity is also important for individuals to have an incentive to report their preferences truthfully. It is easy to see why: if monotonicity is violated it is possible that A would
be selected if person i prefers B to A, but B would be selected if i prefers A to B. Clearly
in that situation person i would have an incentive to mis-report her preferences.15
Before proceeding we define precisely what it means for a choice rule to be “manipulable”
by individuals. Basically manipulable means that is possible to get an alternative you like
more by misreporting your true preferences. If a choice rule is not manipulable we call it
non-manipulable or strategy proof: in that their is no benefit to any individual from behaving
strategically rather than truthfully reporting her own preferences. Unlike in Arrow, we look
for a social choice rule that selects one alternative from the choice set, and we assume that
we can uniquely select one outcome for each reported preference profiles. Individuals could
still be indifferent, but the choice rule must specify which alternative is selected in the event
of a tie in a way that is non-random. The key is that this rules out an approach such
as randomly appointing a dictator, which would clearly be strategy proof, but which are
arguably no more desirable than appointing a dictator. Note that in Arrow’s theorem we
also assumed that the ranking was not random.
Definition 9. A social choice rule, f , is manipulable if there exists a voter i and a profile
of preferences, (1 , . . . , n ), such that for some preference profile 0i ,
f (X, (1 , . . . , i−1 , 0i , i+1 , . . . , n )) i f (X, (1 , . . . , n )).
As social choice rule non-manipulable if it is not manipulable.
To be non-manipulable, it must be that no voter can ever benefit from misreporting their
preferences. So it is always a best response to report truthfully, regardless of the preferences
of the other individuals and whether or not they are reporting truthfully. The individuals
then have a very simple decision: they don’t have to know the distribution of others’ preferences or worry about their strategies, reporting truthfully is always optimal. Clearly if a
15
For empirical evidence of how this can arise in practice see Spenkuch (2015), who looks at strategic
voting in an election to the German Bundestag where monotonicity was violated.
28
social choice rule is non-manipulable then, if the decision is made by this choice rule, it will
always be a Nash equilibrium for all individuals to truthfully report their preferences.
We have seen that if independence of irrelevant alternatives or monotonicity are violated
this can sometimes cause individuals to have an incentive to misreport their preferences. As
such, if instead of looking for a social choice rule that satisfies independence of irrelevant
alternatives and monotonicity we look for a choice rule that is non-manipulable, then, analogously to Arrow, we will find that no such rule exists. This leads to the Gibbard-Sattherwaite
theorem, proven independently by Allan Gibbard and Mark Satterthwaite.
Theorem 8. (Gibbard and Satterthwaite 1973) Suppose X has at least three alternatives and
there are a finite number of voters. Then there does not exist a non-manipulatable choice
rule that yields a deterministic winner for any preferences and satisfies pareto optimality that
is non-dictatorial.
The Gibbard-Satterthwaite Theorem demonstrates that any non-dictatorial, pareto optimal, voting system is potentially manipulable. The requirement that the choice rule be
deterministic rules out, for example, the random dictator choice rule, where one individual
is selected at random with her most preferred alternative selection. The random dictator
method is non-manipulable and pareto optimal but does not give a deterministic winner.
While the incentives to misreport preferences are clear in plurality rule elections, in other
settings the incentives are more subtle. In fact, proponents of electoral reform sometimes
argue in the press that a desirable property of other systems, such proportional representation
and single transferable vote (Hare), is that individuals would not have an incentive to vote
strategically. This is incorrect: if all preference profiles are possible then single transferrable
vote, proportional representation, or any other system devised, would be manipulable.16
As with Arrow’s Impossibility Theorem, the Gibbard-Satterthwaite Theorem depends
heavily on the possibility of any preference combination.17 As we have discussed, the logic of
Gibbard and Satterthwaite theorem is similar to Arrow’s result, and the proof closely mimics
the proof of Arrow’s Impossibility Theorem, replacing the use of independence of irrelevant
alternatives and monotonicity with non-manipulatability at the appropriate point in the
proof. As such we will not go through the proof in this class. The Gibbard-Satterthwaite
theorem is arguably as important, or even more important than, the Impossibility Theorem.
It was around the time that economists started taking information, and the fact that individuals may have private information: about their preferences, what they have done, their
risk or ability level, or about an object they may be selling. It was a core result establishing
the limit on what can be accomplished given the strategic incentives of the individuals in
the setting we are analyzing.
In the application we have considered, it is a voting mechanism that determines which
of a set of alternatives to select. Presumably we care about which alternative is selected
16
The incentives to mis-report in proportional representation is more complicated, since the resulting
policy results from bargaining between the different parties in the legislature depending on the number of
seats each party received. However, no matter what the mapping between seat share and policy outcomes is,
Gibbard-Satterthwaite demonstrates that there would be room for manipulation at least for some preferences.
For example, in the Austen-Smith and Banks (1988) model of government formation the influence of a party
can be monotonic in the share of votes it receives, so it is possible to gain from mis-reporting preferences.
17
When preferences are single-peaked, Condorcet’s method is non-manipulable and always yields a deterministic winner.
29
because it affects the allocation in the economy. In the remainder of the class we consider
how to design systems for allocating objects taking into account that individuals may have an
incentive to misreport their preferences. We will consider this in a simpler environment than
designing voting rules. As Arrow’s Impossibility Theorem demonstrates, it is not clear what
the objective should be even if we knew everyone’s preferences. In many other environments,
however, there is a natural objective. We will consider the limits on what is possible in those
environments.
13
Designing Institutions: What Can We Achieve?
In Econ 201, and in the models of political competition we have considered, we looked at
a specific game and solved for the equilibrium. A different question is: suppose we could
choose the game we could have the individuals play. How would we design the system to
ensure a desirable outcome is obtained? This is similar to the question asked in the Arrow
and Gibbard-Satterthewaite Theorems when we considered whether it is possible to design
a choice procedure that satisfies what we consider to be desirable properties. There we are
asking, if we could choose the electoral system to be used, could we design one that is immune
to the issues we have discussed. We saw that this was not possible, and such theorems serve
to establish the limits of what we can accomplish. This is important for evaluating any
procedure we might use.
If the preferences of all individuals were known, then conditional on determining which
outcome is desirable (itself a difficult problem as we have seen), designing the game to induce
this behavior is trivial—we could specify the game where the only action available to each
player is the action that we want them to take. However, if preferences are unknown, then
we won’t know what the most desirable action is and the problem becomes more difficult.
Moreover, when there are information asymmetries, decentralized markets may not ensure
a socially efficient outcome.
Of course, as we’ve seen, even if we were to know the preferences of all individuals it is not
always clear which alternative should be selected. However, there may be some alternatives
that are clearly suboptimal (e.g. if the alternative is pareto dominated) or some sensible
objective to focus on. For example, when determining how much of a public good to provide
we want to ensure that the outcome is pareto optimal—which as we saw in Econ 201 may not
be the case when left to voluntary contributions of individuals. Another application would
be when selling one (or multiple) object(s). One goal could be to generate as much revenue
for the seller as possible. In this case we are considering second-degree price discrimination
or the design of auctions and we could be trying to generate revenue for the government or a
company. Alternatively we might try to allocate objects to the individuals who value them
most highly, ensure an efficient match between individuals and institutions, or mediate a
dispute to prevent a destructive outcome such as war or the costs of a trial. We begin by
looking at a setting in which there is one object that must be allocated to either the buyer
or the seller. We will then consider the problem of public good provision.
We now turn to the problem of facilitating trade. In this case we would think of the
mechanism designer as a broker, mediator, or arbitrator whose objective is to make efficient
trades happen (assume that this mediator is disinterested third party who only wants to
30
ensure that efficient trades happens and inefficient ones don’t). The deal in question could
be a salary negotiation, one firm could be considering acquiring another (privately held)
firm, or a buyer could be considering a major purchase such as a home. In each case the
seller of the worker’s time, the firm, or the house, has some valuation from walking away and
the buyer has some valuation of owning it, and this valuation is private information. The
designer’s task is to design the rules by which the bargaining should take place for a given
objective.
Suppose the buyer has valuation vb for the object, and the seller has valuation vs for
the object. As we may not know the valuation of the buyer or the seller, assume that
vs ∈ [v s , v̄s ] and vb ∈ [v b , v̄b ], where v s < v̄s and v b < v̄b . We assume that the distribution
for both the seller and the buyer has full support on their interval. That is, P r(vb < v) is
strictly increasing in v on [v b , v̄b ] and P r(vs < v) is strictly increasing in v on [v s , v̄s ]. In
essence, full support says that the density from which vs and vb are drawn from is always
strictly positive.
The buyer and seller each knows their own valuation, but both only know the distribution
of the other’s valuation. Similarly, whoever is designing the mechanism only knows the
distribution of valuations for each individual. Assume also for simplicity, that vb and vs are
independent: this means that there is no information revealed to the buyer from the seller’s
valuation. There are then two things to determine: (1) does the buyer get the object or does
it stay with the seller? (2) how much should the seller receive from the buyer? Let x ∈ [0, 1]
denote the probability the object is transferred to the buyer. That is, if x = 1 sale happens
with certainty, x = 0 means it never does, and x ∈ (0, 1) means the object is sold with some
positive probability. Suppose that the buyer transfers amount t to the seller. The surplus of
the buyer and seller from participating in the transaction are
ub (vb , x, t) = vb x − t,
us (vs , x, t) = t − vs x.
Note that we are normalizing the payoff to not trading to be equal to 0, and assuming that
both the buyer and seller are risk neutral. In this simple setting, a natural goal is to ensure
that trade happens if and only if vb > vs (i.e. x = 1 if vb > vs and x = 0 if vb < vs . If vb > vs
then it cannot be pareto efficient for trade not to happen: if the buyer and seller agree to
trade at any price t ∈ (vs , vb ) both are made better off. Similarly, if vs > vb it cannot be
pareto efficient for trade to happen: if the buyer were to sell the object back to the seller for
any price in (vb , vs ) this would make both better off. From the perspective of the mechanism
designer the price may not be important part of the objective as it simply reflects a transfer
from one player to another and so different prices cannot be pareto ranked.
Since don’t know ex-ante, when designing the mechanism, the realized valuations of the
buyer and seller respectively we seek to design an allocation rule such that the individuals
trade if and only if vb > vs . We say that the mechanism is ex-post efficient if it ensures
trade whenever trade is socially efficient given the realized vs and vb . Ex-post means that
the mechanism we designed before the valuations of the individuals are determined will be
efficient after the valuations have been realized. Note that if we cannot compel the buyer
or the seller to trade then each must prefer to do so rather than walk away. So, for any
valuation the seller/buyer could have they cannot receive negative utility. We call this
31
individual rationality, and our goal is to find an individually rational trading mechanism
under which the final allocation is ex-post efficient.
One mechanism to allocate the object is to allow the seller to set a price, then the buyer
decides whether to accept the price or not. We can quickly see that this may not generate an
equilibrium that is ex-post efficient. In order to calculate the equilibrium we need to assume
the specific distribution of buyer and seller valuations. Let us assume that the buyer and
seller valuations are both uniform on [0, 1]. If we allow the seller to set the price then the
seller would set a price t(vs ): t is the transfer or price she receives from the buyer in the
event of a sale. Note that the seller knows her own valuation and so can condition the price
she sets on her valuation. We assume that the buyer and seller valuations are independent so
the seller’s valuation does not affect the buyer’s valuation: this is different from the market
for lemons problem in which the seller’s valuation reveals information to the buyer—while it
is possible to address that case, it is more complicated and so we focus on the simplest case
in which valuations are independent.
What price would the seller set? Note that if she sets price t then a buyer will purchase
the object if and only vb ≥ t. Assuming that vb is uniformly distributed on [0, 1] this happens
with probability 1−t. Hence, if the seller sets price t she sells the object at price t, generating
surplus t − vs , with probability 1 − t, and with probability t she keeps the object generating
0 surplus. Hence the expected utility of the seller from setting price t is
us = (1 − t)(t − vs ).
The seller sets the price t to maximize her utility. This is a straightforward maximization
problem and taking derivative and solving for the maximum shows that she would set price
vs + 1
.
2
But note that then not all efficient trades are realized: the object is sold to the buyer if and
only if vb ≥ vs2+1 , so if vb ∈ (vs , vs2+1 ) trade would be efficient but would not take place. This
is an example of market-power: if the seller prices at cost (i.e. her valuation), all efficient
trades take place, but the seller would receive no surplus. If the seller prices above cost, this
prevents some, but not all, efficient trades from occurring, but allows the seller to extract
positive surplus from the trades that do happen.
We have seen an example in which the game induced by the seller setting the price is not
ex-post efficient. But this is just one possible mechanism, and an example with one specific
distribution of valuations. We can imagine many other ways to facilitate trade. Is there
any way that will ensure an ex-post efficient outcome? Myerson and Satterthwaite (1983)
shows that it is not possible to construct a mechanism that is both individuality rational and
ex-post efficient. This impossibility theorem does not depend on the assumption that buyer
and seller valuations are uniformly distributed, but extends to any distribution with strictly
positive density on the respective intervals [v s , v̄s ] and [v b , v̄b ] in which there is overlap in
the possible valuations of the buyer and seller. It is important that the two intervals have a
non-empty intersection: if v̄s ≤ v b then having a third party set price any t ∈ [v̄s , v b ] would
lead to efficient trade.
t(vs ) =
Theorem 9. (Myerson and Satterthwaite 1983) Suppose vs is drawn from a distribution with
full support on [v s , v̄s ] and vb from a distribution with full support on [v b , v̄b ]. Suppose that vs
32
and vb are independent and (v s , v̄s ) ∩ (v b , v̄b ) 6= ∅. Then there does not exist an individually
rational bilateral trading mechanism that is ex-post efficient.
The Myerson-Satterthwaite theorem shows that, when individuals have private information, it is not always possible to ensure a pareto optimal allocation no matter how we design
the system. However, this does not mean any mechanism is equally good, and we can compare different mechanisms by how close they come to ex-post efficiency. For example, we
have seen what happens when the trading mechanism is to allow the seller to make a take it
or leave it offer to the buyer. This is not ex-post efficient, but no mechanism will be ex-post
efficient. Is simply allowing the seller to make a take it or leave it offer to the buyer the
best we can do, or is it possible to increase efficiency with a different mechanism, and, if so,
what mechanism? We turn to answering that question in the next section. While we will
not prove the Myerson-Satterthwaite theorem, we will prove a simpler discretized version.
In this environment we will demonstrate that it is not possible to ensure the ex-post efficient
allocation and will solve for the limit on what we can achieve.
The Myerson-Satterthwaite theorem is applied much more generally than to the sale of
an object. It basically says that when there is private information, it is not possible to
guarantee that an efficient outcome will be chosen. This is often applied to study socially
inefficient events such as war, strikes, or delay in reaching deals on the debt ceiling or the
sequester. It says that in environments where a small number of individuals are interacting,
and preferences are unknown, it will not necessarily be possible to ensure the outcome is
pareto efficient. This shows a limit of the Coase Theorem, since even if we could design
the optimal way for the trade to take place the outcome may be inefficient. When there
are a small number of players (so each player has “market power”) information can prevent
efficiency. We consider this difference in problem set 3.
14
The Revelation Principle and Optimal Mechanisms
We now turn to considering what can be accomplished. We consider what game we could
have the buyer and seller play in order to maximize the total surplus from trade. This is
a complicated problem: we could specify any game we could imagine and need to consider
what the equilibrium would be in that game. Our life is made a lot easier due to an important
result introduced by Roger Myerson known as the revelation principle (Myerson, 1979, 1981).
This was the basis of Myerson’s nobel prize in 2008. The revelation principle, in essence,
says that when designing the mechanism (i.e. choosing the game to have the individuals
play) it is sufficient to restrict attention to games in which the only action the individuals
take is to announce their own valuation. This does not mean we should always use a direct
mechanism, but it is useful for determining the limits on what we can and cannot achieve,
and for evaluating the efficiency of different mechanisms.
In mechanism design we are specifying the rules of the game the individuals will play.
Assume there are n players, and that each player i has a type vi ∈ Vi . This type could
reflect the valuation placed on an object or a public good, or something different such as
the individuals ideological ideal point or preferred school. Each individual i knows their
own type, and everyone else only knows the distribution individual i’s type is drawn from.
We specify the game for each individual to play, G, and each individual has a strategy si
33
which determines the action they would take as a function of their type (e.g. In the price
setting game, the strategy for the seller is what price to set for any valuation, and the buyer’s
strategy is the decision, for any valuation, of whether to accept the seller’s offer). As each
individual (may) not know the types of other players, and so may not know which action
they take (e.g. the seller doesn’t know the buyer’s valuation and so doesn’t know if her offer
will be accepted), each player seeks to maximize their expected utility. A strategy profile
s∗ = (s∗1 , . . . , s∗n ) constitutes a (Bayesian) Nash Equilibrium if, for all individuals i, all types
vi , and all strategies s0i ,
Ev−i [ui (s∗1 (v1 ), . . . , s∗n (vn ))] ≥ Ev−i [ui (s∗1 (v1 ), . . . , s∗i−1 (vi−1 ), s0i (vi ), s∗i+1 (vi+1 ), s∗n (vn ))].
That is, a strategy profile constitutes a Nash equilibrium if each player is maximizing their
expected utility given their type and the probability distribution over the other players’
types.
For any game G we can then calculate the set of equilibria, and resulting outcome of
the game. The revelation principle is the key insight that, if there exists a game G with an
equilibrium s∗ , then we can define a new game G0 as follows. The allowed strategy of each
player is to report their type vi ∈ Vi . We call G0 a direct mechanism, since all we allow the
individuals to do is report their type. Under the game G0 the outcome if each individual
reports vi is the allocation (s∗1 (v1 ), . . . , s∗n (vn )) from the game G. Note that if s∗ is a Nash
equilibrium, then no individual i could do better under a different strategy: including the
strategy of playing s∗i (vi0 ) for some vi0 6= vi when the individual’s valuation is vi . Hence if
s∗ is an equilibrium of G then it is an equilibrium for each individual to truthfully report
their type in the game G0 . We refer to a mechanism for which it is an equilibrium for each
individual to report their type truthfully as incentive compatible.
The above argument shows that we can restrict attention to a mechanism or game, G, in
which each player simply reports their true type, vi , and truthful revelation is incentive compatible. We apply the revelation principle to study the optimal bilateral trading mechanism.
In order to be able to calculate the optimal mechanism, we consider a simpler environment in
which the seller and buyer’s valuations are both binary. Assume that the buyer’s valuation
could be either 1 or 5 each with equal probability, and the seller’s valuation could be 0 or
4, each with equal probability. The buyer and seller’s valuation are independent. To solve
this we seek to maximize our objective—the total gains from trade—given the individual rationality and incentive compatibility constraints. This is an, admittedly more complicated,
version of the constrained maximization problems we saw in Econ 200 and 201.
Such a mechanism can then be represented by the probability of trade x(vs , vb ) for any
reported valuation, and the transfer from the buyer to the seller from this reported valuation t(vs , vb ). As there are two types for each the seller and the buyer, the mechanism
must specify and probability of trade and a transfer for each of the four type combinations.
Hence the mechanism consists of solving for eight numbers: x(0, 1), x(0, 5), x(4, 1), x(4, 5)
and t(0, 1), t(0, 5), t(4, 1), t(4, 5).
Since the buyer doesn’t know the seller’s valuation, if the buyer announces that she
b)
and the exhas valuation vb then the expected trade probability is xb (vb ) = x(0,vb )+x(4,v
2
t(0,vb )+t(4,vb )
pected payment is tb (vb ) =
. Similar calculations apply for the seller, xs (vs ) =
2
x(vs ,1)+x(vs ,5)
t(vs ,1)+t(vs ,5)
and ts (vs ) =
. We need to ensure the incentive compatibility and
2
2
34
individual rationality constraints for the buyer and the seller of each type in any mechanism.
Taking the buyer first we have
5xb (5) − tb (5) ≥ 5xb (1) − tb (1),
xb (1) − tb (1) ≥ xb (5) − tb (5),
and
5xb (5) − tb (5) ≥ 0,
xb (1) − tb (1) ≥ 0.
The first two inequalities are the IC constraints for each type, and the last two inequalities
are the IR constraints. Similarly, for the seller we get
ts (0) ≥ ts (4),
ts (4) − 4xs (4) ≥ ts (0) − 4xs (0),
and
ts (0) ≥ 0,
ts (4) − 4xs (4) ≥ 0.
Note that, to be ex-post efficient the mechanism must involve
x(4, 1) = 0,
and
x(0, 1) = x(0, 5) = x(4, 5) = 1.
We first show that there is no incentive compatible (IC) and individually rational (IR)
mechanism that is ex-post efficient.
Theorem 10. (Discrete Myerson-Satterthwaite) Suppose the buyer’s valuation is vb ∈ {1, 5},
both with equal probability, the seller’s valuation is vs ∈ {0, 4} both with equal probability, and
the buyer and seller valuations are independent. Then there does not exist an individually
rational bilateral trading mechanism that is ex-post efficient.
Proof. By the revelation principle, it is sufficient to show that there is no incentive compatible
and individually rational direct mechanism that is ex-post efficient. To have an ex-post
efficient allocation, the allocation is pinned down for any realization of valuations. So we
must show that there do not exist a set of transfers t(0, 1), t(0, 5), t(4, 1), t(4, 5) consistent
with IC and IR and the ex-post efficient allocation. To show this we will demonstrate that
it is not possible to satisfy the IC constraints when vb = 5 and vs = 0 and the IR constraints
when vb = 1 and vs = 4 simultaneously. To show this we proceed by contradiction, and
will show that it is not possible to simultaneously satisfy the IC constraints for the high
valuation buyer and the low valuation seller while also satisfying the IR constraints for the
low-valuation buyer and the high valuation seller.
Suppose there were, and so both the buyer and seller are reporting truthfully. As the
buyer and seller have independent valuations, regardless of their own valuation they believe
35
each valuation is equally likely for the other player. First consider the buyer, and focus on
the IC constraint of the high valuation buyer. If a seller reports vb = 5 she gets the object
with probability
1
xb (5) = (x(0, 5) + x(4, 5)) = 1
2
and pays
1
tb (5) = (t(0, 5) + t(4, 5))
2
in expectation. If she reports vb = 1 she gets the object with probability
1
xb (1) = (x(0, 1) + x(4, 1)) = 1/2
2
and pays
1
tb (1) = (t(0, 1) + t(4, 1)).
2
So to have the buyer report truthfully when her valuation is vb = 5 we must have
xb (5)5 − tb (5) ≥ xb (1)5 − tb (1)
or equivalently
5
≥ tb (5) − tb (1).
(1)
2
We next consider the IR constraint for the buyer with vb = 1. The mechanism we must
have
xb (1) − tb (1) ≥ 0
or equivalently
tb (1) ≤ 1/2.
Now consider the seller. If she reports vs = 4 the object is sold with probability
1
xs (4) = (x(4, 1) + x(4, 5)) = 1/2
2
and she receives
1
ts (4) = (t(4, 1) + t(4, 5))
2
in expectation. If she reports vs = 0 the object is sold with probability
1
xs (0) = (x(0, 1) + x(0, 5)) = 1
2
and she receives
1
ts (0) = (t(0, 1) + t(0, 5)).
2
Now consider the IC constraint when vs = 0. To be satisfied we must have
ts (0) − xs (0)0 ≥ ts (4) − xs (4)0,
36
(2)
or equivalently
ts (0) − ts (4) ≥ 0.
(3)
For the IR constraint when vs = 4, seller is willing to participate if and only if
ts (4) − xs (4)4 = ts (4) − 2 ≥ 0.
(4)
We now show that it is not possible to simultaneously satisfy (1)–(4). By (2)
t(0, 1) + t(4, 1)
1
≤
2
2
so
t(0, 1) + t(4, 1) ≤ 1.
And from (1) and (2),
5
t(0, 5) + t(4, 5) 1
≥ tb (5) − tb (1) ≥
−
2
2
2
or equivalently
t(0, 5) + t(4, 5) ≤ 6.
Adding up these two constraints, to get the IC satisfied for the high valuation buyer and the
IR for the low valuation buyer we must have that
t(0, 1) + t(0, 5) + t(4, 1) + t(4, 5) ≤ 7.
Now note that by (4),
t(4, 1) + t(4, 5)
≥2
2
so
t(4, 5) + t(4, 1) ≥ 4.
Further by (3) and (4),
ts (0) =
t(0, 1) + t(0, 5)
≥ ts (4) ≥ 2
2
so
t(0, 1) + t(0, 5) ≥ 4.
Adding up these two conditions, to get the IR satisfied for the high valuation seller and the
IC for the low valuation seller we must have
t(0, 1) + t(0, 5) + t(4, 1) + t(4, 5) ≥ 8.
This, obviously contradicts that t(0, 1) + t(0, 5) + t(4, 1) + t(4, 5) ≤ 7 and so the constraints
for the buyer and seller cannot be satisfied simultaneously. Hence there cannot exist an
ex-post efficient mechanism.
37
The proof of the discrete Myerson-Satterthwaite Theorem illustrates the fundamental
friction which makes it impossible to ensure ex-post efficiency. The IR constraints put an
upper bound on how much the lowest-valuation buyer can be expected to pay, and the IC
constraints but a bound on how fast the expected payment can be increasing in the buyer’s
valuation. In particular, since a buyer with a higher valuation could always mimic one with
a lower valuation, such a buyer must expect to pay less than their valuation when from
reporting truthfully. In the discrete setting,
xb (5)5 − tb (5) ≥ xb (1)5 − tb (1) > xb (1) − tb (1) ≥ 0,
and so
tb (5) < 5xb (5).
Reversing these arguments, the seller must expect to receive more than her valuation from
reporting truthfully (unless she has the highest possible valuation). This means that when
the seller’s valuation is slightly lower than the buyer’s valuation, so trade will be efficient, it
will not be possible to ensure it happens.
We now turn to the optimal mechanism. We define the objective of the designer to be
to maximize the total gains from trade:
X
E[x(vs , vb )(vb − vs )] =
p(vs , vb )x(vs , vb )(vb − vs )
vs ,vb
x(0, 1)(1 − 0) + x(0, 5)(5 − 0) + x(4, 1)(1 − 4) + x(4, 5)(5 − 4)
4
5x(0, 5) + x(0, 1) + x(4, 5) − 3x(4, 1)
.
=
4
=
Now that we have our objective we now look at the constraints that must be satisfied. We
need to look at the IC and IR constraints for the seller and buyer of each type. Combining
the IC and IR constraints we have a total of eight constraints. In addition we need to
make sure that x(vs , vb ) ∈ [0, 1] for each vs and vb . We then must maximize a our objective
5x(0, 5) + x(0, 1) + x(4, 5) − 3x(4, 1) subject to (1) the IC and IR constraints (2) x(vs , vb ) ∈
[0, 1] for each vs and vb .
This is now a standard problem like we have seen in Econ 200 and Econ 201: We can
set up the Lagrangian to maximize the objective, with a Lagrange multiplier for each of
the (sixteen) constraints. While not all constraints bind, we can solve for the Kuhn-Tucker
conditions to determine which constraints bind which don’t. Alternatively we can reason
out which constraints will bind, use this to reduce the number of constraints. Even with
these simplifications, however, since there are so many constraints the computations become
rather involved. So, in the interest of time, we will not explicitly work through solving the
optimal mechanism, but rather turn to a simpler problem in the next section.
While we have set up the problem with discrete valuations, we could do the same thing
when the preferences of the buyer and the seller are both uniformly distributed as in the
previous section. While it is possible for us to solve for the surplus maximizing mechanism, it
is more complicated than the constrained maximization problems we have seen. The reason
for this this is that we now must maximize over the functions x(vs , vb ) and t(vs , vb ) which
38
each have a continuum as their domain. For this reason we will not set up or solve this
problem in this class. The optimal mechanism given the incentive constraints is solved for
in Myerson and Satterthwaite (1983) and the solution is quite intuitive.
The optimal mechanism involves allocation rule
1 if vb > vs + 1/4,
x(vs , vb ) =
0 otherwise.
This allocation rule is implemented with transfers
vs +vb +0.5
if vb > vs + 1/4,
3
t(vs , vb ) =
0
otherwise.
There are two important take-away points from this solution. First, we can see that
it is not possible to design a mechanism that is guaranteed to generate an ex-post efficient
allocation: when vs < vb < vs +1/4 it would be efficient to have the object sold, but under the
optimal mechanism it isn’t. This is shows the Myerson-Satterthwaite theorem for the case
with uniform distribution. That we cannot achieve the first best shows us that informational
constraints are real constraints applying to the problem that must be taken seriously and
that limit what we can achieve. We often refer to the an optimal mechanism, given the
incentive constraints, as “second best”. Second, for any mechanism that is proposed, since
we can calculate the optimal mechanism, we can determine whether that mechanism achieves
the “second best” or not. We can see that the trading mechanism of having the seller make
a take it or leave it offer to the buyer does not achieve the second best. When the seller
makes it a take it or leave it offer, the result is that
1 if vb > vs2+1 ,
x(vs , vb ) =
0 otherwise.
As this differs from the optimal, we can see that having the seller make a proposal to buyer
is not the “second best” mechanism.
While the revelation principle says that there is a direct mechanism that implements the
second-best allocation, it is possible to implement this allocation with a mechanism that
seems more likely in practice. Suppose we use the following mechanism: the buyer and seller
b
each simultaneously set a price, ps and pb , and the mechanism is to trade at price t = ps +p
2
if ps ≤ pb and to not trade otherwise. As we will see on problem set 4, that mechanism is
not incentive compatible, and the seller will set a price ps > vs and the buyer will set a price
pb < vb . However, in the equilibrium of this game, the second-best allocation characterized
above is attained: we will solve for the equilibrium of this game in problem set 4. So we
can verify that the seller making a take it or leave it offer to the buyer is not optimal, but
allowing the buyer and seller to each set price and trading at the average of the two prices
is.
What are the economics behind why it is better to give both the buyer and the seller
an opportunity to set a price rather than simply allow the seller to set the price? Recall
that, when the seller sets the price, she has market power and so will set a price higher than
her valuation in an effort to extract surplus. As we have seen in Econ 201, the deadweight
loss associated with market power is quadratic in the amount of market power the seller
39
possesses. By giving both the buyer and the seller proposal power this is split between the
two individuals, meaning that the marginal effect of each agent exerting market power is
decreased.
15
Regulating a Monopolist
In the last section we considered the design of an institution to facilitate trade between a
buyer and a seller. We now consider the possibility of regulating a market. As we have seen
in Econ 201, a market in the absence of government intervention may or may not be efficient:
Efficiency may break down due to externalities (which we consider in the next section by
looking at public good provision) or when there is market power, which we consider in
this section. If the government has all relevant information, and the authority to regulate
the market, it can simply force everyone to take actions that ensure the outcome is pareto
efficient. If there is some information the government lacks (as is generally the case) then its
problem becomes more complicated. Here we consider a simplified version of the problem,
introduced in Baron and Myerson (1982), of regulating a monopolist with unknown costs.
Suppose the monopolist could either have high or low cost, each with probability 1/2.
To make the problem as simple as possible, assume the firm has linear costs
c(q) = cq
where c ∈ {1, 2} is the firm’s cost of producing each unit and q is the quantity produced.
We say that a firm is low cost if c = 1 and high cost if c = 2. The firm knows its own cost, c,
but the government only knows the distribution of costs. Both the firm and the government
know the distribution of buyer preferences. The demand for the product the monopolist sells
is
q(p) = a − p
where a > 3.18
From Econ 201 we know that the monopolist, if not regulated, will set p > c, generating
deadweight loss and the possibility that government regulation could improve efficiency. We
assume that the government can choose the price to regulate the firm at, p, and also the
transfer (positive or negative), t, to make to the firm. If it sets t > 0 then it is subsidizing
the firm, whereas if t < 0 it is taxing (charging a fee) for the firm to be able to operate. The
price in the market then determines the quantity based on consumer demand. We assume
the objective of the government is to maximize the consumer surplus less the amount of the
transfer paid to the monopolist,
Z q(p)
(a − p)2
− t.
CS(p) − t =
(a − q − p)dq − t =
2
0
Hence the government doesn’t care about the firm’s profit, but the firm has the option to
shut down if it were to earn negative profits.19
18
Assuming that a > 3 will ensure that it will be optimal for the government to allow both the high and
low cost firm to operate; when a < 3 it will be optimal to force the high cost firm to shut down. Why a > 3
is the necessary condition will become clear later.
19
We could allow the government to care about firm profits/producer surplus as well. However we want to
incorporate that the government would prefer not to make transfers to the firm if would operate anyways.
40
The firm’s profits are then
π(c, p, t) = t + (p − c)q(p) = t + (p − c)(a − p)
and the IR constraint is that π(c, p, t) ≥ 0. If the government knew the firm’s cost c then it
would choose −t = (p − c)(a − p) to capture all surplus, and then choose p to maximize
(a − p)2
+ (p − c)(a − p).
2
Taking FOC,
−(a − p) + (a − p) − (p − c) = 0
so it would be optimal to set p = c (hence t = 0). This is familiar from Econ 201: to maximize
the total surplus we would want to regulate the firm to price at cost, and here we can make
sure all of that surplus goes to consumers. We now consider the optimal mechanism when
the government does not know the firm’s cost.
By the revelation principle it is sufficient to restrict attention to direct mechanisms which
are incentive compatible and individually rational. In the direct mechanism the government
simply asks the firm to report its cost c ∈ {1, 2} then chooses the price pc and transfer tc .
Since each cost is equally likely the government’s objective is then to maximize
(a − p2 )2
1 (a − p1 )2
− t1 +
− t2 .
(5)
2
2
2
The first thing the government must do is ensure the firm will continue to operate. This
gives the IR constraints for the low and high cost firms respectively:
π(1, p1 , t1 ) = t1 + (p1 − 1)(a − p1 ) ≥ 0,
(6)
π(2, p2 , t2 ) = t2 + (p2 − 2)(a − p2 ) ≥ 0,
(7)
The government also needs to make sure that both types of firm truthfully report their cost.
That is we must have π(1, p1 , t1 ) ≥ π(1, p2 , t2 ) and π(2, p2 , t2 ) ≥ π(2, p1 , t1 ). Substituting
into these equations we get the IC constraints:
t1 + (p1 − 1)(a − p1 ) ≥ t2 + (p2 − 1)(a − p2 ),
(8)
t2 + (p2 − 2)(a − p2 ) ≥ t1 + (p1 − 2)(a − p1 ).
(9)
We are then left with four constraints. Note, however, that some of them are redundant.
In particular, we do not have to worry about the low cost firm dropping out of the market:
the payoff to a low cost firm of selling at the high cost firm’s price is always higher than
to the high cost firm, but the high cost firm’s profits are non-negative or its IR constraint
would be violated. So combining (8) and (7) we have
t1 + (p1 − 1)(a − p1 ) ≥ t2 + (p2 − 1)(a − p2 ) ≥ t2 + (p2 − 2)(a − p2 ) ≥ 0
so the IR constraint for the low cost firm is automatically satisfied if the other constraints
are satisfied. Next note that, since the IR constraint for the low cost firm is redundant, the
41
IC constraint must hold with equality. Why? If not, the government could decrease t1 and
still have the firm operate and select the correct contract. But if the IC constraint for the
low cost firm is satisfied with equality and p2 ≥ p1 , the IC for the high cost firm holds for
free. When (8) holds with equality,
t1 + (p1 − 2)(a − p1 ) = t1 + (p1 − 1)(a − p1 ) − (a − p1 )
= t2 + (p2 − 1)(a − p2 ) − (a − p1 )
≤ t2 + (p2 − 1)(a − p2 ) − (a − p2 )
= t2 + (p2 − 2)(a − p2 )
so the IC constraint for the high cost firm is automatically satisfied as well. This argument
means that only two of the constraints, the IC for the low cost firm and the IR for the high
cost firm, are relevant and we only have two constraints to worry about.20
So we are left maximizing (5) subject to (7) and (8). Letting λ1 and λ2 be the Lagrange
multipliers associated with (8) and (7) the Lagrangian for this problem is,
1 (a − p1 )2
(a − p2 )2
L(t1 , t2 , p1 , p2 , λ1 , λ2 ) =
− t1 +
− t2 +
2
2
2
λ1 (t1 + (p1 − 1)(a − p1 ) − t2 − (p2 − 1)(a − p2 )) + λ2 (t2 + (p2 − 2)(a − p2 )).
Taking First Order Conditions,
1
∂L(t1 , t2 , p1 , p2 , λ1 , λ2 )
= λ1 − = 0,
∂t1
2
∂L(t1 , t2 , p1 , p2 , λ1 , λ2 )
1
= −λ1 + λ2 − = 0,
∂t2
2
a − p1
∂L(t1 , t2 , p1 , p2 , λ1 , λ2 )
=−
+ λ1 (a + 1 − 2p1 ) = 0,
∂p1
2
∂L(t1 , t2 , p1 , p2 , λ1 , λ2 )
a − p2
=−
− λ1 (a + 1 − 2p2 ) + λ2 (a + 2 − 2p2 ) = 0,
∂p2
2
∂L(t1 , t2 , p1 , p2 , λ1 , λ2 )
= (t1 + (p1 − 1)(a − p1 ) − t2 − (p2 − 1)(a − p2 ) = 0,
∂λ1
∂L(t1 , t2 , p1 , p2 , λ1 , λ2 )
= t2 + (p2 − 2)(a − p2 ) = 0.
∂λ2
From the first two FOCs we get that
λ1 = 1/2,
λ2 = 1.
20
The argument for why we need only consider the IR for the high cost firm and the IC for the low cost
firm is the same as the argument for why, in the Discrete Myerson-Satterthwaite proof, we looked at the
IR constraints for only the low valuation buyer and the high valuation seller, and the IC constraints for the
high valuation buyer and the low valuation seller.
42
Substituting this into the third and fourth FOC we see
p1 = 1,
and
p2 = 3.
Finally from the last two FOCs we have
π(2, t2 , p2 ) = t2 + (3 − 2)(a − 3) = 0,
so t2 = 3 − a, and
π(1, t1 , p1 ) = t1 = 3 − a + (3 − 1)(a − 3) = a − 3 > 0.
Comparing the optimal mechanism when the government is unsure of the firm’s cost to
the case where costs are known we see the following:
1. It is not possible for the government to implement the full information solution when
the monopolist privately knows its own cost.
2. The optimal regulation involves the government setting the efficient price for the low
cost firm (p1 = 1) but giving the low-cost monopolist a subsidy, t1 = a − 3 > 0, that
allows them to earn positive profits. This is because, otherwise, the low cost firm would
mimic the high cost firm and earn positive profits that way.
3. In the optimal regulation, the high cost firm is made exactly indifferent between operating and not operating, so receives none of the surplus. However the price set for the
high cost firm is not efficient. The government forces the firm to price above marginal
cost, then takes the surplus in the form of a fee collected for the firm to be able to
operate. The reason for this that, by increasing the price/decreasing the quantity it
reduces the surplus the low cost firm would receive when the high cost firm earns 0
profits. So the government creates an inefficiency when the firm is high cost so as to
be able to extract more surplus when the firm is low cost.
For those who studied second degree price discrimination in Econ 201 the logic is similar.
In that case the firm extracts all surplus from the low valuation consumer, but has to give
the consumer with a higher valuation some surplus in order for them to select the contract
meant for them. The seller also degrades the quality offered to the low valuation consumer
to reduce the amount of surplus it must share with the high valuation consumer. Here the
government extracts all surplus from the high cost (i.e. low profitability) firm, and forces it
to produce less than optimal, to reduce the amount of surplus it has to share with the low
cost (i.e. high profitability) firm.
43
16
The Provision of Public Goods
We now consider another classic problem: how to ensure an efficient amount of public good
provision. The provision of public goods is difficult since, as we saw in Econ 201, there is a
positive externality associated with contributing to a public good, and so each individual will
contribute less than the socially optimal amount. Hence we do not expect that voluntary
contributions of individuals to lead to a pareto efficient level of public good provision. If we
know how much each individual values the public good then calculating the efficient level
of public good provision is straightforward. Then we can specify the game in which each
individual must contribute the appropriate amount in order to pay for the efficient level of
the public good. However, it is difficult to get people to reveal their valuation of the public
good truthfully. Since all individuals value the public good, they must pay the part of the
cost or else they would have an incentive to announce a very high valuation. However if
the amount the individual must contribute increases too much based on their announced
valuation, they will have an incentive to under-report their true valuation and free-ride on
the contributions of others. This was the case in the example we studied on problem set 3.
A class of mechanisms that allows us to ensure that the social efficient level of public good
is provided are the Vickrey-Clarke-Groves mechanisms. William Vickrey who developed the
idea—that was later generalized independently by Clarke and Groves—in the context of
auctions was awarded a Nobel prize for this work in 1996. The idea behind the VickreyClarke-Groves (VCG) mechanism is simple. In order to make it incentive compatible for
individuals to truthfully report their preferences their contribution to the public good must
reflect their own value of increasing the public good, but not force the individual to pay for
the externality that increasing the public good has on others. If the game can be set up
in this way we can have individuals reveal their valuation and provide the efficient level of
public good.
Suppose there are n individuals i = 1, . . . , n with utility functions
ui (G, t) = vi log(G) + w − ti ,
where G is the amount of public good, and ti is the amount of the transfer individual i
must make the government to pay for the provision of the public good. Assume that each
individual’s valuation of the public good vi is their own private information. Note that, as
we have discussed, the pareto optimal allocation are those which maximize the sum of the
individuals’ utilities. This depends heavily on the fact that it is possible to make transfers
between the individuals (adjust the ti ’s). If this were not the case, allocations which do
not maximize the sum of utilities could still be pareto optimal. If we set up the problem of
maximizing
n
X
vi log(G) + w − ti
i=1
subject to the constraint
G≤
n
X
i=1
44
ti
we see immediately that any solution must have
G=
n
X
vi .
i=1
P
If our goal is to choose the socially efficient level of public good provision G = ni=1P
vi then
we need to look for an incentive compatible direct mechanism under which G = ni=1 vi .
The difficulty lies in constructing the transfers to make this mechanism incentive compatible.
Under a VCG mechanism each individual
Pn reports their valuation vi of the public good, the
level of public good provision is G = i=1 vi and each individual makes transfer
"
!
!#
n
X
X
X
X
ti (v1 , . . . , vn ) = t̄ + vi −
vj log
vk −
vj log
vk
.
j6=i
k=1
j6=i
k6=i
The amount t̄ is a constant determined by the mechanism designer that must be paid regardless of the reported valuation and so does not affect the individual’s incentives. Note
that vi reflects the increase in the optimal amount
of the public
good because individual i
P
P
Pn
P
values it, and j6=i vj log ( k=1 vk ) − j6=i vj log
k6=i vk is the amount that individuals
other than i would be willing to pay to increase the amount of public good by vi . The idea of
the VCG mechanism is as follows. Because the higher individual i’s valuation is the higher
the level of public good, and because increasing the amount of the public good provides a
benefit to other individuals, the VCG mechanism requires each individual to pay for their
own value of increasing the public good, but not the effect of this increase on the utility of
other individuals. In general, in a world with externalities, in order to make a mechanism
incentive compatible each individual must only pay for their own marginal valuation and not
the externality component. Note that here, regardless of the reports of the other individuals,
person i has an incentive to truthfully report her own preferences, vi . If individual i with
valuation vi reports valuation vi0 her payoff is
!
X
vi log
vj + vi0 + w − ti (v1 , . . . , vi−1 , vi0 , vi+1 , . . . , vn ) =
j6=i
!
vi log
X
j6=i
vj + vi0
!
+ w − t̄ − vi0 +
X
vj log
j6=i
X
vk + vi0
!
−
k6=i
X
j6=i
vj log
X
vk
.
k6=i
Taking derivative with respect to vi0 , and setting equal to 0 we get first order condition
P
vi
j6=i vj
P
−1+ P
= 0.
0
0
k6=i vk + vi
k6=i vk + vi
So we can see that the vi0 that maximizes individual i’s payoff is
vi0 = vi .
This means that it is a Nash equilibrium for all individuals to truthfully report their true
valuation, and so the VCG mechanism is incentive compatible. In fact, since individual i
45
has an incentive to report her valuation truthfully regardless of other’s valuation, the VCG
mechanism is non-manipulable or strategy-proof in the sense we described in section 12.
Note that a mechanism being strategy proof is stronger than it being a Nash equilibrium for
each individual to truthfully report their valuation: to be a Nash equilibrium no individual
can gain from misreporting their preferences in expectation, but to be strategy proof no
individual can benefit from misreporting their valuation no matter what the valuation of
the other individuals happens to be. That is, a mechanism is incentive compatible if it is a
Nash equilibrium for each person to report truthfully, it is strategy-proof if each person has
a dominant strategy to report truthfully. We often like mechanisms that are strategy-proof
since, especially with a large number of individuals in the economy, it may not be clear what
each individual believes about the others’ valuations.
One important feature to note of the VCG mechanism is that the total amount of public
good provided,
n
X
G=
vi ,
i=1
may not equal to the total amount collected
!#
!
"
n
n
n
n
X
X
X
X
X
X
X
.
vj log
vk
vj log
vk −
ti = nt̄ +
vi +
i=1
i=1
i=1
j6=i
k6=i
j6=i
k=1
This is because the amount each individual pays is different from their valuation and
"
!
!#
n
n
X
X
X
X
X
vj log
vk −
vj log
vk
i=1
j6=i
k6=i
j6=i
k=1
depends on the valuation of the individuals. Given that t̄ must be chosen before the vi are
reported we don’t know its value when designing the mechanism. And if t̄ depended on
the reported valuation this would change the incentives of the players and cause them to
misreport their valuations. In fact, there is no way to construct a mechanism that guarantees
both an efficient level of public good provision and that the amount collected will equal to
the amount spent on the public good. We call this budget balancedness.
Theorem 11. There does not exists an incentive compatible, budget-balanced mechanism
that guarantees the efficient level of public good provision.
So there are positive and negative conclusions from the study of public good provision
problem. While we can ensure the socially efficient level of public good, any mechanism that
ensures the socially optimal level of public good provision runs the risk of running budget
deficit or a budget surplus. We can construct mechanisms that ensure that the allocation
will be budget balanced, but those mechanisms will not ensure the socially efficient level
of public good. So again we see that there are tradeoffs. We can construct t̄ so that the
budget is balanced in expectation, but if we insist that budget balancedness must occur for
any profile (v1 , . . . , vn ) then we won’t be able to implement the efficient level of the public
good.
46
17
The Design of Markets and the Gale Shapley Algorithm
We conclude the course by looking at the design of markets in which two distinct sets of
individuals or institutions must be matched with individuals/institutions from the other set,
and both sides have preferences over who they match with. In the most environments the
seller of an object doesn’t care who they sell to but only the price they receive. However
there are many settings where who an individual matches with is very important. Some
examples include: matching students with schools (either college or more commonly due to
the centralized structure, public school systems), matching medical residents with hospitals,
matching kidney donors with those in need of a transplant, and marriage markets. Moreover,
in many of those settings, monetary transfers are either restricted or forbidden.
We begin by defining the Gale-Shapley algorithm for matching, a commonly used approach that is guaranteed to produce a “stable” match introduced in Gale and Shapley
(1962). We will then conclude the class by considering some applications of market design.
The Gale-Shapley algorithm is covered in Chapter 1 of the Gura and Maschler (2009) textbook, available on the library reserves section of Chalk. Lloyd Shapley shared the Nobel
Prize in economics in 2012 with Al Roth: Shapley for his theoretical work on developing this
algorithm, and Roth for applying it and adapting it to the design of different markets. A
survey article on market design written by Roth that highlights many practical issues and
applications of market design is available on Chalk. We will discuss these applications after
describing the algorithm.
Assume there are n individuals/institutions on each side of the market, and that each
individual/institution has strict preferences over the potential matches on the other side. For
concreteness, and to be consistent with the Gale-Shapley terminology, suppose there are n
men, a1 , . . . , an , and n women, b1 , . . . , bn , and that each side of the market matches with one
individual from the other side of the market (all heterosexual monogamous matches with no
monetary transfers). Assume that each individual can rank all potential matches as well as
not being matched. For simplicity, assume that each individual has strict preferences over
each potential match. Let ai and bj be the preferences of the ith man and jth woman
respectively. This preference order is over n + 1 alternatives: the n potential matches of the
other gender and the option to remain unmatched.
In a matching problem each individual/institution is matched with an individual/institution
on the other side of the market.21 Since an individual can’t unilaterally decide to match with
someone else we look for “stable” matches, in which no pair of individuals could agree to
match with each and both do better. If the matching outcome is not stable then two individuals could benefit by leaving their matches and matching with each other. Stability
is important since it ensures that no pair of individuals could decide, after the algorithm
has assigned matches for each pair, to leave the market and match with each other. This is
important when it is not possible to compel people to participate in the clearinghouse. We
define a stable match as follows.
21
A related problem, often referred to as the roommate assignment problem, is similar but instead of
assuming there are two different pools of individuals to match with the other pool, everyone is in the same
pool and we assign an efficient match.
47
Definition 10. A match is stable if there do not exist a man ai and a woman bj who are
not matched together but who prefer each other to the individual (if any) they are matched
with.
The Gale-Shapley Algorithm, sometimes called the Deferred Acceptance Algorithm, is a
simple procedure that works as follows. In the first round, each man “proposes” to the woman
who is his first choice. Any woman who received only one proposal becomes “engaged” to that
man unless she prefers to be unmatched, and any woman who receives multiple proposals
becomes engaged to her first choice among those who proposed to her, and “rejects” the
others. In each future round, every man who was rejected in the previous round proposes
to his most preferred woman who has not rejected him so far. Each woman chooses her
most preferred alternative among those who proposed to her in this round (if any) and the
individual she was engaged to in the previous period (if any) and rejects all but her most
preferred alternative. When we reach a stage at which no man is rejected, because there
is one man proposing to each woman (or deciding not to propose to any of the remaining
who have not rejected), the algorithm terminates. At this point the engagements become
“marriages” and the individuals are matched.
Obviously real marriage/dating markets do not work exactly like this, but this is literally
how students are assigned seats in a school choice program: they (or rather their parents)
submit a choice form detailing their preferences, and the school board puts this list in the
computer with the each school’s ranking of which students have priority for that seat, and the
Gale-Shapley, or a related, algorithm, is applied to assign students to different schools. That
problem is a little different than the one we considered here—in that case several students
are matched to each school and the school may not have strict preferences. However the
basic algorithm can be adapted to such a setting, and that problem is known as the college
assignment problem.
Note that, by construction, the Gale-Shapley algorithm cannot terminate without reaching a stable match: if any man prefers a different woman to the one he matches with he
would have proposed to her, and if that woman prefers him to the man she matched with
she would have accepted his proposal instead (once a woman is engaged this engagement
can only be broken if she gets a proposal she prefers). Moreover, since there are n men, and
each man can only propose to each woman at most once, it is clear that the Gale-Shapley
algorithm must terminate in no more than n2 rounds and so ultimately settles into a stable
match in finite time.
A more difficult issue is whether individuals have an incentive to truthfully reveal their
preferences and whether the outcome is efficient. Roth (1982) showed that each man has
an incentive to truthfully report his preferences, but that a woman may benefit from misreporting her true preferences. The reason for this is that the Gale-Shapley algorithm produces the optimal stable allocation from the perspective of the men, but not necessarily from
the perspective of the women. Intuitively, it gives the side that makes the proposals (men)
the ability to propose to anyone, but the women can only choose among those who have
proposed to them. We can see this from the following simple example.
Example 6. Suppose there are two men and two women. Man 1 prefers woman 1, man 2
prefers woman 2. The women have opposite preferences (woman 1 prefers man 2 to man
1 and woman 2 prefers man 1 to man 2), and everyone prefers matching with their second
48
choice to not matching. Under the Gale-Shapley algorithm, man 1 proposes to woman 1 and
man 2 proposes to woman 2. Since each man has only proposed to one woman, and the
women prefer to match, neither proposal is rejected. This means that we have a stable match
and the algorithm terminates. This is the optimal stable match for the men, but not for the
women: man 1 matching with woman 2 and man 2 with woman 1 is also stable, and both
women prefer it.
This also demonstrates that it is possible for the women to benefit from misreporting their
preferences. If woman 1 reported that she preferred not matching to matching with man 1,
then man 1 would be rejected and unmatched after round 1 and propose to woman 2. Woman
2 would then reject man 2 for man 1, and man 2 would propose to woman 1, who accepts and
the algorithm terminates. Note that by misreporting her preferences woman 1 gets a match
she prefers to the case where she reports truthfully.
While women may have an incentive to misreport their preferences, and there may exist
a different stable match that would make all women better off, there is no scope for men
to benefit from misreporting. Moreover, there is no other stable match that is preferred
by anyone on the men’s side. We say that a stable match is the optimal stable match
for the proposers if no proposer receives a higher utility in any other stable match. The
Gale-Shapley algorithm terminates at the proposer optimal stable match, since each man
proposes in decreasing order of preference, and once any stable match is reached the algorithm
terminates.
Theorem 12. The Gale-Shapley algorithm terminates in finite time and, if both sides have
reported their preferences truthfully, always produces a stable match. Moreover it is strategyproof for the proposers and generates the optimal stable match for the proposers.
The Gale-Shapley algorithm is commonly used in practice because it has many properties
that are considered desirable: namely that it always produces a stable match and is strategy
proof for the proposers. It has the drawback that it is not immune to manipulation on the
woman side, and that the match is only optimal on the proposer side. How the market
designer trades off different objectives, and how concerned we’d be about this, depends on
the specifics of the market considered. We consider some of these applications in the next
section.
18
Applications of Market Design
We now consider some applications of market design, beginning with school choice. See
Abdulkadiroğlu and Sönmez (2003), available on Chalk, for a discussion of some of these
issues related to school choice. In school choice there will be many students matched to each
school. While schools do not have utility functions, they have priority rankings over students.
These priority rankings typically take the form of giving higher priorities to students who
have a sibling currently attending the school, who lives in the walk zone, and to lower
income students (e.g. those who qualify for free or reduced lunch). As there are usually
many students in each priority ranking group, the students are also assigned a random
lottery number to generate a strict priority ranking over students.
49
How concerned we are about the possibility of misreporting on the “woman” side depends
on the application. If we are considering a setting in which students propose to schools,
which have publicly stated priority protocols, and the matching algorithm administered by
the school district which sets the priority, then we might not be too concerned about the
priority rankings being mis-reported.
It is however very important that the mechanism used be strategy proof on the “man”
or student side. First, when considering such a large market (such as the market for public
schools in New York city) it is very difficult to anticipate the beliefs individuals will have
about which schools others will apply to, and so determine the equilibrium of the game.
Second, if we want to know the preferences of individuals—either because we want to see
what characteristics of schools are valued by parents, or because we want to evaluate how well
the program is doing in assigning students to schools they want to attend—it is important
for individuals to have an incentive to report truthfully.22 Finally, when school districts have
used mechanisms that are not strategy proof (such as the so-called Boston mechanism that
gave higher priority to those who ranked a certain school as their first choice) websites and
parent groups have sprung up collecting data about admissions rates and trying to figure
out the optimal way to rank schools.23 Many school districts insist on a strategy-proof
mechanism since they are afraid that the children of less involved or strategic parents will
be placed at a disadvantage in terms of getting their students into their preferred school.
In school choice we generally evaluate the effectiveness of the match based on how well it
does for students, not on whether schools admit students with high priority rankings. So we
are unlikely to be troubled that the student optimal rather than school optimal mechanism
is used. We would also like the mechanism to ensure pareto optimality for the students, but
unfortunately Gale-Shapley does not guarantee that. In particular, pareto optimality may
conflict with stability, and the Gale-Shapley algorithm selects the stable allocation in this
case.
Example 7. Suppose there are three Students and three Schools. The preferences of Students
1, 2, and 3 are
b2 a1 b1 a1 b3
b1 a2 b2 a2 b3
b1 a3 b2 a3 b3
The Priority Rankings at schools 1, 2, and 3 are
a1 b1 a3 b1 a2
a2 b2 a1 b2 a3
a2 b3 a1 b3 a3
22
See Abdulkadiroğlu et al. (2015) for an empirical study of the welfare gains from improving the design
of markets in NYC. They find very large welfare gains from moving from a Gale Shapley style mechanism.
23
Under the Boston mechanism, the first slots in a school are allocated to those with a high priority who
ranked the school first. By advantaging those who rank an alternative first students have an incentive to
conserve their first place ranking by not using it on a school they know they are unlikely to get accepted by.
50
Under the Gale-Shapley algorithm, Student 1 proposes to School 2, and Student 2 and 3 each
propose to School 1, who accepts Student 3. Student 2 is then unmatched and so applies to
his second choice, School 2, who then rejects Student 1 and accepts Student 2. Student 1 then
proposes to School 1 who then rejects Student 3 for School 1. Finally Student 3 matches with
School 3, and the algorithm terminates. This leaves matches (a1 , b1 ), (a2 , b2 ) and (a3 , b3 ).
Notice however that, from the students’ perspective (a1 , b2 ), (a2 , b1 ) and (a3 , b3 ) is a pareto
improvement. However it is not stable: Student 3 prefers School 1 to the school it’s matched
with and has a higher priority for School 1 than student 2 does.
We are, of course, concerned about pareto optimality—there are, other algorithms that
are guaranteed to be pareto optimal—see, in particular, the top trading cycles mechanism
described in Abdulkadiroğlu and Sönmez (2003). But, as usual with these problems, there
are other tradeoffs. In particular, for a student assignment mechanism where only the
proposers have preferences, there is no strategy-proof algorithm that guarantees a stable,
pareto efficient match.24 While we have extensively discussed the value of pareto optimality,
stability is also considered important: it ensures no student is rejected by a school that they
preferred to the one they were assigned to, when they have a higher priority for acceptance
to the rejecting school than at least one of the accepted students. This is sometimes referred
to as “justified envy” and school boards may be constrained to satisfy it.
Another market in which matching is important, and monetary payments are forbidden, is
the market for kidneys. The demand for kidneys far exceeds the supply of kidney donations,
and an individual who needs a kidney will probably die if they don’t receive one.25 Donations
typically either come from cadavers or from individuals willing to done to a specific individual
they are close to. The standard economic response when demand exceeds supply is that the
price in the market should adjust to the market clearing price. While some economists (most
notably Gary Becker) have advocated for it to be legal to buy and sell kidneys there is federal
law that forbids the sale of kidneys or other organs,26 and we must deal with this constraint.
The kidney exchange market is an important area for market design because lives are at
stake, monetary transfers are prohibited, and matching process is complicated because the
donor and the recipient must be an appropriate match for each other or the recipient will
reject the transplant.
In some cases an individual would be willing to donate a kidney to an individual but may
not be a match to donate to that individual. For example, person A would like to donate to
B, and person C would like to donate to D, but A is only compatible with D and C is only
compatible with B. While monetary payments are illegal, pairwise donations can be done
legally. In a pairwise donation, A and C agree that they would each donate a kidney to the
person they are compatible with. Obviously, in practice, finding a suitable pair with which
to make such an exchange is difficult. But matching algorithms can and have been used to
help match sets of non-compatible donors and facilitate pairwise donation. It can also be
24
When both sides of the market have strict preferences the Gale-Shapley algorithm ensures the there is
no pareto improvement possible on both sides.
25
Approximately 7000 people on the waitlist for a kidney either day or are removed from the waitlist for
being to sick for surgery. It is likely a significant fraction of those people could have been saved if kidneys
were available earlier.
26
The market for kidneys is an example of what is called a “repugnant market” because people are
uncomfortable or disgusted by the idea of internal organs being sold for a price.
51
to identify cycles involving more than two simultaneous donations (e.g. A donates to D, C
donates to F, and E donates to B).
The New England Program for Kidney Exchange, created in 2004 in order to facilitate
matching kidney donations. Other similar programs to facilitate kidney exchange have since
been created. There are a lot of additional logistical complications that come up in this
setting of course. For example, since donors have the right to back out of the donation at
any time, the surgeries all have to take place simultaneously. A method for finding matches
is based on the top trading cycles algorithm. Each potential recipient has a ranking of the
possible kidneys they could receive and the have a preference ordering over them (closer
genetics to the donor make the match more likely to succeed). Simultaneously each kidney
is “pledged” to an individual. Essentially each potential recipient points at their first choice
kidney, and each kidney points at the person it is pledged to. If we find a cycle, we remove
those individuals and execute the exchange. Of course many issues come up, including
merging the live donations with the waitlist from cadavers. See the lecture notes by Tayfun
Sömnez posted on Chalk for additional details.
Until recently the most common and famous application of market design was to the
matching of medical residents to hospitals. Residents have preferences for which hospital to
work at (hospitals differ in terms of quality but also in areas of expertise and in terms of the
location, so residents have heterogenous preferences), and since residents have specialized
in different areas in their schooling, hospitals have heterogenous preferences over residents.
The difficulty in the market for medical residents was that each hospital would attempt to
hire many new graduates at once, over a short period of time. Since they didn’t know who
would accept, and if offers were left open for too long the hospital would lose their next
choice, hospitals tended to make exploding offers forcing the residents to make a decision
before they knew which other hospitals would make them an offer. As it was understood
by the interested parties—namely medical schools and hospitals—that the market was not
working well, medical schools and hospitals agreed to create a clearinghouse 1952. This
clearinghouse was set up to match all residents and hospitals at once through a variation of
the Gale-Shapley algorithm with residents proposing to hospitals. Today this is called the
National Resident Matching Program (NRMP). Over time this program has been redesigned
to accommodate other needs (a large increase in the number of couples who graduate at the
same time and want to be sent to the same location) and to improve the way the mechanism
works in this specific market. One difficulty relative to the school choice example however is
that hospitals have preferences over residents, and we know that the Gale-Shapley algorithm
is only strategy proof on one side, so hospitals could potentially benefit from mis-reporting
their preferences. However more recent research has shown that the probability of benefitting
from misreporting on the “woman” side of the market becomes limited as the market becomes
large, so this may not be that great a concern. See Roth (2008) for a discussion.
References
Abdulkadiroğlu, A., N. Agarwal, and P. A. Pathak (2015): “The Welfare Effects
of Coordinated Assignment: Evidence from the NYC HS Match,” unpublished.
52
Abdulkadiroğlu, A. and T. Sönmez (2003): “School Choice: A Mechanism Design
Approach,” The American Economic Review, 93, 729–747.
Alesina, A. (1988): “Credibility and Policy Convergence in a Two-Party System with
Rational Voters,” American Economic Review, 78, 796–805.
Austen-Smith, D. and J. Banks (1988): “Elections, Coalitions, and Legislative Outcomes,” American Political Science Review, 82, 405–422.
Baron, D. P. and R. B. Myerson (1982): “Regulating a Monopolist with Unknown
Costs,” Econometrica, 911–930.
Besley, T. and S. Coate (1997): “An Economic Model of Representative Democracy,”
Quarterly Journal of Economics, 112, 85–114.
Dasgupta, P. and E. Maskin (2008): “On the Robustness of Majority Rule,” Journal
of the European Economic Association, 6, 949–973.
Downs, A. (1957): An Economic Theory of Democracy, New York: Harper and Row.
Gale, D. and L. S. Shapley (1962): “College Admissions and the Stability of Marriage,”
The American Mathematical Monthly, 69, 9–15.
Geanakoplos, J. (2005): “Three Brief Proofs of Arrow’s Impossibility Theorem,” Economic Theory, 26, 211–215.
Gura, E.-Y. and M. Maschler (2009): Insights into Game Theory, Cabridge Univ.
Press, Cambridge.
Lee, D. S., E. Moretti, and M. J. Butler (2004): “Do Voters Affect or Elect Policies?
Evidence from the US House,” Quarterly Journal of Economics, 807–859.
McCarty, N., K. Poole, and H. Rosenthal (2008): Polarized America: The Dance
of Ideology and Unequal Riches, Cambridge: MIT Press.
Myerson, R. B. (1979): “Incentive Compatibility and the Bargaining Problem,” Econometrica, 61–73.
——— (1981): “Optimal Auction Design,” Mathematics of Operations Research, 6, 58–73.
Myerson, R. B. and M. A. Satterthwaite (1983): “Efficient Mechanisms for Bilateral
Trading,” Journal of Economic Theory, 29, 265–281.
Osborne, M. J. and A. Slivinski (1996): “A Model of Political Competition with
Citizen-Candidates,” Quarterly Journal of Economics, 111, 65–96.
Roemer, J. E. (2009): Political Competition: Theory and Applications, Harvard University
Press.
Roth, A. E. (1982): “The Economics of Matching: Stability and Incentives,” Mathematics
of Operations Research, 7, 617–628.
53
——— (2008): “What Have We Learned from Market Design?” The Economic Journal,
118, 285–310.
Snyder, J. and D. Stromberg (2010): “Press Coverage and Politcal Accountability,”
Journal of Political Economy, 118, 355–408.
Spenkuch, J. (2015): “Please Don’t Vote For Me: Voting in a Natural Experiment with
Perverse Incentives,” Economic Journal, 125, 1025–1052.
Taylor, A. D. and A. M. Pacelli (2008): Mathematics and Politics: Strategy, Voting,
Power, and Proof, Springer.
54