Game Theory - Bradley University

Game Theory
OLLI
Fall, 2015
Day One
Game Theory:
• Not just about games
• About strategies used in daily life
• Concerns cooperation rather than confrontation
• Collaboration rather than competition
• Goes back thousands of years – Talmud and Sun Tzu’s writings
• Modern theory credited to John von Neuman and Oskar
Morgenstern
• John Nash generalized the results and provided a basis for the
modern field
• Game theory builds mathematical models and draws conclusions
from those models.
Rock-Paper-Scissors
Paper covers rock – so paper wins.
Rock breaks scissors – so rock wins.
Scissors cuts paper – so scissors wins.
Rock-Paper
Paper covers rock – so paper wins.
No winner
Ro
C
Paper
Sc
i
o
c
ck
ss
or
Paper covers rock - C wins
s
k
Rock breaks scis sors - R wins
R
Paper covers rock - R wins
ck
o
R
Paper
R
C
Paper
Sc
i
S
c
i
ss
o
No winner
rs
Scissors cuts paper - C wins
s
s
o
r
Rock breaks scis sors - C wins
ck
o
R
s
Paper
C
Sc
i
ss
o
Scissors cuts paper - R wins
rs
No winner
No winner
Ro
C
Paper
Sc
i
o
c
ck
ss
or
Paper covers rock - C wins
s
k
Rock breaks scis sors - R wins
R
Paper covers rock - R wins
ck
o
R
Paper
R
C
Paper
Sc
i
S
c
i
ss
o
No winner
rs
Scissors cuts paper - C wins
s
s
o
r
Rock breaks scis sors - C wins
ck
o
R
s
Paper
C
Sc
i
ss
o
Scissors cuts paper - R wins
rs
No winner
No winner
Ro
C
Paper
Sc
i
o
c
ck
ss
or
Paper covers rock - C wins
s
k
Rock breaks scis sors - R wins
R
Paper covers rock - R wins
ck
o
R
Paper
R
C
Paper
Sc
i
S
c
i
ss
o
No winner
rs
Scissors cuts paper - C wins
s
s
o
r
Rock breaks scis sors - C wins
ck
o
R
s
Paper
C
Sc
i
ss
o
Scissors cuts paper - R wins
rs
No winner
Rock, paper, scissors with scores in terms of (Row, Column):
Column
R
P
S
R
(0, 0)
(-1, 1)
(1, -1)
P
(1, -1)
(0, 0)
(-1, 1)
S
(-1, 1)
(1, -1)
(0, 0)
Row
Rock, paper, scissors (scores in terms of Row):
Column
Row
Political Decision Time
Two candidates for political office must decide to
be for, against, or neutral on a certain
referendum.
Political Decision Time
Pollsters have determined that if candidate R
comes out for the referendum, then he will gain
8000 votes if candidate C also comes out for the
referendum, will lose 1000 votes if candidate C
comes out against, and will gain 1000 votes if
candidate C comes out neutral.
Political Decision Time
If candidate R comes out against, then he will,
respectively, lose 7000 votes, gain 4000 votes,
lose 2000 votes if candidate C comes out,
respectively, for, against, neutral on the
referendum. If candidate R is neutral, then he will
gain 3000 votes if C is for or against and will gain
2000 votes if C is neutral.
Political Decision Time
Pollsters have determined that if candidate R comes out for the
referendum, then he will gain 8000 votes if candidate C also
comes out for the referendum, will lose 1000 votes if candidate
C comes out against, and will gain 1000 votes if candidate C
comes out neutral. If candidate R comes out against, then he
will, respectively, lose 7000 votes, gain 4000 votes, lose 2000
votes if candidate C comes out, respectively, for, against,
neutral on the referendum. If candidate R is neutral, then he will
gain 3000 votes if C is for or against and will gain 2000 votes if C
is neutral.
R gains 8000
Fo
C
Agains t
Ne
o
r
ut
ra
l
R gains 1000
r
F
R gains 3000
Fo
R
Agains t
C
r
Agains t
Ne
N
e
u
C gains 1000
R gains 3000
ut
ra
l
R gains 2000
t
r
a
C gains 7000
l
Fo
r
Agains t
C
Ne
R gains 4000
ut
ra
l
C gains 2000
Vote gain or loss in terms of R:
C
For
Agains t
Neutral
+ 8000
- 1000
+ 1000
Agains t
+ 3000
+ 3000
+ 2000
Neutral
- 7000
+ 4000
- 2000
For
R
Vote gain or loss in terms of (Row, Column):
C
For
Agains t
Neutral
(+ 8000, - 8000)
(- 1000, + 1000)
(+ 1000, - 1000)
Agains t
(+ 3000, - 3000)
(+ 3000, - 3000)
(+ 2000, - 2000)
Neutral
(- 7000, + 7000)
(+ 4000, - 4000)
(- 2000, + 2000)
For
R
Vote gain or loss in terms of R:
C
R
For
Agains t
Neutral
+ 8000
- 1000
+ 1000
Agains t
+ 3000
+ 3000
+ 2000
Neutral
- 7000
+ 4000
- 2000
For
The best strategy is for R to be against and
C to be neutral.
Two television stations, WMBD and WHOI, each have a
game show and a situation comedy to schedule for their
1:00 and 2:00 afternoon time slots. If both schedule
game shows at 1:00, then WMBD will take $5,000 in
advertising revenue away from WHOI. If both schedule
the game show at 2:00, then WHOI will take $3,000 in
advertising revenue from WMBD. If they choose
different hours for their game show, then WMBD will
take $2,000 from WHOI by scheduling the show at 2:00
and will take $1,000 from WHOI if scheduled at 1:00.
What is the best strategy for the television stations?
Two television stations, WMBD and WHOI, each have a
game show and a situation comedy to schedule for their
1:00 and 2:00 afternoon time slots. If both schedule
game shows at 1:00, then WMBD will take $5,000 in
advertising revenue away from WHOI.
If both schedule the game show at 2:00, then WHOI will
take $3,000 in advertising revenue from WMBD. If they
choose different hours for their game show, then
WMBD will take $2,000 from WHOI by scheduling the
show at 2:00 and will take $1,000 from WHOI if
scheduled at 1:00. What is the best strategy for the
television stations?
Scheduling game shows; revenue in terms of WMBD:
The best strategy is for WMBD to schedule the
game show at 1:00 and WHOI to schedule their
game show at 2:00.
Two merchants are planning on building competing stores to serve a
region containing three small cities. The fraction of the total
population that live in each city is shown in the figure below. (Sam’s
and Costco??)
City 1
30%
City 3
25%
City 2
45%
Two merchants are planning on building competing stores to serve a
region containing three small cities. The fraction of the total
population that live in each city is shown in the figure below. If both
merchants locate in the same city, Sam’s will get 65% of the total
business in all three cities.
City 1
30%
City 3
25%
City 2
45%
Two merchants are planning on building competing stores to serve a
region containing three small cities. The fraction of the total
population that live in each city is shown in the figure below. If both
merchants locate in the same city, Sam’s will get 65% of the total
business. If the merchants locate in different cities, each will get 80%
of the business in the city it is in, and Sam’s will get 60% of the
business from the city not containing Costco.
City 1
30%
City 3
25%
City 2
45%
Percentage of business in terms of Sam’s.
Costco
City 1
City 1
City 2
City 3
65%
Sam’s
City 2
City 3
65%
65%
Percentage of business in terms of Sam’s.
Costco
City 1
Sam’s
City 2
City 3
City 1
City 2
65%
80% of 1
20% of 2
60% of 3
City 3
65%
65%
Percentage of business in terms of Sam’s.
Costco
City 1
City 2
City 3
City 1
65%
80% of 1
20% of 2
60% of 3
80% of 1
60% of 2
20% of 3
City 2
20% of 1
80% of 2
60% of 3
65%
60% of 1
80% of 2
20% of 3
City 3
20% of 1
60% of 2
80% of 3
60% of 1
20% of 2
80% of 3
65%
Sam’s
Adding in percentage of local population in the three cities:
Costco
City 1
City 2
City 3
City 1
30%
65%
80% of 1
20% of 2
60% of 3
80% of 1
60% of 2
20% of 3
City 2
45%
20% of 1
80% of 2
60% of 3
65%
60% of 1
80% of 2
20% of 3
City 3
25%
20% of 1
60% of 2
80% of 3
60% of 1
20% of 2
80% of 3
65%
Sam’s
Doing the math:
Costco
City 1
City 2
City 3
80% of 1
20% of 2
60% of 3
80% of 1
60% of 2
20% of 3
20% of 1
80% of 2
60% of 3
65%
60% of 1
80% of 2
20% of 3
20% of 1
60% of 2
80% of 3
60% of 1
20% of 2
80% of 3
65%
City 1
30%
65%
City 2
45%
City 3
25%
Sam’s
S in City 1 and C in City 2:
80%(30%)
20%(45%)
+ 60%(25%)
48%
Doing the math:
Costco
City 1
City 1
30%
65%
City 2
City 3
S in City 1 and C in
City 2:
48%
Sam’s
City 2
45%
City 3
25%
65%
65%
80%(30%)
20%(45%)
+ 60%(25%)
48%
Percentage of area business in terms of Sam’s:
Costco
City 1
City 2
City 3
City 1
30%
65%
48%
56%
City 2
45%
57%
65%
59%
City 3
25%
53%
47%
65%
Sam’s
Percentage of area business in terms of Sam’s – above or below 50%:
Costco
City 1
City 2
City 3
City 1
30%
+ 15%
- 2%
+ 6%
City 2
45%
+ 7%
+ 15%
+ 9%
City 3
25%
+ 3%
- 3%
+ 15%
Sam’s
Percentage of area business in terms of Sam’s – above or below 50%:
Costco
Sam’s
City 1
City 2
City 3
City 1
30%
+ 15%
- 2%
+ 6%
City 2
45%
+ 7%
+ 15%
+ 9%
City 3
25%
+ 3%
- 3%
+ 15%
Doing the “math” to determine the chance that each merchant will choose a particular city:
City 1 S chooses 27% C chooses 39%
City 2 S chooses 64% C chooses 9%
City 3 S chooses 9% C chooses 52%
Day Two
Following up on three comments/questions:
Thomas Dewey – his political life: MOST interesting;
someone needs to do an OLLI class just on him!
Kroger: who did the research? Why the decision? Marketing
plan? Evidently they don’t want us to know – or I just
couldn’t find it.
Political decision time: really strictly determined? Wouldn’t it
be better to choose each vote 1/3 of the time? (More to
come.)
Game Theory is the logical analysis of situations of conflict and
cooperation.
The elements of a game:
1) Players: how many (there are at least two); how they are chosen
2) Possible actions: each player has a number of possible strategies.
3) Information available to players
4) Payoffs/consequences/outcomes
5) Players’ preferences over payoffs
Obstacles found in games:
1) Any real-world game is very complex.
2) Game theory deals with “rational” moves.
3) Game theory cannot give unique prescriptions for players or play just allows for analysis and strategies.
Strictly determined: game has saddle
point; optimal strategies for both
players
Not strictly determined: game has no
saddle point; mixed strategies for both
players
Political Decision Time
If candidate R comes out against, then he will, respectively,
lose 7000 votes, gain 4000 votes, lose 2000 votes if
candidate C comes out, respectively, for, against, neutral
on the referendum. If candidate R is neutral, then he will
gain 3000 votes if C is for or against and will gain 2000
votes if C is neutral.
Vote gain or loss in terms of R:
C
R
For
Agains t
Neutral
+ 8000
- 1000
+ 1000
Agains t
+ 3000
+ 3000
+ 2000
Neutral
- 7000
+ 4000
- 2000
For
The best strategy is for R to be against and
C to be neutral. (Strictly determined.)
Was suggested that each choice should be
given 1/3 probability.
The math:
1
8000  1000  1000  3000  3000  2000  7000  4000  2000
9
1
 11, 000   1, 222.22
9
This is the GAIN to Row (and the loss to Column).
How about if Row chooses each with 1/3
probability and Column always chooses
against?
The math:
1
 1000  3000  4000 
3
1
  6, 000   2, 000
3
One more time:
Row chooses each with 1/3 probability and
Column chooses against 2/3 and the others 1/6
each.
The math:
1 2
  1000  3000  4000 
3 3
1 1
   8000  3000  7000  1000  2000  2000 
3 6
2
1
1
7
1
  6, 000    5000   1333  277  1611
9
18
3
9
9
Vote gain or loss in terms of R:
C
For
Agains t
Neutral
+ 8000
- 1000
+ 1000
Agains t
+ 3000
+ 3000
+ 2000
Neutral
- 7000
+ 4000
- 2000
For
R
Both choose each option 1/3 of time: +$1222.22 (Row advantage)
Row chooses each option 1/3 of time, Column always is against: +$2000
(Row advantage)
Row chooses each option 1/3 of time, Column is against 2/3 of time: +$1611 1/9
(Row advantage)
Strictly determined: game has saddle
point; optimal strategies for both
players
Not strictly determined: game has no
saddle point; mixed strategies for both
players
Remember: “strictly determined” is the
game theory designation that this is
the best strategy for both involved. It
isn’t necessarily what they actually do.
You are the owner of several greeting-card
stores must decide in December about the
type of displays to emphasize for Mother’s
Day in May. You have three possible
choices:
• emphasize chocolates
• emphasize collectible gifts
• emphasize gifts that can be engraved
Your success is dependent of the state of the
economy in May (this decision is made in
December).
• If the economy is strong, you will do well
with the collectible gifts.
• In a weak economy, the chocolates will do
very well.
• In a mixed economy, the gifts that can be
engraved will do well.
You prepare a matrix for the possibilities, with the
numbers representing your profits in thousands of
dollars.
Weak Econom y
Mixed
Strong Economy
Chocolates
85
30
75
Collectibles
45
45
110
Engraved
60
95
85
What is your strategy?
If you’re an optimist…
Weak Econom y
Mixed
Strong Economy
Chocolates
85
30
75
Collectibles
45
45
110
Engraved
60
95
85
If you’re a pessimist…
Weak Econom y
Mixed
Strong Economy
Chocolates
85
30
75
Collectibles
45
45
110
Engraved
60
95
85
You hear on the news that leading experts believe
there’s a 50% chance of a weak economy in May, a
20% chance of a mixed economy, and a 30% chance of
a strong economy.
Weak Econom y
Mixed
Strong Economy
Chocolates
85
30
75
Collectibles
45
45
110
Engraved
60
95
85
Weak Econom y 50%
Mixed 20%
Strong Economy 30%
Chocolates
85
30
75
Collectibles
45
45
110
Engraved
60
95
85
Chocolates: 85(.50)  30(.20)  75(.30)  71
Collectibles: 45(.50)  45(.20)  110(.30)  64.5
Engraved:
60(.50)  95(.20)  85(.30)  74.5
A classic game from the annals of military
strategy is the Battle of the Bismarck Sea. In early
1943, US military intelligence determined that a
Japanese convoy would be moving from New
Britain to New Guinea. It was up to General
MacArthur's command to intercept the convoy
and do damage.
The Japanese commander had the choice of
sending his convoy either north or south of New
Britain; either route required three days to reach
its destination. Weather predictions were that:
► the northern route would be rainy with poor
visibility
► the southern route would be clear.
MacArthur's general in charge of the mission
could either concentrate most of his search
aircraft on the northern route or on the southern
route. It was up to this commander to find the
convoy and expose it to bombing by as many
aircraft as possible. The Japanese commander, of
course, wanted to limit the exposure to bombing.
The decisions of the two commanders were
completely independent of one another.
The following indicates the reasoning of the two commanders (what
happens during the three-day period:
North: rainy, poor visibility
Japanese convoy goes North
North: rainy,
poor visibility
South: clear
South: clear
Japanese convoy goes South
Weather interferes with the
search, but two days of
US Aircraft goes North bombing is possible because
of greater concentration of
aircraft.
The convoy is in clear
weather,
but most of search was to the
North, so only have time for
two days of bombing.
Poor visibility and loss of time
because of misdirected
US Aircraft goes South aircraft limits bombing time
to one day.
Convoy in clear weather and
search aircraft concentrated
in area provides three days
of bombing.
Japanese convoy goes North
US Aircraft goes North
US Aircraft goes South
Japanese convoy goes South
Weather interferes with the
search, but two days of
bombing is possible because
of greater concentration of
aircraft.
The convoy is in clear weather,
but most of search was to the
North, so only have time for two
days of bombing.
Poor visibility and loss of time
because of misdirected
aircraft limits bombing time
to one day.
Convoy in clear weather and
search aircraft concentrated
in area provides three days
of bombing.
The history books tell us that both commanders
chose the northern route.
A company is being sued (product liability).
Do they try to settle for $1.2 million, or do
they go to trial where they could possibly lose
as much as $4 million?
What do the strategists for the company
being sued know?
- $1.2 m illion
Settle
Go
to
trial
innocent
finds
Jury
Jury
finds
guilty
$0
- $1.2 m illion
Settle
Go
to
trial
innocent
finds
Jury
$0
- $4 m illion
Jury
finds
guilty
- $2 m illion
- $1 m illion
- $1.2 m illion
Settle
Go
to
trial
innocent
finds
Jury
25%
$0
- $4 m illion
Jury 75%
finds
guilty
- $2 m illion
- $1 m illion
- $1.2 m illion
Settle
Go
to
trial
innocent
finds
Jury
25%
$0
- $4 m illion
Jury 75%
finds
guilty
30%
40%
- $2 m illion
30%
- $1 m illion
- $1.2 m illion
Settle
Go
to
trial
innocent
finds
Jury
25%
$0
- $4 m illion
Jury 75%
finds
guilty
30%
40%
- $2 m illion
30%
- $1 m illion
What are the chances of the $4 million settlement?
75% X 30% = 22.5%
- $1.2 m illion
Settle
Go
to
trial
innocent
finds
Jury
25%
$0
- $4 m illion
Jury 75%
finds
guilty
30%
40%
- $2 m illion
30%
- $1 m illion
What are the chances of the $1 million settlement?
75% X 30% = 22.5%
- $1.2 m illion
Settle
Go
to
trial
innocent
finds
Jury
25%
$0
- $4 m illion
Jury 75%
finds
guilty
30%
40%
- $2 m illion
30%
- $1 m illion
Expected value if jury finds company guilty: - $2.3 million
Multiply by 75%: -$1.725 million
Strategic Choices
Airlines: sometimes people don’t show and the airline can
sell their ticket (get twice the revenue for the
ticket)
: should the airline overbook?
Price of one-way ticket from Peoria to Orlando: $130
Offer to “overbooked” person to take another flight: $100
Price of one-way ticket from Peoria to Orlando: $130
Offer to “overbooked” person to take another flight: $100
If the airline overbooks by one person, what can happen?
(Assume can sell the seat of a no-show.)
0 no-shows: have to pay overbooked person $100
(- $100 for seat)
1 no-show: can sell the seat again for $130
(+$130 for seat)
Price of one-way ticket from Peoria to Orlando: $130
Offer to “overbooked” person to take another flight: $100
If the airline overbooks by two people, what can happen?
(Assume can sell the seats of any no-shows.)
0 no-shows: have to pay overbooked persons $100 each
(- $200 for two seats)
1 no-show: can sell one seat again for $130; pay
overbooked person $100
(+$130 – $100 = +$30 net for two seats)
2 no-shows: can sell two seats again for $130 each
(+$260 net for two seats)
Price of ticket:
$130
Penalty for overbook:
-$100
Number of no-shows
Don't overbook
Overbook by 1
Overbook by 2
0
0
-100
-200
1
0
+130
+30
2
0
+130
+260
Experience has shown that:
The probability that there WON’T
be a no-show is 25%.
The probability that there will be 1
no-show is 40%.
The probability that there will be 2
no-shows is 35%.
Price of ticket:
$130
Penalty for overbook:
-$100
Number of no-shows
0: 25% 1: 40% 2: 35%
Don't overbook
0
0
0
Overbook by 1 -100 +130 +130
Overbook by 2 -200
+30 +260
Overbook by 1, Expect: 25%(-100) + 40%(+130) + 35%(+130) = $72.50
Overbook by 2, Expect: 25%(-200) + 40%(+30) + 35%(+260) = $53.00
The application of game theory to political science is
focused in the overlapping areas of fair division,
political economy, public choice, war bargaining,
positive political theory, and social choice theory. In
each of these areas, researchers have developed
game-theoretic models in which the players are often
voters, states, special interest groups, and politicians.
Unlike those in economics, the payoffs for games in
biology are often interpreted as corresponding to
fitness. In addition, the focus has been less on
equilibria that correspond to a notion of rationality and
more on ones that would be maintained by
evolutionary forces. The best known equilibrium in
biology is known as the evolutionarily stable
strategy (ESS), first introduced in 1973. Although its
initial motivation did not involve any of the mental
requirements of the Nash equilibrium, every ESS is a
Nash equilibrium.
In biology, game theory has been used as a model to understand
many different phenomena. It was first used to explain the
evolution (and stability) of the approximate 1:1 sex ratios:
suggested that the 1:1 sex ratios are a result of evolutionary
forces acting on individuals who could be seen as trying to
maximize their number of grandchildren.
Additionally, biologists have used evolutionary game theory and
the ESS to explain the emergence of animal communication. For
example, the mobbing behavior of many species, in which a large
number of prey animals attack a larger predator, seems to be an
example of spontaneous emergent organization. Biologists have
used the game of chicken to analyze fighting behavior and
territoriality.
Game theory has come to play an increasingly important role in
logic and in computer science. Computer scientists have used
games to model interactive computations.
The emergence of the internet has motivated the development of
algorithms for finding equilibria in games, markets,
computational auctions, peer-to-peer systems, and security and
information markets.
Game theory has been put to several uses in philosophy. In
ethics, authors have attempted to pursue the project of deriving
morality from self-interest. Since games like the prisoner’s dilemma
present an apparent conflict between morality and self-interest,
explaining why cooperation is required by self-interest is an important
component of this project.
Paradox: reasoning leading to a conclusion
that seems logically
wrong/unacceptable/contradictory
Dilemma: situation with difficult choices
between two or more alternatives (often
equally undesirable)
Monty Hall Problem
Let’s Make a Deal
Monty Hall Problem
Suppose you're on a game show, and you're given the
choice of three doors: Behind one door is a car; behind the
others, goats. You pick a door, say No. 1, and the host, who
knows what's behind the doors, opens another door, say
No. 3, which has a goat. He then says to you, "Do you want
to pick door No. 2?" Is it to your advantage to switch your
choice?
You pick door No. 1, and the host, who knows
what's behind the doors, opens another door, No. 3,
which has a goat. Do you stay with No. 1 or do you
switch?
You pick door No. 1, and the host, who knows
what's behind the doors, opens another door, No. 3,
which has a goat. Do you stay with No. 1 or do you
switch?
Day Three
You pick door No. 1, and the host, who knows
what's behind the doors, opens another door, No. 3,
which has a goat. Do you stay with No. 1 or do you
switch?
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
1
3
Door 3
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
1
3
Door 3
Host
shows
Car
Location
Door 1
1
3
1
2
1
2
1
3
Player
picks
Door 1
Host
shows
Door 2
1
1
3
1
Door 3
Door 2
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
1
2
Door 2
1
2
Door 3
1
1
3
1
Door 3
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
1
2
Door 2
1
2
Door 3
1
1
3
1
Door 3
Door 3
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
1
2
Door 2
1
2
Door 3
1
Door 3
1
3
1
Door 3
Door 2
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
Stay
1
2
1
Door 2 6
1
2
Door 3
1
1
6
Door 3
1
3
Door 2
1
3
1
3
1
Door 3
Switch
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
Stay
1
2
1
Door 2 6
1
2
Door 3
1
1
6
Door 3
1
3
Door 2
1
3
1
3
1
Door 3
Car
Switch
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
1
2
1
Door 2 6
1
2
Door 3
1
1
6
Door 3
1
3
Door 2
1
3
1
3
1
Door 3
Stay
Switch
Car
Goat
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
Switch
Goat
1
2
1
Door 2 6
Car
1
2
Door 3
1
6
Car
1
Door 3
1
3
Door 2
1
3
1
3
1
Door 3
Stay
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
Switch
1
2
1
Door 2 6
Car
Goat
1
2
Door 3
1
6
Car
Goat
1
Door 3
1
3
Door 2
1
3
1
3
1
Door 3
Stay
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
Switch
1
2
1
Door 2 6
Car
Goat
1
2
Door 3
1
6
Car
Goat
1
Door 3
1
3
Door 2
1
3
1
3
1
Door 3
Stay
Goat
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
Switch
1
2
1
Door 2 6
Car
Goat
1
2
Door 3
1
6
Car
Goat
Goat
Car
1
Door 3
1
3
Door 2
1
3
1
3
1
Door 3
Stay
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
Stay
Switch
1
2
1
Door 2 6
Car
Goat
1
2
Door 3
1
6
Car
Goat
Door 3
1
3
Goat
Car
Door 2
1
3
Goat
1
1
3
1
Door 3
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
Stay
Switch
1
2
1
Door 2 6
Car
Goat
1
2
Door 3
1
6
Car
Goat
Door 3
1
3
Goat
Car
Door 2
1
3
Goat
Car
1
1
3
1
Door 3
Car
Location
Door 1
1
3
Player
picks
Door 1
1
3
Door 2
Host
shows
Stay
Switch
1
2
1
Door 2 6
Car
Goat
1
2
Door 3
1
6
Car
Goat
Door 3
1
3
Goat
Car
Door 2
1
3
Goat
Car
1
1
3
1
Door 3
A similar paradox to the Monty Hall Problem
is Bertrand’s Box Paradox.
Bertrand's box paradox is a classic paradox of elementary
probability theory. It was first posed by Joseph Bertrand in 1889.
There are three boxes:
a box containing two gold coins,
a box containing two silver coins,
a box containing one gold coin and one silver coin.
GG
SS
GS
GG
SS
GS
One box is chosen at random, and a coin is withdrawn (without
peeking at the other coin).
The coin is gold.
What are the chances that the other coin is gold, too?
After choosing a box at random and withdrawing
one coin at random, if that happens to be a gold
coin, it may seem that the probability that the
remaining coin is gold is 1⁄2; in fact, the probability
is actually 2⁄3.
GG
SS
GS
One explanation:
The coin chosen is gold. That means it came from the left or right
box.
With a gold coin taken from the left or right box, there are 3 coins
left in the two boxes: 2 gold and 1 silver. The probability is 2/3
that the remaining coin in the box is gold.
GG
SS
GS
Have drawn one gold coin. Now, the probability that the second
coin drawn (from any box) will be gold:
Probability (draw gold if GG)
Probability (draw gold if GG)  Probability (draw gold if SS)  Probability (draw gold if GS)
1
1
2



3
3
1 0  1
2
2
Social Dilemmas
Social dilemmas:
• Prisoner’s Dilemma
• Tragedy of the Commons
• The Free Rider
• Chicken
• The Volunteer’s Dilemma
• The Battle of the Sexes
• Stag Hunt
Understanding of social dilemmas came when John Nash discovered that all of them arise from the same basic
logical trap. This is called the Nash Equilibrium.
It is a position in which both sides have selected a strategy and neither side can then independently change its
strategy without ending up in a less desirable position. It’s a point of balance in a social situation, from which
neither side can escape without loss.
The secret to resolving such situations is for the parties find some way of agreeing to coordinate their actions
and for all parties to stick to the agreement.
.
Prisoners’ Dilemma:
Two thieves, Bill and Fred, have been caught by the police, but the
prosecutor has only enough evidence to sentence them both to two years,
on a charge of carrying concealed weapons, rather than the maximum
penalty of ten years for burglary. As long as they both plead not-guilty,
they will both get two years.
Both plead not-guilty: both get two years
The prosecutor approaches Bill and points out that if Fred pleads guilty,
but Bill does not, then Fred will receive a reduced sentence of four years
for pleading guilty, but Bill will get the maximum sentence of ten years. So
Bill’s best bet, if he believes that Fred will plead guilty, is to also plead
guilty and receive four years, rather than ten.
Both plead guilty: both get four years
The prosecutor also adds that he will give Bill the deal that if he
pleads guilty and Fred doesn’t , then he can go free for turning
state’s evidence. Bill reasons that he can do better for himself if
he pleads guilty, no matter what Fred does. The trouble is that
the prosecutor has made the same deal with Fred, who has come
to the same conclusion. So they both plead guilty, and both get
four years. If they’d both kept quiet, then they would both have
received two years.
Fred
Guilty
Not Guilty
Guilty
Bill
Not Guilty
Both 4
years
Bill 0
Fred 10
Bill 10
Fred 0
Both 2
years
Diner's dilemma is an n-person Prisoner’s dilemma.
Several individuals go out to eat, and prior to ordering,
they agree to split the check equally between all of them.
Each individual must now choose whether to order the
expensive or inexpensive dish. It is presupposed that the
expensive dish is better than the cheaper.
Each individual reasons that the expense s/he adds to
their bill by ordering the more expensive item is very
small (if they’re the only one), and thus their personal
improved dining experience is worth the money.
However, having all reasoned thusly, they all end up
paying for the cost of the more expensive meal, which by
assumption, is worse for everyone than having ordered
and paid for the cheaper meal.
Tragedy of the Commons
Explained by parable of the herders:
A group of herders, each grazing his own animals on
common land, with one herder thinking about adding an
extra animal to his herd. An extra animal will yield a tidy
profit, and the overall grazing capacity of the land will
be only slightly diminished, so it seems perfectly
logical… The tragedy comes when all the other herders
think the same way. They all add extra animals, the land
becomes overgrazed, and soon there is no pasture left.
Tragedy of the Commons
Also explained by teaspoons:
Teaspoons are disappearing from the common coffee
area. The teaspoon users make decisions that their own
utility (benefit to themselves) is improved by removing a
teaspoon from the common coffee area for personal
use, while everyone else’s utility is reduced by only a
fraction per head, since there are so many teaspoons
available. As more and more teaspoon users come to
the same conclusion, the teaspoons in the common area
are soon depleted.
The Tragedy of the Commons
Another example
After the 2004 tsunami in Sri Lanka, funds had been
donated to help people who lived in the affected
areas to move out or to rebuild their houses. Some
people from outside these areas actually moved into
them so they could claim a share of the benefits.
Doing so, they each took a small slice designated for
those who lived there originally.
The Tragedy of the Commons
Another example
Vegetable farmers are restricted in how much water they are allowed to use as a
result of a drought in the area. If they cooperate with the restriction, they will
get a lower yield per acre. (For purposes of illustration, let this yield be 5 tons per
acre rather than the usual 10 tons per acre.) If a few cheat by using water freely,
they could still get 10 tons per acre. If most of them cheat, the reservoirs would
run low, and their yields could drop to 2 tons per acre. More severe restrictions
could also come into force, and individuals who cooperated with the new
restrictions might only then get 1 ton per acre. The outcome depends on how
most of the farmers see themselves: as members of a cooperative group, or as
competing individuals.
The Tragedy of the Commons
Other examples:
DVD piracy, overfishing, pollution and global warming,
benefit cheating, etc.
The Free Rider
• Examples: leaving a mess for others to clean up in a shared
accommodation, the choice between remaining seated or standing
to get up to get a better view, refusing to join a labor union but still
accepting the benefits won by negotiations of members, credit card
fraud, disarmament, etc.
• A new steeple on a church would cost $100,000; everyone is asked
to contribute $100. If a person decides not to contribute and just
reap the benefit of a new steeple, they could assign their benefit
worth $200.
Chicken
For example, in the movie Rebel Without a Cause, the
two characters race their cars toward an abyss. The
first one to jump out is the loser (and the chicken).
Ironically, in this movie, the loser lost even more
when his leather jacket got hung up on the car door
and he went over the cliff with the car.
Chicken
Another example could be the Cuban Missile Crisis –
which could have ended in disaster – when
Khrushchev refused to stop bringing (remove) Soviet
missiles to Cuba and President Kennedy refused to lift
the naval blockade.
Chicken
Two people walking along a sidewalk toward each
other have the choice of stepping to one side or not.
If they both step aside, then the outcome is “neutral”.
If neither steps aside, then it is a “bad” outcome. The
“good” outcome (for one or the other it’s good) is if
one steps aside and the other doesn’t.
Don't Step
Aside
Don't Step
Aside
Step Aside
Step Aside
The hawk-dove version of the game imagines two players
(animals) contesting an indivisible resource who can choose
between two strategies, one more escalated than the other.
They can use threat displays (play Dove), or physically attack
each other (play Hawk). If both players choose the Hawk
strategy, then they fight until one is injured and the other wins.
If only one player chooses Hawk, then this player defeats the
Dove player. If both players play Dove, there is a tie, and each
player receives a payoff lower than the profit of a hawk
defeating a dove.
The Volunteer’s Dilemma
Group situations in which the person making the
first move risks losing out – while the others
gain. But if no one volunteers, then the loss can
be disastrous.
The Volunteer’s Dilemma
Who should jump out of the life boat to keep it from sinking?
Who should take the blame for a group offense so not all are
punished?
Migrating wildebeest herds coming to a river with crocodiles.
A person throws himself on a grenade to save the others.
The Volunteer’s Dilemma
A science magazine conducted an experiment: they
invited readers to send a card requesting either $20
or $100. The offer was for everyone to receive what
they asked for – as long as no more than 20% of the
requests were for $100 – in which case no one would
get anything. (The magazine backed out of their offer
before the results came in, but they would have been
fine: 35% asked for $100.)
The Battle of the Sexes
A couple decide independently whether to go to a
movie or a ball game. Each person likes to do
something together with the other, but one of them
prefers movies and the other ball games. The payoffs
are in terms of satisfaction: being at the preferred
place gives satisfaction, and being with the other
person gives even more satisfaction.
Movie
Ball Game
Ball Game
Ball Game
Movie
Movie
The Stag Hunt
• The “inverted” Prisoner’s Dilemma; with the Prisoner’s Dilemma,
the reward to the individual is always greater for cheating/telling.
With the Stag Hunt, the reward is greater for cooperating/not
cheating.
• A group of villagers are hunting a deer. It’s agreed that, in order to
get the deer, everyone has to stay at their assigned place. The
temptation comes when a hare comes within reach of a villager. If
he goes for the hare (a sure thing), then the others will lose the
chance to get a deer (from the commotion).
Some Other Games
The Three Prisoners
Three prisoners, A, B and C, are in separate cells and sentenced to death. The governor has
selected one of them at random to be pardoned. The warden knows which one is pardoned,
but is not allowed to tell. Prisoner A begs the warden to let him know the identity of one of the
others who is going to be executed. "If B is to be pardoned, give me C's name. If C is to be
pardoned, give me B's name. And if I'm to be pardoned, flip a coin to decide whether to name
B or C."
The warden tells A that B is to be executed. Prisoner A is pleased because he believes that
his probability of surviving has gone up from 1/3 to 1/2, as it is now between him and C.
Prisoner A secretly tells C the news, who is also pleased, because he reasons that A still has
a chance of 1/3 to be the pardoned one, but his chance has gone up to 2/3. What is the
correct answer?
The answer is that prisoner A didn't gain information about his own fate. Prisoner A, prior to
hearing from the warden, estimates his chances of being pardoned as 1/3, the same as both
B and C. As the warden says B will be executed, it's either because C will be pardoned (1/3
chance), or A will be pardoned (1/3 chance) and the B/C coin the warden flipped came up B
(1/2 chance; for a total of a 1/6 chance B was named because A will be pardoned). Hence,
after hearing that B will be executed, the estimate of A's chance of being pardoned is half that
of C. This means his chances of being pardoned, now knowing B isn't, again are 1/3, but C
has a 2/3 chance of being pardoned.
The ultimatum game is a game in economic experiments. The first player (the proposer) receives a sum
of money and proposes how to divide the sum between himself and another player. The second player
(the responder) chooses to either accept or reject this proposal. If the second player accepts, the money
is split according to the proposal. If the second player rejects, neither player receives any money. The
game is typically played only once so that reciprocation is not an issue.
Two people, Alice and Bob, play the game. An experimenter puts 100 one dollar bills on a table in front of
them. Alice may divide the money between herself and Bob however she chooses. Bob then decides
whether to accept her division, in which case each keeps the money as Alice divided it, or to reject the
division, in which case neither receives any money.
For example, Alice divides the money into one stack worth 65 dollars and one worth 35 dollars. She offers
the smaller amount to Bob. If he accepts, he keeps 35 dollars and Alice keeps 65 dollars. If Bob rejects the
division, neither he nor Alice receive anything.
If Bob acts rationally according to Rational choice theory, he should accept any division in which Alice
offers him at least one dollar, since doing so leaves him with more money than he would have had
otherwise. Even a division which gives Alice 100 dollars and Bob zero costs Bob nothing, so he has no
purely rational reason to reject it. If Alice knows that Bob will act rationally, and if she acts rationally
herself, then she should offer Bob one dollar and keep 99 for herself. In practice, divisions which Bob
regards as unfair are generally rejected.
Blotto games (or Colonel Blotto games, or "Divide a Dollar" games) constitute a class of two-person
zero-sum games in which the players are tasked to simultaneously distribute limited resources over
several objects (or battlefields). In the classic version of the game, the player devoting the most
resources to a battlefield wins that battlefield, and the gain (or payoff) is then equal to the total
number of battlefields won.
The Colonel Blotto game was first proposed and solved by Emile Borel in 1921, as an example of a
game in which "the psychology of the players matters". It was studied after the Second World War by
scholars in Operation Research, and became a classic in Game Theory.
The game is named after the fictional Colonel Blotto from Gross and Wagner's 1950 paper. The
Colonel was tasked with finding the optimum distribution of his soldiers over N battlefields knowing
that:
1.on each battlefield the party that has allocated the most soldiers will win, but
2.both parties do not know how many soldiers the opposing party will allocate to each battlefield,
and:
3.both parties seek to maximize the number of battlefields they expect to win.
As an example Blotto game, consider the game in which two players each write down
three positive integers in non-decreasing order and such that they add up to a
pre-specified number S.
Subsequently, the two players show each other their writings, and compare
corresponding numbers. The player who has two numbers higher than the
corresponding ones of the opponent wins the game.
For S = 6 only three choices of numbers are possible: (2, 2, 2), (1, 2, 3) and (1, 1, 4).
It is easy to see that:
Any triplet against itself is a draw
(1, 1, 4) against (1, 2, 3) is a draw
(1, 2, 3) against (2, 2, 2) is a draw
(2, 2, 2) beats (1, 1, 4)
It follows that the optimum strategy is (2, 2, 2) as it does not do worse than breaking
even against any other strategy while beating one other strategy.
There are however several Nash equilibria. If both players choose the strategy
(2, 2, 2) or (1, 2, 3), then none of them can beat the other one by changing strategies,
so every such strategy pair is a Nash equilibrium.
Day Four
Matching Problems
Matching problems can include:
Matching men and women – when there are the same
number of each
Matching volunteers with jobs at a “work day”
Matching professional teams and “draft picks”
Matching medical schools and applicants
A stable matching system is not necessarily a
system under which everyone is satisfied. A
matching system is stable when no unmatched pair
will find it beneficial to deviate from the matching and
form their own match.
For any preference structure, there is at least one
stable matching system.
The method used: Gale-Shapley algorithm
Gale-Shapley Algorithm
First Stage: Every “chooser” turns to the first on their list and issues an
“invitation”. Everyone who receives more than one invitation selects their
favorite of the “choosers” and tells the others that they will never be a choice.
Second Stage: Every “chooser” who has been rejected selects the second on
their list and makes an invitation. Everyone who receives more than one
invitation (including invitations from the previous stage) chooses their favorite
and rejects the others.
Third Stage: Every “chooser” who has been rejected now goes to the next on
their list and is screened through the same process.
The procedure continues until no “chooser” is rejected – everyone has a
match.
Consider the four teams: Atlanta (Hawks), Bulls, Celtics, and Detroit (Pistons)
who are trying to draft: Julius Erving, Kareem Abdul-Jabbar, Larry Bird, and
Michael Jordan. The following charts show the preferences of the teams and
the players.
Atlanta
Bulls
Celtics
Detroit
Julius
Kareem
Larry
Michael
1
Julius
Julius
Kareem
Michael
1
Detroit
Bulls
Detroit
Celtics
2
Kareem
Michael
Julius
Kareem
2
Celtics
Detroit
Atlanta
Bulls
3
Larry
Larry
Larry
Larry
3
Atlanta
Atlanta
Bulls
Atlanta
4
Michael
Kareem
Michael
Julius
4
Bulls
Celtics
Celtics
Detroit
First Stage: Every “chooser” turns to the first on their list and issues an
“invitation”. Everyone who receives more than one invitation selects their
favorite of the “choosers” and tells the others that they will never be a
choice.
Julius
Kareem
Larry
Michael
First Stage: Every “chooser” turns to the first on their list and issues an
“invitation”. Everyone who receives more than one invitation selects their
favorite of the “choosers” and tells the others that they will never be a
choice.
Julius
Kareem
Atlanta
Bulls
Celtics
Larry
Michael
Detroit
First Stage: Every “chooser” turns to the first on their list and issues an
“invitation”. Everyone who receives more than one invitation selects their
favorite of the “choosers” and tells the others that they will never be a
choice.
Julius
Kareem
Atlanta
Bulls
Celtics
Larry
Michael
Detroit
Julius chooses Atlanta over the Bulls.
Julius
Kareem
Atlanta
Bulls
Celtics
Larry
Michael
Detroit
Second Stage: Every “chooser” who has been rejected selects the second on their list
and makes an invitation. Everyone who receives more than one invitation (including
invitations from the previous stage) chooses their favorite and rejects the others.
Julius
Kareem
Atlanta
Bulls
Celtics
Larry
Michael
Detroit
Second Stage: Every “chooser” who has been rejected selects the second on their list
and makes an invitation. Everyone who receives more than one invitation (including
invitations from the previous stage) chooses their favorite and rejects the others.
Julius
Kareem
Larry
Michael
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Second Stage: Every “chooser” who has been rejected selects the second on their list
and makes an invitation. Everyone who receives more than one invitation (including
invitations from the previous stage) chooses their favorite and rejects the others.
Julius
Kareem
Larry
Michael
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Michael chooses the Bulls over Detroit.
Julius
Kareem
Larry
Michael
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Detroit makes its second choice, since Michael chose the Bulls over that team.
Julius
Kareem
Larry
Michael
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Atlanta
Bulls
Celtics
Detroit
Detroit
Bulls
Kareem chooses Detroit over the Celtics.
Julius
Kareem
Larry
Michael
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Atlanta
Bulls
Celtics
Detroit
Detroit
Bulls
Now the Celtics make their second choice.
Julius
Kareem
Larry
Michael
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Atlanta
Bulls
Celtics
Detroit
Detroit
Bulls
Atlanta
Bulls
Celtics
Celtics
Detroit
Detroit
Bulls
Julius chooses the Celtics over Atlanta.
Julius
Kareem
Larry
Michael
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Atlanta
Bulls
Celtics
Detroit
Detroit
Bulls
Atlanta
Bulls
Celtics
Celtics
Detroit
Detroit
Bulls
So now Atlanta needs to make their second choice.
Julius
Kareem
Larry
Michael
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Atlanta
Bulls
Celtics
Detroit
Detroit
Bulls
Atlanta
Bulls
Celtics
Celtics
Detroit
Detroit
Bulls
Atlanta
Bulls
Celtics
Celtics
Detroit
Atlanta
Detroit
Bulls
Kareem chooses Detroit over Atlanta – stays with Detroit.
Julius
Kareem
Larry
Michael
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Atlanta
Bulls
Celtics
Detroit
Detroit
Bulls
Atlanta
Bulls
Celtics
Celtics
Detroit
Detroit
Bulls
Atlanta
Bulls
Celtics
Celtics
Detroit
Atlanta
Detroit
Bulls
So Atlanta makes their third choice, Larry.
Julius
Kareem
Larry
Atlanta
Bulls
Celtics
Detroit
Atlanta
Bulls
Celtics
Detroit
Bulls
Atlanta
Bulls
Celtics
Detroit
Detroit
Bulls
Atlanta
Bulls
Celtics
Celtics
Detroit
Detroit
Bulls
Atlanta
Bulls
Celtics
Celtics
Detroit
Atlanta
Detroit
Bulls
Celtics
Detroit
Atlanta
Michael
Bulls
The final matching:
Julius
Kareem
Celtics Detroit
Larry
Michael
Atlanta Bulls
A stable matching system is not necessarily a system under
which everyone is satisfied.
Julius
Kareem
Larry
Michael
(Celtics choice 2)
(Detroit choice 2)
(Atlanta choice 2)
(Bulls choice 2)
Celtics
Detroit
Atlanta
Bulls
(Julius choice 2)
(Kareem choice 2)
(Larry choice 3)
(Michael choice 2)
Medical School Admissions Problem
In many countries a large number of applicants seek
admission to a small number of medical schools. The
competition for this limited number of places poses
several problems. Because the number of
candidates is greater than the medical school
admissions quotas, many candidates apply to
several medical schools.
Medical School Admissions Problem
The number of applications submitted to the
admissions office of each school often exceeds its
quota. The offices must evaluate all applicants and
decide which ones to accept and which to reject.
They will then offer to some applicants if there are
vacancies (first acceptances don’t commit) and
refuse admission to other (less-qualified) even if
there are vacancies.
Medical School Admissions Problem
One cannot assume that all who are offered admission
won’t accept an offer from what they consider to be better
offers from other schools. The admission quota of a school
who gets rejections from first offers won’t meet their quota.
The medical schools, to fill their quotas, may offer
admission to a number of applicants greater than their
quota. But, if enough students don’t decline, the number of
acceptances will exceed the quota.
Proposed solution to medical school problem:
Create an independent “placement center” where data is
collected and assignments made in the following manner:
1) Each applicant submits a list of schools he is willing to
accept.
2) Each applicant ranks the schools in order of preference.
3) The placement center sends each medical school a list of all
who have applied to that school.
Proposed solution to medical school problem:
4) Each school announces its quota.
5) Each medical school submits to the placement center a list
of the applicants ranked in order of preference and lists those
applicants it will not accept, even if there are vacancies.
Proposed solution to medical school problem:
6) The placement center implements the Gale-Shapley algorithm by:
a. All applicants are referred to the medical school of their first choice
according to the new preference order. If the number of candidates
is greater than the quota, the applicants are selected according to
the school’s preference order until the quota is filled. Any applicants
left over are “rejected”.
b. Rejected applicants are then referred to their next-choice medical
school. They are added to the list from the previous stage. If this
creates a list at the medical school that exceeds the quota, then the
school’s preference order is used and those exceeding the quota
rejected.
c. The process is repeated until the quota at each school is filled. At
this point, all the medical schools admit everyone on their current
list.
Game Theory in the
Movies and on TV
Rebel Without a Cause: the two characters race their cars toward an
abyss. The first one to jump out is the loser (and the chicken).
Ironically, in this movie, the loser lost even more when his leather
jacket got hung up on the car door and he went over the cliff with the
car.
Princess Bride
The story centers on Buttercup, a former farm girl who has
been chosen as the princess bride to Prince Humperdinck of
Florian. Buttercup does not love him, she who still laments
the death of her one true love, Westley, five years ago.
Westley was a hired hand on the farm, his stock answer of
"as you wish" to any request she made of him which she
came to understand was his way of saying that he loved her.
But Westley went away to sea, only to be killed by the Dread
Pirate Roberts. One scene humorously highlights both
strategic manipulation of the rules of the game and the
unrealistic assumption of ”common knowledge”.
An item of information in a game is common
knowledge if all of the players know it and all
of the players know that all other players
know it and all other players know that all
other players know that all other players know
it, and so on. This is much more than simply
saying that something is known by all, but
also implies that the fact that it is known is
also known by all, etc.
Common Knowledge
Consider a simple example of two allied armies situated on
opposite hilltops waiting to attack their foe. Neither commander
will attack unless he is sure that the other will attack at exactly
the same time. The first commander sends a messenger to the
other hilltop with the message "I plan to attack in the morning."
The messenger's journey is perilous and he may die on the way
to delivering the message. If he gets to the other hilltop and
informs the other commander - can we be certain that both will
attack in the morning? Note that both commanders now know the
message, but the first cannot be sure that the second got the
message. Thus, common knowledge implies not only that both
know some piece of information, but can also be absolutely
confident that the rest no it, and that the rest know that we know
it, and so on.
In the Princess Bride
Our hero Westley, in the guise of the Dread Pirate
Roberts, confronts his foe-for-the-moment, the Sicilian,
Vizzini. Westley challenges him to a Battle of Wits. Two
glasses are placed on the table, each containing wine
and one purportedly containing poison. The challenge,
simply, is to select the glass that does not lead to
immediate death.
Roberts: “All right: where is the poison? The battle of wits has begun. It ends
when you decide and we both drink, and find out who is right and who is
dead.”
Vizzini: “But it's so simple. All I have to do is divine from what I know of you.
Are you the sort of man who would put the poison into his own goblet, or his
enemy's? Now, a clever man would put the poison into his own goblet,
because he would know that only a great fool would reach for what he was
given. I'm not a great fool, so I can clearly not choose the wine in front of you.
But you must have known I was not a great fool; you would have counted on
it, so I can clearly not choose the wine in front of me.”
Roberts: “You've made your decision then?”
Vizzini: “Not remotely. Because iocane comes from Australia, as everyone
knows. And Australia is entirely peopled with criminals. And criminals are used
to having people not trust them, as you are not trusted by me. So I can clearly
not choose the wine in front of you.”
Roberts: “Truly, you have a dizzying intellect.”
The scene, beyond providing some comic relief on the theme of common
knowledge, also has an important lesson on strategic moves; if the rules of the
game may be changed, then the game can be rigged to one player's advantage:
Vizzini: “Let's drink -- me from my glass, and you from yours.”
[allowing Roberts to drink first, he swallows his wine]
Roberts: “You guessed wrong.”
Vizzini (roaring with laughter): “You only think I guessed wrong -- that's what's so
funny! I switched glasses when your back was turned. You fool. You fell victim to
one of the classic blunders. The most famous is ‘Never get involved in a land war
in Asia.’ But only slightly less well known is this: ‘Never go in against a Sicilian
when death is on the line.’”
[He laughs and roars and cackles and whoops until he falls over dead.]
[Roberts begins to rescue Buttercup, the girl over whom this battle was staged in
the first place]
Buttercup: “To think -- all that time it was your cup that was
poisoned.”
Roberts: “They were both poisoned. I spent the last few years
building up an immunity to iocane powder.”
The movie contains several other scenes with game-theoretic themes,
including many on bluffing.
The Good, the Bad and the Ugly
In the final scene: a game set up so that it cannot be lost, and can we really
trust Clint Eastwood?
The final scene in this Clint Eastwood movie is an example of game
theory. Three men in a triangle -- each with a gun, a rock at the center
of the three. It is up to each man to evaluate his situation. All are
excellent shots. Who do they shoot?
Clint has supposedly put a message on a rock that holds the key to
everything, but do the other two trust Clint to have actually written the
correct answer? As the other two evaluate the situation, they realize
they can't trust Clint to have written the answer on the rock -- therefore
they can't shoot Clint who likely still has the answer. That means the
other two can only shoot each other, but only one will likely hit before
the other.
What they don't know is that Clint has given one an unloaded gun...
Clint can ignore this one. The one Clint has to worry about with the
loaded gun will try to kill the one with the unloaded gun. Neither will fire
at Clint. Clint will fire at the one with the loaded gun. As the camera
passes from one face to the other the audience is meant to figure out
what each would do.
The guy with the loaded gun shoots at the guy with the unloaded gun -Clint shoots the guy with the loaded gun. Game over. As with the
hangings in the movie, he has dangled Duco out as bait while Clint
takes the money.
The game is decided before it starts.
Clint sets up a situation where each evaluates their possible moves,
but in reality, Clint has already won the game. Its a brilliant example of
people making the best decisions based on the information available to
them...and somebody manipulating the information available to them.
Butch Cassidy and the Sundance Kid
The leaders of the Hole-in-the-Wall Gang, Butch and Sundance, are fond of robbing
trains with a particular fancy for a rail line run by Mr. E. H. Harriman. They learn that
Harriman, having had enough, hired a group to pursue the robbers. "He resents the
way you've been picking on him so he's outfitted a special train and hired special
employees," they are told by a friend who also suggests that the gesture is quite
flattering.
After eluding the gang, Butch recognizes how the outcome is Pareto dominated,
though using a different vocabulary:
"A set-up like that costs more than we ever took...That crazy Harriman. That's
bad business. How long do you think I'd stay in operation if every time I pulled
a job, it cost me money? If he'd just pay me what he's spending to make me
stop robbin' him, I'd stop robbin' him."
[Screaming out the door at E. H. Harriman.] "Probably inherited every penny
you got!"
"Those inherited guys - what the hell do they know?"
The Pareto principle (also known as the 80–20 rule, the law of the vital
few, and the principle of factor sparsity) states that, for many events,
roughly 80% of the effects come from 20% of the causes.
Essentially, Pareto first showed that approximately 80% of the land in
Italy was owned by 20% of the population; Pareto developed the
principle by observing that 20% of the peapods in his garden
contained 80% of the peas. And it is a common rule of thumb in
business; e.g., "80% of your sales come from 20% of your clients."
Ransom
Irrationality improves Mel Gibson's bargaining power as he turns the tables on
the kidnappers of his son.
Movie scene: "Instead, I am offering this money as a reward on your head."
Ransom
You only want to pay the ransom if you trust the kidnappers are trustworthy to spare the loved one.
A 2-person Kidnapping is a classic example of a finite sequential game with imperfect information.
Let’s assume the following:
1. You have seen the kidnapper’s face, or the kidnapper has other reasons to believe that murdering you
poses less risk of capture than releasing you.
2. You have the money to pay the kidnapper, and will be able to transfer payment to the kidnapper without
any issue.
3. Your public Image is important (known to both players), and/or you have some skeletons in the closet
you’d rather stay hidden (known generally by both players).
Don Eppes is an FBI agent of the all I need to know I learned on the streets ilk. His
brother, Charlie, is a brilliant, reclusive university professor brother. Charlie helps
solve crimes by relating the crime to seemingly complicated (though generally
nonsensical) mathematical theories. Two episodes of Numb3rs involved game
theoretic themes directly.
One of the episodes is "Assassin" (Season 2, Episode 5)
Charlie and Don are in pursuit of an assassin, attempting to calculate where the killer
would strike.
Charlie: “Hide-and-seek.”
Don: “What are you talking about, like the kids' version?”
Charlie: “A mathematical approach to it, yes. See, the assassin must hide in order to
accomplish his goal, we must seek and find the assassin before he achieves that
goal.”
Megan: “Ah, behavioral game theory, yeah, we studied this at Quantico.”
Charlie: “I doubt you studied it the way that Rubinstein, Tversky and Heller studied
two-person constant-sum hide-and-seek with unique mixed strategy equilibria.”
Megan: “No, not quite that way.”
Don: “Just bear with him.”
Simpsons
Rock-Paper-Scissors is often employed on The
Simpsons to settle disputes. In one game (episode
9F16), Bart thinks "Good ol' rock. Nuthin' beats that!"
which Lisa of course predicts, highlighting the
importance of unpredictability in mixed strategies.
In episode 7F21, Bart and two friends are deciding how to share among them a single copy of
Radioactive Man #1, a comic book of great value to all three:
Martin: How about this, guys? Bart can have it Mondays and Thursdays, Milhouse will get it
Tuesdays and Fridays, and yours truly will take it Wednesdays and Saturdays.
Bart: Perfect!
Milhouse: Wait a minute! What about Sundays?
Bart: [suspiciously] Yeah, what about Sundays?
Martin: Well, Sunday possession will be determined by a random number generator. I will take the
digits 1 through 3, Milhouse will have 4 through 6, and Bart will have 7 through 9.
Bart: Perfect!
Milhouse: Wait a minute! What about 0?
Bart: [suspiciously] Yeah, what about 0?
Milhouse: Yeah!
Martin: Well, in the unlikely event of a 0, possession will be determined by Rock Scissors Paper
competition, best 3 out of 5. How's that?
Bart: Oh, okay.
Milhouse: Yeah, all right.
Big Bang Theory
"The Dumpling Paradox" is the seventh episode of the
first season. When the gang orders Chinese, their order of
dumpling appetizer comes with four dumplings which is
non-divisible by the three of them.
"The Lizard-Spock Expansion" is the eighth episode of the second
season. This episode first aired on Monday, November 17, 2008.
Social Justice
Bankruptcy
Analysis of a Bankruptcy Problem from the Talmud
Many times one encounters a bankruptcy situation where there are claims
against a given estate, and the sum of the claims against the estate exceeds its
worth. In such situations, one would like to know what would be a “fair” way of
dividing the estate among the claimants.
What seems fair in one case may seem less so in another.
Analysis of a Bankruptcy Problem from the Talmud
The following is an interesting method of division that has its origin in the Babylonian Talmud. It
involves a man who married three women and promised them in their marriage contract the
sums of 100, 200 and 300 units of money to be given them upon his death. The man died, but
his estate amounted to less than 600 units. The recommendation in the Mishna (one portion of
the Talmud) was the following:
Estate Worth →
100
200
300
100
33 1/3
50
50
200
33 1/3
75
100
300
33 1/3
75
150
Claims ↓
For many years, this arrangement was not understood. Some
thought this division reflected special circumstances whose
description was neglected. No solid explanation was found
until quite recently. Two game theorists (Aumann and
Maschler) examined the rule. They found that one solution
concept, called the nucleolus, gave precisely the numbers in
the table. The problem was, though, that the nucleolus was not
discovered until 1969. Instead, a consistency axiom gives a
hint as to how this was created and shows how more problems
with creditors and various claims can be resolved.
Consider the case where a garment is worth 100 units
of money.
One claims that his share is 50 units.
The other claims that his share is 80 units.
The claimant asking for 50 units of money declares, in
effect, that he has no claim to the second 50 units and,
as far as he is concerned, the other claimant can have
them.
The claimant to 80 units declares that he has no claim to the
remaining 20 units and, as far as he is concerned, the first
claimant can have them.
Value of garment
The two claims
100
80
Says other can
have 20
Uncontested division
Gets 50
50
Says other can
have 50
Gets 20
That leaves 30 remaining (100 – 50 – 20 = 30)
Value of garment
The two claims
100
80
Says other can
have 20
Uncontested division
50
Says other can
have 50
Gets 50
Gets 20
That leaves 30 remaining (100 – 50 – 20 = 30)
Divide the remaining 30 equally: 15 and 15
Add to uncontested.
Value of garment
The two claims
100
80
Says other can
have 20
Uncontested division
Gets 50
50
Says other can
have 50
Gets 20
Divide the remaining 30 equally: 15 and 15
Add to uncontested.
50
20
+ 15
+ 15
65
35
Analysis of a Bankruptcy Problem from the Talmud
The three women were promised 100, 200 and 300 units of money. The recommendation in the
Mishna (one portion of the Talmud) was the following. The first column (100) is an equal
division. The third column (300) is in proportion to the promise. The middle column follows the
nucleolus.
Estate Worth →
100
200
300
100
33 1/3
50
50
200
33 1/3
75
100
300
33 1/3
75
150
Claims ↓
Aumann and Maschler show that there is in fact only one division that is
consistent. The solution can be described by the following seven-step algorithm:
1. Order the creditors from lowest to highest claims.
2. Divide the estate equally among all parties until the lowest creditor receives
one-half of the claim.
3. Divide the estate equally among all parties except the lowest creditor until the
next lowest creditor receives one-half of the claim.
4. Proceed until each creditor has reached one-half of the original claim.
5. Now, work in reverse. Start giving the highest-claim money from the estate
until the loss, the difference between the claim and the award, equals the loss
for the next highest creditor.
6. Then divide the estate equally among the highest creditors until the loss of the
highest creditors equals the loss of the next highest.
7. Continue until all money has been awarded.
Analysis of a Bankruptcy Problem from the Talmud
Estate Worth →
100
200
300
100
33 1/3
50
50
200
33 1/3
75
100
300
33 1/3
75
150
Claims ↓
Value of estate
200
The claims
100
Lowest gets
one-half of
claim
50
200
300
Divide the remaining between the other two: 200 – 50 = 150
50
Can stop. No more creditors
75
75
Social Justice - Voting Paradox
A certain amount of municipal budget is unspent, and the city council must decide how to
invest it. There are three options: investment in education, investment in security, investment
in health. On the council are representatives of three parties:
Left party:
3 members
Center party:
4 members
Right party:
5 members
The parties’ list of preferences (in order from top to bottom):
Center (4)
health
security
education
Left (3)
education
health
security
Right (5)
security
education
health
No issue has a majority if everyone votes for their favorite.
If there were votes in “pairs” - choose two of the three items to vote on:
security vs. education: vote is 9 to 3
(center and right put security over education)
health vs. security:
vote is 7 to 5
(center and left put health over security)
education vs. health:
vote is 8 to 4
(left and right put education over health)
Center (4)
health
security
education
Left (3)
education
health
security
There is a cyclic preference relation:
security
education
Right (5)
security
education
health
health
security
This paradox was first noted in 1785 by mathematician and philosopher Marquis de
Condorcet. A decision rule in the form of a social choice function determines the
preference order of the society with regard to all the alternatives under discussion.
The task is not just to choose the most-preferred from
among the alternatives, but also to determine the
preference order of the society with regard to all the
alternatives under discussion.
There is not always a decisive answer. There is no social
choice function that satisfies all the constraints and goals.
Cooperative Game
Buyer-Seller Exchange Game
Player A has a house he is willing to sell, and Player B is interested in
buying the house. A deal can be made only if both parties “profit” (the
selling price is higher than the lowest the seller will take and the selling
price is lower than the highest the buyer will pay).
Suppose a house is worth $100,000. The seller will not sell it for lower
than this amount but hopes to get a higher payment.
Also, suppose the worth of the house to the buyer is $150,000. He will
not pay more than this but hopes to pay less.
A deal can be made if the house is sold for $120,000. The seller will
“profit” $20,000, and the buyer will “profit” $30,000.
The Traveler’s Dilemma (Nash equilibrium)
An airline loses two suitcases belonging to two different travelers. Both
suitcases happen to be identical and contain identical items. An airline
manager tasked to settle the claims of both travelers explains that the airline is
liable for a maximum of $100 per suitcase (he is unable to find out directly the
price of the items), and in order to determine an honest appraised value of the
suitcases the manager separates both travelers so they can't confer, and asks
them to write down the amount of their value at no less than $2 and no larger
than $100. He also tells them that if both write down the same number, he will
treat that number as the true dollar value of both suitcases and reimburse both
travelers that amount. However, if one writes down a smaller number than the
other, this smaller number will be taken as the true dollar value, and both
travelers will receive that amount along with a bonus/plus: $2 extra will be paid
to the traveler who wrote down the lower value and a $2 deduction will be taken
from the person who wrote down the higher amount. The challenge is: what
strategy should both travelers follow to decide the value they should write
down?
Application of Dilemmas (Nash equilibrium)
Write down the amount of their value at no less than $2 and no larger than
$100.
If both write down the same number, he will treat that number as the true dollar
value of both suitcases and reimburse both travelers that amount.
If one writes down a smaller number than the other, this smaller number will be
taken as the true dollar value, and both travelers will receive that amount along
with a bonus/plus: $2 extra will be paid to the traveler who wrote down the
lower value and a $2 deduction will be taken from the person who wrote down
the higher amount.
One might expect a traveler's optimum choice to be $100; that is, the traveler
values the suitcase at the airline manager's maximum allowed price.
Remarkably, and, to many, counter-intuitively, the traveler's optimum choice is
in fact $2; that is, the traveler values the suitcase at the airline manager's
minimum allowed price.
List of Games