SOLUTION CONCEPT IN THE BASIC STRATEGIC GAME by John

SOLUTION CONCEPT IN THE BASIC STRATEGIC GAME
by
John Bryant*
Rice University
August, 2014
© 2014
This paper first considers the extension of maximization in simple choice to the basic
strategic game. Iterated elimination of strictly set and (standard) vector dominated strategies are
examined and compared in their roles as solution concepts, exploiting the Prisoner’s Dilemma.
Unfortunately, these dominance criteria typically eliminate only a fraction of the strategies.
Hence attention turns to the overwhelming favorite solution concept for strategic games, Nash
equilibrium, which typically eliminates all but a few strategies. Nash equilibrium is not an
extension of maximization like the dominance criteria, and the use of Nash equilibrium presents
three basic problems, whether Nash equilibrium is indeed self-enforcing in the relevant sense,
why players play Nash equilibrium strategies to begin with, and, if there are multiple Nash
equilibria, how players coordinate on one. The role of the self-enforcing property of Nash
equilibrium in enabling, but not generating, a solution is described, exploiting the Prisoner’s
Dilemma and Bryant’s “min rule” game as examples. Motivated by relations between the
dominance criteria and Nash equilibrium, conditions are provided under which the assumptions
of the basic strategic game might be augmented so that pure strategy Nash equilibrium outcomes
are implied. Finally it is suggested how such augmentations might be developed by expanding
upon well known assumptions in the literature, examples being common knowledge, deviation
avoidance, and, in the case of mixed strategy equilibrium, stability. Given the importance of
Nash equilibrium in game theory, and the importance of game theory for strategic analysis, this
may be a fruitful approach.
Keywords: Game Theory, Strategic Games, Solution Concept, Dominance, Nash Equilibrium
JEL codes: C72
*John Bryant, Dept. of Economics-MS 22, Rice University, P. O. Box 1892, Houston, Texas
77251-1892, USA, [email protected], phone 713-348-3341, fax 713-348-5278
(kindly clearly mark “for John Bryant”), total word count - 8217.
I. INTRODUCTION
Probably everyone who studies Game Theory at one point or another wonders “why
‘solution concept,’ what’s wrong with ‘solution?’” This paper revisits this question, and attempts
to clarify, or at least expound on, the matter. In doing so, full use is made of well known game
theory texts, as excellent sources for widely accepted content, and masters of clarity. Even in the
texts, however, ambiguity over solution versus solution concept can creep in, and perhaps
appropriately so. And this is not a matter of mere semantics, but a deep, substantive issue at the
very core of Game Theory.
Game Theory starts with a conundrum. As it models strategic interaction between
rational, maximizing players Game Theory immediately presents a fundamental conceptual
problem, how to extend maximization from simple choice to strategic interaction, as emphasized
by Von Neumann and Morgenstern themselves (c. 1944, Chapter I, 2. -2.2.4).
“Thus each participant [in a “social exchange economy”] attempts to maximize a function … of
which he does not control all the variables. This is certainly no maximum problem, but a
peculiar and disconcerting mixture of several conflicting maximum problems.
… the conceptual differences between the original ([Robinson] Crusoe’s) maximum
problem and the more complex problem [of a “social exchange economy”]. … we face here and
now a really conceptual -- and not merely technical -- difficulty. And it is this problem which the
theory of games of strategy is mainly devised to meet” (Von Neumann and Morgenstern, 1967
ed., pp. 11, 12).
This “conceptual difficulty” which game theory “is mainly devised to meet” is no small
matter. It may be one of the most fundamental aspects of social interaction, and perhaps of
decentralized systems generally. It also bears on the nature of knowledge and uncertainty in
situations of strategic interaction. More narrowly, it explains the use of “solution concepts” in
game theory. If maximization were directly applicable to strategic interaction then the solution
of the game would just be the maximizations for all the players.
The nature of this conceptual difficulty is best studied in a basic strategic game, which
serves to isolate strategic interaction and the particular problems it poses. This strategic game
assumes a finite number of players, that their respective strategy spaces, consisting of a finite
number of strategies, and their respective payoffs to the cross product of all the players
strategies, or outcomes, are known. These are the only payoffs, and, in particular, payoffs do not
depend upon the structure of the game itself, or anything other than the outcomes. These payoffs
can be in ordinal utility or in Von Neumann and Morgenstern cardinal utility. The players are
isolated, and they are inferring what this game means for their behavior, strategy choice, taking
into account that they all are so inferring, and are known to be rational payoff maximizing agents
with full information on the structure of the game. Then “All players choose simultaneously
and independently of each other a strategy from their strategy sets and the game is
over.” (Aliprantis and Chakrabarti, 2011, p. 53) It is a one shot game. To avoid indeterminacies
caused by indifference, that is by equal payoffs, the assumption is added here that every player’s
payoffs imply a strict order over the outcomes. Finally, it is assumed that the specifications of
2
these components comprise a complete description of the relevant environment. We are thinking
of people facing an environment, thinking about it as if conceptualizing a basic strategic game,
and playing a strategy.
These players are not facing a maximization problem because the payoffs depend upon
the cross product of the players’ strategies. Payoffs are jointly determined. Hence the call for an
extension of maximization to strategic interaction for these rational maximizing players. First set
and vector dominance, which are explicit extensions of maximization, are treated. Then, with
the limitations of these approaches in mind, the basis for the use of Nash equilibrium, the most
widely used and fundamental “solution concept ” of game theory, is addressed. Nash
equilibrium does not involve an extension of maximization per se, as do the dominance criteria,
but the exploitation of maximizing behavior is critical to it.
II. DOMINANCE SOLUTION CONCEPTS
Set dominance is the least heroic, and least “powerful,” extension of maximization to
strategic interaction. (Bryant, 1987) Indeed, one could argue that it is not an extension at all.
Consider two strategies, 1 and 2, of player A, say. If the minimum payoff of the set of player A’s
payoffs associated with strategy 1 is larger than the maximum payoff of the set of player A’s
payoffs associated with strategy 2, then strategy 1 set dominates strategy 2. No matter what
happens should he play strategy 1, and no matter what happens should he play strategy 2,
strategy 1 yields a higher payoff. Hence player A drops strategy 2 from consideration. Being
rational maximizers the players can infer the dropping of strategies by each other. Moreover
when strategies, and their associated payoffs, are deleted this can cause set dominant relations
among the truncated remaining strategies to emerge. When this rational processing is complete
there remain strategies none of which are set dominated by another, including possibly a single
strategy, for each player. Usually in the context of (standard) vector dominance, this rational
processing is often referred to as the iterated elimination of dominated strategies.
There are two additional aspects of set dominance worth noting here. In the first place,
set dominance exploits only ranking, so it only requires ordinal utility, not cardinal utility,
payoffs. This may be of significance as one may encounter circumstances in which just ranking
outcomes is problem enough. Indeed cardinal utility may in reality be a very strong assumption
for some contexts. Secondly, set dominance does not exploit the independence of the, isolated,
players’ choices of strategy. It would not influence the player’s behavior, as characterized by set
dominance, if he believed the other players could, or possibly could, make their decisions
contingent on his. This possible dependence also may characterize some circumstances one
encounters. This second point on independence is clarified by turning to (standard) vector
dominance, which does exploit independence.
(Standard) vector dominance is a somewhat more “powerful” extension of maximization
to strategic interaction in that it eliminates all of the strategies eliminated by set dominance and,
typically, more. Again consider two strategies, 1 and 2, of player A. If for every cross product of
strategies of the other players the payoff to strategy 1 is higher than the payoff for strategy 2 for
player A then strategy 1 vector dominates strategy 2. As with set dominance the rational
maximizing players then engage in the iterated elimination of dominated strategies. When this
3
rational processing is complete there remain strategies none of which are vector dominated by
another, including possibly a single strategy, for each player.
Vector dominance, as well, exploits only ranking of payoffs, so it also only requires
ordinal utility, not cardinal utility, payoffs. However, the ranking is done for each given cross
product of strategies of the other players separately, and thus exploits independence. For
example, the payoff from strategy 1 and one cross product of strategies of the other players is not
compared to the payoff from strategy 2 and a different cross product of strategies of the other
players, as might be called for if the other players could, or possibly could, make their decisions
contingent on player A’s decision. Thus set dominance implies vector dominance, but not the
reverse. The iteratively non-vector dominated strategies are a subset (not necessarily proper) of
the iteratively non-set dominated strategies. For example consider that famous, and insight rich,
game theoretic conundrum the Prisoner’s Dilemma, in standard matrix form (where 4 is the best
outcome).
FIGURE I
Prisoner’s Dilemma
The (2,2) outcome is associated with vector dominant strategy 1, and is a Nash equilibrium as
well, but it is Pareto dominated by the (3,3) outcome, hence the fame of the Prisoner’s Dilemma.
However, strategy 1 is not set dominant as 3 is preferred to 2, and none of the strategies are ruled
out by set dominance. This illustrates that dominance criteria and Nash equilibrium do not bear a
simple relation with welfare, and portrays the gulf between maximization in a simple choice
setting and strategic interaction. It also illustrates that set and vector dominance are not above
challenge as extensions of maximization to strategic interaction, set dominance for admitting
unreasonable behavior, and vector dominance for eliminating reasonable behavior.
There is a further characterization of vector dominance that sheds a familiar light on the
independence condition, and on ordinal and cardinal payoffs as well. Let us first suppose that the
payoffs are ordinal and not cardinal, and that strategy 1 of player A vector dominates strategy 2
of player A, as in the Prisoner’s Dilemma. Now consider instead corresponding strategies 1* and
2* that are any strategies with Von Neumann and Morgenstern cardinal utility payoffs that
separately and together imply the same strict ranking of the outcomes involving strategies 1 and
2 as do strategies 1 and 2 themselves. Then strategy 1* vector dominates strategy 2*, and, for
4
any given probability distribution on the strategies of the other players, player A’s expected
utility of strategy 1* is greater than his expected utility of strategy 2*. Vector dominance may
not be beyond challenge, but it is not defenseless either.
III. THE SOLUTION CONCEPT AND NASH EQUILIBRIUM
The example of the Prisoner’s Dilemma not withstanding, vector dominance is a
demanding criterion to meet. There can remain many strategies and outcomes after the iterated
elimination of vector dominated strategies. This suggests that the assumptions of the basic
strategic game, supplemented with an extension of maximization, may be insufficient to
determine behavior, to logically imply an outcome, in many circumstances. Thus this forcefully
brings up the issue of the solution concept of the basic strategic game. Consider a broad
definition of solution concept advocated by Roger Myerson:
“In general, a solution concept is any rule for specifying predictions as to how players might be
expected to behave in any given game.” (Myerson, 1991, p. 107) “The simplest kind of solution
concept is one that specifies the set of strategies that each player might reasonably be expected to
use, without making any attempt to assess the probability of the various strategies .” (Myerson,
1991, p. 88)
The above should not be interpreted to mean that just any rule should do for “solution
concept,” as Myerson himself suggests. Indeed, one view is that what one should mean by
“solution concept,” or perhaps “solution,” is the logical implications for behavior of the
assumptions of the basic strategic game, as supplemented with an extension of maximization
presumably. If there are no further implications than those derived from the iterated elimination
of vector dominated strategies, then that solution concept is that the players do not play the
eliminated strategies and do play one of their remaining strategies. However, this strategy
choice, among the iteratively undominated strategies, presumably is caused by something, and
we have a violation of the complete description assumption. Of course, it could just be that the
basic strategic game does capture all the relevant strategic elements, but there are non-strategic
elements at work, and perhaps these non-strategic elements are hard to specify. Naturally one
would hope instead that somehow the solution is that one particular unique outcome is implied
by a strategic structure, or, barring that, at least fewer outcomes than those of iteratively
undominated strategies, and hopefully, probability distributions over players’ strategies as
Myerson suggests. There needs to be some “mechanism” beyond what we have seen thus far to
produce this result, some stronger solution concept. Ideally this would involve alterations or
supplements to the assumptions of the basic strategic game and the resulting logical implications
for behavior.
The prime candidate in the literature for solution concept for the basic strategic game is
the Nash equilibrium, which does not, of itself, take this ideal form. This mathematical object,
Nash equilibrium, is even more directly tied to maximization than are dominance criteria, and
sheds light on the nature of the concept of solution concept itself. The role of this solution
concept (and it is not a dynamic object, formally anyway) is a major subject of interest indeed in
5
game theory, consider: “Nash equilibrium is the solution concept used to solve simultaneousmove games;” (Dixit and Skeath, c. 2004, p. 113) “The Nash equilibrium will be the basic
concept throughout this book. ... Nash equilibrium (or simply an equilibrium or even a
solution) of a strategic game …” (Aliprantis and Chakrabarti, 2011, pp. 48, 54) “Nash’s (1951)
concept of equilibrium is probably the most important solution concept in game
theory.” (Myerson, 1991, p. 105) and “...the general n-person noncooperative game, presented in
finite strategic form. The fundamental solution concept for such games was formulated by John
Nash.” (Shubik, c. 1982, p. 240). The following discussion treats pure strategy Nash
equilibrium, leaving the consideration of mixed strategies to the end.
A pure strategy Nash equilibrium is an outcome such that, for every player, if the other
players play their Nash equilibrium strategies, then the player in question maximizes his payoff
by playing his Nash equilibrium strategy as well: this is often referred to as Nash equilibrium
being self-enforcing. A Nash equilibrium is an individual best reply strategy for all players. The
player contemplates deviating from the Nash equilibrium and observes that that reduces his
payoff. Thus Nash equilibrium itself is not an extension of maximization, but involves a direct
use of maximization by the players. Still it is in the spirit of vector dominance as a vector
dominating strategy has a higher payoff than the dominated strategy for every given play by the
other players, while the player’s Nash equilibrium strategy has the highest payoff for Nash
equilibrium play by the other players. Moreover, notice that Nash equilibrium does just depend
upon the structure of the game as specified by the basic strategic game, which would seem to
support its status as a solution concept for that game. It seems as though there are no alterations
or supplements to the assumptions of the basic strategic game involved. But how can this be
possible in a situation of strategic interaction? Were Von Neumann and Morgenstern wrong
about their conceptual difficulty after all? Or, alternatively, where is the sleight of hand, if you
will? If the players all expect the others to play their respective Nash equilibrium strategy they
will do so themselves, as rational maximizing agents. But how do they as isolated rational
maximizing agents come to that common expectation as THAT itself is not in general implied by
the structure of the game? At least in the first instance players are pursuing good outcomes not
common expectations. And we have already seen in the Prisoner’s Dilemma that common
expectations and Nash equilibrium need not have good welfare properties, should such properties
be an element in a solution concept. (This approach of Nash equilibrium, by the way, is an
example of a tried and true technique used in mathematics, namely if you do not know how to
solve a problem show how it is really a problem you do know how to solve.)
Nash equilibrium as a solution concept poses three basic problems. Is Nash equilibrium
indeed self-enforcing, in the appropriate sense, and what role does this play? Why do players
play a Nash equilibrium to begin with? Finally, if there are multiple Nash equilibria how do
players coordinate on the same one?
The above characterization of Nash Equilibrium as being self-enforcing is widely
observed and warrants scrutiny. It is worth noting starting off that this payoff reducing
“deviating” is not physically dynamic but a purported thought process by players in the strategic
game. Now suppose the players all imagine all imagining playing their strategy of a, and the
same, Nash equilibrium somehow. Next the individual player is imagining taking as given that
the others will be playing their Nash equilibrium strategies, and is thinking about the
6
consequences of deviating. But cannot he then imagine the others considering deviating, and
they the same, and then the possibility of a number of them doing so such that they are not
necessarily made worse off by deviating, this being considered? Such contemplation of others’
deviating does not violate independence. Common expectation may, then, be even more of a
sticky wicket, given the basic strategic game’s rational maximizing players inferring what the
game means for their behavior, strategy choice, taking into account that they all are so inferring,
and freely choosing. On the other hand, such implicit “communal” deviation may seem, to the
individual players, to be a bad bet relative to playing the Nash equilibrium strategy. More
basically, though, imagining being “at” a Nash equilibrium and contemplating deviating is not
actually being “at” a Nash equilibrium, which, after all, can only occur when the “players choose
simultaneously and independently of each other a strategy and the game is over.” (Aliprantis and
Chakrabarti, 2011, p. 53) Self- enforcing Nash equilibrium would not resolve the problem of the
source of common expectations, of why players play a Nash equilibrium to begin with, but rather
it enables the conceivable possibility of a resolution.
For a simple and powerful example, consider again the Prisoner’s Dilemma. Both
players deviating generates an outcome Pareto superior to the Nash equilibrium, but if they knew
the other would deviate they would play their Nash equilibrium strategy. But that does not imply
that they know what the other will do. The Prisoner’s Dilemma particularly clearly exhibits
potential benefits to deviation from a Nash equilibrium. While such deviation is ruled out by
vector dominance, it is not by set dominance. Do, then, people tacitly cooperate in this
circumstance? As to what experimental games reveal about behavior in this conundrum, the
experimental results are mixed. “There is typically a mixture of cooperation [both deviating] and
defection [playing the Nash equilibrium strategy], with the mix being somewhat sensitive to
[cardinal] payoffs and procedural factors, with some variation across groups.” (Holt c. 2007, p.
38). The Nash equilibrium in the Prisoner’s Dilemma not being self-enforcing provides one
possible explanation for these results, including there being mixed results.
There is another, or complementary, explanation for a non self-enforcing Nash
equilibrium in the Prisoners Dilemma. Perhaps people are “socialized” into playing the Pareto
dominating outcome in Prisoner’s Dilemma situations. Such “socialization” would seem a good
idea, in the context of the game anyway, if people can be so “socialized.” Perhaps this is
illuminating the distinction between “narrow self interest” and “enlightened self interest!”
Likely such “socialization” violates the assumptions of the basic strategic game that payoffs
determine behavior and that the payoffs themselves are just determined by the outcomes alone,
and are not influenced by the structure of the game. This source of violation of the assumptions
of the basic strategic game, “socialization” in the Prisoner’s Dilemma, may be a deep insight
into the functioning of societies (and religions?) provided by game theory. To explain the
experimental results it would have to be imperfect or non-uniform socialization. In any case, the
purported self-enforcing property of Nash equilibrium is an important matter bearing on the issue
of the Nash equilibrium as solution concept itself. For is not a non self-enforcing Nash
equilibrium not an equilibrium at all?
7
IV. FOUR PROPOSITIONS ON NASH EQUILIBRIUM AS SOLUTION CONCEPT
As a start on further considering the Nash equilibrium as solution concept, first its
relationship to set and vector dominance is addressed. We have already seen in the Prisoner’s
Dilemma that the outcome associated with the vector dominant strategies and the Nash
equilibrium outcome are the same outcome, which certainly is suggestive. The first proposition
generalizes this.
Proposition I: If after the iterated elimination of set or of vector dominated strategies only a
single outcome remains it is a pure strategy Nash equilibrium.
Proof: Suppose this unique iteratively undominated outcome is not a pure strategy Nash
equilibrium. Then for at least one player, given that the others play their iteratively undominated
strategy, the player is better off not playing his iteratively undominated strategy, but another
strategy. Consider the strategy of his which would have the highest payoff for him when the
others play their iteratively undominated strategies. Then that strategy is not dominated and
cannot have been eliminated. (See also Aliprantis and Chakrabarti, 2011, p. 64, and Dixit and
Skeath, c. 2004, p. 96)
This certainly seems supportive of the idea of Nash equilibrium as a solution concept,
taking set and vector dominance as bona fide extensions of maximization to strategic interaction.
But here iterated elimination of dominated strategies yields a common expectation. The isolated
rational maximizing agents do come to that common Nash equilibrium expectation, as here it is
implied by the structure of the game, which is known to them.
Proposition II: Every Nash equilibrium outcome of the original game remains as a Nash
equilibrium outcome after the iterated elimination of set or vector dominated strategies.
Proof: A Nash equilibrium has the property that for every player if the other players play
their Nash equilibrium strategies the player in question gets his highest payoff possible, given
their play, by playing his Nash equilibrium strategy. Then those Nash equilibrium strategies
cannot be dominated and cannot be eliminated. (Aliprantis and Chakrabarti (2011, p. 62) in
“Theorem 2.13 (The Iterated Elimination of Strictly Dominated Strategies)” provide a broader
result for vector dominance).
Notice that Proposition II implies that the Proposition I outcome is a unique Nash
equilibrium outcome in the original game. However, this is a necessary, but not sufficient,
condition for iterated dominance yielding a single outcome.
In terms of supporting Nash equilibrium as a solution concept Proposition II is a mixed
bag. On the one hand, by Proposition II the Nash equilibria are not eliminated by the dominance
criteria, which non-elimination should be true of a solution, taking set and vector dominance as
bona fide extensions of maximization to strategic interaction, once again, anyway. On the other
hand, if there are multiple Nash equilibria, and there well can be, then the dominance criteria do
not reduce this set. This adds a wrinkle to the basic issue of why the players would play a Nash
equilibrium to begin with. With multiple Nash equilibria how can isolated players coordinate on
one of them, and which one? To take a canonical case:
8
“To focus the analysis consider the following tacit coordination game, which is a strategic form
representation of John Bryant’s (1983) Keynesian coordination game.” (Van Huyck, Battalio and
Beil, 1990, p. 235) For further discussion see Cooper (1999) and Holt (c. 2007, Chapter 26
“Coordination Games,” pp. 325-337).
U(i) = A { minimum [e(1), ...,e(N)] } - B {e(i)},
i = 1, ...,N; A > B> 0; 0 ≤ e(i) ≤ E.
[Symmetric, Summary Statistic]
(Vincent Crawford first saw parallels between this game and J. J. Rousseau’s Stag Hunt parable,
but the original motivation was the work of J. M. Keynes.)
This specification is of a multiple player multiple strategy team game. Numerous
individuals’ efforts are complements. There are N>1 individuals, labeled as individual 1 through
N, and individual i is one particular one of those N individuals. Each individual i chooses an
“effort level” e(i) that is between 0 and E, E>0. The payoff to individual i is called U(i) and is
increasing in the minimum effort level chosen in the whole group of N individuals, and
decreasing, but less strongly, in the individual’s own effort. This strategic game has a whole
continuum of Pareto ranked strict Nash equilibria, where all individuals pick the same effort
level, at any level between 0 and E. Like the Prisoner’s Dilemma, this graphically illustrates that
common expectations and Nash equilibrium need not have good welfare properties. All but the
highest effort level, Pareto optimal, equilibrium are Pareto dominated by other equilibria. But
here every possible strategy choice of an individual is the strategy choice associated with one of
the Nash equilibria, the dominance criteria eliminate no strategies, and the Nash equilibria are of
linear measure zero in the cross product of strategy sets. In short it is an extreme coordination
problem. Indeed, in Van Huyck, Battalio and Beil’s (1990), replicated and robust, experiments,
with a finite strategy version of this game, a Nash equilibrium is never observed.
While yielding an extreme coordination problem, this case may not be special. To model
team production, where “moving together” beats “going it alone,” with constant returns to scale
and continuity in “effort,” then, by Euler’s theorem, the payoff function must be nondifferentiable, as with the “min rule.” This suggests that such coordination problems, while
extreme, might be prevalent with isolated players, particularly as specialization, and
interdependence generally, grows. For example, interpret the e(i) as component parts production.
Thus the no surplus condition (a condition of Walrasian equilibrium not met in team production)
is not a mere technicality. Indeed such coordination problems may play an important role in
macroeconomic and financial instability as well as in industrial organization, see Russell
Cooper’s (1999) seminal book Coordination Games: Complementarities and Macroeconomics.
At the same time, there is a refinement of the Nash equilibrium solution concept that is
applicable here even though these equilibria are strict. In particular, imperfect information,
global games, versions of this model provide interesting insights into the possible elimination of
the continuum of equilibria and the coordination problem. The non-differentiability of the “min
9
rule” is critical to the continuum of equilibria in this game, as it means that team production
obtains even for infinitesimal incremental increases in effort. In the very interesting global
games “min rule” paper of Carlsson and Ganslandt (1998) the realized effort levels are noisy,
which “smooths out” the kink so that differentiability obtains, and the continuum of equilibria is
eliminated. On the other hand, in the also very interesting global games “min rule” paper of
Asheim and Yoo (2008), the uncertainty is in individual specific payoff parameters, analogous to
A above, which maintains the “min rule” kink, and the continuum of equilibria and coordination
problem remains. Thus the choice of how to introduce uncertainty is not innocuous, and the
environment being modeled may have implications for it. Further, this refinement does not
address the basic issue of why players play a Nash equilibrium to begin with.
There is, in any case, a further possible interpretation of Van Huyck, Battalio and Beil’s
experimental results, stemming from non self-enforcing Nash equilibrium, that may be more
compelling. Indeed, it may provide a more compelling case of non self-enforcing Nash
equilibria than does the Prisoner’s Dilemma itself. The outcome in which all players choose the
maximum effort, E, is a strict Nash equilibrium, and the unique Pareto optimal outcome, and
hence naturally focal. (Schelling, 1980) With the focal property, it is as if players do “start at,”
mentally, that optimal Nash equilibrium. But that equilibrium is never observed in the
experiments. Notice that effort above the minimum effort level of the players is wasted, this is a
“weakest link” technology. Thus it seems that many players are “undercutting,” choosing a
lower effort level than E, out of their fear of someone else undercutting, out of his fear of
someone else undercutting, and so on. In other words, accepting the focal argument, in this
weakest link environment Nash equilibrium is not self-enforcing, and the outcome not
determinate as a consequence.
Further, behavior in their repeated experiments exhibited a downward cascade. In
their experimental form Van Huyck, Battalio and Beil used seven levels of “effort,” and payoffs
were in dollars. In the first experiment there were seven groups of 14-16 subjects playing ten
repeated games with the same group. In the first play almost 70% did not play the high, best for
all, if all do so, effort level, but only 2% played the lowest effort level. Only 10% predicted there
would be an equilibrium outcome. In subsequent repetitions, many of the players played below
the minimum effort level of the previous repetition, that is learning, or at least reaction to
previous play, was observed. Perhaps the preceding minimum effort level became focal for
some, but not self-enforcing due to the fear of undercutting. By the tenth period 72% of the
subjects played the minimum effort level, but a Nash equilibrium was never observed. Thus
while common expectation, or “common knowledge,” was not achieved in this interaction,
greater concentration in behavior did develop. It is worth noting that while this “min rule”
specification imposes more severe tests of self-enforcement of Nash equilibrium as the number
of players grows, qualitatively similar results can even be generated in 2x2 versions of the
experimental game, and “To summarize, it is not appropriate to assume that behavior will
somehow converge to the best equilibrium for all, even when it is a Nash equilibrium” (Holt, c.
2007, p. 41).
Given the downward cascade of effort levels particularly, such models do seem to have the
potential for contributing to our understanding of instability. At the same time, that the
Prisoner’s Dilemma and the “min rule” may provide examples of non self-enforcing Nash
10
equilibria does not rule out Nash equilibria typically being self-enforcing.
Generalizations of Proposition I provide further support to Nash Equilibrium as a solution
concept.
Proposition III: Suppose that the assumptions of the basic strategic game do somehow imply
some single outcome for every such game that has a pure strategy Nash equilibrium. Then that
outcome is a pure strategy Nash equilibrium.
Proof: The assumptions of the basic strategic game include that the players are rational
maximizers choosing freely among their strategies and with full information of the structure of
the game. If the solution were not a Nash equilibrium then at least one of the players could
increase his payoff, if the others did play their purported solution strategies, by deviating. We
cannot rule out such deviation by a rational maximizer choosing freely among his strategies and
with full information of the structure of the game. (Nor then can we in general rule out other
players deviating.)
Originally it may have been hoped that the assumptions of the basic strategic game would
generate a solution in the sense of predictions of particular unique outcomes. If so, pending
figuring out how it does so, Nash equilibrium provides all the possibilities. The approach that
behavior in simple choice settings is sufficient for determining social behavior has been a very
fruitful one in Economics, Walrasian equilibrium being the lead example. So such hope for the
basic strategic game would have been natural.
Proposition IV: Suppose the assumptions of a particular basic strategic game, that has a pure
strategy Nash equilibrium, can be augmented with assumptions on the environment that leave the
strategy spaces and payoffs unchanged, and the players rational maximizers choosing freely
among their strategies and with full information of the structure of the game. If this augmented
game implies a single outcome it is a pure strategy Nash equilibrium of the original game.
Proof: (See the proof of Proposition III)
Paraphrasing, Proposition IV says that, under the right conditions, if there is a solution, in
the sense of an implied single outcome, it is a Nash equilibrium.
Notice further that if the Nash equilibria in a game are not self-enforcing then under the
conditions of Propositions III and IV there can be no single outcome that is implied. For then, by
definition, deviation from a Nash equilibrium cannot be ruled out.
Proposition IV’s assumptions violate the complete description assumption, but they do
preserve the property that Nash equilibria provide all the possibilities for solution, given that
there is a solution concept as characterized in the Proposition. Of course this Proposition leaves
open what, and how situation specific, the additional assumptions are, if there exist such
assumptions. It is worth keeping in mind again that in a real world situation, of the sort being
modeled with strategic games, just one outcome happens, in a single play anyway, and,
presumably, something makes this so. The question is the nature of the something, namely
whether it meets the Proposition IV assumptions. Pending figuring out the additional
assumptions on the environment, Nash equilibrium provides the possible candidates for the
11
implied single outcome from additional elements of the environment meeting the conditions of
Proposition IV, if such exist.
Indeed, pending having a set of additional assumptions as characterized in Proposition IV
in hand, or, on the other side, pending a demonstration of their impossibility, Nash equilibrium
might well be thought of as a solution concept. Notice, however, that even if the original basic
strategic game has a unique pure strategy Nash equilibrium this uniqueness, of itself, does not
resolve the fundamental issue of why the players would play a Nash equilibrium. The additional
assumptions of Proposition IV are doing some heavy lifting, not only in determining which Nash
equilibrium, but also why equilibrium at all, and arguably should be a major component in the
modeling. Just using Nash equilibrium on a wing and a prayer is open to question if there is an
alternative. After all, the Proposition IV assumptions are purely hypothetical. Still, on the other
hand, the additional factors generating Nash equilibrium play may be very real, but hard to
specify.
Still, Proposition IV may provide the strongest case for Nash Equilibrium as a solution
concept for the basic strategic game, indeed possibly even an ideal solution concept involving
alterations or supplements to the assumptions of the basic strategic game and the resulting logical
implication for behavior being a unique outcome. Moreover, the basic strategic game does seem
to describe the strategic elements of at least some social interactions, else how would it be
possible to structure experiments on them, see e.g. Holt (2007)? Further, experiments may strip
away the Proposition IV additional environmental elements or features that generate a unique
outcome.
V. COMMON KNOWLEDGE
There are indeed approaches in the literature, implicitly somewhat in the vein of
Proposition IV, to solving the basic strategic game, as succinctly summarized by Aliprantis and
Chakrabarti (2011, pp. 48, 55) :
“What should be the notion of a ‘solution’ {for a matrix [two player strategic] game}? … Nash
equilibrium, a fundamental notion for game theory. … it is in the [unambiguous] interest of a
player to play a Nash equilibrium only if the player is quite certain that the others are going to
play the Nash equilibrium. Often it requires that each player know this, that every player know
that every player knows this, and so on ad infinitum. That is it must be common knowledge that
the players are going to play the Nash equilibrium. ... a Nash equilibrium strategy profile is
self-enforcing. Hence, if the players are searching for outcomes or solutions from which no
player will have an incentive to deviate, then the only strategy profiles that satisfy such a
requirement are the Nash equilibria.”
The necessity of common knowledge to justify Nash equilibrium is widely accepted as
Aliprantis and Chakrabarti point out. Common knowledge is assumed to imply that all the
players know how the others will play (Osborne, 2004, p. 21), which seems to go a long way
towards assuming away the heart of the problem. A critical question then is how the players
come to have common knowledge. Often common knowledge is justified by prior repeated
interaction among the players in the game. (Dixit and Skeath, c. 2004, p. 89, Osborne, 2004, p.
12
21) While common knowledge was not achieved in Van Huyck, Battalio and Beil’s famous
(1990) repeated experiments, for example, greater concentration in behavior did develop. This
leaves open the question of when and how repeated interaction, of what form, generates common
knowledge. For one thing, in some settings, players may have motive to try to give the others
misconceptions about their behavior prior to the game at hand. Also, prior to achieving common
knowledge, players presumably would infrequently encountered Nash equilibrium play. Further,
game theory based on the Nash equilibrium solution concept can only be justified in this manner
when such extensive player interaction has actually occurred. This approach, of course, violates
the complete description assumption, and the assumption of isolated players given that their prior
interaction affects behavior. Ideally, one might also have the explicit, rigorous modeling of the
interaction as a, perhaps the, critical component of a game theoretic model so motivated, that is a
dynamic model of learning perhaps. In any case, if common knowledge is achieved, the
outcome is a Nash equilibrium of the basic strategic game. Common knowledge of pure
strategies implies a single outcome, and by Proposition IV it is a pure strategy Nash equilibrium.
Just achieving common knowledge does not specify which Nash equilibrium if there are multiple
ones however. Also, whether this can be thought of as a solution to the basic strategic game is
another matter, given the violation of the isolated players assumption. Of course one might also
wonder, for example, whether such extensive and intensive interaction might not possibly result
in non-Nash tacit cooperation in the Prisoner’s Dilemma for example!
VI. AVOIDING DEVIATION
Another justification for Nash equilibrium as solution concept is the above Aliprantis and
Chakrabarti (2011, p. 55) “searching for outcomes or solutions from which no player will have
an incentive to deviate.” If referring to the strategic game itself, this “searching” must be a
metaphor for individual mental activity, done in isolation prior to play. Why do players do this,
particularly as a Nash equilibrium need not have good welfare properties? An immediate
reaction after all is that our players may eat bread, but they do not eat non-deviation. In the basic
strategic game what they are doing is trying to rationally maximize their payoff, and payoffs
depend only upon the outcomes. Yet a player supposedly is going to pick a strategy of some
Nash equilibrium on the chance that the others pick their strategy of that Nash equilibrium,
because no one would have an incentive to deviate from it if they knew the others would do it,
which they do not know, unless this argument implies that they do know, but it does not, for that
is circular reasoning: “if they all thought they all would do it, they all would do it, therefore they
all do it.” This deviation, too, is not a dynamic event, but a thought experiment. And a thought
experiment that does not answer the observation that if a player’s play did, in fact, turn out to be
a deviation from play of a particular Nash equilibrium by the other players, and if the player had
known what was going to happen he would have acted differently, but he did not know what was
going to happen. Also, how do the players “converge” (so to speak) on the same Nash
equilibrium outcome when there are multiple Nash equilibria? At least as described above, this
deviation avoidance mechanism might require further development to reach the high bar of
producing common knowledge.
Perhaps a sketch of a simple environment that does exhibit deviation avoidance of a sort
will help illustrate the challenges associated with modeling deviation avoidance. One of the
13
players in an, otherwise, basic strategic game has a technology, a magic wand, for making any
one of the outcomes focal (which violates the isolated players assumption). (Myerson, 1991, p.
111) That is to say, unless there is a reason for some player(s) individually and alone not to do
so, each player will assume that the others will play their strategy of the focal outcome, and
consequently do so themselves. Thus the magic wand insures that, if chosen, a Nash equilibrium
is self-enforcing. Then, if it is a good outcome for him, the player with the focal technology will
choose the Nash equilibrium in which his payoff is highest. Notice that this mechanism resolves
both which equilibrium and why equilibrium at all, if the focal arbitrator acts. Perhaps this
sketch is enough to suggest that further development is both necessary and worthwhile to
produce a rigorous Proposition IV solution concept in this deviation avoidance vein.
Admittedly, the focal approach for determining which equilibrium, and for resolving the
more basic question of why to play equilibrium strategies to begin with, does not require the
device of a focal arbitrator. In some games, like Bryant’s “min rule,” the very structure of the
game itself determines a focal outcome, but in many strategic interactions the focal could depend
upon aspects of the environment outside the strategic structure, as emphasized by Schelling
(1980). Still, elaborations of the focal arbitrator could provide interesting insights into the
purposeful generation of the focal.
VII. VON NEUMANN AND MORGENSTERN CARDINAL UTILITY AND THE MIXED
STRATEGY NASH EQUILIBRIUM SOLUTION CONCEPT
So far the issue of games without pure strategy Nash equilibria has been skirted, and
there most certainly are such games. For example, nine of the seventy-eight completely ordered
2x2 ordinal games do not have pure strategy Nash equilibria. The standard approach to these
games is to assume Von Neumann Morgenstern [vNM] cardinal utility, not ordinal utility, and
treat probability distributions over the pure strategies, that is treat mixed strategies. Cardinal
utility and probabilities together enable expected utility analysis, and the ranking of mixed
strategies thereby. To an extent, this works on games without pure strategy Nash equilibria,
consider Nash’s famous theorem:
“Proposition 119.1 (Existence of mixed strategy Nash equilibrium in finite games) Every
strategic game with vNM preferences in which each player has finitely actions has a [at least
one] mixed strategy Nash equilibrium.” (Osborne, 2004, p. 119)
In particular, “every” includes games without pure strategy Nash equilibria. A mixed
strategy equilibrium meets Myerson’s (1991, pp. 88) goal of enabling one to “assess the
probability of the various strategies,” in that equilibrium anyway.
VIII. STABILITY
Yet this mixed strategy approach to the problem of no pure strategy Nash equilibria only
works to an extent:
14
“… the [non-degenerate] mixed-strategy equilibria are Nash equilibria only in a weak sense [that
is, they are non-strict].
This property seems to undermine the basis for mixed-strategy Nash equilibria as the
solution concept for games. … Why not just do the simpler thing [than mixing] by choosing one
of his pure strategies [in the mixture]? After all, the expected payoff is the same. The answer is
that to do so would not be a Nash equilibrium; it would not be a stable outcome, because then the
other player would not choose to use his mixture.” (Dixit and Skeath, 2004, p. 244)
That the mixed strategy Nash equilibria are necessarily weak does seem to weaken the
case for them being a solution concept for the basic strategic game. Deviation from a nondegenerate mixed strategy equilibrium is not costly and cannot be ruled out in the static setting of
the basic strategic game, as the players are rational maximizers choosing freely among their
strategies and with full information of the structure of the game. In brief, these equilibria are not
self-enforcing in a static setting, and, hence, direct analogues to Propositions III and IV are not
available.
Dixit and Skeath suggest that players play mixed strategy Nash equilibrium strategies,
which gains them nothing in the immediate sense, because otherwise they would not be playing a
Nash equilibrium. For “then the other player would not choose to use his mixture” and it is “not
a stable outcome.” These suggested reasons seem to demand a dynamic game setting in which
current play affects future play. Ideally, then, one might have the explicit, rigorous modeling of
the setting of the dynamic game generating the equilibrium play as a critical component of the
mixed strategy Nash equilibrium model so motivated. It would also seem that such support of
mixed strategy Nash equilibrium as solution concept might be valid only in the context of such a
dynamic game setting.
IX. CONCLUSION
Iterated elimination of set and vector dominated strategies are derived from explicit
extensions of maximization in simple choice settings to strategic interaction, as suggested by Von
Neumann and Morgenstern (c. 1944, Chapter I, 2. -2.2.4). Thereby they provide the simplest
solution concepts for basic strategic games, but they often only imply modest restrictions on
strategy choice. Consequently the solution concept of choice for such games has been Nash
equilibrium, which typically further restricts behavior. Nash equilibrium not evolving from such
an extension of maximization to strategic interaction, there is a tradeoff in the use of Nash
equilibrium as a solution concept however. Specifically, it requires that the players have
common expectations without the structure of the game itself necessarily providing a mechanism
for this occurring. Consequently Nash equilibrium really seems to require further support for its
use as solution concept.
A promising approach to the support of the pure strategy Nash equilibrium as solution
concept is the possibility for supplements to the assumptions of the basic strategic game that
yield the prediction of a single outcome. If these new assumptions leave the strategy spaces and
payoffs unchanged, and do not impede the players being rational maximizers choosing freely and
with full information of the structure of the game, as in the basic strategic game itself, then that
predicted outcome is a pure strategy Nash equilibrium of the original game. Further, prominent
discussions in the literature supportive of pure strategy Nash equilibrium as solution concept, and
15
implicitly in this vein of new assumptions for the basic strategic game, are promising and
warrant further attention and development. In particular, these can suggest dynamic
environments that may be worth explicitly including as an integral component of the game
theoretic models themselves. Finally, mixed strategy Nash equilibria are weak, and thus not selfenforcing, and may consequently be of a more provisional character as a solution concept. Still,
mixed strategy Nash equilibrium may be supported in a dynamic game setting, which too might
fruitfully be explicitly included as an integral component of the models. This approach may
improve our understanding of environments that are conducive to equilibrium play, particularly
important given the extensive role for Nash equilibrium in game theory, and for game theory
itself, that game theorists advocate for them.
At the same time, it could just be, for example, that the basic strategic game does capture
all the relevant strategic elements, but there are non-strategic elements at work helping determine
the outcome, and these non-strategic elements are hard to identify and specify. If these nonstrategic elements also leave the strategy spaces and payoffs unchanged, and do not impede the
players being rational maximizers choosing freely and with full information of the structure of
the game, then pure strategy Nash equilibrium provides all the candidates for solution, in the
sense of predictions of particular unique outcomes. This should suffice for “solution concept”
status.
“To know game theory is to change your lifetime way of thinking. [Dixit and Skeath] is a
delightful skeleton key to the twenty-first century’s emerging culture.” (Paul A. Samuelson,
Dixit and Skeath, c. 2004, back cover)
RICE UNIVERSITY
16
Bibliography
Aliprantis, Charalambos D., and Subir K. Chakrabarti, Games and Decision Making, (New York,
NY: Oxford University Press, 2011).
Asheim, Geir B. and Seung Han Yoo, “Coordinating Under Incomplete Information,” Review of
Economic Design, 12 (2008), 293-313.
Bryant, John, “A Simple Rational Expectations Keynes-Type Model,” The Quarterly Journal of
Economics, 98 (1983), 525-528, reprinted in Coordination Failures and Real Rigidities, New
Keynesian Economics Vol. II, N. Mankiw and D. Romer, eds., MIT Press Readings in
Economics, B. Friedman and L. Summers, eds. (Cambridge, MA: M.I.T. Press, 1991).
_____, “The Paradox of Thrift, Liquidity Preference and Animal Spirits,” Econometrica, 55
(1987), 1231-1235.
Carlsson, Hans and Mattias Ganslandt, “Noisy Equilibrium Selection in Coordination Games,”
Economics Letters, 60 (1998), 23-34.
Cooper, Russell W., Coordination Games: Complementarities and Macroeconomics,
(Cambridge, UK: Cambridge University Press, 1999).
Dixit, Avinash and Susan Skeath, Games of Strategy, 2nd ed. (New York, NY: W. W. Norton, c.
2004).
Holt, Charles A., Markets, Games, & Strategic Behavior, (New York, NY: Pearson Addison
Wesley, c. 2007).
Myerson, Roger B., Game Theory: Analysis of Conflict, (Cambridge, MA: Harvard University
Press, 1991).
Nash, John F., “Noncooperative Games,” Annals of Mathematics, 54 (1951), 285-295.
Osborne, Martin J., An Introduction to Game Theory, (New York, NY: Oxford University Press,
2004).
Samuelson, Paul A., in Avinash Dixit and Susan Skeath, Games of Strategy, 2nd ed. (New York,
NY: W. W. Norton, c. 2004, back cover).
Schelling, Thomas C., The Strategy of Conflict, (Cambridge, MA: Harvard University Press,
1960, 1980).
17
Shubik, Martin, Game Theory and the Social Sciences: Concepts and Solutions, (Cambridge,
MA: The MIT Press, c. 1982).
Van Huyck, John B., Raymond C. Battalio and Richard O. Beil, “Tacit Coordination Games,
Strategic Uncertainty, and Coordination Failure,” American Economic Review, 80 (1990),
234-248.
Von Neumann, John and Oskar Morgenstern, Theory of Games and Economic Behavior, (New
York, NY: John Wiley & Sons, Science Edition 3rd printing, 1967, c. 1944).
18
FIGURE I
Prisoner’s Dilemma
19