Asynchronous Revision Games with Deadline: Unique Equilibrium

Asynchronous Revision Games with Deadline: Unique
Equilibrium in Coordination Games
Yuichiro Kamada and Takuo Sugaya
June 7, 2010
Abstract
Players prepare their actions before they play a normal-form game at a predetermined deadline. In
the preparation stage, each player stochstically obtains opportunities to revise their actions, and …nallyrevised action is played at the deadline. We show that, (i) a strictly Pareto-dominant Nash equilibrium,
if there exists, is the only equilibrium in the dynamic game; and (ii) in “battle of the sexes” games,
(ii-a) the equilibrium payo¤ set is a full-dimensional subset of the feasible payo¤ set under symmetric
structure, but (ii-b) a slight asymmetry can select a unique equilibrium.
Keywords: Revision games, …nite horizon, equilibrium selection, asynchronous moves
JEL codes: C72, C73
1
Introduction
Strategic interactions often require the preparation of the strategies in advance and even though players
interact once and for all, there can be a long period for the preparation. For example, consider two …rms
which have an investment opportunity. The date to make the actual investment is …xed. However, before the
actual investment, two …rms need to make the preparation: negotiating with banks to have enough liquidity,
allocating agents working for the project, and so forth.
It is natural to think that whether or not the investment is pro…table to one …rm depends on whether or
not the other …rm does the investment. Moreover, while the …rms make the preparation for the investments,
it is often the case that the …rms cannot always change their past preparation. That is, they have several
constraints for changing preparations, such as the administrative procedure to go through for changing
the past plan, obligations due to another project, and so forth. Rather, opportunities for revision of the
preparation may arrive stochastically over time.
This stochastic opportunity for the revisions allows players to credibly commit to the strategy. In the
above example, whether or not one …rm should invest depends on whether or not the opponent will invest.
If one …rm observes the opponent is making preparation for the investment, there is a probability that the
opponent will not have an opportunity to revise its decision and will invest, which makes the investment
more pro…table for the observant.
The present paper models such situations as follows: An n-player normal-form game is played once and
for all at a predetermined deadline, before which players have opportunities to revise their actions. These
opportunities arrive stochastically and independently among players. To model the preparation of actions,
we use the framework of a …nite horizon version of “revision games,” proposed by Kamada and Kandori
(2009). In revision games, players prepare their actions at opportunities in continuous time that arrive
Kamada: Department of Economics, Harvard University, Cambridge, MA, 02138, e-mail: [email protected]; Sugaya: Department of Economics, Princeton University, Princeton, NJ, 08544, e-mail: [email protected]; We thank Dilip
Abreu, Gabriel Carrol, Sylvain Chassang, Drew Fudenberg, Michihiro Kandori, Fuhito Kojima, Barton Lipman, Stephen Morris,
Satoru Takahashi, and seminar participants at Research in Behavior in Games seminar at Harvard for helpful comments.
1
with a Poisson process until a predetermined deadline. The actions that are prepared most recently at the
deadline are played once and for all. They consider the limit that the length of the preparation stage goes
to in…nity, so that there is enough number of opportunities by the deadline with a high probability. This is
equivalent to saying that they consider the limit that the arrival rate becomes very high for a …xed length of
preparation stage, that is, they consider the limit that players can revise their actions very frequently. This
assumption is also maintained in the present paper. We also assume that the Poisson arrival of the revision
opportunity is independent among players.
Kamada and Kandori (2009) show that, with certain conditions such as continuous strategy space, nonNash “cooperative”action pro…les can be played at the deadline. As they have shown, this result would not
be possible with a …nite game with a unique pure Nash equilibrium.1 Hence their focus is on expanding the
set of equilibria when the static Nash equilibrium is ine¢ cient relative to non-Nash pro…les. We ask a very
di¤erent question in this paper: we consider games with multiple e¢ cient static Nash equilibria, and ask
which of these equilibria is selected.2
We show that, even though the revision opportunity arrives with a very high probability, the small
probability of not being able to change the prepared action has surprisingly strong e¤ects on the equilibrium
selection: (i) if there exists a Nash equilibrium that strictly Pareto-dominates all the other action pro…les,
then it is the only equilibrium of the corresponding revision game; and (ii) in “battle of the sexes”games in
which Nash equilibria are not Pareto ranked, (ii-a) while with perfectly symmetric structure, the equilibrium
payo¤ set is a full-dimensional subset of the feasible payo¤ set, (ii-b) a slight asymmetry is enough to select
a unique equilibrium, which corresponds to the Nash equilibrium in the static game that gives the highest
payo¤ to the “strong”player. Actually, for (ii), it will be turned out that our result is valid for all the 2 x 2
games with two pure strategy Nash equilibrium. However, for the simple explanation, let us concentrate on
the battle of the sexes in the Introduction.
The rough intuition for our results is as follows: First, consider a game with strictly Pareto-dominant
Nash equilibrium. Firstly, we show that once all the players prepare the strictly Pareto-dominant Nash
equilibrium strategy, they will not escape from that state. Expecting that, if a unilateral change of the
preparation by player i can induce the strictly Pareto-dominant Nash equilibrium, she will do so. In turn,
player j whose unilateral change in the preparation can induce the situation where player i’s unilateral
change can induce the strictly Pareto-dominant Nash equilibrium, she will also do so if the deadline is far
expecting that player i will go to strictly Pareto dominant Nash equilibrium.3
Second, consider the “battle of the sexes”game with symmetric payo¤s and homogeneous arrival rate of
the revision opportunities. We prove that there are, in particular, three types of equilibria: The …rst one is
such that the process sticks to the Nash pro…le that player 1 prefers, from the beginning until the end of the
game. The second one is such that the process sticks to the Nash pro…le that player 2 prefers, again from
the beginning until the end of the game. The most important is the third one, in which the process starts at
an ine¢ cient non-Nash pro…le that would give a player the best payo¤ if the opponent switches to another
action. In this equilibrium, there is a “cuto¤” time such that if a player obtains an opportunity after this
cuto¤ while the opponent still sticks to the ine¢ cient action, the player switches her action. Once the action
pro…le reaches a Nash equilibrium, players will stick to that pro…le. Notice that, because of the symmetry,
two players’cuto¤s must be the same, and the indi¤erence condition at the cuto¤ implies that each player
expects their respective “worse” Nash equilibrium payo¤ in this equilibrium.4 We can then show that all
the equilibrium payo¤ set is such that at each point in the set, at least one player gets her “worse” payo¤.
1 The possibility of cooperation in …nite horizon in Kamada and Kandori (2009) is closely related to that of …nitely repeated
games with multiple Nash equilibria (Benoit and Krishna, 1985).
2 See also Ambrus and Lu (2009) for a variant of revision games model of bargainig in which the game ends when an o¤er is
accepted.
3 Similar results are obtained in the literature. See Farrell and Saloner (1985) and Dutta (2003) for early works on this topic.
Takahashi (2005) proves these results in a very general context.
4 The intution is similar to the one of the “war of attrition." The war of attrition is analyzed in Abreu and Gul (2000) and
Abreu and Pearce (2000). They consider the war of attrition problem in the context of bargaining. They show that if there
is an “irrational type" with a positive probability, then the agreement delays in equilibrium because rational players try to
imitate the irrational types. Players give in at the point where imitation is no longer pro…table. Although the war of attrition
is a common feature, the focus of their work and ours is quite di¤erent.
2
Hence, the equilibrium payo¤ set is a full-dimensional subset of the feasible payo¤ set.
Third, consider the “battle of the sexes”game with asymmetric payo¤s or asymmetric arrival rates of the
revision opportunities. We prove that the only equilibrium payo¤ is the one that corresponds to the Nash
equilibrium that the “strong” player (player 1) prefers, where by the “strong” player we mean the one who
expects more in his preferred Nash equilibrium than the opponent (player 2) does in the other pure Nash
equilibrium or the player with the lower arrival rate for the revision, i.e., the one who can more credibly
commit to the currently prepared action. The reason is simple. Consider the “chicken race equilibrium,”
the third type of the equilibrium discussed in the case of the symmetric “battle of the sexes”game. Because
we have asymmetry now, the cuto¤s of two players must di¤er. In particular, the strong player has to
have a cuto¤ closer to the deadline than that of the weak player. This implies that, in the chicken race
equilibrium, if it exists, the strong player expects strictly more than his “worse” Nash equilibrium payo¤,
while the weak player expects strictly less than her “worse” Nash equilibrium payo¤. Hence, the strong
player would not want to stick to the “worse”Nash equilibrium, which rules out the possibility of the second
type of equilibrium in the symmetric case. Also, the weak player would not want to stick to the chicken
race equilibrium, which rules out the third type of equilibrium in the symmetric case. Therefore the only
possibility is the …rst type of equilibrium. Thus a unique pro…le is selected.
We provide a detailed explanation of the intuition in Section 4.2 in a simpli…ed setting where the two Nash
equilibria are absorbing states. Note that, although the intuition looks simple, proving that the equilibrium
is unique is far less simple. This is because, no Nash action pro…le being a priori absorbing, hence when the
deadline if far away, players might have incentives to move away from the Nash pro…le. The above explanation
captures a part of this e¤ect because player 1 does not want to stick to his “worse” Nash equilibrium. We
will show that even player 2 does not want to stick to this pro…le (which he likes the best) when the deadline
if far away.
Finally, we note that the selection of equilibrium depends also on the heterogeneity of two players’arrival
rates. Naturally, a player with small arrival rate, i.e. the player with infrequent revision opportunities,
generally expects more from the game, with everything else symmetric. The reason is again simple. When
the arrival rate is small, the player can e¤ectively “commit” to his action for a while, so he can behave as a
“Stackelberg leader.”
Our results crucially hinges on asynchronicity of the revision process. If the revision process were synchronous, the very same indeterminancy among multiple strict Nash equilibria would be present. The result
that asynchronous moves select an equilibrium is not new, but the cases in which the asynchronous moves
result in equilibrium selection is very limited. Laguno¤ and Matsui (1997) show that in pure coordination
games (games in which all players’ payo¤ functions are the same) the Pareto e¢ cient outcome is chosen,
while Yoon (2001) shows that this result is nongeneric.5 Our results, although in a bit di¤erent context in
which the game is played only once at a deadline, show that the payo¤ structure can be anything in order
to obtain the best outcome as a unique equilibrium.6 ;7
Let us compare our work with the large literature on equilibrium selection. The key di¤erence between
our work and the past works in the literature is that, we consider a di¤erent class of situations than the ones
considered in the literature. Speci…cally, four features distinguish our model from the existing literature on
equilibrium selection. That is, we assume that (i) players are rational, (ii) they are nonanonymous, (iii) the
structure of the game is common knowledge, and (iv) the game is played once and for all. Let us overview
the related literature below. All of these four features seem to be present in the example mentioned in the
very …rst paragraph of this Introduction.
Kandori, Mailath, and Rob (1993) and Young (1993) consider an evolutionary learning model in which
5 Laguno¤ and Matsui (2001) argue that this nongenericity result curucially depends on the order of the limits with respect
to the discount factor and the purity of coordination.
6 For asynchronicity with respect to the timing of signals in imperfect monitoring setting, see Fudenberg and Olszewski
(2009).
7 Caruana and Einav (2008) consider a similar setting as ours, with …nite horizon and asynchronous moves, and show that
the equilibrium in generic 2 x 2 game there is a unique equilibrium, irrespective of the order and timing of moves and the
speci…cation of the protocol. They assume that players incur a cost whenever they revise their actions, which substantially
simplify their problem. In particular, on the equilibrium path, each player revises their actions at most once.
3
players interact repeatedly, and each player’s action at each period is stochastically perturbed. The key
di¤erence between their assumptions and ours is that in their model players are assumed to be boundedly
rational –players are assumed to play myopically in repeated interactions, following a rule-of-thumbs type
of decision rule, while we assume completely rational players. In addition, the game is repeated in…nitely in
their models, while the game is played once and for all in our model.
An evolutionary model called “perfect foresight dynamics” drops the bounded rationality assumption.
Matsui and Matsuyama (1994) and Oyama, Takahashi, and Hofbauer (2008) consider a model in which
patient players take into account the future path of the population dynamics. The key assumption to select
an equilibrium is that there is a population of agents who are randomly and anonymously matched over
time. In our model, on the other hand, …xed …nite players play a …nite horizon revision game.
Global games are another successful way to select an equilibrium. Rubinstein (1981), Carlsson and van
Damme (1993), Morris and Shin (1998), Sugaya and Takahashi (2009) show that non-existence of almost
common knowledge due to the incomplete information can select an equilibrium. Without almost common
knowledge, it is possible that a player thinks that his opponent thinks that the player thinks ... that the
game is very di¤erent from the one that the player believes, which is the key to rule out risk-dominated
equilibria. Thus, in particular, it is impossible to have a common knowledge that the game is in a small
open neighborhood of some a priori speci…ed one.
All of these three lines of the literature result in a selection of risk-dominant equilibrium of Harsanyi and
Selten (1988) in 2 x 2 games.8 In our model, however, a di¤erent answer is obtained: a Pareto-dominant
strict Nash equilibrium is played even if it is risk-dominated. Roughly speaking, since we assume complete
information with nonanonymous players, there is no “risk” of mis-coordination, which ensures that rational
players select a Pareto-dominant equilibrium.
Another line of the literature analyzes the e¤ect of the switching costs on the equilibrium selection.
Lipman and Wang (2000) consider a …nite horizon repeated games with small switching costs, in which the
per-period length (hence the per-period payo¤) is very small relative to the costs. They too use backward
induction argument to select a unique equilibrium in games that we consider in the present paper. However,
there are three important di¤erences. First, as the frequency of revision increases, it is potentially possible
in our model that players change their actions very often, while in Lipman and Wang (2000), the switching
cost essentially precludes this possibility. Second, the game is repeatedly played in their model, while it
is played only once at the deadline in our model. Hence, in particular, their model cannot have “chicken
race equilibrium” type of strategies in equilibrium. Third, in their model, the prediction of the game is not
robust to a¢ ne transformation of payo¤s.9 Hence, in some payo¤ speci…cations, their selection matches ours,
while in other cases it does not. The reason for this scale-dependence is that the switching cost is directly
payo¤-relevant. In our case, detailed speci…cations of the preparation stage (such as the arrival rate) is not
directly payo¤-relevant, so our result is robust to the a¢ ne transformation of the payo¤s.
Caruana and Einav (2008) consider a game where players play the actual game once and for all at the
…xed date and before playing the case, players commit to the actions. Though players can change their
commitment before the deadline, they have to pay the switching cost. Our paper is di¤erent from theirs in
two aspects. First, in our model, players can revise the preparation without paying the cost when they get
the opportunity. Second, Caruana and Einav consider the perfect information game, that is, at a …xed date,
only one player can change the past commitment. In their model, the order of the movement at each date
is predetermined. In our model, the opportunity of the movement is stochastic. With the symmetric arrival
rate, we can consider the situation that at each date, each player can move equally likely.
In the context of revision games with …nite strategy sets, Kamada and Sugaya (2010a) consider a model
8 As an exception, Young (1998) shows that in the context of contracting, his evolutionary model does not necessarily lead
to risk-dominant equilibrium (p-dominant equilibrium in Morris, Rob and Shin (1995)). But he considers a large anonymous
population of players and repeated interaction, so the context he focuses on is very di¤erent from the one we focus on.
9 This nonrobustness is …ne for their purpose since they focus on pointing out that the set of outcomes discoutinuously
changes by the introdction of a small switching cost. However, it would not be very ideal in our context because we focus on
identifying which outcome is selected by the introduction of the preparation stage. For that purpose, we want to minimize the
information necessary for the selection: the preference of each player is su¢ cient information to identify the outcome and the
information to compare di¤erent players’payo¤s is not necessary.
4
of election campaign with three possible actions (Left, Right, and Ambiguous). The main di¤erence from
their work is that they assume that once a candidate decides which of Left and Right to take, she cannot
move away from that action. Thus the characterization of the equilibrium is substantially more di¢ cult in
the model of the present paper, because in our model an action a player has escaped from can be taken again
by that player in the future.
2
Model
We consider a normal-form
game (I; (Xi )i ; ( i )i ) where I is the set of players, Xi is the …nite set of player
Q
i’s actions, and i : j Xj ! R is player i’s utility function. Before players actually take actions, they need
to “prepare” their actions. We model this situation as in Kamada and Kandori (2009): time is continuous,
t 2 [ T; 0] with T = 1, and the normal form game (referred to as a “component game”) is played once and
for all at time 0. The game proceeds as follows. First, at time 1, players simultaneously choose actions.
Between time 1 and 0, each player independently obtains opportunities to revise their prepared action
according to a Poisson process with arrival rate ( i )i . At t = 0, the action pro…le that has been prepared
most recently by each player is actually taken and each player receives the payo¤ that corresponds to the
payo¤ speci…cation of the component game. Each player has perfect information about past events at any
moment of time. No discounting is assumed, although this assumption does not change any of our results.10
We consider the limit of the set of subgame-perfect equilibrium payo¤s of this game as the arrival rate
goes to in…nity.
Let ( i )i ( ) be the set of subgame-perfect equilibrium payo¤s given arrival rate ( i )i .
De…nition 1 A payo¤ set S
lim i !1 for al l i ( i )i ( ). If
i = n =ri
R is a revision equilibrium payo¤ set of
if S = (ri )i ( ) :=
( ) is a singleton, we say its element is a revision equilibrium payo¤ .
(ri )i
for al l i
The set of action pro…les that correspond to the revision equilibrium payo¤ set is a revision equilibrium
set. If the revision equilibrium set is a singleton, we say its element is a revision equilibrium.
That is, a revision equilibrium payo¤ set is the set of payo¤s achievable by the revision game de…ned in
this section. Note that ri represents the relative frequency of the arrivals of the revision opportunities for
player i compared to player n. It will turn out in the sequel that this set is often a singleton. The term
“revision equilibrium payo¤” is used for a convenient abbreviation that represents the element of such a
singleton set. “Revision equilibrium set” and “revision equilibrium” are analogously de…ned.
Note well that whenever we refer to some action pro…le (resp. a payo¤ pro…le) as a revision equilibrium
(resp. a revision equilibrium payo¤), we implicitly mean that it is the unique element of the revision
equilibrium set (resp. revision equilibrium payo¤ set).
3
Pareto-Dominant Strict Nash Equilibrium in a Component Game
In this section we consider a component game with a Nash equilibrium that Pareto-dominates all other
action pro…les. Note that this condition is stronger than Pareto-ranked Nash equilibria. We will show that
this strategy pro…le is selected in our model.
De…nition 2 A strategy pro…le x is
Q said to be strongly Pareto-dominant strategy pro…le if there exists > 0
such that for all i and all x 2 X
i Xi with x 6= x , i (xi ) > i (x) + .
Proposition 3 Suppose that x is strongly Pareto-dominant strategy pro…le and inf
i. Then, x is a unique revision equilibrium.
1 0 Discounting
only scales down the payo¤ at time 0.
5
i
=:
i
>
1 for all
Intuitive explanation goes as follows. Firstly, we show that once all the players prepare the strongly
Pareto-dominant Nash equilibrium strategy, they will not escape from that state. Expecting that, if a
unilateral change of the preparation by player i can induce the strongly Pareto-dominant Nash equilibrium,
she will do so. In turn, player j whose unilateral change in the preparation can induce the situation where
player i’s unilateral change can induce the strictly Pareto-dominant Nash equilibrium, she will also do so
if the deadline is far expecting that player i will go to strongly Pareto dominant Nash equilibrium. By
induction, we can show that the strongly Pareto-dominant action will be selected. The formal proof requires
more solid arguments since player i takes into account the possibility that a player m taking xm now will
prepare some other action when she moves.
Notice that the above equilibrium selection always selects the strong Pareto-dominant equilibrium in
…nite component games. In particular, even if the better Nash equilibrium is risk-dominated by a worse
Nash equilibrium, the asynchronous revision process leads to the better equilibrium. The key is that, if the
remaining time is su¢ ciently long, since it is almost common knowledge that the opponent will move to the
Pareto-dominant equilibrium afterwards, the risk of mis-coordination can be arbitrarily small.
Notice also that we do not require that the game is of “pure-coordination,” in which all players’payo¤
functions are the same. This result is in a stark di¤erence from Laguno¤ and Matsui (1997), in which they
need to require that the game is of pure coordination, as otherwise their result would not hold (Yoon, 2001).
4
Pareto-Unranked Nash Equilibria in a Component Game: Battle of the Sexes
The result in the previous section suggests that the Pareto-dominant equilibrium is selected in games in
which such a pro…le exists. But what happens if there are multiple Pareto-optimal Nash equilibria? An
example of this class of games is the game of “battle of the sexes,” as follows:
g( ) = U
D
L
R
a + ; 1 0; 0 ;
0; 0
1; a
with a > 1 and
0. Observe that there are two pure strategy Nash equilibrium, (U; L) and (D; R), where
player 1 prefers the former equilibrium while player 2 prefers the latter.
If 1 = 2 = and = 0, then two players are perfectly symmetric. Then it is obvious that the set of
equilibrium payo¤ is symmetric. In Subsection 4.1, we show that this set is a full-dimensional subset of the
feasible payo¤ set. This set for the case with a = 2 is depicted in Figure 1.
However, if is strictly positive or 1 6= 2 , then what would happen is no longer obvious. We show in
Subsection 4.3 that for 1 = 2 and > 0, (U; L), which corresponds to the payo¤ (a + ; 1), is a revision
equilibrium. The revision equilibrium payo¤ in this case with a = 2 is depicted in Figure 2. Similarly, if
1 < 2 and = 0, (a; 1) is a revision equilibrium.
4.1
Anti-Equilibrium Selection in Symmetric Battle of the Sexes Game
In this subsection, we consider a 2 x 2 game g with
L
U
D
R
1 (U; L);
2 (U; L)
1 (U; R);
2 (U; R)
1 (U; D);
2 (U; D)
1 (D; R);
2 (D; R)
with two pure strategy equilibrium (U; L) and (D; R) and
1 (U; L)
>
(U;
L)
<
2
6
1 (D; R)
2 (D; R)
(D; R)
(U; L)
2 (U; L)
2 (D; R)
(U; R)
+
(D; R)
2 (U; R)
+
2 (U; L)
1
1
1
1
=
2
2
1
1
(U; L)
(U; L)
2 (D; R)
2 (D; R)
(U; R)
(D; R)
1 (U; R)
:
2 (U; L)
1
1
1
1
Since there are only two players, we de…ne r
r1 for simplicity. The battle of the sexes with r = 1 and
= 0 is a special case.
The purpose of this section is to highlight the di¤erence between two cases: A nongeneric case in which
a very special joint condition on the arrival rates and payo¤ structure is satis…ed (as in the case with r = 1
and = 0) and a generic case that encompasses all the other cases.
Proposition 4 Suppose that
(D; R)
1 (U; L)
2 (U; L)
2 (D; R)
(U; R)
+
1 (D; R)
2 (U; R)
+
2 (U; L)
1
1
=
2
(U; L)
1 (U; L)
2 (D; R)
2 (D; R)
1
1
2
1
(U; R)
1 (D; R)
1 (U; R)
2 (U; L)
1
in the game g. Then, the revision equilibrium payo¤ set satis…es
r
(g(0))
f(y1 ; y2 ) 2 co f(
1
(D; R) ;
2
(U; L)) ; (
1
(U; L) ;
2
(U; L)) ; (
1
(U; L) ;
2
The basic logic can be explained by our illustrative example of the battle of sexes with
7
(U; L))gg:
1
=
2
=
and
= 0. Figure 1 below depicts the set
1
(g(0)), for the case of a = 2.
Feasible payo¤ set and revision equilibrium payo¤ set in symmetric “battle of the
sexes” game with a = 2.
Let us …rst discuss how to obtain the extreme points of the sets, assuming a = 2 for the sake of concreteness. Point (2; 1) can be obtained by just sticking to (U; L) on the path of the play. In a perfectly symmetric
way, (1; 2) can be obtained by just sticking to (D; R) on the path of the play. To obtain the payo¤ pro…le
(1; 1), we construct an equilibrium as follows: When t < t for some appropriately chosen t , players 1
and 2 stick to U and R, respectively. For time t
t , if a player gets an opportunity, he gives up and
change his action (to D if player 1 gives up and to R if player 2 gives up). From now on we refer to this
equilibrium by “chicken race equilibrium,” as the equilibrium has a favor of the “chicken race” game,” in
which two drivers drive their cars towards each other until one of them gives in, while if both do not give in
then the cars crash and the drivers die.
The key feature of the game to sustain this “chicken race equilibrium” is that, since two players are
symmetric, their “cuto¤s,” t , must be the same. Note that the structure of the chicken race equilibrium is
close to war of attrition. Having the same t implies that both of the players are equally patient to give in.
8
Actually, the condition
(D; R)
1 (U; L)
2 (U; L)
2 (D; R)
1
1
=
2
(U; R)
+
(D;
R)
1
2 (U; R)
+
2 (U; L)
1
(U; L)
(U;
L)
1
2 (D; R)
2 (D; R)
1
2
1
(U; R)
(D;
R)
1
1 (U; R)
2 (U; L)
1
is the generalized condition so that two players have the same “giving-up” cuto¤. The natural conjecture
from the war of attrition is that, if we break the symmetry, we can destroy the chicken race equilibrium.
This conjecture will be shown formally in the next section. That is, with
(D; R)
1 (U; L)
2 (U; L)
2 (D; R)
1
1
6=
2
(U; R)
+
1 (D; R)
2 (U; R)
+
2 (U; L)
(U; L)
1 (U; L)
2 (D; R)
2 (D; R)
1
1
2
1
(U; R)
1 (D; R)
1 (U; R)
2 (U; L)
1
we show that the equilibrium is essentially unique.
Let us …nish this section by explaining how to obtain the payo¤s other than the extreme points of the
equilibrium payo¤ set de…ned in Proposition 4. Because we are not assuming any public randomization
device, it is not immediately obvious that these payo¤s can be achieved. However, we can use the Poisson
arrivals when t is large as if it were a public randomization device: players start their play at (U; R), and
“count” the numbers of Poisson arrivals to each player until t < t and then decide which of three
equilibria to play after t. Since the length of the time interval during which players count the arrivals can
be arbitrary rational and irrational numbers, we can sustain all the payo¤s in the equilibrium payo¤ set
de…ned in Proposition 4.
4.2
Illustration: Three State Example
In the next subsection, we show that in generic coordination games, a unique equilibrium can be selected.
To gain some intuition for why we have equilibrium selection in generic coordination games, this subsection
considers a simpli…ed setting of a revision game-like model. The example also reveals the di¢ culty of the
analysis of our general setting.
Suppose that there exist three states, s0 , s1 and s2 , in which payo¤s are, (0; 0), (2 + ; 1) and (1; 2),
respectively. Time is t 2 [ T; 0] as before, and the initial state (state at t = T ) is predetermined and
s0 . The Poisson arrival rate is symmetric and 1 = 2 = > 0. Each player i = 1; 2 has two actions, Ii
(Insist) and Yi (Yield). Starting from state s0 , action Ii does not change the state, while action Yi changes
the state from s0 to s i . Thus, i’s action Yi changes the state that the opponent i favors the most. We
suppose (only in this subsection) states s1 and s2 are “absorbing states.”That is, once the state si , i = 1; 2,
is reached, the state cannot change to any other states (See Figure 2).
The transition of 3-state example
As in our general model, we suppose that each player has a chance to revise their actions according to a
Poisson process with arrival rate , and the …nal state will be realized. As in the analysis of the general
9
model, we consider the limit case of T ! 1, which is equivalent to the case when the frequency of revision
( ) goes to in…nity with …xed T > 0.
The following proposition holds:
Proposition 5 There is a unique subgame perfect equilibrium. In this equilibrium, the action pro…le at time
t = T is (I1 ; Y2 ). Hence state s1 is reached immediately.
Since this proposition is a straightforward corollary of our general result, we do not provide a proof.
Rather, we explain the intuition of this result. First, it is intuitive that players use cuto¤ strategies: there
exists ti for each i such that if the state is still in s0 at time t, for t < t1 , player i takes Ii , for t = ti ,
is indi¤erent between Ii and Yi and for t > ti , takes action Yi .
Suppose …rst that two players have exactly the same cuto¤ t1 = t2 . Then, at this cuto¤, if player i is
given a chance to revise her action, her expected payo¤ would be exactly equal to 1, as she must be indi¤erent
between Ii and Yi . But this is impossible, because the payo¤ structure is asymmetric. More precisely, if
player 1 is indi¤erent, player 2 must strictly prefer taking Y2 . On the other hand, if player 2 is indi¤erent,
player 1 must strictly prefer taking I1 . Notice that this argument suggests that t2 < t1 . That is, player
2’s cuto¤ must be further away from the deadline than player 1’s cuto¤. This is intuitive: player 1 expects
more from insisting (since 2 + > 2), so he has a larger incentive to wait.
Given t2 < t1 , we argue that 1 < t1 < 0. This is straightforward: note that player 2 will take Y2
as soon as possible after t1 . We have 1 < t1 : suppose not. By instead waiting until a …nite but large
(in absolute value) T , he can guarantee approximately 2 + as by player 2 will take Y2 as soon as possible.
Hence, t1 = 1 is not optimal. We also have t1 < 0 because the payo¤ from taking I1 at time
will
converge to 0 as goes to 0, while the payo¤ from taking Y1 is always 1.
Finally, we argue that for any t2 2 [ T; t1 ], the payo¤ from taking I2 (assuming that she takes Y2
whenever she gets a chance to revise at later periods; the same quali…cation applies whenever we consider
the expected payo¤ of taking Y2 in this paragraph) is strictly less than 1, hence taking Y2 is a unique best
response at time T from the following reasons. First, we show that player 2’s payo¤ from taking action I2
at time t1 is strictly less than 1. Since player 1 is indi¤erent between I1 and Y1 at t1 , player 1’s expected
payo¤ from I1 at t1 is exactly equal to 1. But the strategies after time t1 is perfectly symmetric, while
the only di¤erence between players 1 and 2 is that the payo¤ that they obtain when they yield is higher
for player 1. This implies that player 2 expects strictly less than 1 at time t1 from I2 . Let this payo¤ be
v2 < 1. Now, consider any time t2 < t1 . At this time, the expected payo¤ from playing I2 is a convex
combination of (i) 1 with a weight that is equal to the probability that she obtains a chance to revise her
action between time t2 and t1 , and (ii) v with a weight that is equal to 1 minus the probability speci…ed
in case (i). Since the weight on (ii) is strictly positive, the overall payo¤ is strictly less than 1 for any t2 ,
which implies t2 = T as desired.
Therefore, we conclude that the action pro…le at time t = T is (I1 ; Y2 ). We note that the analysis
of the general setting is substantially more di¢ cult than the case we discussed here. The reason is that
not both s1 and s2 are a priori absorbing states since players can revise their actions in s1 and s2 . Indeed,
it will turn out that the action pro…le that corresponds to state s2 is not absorbing in the general model.
Nonetheless, the example in this section captures an essential argument for our equilibrium selection: a
slight payo¤ asymmetry leads to a slight di¤erence of the incentives, which causes a huge di¤erence in the
equilibrium behavior in the dynamic setting.
4.3
Equilibrium Selection in Generic Coordination Games
In this subsection we consider generic coordination games with payo¤ matrix as follows:
L
U
D
R
1 (U; L);
2 (U; L)
1 (U; R);
2 (U; R)
1 (U; D);
2 (U; D)
1 (D; R);
2 (D; R)
10
with two pure strategy equilibrium (U; L) and (D; R) and
player 1 prefers (U; L) to (D; R) :
player 1 prefers (D; R) to (U; L) :
(D; R)
1 (U; L)
2 (U; L)
2 (D; R)
1
1
6=
2
(U; R)
+
1 (D; R)
2 (U; R)
+
2 (U; L)
1
For example, the “ Battle of the Sexes” game with
1
=
1 (U; L)
>
(U;
L)
<
2
(U; L)
1 (U; L)
2 (D; R)
2 (D; R)
1
2
1
2
1 (D; R)
2 (D; R)
(U; R)
1 (D; R)
1 (U; R)
:
2 (U; L)
1
and " > 0 is a generic game.
Proposition 6 Suppose that
(D; R)
1 (U; L)
2 (U; L)
2 (D; R)
1
1
<
2
(U; R)
+
(D;
R)
1
2 (U; R)
+
2 (U; L)
(U; L)
(U;
L)
1
2 (D; R)
2 (D; R)
1
1
2
1
(U; R)
(D;
R)
1
1 (U; R)
:
2 (U; L)
1
Then, (U; L) is a unique revision equilibrium. On the other hand, Suppose that
(D; R)
1 (U; L)
2 (U; L)
2 (D; R)
1
1
>
2
(U; R)
+
(D;
R)
1
2 (U; R)
+
2 (U; L)
(U; L)
(U;
L)
1
2 (D; R)
2 (D; R)
1
1
2
1
(U; R)
(D;
R)
1
1 (U; R)
:
2 (U; L)
1
Then, (D; R) is a unique revision equilibrium.
The proposition says that the revision equilibrium payo¤ set in 2 x 2 games is generically a singleton.
Let us consider the chicken race equilibrium, that is, consider what players take if they have a chance to
revise the action when the current preparation is (U; R). For simple explanation, we assume (only for this
intuitive explanation) that (D; R) and (U; L) are absorbing. Without loss of generality, we assume
(D; R)
1 (U; L)
2 (U; L)
2 (D; R)
1
1
<
2
(U; R)
+
(D;
R)
1
2 (U; R)
+
2 (U; L)
1
(U; L)
(U;
L)
1
2 (D; R)
2 (D; R)
1
2
1
(U; R)
(D;
R)
1
1 (U; R)
:
2 (U; L)
1
since the other case is completely symmetric. Consider player 1’s incentive to give in at time
both players will give in after t. If player 1 skips the opportunity, player 1 gets
(D; R)
1 (U; L)
1
t given that
(U; R) if player 1 moves …rst,
1 (U; R) if player 2 moves …rst.
1
Note that (i) if 1 (U; L)
1 (U; R) becomes larger, then the gain of waiting goes up and that (ii) 2 is larger,
then player 2 is more likely to move …rst and player 1 will get higher gain 1 (U; L)
1 (U; R) and hence the
gain of waiting goes up. Therefore, as we explained in the previous section, in the chicken race equilibrium,
player 1 becomes more tough to give in, which makes player 2 gives in earlier. The above inequality shows
that in these two cases, it is more likely that the payo¤ of the equilibrium (U; L) preferable to player 1 will
be selected.
11
Actual equilibrium path is illustrated by the following …gures. If
prepares the static best response to the other:
t is su¢ ciently close to 0, every player
Then, suppose (U; R) is prepared at time t. As we explained in Section 4.2, the “strong”player 1 sticks
to (U; R) if t is su¢ ciently far away from 0, knowing that the “weak” player 1 gives in and goes to (U; L)
if player 2 can move afterwards:
There is more complication that arises when t is further way from the deadline 0. Suppose (D; R) is
prepared. As we mentioned in Section 4.2, (D; R) is not an absorbing state, that is, player 2 does not want
to stick (D; R) when t is su¢ ciently far away from the deadline. The reason is that (D; R) is a “risky”
choice. If player 1 can move, player 1 will prepare (U; R). Then, if player 2 moves …rst, they go to (U; L).
If player 1 moves …rst after player 1 starts to give in, they go to (D; R). Otherwise, they stay at (U; R).
Since player 1 needs to move …rst to go to (U; R), the time they spend after going to (U; R) depends on how
fast player 1 will have an opportunity to move. Hence, if player 1 has an opportunity late, then there is a
probability that they will end up with (U; R), a bad outcome. On the other hand, if player 2 prepares (D; L)
today, then only if player 1 gets one opportunity to move, they will end up with (U; L).
Let us go back to the illustrative example of the battle of the sexes. Let us consider the case with the
same arrival rate 1 = 2 = but a slight asymmetry in the payo¤ matrix:
L
R
a + ; 1 0; 0
0; 0
1; a
U
D
The proposition argues that even a very small
(D; R)
1 (U; L)
2 (U; L)
2 (D; R)
1
<
is su¢ cient to choose a unique equilibrium and, if
(U; R)
+
(D;
R)
1
2 (U; R)
+
2 (U; L)
1
12
(U; L)
(U;
L)
1
2 (D; R)
2 (D; R)
1
(U; R)
(D;
R)
1
1 (U; R)
:
2 (U; L)
1
> 0,
and (U; L) will be selected. Figure 3 depicts the revision equilibrium payo¤, for the case of a = 2. Therefore,
the slight asymmetry is enough to choose a unique equilibrium.
Similarly, if 1 < 2 with = 0, we also have
(D; R)
1 (U; L)
2 (U; L)
2 (D; R)
1
1
<
2
(U; R)
+
1 (D; R)
2 (U; R)
+
2 (U; L)
1
(U; L)
1 (U; L)
2 (D; R)
2 (D; R)
1
2
1
(U; R)
1 (D; R)
1 (U; R)
:
2 (U; L)
1
since player 1 has more commitment power. Hence, (U; L) is selected.
Feasible payo¤ set and revision equilibrium payo¤ in asymmetric “battle of the
sexes” game with a = 2.
5
Concluding Remarks
We analyzed the situation in which two players prepare their actions before they play a normal-form coordination game at a predetermined date. In the preparation stage, each player stochastically obtains
opportunities to revise their actions, and …nally-revised action is played at the deadline. We showed that,
(i) If there exists a Nash equilibrium that strictly Pareto-dominates all the other action pro…les, then it is
13
the only equilibrium; and (ii) in “battle of the sexes”games in which Nash equilibria are not Pareto ranked,
(ii-a) while with perfectly symmetric payo¤ structure, the equilibrium set is a full-dimensional subset of the
feasible payo¤ set, (ii-b) a slight asymmetry is enough to select a unique equilibrium, which corresponds to
the Nash equilibrium in the static game that gives the highest payo¤ to the “strong” player.
Let us mention possible directions of future research. First, our analysis has been restricted to 2 x 2
games when we have dealt with the case with multiple Pareto optimal equilibria, but the basic intuition
seems to extend to more general cases. Second, even in a 2 x 2 games, we have not yet considered all the
cases. For example, in the game as follows, what happens in revision games seems not straightforward:
U
D
L
10; 11
11; 1
R
1; 10
0; 0
In this game, D is a dominant action for player 1, while player 2’s best response depends on player 1’s action.
Although (D; R) is the unique Nash equilibrium, its payo¤ is Pareto-dominated by that of (U; L). Hence, it
is not obvious whether players want to switch to the Nash action as soon as possible.
Third, it would be interesting to consider the case in which there exists only one mixed equilibrium. For
example, in a symmetric “matching pennies” game, it is obvious that the probability distribution over the
outcome is the same as in the mixed strategy equilibrium of the component game. A question is whether
this is true for asymmetric case.
Fourth, it would be interesting to see the hybrid version of synchronized and asynchronized revision
games.
Fifth, in Kamada and Sugaya (2010b), we are working on a model in which players can choose the
frequency of revision (the arrival rate) at the beggining of the game. The result of this paper seems to
suggest that players would be better o¤ by choosing small arrival rates. However, this is not an entirely
correct intuition, because lowering one player’s arrival rates have two e¤ects: On the one hand, it makes
the opponent more willing to lower her arrival rate, to maintain her “strength” as a Stackelberg leader. On
the other hand, it makes her less willing to lower it because she does not want to end up in sticking at the
nonequilibrium state that gives her a bad payo¤. It turns out that the strength of these two e¤ects depend
in a nontrivial way on the payo¤ structure.
These possibilities are out of the scope of this paper, but we believe that the present paper laid out
motivations that are enough to pursue these generalizations.
References
[1] Abreu, Dilip and Faruk Gul “Bargaining and Reputation” Econometrica, 2000, 68, 85-117.
[2] Abreu, Dilip and David Pearce “Bargaining, Reputation and Equilibrium Selection in Repeated
Games” Princeton Economic Theory Papers, 2000.
[3] Ambrus, Attila and Shih-En Lu, “A continuous-time model of multilateral bargaining,” 2009,
mimeo.
[4] Benoit, Jean-Pierre and Vijay Krishna “Finitely Repeated Games,”Econometrica, 1985, 53, 905922.
[5] Carlsson, Hans and Eric van Damme “Global Games and Equilibrium Selection,” Econometrica,
1993, 61, 989-1018.
[6] Caruana, Guillermo and Liran Einav “A Theory of Endogenous Commitment,” Review of Economic Studies, 2008, 75, 99-116.
[7] Dutta, Prajit K. “Coordination Need Not Be A Problem,” 2003, mimeo.
14
[8] Farrell, J., and Saloner, G., “Standardization, Compatibility, and Innovation,” RAND Journal of
Economics, 1985, 16, 70-83.
[9] Fudenberg, Drew and Wojciech Olszewski “Repeated Games with Asynchronous Monitoring”
2009, mimeo.
[10] Harsanyi, John, C. and Reinhard Selten “A general Theory of Equilibrium Seletion in Games,”
MIT press, Cambridge, MA. 1988.
[11] Kamada, Yuichiro and Michihiro Kandori, “Revision Games,” 2009, mimeo.
[12] — and Takuo Sugaya, “Policy Announcement Game: Valence Candidates and Ambiguous Policies,”
2010a, mimeo.
[13] — and Takuo Sugaya, “Who Wants to Tie the Hands?,” 2010b, mimeo.
[14] Kandori, Michihiro, George J. Mailath, and Rafael Rob “Learning, Mutation, and Long Run
Equilibria in Games,” Econometrica, 1993, 61, 29-56.
[15] Lipman, Barton, L. and Ruqu Wang “Switching Costs in Frequently Repeated Games,” Journal
of Economic Theory, 2000, 93, 149-190.
[16] Matsui, Akihiko and Kiminori Matsuyama “An Approach to Equilibrium Selection,” Journal of
Economic Theory, 1995, 65, 415-434.
[17] Morris, Stephen, Rafael Rob, and Hyun Song Shin “p-Dominance and Belief Potential,”Econometrica, 1995, 63, 145-167.
[18] Morris, Stephen and Hyun Song Shin “Unique Equilibrium in a Model of Self-Ful…lling Currency
Attacks” American Economic Review, 1998, 88, 587-597.
[19] Oyama, Daisuke, Satoru Takahashi, and Josef Hofbauer “Monotone Methods for Equilibrium
Selection under Perfect Foresight Dynamics,” Theoretical Economics, 2008, 3, 155-192.
[20] Rubinstein, Ariel “The Electronic Mail Game: A Game with Almost Common Knowledge,”American
Economic Review, 1989, 79, 389-391.
[21] Sugaya, Takuo and Satoru Takahashi “Coordination Failure in Repeated Games with Private
Monitoring,” 2009, mimeo
[22] Satoru Takahashi “In…nite Horizon Common Interest Games with Perfect Information,” Games and
Economic Behavior, 2005, 53, 231-247.
[23] Young, H. Peyton “The Evolution of Conventions,” Econometrica, 1993, 61, 57-84.
[24] Young, H. Peyton “Conventional Contracts,” Review of Economic Studies, 1998, 65, 773-792.
A
Appendix: Proof of Propositions
Since …xing T and letting converge to in…nity is equivalent to …xing and letting T converge to in…nity,
we consider the latter formulation for the sake of “backward induction.” The following notations are useful:
let be a strategy in the revision game. As we will see, only depends on the calendar time t and the action
pro…le most recently prepared by the player. Therefore, letting xi be the action most recently prepared by
player i at t, we can write ti (x) 2 Ai .11 Finally, let BRi (x i ) be a static best response to x i by player i
in the component game.
1 1 We
can concentrate on the pure strategies without loss of generality.
15
A.1
Proof of Proposition 3
Firstly, we prove the following two lemmas:
Lemma 7 Suppose that x is strongly strictly Pareto-dominant strategy pro…le and inf i =: i > 1.
Then if a subgame starts with a pro…le x , there is a unique subgame perfect equilibrium in that subgame and
this equilibrium designates actions xi for each i on the equilibrium path of the play.
Proof. Let be a game with strongly strictly Pareto e¢ cient strategies. Let the Pareto-dominant action
pro…le in the component game be x . Take an arbitrary > 0 such that for all i and all x 2 X with x 6= x ,
i (xi ) > i (x) + .
The lower bound of the payo¤ from taking action xi at time > 0 given the opponent’s current action
x i is
P
P
exp
exp
i (x ) + 1
i:
j j
j j
The upper bound of the payo¤ from taking action x
^i 6= xi at time
x i is
P
exp
xi ; x i ) + 1 exp
i (^
j j
Hence taking xi is strictly better at time
P
exp
P
> exp
P
() exp
i (x
j
j
i (x
0
(= > @
)+ 1
xi ; x i )
i (^
j
j
xi ; x i )
i (^
exp
1
1
P
j
j
j
<P
ln
j
j
P
j
j
exp
> 1
i
P
P
exp
1
(x )
(x )
(x )
j
j
j
1A [ (x )
j
Notice that the right hand side is strictly positive.
Let
(
1
= min P
ln
i2I
(x ):
j
j
exp
+ 1
)
()
P
conditional on any history if
j
j
> 0 given the opponent’s current action
[ (x )
j
i] :
i] :
+1 :
i
+1
i
)
:
Notice that
is strictly positive. Also notice that for all t 2 [0; ), each player i chooses xi at t conditional
on the opponent’s current action x i .
Now we make a backward induction argument. Suppose that for all time t 2 [0; k ), each player i chooses
xi at t conditional on the opponent’s current action x i . We show that for all time t0 2 [k ; (k + 1) ),
each player i chooses xi at t0 conditional on the opponent’s current action x i .
The lower bound of the payo¤ from taking action xi at time t with t 2 [k ; k + 0 ) with 0 > 0 is
P
0
exp
j
j
i (x
)+ 1
exp
0
exp
0
P
j
j
xi ; x i )
i (^
+ 1
16
exp
j
0
The upper bound of the payo¤ from taking action x
^i 6= xi at time
0
P
j
i:
> 0 is
P
j
j
(x ):
0
Hence taking xi is strictly better at time
P
0
exp
P
0
> exp
i (x
j
j
j
j
(=
0
conditional on the opponent’s current action x
)+ 1
xi ; x i )
i (^
<P
1
+ 1
ln
j
j
if
i
P
0
exp
(x )
j
j
+1 :
(x )
j
j
P
0
exp
i
i
Thus each player i strictly prefers playing xi at all time t0 < k
n
+ mini2I
P1
j
j
ln
(x )
i
+1
o
=
(k + 1) conditional on the opponent’s current action x i . Since we have already proven that each player
i plays xi at all time t 2 [0; ) conditional on the opponent’s current action x i , the backward induction
argument is complete. This completes the proof.
Let V i ( jht ) be player i’s value at history ht . The following Lemma shows that, if there is a player
whose value is close to i (x ), then for all the other players, the value is close to i (x ).
Lemma 8 (Takahashi (2005)) Suppose that x is strongly strictly Pareto-dominant strategy pro…le. Then,
there exists K < 1 such that for any > 0, for any a and t, if there exists i with V i ( jht )
,
i (x )
then
V j ( jht )
K for all j:
j (x )
Now we prove the proposition by mathematical induction. Let V im (t) be the in…mum of player i’s payo¤
such that at least m players take xj at t. We will show, by mathematical induction with respect to m,
that V im (t) is su¢ ciently close to i (x ) when t is su¢ ciently large.
By Lemma 7, when m = n, V im (t) = i (x ) for all i; m.
Suppose limt!1 V im (m + 1; t) = i (x ). Take > 0 arbitrarily. There exists T0 ( ) such that
V im+1 (t)
i
(x )
for all t T0 ( ). Consider the situation where m players take xj at t with t = T0 ( ) + . Then, if player
j that is not taking xj at time t will move next by T0 ( ), then player j will yield at least j (x )
by
taking xj . Hence, Lemma 8 implies each player i will yield i (x ) K in this case. Therefore,
min
P
V im (t)
j
j
j
+ 1
n
1
min
P
j
j
j
min
P
j
+ 1
j
j
P
exp
n
1
j
n
1
P
j
j
j
j
n
1
min
P
j
1
j
j
min
P
j
n
1
j
j
exp
n
1
exp
P
j
j
P
j
o
j
(
i
(x )
j
o
K )
!
o
j
j
exp
for all i. Note that there exists T ( ) such that for all
o
P
exp
exp
min
P
j
i
(x )
i
P
!
o
j
j
K
i
T ( ),
i
(x )
!
o
17
min
P
j
i
1
j
i
(x )
j
min
P
j
j
j
!
2
i
2
;
which implies
V
i
m
min
P
(t)
j
i
(x ) +
min
P
1
j
j
j
j
j
!
i
(K + 1)
for all i. Consider V im (t) with t = T0 ( ) + T ( ) + . Then, if player j that is not taking xj at time t will
move next by T0 ( ) + T ( ), then she will yields at least j (x )
by taking xj . Again, Lemma 8 implies
each player i will yield i (x ) K in this case. Otherwise, from above, player i’s value at T0 ( ) + T ( ) is
min
min
(K + 1) . Therefore,
at least P jj i (x ) + 1 P jj
i
j
j
min
P
V im (t)
j
j
j
+ 1
n
1
min
P
j
j
j
min
P
j
j
j
n
1
j
j
j
(K + 1)
Then, for
n
1
o
j
j
i
o
j
j
i
K )
!
o
min
P
j
(x )
P
exp
(x )
j
j
P
n
1
(
P
exp
exp
min
P
+ 1
P
exp
!
o
j
j
min
P
j
j
i
(x ) +
min
P
1
j
j
j
i
(x ) +
min
P
1
j
j
j
j
j
!
!
(K + 1)
i
i
!
!
T ( ),
V
i
m
(t)
(
min
P
j
j
+
j
1
min
P
j
j
j
!
min
P
j
j
j
)
i
(x ) +
1
min
P
j
j
j
!2
Recursively, we can show that for any , for any l, for all i and t T0 + lT ( ),
(l 1
!m
)
!l
X
min j
min j
min j
i
P
P
1 P
V m (t)
i (x ) + 1
m=0
j
j
j
j
j
Therefore, for any " > 0, there exists l such that for any l l,
(l 1
!m
)
X
min j
min j
P
1 P
i (x )
m=0
j
j
j
1
"
,
3(K+l)
we have, for any i and t
i
min
P
j
j
!l
i
j
(x )
j
j
Taking
j
i
i
(K + 2) :
(K + l) :
"
:
3
"
:
3
T0 + lT ( ),
V im (t)
i
(x )
"
as desired.
A.2
Proof of Proposition 4
Firstly, we prove the existence of an equilibrium with expected payo¤ ( 1 (D; R) ; 1 (U; L)) by construction:
let t be
1
1 1 (D; R)
1 (U; R)
1 (U; L)
1 (U; R)
log
+
(1)
t =
+
(U;
L)
(D;
R)
(U;
L)
1
2
2 1
1
1
1 (D; R)
1
2 2 (U; L)
2 (U; R)
2 (D; R)
1 (U; R)
=
log
+
:
+
(D;
R)
(U;
L)
(D;
R)
1
2
1 2
2
2
2 (U; L)
18
Note that at t , staying at (U; R) gives ( 1 (D; R) ; 1 (U; L)) to both players, given that for
t
i (x) = BRi (x). We verify that the following strategy pro…le constitutes an equilibrium:
Player 1 takes U at
T.
Player 2 takes R at
T.
t 2 ( t ; 0],
Player 1 takes the following Markov strategy:
– For
– For
t 2 [ T; ; t ),
t 2 [ t ; 0],
t
t
1 (U; R) = 1 (D; R)
t
1 (x) = BR1 (x).
= U and
t
1 (U; L)
=
t
1 (D; L)
= U.
Player 2 takes the following Markov strategy:
– For
– For
t
t
2 (U; L) = 2 (U; R)
t
2 (x) = BR2 (x).
t 2 [ T; t ),
t 2 [ t ; 0],
= R and
t
2 (D; L)
=
t
2 (D; R)
= R.
Then, (1) gives the condition that, given that players follow the above strategy after t , player 1 is
indi¤erent between U and D at x = (U; R) and player 2 is indi¤erent between R and L at x = (U; R). Given
this, at (U; R) and (D; R) at t with t
t , (1) ensures that the expected payo¤ for player 1 is 1 (D; R).
Similarly, at (U; L) and (U; R) at t with t t , the expected payo¤ for player 2 is 1 (U; L). Therefore,
At
T , given player 2 takes R, both U and D are optimal.
At
T , given player 1 takes U , both L and R are indi¤erent.
For t 2 (t ; T ), at (U; R) and (D; R), both U and D are optimal. At (U; L) and (D; L), U is strictly
optimal.
For t 2 (t ; T ), at (U; L) and (U; R), both L and R are optimal. At (D; L) and (D; R), R is strictly
optimal.
For t 2 [t ; 0],
t
i (x)
= BRi (x) is optimal.
Secondly, we prove the existence of an equilibrium with expected payo¤ close to ( 1 (U; L) ; 2 (U; L)) by
construction. Let T > 0 be a large number. We verify that the following strategy pro…le constitutes an
equilibrium:
Player 1 takes U at
T.
Player 2 takes R at
T.
Player 1 takes the following Markov strategy:
– for
– for
t 2 [ T; t ),
t 2 [ t ; 0],
t
t
1 (U; R) = 1 (D; R)
t
1 (x) = BR1 (x).
= U and
t
1 (U; L)
=
t
1 (D; L)
= U.
Player 2 takes the following Markov strategy:
– for
– for
t 2 [ T; T ),
t 2 [ T ; 0],
t
2 (U; L)
t
2 (x)
=
t
2 (U; R)
= R and
t
2 (D; L)
=
t
2 (D; R)
= R.
= BR2 (x).
Note that for su¢ ciently large T , the expected payo¤ is su¢ ciently close to ( 1 (U; L) ; 2 (U; L)). We
verify that the above strategy pro…le constitutes an equilibrium: Note that at (U; R) and (D; R) at t with
t
t , T < t ensures that the expected payo¤ for player 1 is higher at (U; R). At (U; L) and (U; R)
at t with t
t , the expected payo¤ for player 2 is 2 (U; L). Therefore,
19
At
T , given player 2 takes R, U is optimal.
At
T , given player 1 takes U , both L and R are indi¤erent.
For
t 2 [ T; t ], at all x, U is optimal.
For
t 2 [ T; t ], at (U; L) and (U; R), both L and R are optimal. At (D; L) and (D; R), R is optimal.
For
t 2 [ t ; 0], each player i takes
t
i (x)
= BRi (x).
Symmetrically, the following equilibrium approximates (
Player 1 takes U at
T.
Player 2 takes R at
T.
1
(D; R) ;
2
(D; R)).
Player 1 takes the following Markov strategy:
– For
– For
t 2 [ T; T ),
t 2 [ T ; 0],
t
1 (U; R)
t
1 (x)
=
t
1 (D; R)
= U and
t
1 (U; L)
=
t
1 (D; L)
= U.
= R and
t
1 (D; L)
=
t
1 (D; R)
= R.
= BR1 (x).
Player 2 takes the following Markov strategy:
– For
– For
t 2 [ T; t ),
t 2 [ t ; 0],
t
2 (U; L)
t
2 (x)
=
t
1 (U; R)
= BR2 (x).
Therefore, we construct equilibria approximating ( 1 (D; R) ; 2 (U; L)), ( 1 (U; L) ; 2 (U; L)), and ( 1 (U; L) ;
Note that the three equilibrium, nobody moves until T and the following is also an equilibrium for any
N; M : with T > T ,
Player 1 takes U at
T.
Player 2 takes R at
T.
Player 1 takes the following Markov strategy:
– For
t 2 [ T; T ),
t
1 (U; R)
=
t
1 (D; R)
= U and
t
1 (U; L)
t
1 (D; L)
=
= U.
– If the number of chances where player 1 can move for [ T; T ) is less than N , going to the …rst
equilibrium, that is,
For
For
t 2 [ T ; t ), t1 (U; R) = t1 (D; R) = U and
t 2 [ t ; 0], t1 (x) = BR1 (x).
t
1 (U; L)
=
t
1 (D; L)
= U.
– If the number of chances where player 1 can move for [ T; T ] is no less than N and no more
than M , going to the second equilibrium, that is,
For
For
t 2 [ T ; t ), t1 (U; R) =
t 2 [t ; 0], t1 (x) = BR1 (x).
t
1 (D; R)
= U and
t
1 (U; L)
=
t
1 (D; L)
= U.
– If the number of chances where player 1 can move for [ T; T ] is no more than M , going to the
third equilibrium, that is,
For
For
t 2 [ T ; T ), t1 (U; R) = t1 (D; R) = U and
t 2 [ T ; 0], t1 (x) = BR1 (x).
t
1 (U; L)
=
t
1 (D; L)
= U.
Player 2 takes the following Markov strategy:
– For
t 2 [ T; T ),
t
2 (U; L)
=
t
2 (U; R)
= R and
20
t
2 (D; L)
=
t
2 (D; R)
= R.
2
(U; L)).
– If the number of chances where player 1 can move for [ T; T ] is less than N , going to the …rst
equilibrium, that is,
For
For
t 2 [ T ; t ), t2 (U; L) = t2 (U; R) = R and
t 2 [ t ; 0], t2 (x) = BR2 (x).
t
2 (D; L)
t
2 (D; R)
=
= R.
– If the number of chances where player 1 can move for [ T; T ] is no less than N and no more
than M , going to the second equilibrium, that is,
For
For
t 2 [ T ; T ), t2 (U; L) = t2 (U; R) = R and
t 2 [ T ; 0], t2 (x) = BR2 (x)
t
2 (D; L)
t
2 (D; R)
=
= R.
– If the number of chances where player 1 can move for [ T; T ] is no more than M , going to the
third equilibrium, that is,
For t 2 [ T ; t ), t2 (U; L) = t2 (U; R) = R and
for t 2 [ t ; 0], t2 (x) = BR2 (x).
t
2 (D; L)
=
t
2 (D; R)
= R.
It is straightforward to show that this is an equilibrium and with appropriate choices of T , N , M , and
su¢ ciently large T
T , we can attain any payo¤ pro…le that can be expressed by a convex combination
of ( 1 (D; R) ; 2 (U; L)), ( 1 (U; L) ; 2 (U; L)), and ( 1 (U; L) ; 2 (U; L)).
A.3
Proof of Proposition 6
We assume that
(U; R) 1
>
(D;
R) 1
1
(D; R)
1 (U; L)
1
(U; L)
(D;
R)
2
(D; L) 1
:
2 (U; L) 2
2
1
2
The other case is completely symmetric.
We use backward induction to derive the (essentially) unique subgame perfect equilibrium. Let xti be the
prepared action by player i at t. In addition, let Vxi (t) be player i’s expected payo¤ when players prepare
an action pro…le x at t and no player has a chance to move at t. Firstly, note that there exists
>0
such that, for each t 2 (
; 0], given player i’s prepared action xt i , player i prepares the best response
to xt i , that is,
t
t
t (x i ) = BR(x i ):
Suppose the players “know”the game proceeds as explained above after
action pro…le, each player’s value is given as follows:
t. Then, at
t, at each prepared
At (U; L),
VU L (t) = (
1
At (U; R), if player 2 moves …rst, then 1 (U; L) ;
and otherwise, 1 (U; R) ; 2 (U; R). Therefore,
VU1R (t)
2
=
|
2
(1
1+
(U; L) ;
2
2
(U; L)) :
(U; L), if player 1 moves …rst, then
exp ( (
{z
1
+
2 ) t)) 1
player 2 m oves …rst
1
+
+
|1
(1
2
exp ( (
{z
1
+
player 1 m oves …rst
+exp ( ( 1 +
|
{z
2 ) t) 1
nob o dy m oves
21
}
(U; R)
}
(U; L)
2 ) t)) 1
}
(D; R)
1
(D; R) ;
2
(D; R),
and
VU2R (t)
=
2
2
(U; L)
+
1
+
2
(1
1
(D; R)
exp ( (
1
+
exp ( (
1
2 ) t))
2
+ 2
+ 2 (U; R) exp ( (
(1
+
2 ) t))
1
At (D; L), if player 1 moves …rst, then 1 (U; L) ;
and otherwise, 1 (D; L) ; 2 (D; L). Therefore,
1
VDL
(t)
1
=
1
+
(1
+
2 ) t) :
(U; L), if player 2 moves …rst, then
exp ( (
1
+
2 ) t))
1
1
(D; R) ;
(U; L)
2
2
+
2
1
1+ 2
+ exp ( (
(1
1
exp ( (
+
2 ) t)
+
1
2 ) t))
1
(D; L)
1
+
1
(D; R)
and
2
VDL
(t)
1
=
1
+
(1
2
+
exp ( (
2 ) t))
2
(U; L)
2
1+ 2
+ exp ( (
(1
1
exp ( (
+
2) t 2
+
1
2 ) t))
2
(D; R)
(D; L)) :
At (D; R),
VDR (t) = (
Let t1 be the solution for
1
(D; R) ;
1
(D; R) = VU1R (t)
2
(U; L) = VU2R (t) :
and t2 be that for
2
(D; R))
That is,
t1 =
1
+
1
t2 =
1
+
1
and
log
2
(D; R)
1 (U; L)
1
1
2
log
2
By assumption, t1 < t2 .
Then, for t 2 ( t1 ; 0], noting that
(U; L)
(D;
R)
2
2
2
1
2
(U; R)
+
(D;
R)
1
(U; L)
(U;
L)
1
1
(U; R)
+
2 (U; L)
2
1
(D; R)
2 (D; R)
2
(U; R)
(D;
R)
1
1
(U; R)
2 (U; L)
1
(U; L) > VU2R (t1 ) since t2 > t1 , we have the following:
If xt2 = L, player 1 will take U since
1
1
(U; L) > VDL
(t) :
1
(D; R) > VU1R (t)
2
(U; L) > VU2R (t)
If xt2 = R, player 1 will take D since
If xt1 = U , player 2 will take L since
22
2
(D; R),
If xt1 = D, player 2 will take R since
2
2
(D; R) > VDL
(t) :
At t = t1 , player 1 is indi¤erent between U and D when player 2 takes R.
For slightly before t1 , since every inequality except for 1 (D; R) = VU1R (t1 ) is strict at
t1 ,
If xt2 = L, player 1 will take U .
If xt2 = R, player 1 will take U .
If xt1 = U , player 2 will take L.
If xt1 = D, player 2 will take R.
While players follow the above strategy, for
t<
t1 ,
If xt2 = L, player 1 will take U .
If xt2 = R, player 1 will take U .
If xt1 = U , player 2 will take L. Suppose player 2 is indi¤erent between L and R at t. If player 2
takes L, his expected payo¤ is de…nitely 2 (U; L). If player 2 takes R, it is a convex combination of
2 (U; R) and 2 (U; L), with a strictly positive weight on the former. Since 2 (U; R) < 2 (U; L) by
assumption, there does not exist such t.
If xt1 = D, there exists T such that player 2 will become indi¤erent between L and R.
We need to verify the fourth argument. Assuming that player 2 will take R after t, the
L is
8
(1 exp ( 2 (t t1 s))) 2 (U; L)
>
>
{z
}
> |
Z t t1
<
player 2 secondly m oves by t1
2
VDR (t) =
1 exp(
1 s)
+ exp ( 2 (t t1 s)) VU2R (t1 )
>
{z
}
|
0
>
>
|
{z
}
player 1 …rstly m oves by t :
1
+
exp (
|
1 (t
{z
t1 ))
}
player 2 do es not m ove by t1
2 (D; R)
player 1 does not m ove by t1
and the payo¤ after taking to L is
1
|
+
1
Z
+
(1
exp ( (
{z
2
+
2 ) (t
t1 )))
}
2
(U; L)
player 1 …rstly m oves by t1
T
t1
exp( 1 s)
| {z }
0
|
1
|
2
exp(
{z
2 s)
}
player 1 do es not m ove by splayer 2 m oves at s
{z
player 2 …rstly m oves by t1
+exp( (
|
1
+ 2 ) (t
{z
nob ody m oves by t1
2
t1 ))VDL
(t1 ) :
}
23
2
VDR
(t
s)ds
}
value of taking
9
>
>
>
=
>
>
>
;
ds
Note that
2
VDR
(t)
=
Z
t t1
1
exp(
1 s) 2
(U; L) ds
0
Z
t t1
exp (
1
1s
2
(t
t1
s))
2
(U; L)
VU2R (t1 ) ds
0
+ exp ( 1 (t t1 )) 2 (D; R)
= (1 exp( 1 (T t1 ))) 2 (U; L)
exp ( 1 (T t1 )) exp ( 2 (T
t1 ))
1
2
+ exp (
(T
1
t1 ))
2
2
VU2R (t1 )
(U; L)
1
(D; R) :
and the payo¤ after taking L is
1
+
Z T
1
+
(1
exp ( (
1
+
2 ) (T
2
t1
2
exp ( (
1
+
2 ) s)
0
+ exp ( (
1
=
1
1
(1
+
+
2 ) (T
1
+
+
exp (
1
(T
exp( (
1
+
2 ) (T
exp(
exp( (
1 2
exp( (
1
1 + 2 )(T
1+
(U; L)
2 (U; L)
t1 )) 2 (D; R)
1
1
(T
s
t1 )))
2
2
+
VU2R (t1 ) A ds
(U; L)
(U; L)
1
2
1 (T
2
t1 )) exp(
2 (T
1
2
(U; L)
!
t1 ))
2
t1 ))
(U; L)
VU2R (t1 )
(exp ( ( 1 + 2 ) (T t1 )) exp ( 1 (T t1 ))) 2 (D; R)
2
+ exp ( ( 1 + 2 ) (T t1 )) VDL
(t1 )
= (1 exp ( 1 (T t1 ))) 2 (U; L)
t1 ))
1 exp (
1 (T
2 exp (
exp ( ( 1 + 2 ) (T t1 )) +
2
(exp ( ( 1 + 2 ) (T t1 )) exp (
2
+ exp ( ( 1 + 2 ) (T t1 )) VDL
(t1 )
1
2
(t1 )
t1 )) exp(
2 )(T
s t1 )))
1 (T
s t1 )) exp(
2 (T s t1 ))
2
2
1
1 (T
exp(
+ exp (
t1 ))
t1 ))
(U; L)
1
2 ) (T
2
2
(1
@
2
t1 )) VDL
exp ( (
2
1
2
0
2
+
+
t1 )))
1
(T
t1 )))
2
2
(T
t1 ))
2
(U; L)
1
VU2R (t1 )
(D; R)
The di¤erence is given by
2
(U; L)
VU2R (t1 ) +
2
VDL
(t1 )
2 (D; R)
2
(U;
L)
V
(t
2
UR 1)
exp (
1
(t
t1 ))
exp (
2
(t
t1 ))
2
Therefore, since 2 (D; R) VDL
(t1 ) > 0, there exists unique T such that the di¤erence is positive
for t 2 [ T ; t1 ] and negative for t < T .
Therefore, for
t 2 ( T ; t1 ), we have
If xt2 = L, player 1 will take U .
If xt2 = R, player 1 will take U .
24
If xt1 = U , player 2 will take L.
If xt1 = D, player 2 will take R.
Since every incentive except for player 2 with x1 = D is strict, for a while, the equilibrium strategy is
If xt2 = L, player 1 will take U .
If xt2 = R, player 1 will take U .
If xt1 = U , player 2 will take L.
If xt1 = D, player 2 will take L.
As long as players follow the above strategy,
If xt2 = L, player 1 will take U .
If xt2 = R, player 1 will take U
If xt1 = U , player 2 will take L since, assuming player 2 will take L after that,
VU2L (t)
=
2 (U; L)
> (1 exp ( 2 (t
{z
|
t1 )))
}
2
(U; L) + exp (
t1 )) VU2R (t1 )
(t
player 2 can m ove by t1
=
VU2R
(t) :
since
2
(U; L) > VU2R (t1 ) :
If x1 = D, player 2 will take L from above.
We need to verify the second argument: at
payo¤ of taking U is given by
VU1R (t) = (1
|
exp (
{z
2
(t
t, if players follow the strategy above afterwards, the
t1 )))
}
1
player 2 can m ove by t1
25
(U; L) + exp (
2
(t
t1 )) VU1R (t1 ) :
On the other hand, the payo¤ of taking D is given by
Z
t T
exp ( ( 1 +
{z
|
0
+
player 1 m oves …rst by T
(U;R)
Z
t T
exp ( ( 1 +
|
{z
0
1
(t
1V
} UR
2 ) s)
s) ds
2
player 2 m oves …rst by T
(D;L)
2
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
6
4
2 ) s)
(1
|
}
exp (
(t
{z
T
1
s)))
}
player 1 m oves by T
(U;L)
+ exp (
|
1
(t T
{z
1
s))
}
3
(U; L)
player 1 do es not m ove by T
(D;L)
8
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
<
1
|1
0
+
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
:
B
B
B
B
B
B
t1 B
B
B
B
B
B
B
@
RT
0
+exp ( (
|
1
1+
2 ) (T
player 1 m oves …rst by t1
(U;L)
exp ( ( 1 +
{z
|
2)
)
t1 )))
}
1 (U; L)
2
}
player 2 m oves …rst by t1
(D;R)
exp (
|
1
(T
{z
t1
))
}
player 1 do es not m ove until t1
(D;R)
t1
exp ( 1 ) 1 VU1R
|
{z
}
1
1
(D; R)
(T
C
C
C
C
C
C
C
A
)d
player 1 m oves by t1
(D;R)
+exp ( (
|
The …rst term is
Z
2
B
B
B
B R
B
B + 0T
B
@
+ 2 ) (t
{z
(U; L)
+
exp ( (
{z
0
nob ody m oves by T
(D;R)
1
(1
1
+ 2 ) (T
{z
nob o dy m oves by t1
(D;L)
8
>
>
>
>
>
>
>
<
exp (
|
(T
{z
1
(t1 )
t1 ))VDL
}
t1 ))
}
1
1
player 1 do es not m ove by t1
(D;R)
R T t1
+ 0
exp ( 1 ) 1 VU1R
T ))
}>
>
>
>
>
>
>
:
|
{z
}
9
>
>
>
>
>
>
>
=
(D; R)
(T
>
>
>
>
>
>
>
;
)d
player 1 m oves by t1
(U;R)
1
C
C
C
C
C
C
C
Cd
C
C
C
C
C
A
:
t T
1
exp ( (
0
1 (U; L)
VU1R (t1 )
Z
1
+
2 ) s)
t T
1
exp ( (
1
+
2 ) s) exp (
2
(t
t1
s)) ds
(t
T )))
0
1
=
2
+
1
(1
exp ( (
1
+
2 ) (t
T )))
1
(U; L)
1
(U; L)
VU1R (t1 ) (exp (
2
(t
26
t1 ))
exp (
2
(t
t1 )
1
9
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
=
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
;
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7 ds
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
7
5
The second term is
1
(U; L)
2
+
1 (t
1
+ (exp (
(
(1
exp ( (
+
1
2 ) (t
T ))
exp ( (
exp ( ( 1 +
+ ( 1 (U; L)
1 (D; R))
1
+
2 ) (t
2 ) (T
exp(
2
T
t1
exp (
1
)
1
1 VU R
exp (
1
)
1 1
(T
t1 ))
1 (T
T )))
1 (U; L)
t1 ))+exp(
2
The calculation goes as follows: the integral with respect to
Z
T )))
2
2 (T
t1 ))
1
)
yields
)d
0
=
Z
T
t1
0
1
(U; L)
VU1R (t1 )
Z
(U; L)
T
t1
1
exp (
1
) exp (
2
(T
t1 )) d
0
=
1
(U; L) (1
1
(U; L)
exp (
1
(T
VU1R (t1 )
t1
)))
1
2
(exp (
1
27
1
(T
t1
))
exp (
2
(T
t1
))) :
Substituting this to the integral with respect to yields
0
1 (U; L)
Z T t1
B
( 1 (U; L)
t1
))
1 (D; R)) exp (
1 (T
B
exp ( ( 1 + 2 ) ) 2 @
exp ( 1 (T
t1
))
1
1
0
VU R (t1 ) 2 1
1 (U; L)
exp ( 2 (T
t1
))
=
1
(U; L)
+(
1
1
exp ( (
+ 2 ) (T
t1 ))
2+ 1
1 (D; R)) (exp ( ( 1 + 2 ) (T
2
(U; L)
1
exp( (
+
=
1
2
(U; L)
2
=
+(
1
(U; L)
+
1
(U; L)
1
2
+
+
(1
1
1
exp ( (
1
1
(1
+
exp (
(U; L)
1
+
exp(
1
1
(D; R) (exp ( (
1
(T
1 (T
+
1
1
2
(U; L)
1
1
=
1
+
B
(D; R) B
@
++(
VU1R
1
(1
exp ( (
1
(T
+
exp(
2
2
exp(
2 (T
1
+
2 ) (T
1 (T
+
(T
1
(1
1
1
t1 ))
+
1
(D; R))
2
exp(
2
(T
2 (T
1
t1 ))
t1 ))
(T
!
t1 )))
+
2 (T
t1 ))
t1 ))
2 ) (T
1
t1 )))
t1 ))
1
2 ) (T
exp (
2
exp (
(T
!
1
1
C
C
A
t1 )))
1
1
t1 ))
t1 )))
(D; R)
exp ( (
2 (T
t1 ))
t1 ))
2 ) (T
t1 ))+
2
exp ( (
2
(U; L)
2 ) (T
exp (
2
(t1 ) =
2+
+
t1 ))
2
1
exp(
1
exp ( ( 1 + 2 ) (T
t1 ))
exp ( 1 (T
t1 ))
1
2
t1 )) exp(
t1 )))
!
t1 ))
t1 )))
t1 )) + exp (
2
0
(U; L)
t1 ))+
1
(T
1 (T
1
2 ) (T
1 (T
2
2
1
1
exp (
(U; L)
since
exp(
+
exp ( (
2
+
1
t1 ))
t1 ))+
2 ) (T
2
=
2 ) (T
t1 )) + exp (
2
1
t1 )))
1
1
VU1R (t1 )
1 + 2 )(T
2 ) (T
exp ( (
exp ( (
exp (
t1 )) exp(
1
(D; R)) (exp ( (
VU1R (t1 )
+
exp( (
2
1
2
(U; L)
2 1
VU1R (t1 )
(U; L)
1
t1 ))
1 + 2 )(T
(T
t1 )) + exp (
2
28
1
2
(T
t1 ))
:
1
C
Cd
A
Substituting this into the integral with respect to s yields
Z
t T
exp ( (
+
1
2 ) s)
2
0
2
6
6
6
6
6 + exp (
4
=
1 (U; L)
(
(t
(t T
s))) 1 (U; L)
1
(1
exp
(
( 1 + 2 ) (T
t1 ))) 1 (U; L)
1+ 2
t1 ))) 1 (U; L)
+ 2 +2 1 (1 exp ( ( 1 + 2 ) (T
s))
exp(
t1 ))+exp(
1 (T
2 (T
>
+ ( 1 (U; L)
1 (D; R)) 2
>
2
1
>
:
1
+ exp ( ( 1 + 2 ) (T
t1 )) VDL (t1 )
T
t T
exp ( (
1
+
exp (
2 ) s)
+ 2 ) (T
1 (D; R))
t1 ))
2 ds
t1 )) (
1
3
9
7
>
>
7
>
= 7
7 ds
7
>
5
>
>
;
1
0
exp ( (
+ ( 1 (U; L)
+
Z
Z
1
8 (1
>
>
>
<
1 (U; L)
1 (D; R))
t1 ))+exp(
t1 ))
1 (T
2 (T
exp(
2
2
1
)
t T
exp ( (
+
1
2 ) s)
2
exp (
(t
1
T
s))
0
=
1
2
(U; L)
(1
+
1 (t
1
+ (exp (
(
exp ( (
+
1
2 ) (t
T )))
2
T ))
exp ( (
+ ( 1 (U; L)
exp ( (
+ 2 ) (T
1 (D; R))
+
1
2 ) (t
1
VDL
1 (U; L)
t1 ))+exp(
1 (T
t1 ))
1
T )))
exp(
2
2
(t1 )
2 (T
t1 ))
1
)
The third term is
exp ( (
=
=
exp ( (
1
+
1
+
+ exp ( (
1
exp ( (
1
+
+ exp ( (
1
2 ) (t
T ))
2 ) (t
+
T )) exp (
2 ) (t
2 ) (t
+
T ))
T ))
2 ) (t
since VU1R (t1 ) =
+
1
RT
exp ( 1 (T
t1
exp (
0
1
(T
t1 ))
+ VU1R (t1 )
1
T ))
(U; L)
2 exp (
1
1
t1 )) 1 (D; R)
)d
) 1 VU1R (T
(D; R)
exp (
1 (U; L) (1
1
(exp
(
(U;
L)
1
2
1
1
(T
t1 ))
1
2
1
exp (
2
t1 )))
1 (T
(T
t
))
exp (
1
1
(T
t1 ))
(
1
(D; R)
(D; R)
since
Z
T
t1
exp (
0
=
1
(U; L)
Z
T
1
)
1
1 VU R
)d
t1
exp (
0
+ VU1R (t1 )
(T
1 (U; L)
)
1
Z
1d
T
t1
exp (
)
1
1
exp (
2
(T
t1 )) d
0
=
1
(U; L) (1
+ VU1R (t1 )
exp (
1
1
(T
t1 )))
1
(U; L)
2
(exp (
1
29
1
(T
t1 ))
exp (
2
(T
t1 ))) :
2
(T
1
t1 )))
(U; L))
Therefore, in total, the value of taking D is
1
(1
2+
1
+
1
exp ( (
1
+
2 ) (t
T )))
1
(U; L)
1
VU1R (t1 ) (exp (
(U; L)
(U; L)
2
(1
+ 2
T ))
1 (t
2
(t
t1 ))
exp (
exp ( (
1
+
2 ) (t
T )))
exp ( (
1
+
2 ) (t
T )))
2
(t
t1 )
(t
T )))
(T
t1 ))
1
1
+ (exp (
(
exp ( (
+ ( 1 (U; L)
=
1
+ 2 ) (T
1 (D; R))
+ exp ( (
1
+
2 ) (t
T ))
exp ( (
1
+
2 ) (t
T ))
t1 ))
exp(
2
1
1
VDL
(t1 )
1 (U; L)
(T
t
))
exp(
t1 ))
2
1 (T
1
2
(U; L)
2 exp (
1
(T
1
t1 ))
1
2
1
)
exp (
2
(U; L)
VU1R (t1 ) (exp ( 2 (t t1 )) exp ( 2 (t t1 )
1 (U; L)
1 (t
+ (exp ( 1 (t T )) exp ( ( 1 + 2 ) (t T )))
)
(
1
(t1 )
exp ( ( 1 + 2 ) (T
t1 )) 1 (U; L) VDL
exp(
t1 )) exp(
t1 ))
2 (T
1 (T
+ ( 1 (U; L)
1 (D; R)) 2
2
1
(
1
(U; L)
1
(D; R))
(
1
(U; L)
1
(D; R))
1
exp ( (
1
+
2 ) (t
2
T ))
exp (
1
(T
t1 ))
1
2
1
exp (
2
(T
T )))
t1 ))
Subtracting VU1R (t) yields
(
1
(U; L)
+ (exp (
(
+
1
1
(t
(D; R))
exp(
exp ( (
T ))
exp ( (
2
exp (
1
2 (T
exp ( (
+
2 ) (T
t1 )) exp(
2
1
+
1
+
2 ) (t
2 ) (t
2
(t
T ))
t1 )
2
1 (t
1 (T
exp(
T )))
1
(U; L) VDL
(t1 )
t1 ))
( 1 (U; L)
1 (D; R))
t1 ))
1 (T
1
1
)
T ))
t1 ))
1
2
1
exp(
2 (T
t1 ))
:
Since
(exp ( 1 (t
exp ( ( 1 +
exp (
T ))
2 ) (T
2 (T
exp ( ( 1 + 2 ) (t T ))) > 0
1
t1 )) 1 (U; L) VDL
(t1 ) < 0
t1 )) exp ( 1 (T
t1 ))
< 0;
2
it su¢ ces to show that F (t)
F (t)
:
= exp (
1
0 with
2
exp ( (
(t
1
+
t1 )
1
(t
2 ) (t
T ))
T ))
2 exp (
1
At the limit where t goes to in…nity, we have
lim F (t) = 0.
t!1
Hence, it su¢ ces to show that
F 0 (t) > 0.
30
(T
t1 ))
1
2
1
exp (
2
(T
t1 ))
!
Since
F 0 (t)
1+ 2
=
exp (
2
+ exp ( (
=
+
if
2
>
(
1,
1
+
1
t1 )
+
exp ( 2 (t
2 exp (
1
2 ) (t
(t
2
T ))
t1 )
1 (t
T )
2 (t
F 0 (t)
=
2
exp (
2
(t
2
+
=
as desired. Symmetrically, if
exp (
1
(T
T ))
t1 ))
1 (t
1
2
1
exp (
2 exp (
(t
2 (t
2
2
<
2
1,
t1 )
1
T )
t1 )
(t t1 ))
T )) +
1 (t
t1 ))
1
2
1
exp (
1
(t
exp (
T )
2
2
(T
(t
t1 ))
t1 ))
;
(t
T )) +
1
1
exp (
exp (
2 exp (
1
2
(t
(t
1 (t
1
t1 )
1
(t
T ))
T )
t1 )
(t t1 ))
T )) > 0
2 (t
2
we have
(
1)
2
1
to
T ))
we have
1)
2
(t
+
F 0 (t) < 0:
2
When 2 = 1 , our model allows the continuity of the cases between
1 . Therefore, the same result holds.
31
2
=
1
and limit where
2
converves