Mailath - Economics703 Microeconomics II Modelling Strategic Behavior
Mailath - Economics703 Microeconomics II Modelling Strategic Behavior
George J. Mailath
Department of Economics
University of Pennsylvania
1
Copyright January 18, 2016 by George J. Mailath.
Contents
Contents i
i
ii CONTENTS
3.2 Purification . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
3.3 Auctions and Related Games . . . . . . . . . . . . . . . . 57
3.4 Games of Incomplete Information . . . . . . . . . . . . . 70
3.5 Higher Order Beliefs and Global Games . . . . . . . . . . 74
3.6 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
4 Nash Equilibrium 85
4.1 Existence of Nash Equilibria . . . . . . . . . . . . . . . . . 85
4.2 Foundations for Nash Equilibrium . . . . . . . . . . . . . 92
4.2.1 Boundedly Rational Learning . . . . . . . . . . . . 93
4.2.2 Social Learning (Evolution) . . . . . . . . . . . . . . 94
4.2.3 Individual learning . . . . . . . . . . . . . . . . . . 100
4.3 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
6 Signaling 135
6.1 General Theory . . . . . . . . . . . . . . . . . . . . . . . . . 135
6.2 Job Market Signaling . . . . . . . . . . . . . . . . . . . . . 138
6.2.1 Full Information . . . . . . . . . . . . . . . . . . . . 139
6.2.2 Incomplete Information . . . . . . . . . . . . . . . 139
6.2.3 Refining to Separation . . . . . . . . . . . . . . . . 144
6.2.4 Continuum of Types . . . . . . . . . . . . . . . . . 145
6.3 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
9 Bargaining 223
9.1 Axiomatic Nash Bargaining . . . . . . . . . . . . . . . . . 223
9.1.1 The Axioms . . . . . . . . . . . . . . . . . . . . . . . 223
9.1.2 Nashs Theorem . . . . . . . . . . . . . . . . . . . . 224
9.2 Rubinstein Bargaining . . . . . . . . . . . . . . . . . . . . . 225
9.2.1 The Stationary Equilibrium . . . . . . . . . . . . . 226
9.2.2 All Equilibria . . . . . . . . . . . . . . . . . . . . . . 228
9.2.3 Impatience . . . . . . . . . . . . . . . . . . . . . . . 230
9.3 Outside Options . . . . . . . . . . . . . . . . . . . . . . . . 230
9.3.1 Version I . . . . . . . . . . . . . . . . . . . . . . . . . 231
9.3.2 Version II . . . . . . . . . . . . . . . . . . . . . . . . 233
9.4 Exogenous Risk of Breakdown . . . . . . . . . . . . . . . 236
9.5 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237
10 Appendices 243
10.1 Proof of Theorem 2.4.1 . . . . . . . . . . . . . . . . . . . . 243
10.2 Trembling Hand Perfection . . . . . . . . . . . . . . . . . 246
10.2.1 Existence and Characterization . . . . . . . . . . . 246
10.2.2 Extensive form trembling hand perfection . . . . 249
10.3 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 250
References 251
Index 257
iv CONTENTS
Chapter 1
II
Confess Dont confess
I Confess 6, 6 0, 9
Dont confess 9, 0 1, 1
E S
E 2, 2 1, 3
S 3, 1 0, 0
1
Copyright January 18, 2016 by George J. Mailath
1
2 CHAPTER 1. NORMAL AND EXTENSIVE FORM GAMES
2
(yxx) (yxy) (yzx) (yzy) (zxx) (zxy) (zzx) (zzy)
x 0,0 0,0 0,0 0,0 1,2 1,2 1,2 1,2
1 y 0,0 0,0 2,1 2,1 0,0 0,0 2,1 2,1
z 1,2 2,1 1,2 2,1 1,2 2,1 1,2 2,1
Figure 1.1.1: The normal form of the voting by veto game in Example
1.1.6.
Ui (si0 , si ) Ui (si00 , si ),
0
and si Si ,
Ui (si0 , si
0
) > Ui (si00 , si
0
).
A strategy is said to be strictly or weakly undominated if it is not
strictly or weakly dominated by some other strategy. If the adjec-
tive is omitted from dominated (or undominated), weak is typically
meant (but not always, unfortunately). A strategy is weakly domi-
nant if it weakly dominates every other strategy.
January 18, 2016 5
But this implies that si0 cannot weakly dominate si , and so si is the
only weakly dominant strategy.
Remark 1.1.2 (Warning). There is also a notion of dominant strat-
egy:
Definition 1.1.4. si0 is a dominant strategy if si00 Si , si Si ,
Ui(si0 , si ) Ui (si00 , si ).
Two cases:
P P
vi + ji vj , if vi + ji vj 0,
Agent is payoff(vi ) = P
0, if vi + ji vj < 0.
P
Dominant strategy to announce vi = vi : If vi + ji v Pj > 0, an-
nouncing vi = vi ensures good is provided, while if vi + ji vj < 0,
announcing vi = vi ensures good is not provided. Moreover, con-
ditional on provision, announcement does not affect payoffnote
similarity to second price auction.
No payments if no provision, but payments large if provision:
P P
Total payments to agents when provision = i ji vj = (n
P
1) i vi . Taxing agent i by an amount independent
P of is behavior
has no impact, so tax i the amount max{ ji vj , 0}. Result is
P P
vi , if j vj 0 and ji vj 0,
P P P
vi + ji vj , if j vj 0 and ji vj < 0,
payoff to i = P P P
ji vj , if j vj < 0 and ji vj 0,
P P
0, if j vj < 0 and ji vj < 0.
L M R
T 1, 0 1, 2 0, 1
B 0, 3 0, 1 2, 0
2
(z, z, x)
1 y 2,1
z 1,2
January 18, 2016 9
: T \{t0 } A,
(a) t 6 t 0 , t 0 6 t,
(b) A(t) = A(t 0 ) A(h), and
(c) perfect recall (every player knows whatever he knew pre-
viously, including own previous actions).6
6. Payoffs, ui : Z R.
1
t0
x
y z
2 2
2
t1 t2 t3
y z x z x y
t4 t5 t6 t7 t8 t9
0 1 0 2 1 2
0 2 0 1 2 1
Figure 1.3.1: The extensive form for Example 1.1.6. An arrow connects a
node with its immediate predecessor, and is labelled by the
action associated with that node.
I
t0
S E
II
t1 t2
S E S E
t3 t4 t5 t6
0 3 1 2
0 1 3 2
Figure 1.3.2: The extensive form for the prisoners dilemma. Player IIs
information set is indicated by the dashed line.
14 CHAPTER 1. NORMAL AND EXTENSIVE FORM GAMES
Go II Go I Go1 II Go1 z5
I
The strategy Stop for player I, for example, in the reduced normal
form should be interpreted as the equivalence class of extensive
form strategies {Stop,Stop1 , Stop,Go1 }, where the equivalence rela-
tion is given by strategic equivalence.
January 18, 2016 15
The importance of the reduced normal form arises from the re-
duction illustrated in Example 1.3.3: Whenever a players strategy
specifies an action that precludes another information set owned
by the same player, that strategy is strategically equivalent to other
strategies that only differ at such information sets. As for Exam-
ple 1.3.3, this is true for arbitrary specifications of payoffs at the
terminal nodes.7
When describing the normal form representation of an extensive
form, it is common to (and we will typically) use the reduced normal
form.
An extensive form strategy always has the form given by (1.3.1),
while a normal form strategy may represent an equivalence class
of extensive form strategies (and a reduced normal form strategy
always does).
1.4 Problems
1.4.1. Describe a social choice rule that is unanimous, nondictatorial, and
strategy proof when there are two alternatives.
1.4.2. Consider the following social choice rule over the set X = {x, y, z}.
There is an exogenously specified order x y z, and define
S(X 0 ) {a X 0 : a a0 a0 X 0 \ {a}}. Then,
1.4.4. (a) Prove that the order of deletion does not matter for the process
of iterated deletion of strictly dominated strategies in a finite
game (Remark 1.2.1 shows that strictly cannot be replaced by
weakly).
(b) Show that the order of deletion matters for the process of iter-
ated deletion of strictly dominated strategies for the following
infinite game: S1 = S2 = [0, 1] and payoffs
si , if si < 1,
ui (s1 , s2 ) = 0, if si = 1, sj < 1,
s , if s = s = 1.
i i j
(a) Prove that each player is information sets Hi are strictly par-
tially ordered by the follows relation (recall footnote 3).
(b) Perfect recall implies that the set of is information sets satisfy
Property 2(c) of Definition 1.3.1, i.e., for all h, h0 , h00 Hi , if
h h00 and h0 h00 then either h h0 or h0 h. Give an
intuitive argument.
(c) Give an example showing that the set of all information sets is
not similarly strictly partially ordered by .
(d) Prove that if h0 h for h, h0 Hi , then for all x h, there
exists x 0 h0 such that x 0 x. (In other words, an individual
players information is refined through play in the game. Prove
should really be in quotes, since this is trivial.)
Chapter 2
ui (si , si
) ui (si , si ).
17
18 CHAPTER 2. A FIRST LOOK AT EQUILIBRIUM
I
t0
L R
II II
t1 t2
` r `0 r0
t3 t4 t5 t6
2 4 1 3
3 2 0 1
II
``0 `r 0 r `0 rr0
I L 2,3 2,3 4,2 4,2
R 1,0 3,1 1,0 3,1
Figure 2.1.2: The normal form of the extensive form in Figure 2.1.1.
January 18, 2016 19
s (s ).
20 CHAPTER 2. A FIRST LOOK AT EQUILIBRIUM
q2
ac
(q1 , q2 )
(a c)/2 1
(a c)/2 ac q1
Figure 2.1.3: The reaction (or best reply) functions for the Cournot game.
a c q2
q1 (q2 ) = .
2
1
i (qj ) = max{ (a c qj ), 0}.
2
q1 =1 (q2 ),
q2 =2 (q1 ).
January 18, 2016 21
Example 2.1.4 (Example 1.1.7 cont., sealed bid 2nd price auction).
Suppose v1 < v2 , and the valuations are commonly known. There
are many Nash equilibria: Bidding vi for each i is a Nash equilib-
rium (of course?). But so is any bidding profile (b1 , b2 ) satisfying
b1 < b2 , b1 v2 and v1 b2 (Why? Make sure you understand
why some inequalities are weak and some are strict). Are there any
other equilibria?
Example 2.1.5 (Example 1.1.3 cont., sealed bid 1st price auction).
Suppose v1 = v2 = v, and the valuations are commonly known.
The unique Nash equilibrium is for both bidders to bid bi = v. But
this eq is in weakly dominated strategies. But what if bids are in
pennies?
22 CHAPTER 2. A FIRST LOOK AT EQUILIBRIUM
Entrant
Out In
Incumbent
0
4 Fight Accommodate
1 1
1 2
Kidnapper
kill release
Victim
1
100 dont reveal reveal
10 5
1 2
Go II Go I Go1 10
I 1, 000
Stop Go
Stop 1, 0 1, 0
Go,Stop1 0, 10 100, 1
Go,Go1 0, 10 10, 1000
Figure 2.2.4: The reduced normal form for the short centipede in Figure
2.2.3
R1 II r I R2 10
I 1, 000
D1 d D2
15 20 9
0 10 1
L R
II
2
2 ` r
0 2
1 1
Figure 2.2.7: An example with ties. The backward induction solutions are
L`, Lr , and Rr . But R is weakly dominated for I.
Example 2.2.5 (An example with ties). If the game has ties, then
iteration deletion of weakly dominated strategies can be stronger
than backward induction (see Figure 2.2.7).
3
This property is generic in extensive form payoffs in the following sense: Fix
an extensive game tree (i.e., everything except the assignment of payoffs to ter-
minal nodes z Z). The space of games (with that tree) is then the space of
all payoff assignments to the terminal nodes, Rn|Z| . Note that, viewed this way,
the space of games is a finite dimensional Euclidean space. The set of payoff
assigments that violate the property of no ties is a subset of a closed Lebesgue
measure zero subset of Rn|Z| .
28 CHAPTER 2. A FIRST LOOK AT EQUILIBRIUM
I
L
R
2
1 II
1
T B
III
` r ` r
0 0 3 0
0 0 1 1
0 1 1 0
Define S(t) {t 0 T : t t 0 }.
h T t h T t.
E S P
E 2, 2 1, 3 1, 1
S 3, 1 0, 0 1, 1
P 1, 1 1, 1 2, 2
L C R
T 4, 6 0, 0 9, 0
.
M 0, 0 6, 4 0, 0
B 0, 0 0, 0 8, 8
(T , L) and (M, C) are both Nash eq of the stage game. The follow-
ing profile (s1 , s2 ) of the once-repeated game with payoffs added is
subgame perfect:
s11 = B,
(
M, if x = B,
s12 (x, y) =
T, if x B,
30 CHAPTER 2. A FIRST LOOK AT EQUILIBRIUM
I
L
R
2
0 I
T B
II
` r ` r
1 4 0 5
1 0 0 1
s21 = R,
(
2 C, if x = B, and
s2 (x, y) =
L, if x B.
s11 = B,
s12 (y) = M, for all y,
s21 = R,
(
C, if x = B, and
s22 (x) =
L, if x B.
I
L
2
0 T B
II
` r ` r
1 4 0 5
1 0 0 1
The game has three Nash eq: (RB, r ), (LT , `), and (LB, `). Note
that (LT , `), and (LB, `) are distinct extensive form strategy pro-
files.
The only subgame perfect equilibrium is (RB, r ).
But, (L, `) also subgame perfect in the extensive form in Figure
2.3.3.
Both games have the same reduced normal form, given in Figure
2.3.4.
` r
L 2, 0 2, 0
T 1, 1 4, 0
B 0, 0 5, 1
Figure 2.3.4: The reduced form for the games in Figures 2.3.2 and 2.3.3.
32 CHAPTER 2. A FIRST LOOK AT EQUILIBRIUM
2.4 Mixing
2.4.1 Mixed Strategies and Security Levels
Example 2.4.1 (Matching Pennies). A game with no Nash eq:
H T
H 1, 1 1, 1
T 1, 1 1, 1
January 18, 2016 33
The greatest payoff that player 1 can guarantee himself may ap-
pear to be 1 (the unfortunate result of player 2 correctly antici-
pating 1s choice).
But suppose that player 1 flips a fair coin so that player 2 cannot
anticipate 1s choice. Then, 1 should be able to do better.
Writing X Y
ui (si , i ) = ui (si , si ) j (sj ),
si Si ji
we then have
X
ui (i , i ) = ui (si , i )i (si ).
si Si
v i = sup Q
inf ui (i , i ).
i (Si ) i ji (Sj )
L R
T 3 0
M 2 2
B 0 3
L R
T 5 0
M 2 2
B 0 5
ui (i , si ) > ui (si0 , si ) si Si .
January 18, 2016 35
For more than two players, it is possible that a strategy can fail
to be a best reply to any mixed profile for the other players, and
yet that strategy is not strictly dominated (see Problem 2.6.17 for
an example).
s1
x(, s2 )
x(, s2 )
x(, s20 )
X
|S |1 s1
R 1
x(, s2 )
|S |1
Figure 2.4.1: The sets X and R 1 .
ui (i , i
) ui (i , i ). (2.4.2)
ui (i , i
) ui (si , i ),
since X
ui (i , i )= ui (si , i )i (si ) ui (si0 , i
),
si Si
ui (i , i
) ui (si , i ).
Example 2.4.3.
L R
T 2, 1 0, 0
B 0, 0 1, 1
{1}, if p > 12 ,
2 (p) = [0, 1], if p = 12 ,
1
{0}, if p < 2 .
1
2
1
3
0 1 1 p
2
L R
U 2, 1 0, 0
D 2, 0 1, 1
A a b
A 2, 2 0, 0 0, 0
a 0, 0 0, 0 2, 2
b 0, 0 2, 2 0, 0
` r
T 2, 2 0, 0
B 0, 0 1, 1
January 18, 2016 43
s11 = B,
(
M, if x = B,
s12 (x, y) =
T , if x B,
s21 = R,
(
C, if x = B, and
s22 (x, y) =
L, if x B.
A B
A 9, 9 0, 5
B 5, 0 7, 7
2.6 Problems
2.6.1. Suppose {(Si , Ui )n
i=1 } is a normal form game, and s1 S1 is a weakly
dominated strategy for player 1. Let S10 = S1 \ {s1 }, and Si0 = Si for
i 1. Suppose s is a Nash equilibrium of {(Si0 , Ui )n
i=1 }. Prove that s
is a Nash equilibrium of {(Si , Ui )n
i=1 }.
2.6.3. Consider (again) the Cournot example (Example 1.1.4). What is the
Nash Equilibrium of the n-firm Cournot oligopoly? [Hint: To cal-
culate the equilibrium, first solve for the total output in the Nash
equilibrium.] What happens to both individual firm output and total
output as n approaches infinity?
2.6.4. Consider now the Cournot duopoly where inverse demand is P (Q) =
a Q but firms have asymmetric marginal costs: ci for firm i, i =
1, 2.
(a) What is the Nash equilibrium when 0 < ci < a/2 for i = 1, 2?
What happens to firm 2s equilibrium output when firm 1s
costs, c1 , increase? Can you give an intuitive explanation?
(b) What is the Nash equilibrium when c1 < c2 < a but 2c2 > a+c1 ?
2.6.5. Consider the following Cournot duopoly game: The two firms are
identical. The cost function facing each firm is denoted by C(q),
is continuously differentiable with C(0) = 0, C 0 (0) = 0, C 0 (q) >
0 q > 0. Firm i chooses qi , i = 1, 2. Inverse demand is given
January 18, 2016 45
2.6.6. (easy) Prove that the information set containing the initial node of a
subgame is necessarily a singleton.
2.6.7. In the canonical Stackelberg model, there are two firms, I and II, pro-
ducing the same good. Their inverse demand function is P = 6 Q,
where Q is market supply. Each firm has a constant marginal cost
of $4 per unit and a capacity constraint of 3 units (the latter restric-
tion will not affect optimal behavior, but assuming it eliminates the
possibility of negative prices). Firm I chooses its quantity first. Firm
II, knowing firm Is quantity choice, then chooses its quantity. Thus,
firm Is strategy space is S1 = [0, 3] and firm IIs strategy space is
S2 = {2 | 2 : S1 [0, 3]}. A strategy profile is (q1 , 2 ) S1 S2 ,
i.e., an action (quantity choice) for I and a specification for every
quantity choice of I of an action (quantity choice) for II.
(a) What are the outcome and payoffs of the two firms implied by
the strategy profile (q1 , 2 )?
(b) Show that the following strategy profile does not constitute a
Nash equilibrium: ( 12 , 2 ), where 2 (q1 ) = (2 q1 )/2. Which
firm(s) is (are) not playing a best response?
46 CHAPTER 2. A FIRST LOOK AT EQUILIBRIUM
(c) Prove that the following strategy profile constitutes a Nash equi-
1 3 1
librium: ( 2 , 2 ), where 2 (q1 ) = 4 if q1 = 2 and 2 (q1 ) = 3 if
q1 12 , i.e., II threatens to flood the market unless I produces
exactly 12 . Is there any other Nash equilibrium which gives the
outcome path ( 21 , 34 )? What are the firms payoffs in this equi-
librium?
(d) Prove that the following strategy profile constitutes a Nash equi-
librium: (0, 2 ), where 2 (q1 ) = 1 if q1 = 0 and 2 (q1 ) = 3 if
q1 0, i.e., II threatens to flood the market unless I produces
exactly 0. What are the firms payoffs in this equilibrium?
(e) Given q1 [0, 2], specify a Nash equilibrium strategy profile
in which I chooses q1 . Why is it not possible to do this for
q1 (2, 3]?
(f) What is the unique backward induction solution of this game?
2.6.8. Player 1 and 2 must agree on the division of a pie of size 1. They
are playing a take-it-or-leave-it-offer game: Player 1 makes an offer
x from a set S1 [0, 1], which player 2 accepts or rejects. If player 2
accepts, the payoffs are 1 x to player 1 and x to player 2; if player
2 rejects, both players receive a zero payoff.
L R
I II
L0 R0 `0 r0
II
2 0
` r ` r 1 0
3 4 1 5
1 0 0 1
2.6.10. Consider the following game G between two players. Player 1 first
chooses between A or B, with A giving payoff of 1 to each player,
and B giving a payoff of 0 to player 1 and 3 to player 2. After player
1 has publicly chosen between A and B, the two players play the
following bimatrix game (with 1 being the row player):
L R
U 1, 1 0, 0
D 0, 0 3, 3
Payoffs in the overall game are given by the sum of payoffs from 1s
initial choice and the bimatrix game.
(a) What is the extensive form of G?
(b) Describe a subgame perfect equilibrium strategy profile in pure
strategies in which 1 chooses B.
(c) What is the reduced normal form of G?
(d) What is the result of the iterated deletion of weakly dominated
strategies?
48 CHAPTER 2. A FIRST LOOK AT EQUILIBRIUM
2.6.12. Prove that player i s security level (Definition 2.4.2) is also given by
vi = sup inf
Q
ui (i , si ).
i (Si ) si ji Sj
Prove that
v i sup inf
Q
ui (si , si ),
si Si si ji Sj
2.6.13. Suppose the 2 2 normal form game G has a unique Nash equi-
librium, and each players Nash equilibrium strategy and security
strategy are both completely mixed.
2.6.15. (a) Prove that Lemma 2.4.1 also holds for mixed strategies, i.e.,
prove that 1 (S1 ) is strictly dominated by some other strat-
egy 10 (i.e., u1 (10 , s2 ) > u1 (1 , s2 ), s2 S2 ) if and only if 1
is not a best reply to any mixture 2 (S2 ).
January 18, 2016 49
L R
T 5, 0 0, 1
C 2, 6 4, 0
B 0, 0 5, 1
1 1
(b) For the game illustrated in Figure 2.6.2, prove that 2 T + 2 B
is not a best reply to any mixture over L and R. Describe a
strategy that strictly dominates it.
2.6.16. Suppose {(S1 , u1 ), (S2 , u2 )} is a two player finite normal form game
and let Sb2 be a strict subset of S2 . Suppose s10 S1 is not a best
reply to any beliefs with support Sb2 . Prove that there exists > 0
such that s10 is not a best reply to any beliefs (S2 ) satisfying
(Sb2 ) > 1 . Is the restriction to two players important?
2.6.17. Consider the three player game in Figure 2.6.3 (only player 3s pay-
offs are presented).
2.6.19. Suppose player 1 must publicly choose between the game on the left
and the game on the right in Figure 2.6.4 (where player 1 is choosing
rows and player 2 is choosing columns. Prove that this game has no
Nash equilibrium in pure strategies. What is the unique subgame
perfect equilibrium (the equilibrium is in behavior strategies).
` r ` r ` r
t 3 0 t 2 4 t 0 0
b 0 0 b 4 2 b 0 3
L M R
Figure 2.6.3: The game for Problem 2.6.17. Player 1 chooses rows (i.e.,
s1 {t, b}), player 2 chooses columns (i.e., s2 {`, r }), and
player 3 chooses matrices (i.e., s3 {L, M, R}). Only player
3s payoffs are given.
t h T H
t 1, 1 1, 1 T 3, 1 1, 1
h 1, 1 1, 1 H 1, 1 1, 1
L C R
T 2, 2 1, 1 0, 0
B 2, 0 0, 0 4, 2
` r ` r
T 1, 1, 1 1, 0, 1 T 1, 1, 0 0, 0, 0
B 1, 1, 1 0, 0, 1 B 0, 1, 0 1, 0, 0
L R
Figure 2.6.6: The game for Problem 2.6.21(b). Player 1 chooses a row (T or
B), player 2 chooses a column (` or r ), and player 3 chooses
a matrix (L or R). In each cell, the first payoff is player 1s,
the second is player 2s, and the third is player 3s.
(a) For the game in Figure 2.6.5, prove that T L is trembling hand
perfect by explicitly describing the sequence of trembles.
(b) The game in Figure 2.6.6 has an undominated Nash equilibrium
that is not trembling hand perfect. What is it?
n o
2.6.22. Say a profile is robust to all trembles if for all sequences k of
k
completely mixed strategy profiles converging to , i is eventually
k
a best reply to every i in the sequence.6
6 k
By eventually, I mean there exists K such that i is a best reply to i for all
k > K. Note that K depends on the sequence.
52 CHAPTER 2. A FIRST LOOK AT EQUILIBRIUM
L M R
T 1, 5 2, 3 0, 0
B 1, 5 0, 0 3, 2
(a) Prove that no profile in the game in Figure 2.6.7 is robust to all
trembles.
(b) There is an extensive form with a nontrivial subgame that has
as its normal the game in Figure 2.6.7. This extensive form
game has two subgame perfect equilibria. What are they? Com-
pare with your analysis of part 2.6.22(a).
2.6.23. It is not true that every trembling hand perfect equilibrium of a nor-
mal form game induces subgame perfect behavior in an extensive
game with that normal form. Illustrate using the game in Figure
2.6.8.
L M R
T 1, 5 2, 6 0, 0
B 1, 5 0, 3 3, 2
+ (1 )[(a qL q2 ) cL ]qL .
1
Copyright January 18, 2016 by George J. Mailath
53
54 CHAPTER 3. GAMES WITH NATURE
+ (1 )[(a qL q2 ) c2 ]q2
= arg max [(a qH (1 )qL q2 ) c2 ]q2
q2
1
= a c2 qH (1 )qL .
2
Solving,
a 2cH + c2 1
qH = + (cH cL )
3 6
a 2cL + c2
qL = (cH cL )
3 6
a 2c2 + cH + (1 ) cL
q2 =
3
3.2 Purification
Player is mixed strategy i in a complete information game G is
said to be purified if in an incomplete information version of G
(with player is type space given by Ti ), that players behavior can
be written as a pure strategy si : Ti Ai such that
i (ai ) = Pr{si (ti ) = ai },
January 18, 2016 55
A B
A 9, 9 0, 5
B 5, 0 7, 7
Example 3.2.1. The game in Figure 3.2.1 has two strict pure strat-
egy Nash equilibria and one symmetric mixed strategy Nash equi-
librium. Let p = Pr {A}, then (using Lemma 2.4.2)
9p = 5p + 7 1 p
a 9p = 7 2p
7
a 11p = 7 a p = .
11
Trivial purification: give player i a payoff irrelevant type ti where
ti U([0, 1]), and t1 and t2 are independent. Then, the mixed
strategy eq is purified by many pure strategy eq in the incomplete
information game, such as
(
B, if ti 4/11,
si (ti ) =
A, if ti 4/11.
A B
A 9 + t1 , 9 + t2 0, 5
B 5, 0 7, 7
Figure 3.2.2: The game G() for Example 3.2.1. Player i has type ti , with
ti U([0, 1]), and t1 and t2 are independent.
while from B is
U1 (B, t1 , s2 ) = 5 Pr t2 t2 + 7 Pr t2 < t2
= 5(1 t2 ) + 7t2
= 5 + 2t2 .
(9 + t1 )(1 t2 ) 5 + 2t2
i.e.,
11t2 4
t1 .
(1 t2 )
Thus the best reply to the cutoff strategy s2 is a cutoff strategy with
t1 = (11t2 4)/(1 t2 ).2 Since the game is symmetric, try for a
symmetric eq: t1 = t2 = t. So
11t 4
t = ,
(1 t)
or
t 2 + (11 )t 4 = 0. (3.2.1)
2
Indeed, even if player 2 were not following a cutoff strategy, player 1s best
reply is a cutoff strategy.
January 18, 2016 57
Let t() denote the value of t satisfying (3.2.1). Note first that
t(0) = 4/11, and that writing (3.2.1) as g(t, ) = 0, we can apply the
implicit function theorem to conclude that for > 0 but close to 0,
the cutoff type t() is close to 4/11, the probability of the mixed
strategy eq in the unperturbed game. In other words, for small,
there is a symmetric equilibrium in cutoff strategies, with t (0, 1).
This equilibrium is not only pure, but almost everywhere strict!
The interior cutoff equilibrium of G() approximates the mixed
strategy equilibrium of G(0) in the following sense: Let p() be
the probability assigned to A by the symmetric cutoff equilibrium
strategy of G(). Then p(0) = 7/11 and p() = 1 t(). Since we
argued in the previous paragraph that t() 4/11 as 0, we
have that for all > 0, there exists > 0 such that
i.e.,
(v v) 0 (v) + (v) = v.
But,
d
v v (v) = v v 0 (v) + (v) ,
dv
so
v2
v v (v) = + k,
2
where k is a constant of integration. Moreover, evaluating both
sides at v = v shows that k = v 2 /2, and so
v2 v2
v v (v) =
2
1
= (v) = v +v . (3.3.2)
2
Summarizing the calculations till this point, I have shown that
if ( , ) is a Nash equilibrium in which is a strictly increasing
and differentiable function, and (v) is interior (which here means
strictly positive), then it is given by (3.3.2).
It remains to verify the hypotheses. It is immediate that the
function in (3.3.2) is strictly increasing and differentiable. More-
over, for v > 0, (v) is strictly positive. It remains to verify the
optimality of bids (i.e., we still need to verify that bidding b1 =
(v1 ) is optimal when bidder 2 is bidding according to ; given
the symmetry, this is enough). Bidder 1s interim payoff is (since
1
(v2 ) [v, v + 2 ] for all v2 [v, v + 1])
U1 (b1 , v1 ; 2 ) = (v1 b1 )F2 (21 (b1 ))
1
v1 b1 ,
if b1 > v + 2 ,
1
= (v1 b1 )(2b1 2v), if b1 [v, v + 2 ],
0, if b1 < v.
1
This function is strictly concave for b1 [v, v + 2 ], and it is clearly
not optimal to bid outside the interval, and so
bidding b = (v )
1 1
is optimal.
As an illustration of the kind of arguments that are useful, I
now argue that every Nash equilibrium must be in nondecreasing
strategies.
January 18, 2016 61
Note that
and so substituting,
1 0
(v v100 ) Pr{2 (v2 ) = b10 } + (v10 v100 ) Pr{2 (v2 ) < b10 }
2 1
1 0
(v1 v100 ) Pr{2 (v2 ) = b100 } + (v10 v100 ) Pr{2 (v2 ) < b100 },
2
and simplifying (and dividing by (v10 v100 ) < 0) we get
This implies
That is, bidder 2 does not make a bid between b100 and b10 , and there
are no ties at b10 or b100 . A bid of b10 and b100 therefore wins with the
same probability. But this implies a contradiction: Since b10 wins
with positive probability, v10 strictly prefers to win with the same
probability at the strictly lower bid of b100 .
0 = F n1 ( 1 (bi ))
d 1 (bi )
+ (vi bi )(n 1)F n2 ( 1 (bi ))f ( 1 (bi )) ,
dbi
and simplifying (similarly to (3.3.1)), we get
that is,
d
(v)F n1 (v) = v(n 1)F n2 (v)f (v),
dv
or (where the constant of integration is zero, since F (v) = 0),
Zv
1
(v) = n1 x dF n1 (x).
F (v) v
That is, each bidder bids the expectation of the maximum of all the
other bidders valuation, conditional on that valuation being less
than his (i.e., conditional on his value being the highest). Equiv-
alently, the bidder bids the expected value of the (n 1)th order
statistic of values, conditional on his value being the nth order
statistic.
v 0 + 1/2 t+1
private value (t) = (v 0 ) = = > t = common value (t).
2 2
This illustrates the winners curse: E[v | t1 ] > E[v|t1 , winning]. In
particular, in the equilibrium just calculated,
E[v | t1 , winning] = E[t1 + t2 | t1 , t2 < t1 ]
Z t1
1
= t1 + t2 dt2
Pr {t2 < t1 } 0
1 h 2
it1 3t1
= t1 + (t2 ) /2 = ,
t1 0 2
January 18, 2016 65
1
while E[v | t1 ] = t1 + 2
> 3t1 /2 (recall t1 [0, 1]).
U1 (s1 , t1 ; 2 )
Z
1
= s1 Pr t2 2 (s1 ) + (t1 2 (t2 )) f2 (t2 ) dt2
{t2 <21 (s1 )}
Z 21 (s1 )
1
= s1 1 F2 2 (s1 ) + (t1 2 (t2 )) f2 (t2 ) dt2 .
0
tf (t)
0 (t) = .
1 F (t)
Since (0) = 0, Zt
f ()
(t) = d.
0 1 F ()
If f (t) = e , then F (t) = 1 et , and
t
Zt
e
(t) = d = t 2 /2.
0 e
68 CHAPTER 3. GAMES WITH NATURE
So,
if pb as ,
0,
pb as
Pr vs : pb p
es (vs ) = , if as pb as + cs ,
cs
1, if pb as + cs ,
January 18, 2016 69
and so
pb as
E p
es (vs ) vs : pb p
es (vs ) = as + cs E vs vs : vs
cs
(and if as pb as + cs )
1 pb as pb + as
= as + cs = .
2 cs 2
So,
not defined, if pb < as ,
pb +as
es (vs ) vs : pb p
E p es (vs ) = , if as pb as + cs ,
2
a + 1 c , if pb as + cs .
s 2 s
and
0, if pb as ,
pb as 3 1
Ub pb , vb ; p
es , = vb 4 pb 4 as , if as pb as + cs ,
cs
v 1 p 1 a 1 c , if pb as + cs .
b 2 b 2 s 4 s
L R L R
T 1, 1 0, 0 T 1, x 0, 1
B 0, 1 1, 0 B 0, 0 1, 1
1 2
p
es (vs ) = + vs
4 3
1 2
p
eb (vb ) = + vb
12 3
Efficient trade requires trade if vs < vb , and no trade if vs > vb .
Under linear equilibrium, trade occurs iff p es (vs ) p
eb (vb ), which
requires
1
vs + vb .
4
n o
1
Thus, for valuations in the set (vs , vb ) | vs < vb < vs + 4 , trade is
efficient but does not occur in equilibrium.
Note: there are other equilibria (see Problem 3.6.7).
his own payoffs but player I thinks there is probability that pay-
offs are given by the left matrix, probability 1 that they are
given by the right matrix. Say that player II is of type 1 if payoffs
are given by the left matrix, type 2 if payoffs are given by the right
matrix. Clearly equilibrium must have: II plays L if II is of type 1,
R if II is of type 2; I plays T if > 21 , B if < 21 .
But now suppose x = 2. Should player II still feel comfortable
playing B if he is of type 2? The optimality of IIs action choice
of B depends on his believing that I will play T , which only occurs
if I assigns probability at least 12 to the right matrix. But suppose
II does not know Is beliefs . Then, II has beliefs over Is beliefs
1
and so II finds R optimal if he assigns probability at least 2 to I
1
assigning probability at least 2 to the right matrix.
But, we are not done: how confident is I that II will play R in the
right matrix? Now we need to deal with Is beliefs over IIs beliefs.
Player I will only play B if he assigns probability at least 21 to II
assigning probability at least 21 to I assigning probability at least 21
to the right matrix. And so on.
This leads us into an infinite regress, since now Is beliefs about
IIs beliefs about Is beliefs become relevant! So, how to analyze
such problems in general?
A strategy for i is
si : Ti Ai .
Let s(t) (s1 (t1 ), . . . , sn (tn )), etc.
72 CHAPTER 3. GAMES WITH NATURE
Eti [ui (s(t), t)] Eti [ui (ai , si (ti ), t)], ai Ai , (3.4.1)
and
si : Ti Ai .
74 CHAPTER 3. GAMES WITH NATURE
A B
A , 9, 5
B 5, 9 7, 7
Figure 3.5.1: The game for Example 3.5.1. For = 9, this is the game
studied in examples 3.2.1 and 2.5.5.
!0
!1 !2
!3 !4
!5 !6
!7 !8
!9
Figure 3.5.2: The solid lines describe player 1s information sets, while the
dashed lines describe player 2s. Player 1 knows the row and
player 2 knows the column containing the realized state.
In state 7 , both players know = 9, player 1 knows that
player 2 knows that = 9, but player 2 assigns probability
1/2 to 8 , and so to the event that player 1 does not know
that = 9.
In state 6 , both players know = 9, both players know
that both players know that = 9, player 2 knows that player
1 knows that player 2 knows that = 9, but player 1 does
not know that player 2 knows that player 1 knows that = 9.
76 CHAPTER 3. GAMES WITH NATURE
we also have s2 t2` = B.
and
(1 1
t1`1 + 2 t1` , if ` < (K + 1)/2,
p2 (t2` ) = 2
(K1)/2
1 t2 , if ` = (K + 1)/2.
78 CHAPTER 3. GAMES WITH NATURE
si : [, 20 + ] {A, B}.
Since
f (x, )
h(x | ) = ,
g()
where f is the joint density, we have
( 1
, if x [ , + ] and [0, 20],
f (x, ) = 40
0, otherwise.
January 18, 2016 79
A B
A 9, 9 0, 5
B 4, 0 7, 7
3.6 Problems
3.6.1. There are two firms, 1 and 2, producing the same good. The inverse
demand curve is given by P = q1 q2 , where qi R+ is firm
is output. (Note that we are allowing negative prices.) There is de-
mand uncertainty with nature determining the value of , assigning
probability (0, 1) to = 3, and complementary probability 1
to = 4. Firm 2 knows (is informed of) the value of , while firm 1
is not. Finally, each firm has zero costs of production. As usual, as-
sume this description is common knowledge. Suppose the two firms
choose quantities simultaneously. Define a strategy profile for this
game. Describe the Nash equilibrium behavior and interim payoffs
(which may be unique).
resolved by a fair coin). Suppose there are two bidders, with bidder
is value vi randomly drawn from the interval [vi , vi ] according to
the distribution function Fi , with density fi .
3.6.4. Consider the two bidder sealed bid auction of Example 3.3.1, when
the two bidders valuations are independent uniform draws from the
interval [0, 1].
In a first price sealed bid auction with reserve price r , the highest
bidder only wins the auction (and pays his bid) if the bid is at least
r.
(b) What is the symmetric Nash equilibrium of the first price sealed
bid auction with reserve price r [0, 1]? [Hint: Since any win-
ning bid must be no less than r , bidders with a value less than
r can be assumed to bid 0 (there are multiple symmetric equi-
libria that only differ in the losing bids of bidders with values
less than r ). It remains to determine the bidding behavior of
those bidders with values r . Prove that any bidder with value
r must bid r . While the equilibrium is not linear in valuation,
the recipe from Example 3.3.1 still works.]
(c) What reserve price maximizes expected revenue?
3.6.5. This question asks you to prove a claim made in Example 3.3.4 as
follows:
January 18, 2016 83
(a) Suppose s2 < s1 , and set Pr{s2 < 1 (t1 ) s1 } > 0. Prove
that there exists s1 satisfying Pr{1 (t1 ) > s1 } < /2. [Hint: This
is trivial if s1 < (why?). The case where s1 = uses a basic
continuity property of probability.]
(b) Show that a deviation by type t2 > 2s1 to a stopping time s2 > s1
(which implies that t2 wins the war of attrition with probability
of at least /2) satisfying s2 < t2 /2 is strictly profitable.
3.6.6. This question concerns Example 3.1.1, the Cournot game with in-
complete information. The idea is to capture the possibility that
firm 2 may know that firm 1 has low costs, cL . This can be done as
follows: Firm 1s space of uncertainty (types) is, as before, {cL , cH },
while firm 2s is {tL , tU }. Nature determines the types according to
the distribution
,
if (t1 , t2 ) = (cL , tL ),
Pr(t1 , t2 ) = , if (t1 , t2 ) = (cL , tU ),
1 , if (t , t ) = (c , t ),
1 2 H U
3.6.7. The linear equilibrium of Example 3.3.5 is not the only equilibrium
of the double auction.
(d) Compare the expected gains from trade under these fixed price
equilibria with the linear equilibrium of Example 3.3.5.
3.6.9. This question asks you to fill in some of the details of the calcula-
tions in Example 3.5.2.
s (s ).
85
86 CHAPTER 4. NASH EQUILIBRIUM
2. ui : S1 Sn R is continuous in s S1 Sn and
quasiconcave in si ,
and so
ui (si0 + (1 )si00 , si ) max ui (si , si )
so that si0 + (1 )si00 arg max ui (si , si ).
The theorem then follows from Kakutanis fixed point theorem
by taking X = S1 Sn and F = (1 , . . . , n ).
Theorem 4.1.3 (Nash). Every finite normal form game, (Si , ui )i , has
a mixed strategy Nash equilibrium.
Q
Proof. Define Xi = (Si ) and X i Xi . Then X R|Si | and X is
nonempty, convex and compact.
In this case, rather than appealing to the maximum theorem, it
is an easy (and worthwhile!) exercise to prove that i is convex-
valued and has a closed graph.
In some applications, we need an infinite dimensional version of
Kakutanis fixed point theorem.
Theorem 4.1.4 (Fan-Glicksberg fixed point theorem). Suppose X is
a nonempty compact convex subset of a locally convex Hausdorff
space, and suppose F : X X. Suppose
lem 4.3.6). Even though the game has discontinuous payoffs (why?),
it does have an equilibrium in mixed strategies. It is also worth not-
ing that rather than working directly with mixed strategies (which
are randomizations over pairs of bids (b0 , b00 ), where b0 is the bid
of v and b00 is the bid of v), it is convenient to work with behavior
strategies. 3
In a behavior strategy, the randomizations over bids
are specified separately for the bidder of value v and v.
Suppose bidder 2 follows the strategy 2 of bidding v if v2 = v
(why is that a reasonable guess?), and of bidding according
to the
distribution function F2 (b) if v2 = v. Then, assuming there are no
atoms in F2 , player 1 has interim payoffs from b > v given by
1 1
U1 (b, v, 2 ) = (v b) + (v b)F2 (b)
2 2
1
= (v b)(1 + F2 (b)).
2
Note that the minimum of the support of F2 is given by v (why?).
Denote the maximum of the support by b.
Suppose 1 is also randomizing over the set of bids, (v, b]. The
indifference condition requires that 1 is indifferent over all b
(v, b]. The bid b = v is excluded because there is a positive prob-
ability
of a tie at v (from the low value bidder) and so it cannot be
optimal for v to bid v. That is, for all > 0,
1
(v b)(1 + F2 (b)) = U1 (v + , v, 2 ),
2
and
1
U1 (v + , v, 2 ) = (v v )(1 + F2 (v + )).
2
3
Moreover, behavior strategies are also more natural when viewing the game
as one of incomplete information.
January 18, 2016 89
bv
F2 (b) = , b (v, b].
v b
Note that F2 (v) = 0.
Moreover, F2 (b) = 1 implies b = (v+v)/2. An alternative deriva-
tion of the value of b is to note that bidder 1 should receive the
1
payoff 2 (v v) from bidding b (at which 1 wins for sure, since
there are no atoms), or
1
(v v) = v b,
2
and then solving for b.
It is straightforward to verify that the symmetric profile in which
each bidder bids v if v = v, and bids according to the distribution
function F (b) = (b v)/( v b) if v = v is a Nash equilibrium.
Note that the high value bidder is not indifferent over every bid
in the support [v, b]:
v b, if b > b,
1
(v v), if b (v, b],
U1 (b, v, 2 ) = 12
(v v), if b = v,
4
0, if b < v.
See Figure 4.1.1 for a graph of this payoff function.
U1
1
v v)
2 (
1
v v)
4 (
v 1
v + v) v b
2 (
v2 = 3 v2 = 4
1 1
v1 = 3 2 4
1
v1 = 4 4
0
1
probability 2 . An intuition for nonexistence can be obtained by
observing that since there is positive probability that 2 has value
3, neither bidder will bid over 3 when their value is 3 (since he
may win). Moreover, neither bidder can randomize on bids below
3 when their value is 3 for standard reasons. But if bidder 1 bids
3, then bidder of value 4 does not have a best reply. Existence is
restored if ties are broken in favor of bidder 2.
Suppose now the joint distribution over bidder valuations is
given by the table in Figure 4.1.2 (a symmetric version of the ex-
ample in the previous paragraph). The critical feature of this joint
distribution is that the bidder with value 4 knows that the other
bidder has value 3 (and so the valuations are not independent). For
any tie breaking rule, the value 4 bidder has no best reply, and so
there is no equilibrium (in pure or mixed strategies). Existence can
only be restored by using a tie breaking rule that awards the item
to the highest value bidder. For (much) more on this, see Jackson,
Simon, Swinkels, and Zame (2002).
Common thread:
94 CHAPTER 4. NASH EQUILIBRIUM
0 1 1
2
behavior is myopic,
dynamic analysis, focusing on asymptotic properties, and
focus on interaction of learning with evolution of system.
A B
A 1 0
B 0 1
No role identification, so the payoff represents the payoff to a
player who chooses the row action, when facing the column action.
If is fraction of population playing A, then
u(A; ) =,
and u(B; ) =1 .
Then
d 1
= >0 iff u(A; ) > u(B; ) iff > ,
dt 2
and
1
< 0 iff u(A; ) < u(B; ) iff < .
2
The dynamics are illustrated in Figure 4.2.1.
1 1
The symmetric mixed strategy equilibrium 2 A+ 2 B is unstable.
January 18, 2016 95
0 1 1
2
A B
A 1 2
B 2 1
u(A; ) = + 2(1 ) = 2 ,
and u(B; ) =2 + 1 = 1 + .
Then
1
> 0 iff 2>1+ iff > ,
2
and
1
< 0 iff 2<1+ iff < .
2
The dynamics are illustrated in Figure 4.2.2.
Dynamics:
F : (S) R+ (S),
with
F ( , t 0 + t) = F (F ( , t 0 ), t).
= F ( , t) t.
xs (t + 1) = xs (t)u(s, t ),
where
xs (t) xs (t)
t (s) = P .
r xr (t) x(t)
Then, since x(t + 1) = x(t)u( t , t ),
t+1 tu(s, t )
(s) = (s) ,
u( t , t )
so the difference equation is
t+1 t u(s, t ) u( t , t )
t
(s) (s) = (s) .
u( t , t )
98 CHAPTER 4. NASH EQUILIBRIUM
Thus, t+1 (s) > (<) t (s) iff u(s, t ) > (<)u( t , t ). In continu-
ous time, this is
u(s, ) u( , )
(s) = (s) .
u( , )
This has the same trajectories as
Note that under the replicator dynamic, every pure strategy pro-
file is a rest point: if (s) = 0 then (s) = 0 even when u(s, ) >
u( , ).
Idea extends in straightforward fashion to games with role iden-
tification. In that case, we have
L R
T 1, 1 1, 0
B 1, 1 0, 0
p = p(1 p)(1 q)
and q = q(1 q).
0 p 1
row
column
50
50 N Y
0 80
0 20
N Y
equal split 50, 50 50, 50
small offer 0, 0 80, 20
0 B p 1
Figure 4.2.5: The phase diagram for the simplified ultimatum example. A
is the nonsubgame perfect equilibrium (1, 3/8), and B is the
subgame perfect equilibrium.
102 CHAPTER 4. NASH EQUILIBRIUM
0
strategies not played by i between t and t 0 in ht .
Examples:
Cournot dynamics
4.3 Problems
4.3.1. Consider the correspondence F : [1, 1] [1, 1] defined as
(0, 1) if x 0,
F (x) =
(0, x ] if x > 0.
2
and
(0, 1), if x 0,
H(x) = h x x i
, , if x > 0.
2 2
4.3.3. (a) Prove that every finite extensive form game has a subgame per-
fect equilibrium.
(b) (Harris, Reny, and Robson, 1995, Section 2.1) Verify that the
following two-stage extensive-form game does not have a sub-
game perfect equilibrium (it is worth noting that this game has
continuous payoffs and compact actions spaces). There are
four players. In stage 1, players 1 and 2 simultaneously choose
a1 [1, 1] and a2 {L, R} respectively. In stage 2, players
3 and 4 are informed of the choices of players 1 and 2, and
then simultaneously choose a3 {L, R} and a4 {L, R} re-
spectively. Payoffs are
u1 (a1 , a2 , a3 , a4 ) =
|a1 | 12 |a1 |2 , if a2 = a3 and a3 = a4 ,
|a | 1 |a |2 , if a2 a3 and a3 = a4 ,
1 2 1
|a1 | 10 1 |a1 |2 ,
if a2 = a3 and a3 a4 ,
2
|a1 | 10 1 |a1 |2 ,
2 if a2 a3 and a3 a4 ,
1, if a2 = a3 = L,
1, if a2 = L, a3 = R,
u2 (a1 , a2 , a3 , a4 ) =
2, if a2 = a3 = R,
2, if a2 = R, a3 = L,
(
a1 , if a3 = L,
u3 (a1 , a2 , a3 , a4 ) =
a1 , if a3 = R,
and (
a1 , if a4 = L,
u4 (a1 , a2 , a3 , a4 ) =
a1 , if a4 = R.
4.3.4. Consider a first price sealed bid auction with private values. There
are two bidders with values v1 < v2 . These values are common
knowledge. Prove that this auction has no pure strategy equilibrium.
Characterize the set of mixed strategy equilibria. [Hint: In these
equilibria, bidder 2 plays a pure strategy and wins with probability
1.]
106 CHAPTER 4. NASH EQUILIBRIUM
4.3.6. This question asks you to fill in the details of Example 4.1.1.
(a) Prove that in any equilibrium, any bidder with value v must bid
v.
(b) Prove that there is no equilibrium in pure strategies.
(c) Prove that in any mixed strategy equilibrium, the minimum of
the support of F2 is given by v.
(d) Prove that it is not optimal for v to bid v.
January 18, 2016 107
(e) Prove that the symmetric profile in which each bidder bids v
if v = v, and according to the distribution function F (b) =
v b) if v = v is a Nash equilibrium.
(b v)/(
4.3.7. A variant of Example 4.1.1. The two bidders have independent pri-
vate values drawn independently and identically from the common
three point set {v1 , v2 , v3 }, with v1 < v2 < v3 . The common dis-
tribution assigns probability pk to value vk , k = 1, 2, 3. There is a
symmetric equilibrium in which the low value bidders bid v1 , the
mid value bidders randomize over a support [v1 , b2 ], and the high
value bidders randomize over a support [b2 , b3 ]. Fully describe the
equilibrium, and verify that it is indeed an equilibrium.
4.3.8. Consider again the two bidder sealed bid auction of Example 3.3.1,
but with a different valuation structure. In particular, the two bid-
ders valuations are independent draws from the interval [1, 2], with
each bidders valuation being a uniform draw from that interval with
probability p (0, 1) and with complementary probability 1 p
equaling some v [1, 2].
4.3.9. Consider the following variant of a sealed bid auction: There are two
bidders who each value the object at v, and simultaneously submit
bids. As usual, the highest bid wins and in the event of a tie, the
object is awarded on the basis of a fair coin toss. But now all bidders
pay their bid. (This is an all-pay auction.)
4.3.11. An asymmetric variant of Example 4.1.1. The two bidders have inde-
pendent private values drawn independently from the common two
point set {v, v}, with v < v. The probability that bidder i has value
1), and p < p . Ties are broken in favor of bidder 2.
v is pi (0, 1 2
(a) Suppose firm i only incurs the fixed cost of production when
qi > 0. Then, firm is profits are given by
(
(P (q1 + q2 ) c)qi , if qi > 0,
Ui (q1 , q2 ) =
0, if qi = 0.
This game has two pure strategy Nash equilibria. What are
they? Why doesnt this game have a symmetric pure strategy
Nash equilibrium? Since firm payoffs are not continuous func-
tions (why not?), the existence of a symmetric Nash equilibrium
in mixed strategies is not implied by any of the theorems in Sec-
tion 4.1.
January 18, 2016 109
(b) Suppose the fixed cost is an entry cost (and so is sunk). The
game is still a simultaneous move game, but now firm is strat-
egy space is {1}R+ , with the strategy 1 meaning dont en-
ter, and a number qi R+ meaning enter and choose qi 0
(note the weak inequality). Then, firm is profits are given by
(P (q1 + q2 ) c)qi , if (s1 , s2 ) = (q1 , q2 ),
Ui (s1 , s2 ) = (P (qi ) c)qi , if si = qi and sj = 1,
0, if si = 1.
4.3.13. Prove that the phase diagram for example 4.2.3 is as portrayed in
Figure 4.2.3. [This essentially asks you to give an expression for
dq/dp.]
t (s) 1,
where t F ( 0 , t).
January 18, 2016 111
A B C
A 1 1 0
B 0 1 1
C 0 0 1
1
I A II a
1
1
D d
III
L R L R
0 3 0 4
0 3 0 4
0 2 1 0
1
Copyright January 18, 2016 by George J. Mailath
113
114 CHAPTER 5. DYNAMIC GAMES
Pr(I plays A) = p1 ,
Pr(II plays a) = p2 , and
Pr(III plays L) = p3.
Consider the Nash equilibrium profile (0, 1, 0) (i.e., DaR). This pro-
file is subgame perfect, and yet player II is not playing sequentially
rationally. It is also not trembling hand perfect (Definition 2.5.1):
playing a is not optimal against any mixture close to DR.
The only trembling hand perfect equilibrium outcome is Aa. The
3
set of Nash equilibria with this outcome is {(1, 1, p3 ) : 4 p3 1}.
In these equilibria, player IIIs information set is not reached, and
so the profile cannot be used to obtain beliefs for III. However,
each Nash equilibrium in the set is trembling hand perfect: Fix an
equilibrium (1, 1, p3 ). Suppose first that p3 [ 34 , 1) (so that p3 1!)
and consider the completely mixed profile
1
p1n =1 ,
n
2
p2n =1 ,
(n 1)
and p3n =p3 .
1 (n 1) 2
+ = 3/n,
n n (n 1)
and so the induced beliefs for III at his information set assign prob-
1 2
ability 3 to the left node and 3 to the right. Player III is therefore
indifferent and so willing to randomize.
The same argument shows that (1, 1, 1) is trembling hand per-
January 18, 2016 115
1
p1n =1 ,
n
2
p2n =1 ,
(n 1)
1
and p3n =1 .
n
Indeed, any sequence of trembles satisfying p1n 1, p2n 1, and
p3n 1 will work, providing
(1 p1n ) 1
lim sup n n .
n (1 p1 p2 ) 3
(It is not even necessary for (1 p1n )/(1 p1n p2n ) to have a well-
defined limit.)
Example 5.1.2. Consider the profile (LR, UD) in the game displayed
in Figure 5.1.2. The label [p] indicates that nature chooses the node
t1 with probability p (so that (t1 ) = p and (t2 ) = 1 p). The
induced distribution Pb on Z is p z1 + (1 p) z8 .
116 CHAPTER 5. DYNAMIC GAMES
z1 U U z5
L t1 R
[p]
z2 D D z6
II II
z3 U U z7
L t2 R
[1 p]
z4 D D z8
2
More formally, for z Z(h), there is a unique path from some x 0 h to z.
Denote the unique sequence of actions on the path from x 0 to z by a1 , . . . , aL
(where a` may be an action of nature), with player (`) choosing the action a` .
Then, for all z h,
YL
P,b (z | h) = h (x 0 ) `=0 b(`) (a` ).
January 18, 2016 117
I A1 II A2 I A1
0
0
h h0 1
D1 D2 D1
0
5 1 10
0 0 1
1
A01 2
[1] 0
U2 D10 2
I
I A1 II A2 A01 0
h [0] 1
D1 D2 D10
1 5 10
0 0 1
Figure 5.1.4: The game for Example 5.1.4. Player I is not playing sequen-
tially rationally at h. Nonetheless, player I does not have a
profitable one-shot deviation at any information set (given
the beliefs specified).
I
L
R
2
0 I
T B
[0] [1]
II
` r ` r
1 4 0 5
1 0 0 1
Example 5.2.1. Recall the extensive form from Example 2.3.4, re-
produced in Figure 5.2.1. The label [p] indicates that the player
owning that information set assigns probability p to the labeled
January 18, 2016 121
node. The profile RBr (illustrated) is Nash and satisfies the condi-
tions of the theorem.
Pb (x)
(x) = x h.
Pb (h)
P,b (x | h0 ) X
(x) = (x) x hh0 .
P,b (hh0 | h0 ) xhh0
I
L
R
2
0 I
T B
[1] [0]
II
` r ` r
1 4 0 5
1 0 0 1
Figure 5.2.2: The profile LB` (illustrated) is weak perfect Bayesian, but not
almost perfect Bayesian.
124 CHAPTER 5. DYNAMIC GAMES
I x1
L
C R
x2 x3
II
` `
c r c r
x4 x5 x6 x7
III
k
Since bk b, from (5.2.1) we have Pb (x) Pb (x), and so if h is on
the path of play of b (i.e., Pb (h) > 0), we immediately have
k
k Pb (x) Pb (x)
(x) = bk - b ,
P (h) P (h)
k k k k
(x2 ) = k and (x3 ) = k .
+ k + k
126 CHAPTER 5. DYNAMIC GAMES
k k
k (x4 ) = ,
(k + k )(k + k )
k k
k (x5 ) = ,
(k + k )(k + k )
k k
k (x6 ) = , and
(k + k )(k + k )
k k
k (x7 ) = .
(k + k )(k + k )
k k k + k k
k (x2 ) k ({x6 , x7 }) =
(k + k ) (k + k )(k + k )
k k
= ,
(k + k ) (k + k )
which equals k (x3 ) k ({x4 , x5 }). That is, (since all probabilities are
positive along the sequence, and so we can divide), we have that for
all k,
k (x2 ) k ({x4 , x5 })
= k ,
k (x3 ) ({x6 , x7 })
which must be preserved in the limit (and is the desirable property
suggested in the caption of Figure 5.2.3).
k k
YL
k
Pb (x) = Pb (x 0 ) `=0
b(`) (a` ),
and so
k QL k
bk 0
Pb (x 0 ) `=0 b(`) (a` )
P (x | h ) =
Pbk (h0 )
k
YL
k
= Pb (x 0 | h0 ) `=0 b(`) (a` )
YL
k
= k (x 0 ) `=0 b(`) (a` )
YL
(x 0 ) `=0 b(`) (a` ) = P,b (x | h0 ).
P
If xhh0 (x) = 0, then the second condition in the definition
trivially holds.
128 CHAPTER 5. DYNAMIC GAMES
P
Suppose xhh0 (x) > 0. Then, for all x hh0 ,
k k
k (x) Pb (x)/Pb (h)
P = P P
xhh0 (x) xhh0 P (x)/ xh P (h)
k bk bk
k
Pb (x)
=P
xhh0 P (x)
bk
k
Pb (x)
= bk
P (hh0 )
k k
Pb (x)/Pb (h0 )
= bk
P (hh0 )/Pbk (h0 )
k
Pb (x | h0 )
= bk .
P (hh0 | h0 )
The equality is preserved when taking limits (since the limits of the
first and last denominator are both strictly positive), and so the
second condition in the definition again holds.
Another instance of the one-shot deviation principle, which we
first saw in Theorem 5.1.1:
Theorem 5.3.2. In a finite extensive form game, suppose is consis-
tent with a profile b. The profile b is sequentially rational given
(and so a sequential equilibrium) if and only if there are no profitable
one-shot deviations from b (given ).
Proof. Lemma 5.1.1 is the easy direction.
Suppose b is not sequentially rational given . Then there is a
player, denoted i, with a profitable deviation. Denote the profitable
deviation (by player i) by bi0 and the information set h0 . Player i
information sets Hi are strictly partially ordered by precedence in
the obvious way (see Problem 1.4.6). Let Hi (h0 ) denote the finite
(since the game is finite) collection of information sets that follow
h0 . Let K be the length of the longest chain in Hi (h0 ), and say an
information set h Hi (h0 ) is of level k if the successor chain from
h0 to h has k links (h0 is 0-level and its immediate successors are
all 1-level). If i has a profitable deviation from bi (given ) at any
K-level information set, then that deviation is a profitable one-shot
deviation (given ), and we are done.
January 18, 2016 129
130 CHAPTER 5. DYNAMIC GAMES
5.4 Problems
5.4.1. This problem concerns the game given in Figure 5.4.1.
5.4.4. This problem illustrates how badly behaved the follows relation
on information sets can be, and the reason for the introduction
of hh0 into the definition of almost perfect Bayesian equilibria. Con-
sider the game in Figure 5.4.2.
0 0
1 1
0
A A 1
B
U B 10
1
10 II I III 1
1
1
A D
0 A 0
1 B B 0
1 1
5.4.8. We return to the environment of Problem 3.6.1, but with one change.
Rather than the two firms choosing quantities simultaneously, firm
1 is a Stackelberg leader: Firm 1 chooses its quantity, q1 , first. Firm
2, knowing firm 1s quantity choice then chooses its quantity. De-
scribe a strategy profile for this dynamic game. What is the ap-
propriate equilibrium notion for this game and why? Describe an
equilibrium of this game. Is it unique?
I
x1
L R
C
II
` r ` r ` r
III
1 1
1 0 `0 r0 `0 r0 `0 r0 `0 r0
0 1
2 0 1 1 2 0 0 0
0 0 0 1 0 0 0 1
1 0 0 1 1 x 0 1
Signaling1
1. for all t T ,
135
136 CHAPTER 6. SIGNALING
1, 1 f f 2, 1
b t1 q
[0.5]
1, 0 r r 0, 0
II II
2, 1 f f 3, 1
b t2 q
[0.5]
0, 0 r r 1, 0
Since the different information sets for player II are not ordered
by (recall Problem 1.4.6), consistency places no restrictions on
beliefs at different information sets of player II. This implies the
following result (which Problem 6.3.3 asks you to prove).
Theorem 6.1.1. Suppose T , M, and R are finite. A profile is a perfect
Bayesian equilibrium if, and only if, it is a sequential equilibrium.
Example 6.1.1 (Separating equilibria). In the game given in Figure
6.1.1, (bq, f r ) is a separating eq.
3, 1 f f 2, 1
b w q
[0.1]
1, 0 r r 0, 0
II II
2, 1 f f 3, 1
b s q
[0.9]
0, 0 r r 1, 0
2 c(e, )
< 0.
e
Also assume c(e, ) 0, ce (e, ) c(e, )/e 0, ce (0, ) = 0,
cee (e, ) > 0, and lime ce (e, ) = .
Two identical firms competing for worker. Each firm values
worker of type with education e at f (e, ). In any discretization
of the game, in any almost perfect Bayesian equilibrium, after any
e, firms have identical beliefs () about worker ability (see
Problem 6.3.5). Consequently, the two firms are effectively playing
a sealed bid common value first price auction, and so both firms
bid their value E f (e, ). To model as a game, replace the two
firms with a single uninformed receiver (the market) with payoff
(f (e, ) w)2 .
January 18, 2016 139
de fe (e, ) ce (e, )
= > 0.
d fee (e, ) cee (e, )
(e(), w(e())) .
U( 0 , 0 , e( 0 )) U( 0 , 00 , e( 00 )), 0 , 00 . (6.2.3)
and
w(e) = E f (e, ).
e
U ( 0 , , e) = k0
U ( 00 , , e) = k0
U ( 00 , , e) = k00
00
E[|e]
e
e ( 0 ) e ( 00 ) e00
w(e) = f (e, ).
Note that (6.2.2) and (6.2.5) differ in two ways: the set of possible
maximizers and how e enters into the objective function.
January 18, 2016 143
e
U ( 0 , , e) = k0
U ( 00 , , e)
00
e100 e200 e
e ( 0 )
e
U ( 0 , , e) = k0p
U ( 0 , , e) = k0
U ( 00 , , e) = k00
p
00
E[]
E[|e]
e ( 0 ) ep e e
Indeed, for two types, the intuitive criterion selects the Riley
separating outcome, i.e., the separating outcome that minimizes
the signaling distortion. Consider a separating outcome this is not
the Riley outcome. Then, e( 00 ) > e100 (see Figure 6.2.2). Consider an
out-of-equilibrium message e (e100 , e( 00 )). Since e > e100 , D(e) =
{ 0 }. Then, as above, the market wage after e must be 00 , and so
since e < e( 00 ),
U( 00 , 00 , e) > U( 00 , 00 , e( 00 )),
0 1 d(e)1 (e)
U ( , (e) (e), e) + Ue ( 0 , (e)1 (e), e)
de
!1
0 1 de()
= U ( , (e) (e), e) + Ue ( 0 , (e)1 (e), e).
d =(e)
1 (e)
de( 0 ) U ( 0 , 0 , e0 ) f (e0 , 0 )
= = .
d Ue ( 0 , 0 , e0 ) fe (e0 , 0 ) ce (e0 , 0 )
146 CHAPTER 6. SIGNALING
6.3 Problems
6.3.1. Show that for the game in Figure 6.3.1, for all values of x, the out-
come in which both types of player I play L is sequential by explic-
itly describing the converging sequence of completely mixed behav-
ior strategy profiles and the associated system of beliefs. For what
values of x does this equilibrium pass the intuitive criterion?
6.3.2. (a) Prove that the signaling game illustrated in Figure 6.3.2 has no
Nash equilibrium in pure strategies.
(b) Definition 6.1.1 gives the definition of a pure strategy perfect
Bayesian equilibrium for signaling games. How should this def-
inition be changed to cover mixed strategies? More precisely,
give the definition of a perfect Bayesian equilibrium for the case
of finite signaling games.
1
(c) For the game in Figure 6.3.2, suppose p = 2 . Describe a perfect
Bayesian equilibrium.
U 6, 3
L t1 R
2, 0
1
[2]
D 0, 0
II
U x, 0
L t2 R
2, 0
1
[2]
D 0, 2
Figure 6.3.1: Game for problem 6.3.1. The probability that player I is type
t1 is 1/2 and the probability that he is type t2 is 1/2. The
first payoff is player Is payoff, and the second is player IIs.
2, 0 3, 0
TL TR
L t1 R
2, 2 ML
[p]
BL BR
1, 3 3, 1
II II
1, 3 3, 1
TL TR
L t2 R
2, 2 ML
[1 p]
BL BR
2, 0 3, 0
6.3.4. The canonical signaling game has a sender with private information,
denoted R choosing a message m R, where is compact.
A receiver, observing m, but not knowing then chooses a response
r R. The payoff to the sender is u(m, r , ) while the payoff to the
receiver is v(m, r , ). Assume both u and v are C 2 . Assume v is
strictly concave in r , so that v(m, r , ) has a unique maximizer in
r for all (m, ), denoted (m, ). Define
U (, , m) = u(m, (m, ), ).
(d) Prove that the messages sent by the sender in any separating
Nash equilibrium are strictly increasing in type.
January 18, 2016 149
6.3.5. Prove that in any discretization of the job market signaling game,
in any almost perfect Bayesian equilibrium, after any e, firms have
identical beliefs about worker ability.
6.3.7. The owner of a small firm is contemplating selling all or part of his
firm to outside investors. The profits from the firm are risky and the
owner is risk averse. The owners preferences over x, the fraction of
the firm the owner retains, and p, the price per share paid by the
outside investors, are given by
u(x, , p) = x x 2 + p(1 x),
where is the value of the firm (i.e., expected profits). The quadratic
term reflects the owners risk aversion. The outside investors are
risk neutral, and so the payoff to an outside investor of paying p
per share for 1 x of the firm is then
(1 x) p(1 x).
There are at least two outside investors, and the price is determined
by a first price sealed bid auction: The owner first chooses the frac-
tion of the firm to sell, 1 x; the outside investors then bid, with
the 1 x fraction going to the highest bidder (ties are broken with a
coin flip).
(P 1 c1 )q11 + (P 2 c1 )q12 ,
(P 2 c2 )q22 .
(b) Suppose now that firm 1s costs are private to firm 1. Firm 2
does not know firm 1s costs, assigning prior probability p
(0, 1) to cost c1L and complementary probability 1 p to cost
c1H , where c1L < c1H < A/2.
i. Define a pure strategy almost perfect Bayesian equilibrium
for this game of incomplete information . What restric-
tions on second period quantities must be satisfied in any
pure strategy almost perfect Bayesian equilibrium? [Make
152 CHAPTER 6. SIGNALING
(c) Does the equilibrium from part 6.3.9(b) pass the intuitive cri-
terion? Why or why not? If not, describe a separating perfect
Bayesian equilibrium that does.
6.3.10. We continue with the setting of Problem 3.6.1, but now suppose that
firm 2 is a Stackelberg leader who has the option of not choosing be-
fore firm 1: Firm 2 either chooses its quantity, q2 , first, or the action
W (for wait). If firm 2 chooses W , then the two firms simultaneously
choose quantities, knowing that they are doing so. If firm 2 chooses
its quantity first (so that it did not choose W ), then firm 1, knowing
firm 2s quantity choice then chooses its quantity.
(c) Prove that the outcome in which firm 2 does not wait for any
, and firms behave as in the separating outcome of question
6.3.9(b) is not an equilibrium outcome of this game.
154 CHAPTER 6. SIGNALING
Chapter 7
Repeated Games1
and so can write a0 for s 0 . The set of all strategies for player i is Si .
Note distinction between
actions ai Ai and
strategies si : H Ai .
155
156 CHAPTER 7. REPEATED GAMES
(if the arg max is unique for some history ht , s(i(ht ) is uniquely
determined, otherwise make a selection from the argmax). Since
Ui (s ) Ui (si , si ),
January 18, 2016 157
p
and since in every period vi is a lower bound for the flow payoff
, s ), we have
received under the profile (si i
X p p
Ui (s )
Ui (si , si ) (1 ) t vi = vi .
t=0
vi = min
Q max ui (i , i ).
i ji(Aj ) i (Ai )
E S
E 2, 2 1, 3
S 3, 1 0, 0
si|ht Si .
si () =E,
and for t 1,
( 0
E, if at = EE for all t 0 = 0, 1, . . . , t 1,
si (a0 , . . . , at1 ) =
S, otherwise.
January 18, 2016 159
P
Payoff to 1 from (s1 , s2 ) is: (1 ) 2 t = 2.
Consider a deviation by player 1 to another strategy s1 . In re-
sponse to the first play of S, player 2 responds with S in every
subsequent period, so player 1 can do no better than always play
S after the first play of S. The maximum payoff from deviating in
period t = 0 (the most profitable deviation) is (1 )3. The profile
is Nash if
1
2 3(1 ) .
3
Strategy profile is subgame perfect: Note first that the profile
only induces two different strategy profiles after any history. De-
note by s = (s1 , s2 ) the profile in which each player plays S for all
histories, si (ht ) = S for all ht H. Then,2
( 0
si , if at = EE for all t 0 = 0, 1, . . . , t 1,
si|(a0 ,...,at1 ) =
si , otherwise.
si (ht ) = sj (ht ) ht H, i, j.
W is set of states,
w 0 is initial state,
2
This is a statement about the strategies as functions, i.e., for all h H,
(
si (h ), if at = EE for all t 0 = 0, 1, . . . , t 1,
0
si|(a0 ,...,at1 ) (h ) =
si (h ), otherwise.
160 CHAPTER 7. REPEATED GAMES
ES, SE, SS
w0 wEE wSS
vention: the subscript on the state indicates the action profile spec-
ified by the output function (i.e., f (wa ) = a, and we may have
f (wa ) = a and f (wa ) = a); it is distinct from the transition func-
tion.
Then,
|Ui (si , si ) Ui (si , si )| T M < /2,
so that
and
Ui (si|hT 1 , si|hT 1 )
= (1 )ui (si (hT 1 ), fi (w)) + Vi ((w, (si (hT 1 ), fi (w)))).
Hence, if Ui (si|hT 1 , si|hT 1 ) > Ui (si|hT 1 , si|hT 1 ), then we are done,
since player i has a profitable one-shot deviation from (W , w 0 , f , ).
Suppose not, i.e., Ui (si|hT 1 , si|hT 1 ) Ui (si|hT 1 , si|hT 1 ). For the
strategy si defined by
(
t si (ht ), t < T 1,
si (h ) =
si (ht ), t T 1,
we have
TX
2
Ui (si , si ) =(1 ) t ui (at (si , si )) + T 1 Ui (si|hT 1 , si|hT 1 )
t=0
TX
2
(1 ) t ui (at (si , si )) + T 1 Ui (si|hT 1 , si|hT 1 )
t=0
=Ui (si , si ) > Ui (si , si ).
That is, the (T 1)-period deviation is profitable. But then either the
one-shot deviation in period T 1 is profitable, or the (T 2)-shot
deviation is profitable. Induction completes the argument.
E S
E 2, 2 (1 ), 3(1 ) ,
S 3(1 ), (1 ) 0, 0
E S
E 2(1 ), 2(1 ) (1 ), 3(1 ) .
S 3(1 ), (1 ) 0, 0
A B C
A 4, 4 3, 2 1, 1
B 2, 3 2, 2 1, 1
C 1, 1 1, 1 1, 1
Stage game has a unique Nash eq: AA. Suppose 2/3. Then
there is a subgame perfect equilibrium of G() with outcome path
(BB) : (W , w 0 , f , ), where W = {wBB , wCC }, w 0 = wBB , fi (wa ) =
ai , and
(
wBB , if w = wBB and a = BB, or w = wCC and a = CC,
(w, a) =
wCC , otherwise.
The automaton is illustrated in Figure 7.1.3.
Values of the states are
Vi (wBB ) =(1 )2 + Vi (wBB ),
and Vi (wCC ) =(1 ) (1) + Vi (wBB ).
January 18, 2016 165
BB
BB
w0 wBB wCC CC
CC
Solving,
Vi (wBB ) =2,
and Vi (wCC ) =3 1.
A B C
A 4(1 ) + (3 1) 3(1 ) + (3 1) 1 + (3 1)
,
B 2(1 ) + (3 1) 2 1 + (3 1)
C 1 + (3 1) 1 + (3 1) (1 ) + (3 1)
A B C
A 4(1 ) + (3 1) 3(1 ) + (3 1) 1 + (3 1)
,
B 2(1 ) + (3 1) 2(1 ) + (3 1) 1 + (3 1)
C 1 + (3 1) 1 + (3 1) (1 ) + 2
166 CHAPTER 7. REPEATED GAMES
i.e.,
0 2 5 + 32 a 0 (1 )(2 3),
c s
H 3, 3 0, 2
L 4, 0 2, 1
Player I (row player) is long-lived, choosing high (H) or low (L) ef-
fort, player II (the column player) is short lived, choosing the cus-
tomized (c) or standardized (s) product.
Subgame perfect equilibrium described by the two state automa-
ton given in Figure 7.2.1. The action profile Ls is a static Nash equi-
librium, and since wLs is an absorbing state, we trivially have that
Ls is a Nash equilibrium of the associated one-shot game, g wLs .
Note that V1 (wHc ) = 3 and V1 (wLs ) = 2. Since player 2 is short-
lived, he must myopically optimize in each period. The one-shot
game from Corollary 7.1.1 has only one player. The one-shot game
January 18, 2016 167
Hc, Hs
Lc, Ls
w0 wHc wLs
c
H (1 )3 + 3
L (1 )4 + 2
and player I finds H optimal if 3 4 2, i.e., if 1/2.
Thus the profile is a subgame perfect equilibrium if, and only if,
1/2.
Example 7.2.2.
c s
H 3, 3 2, 2
L 4, 0 0, 1
The action profile Ls is no longer a static Nash equilibrium, and
so Nash reversion cannot be used to discipline player Is behavior.
Subgame perfect equilibrium described by the two state automa-
ton in Figure 7.2.2.
Since player 2 is short-lived, he must myopically optimize in
each period, and he is.
Note that V1 (wHc ) = 3 and V1 (wLs ) = (1 )0 + 3 = 3. There
are two one shot games we need to consider. The one-shot game
g wHc associated with wHc is given by
c
H (1 )3 + 3
L (1 )4 + 32
168 CHAPTER 7. REPEATED GAMES
Hc, Hs
Lc, Ls
Lc, Ls
s
H (1 )2 + 32
L (1 )0 + 3
R 10
so x(p, H ) = p di = 10 p. Also,
(
1, if i p + 6, and
iL (p) =
0, if i < p + 6,
R 10
so x(p, L ) = p+6 di = 10 (p + 6) = 4 p.
5
Unique subgame perfect equilibrium of stage game is ((L, 2 ), L ).
Why isnt the outcome path ((H, 6), H (6)) consistent with sub-
game perfection? Note that there are two distinct deviations by
the firm to consider: an unobserved deviation to (L, 6), and an ob-
served deviation involving a price different from 6. In order to deter
an observed deviation, we specify that consumers believe that, in
response to any price different from 6, the firm had chosen Q = L,
leading to the best response i given by
(
1, if p = 6 and i p, or p 6 and i p + 6,
i (p) =
0, otherwise,
implying aggregate demand
(
4, if p = 6,
x(p, ) =
max{0, 4 p}, p 6.
Clearly, this implies that observable deviations by the firm are not
profitable. Consider then the profile ((H, 6), ): the unobserved
deviation to (L, 6) is profitable, since profits in this case are (10
6)(6 1) = 20 > 16. Note that for the deviation to be profitable,
firm must still charge 6 (not the best response to H ).
Eq with high quality: buyers believe H will be produced as long
as H has been produced in the past. If ever L is produced, then L
is expected to always be produced in future. See Figure 7.2.3.
It only remains to specify the decision rules:
(
(H, 6), if w = wH , and
f1 (w) =
(L, 25 ), if w = wL .
and (
, if w = wH , and
f2 (w) =
L , if w = wL .
January 18, 2016 171
(Hp, x)
(Lp, x) (Qp, x)
w0 wH wL
Figure 7.2.3: Grim trigger in the quality game. Note that the transitions
are only a function of Q.
7.3 Applications
7.3.1 Efficiency Wages I
Consider an employment relationship between a worker and firm.
Within the relationship, (i.e., in the stage game), the worker (player
I) decides whether to exert effort (E) for the firm (player II), or to
shirk (S) for the firm. Effort yields output y for sure, while shirking
yields output 0 for sure. The firm chooses a wage w that period. At
the end of the period, the firm observes output, equivalently effort
and the worker observes the wage. The payoffs in the stage game
are given by
(
w e, if aI = E and aII = w,
uI (aI , aII ) =
w, if aI = S and aII = w,
and (
y w, if aI = E and aII = w,
uII (aI , aII ) =
w, if aI = S and aII = w.
Suppose
y > e.
Note that the stage game has (S, 0) as the unique Nash equilib-
rium, with payoffs (0, 0). This can also be interpreted as the pay-
offs from terminating this relationship (when both players receive
a zero outside option).
Grim trigger at the wage w is illustrated in Figure 7.3.1.
January 18, 2016 173
{Ew : w w }
{Ew : w < w } or Sw
w 0 wEw wS0
Ew
wEw
E0 Sw
S0
w0 wE0 wN
y w 0.
The worker has two incentive constraints. In state wE0 , the value to
the worker is
(1 )e + (w e) = w e =: VI (wE0 ).
VI (wE0 ) 0 (1 ) + VI (wE0 ),
that is
VI (wE0 ) 0.
w e (1 )w + VI (wE0 )
= (1 + 2 )w e,
which is equivalent
w e VI (wE0 ) 0.
Q = min{p1 , p2 , . . . , pn },
= (1 )p()( p()),
(7.3.6)
5
Strictly speaking, this example is not a repeated game with perfect monitor-
ing. An action for firm i in the stage game is a vector (pi ()) . At the end
of the period, firms only observe the pricing choices of all firms at the realized
. Nonetheless, the same theory applies. Subgame perfection is equivalent to
one-shot optimality: a profile is subgame perfect if, conditional on each informa-
tion set (in particular, conditional on the realized ), it is optimal to choose the
specified price, given the specified continuation play.
January 18, 2016 177
where
1 X
v = p(0 )(0 p(0 ))q(0 ).
n 0
Inequality (7.3.6) can be written as
nv
p()( p())
(n 1)(1 )
X
= p(0 )(0 p(0 ))q(0 ).
(n 1)(1 ) 0
n1
1 . (7.3.7)
(n 1)(1 ) n
Suppose there are two equally likely states, L < H. In order to
support collusion at the monopoly price in each state, we need :
( )
L2 L2 H 2
+ (7.3.8)
4 (n 1)(1 )2 4 4
( )
H2 L2 H 2
and + . (7.3.9)
4 (n 1)(1 )2 4 4
Since H 2 > L2 , the constraint in the high state (7.3.9) is the relevant
one, and is equivalent to
2(n 1)H 2
,
L2 + (2n 1)H 2
a tighter bound than (7.3.7), since the incentive to deviate is un-
changed, but the threat of loss of future collusion is smaller.
Suppose
n1 2(n 1)H 2
<< 2 ,
n L + (2n 1)H 2
so that colluding on the monopoly price in each state is inconsis-
tent with equilibrium.
178 CHAPTER 7. REPEATED GAMES
Since the high state is the state in which the firms have the
strongest incentive to deviate, the most collusive equilibrium sets
p(L) = L/2 and p(H) to solve the following incentive constraint
with equality:
( )
L2
p(H)(H p(H)) + p(H)(H p(H)) .
(n 1)(1 )2 4
that is,
L2 L L
p(H)(H p(H)) = L .
4 2 2
L
Note that setting p(H) = 2 violates this inequality (since H > L).
Since profits are strictly concave, this inequality thus requires a
lower price, that is,
L
p(H) < .
2
In other words, if there are enough firms colluding (n(1 ) >
1), collusive pricing is counter-cyclical! This counter-cyclicality of
prices arises with also with more states and a less severe require-
ment on the number of firms (Mailath and Samuelson, 2006, 6.1.1).
v =2(1 ) + (EE)
3(1 ) + (SE),
for some (EE), (SE) [0, 2]. The largest value for (EE)
is 2, while the incentive constraint implies the smallest is (1
)/, so that W EE = [3(1 ), 2]. See Figure 7.4.1 for an
illustration.
180 CHAPTER 7. REPEATED GAMES
v = SS
3 0
1 v = 2(1 ) + EE
1
W SS
W EE
v
2 v0 2
3 3
v =0 (1 ) + (SS)
(1)(1 ) + (ES),
for some (SS), (ES) [0, 2]. Since the inequality is satis-
fied by setting (SS) = (ES), the largest value for (SS) is
2, while the smallest is 0, and so W SS = [0, 2].
Observe that
[0, 2] W SS W EE .
[0, 2] W SS W EE .
So, when is
[0, 2] W SS W EE ?
This holds iff 2 3(1 ) (i.e., 35 ).
[0, 2] = W SS W EE ,
V1 (wS0 ) = 0, V1 (wEw ) = w e,
E Nature
[p] [1 p]
II
(w e, y w) (w, y w) (w, w)
Figure 7.5.1: The extensive form and payoffs of the stage game for the
game in Section 7.5.1.
{(y, w) : w w }
Figure 7.5.2: The automaton for the strategy profile in the repeated game
in Section 7.5.1. The transition from the state wEw labelled
{(y, w) : w w } means any signal profile in which output is
observed and the firm paid at least w ; the other transition
from wEw occurs if either the firm underpaid (w < w ), or
no output is observed (0).
January 18, 2016 185
i.e.,
(1 p)w (1 p)e
or
1 p e p(1 )
w e= + e.
(1 p) (1 p)
To understand the role of the imperfect monitoring, compare
this with the analogous constraint when the monitoring is perfect
(7.3.1), which requires w e/..
The firm does not wish to deviate in wEw if
V2 (wEw ) (1 )y + py,
i.e.,
y w (1 )y + py a (1 p)y w .
So, the profile is an equilibrium if
1 p
(1 p)y w e.
(1 p)
Public histories:
t
H
t=0 Y ,
W is set of states,
w 0 is initial state,
: W Y W is transition function.
y y
(3p2q)
(p+2q)
E (pq)
(pq)
3(1r ) 3r
S (qr )
(qr )
Figure 7.5.3: The ex post payoffs for the imperfect public monitoring ver-
sion of the prisoners dilemma from Figure 7.1.1.
See Problem 7.6.13 for the proof (another instance of the one-
shot deviation principle). .
y y
w0 wEE wSS y
y
and
This is eq if
and
or
(p q){V (wEE ) V (wSS )} (1 ).
We can obtain an expression for V (wEE ) V (wSS ) without solv-
ing for the value functions separately by differencing the value re-
cursion equations, yielding
so that
2(1 )
V (wEE ) V (wSS ) = ,
1 (p r )
January 18, 2016 189
and so
1
. (7.5.1)
3p 2q r
Turning to wSS , we have
{r V (wEE ) + (1 r )V (wSS )}
(1 ) (1) + {qV (wEE ) + (1 q)V (wSS )}
or
(1 ) (q r ){V (wEE ) V (wSS )},
requiring
1
. (7.5.2)
p + 2q 3r
Note that (7.5.2) is trivially satisfied if r q (make sure you
understand why this is intuitive).
The two bounds (7.5.1) and (7.5.2) on are consistent if
p 2q r .
The constraint (0, 1) in addition requires
3r p < 2q < 3p r 1.
Solving for the value functions,
1
V (wEE ) 1 p (1 p) 2
=(1 )
V (wSS ) r 1 (1 r ) 0
(1 )
=
1 p (1 (1 r )) 2 1 p r
1 (1 r ) 1 p 2
r 1 p 0
(1 ) 2 (1 (1 r ))
=
(1 ) 1 p r 2r
1 2 (1 (1 r ))
= .
1 pr 2r
190 CHAPTER 7. REPEATED GAMES
7.6 Problems
7.6.1. Suppose G {(Ai , ui )} is an n-person normal form game and GT
is itsQT -fold repetition (with payoffs evaluated as the average). Let
A i Ai . The strategy profile, s, is history independent if for all i
and all 1 t T 1, si (ht ) is independent of ht At (i.e., si (ht ) =
si (ht ) for all ht , ht At ). Let N(1) be the set of Nash equilibria
of G. Suppose s is history independent. Prove that s is a subgame
perfect equilibrium if and only if s(ht ) N(1) for all t, 0 t T 1
and all ht At (s(h0 ) is of course simply s 0 ). Provide examples to
show that the assumption of history independence is needed in both
directions.
7.6.2. Prove the infinitely repeated game with stage game given by match-
ing pennies does not have a pure strategy Nash equilibrium for any
.
(a) Prove that has a unique fixed point. [Hint: Show that is a
contraction.]
(b) Given an explicit equation for the fixed point of .
(c) Interpret the fixed point.
(b) Note that Vei (w) Vi (w) for all w. Denote by wi , the state that
maximizes V ei (w)Vi (w) (if there is more than one, choose one
arbitrarily).
Let aw
i be the action solving the maximization in 7.6.4(a), and
define
Vi (w) := (1 )ui (aw w
i , fi (w)) + Vi ((w, (ai , fi (w)))).
7.6.5. Prove that the profile described in Example 7.1.4 is not a Nash equi-
librium if [1/3, 1/2). [Hint: what is the payoff from always
playing A?] Prove that it is Nash if [1/2, 1).
7.6.6. Suppose two players play the infinitely repeated prisoners dilemma
displayed in Figure 7.6.1.
192 CHAPTER 7. REPEATED GAMES
E S
E 1, 1 `, 1 + g
S 1 + g, ` 0, 0
L R
U 2, 2 x, 0
D 0, 5 1, 1
(a) For what values of the discount factor is grim trigger a sub-
game perfect equilibrium?
(b) Describe a simple automaton representation of the behavior in
which player I alternates between E and S (beginning with E),
player II always plays E, and any deviation results in permanent
SS. For what parameter restrictions is this a subgame perfect
equilibrium?
(c) For what parameter values of `, g, and is tit-for-tat a subgame
perfect equilibrium?
7.6.7. Suppose the game in Figure 7.6.2 is infinitely repeated: Let de-
note the common discount factor for both players and consider the
strategy profile that induces the outcome path DL, U R, DL, U R, ,
and that, after any unilateral deviation by the row player speci-
fies the outcome path DL, U R, DL, U R, , and after any unilat-
eral deviation by the column player, specifies the outcome path
U R, DL, U R, DL, (simultaneous deviations are ignored. i.e., are
treated as if neither player had deviated).
(c) Suppose now x = 4. How does this change your answer to part
7.6.7(b)?
(d) Suppose x = 5 again. How would the analysis in part 7.6.7(b)
be changed if the column player were short-lived (lived for only
one period)?
7.6.9. Consider the prisoners dilemma from Figure 7.1.1. Suppose the
game is infinitely repeated with perfect monitoring. Recall that a
strongly symmetric strategy profile (s1 , s2 ) satisfies s1 (ht ) = s2 (ht )
for all ht . Equivalently, its automaton representation satisfies f1 (w) =
f2 (w) for all w. Let W = {v, v}, v > 0 to be determined, be the
set of continuation promises. Describe a strongly symmetric strat-
egy profile (equivalently, automaton) whose continuation promises
come from W which is a subgame perfect equilibrium for some val-
ues of . Calculate the appropriate bounds on and the value of v
(which may or may not depend on ).
7.6.10. Describe the five state automaton that yields v 0 as a strongly sym-
metric equilibrium payoff with the indicated cycle in Figure 7.4.1.
E S
E 1, 2 1, 3
S 2, 4 0, 0
L R
T 2, 3 0, 2
B 3, 0 1, 1
7.6.12. Consider the stage game in Figure 7.6.4, where player 1 is the row
player and 2, the column player (as usual).
7.6.13. Fix a repeated finite game of imperfect public monitoring (as usual,
assume Y is finite). Say that a player has a profitable one-shot devi-
ation from the public strategy (W , w 0 , f , ) if there is some history
January 18, 2016 195
7.6.15. Player 1 (the row player) is a firm who can exert either high effort (H)
or low effort (L) in the production of its output. Player 2 (the column
player) is a consumer who can buy either a high-priced product, h,
or a low-priced product `. The actions are chosen simultaneously,
and payoffs are given in Figure 7.6.5. Player 1 is infinitely lived, dis-
counts the future with discount factor , and plays the above game
in every period with a different consumer (i.e., each consumer lives
only one period). The game is one of public monitoring: while the
196 CHAPTER 7. REPEATED GAMES
h `
H 4, 3 0, 2
L x, 0 3, 1
actions of the consumers are public, the actions of the firm are not.
Both the high-priced and low-priced products are experience goods
of random quality, with the distribution over quality determined by
the effort choice. The consumer learns the quality of the product af-
ter purchase (consumption). Denote by y the event that the product
purchased is high quality, and by y the event that it is low qual-
quality signal). Assume the
ity (in other words, y {y, y} is the
distribution over quality is independent of the price of the product:
(
p, if a1 = H,
Pr(y | a) =
q, if a1 = L,
(a) Describe the ex post payoffs for the consumer. Why can the ex
post payoffs for the firm be taken to be the ex ante payoffs?
(b) Suppose x = 5. Describe a perfect public equilibrium in which
the patient firm chooses H infinitely often with probability one,
and verify that it is an equilibrium. [Hint: This can be done with
one-period memory.]
(c) Suppose now x 8. Is the one-period memory strategy profile
still an equilibrium? If not, can you think of an equilibrium in
which H is still chosen with positive probability?
201
202 CHAPTER 8. TOPICS IN DYNAMIC GAMES
extracts
ati , if at1 + at2 s t ,
ati = ati
a1 +at2
t st, if at1 + at2 > s t .
It will turn out that in equilibrium, at1 + at2 < s t , so we can ig-
nore the rationing rule (see Problem 8.4.1(a)), and assume player i
derives payoff
ui (s, a) = log ai ,
from (s, a) for all values of (s, a), with the choices restricted to
those satisfying ai < s.
The transition rule is
q(s, a) = 2 max{0, s a1 a2 },
that is, it is deterministic and doubles any leftover stock after ex-
traction. The initial stock is fixed at some value s 0 .
ht = (s 0 , a0 , s 1 , a1 , . . . , s t1 , at1 , s t ) (S A)t S.
Let H t denote the set of all feasible t-period histories (so that s is
consistent with (s 1 , a1 ) for all 1 t). A pure strategy for i
is a mapping
i : t H t Ai .
For any history ht , write the function that identifies the last state
s t by s(ht ). Let G(s) denote the dynamic game with initial state s.
As usual, we have:
Definition 8.1.1. The profile is a subgame perfect equilibrium if
for all ht , |ht := (1 |ht , . . . , n |ht ) is a Nash equilibrium of the dy-
namic game G(s(ht )).
Different histories that lead to the same state are effectively
payoff equivalent. Loosely, a strategy is said to be Markov if at
different histories that are effectively payoff equivalent, the strat-
egy specifies identical behavior. See Maskin and Tirole (2001) for a
discussion of why this may be a reasonable restriction. There are
January 18, 2016 203
i (ht ) = i (ht ).
If the above holds for histories ht and h of possibly different length
(so that t is allowed), the strategy is stationary.
Restricting equilibrium behavior to Markov strategies:
Definition 8.1.3. A Markov perfect equilibrium is a strategy profile
that is a subgame perfect equilibrium, and for which each i is
Markov.
(1 )
= 2V 0 (2(s 2ai (s))).
ai (s)
ai (s) = ks,
1
ai (s) = s.
2
3(1 ) + 0 = 3(1 ),
1
and so the deviation is not profitable if (as before) 3 .
The value when the current state is Sj is
Vi (Sj ) = 0,
while the payoff from a one-shot deviation is (since under the Markov
strategy, a deviation to Ei triggers perpetual E1 E2 ; the earlier devia-
tion is forgiven)
x(1 ) + 2 = (2 + x) x.
(2 + x) x 0
x
.
2+x
Note that
x 1
x 1.
2+x 3
Thus, is an MPE (inducing the outcome path (E1 E2 ) ) if x 1
and
1 x
. (8.1.2)
3 2+x
i.e., chooses p = 1/2, for a payoff of 1/4. This is the optimal seller
mechanism (this can be easily shown using standard mechanism
design techniques).
Suppose now two periods, with common discount factor
(0, 1). If seller chose p 0 = 1/2 in the first period, and buyers with
v > 1/2 buy in period 0, then buyers with value v [0, 1/2] are
left, and then optimal for seller to price p 1 = 1/4. But then buyer
v = 1/2 strictly prefers to wait till period 1 (and so by continuity
so do some buyers with v > 1/2.
Suppose seller makes offers p t in period t, t = 0, 1, and buyers
v < dont buy in period 0. Then, p 1 = /2. If < 1, then
should be indifferent between purchasing in period 0 and period 1,
so that
p 0 = ( p 1 )
= /2
= p 0 = (1 /2) < .
(1 /2)(1 ) + 2 /4.
208 CHAPTER 8. TOPICS IN DYNAMIC GAMES
or
Ws () = max p p + Ws p , (8.2.1)
p
Ws0 () = p() = ,
so that
Ws0 (p()) = p(p()) = p() = 2 .
Substituting,
2 + 2 2 = 0,
or
1 2 + 2 2 = 0. (8.2.2)
Turning to the buyers optimality condition, a buyer with valuation
v = p must be indifferent between accepting and rejecting, so
p p = p p ,
or
1 = (1 ) . (8.2.3)
Solving (8.2.2) for yields
2 4 4 1 1
= = .
2
Since we know < 1, take the negative root,
1 1
= .
Substituting into (8.2.3),
p
= + 1 ,
210 CHAPTER 8. TOPICS IN DYNAMIC GAMES
or
1
= ,
1
so that
p 1 1 1 (1 )
= 1 = .
Note that in this equilibrium, there is skimming: higher valua-
tion buyers buy before lower valuation buyers.
Equilibrium is not unique. It is the only stationary equilibrium.
Let denote real time, the length of a period, and r the rate
of time discount, so that = er . If buyer with valuation v buys
at or after , his utility is no more than er v. Buying in period 0,
he earns v (), and so for close to 1, buyer v buys before .
Note that this is not a uniform statement (since for all and all
there exists v such that v purchases after ).
The Coase conjecture is:
As time between offers shrinks, price charged in first period con-
verges to competitive price, and trade becomes efficient.
8.3 Reputations
Recall the stage game of the the chain store paradox from example
2.2.1, reproduced in Figure 8.3.1.
Two Nash equilibria: (In, Accommodate) and (Out, Fight). Latter
violates backward induction.
Chain store, play the game twice, against two different entrants
(E1 and E2 ), with the second entrant E2 observing outcome of first
interaction. Incumbent receives total payoffs.
Chain store paradox only backward induction (subgame per-
fect) outcome is that both entrants enter (play In), and incumbent
always accommodates.
But, now suppose incumbent could be tough, t : such an incum-
bent receives a payoff of 2 from fighting and only 1 from accom-
modating. Other incumbent is normal, n . Both entrants prior
1
assigns prob (0, 2 ) to the incumbent being t . In second
January 18, 2016 211
Entrant
Out In
Incumbent
0
4 Fight Accommodate
1 1
1 2
Pr{F | t } Pr{t }
Pr{t | F } =
Pr{F }
= ,
+ (1 )
solving
1
=
+ (1 ) 2
gives
= ,
1
1
where < 1 since < 2 .
212 CHAPTER 8. TOPICS IN DYNAMIC GAMES
4, 1 I I 3, 1
F t A
[]
6, 0 O O 5, 0
E2 E2
3, 1 I I 4, 1
F n A
[1 ]
5, 0 O O 6, 0
i.e.,
1
=
.
2
Entrant E1 thus faces a probability of F given by
+ (1 ) = 2.
1
Hence, if < 4 , E1 faces F with sufficiently small probability that
1 1
he enters. However, if ( 4 , 2 ), E1 faces F with sufficiently high
probability that he stays out.
which all entrants stay out, any entry is met with F is a subgame
perfect equilibrium, supported by the threat that play switches
to the always-enter/always-accommodate equilibrium if the incum-
bent ever responds with A. The automaton representation is given
in Figure 8.3.3.
O, IF
IA
w0 wOF wIA
(1 ) + 4 2,
i.e.,
1
.
3
We now consider the reputation game, where the incumbent may
be normal or tough.
The profile in which all entrants stay out, any entry is met with
F is a subgame perfect equilibrium, supported by the threat that
the entrants believe that the incumbent is normal and play switches
2
This is similar to the collusion example discussed in Section 7.3.2. The stage
game is not a simultaneous move game, and so the repeated game does not have
imperfect monitoring. In particular, the incumbents choice between F and A is
irrelevant if the entrant plays O (as the putative equilibrium requires). Subgame
perfection, however, requires that the incumbents choice of F be optimal, given
that the entrant had played I. The principle of one-shot optimality applies here:
the profile is subgame perfect if, conditional on I, it is optimal for the incumbent
to choose F , given the specified continuation play (this is the same idea as that
in footnote 5 in Section 7.3.2).
214 CHAPTER 8. TOPICS IN DYNAMIC GAMES
=(1 )4 + (1 ) + +1 4
=4 3 (1 )
4 3(1 ) = 1 + 3.
January 18, 2016 215
For > 1/3, the outcome in which all entrants enter and the
incumbent accommodates in every period is thus eliminated.
1
(If q > 2 , it is not a best reply for the entrant to play I.) An upper
bound on the number of periods in ht in which an entrant entered
is thus
1
k(t) := #{ : q 2 },
1
the number of periods in ht where q 2 . (This is an upper bound,
and not the actual number, since the entrant is indifferent if q =
1
2
.)
Let := Pr{0 |h } be the posterior probability assigned to 0
after h , where < t (so that h is an initial segment of ht ). If
entrant does not enter, +1 = . If entrant does enter in ht ,
then the incumbent fights and3
Pr{0 , F |h }
+1 = Pr{0 |h , F } =
Pr{F |h }
Pr{F |0 , h } Pr{0 |h }
=
Pr{F |h }
= .
q
Defining
(
q , if there is entry in period ,
q =
1, if there is no entry in period ,
= q +1 ,
1
Note that q < 1 = q = q 2 .
3
Since the entrants action is a function of h only, it is uninformative about
the incumbent and so can be ignored in the conditioning.
January 18, 2016 217
Then,
0 = q0 1 = q0 q1 2
t1
Y
= t q
=0
Y
= t q
{:qt 12 }
k(t)
1
.
2
1
Taking logs, log 0 k(t) log 2 , and so
log 0
k(t) .
log 2
The key intuition here is that since the entrants assign prior pos-
itive probability (albeit small) to the Stackelberg type, they cannot
be surprised too many times (in the sense of assigning low prior
probability to F and then seeing F ). Note that the upper bound is
independent of t and , though it is unbounded in 0 .
Proof. The normal type can guarantee histories in which every entry
results in F by always playing F when an entrant enters. Such be-
havior yields payoffs that are no larger than the incumbents Nash
equilibrium payoffs in any equilibrium (if not, the incumbent has an
incentive to deviate). Since there is positive probability that the in-
cumbent is the Stackelberg type, the history resulting from always
playing F after entry has positive probability. Applying Lemma
218 CHAPTER 8. TOPICS IN DYNAMIC GAMES
8.4 Problems
8.4.1. (a) Suppose (1 , 2 ) is an MPE of the fisheries game from Example
8.1.1 satisfying 1 (s) + (s) < s for all s. Prove that the profile
remains an MPE of the dynamic game where payoffs are given
by
log ai ,
ui (s, a) = n o if a1 + a2 s,
log ai
s , if a1 + a2 > s.
a1 +a2
8.4.2. What is the symmetric MPE for the fisheries game of Example 8.1.2
when there are n players, and the transition function is given by
P
q(s, a) = max 0, s i ai ,
where > 1?
8.4.3. (a) In the MPE calculated in Example 8.1.2, for what values of the
discount factor does the stock of fish grow without bound, and
for which values does the stock decline to extinction?
(b) This MPE is inefficient, involving excess extraction. To see this,
calculate the largest symmetric payoff profile that can by achieved
when the firms choose identical Markov strategies, and prove
that the efficient solution extracts less than does the MPE.
(c) Describe an efficient subgame perfect equilibrium for this game
(it is necessarily non-Markov).
January 18, 2016 219
{1, 2} 9 3 0
{2, 3} 0 9 3
{1, 3} 3 0 9
8.4.6. Consider the model of Section 8.2, but assume the buyers valuation
v can only take on two values, 2 and 3. Moreover, the sellers beliefs
assign probability to the value 2. The sellers cost (value) is zero,
and the buyer and seller have a common discount factor (0, 1].
(a) What are the unique perfect Bayesian equilibria (PBE) of the one
period model (in this model, the seller makes a take-it-or-leave-
it offer to the buyer)?
Consider now the two period model, that is, if the buyer rejects the
offer in the first period, then the seller make a final offer in period
2, after which the game ends.
(c) Prove that, in any PBE, both types of buyer must accept any first
period price strictly smaller than 2.
(d) Prove that, in any PBE, if a first period price p1 > 2 is rejected,
then the sellers posterior in the beginning of the second period
must assign probability at least to the low value buyer.
January 18, 2016 221
1
(e) Suppose = 2 . Describe the unique pure strategy PBE.
1
(f) Suppose = 4 . Prove that there is no pure strategy PBE. [Hint:
Suppose p1 is the first period price in a candidate pure strategy
PBE. How should the seller respond to a rejection of a deviation
to a price p1 p1 ?]
1
(g) Suppose = 4 . Suppose 6/7. Describe the unique PBE
(from part 8.4.6(f), it is necessarily in mixed strategies).4
L R L R
T 2, 3 0, 2 T 3, 3 1, 2
B 3, 0 1, 1 B 2, 0 0, 1
n 0
4
A natural state variable in this game is the posterior probability assigned to
the low value buyer; but with this notion of state, this game does not have a
Markov perfect equilibrium. The notion of weak Markov equilibrium was intro-
duced by Fudenberg, Levine, and Tirole (1985) to accommodate problems like
this. A similar issue arises in the chain store paradox (see Kreps and Wilson
(1982a) and Remark 17.3.1 in Mailath and Samuelson (2006)).
222 CHAPTER 8. TOPICS IN DYNAMIC GAMES
8.4.11. This is a continuation of Problem 7.6.12. Suppose now that the game
with the long-lived player 1 and short-lived player 2s is a game of
incomplete information. With prior probability (0, 1), player 1
is a behavioral type who chooses T in every period, and with prob-
ability 1 , he is a strategic or normal type as described above.
Suppose > 21 . Describe an equilibrium in which the normal type of
player 1 has a payoff strictly greater than 2 for large .
(a) Prove that in any period in which there is entry, the probability
of F , conditional on the incumbent not being 0 is no larger
1
than 2 . Denote this probability by .
(b) Prove that if the incumbent is not 0 , and the entrants never
enter, than with probability one, the entrants posterior proba-
bility on 0 converges to zero.
(c) Let ht be a positive probability period-t history in which every
entry results in F , and there is entry in the last period. Provide
an upper bound on the fraction of periods in ht in which an
entrant enters. [Hint: Express the odds ratio in period + 1
after entry results in F in terms of the odds ratio in period ,
and use part 8.4.12(a).]
Chapter 9
Bargaining1
S 0 = {(1 s1 + 1 , 2 s2 + 2 ) : (s1 , s2 ) S}
1
Copyright January 18, 2016 by George J. Mailath
223
224 CHAPTER 9. BARGAINING
and
d0i = i di + i , i = 1, 2.
Then
fi (S 0 , d0 ) = i fi (S, d) + i , i = 1, 2.
2. SYM (Symmetry)
If d1 = d2 and (s1 , s2 ) S = (s2 , s1 ) S, then
f (S, d) = f (T , d).
f (S, d) s.
1 + d1 d2
s1 =
2
(which is (9.4.1)).
January 18, 2016 225
(s10 , s20 )
( 12 , 12 )
1
xy = 4
x+y =1
1 1
S0 ,
2 2
max{|s10 |, |s20 | : s 0 S 0 }
ht [0, 1]t1 , I1 (ht ) = I1 (ht ) (and similarly for the other strate-
gies). Thus, if a strategy profile is a stationary equilibrium (with
agreement), there is a pair (x , z ), such that I expects x in any
subgame in which I moves first and expects z in any subgame
in which II moves first. In order for this to be an equilibrium, Is
claim should make II indifferent between accepting and rejecting:
1 x = 2 (1 z ), and similarly I is indifferent, so z = 1 x .
[Consider the first indifference. Player I wont make a claim that
II strictly prefers to 1 z next period, so 1 x 2 (1 z ). If
II strictly prefers (1 z ) next period, she rejects and gets 1 z
next period, leaving I with z . But I can offer II a share 1 z this
period, avoiding the one period delay.] Solving yields
x = (1 2 )/(1 1 2 ),
and
z = 1 (1 2 )/(1 1 2 ).
The stationary subgame perfect equilibrium (note that backward
induction is not well defined for the infinite horizon game) is for I
to always claim x and accept any offer z , and for II to always
offer z and always accept any claim x . Note the implicit appeal
to the one-shot deviation principle, which also holds here (the proof
is along similar lines, but simpler than, the proof of Theorem 7.1.3).
and
mi = inf is discounted expected payoff
in any subgame perfect eq of i/j .
Claim 9.2.1. mj 1 i Mi .
Proof. Note first that i must, in equilibrium, accept any offer >
i Mi .
The claim is proved by proving that in every equilibrium, player
js payoff in j/i is at least 1 i Mi . Suppose not, that is, suppose
there exists an equilibrium yielding a payoff uj < 1 i Mi to j.
But this is impossible, since j has a profitable deviation in such an
equilibrium: offer i Mi + , small. Player i must accept, giving j a
payoff of 1 i Mi > uj , for sufficiently small.
Claim 9.2.2. Mj 1 i mi .
= Mj 1 i mi .
Mj 1 i (1 j Mj )
(1 i ) (1 j )
= Mj , Mi .
(1 i j ) (1 i j )
230 CHAPTER 9. BARGAINING
This implies
(1 i ) (1 j )
mi 1 j =
(1 i j ) (1 i j )
and so
(1 j )
mi = Mi = .
(1 i j )
9.2.3 Impatience
In order to investigate the impact of reducing the bargaining fric-
tion intrinsic in impatience, we do the following:
Time is continuous, with each round of bargaining taking
units of time. If player i has discount rate ri ,
i = eri .
1 2 1 er2
x () = =
1 1 2 1 e(r1 +r2 )
and so
1 er2
lim x () = lim
0 0 1 e(r1 +r2 )
r2 er2
= lim
0 (r1 + r2 )e(r1 +r2 )
r2
= ,
r1 + r2
where lHopitals rule was used to get to the second line.
Note that the first mover advantage has disappeared (as it should).
The bargaining is determined by relative impatience.
x 1 [0, 1]
II
O
R A
II
(0, b) (x 1 , 1 x 1 )
x 2 [0, 1]
R A
I
(1 x 2 , 2 (1 x 2 ))
x 3 [0, 1]
Figure 9.3.1: The first two periods when II can opt out only after rejecting
Is proposal.
9.3.1 Version I
Suppose player II can only select outside option when rejecting Is
proposal, and receives b in that period. See Figure 9.3.1 for the
extensive form.
Claim 9.3.1. m2 1 1 M1 .
Proof. Same argument as Claim 9.2.1.
Claim 9.3.2. M1 1 b, M1 1 2 m2 .
Proof. M1 1 b (since II can always opt out). M1 1 2 m2
follows as in case without outside option (Claim 9.2.2).
232 CHAPTER 9. BARGAINING
M2 1 1 m1 = 1 1 (1 2 ) / (1 1 2 ) = (1 1 ) / (1 1 2 ) .
9.3.2 Version II
If II can only select outside option after I rejects (receiving b in that
period, see Figure 9.3.2), then there are multiple equilibria. The
equilibrium construction is a little delicate in this case. In fact, for
many parameter constellations, there is no pure strategy Markov
perfect equilibrium.
There is, however, a Markov perfect equilibrium in behavior strate-
gies. To illustrate the role of the Markov assumption it is useful to
234 CHAPTER 9. BARGAINING
x 1 [0, 1]
II
R A
II
(x 1 , 1 x 1 )
x 2 [0, 1]
R
II A
O
O
I
(0, 2 b) (1 x 2 , 2 (1 x 2 ))
x 3 [0, 1]
Figure 9.3.2: The first two periods when II can opt out only after I rejects
his proposal.
January 18, 2016 235
1 x 1 = 2 (1 x 2 )
and x 2 = 1 x 1 ,
implying the Rubinstein shares for the two players. This can only be
an equilibrium if it is indeed optimal for II not to take the outside
option:
22 (1 1 )
b 2 (1 x 1 ) = .
(1 1 2 )
b 2 (1 x 1 ) = 22 .
Suppose
22 (1 2 )
< b < 22 . (9.3.1)
(1 1 2 )
1 x 1 = 2 (1 x 2 ),
x 2 = 1 x 1 ,
and b = 2 (1 x 1 ).
236 CHAPTER 9. BARGAINING
We then have
b
x1 = 1 ,
2
b
x2 = 1 2 ,
2
22 b
and = .
1 2 (2 b)
It is easy to check that (0, 1) under (9.3.1).
Note the critical role played by the timing of the decision on the
outside option. In particular, even though I strictly prefers that II
never take the outside option, I is not able to prevent it, since Is
offer is only made after II decides not to take it. In Figure 9.3.1, on
the other hand, I always has the option of claiming 1 x 1 > b.
This implies
1 + dj di 1 + di dj
mi 1 (1 ) dj = = Mi
(1 + ) (1 + )
and so
1 + di dj
mi = Mi = .
(1 + )
Now, we are interested in the payoffs as 1, and
1 + di dj
mi ,
2
so that Is share is
1 + d1 d2
x = . (9.4.1)
2
For much more on bargaining, see Osborne and Rubinstein (1990).
9.5 Problems
9.5.1. A three person bargaining problem is a pair < S, d >, where S
R3 is compact and convex, d S, and s S such that si > di
for i = 1, 2, 3. Let B denote the collection of bargaining problems,
and as for the two person case, a bargaining solution is a function
f : B R3 such that f (S, d) S. The Nash axioms extend in the
obvious manner. Prove that if a bargaining solution satisfies INV,
SYM, IIA, and PAR, then
9.5.3. (a) Give the details of the proof of Claim 9.3.4. (A few sentences
explaining why it works is sufficient.)
(b) Give the details of the proof of Claim 9.3.6. (A few sentences
explaining why it works is sufficient.)
9.5.4. We will use the finite horizon bargaining result from question 9.5.2
to give an alternative proof of uniqueness in the Rubinstein model.
January 18, 2016 239
9.5.5. Let G(n) denote Rubinsteins alternating offer bargaining game with
discounting with the following change: the share going to player
1 must be an element of the finite set A(n) = {0, n 1
, . . . , (n1)
n , 1},
where n {1, 2, 3, . . .}. Suppose the two players have identical dis-
count factors .
(a) Show that for any n, there exists 0 (0, 1) such that for any
(0 , 1), the continuum case proof that there is a unique
partition of the pie achievable in a subgame perfect equilibrium
fails for G(n). Show that any share to player 1 in A(n) can be
achieved in a subgame perfect equilibrium of G(n).
(b) Prove or provide a counterexample to the following claim: For
any (0, 1) and any > 0, there exists n0 such that for
any n > n0 the share going to player 1 in any subgame perfect
equilibrium is within of the Rubinstein solution.
(c) Suppose
1 k
k.
1+ n
Prove that this game does not have a symmetric stationary equi-
librium. Describe the two stationary equilibria that, for large n,
approximates the Rubinstein solution.
9.5.6. There is a single seller who has a single object to sell (the sellers
reservation utility is zero). There are two potential buyers, and they
each value the object at 1. If the seller and buyer i agree to a trade at
price p in period t, then the seller receives a payoff of t1 p, buyer
i a payoff of t1 1 p , and buyer j i a payoff of zero. Con-
sider alternating offer bargaining, with the seller choosing a buyer
to make an offer to (name a price to). If the buyer accepts, the
game is over. If the buyer rejects, then play proceeds to the next pe-
riod, when the buyer who received the offer in the preceding period
makes a counter-offer. If the offer is accepted, it is implemented.
240 CHAPTER 9. BARGAINING
If the offer is rejected, then the seller makes a new proposal to ei-
ther the same buyer or to the other buyer. Thus, the seller is free
to switch the identity of the buyer he is negotiating with after every
rejected offer from the buyer.
(a) Suppose that the seller can only make proposals in odd-numbered
periods. Prove that the sellers subgame perfect equilibrium
payoff is unique, and describe it. Describe the subgame perfect
equilibria. The payoffs to the buyers are not uniquely deter-
mined. Why not?
(b) Now consider the following alternative. Suppose that if the
seller rejects an offer from the buyer, he can either wait one
period to make a counteroffer to this buyer, or he can immedi-
ately make an offer to the other buyer. Prove that the sellers
subgame perfect equilibrium payoff is unique, and describe it.
Describe the subgame perfect equilibria. [Cf. Shaked and Sut-
ton (1984).]
2
(d) Suppose 0 < w 1+ . There is a subgame perfect equilib-
rium which gives the same equilibrium outcome as that in part
9.5.8(a). Describe it and verify it is an equilibrium. (Hint: Use
the equilibrium you found in part 9.5.8(c).)
(e) Describe an equilibrium with inefficient delay.
Chapter 10
Appendices1
bi (si (h)) = 1, h Hi ,
and
bi (a) = 0, a hHi {si (h)}.
Natures move is now denoted by b0 . Given b = (b0 , b1 , . . . , bn ),
the game is played as follows: play begins with player (t0 ) ran-
domly selects an action in A(h) according to the distribution b(h) (a) aA(h) .
All distributions are independent. The induced distribution on Z is
calculated as follows. Define recursively p n (t) = p(p n1 (t)) for
1
Copyright January 18, 2016 by George J. Mailath
243
244 CHAPTER 10. APPENDICES
Thus,
n o n o
k k k+1
si Si : h R(si ), si (h ) = ak = si Si : h R(si )
and
Y
bi = bi (ak )
k=1
,
Y X X
= i (si ) i (si )
k=1 si Si s.t. si Si s.t.
hk R(si ) and si (hk )=ak hk+1 R(si )
,
X X
= i (si ) i (si ) .
si Si : si Si :
h` R(si ),si (h` )=a` h1 R(si )
P
Let : Si (0, 1) be a function satisfying si Si (si ) < 1 for
all i. The associated perturbed game, denoted (G, ), is the normal
form game {(R1 , v1 ), . . . , (Rn , vn )} where
Ri = {i (Si ) : i (si ) (si ), si Si }
si BRi (i ) = i (si ) .
248 CHAPTER 10. APPENDICES
P
Since k , there exists k0 such that k > k0 implies si Si k (si ) <
1 for all i.
Let m = min {i (si ) : i (si) > 0, i}. There exists k such that,
00
for all k > k00 , ik (si ) i (si ) < m/2. Suppose k > max {k0 , k00 , 2/m}.
Then ik (si ) k (si ). [If i (si ) = 0, then immediate. If i (si ) > 0,
then ik (si ) > i (si ) m/2 > m/2 > 1/k.]
k k
Since i is a best reply to i , if si is not a best reply to i ,
k
then i (si ) = 0. But this implies that is an equilibrium of (G, k )
(since si is played with minimum probability).
January 18, 2016 249
10.3 Problems
10.3.1. By explicitly presenting the completely mixed trembles, show that
the profile L` is normal form trembling hand perfect in the following
game (this is the normal form from Example 2.3.4):
` r
L 2, 0 2, 0
T 1, 1 4, 0
B 0, 0 5, 1
Carlsson, H., and E. van Damme (1993): Global Games and Equi-
librium Selection, Econometrica, 61(5), 9891018.
Cho, I.-K., and D. Kreps (1987): Signaling Games and Stable Equi-
libria, Quarterly Journal of Economics, 102(2), 179221.
251
252 BIBLIOGRAPHY
Morris, S., and H. S. Shin (2003): Global Games: Theory and Ap-
plications, in Advances in Economics and Econometrics (Proceed-
ings of the Eighth World Congress of the Econometric Society), ed.
by M. Dewatripont, L. Hansen, and S. Turnovsky. Cambridge Uni-
versity Press.
257
258 INDEX
payoff dominance, 42
perfect Bayesian equilibrium, 122
almost, 122
signaling games, 135, 140
weak, 121
perfect information, 12
perfect public equilibrium, 186
purification, 54, 80
Harsanyi, 55
renegotiation, 43
Riley outcome, 142
risk dominance, 43
security level, 34
sequential equilibrium, 125
sequential rationality, 117
strategy
behavior, 39
extensive form, 11
Markov, 203
mixed, 33
formal treatment, 89
reduced normal form, 14
strongly symmetric, 159
subgame, 28
subgame perfect equilibrium, 28
system of beliefs, 115