4 Game Complete
4 Game Complete
Masaki Miyashita
Chapter 4:
Games with Compete Information
What is Game Theory?
• The output is
• Each bidder has a valuation vi ≥ 0 for the good, which is unknown to the others.
• The highest bidder wins the object and pays the second highest bit.
ui : A1 × ⋯ × An → ℝ .
a := (a1, …, an) .
• Each player i has a utility function, de ned over the set of strategy pro les A .
• Timeline:
• Payoff ui(a1, …, an) materializes to each player. The game ends right away.
Example 1: Partnership
N = {1,2}
A1 = A2 = {E, S}
N = {1,2}
A1 = A2 = ℝ+
N = {1,…, n}
A1 = ⋯ = An = ℝ+
vi − maxj≠i bj if bi > maxj≠i bj
{0
ui(bi, b−i) =
if bi < maxj≠i bj
• Static & Simultaneous Choice: players choose their actions only once and do
so without knowing the choices of other players.
• Complete Information: Each player has full knowledge of the game’s details,
including one’s own action set Ai and payoff function ui, as well the action sets Aj
• Moreover, each player knows the fact that the other players also fully
understand these details.
• 3rd order: All agents know “all agents know “all agents are rational.””
⋮
Comments
• Each child can see the faces of others but cannot tell if one’s own face is muddy.
• The children are told that those with muddy faces must step forward (but anyone
with a clean face should not step forward).
• Suppose there are k children with muddy faces. Can they eventually gure out
that they are among those with muddy faces?
fi
Equilibrium
• An action pro le forms an equilibrium when each player’s choice is, in some
sense, “optimal.”
• In contrast, if there exists an action that dominates all other actions for a player,
that action should be choose de nitely.
fi
Dominant Strategy Equilibrium
De nition.
• We say ai strictly dominates bi if ui(ai, a−i) > ui(bi, a−i) for all a−i ∈ A−i .
De nition.
• We say ai weakly dominates bi if ui(ai, a−i) ≥ ui(bi, a−i) for all a−i ∈ A−i
and ui(ai, a′−i) > ui(bi, a′−i) for some a′−i ∈ A−i .
De nition.
• A strategy pro le a* = (a*
1
, …, a*
n ) ∈ A is a strict DSE in G if a*
i
is strictly
• In contrast, both get strictly better off if they choose move to (E, E) .
q*
i
(qj) is derived as the output level that maximizes ui( ⋅ , qj) .
a−c qj
− if qj ≤ (a − c)/b,
q*
i
(qj) = 2b 2
0 if qj > (a − c)/b .
fi
fi
fi
fi
Example 3: Second-price Auction
• Fix any b−i ∈ A−i and let x ≡ maxj≠i bj denote the strongest opponent’s bid.
• If x < vi or x < bi, then the payoff remains unchanged between bi and b*
i
.
• In contrast, if vi < x < bi, the player receives 0 by bidding vi, while the payoff
wold be negative vi − bi < 0 with bi .
fi
fi
Dominant Strategy Equilibrium
• Cournot Duopoly: any output choice qi ∈ [0, (a − c)/2b] can be a unique best
response against some qj ≥ 0. So, there exists no weak DSE.
• The dominance principle can provide strong predictions when a game has a DSE.
• However, it may be too stringent for some games; for instance, in the Cournot
duopoly, no strategy pro le survives the dominance criterion.
fi
Nash Equilibrium
Nash Equilibrium
• The most commonly used equilibrium concept in game theory is Nash equilibrium.
• Captures a steady state of the play in a game, where each player holds the
correct belief about others’ behavior and acts rationally.
De nition.
ui(a*
i
, a*
−i
) ≥ ui(ai, a*
−i
), ∀i ∈ N, ∀ai ∈ Ai .
fi
fi
Example 2: Cournot Duopoly
• Each rm’s best-response strategy was:
a−c qj
− if qj ≤ (a − c)/b,
q*
i
(qj) = 2b 2
0 if qj > (a − c)/b .
and (ii) q*
2
best responds to q*
1
.
( 3 3 )
a−c a−c
(q*, q*) =
1 2
, .
2. NE is weaker than DSE in the sense that if a strategy pro le forms a Nash
equilibrium, then it is a weak DSE (and thus, a strict DSE as well).
• In other words, σi is a lottery over pure strategies, where σi(ai) represents the
probability that player i chooses ai according to σi .
Notation.
• For a mixed strategy pro le σ = (σ1, …, σn) ∈ Σ, we extend each player’s payoff
functions by considering expected utility:
∑
ui(σ) = σ(a)ui(a) where σ(a) = σ1(a1) × ⋯ × σ(an) .
a∈A
• Since pure strategies are a special type of mixed strategies, the next de nition
generalizes the previous one.
De nition.
ui(σ*
i
, σ*
−i
) ≥ ui(σi, σ*
−i
), ∀i ∈ N, ∀σi ∈ Σi .
• Equivalently, σ* is a NE if
• In other words, for any σi, σ′i ∈ Σi and λ ∈ [0, 1], it holds that
form a NE, each player must be indifferent among all pure strategies assigned
with positive probabilities.



fi
fi
Indifference Principle
ui(σ*
i
, σ*
−i
) = ui(ai, σ*
−i
), ∀ai ∈ supp(σ*
i
),
ui(σ*
i
, σ*
−i
) ≥ ui(ai, σ*
−i
), ∀ai ∉ supp(σ*
i
).
fi
fi
Indifference Principle
• A norma-form game is said to be nite when the set of players N and players’
action sets A1, …, An are all nite.
• The proof is based on Fixed Point Theorem; see Appendix for details.
Proof
fi
fi
fi
Existence of Nash Equilibrium
• The theorem guarantees the existence of NE for any nite game, but it does not
address the issue of uniqueness.
• Also, the theorem does not provide any ef cient way to calculate NE.
• No general method for solving games with continuous actions even with two
players, while some classes of games are known to have ef cient computation
algorithms.
fi
fi
fi
fi
fi
Discussion: Nash Equilibrium
• See Chapter 3.2 of Course of Game Theory by Osborne and Rubinstein (1995)
for a comprehensive discussion.
NE as Steady State
• I.e., once the players reach NE, the dynamics cease to evolve.
∑
μi(a−i)u(σi, a−i) .
a−i∈Si
Belief about
strategic uncertainty
NE as Bayesian Rationality
• In NE σ* = (σ*
1
, …, σ*
n) …
μ*
i
(a−i) = σ*(a ) × ⋯ × σ*
1 1
(a ) × σ*
i−1 i−1
(a ) × ⋯ × σ*
i+1 i+1 n (an) .
• In response to μ*
i
, each player chooses the optimal strategy σ*
i
that
maximizes the subjective expected utility:
∑
σ*
i
maximizes μ*
i
(a−i)u(σi, ai) .
a−i∈Si
• The strong assumption: Each player’s belief must coincide perfectly with the
“actual” action distribution, resulting from the play of opponents.
∑
a*
i
∈ arg max μi(a−i)ui(σi, a−i) .
σi∈Σi
a−i∈Si
• I.e., a*
i
is optimal not only for the equilibrium belief but also for all possible beliefs
(akin to Bewley’s representation under Knightian uncertainty).
• Each of two players i ∈ {R, B}, say Red and Black, simultaneously chooses one
of four cards ai ∈ {1, 2, 3, J} (Ace, Two, Three, and Joker).
Red / Black 1 2 3 J
1 0,1 1,0 1,0 0,1
2 1,0 0,1 1,0 0,1
3 1,0 1,0 0,1 0,1
J 0,1 0,1 0,1 1,0
• In words:
• Red wins iff both players choose J, or they choose different numbers.
• Black wins iff exactly one player chooses J, or both choose the same number.
Example 4: O’Neill Game
Black
Experiment
Theoretical prediction
Red
Summary: NE and DSE
• NE and DSE can be related through the lens of Bayesian rationality:
In Nash equilibrium…
players optimize their strategic choices with respect to their equilibrium beliefs,
and those beliefs are consistent with the actual choices of opponents.
In DSE…
players optimize their strategic choices with respect to all possible beliefs.
• DSE provides a robust prediction, but it need not exist in several games.
• Many strategic situations are inherently dynamic, where players make decisions at
multiple (and possibly different) points in time.
• P1 moves rst.
• Their dynamic decision making can be summarized as the following game tree:
• Each node is labeled with a player, indicating who makes a decision at that point.
• Each edge (arrow) represents an action available to the player at each node.
• At each terminal node, the resulting payoffs for all players are speci ed.
fi
Extensive-form Game
where (M, H) describes the P2’s contingent plan to take M when P1 takes M
and H when P1 takes H, etc.
fi
NE in Extensive-form Game
• In the coordination game, there are three (pure strategy) Nash equilibria.
backward induction.
Example 2’: Stackelberg Competition
• A dynamic version of Cournot Competition
Exercise.
• For each rm, compare the equilibrium production levels with those in the NE of
Cournot Competition; are they higher or lower? What about pro ts?
fi
fi
fi
fi
Example 3’: Ascending-price Auction
• The auctioneer continuously increases the price, and at each moment, participants
must decide whether to “stay” in the auction or “leave.”
• As long as multiple participants choose to stay, the price keeps increasing. The
auction terminates when only one participant remains. This last participant is the
winner, who receives the good & pays the nal price.
• In the previous examples, actions are observable & players are perfectly informed
about all past decisions.
• In other situations, each player may have only partial information about actions
taken previously when making a decision.
• When multiple nodes belong to the same information set, it means that the player
cannot distinguish between these nodes.
(Im)perfect Information
• Note that the player has to choose the same action for all decisions nodes
within the same information set.
• The subgame starts at a particular decision node and includes all subsequent
nodes and edges stemming from that node.
• Any information set intersecting the subgame must lie entirely within it.
Exercise. How many subgames does the following game have? Write down the set
of strategies for each player and derive SPE.
(Im)perfect Information
• An extensive-form game is said to be nite when the set of nodes is nite. It has
perfect information if every information set consists of a single decision node.
• For each of the last decision nodes, we choose an optimal action for the player
acting at that node and replace the node with a terminal node that re ects the
resulting payoff pro le from the chosen optimal action.
• Repeat the same procedure, working backward through the game tree, until
reaching the initial node.
fi
fi
fi
fl
fi
Existence of SPE
• Two players, Black and White, alternately put one stone on a rectangle table.
• Players can put a stone anywhere on the table as long as it doesn’t overlap
with any existing stones.
• The game continues until there is no space left to place a new stone. The winner is
the player who placed the last stone.
Question: Which player has a winning strategy, and what is the winning strategy?
Appendix:
Proof of Theorem 1
Go Back
Existence of Nash Equilibrium
• In other words, BRi(σ−i) is the set of mixed strategies that maximize i ’s expected
payoff when other players are expected to play σ−i .
xn → x ⟹ f(xn) → f(x) .
• Then, any continuous function f : X → X admits at least one xed point, i.e.,
• F has a closed graph if for any sequence {xn} and {yn} in X such that
yn ∈ F(xn), if xn → x and yn → y, then y ∈ F(x) .
• Let’s check the Kakutani’s assumptions to apply the xed point theorem!
fi
fi
• It is straightforward to check that:
• Consider any sequences {σ k} and {ρ k} of mixed strategy pro les such that
ui(ρik, σ−i
k k
) ≥ ui(ai, σ−i ), ∀k ≥ 1, ∀i ∈ N, ∀ai ∈ Ai .
• Therefore, Kakutani’s xed point theorem implies there exists σ* ∈ Σ such that
σ* ∈ BR(σ*),
which is a Nash equilibrium.
Go Back
fi
fi