Math 99 Rhandout 2020
Math 99 Rhandout 2020
Peter S. Park
Definition
A choice correspondence c : M(X ) ⊸ X is a correspondence defined on
the menus A ⊆ X that satisfies ∅ =
̸ c(A) ⊆ A, representing the DM’s
possible choices when choosing between the alternatives in A.
Preferences: Preference relation
Definition
A preference relation ≾ is a binary relation (i.e., a subset of X × X ) where
x ≾ y means that the DM weakly prefers y over x.
with the utility function being the most parsimonious, perhaps the
reason for its ubiquity in economics.
Not all choice functions can be represented as preference relations and
not all preference relations can be represented as utility functions.
Let’s try to see precisely when we can do so.
Preferences: Revealed preference
Definition
A preference relation ≾ is complete if and only if any two x, y ∈ X , we
have x ≾ y or y ≾ x.
Preferences: Axiom of transitivity
Even if ≾ is complete, there might not be a maximal element.
For example, suppose X is the list of candidates running for a political
position and ≾ denotes pairwise comparisons for candidates x, y ∈ X .
Define majority voting by the voting system that crowns the candidate
who wins more votes in pairwise comparisons with every other
candidate.
It is well-known that majority voting is not deterministic, i.e.,
{x ∈ X : y ≾ x for all y ∈ X }
x1 ≺ x2 ≺ · · · xn ≺ x1 ,
Definition
A preference relation ≾ is transitive if and only if for every x, y , z ∈ X such
that x ≾ y and y ≾ z, we have x ≾ z.
Definition (WARP)
Whenever for menus A and B and elements x, y ∈ A ∩ B we have x ∈ c(A)
and y ∈ c(B), it follows that x ∈ c(B).
Exercise
Prove that a choice correspondence c satisfies WARP if and only if it
satisfies Sen’s Properties α and β.
Preferences: Representations of rational preference
Theorem (Arrow)
For a choice correspondence c, there exists a complete and transitive
preference ≾ such that c = c≾ if and only if c satisfies WARP. The only
such preference relation is the revealed preference relation.
x ≾ y ⇐⇒ y ∈ c({x, y }).
Theorem (Sen)
Suppose that a choice correspondence c is defined on a subset M ⊆ M(X )
that contains all menus of up to three elements. There exists a complete
and transitive preference ≾ such that c = c≾ if and only if c satisfies
WARP. The only such preference relation is the revealed preference relation.
Exercise
Prove the above theorem. Does it still hold if we only assume that M
contains all menus of up to two elements?
Preferences: Utility representations
Proposition
Suppose X is finite or countably infinite. The preferences that correspond
to utility-function equivalence classes are precisely those that correspond to
complete and transitive preference relations.
Proof. ( =⇒ ): Clear.
Preferences: Rational = utility-representable
( ⇐= ): For every x ∈ X , define
W (x) = {z ∈ X : z ≾ x}.
Exercise
Why can’t uncountably infinitely many positive-measure intervals disjointly
fit into R?
Preferences: Uncountably infinite case, pt. 2
Definition
A preference relation ≾ is continuous if and only if the upper- and
lower-contour sets, {y ∈ X : x ≾ y } and {y ∈ X : y ≾ x}, are closed.
Theorem (Wold)
Let ∼ be an equivalence relation on [0, 1] such that every equivalence class
is a closed interval. Then, there exists a continuous and nondecreasing map
T : [0, 1] → [0, 1] such that T (x) = T (y ) if and only if x ∼ y .
Exercise
Prove this. (Hint: If we have a sequence of continuous and nondecreasing
→ [0, 1] such that |Tn+1 (x) − Tn (x)| ≤ Mn for constants
maps Tn : [0, 1] P
Mn that satisfy ∞ n=0 Mn < ∞, we have a well-defined continuous and
nondecreasing map T (x) = limn→∞ Tn (x) = ∞
P
n=0 n+1 (x) − Tn (x),
T
since the series converges uniformly by the Weierstrass M-test. Construct
the Tn by iteration so that their uniform limit T will have the desired
property that T (x) = T (y ) if and only if x ∼ y .)
Preferences: Continuous utility representations, pt. 2
We define an equivalence relation ∼ on [0, 1]:
Fix a subset S ⊆ [0, 1].
A lacuna of S is an interval with at least two points that is disjoint
from S and has upper and lower bounds in S.
A gap is a maximal lacuna.
If there are two gaps of S of the form [a, b) and (b, c], then label
[a, c] as a equivalence class.
Any half-open, half-closed gap of S that does not satisfy the above
condition has its closure as an equivalence class.
All other points are single-element equivalence classes.
Proof. ( ⇐= ): Immediate.
Observe that f = T ◦ u is another utility representation of ≾.
We now show that f is continuous. First, we show that it is
upper-semicontinuous, i.e., f −1 ([r , ∞)) is closed for all r .
Case 1: r ∈ f (X ), say f (a) = r .
Case 2: r bounds f (X ) from above.
Case 3: r bounds f (X ) from below.
Case 4: The gap of f (X ) containing r is open, say (c, d).
Case 5: The gap of f (X ) containing r is closed, say [c, d].
These exhaust the casework, since no gap of f (X ) is half-open and
half-closed.
The proof of lower-semicontinuity is analogous.
Preferences: Proof of the main theorem, pt. 2
Theorem (Eilenberg)
Suppose X is connected and separable. The preferences that correspond to
continuous-utility-function equivalence classes are precisely those that
correspond to complete, transitive, and continuous preference relations.
Proof. ( =⇒ ): Clear.
( ⇐= ): By the main theorem, it suffices to show that ≾ has a utility
representation (we don’t need to check continuity).
By assumption, X has a countable dense subset Z = {zn : n ∈ Z≥0 }.
For every x ∈ X , define
N(x) = {n ∈ Z≥0 : zn ≺ x}
Let µ be an arbitrary probability distribution on Z>0 whose support is
all of Z>0 .
P
Let’s show that the utility function u(x) = n∈N(x) µ(n) works, i.e.,
x ≾ y if and only if u(x) ≤ u(y ).
Preferences: Eilenberg’s result, pt. 2
Theorem (Debreu)
Suppose X is second-countable. The preferences that correspond to
continuous-utility-function equivalence classes are precisely those that
correspond to complete, transitive, and continuous preference relations.
Proof. ( =⇒ ): Clear.
( ⇐= ): By our main theorem, it suffices to show that ≾ has a utility
representation (we don’t need to check continuity).
By assumption, X has a countable basis {Un }n∈Z>0 .
For every x ∈ X , define
N(x) = {n ∈ Z≥0 : z ≾ x for all z ∈ Un }.
Let µ be an arbitrary probability distribution on Z≥0 whose support is
all of Z≥0 .
P
Let’s show that the utility function u(x) = n∈N(x) µ(n) works, i.e.,
x ≾ y if and only if u(x) ≤ u(y ).
Preferences: Debreu’s result, pt. 2
Definition
A utility function u : ∆(X ) → R is a von Neumann–Morgenstern (vNM)
utility function if and only if there exists U = (U1 , . . . , Un )—corresponding
to the outcomes x1 , . . . , xn comprising X —such that
u(v ) = U · v ∀v ∈ ∆(X ).
i.e., vNM utility functions are precisely the ones that are linear.
In general, linearity is not preserved by a strictly increasing
transformation T : R → R.
Exercise
Suppose that u : ∆(X ) → R is a vNM utility function representing a
preference relation ≾ on ∆(X ). Then, f : ∆(X ) → R is also one if and
only if ∃a and b > 0 such that f (v ) = a + bu(v ) ∀v ∈ ∆(X ).
Preferences: Axiomatizing rationality under risk
αv + (1 − α)v ′′ ∼ v ′ .
v ≾ v ′ ⇐⇒ αv + (1 − α)v ′′ ≾ αv ′ + (1 − α)v ′′ .
Preferences: Allais’ paradox
v ≾ v ′ ⇐⇒ αv + (1 − α)v ′′ ≾ αv ′ + (1 − α)v ′′ .
Exercise
Prove the ⇐= direction.
Proof: ( =⇒ ): By construction of u.
Let v , v ∈ ∆(X ) be the most and least preferred lottery by the
preference ≾.
If v ∼ v , then take u to be any constant function.
The remaining case is v ≺ v .
Preferences: Expected utility representation, pt. 2
Let’s show that for any v ∈ ∆(X ), there exists a unique λv such that
λv v + (1 − λv )v ∼ v .
v ≺ αv + (1 − α)v ≺ βv + (1 − β)v ≺ v
The first and last ≺ follow from the independence axiom, since
v = αv + (1 − α)v and v = βv + (1 − β)v .
The second ≺ also follows from the independence axiom, since
αv + (1 − α)v = (β − α)v + αv + (1 − β)v and
βv + (1 − β)v = (β − α)v + αv + (1 − β)v .
Preferences: Expected utility representation, pt. 3
We show that u(v ) = λv is a vNM utility representation of ≾.
Represents ≾:
w ≾ z ⇐⇒ λw v + (1 − λw )v ≾ λz v + (1 − λz )v ⇐⇒ λw ≤ λz .
vNM: Need to show linearity, i.e., for all α ∈ [0, 1] and w , z ∈ ∆(X ),
αw + (1 − α)z
∼ α (u(w )v + (1 − u(w ))v ) + (1 − α) (u(z)v + (1 − u(z))v )
∼ (αu(w ) + (1 − α)u(z)) v + (1 − αu(w ) − (1 − α)u(z)) v .
P
Updated approach to utilitarianism: max i αi Ui
vNM setting imbues utility with cardinal meaning.
After utility representations Ui of preferences ≾i are fixed, scaling the
moral weights αi may help justify interpersonal comparisons of utility.
Decision-making under risk as if behind the veil of ignorance.
Preferences: Monetary risks
Definition
A vNM utility function U on ∆(X ) is risk-averse if and only if for any
non-point-mass F ∈ ∆(X ), the utility function U prefers, over F , the
point-mass lottery δEA⇝F (A) yielding the mean payoff
Z
EA⇝F (A) = xdF (x)
x∈X
with probability 1.
Theorem (Jensen)
A vNM utility function U on ∆(X ) is risk-averse if and only if u is concave.
Exercise
Why does the above suffice to prove Jensen’s inequality for a general
probability distribution F ∈ ∆(X )?
Preferences: Risk aversion, pt. 3
Proof by induction. Base case is trivial.
Suppose claim holds for all lotteries supported on < n points.
Consider arbitrary F supported on x1 , . . . , xn ∈ X . Then,
n
X
EA⇝F (u(A)) = p(x1 )u(x1 ) + p(xi )u(xi )
i=2
n
X p(xi )
= p(x1 )u(x1 ) + (1 − p(x1 )) u(xi )
1 − p(x1 )
i=2
n
!
Xp(xi )
≤ p(x1 )u(x1 ) + (1 − p(x1 )) u xi
1 − p(x1 )
i=2
n
!
X p(xi )
≤ u p(x1 )x1 + (1 − p(x1 )) xi
1 − p(x1 )
i=2
= u (EA⇝F (A))
by the inductive hypothesis and the definition of concavity. Q.E.D.
Preferences: Measuring risk aversion
Definition
The certainty equivalent c(F , u) is the number of dollars c such that
Z
u(c) = u(x)dF (x).
x∈X
Definition
For C 2 Bernoulli utility functions u, define the Arrow–Pratt coefficient
u ′′ (x)
A(x, u) = − .
u ′ (x)
Preferences: Measuring risk aversion, pt. 3
Theorem (Pratt)
The following are equivalent.
1 Whenever u prefers a lottery F to a point-mass lottery δx , then so
does v .
2 For every F ∈ ∆(X ), we have c(F , u) ≤ c(F , v ).
3 There exists an concave order-preserving transformation T such that
u = T ◦ v.
4 For every x ∈ X , we have A(x, u) ≥ A(x, v ).
Exercise
Prove this.
Preferences: Descriptive decision-making under risk
Risk-seeking behavior.
Gambling addicts repeatedly take negative expected value gambles
that increase variance.
Can a risk-seeking Bernoulli utility function (i.e., with positive
curvature) be consistent with rationality?
Preferences: Descriptive decision-making under risk, pt. 2
We have a lot of useful tools that can deal with utility functions.
The assumption that human decision-making follows a utility
function—the assumption of rationality—makes models parsimonious
and tractable.
Helps explain the omnipresence of utilitarianism in economic theory.
(Economists generally do not make normative moral claims about
what society’s objective should be, e.g., the social choice function.
They only make recommendations on how to achieve such an
objective when one is chosen by others, e.g., voters and politicians.)
We now know of several empirical inconsistencies regarding this
assumption of rationality.
We will explore more of these issues of descriptive decision theory later
in the course.
General equilibrium (GE) theory: Invisible-Hand Hypothesis
such that p t yi ≤ p t ωi .
If the domain is compact and ui is continuous, the maximum of the
above utility maximization program is realized.
GE theory: A notion of efficiency
max ui (xi )
x∈X
I
X I
X
s.t. xi ≤ ωi and uj (xj ) ≥ ūj ∀j ̸= i
i=1 i=1
Theorem (Walras)
Let u be a monotone utility function and p ≫ 0. For any v , w ∈ X that
maximize the utility maximization program
max u(z)
z∈X
such that p t z ≤ c
for some c = cv , cw > 0, we have the following.
1 The constraint is satisfied with equality, i.e., p t v = cv and p t w = cw .
2 We have u(v ) ≤ u(w ) ⇐⇒ cv ≤ cw .
Exercise
Prove Walras’ law.
GE theory: First welfare theorem
Exercise
Prove this.
GE theory: Proof of the second welfare theorem
and
B = B1 + · · · + BI .
Since x is Pareto optimal, it follows that 0 ∈
/ B.
Each Bi is open convex. Indeed, since ui is strictly concave,
Theorem (Negishi)
Suppose U is convex. A Pareto efficient allocation is precisely an argument
maximizing, for some α ∈ RI≥0 \ {0}, the Negishi program
I
X I
X I
X
max αi ui (xi ) s.t. xi ≤ ωi .
x∈X
i=1 i=1 i=1
Exercise
Prove Negishi’s Theorem.
GE theory: Negishi’s theorem and social choice
min f (x)
x∈X
Theorem (Karush–Kuhn–Tucker)
Consider x̄ ∈ X that satisfies all the constraints of the optimization
program. Suppose further that for I, the gradients ∇gi for i ∈ I and ∇hj
for all j are linearly independent. A necessary condition for x̄ to be a local
minimum is that there exists u ≥ 0 and w such that
Consider the Pareto and Negishi programs for which all relevant
functions are assumed to be continuously differentiable.
The negated form of the Pareto program has objective function
−ui ◦ Coordi ,
and
I
X I
X
Gℓ = (xi )ℓ − (ωi )ℓ for all 1 ≤ ℓ ≤ L.
i=1 i=1
GE theory: Deriving the KKT condition, pt. 2
KKT necessary condition restricted to the vector entries corresponding
to agent i:
0
..
X L XL .
KKT
∇ui (x̄i ) = λGℓ ∇i Gℓ (x̄) = 1
λGℓ
ℓ=1 ℓ=1 ..
.
0
where the last vector only has 1 in the ℓth entry.
Thus, λGℓ is the partial derivative of ui with respect to the ℓth entry,
evaluated at x̄.
This means that the KKT necessary condition restricted to the vector
entries corresponding to agent j ̸= i is:
L
X
KKT
λgj ∇uj (x̄j ) = λGℓ ∇j Gℓ (x̄) = ∇ui (x̄i ).
ℓ=1
GE theory: Deriving the KKT condition, pt. 3
Exercise
Derive the KKT necessary condition for the Negishi program. Suppose that
U is convex, so the Pareto and Negishi programs are equivalent. If x̄ is a
solution to both problems, then show that λgj defined above = αj /αi .
(Intuition: At a local optimum, all agents have the same gradient
direction—the direction of maximal utility ascent—since otherwise, some of
them can trade so that they all gain. Furthermore, the magnitude ratio of
the agents’ gradient vectors is α1 : · · · : αI , since otherwise the agent
shortchanged by the gradient magnitude ratio can take goods from agents
who were overrewarded by the gradient magnitude ratio to increase
I
X
αi ui (xi ),
i=1
Theorem (Karush–Kuhn–Tucker)
Consider x̄ ∈ X that satisfies all the constraints of the optimization
program. Suppose u ≥ 0 and w satisfy
Game theory was born in Princeton due to the pioneering work of two
mathematicians named John:
John von Neumann,
who (with economist Oskar Morgenstern) began the study of game
theory
and used fixed point theorems in a way that has since been
fundamental to game theory and economics in general;
and John Nash,
who formulated the notion of a Nash equilibrium.
The Nash equilibrium has had “a fundamental and pervasive impact in
economics and the social sciences which is comparable to that of the
discovery of the DNA double helix in the biological sciences.”
(Myerson 1999)
Game theory: Nash equilibrium
Suspect 2
Stay quiet Confess
Stay quiet −1, −1 −3, 0
Suspect 1
Confess 0, −3 −2, −2
When i faces the others’ actions a−i , her best response is:
If max a−i ≥ vi ,
then the bare-minimum payoff is the best she can do, so her best
response is to bid any number ≤ max a−i .
If max a−i < vi ,
then she gains positive payoff by bidding some number in the interval
(max a−i , vi ).
But if the highest bidder (with the tiebreaker of lowest-index) has
index less than i, then i has no best response.
(Indeed, decreasing choices of ai that converge to max a−i will
continually improve her payoff vi − max a−i ,
but at max a−i the payoff decreases to zero.)
On the other hand, if the highest bidder has index greater than i,
then i has a best response of bidding max a−i , since she would win.
Game theory: First-price sealed-bid auction, pt. 2
Exercise
Prove that in our second-price sealed-bid auction, each player i’s strategy
of bidding vi is a weakly dominant strategy, i.e., for any a−i , player i’s
payoff is at least as large as if she had submitted any other bid. Thus,
(v1 , . . . , vI ) is a Nash equilibrium.
Find all the other Nash equilibria: in particular, there are Nash equilibria in
which the auction winner is not player 1.
Game theory: Strategic games with no Nash equilibria
Player 2 (George)
Even (Two) Odd (One)
Even (Two) 0, 1 1, 0
Player 1 (Jerry)
Odd (One) 1, 0 0, 1
Consider the game odds and evens, in which two players bid either one
(odd) or two (even);
player 1 wins if the sum is odd, and player 2 wins if the sum is even.
There is no Nash equilibrium, since for any action profile a, the losing
player can deviate to win.
However, what if players can use randomized strategies?
Game theory: Mixed strategies
Player 2 (George)
Even (Two) Odd (One)
Even (Two) 0, 1 1, 0
Player 1 (Jerry)
Odd (One) 1, 0 0, 1
Theorem (Nash)
The strategic game ⟨I , (Ai ), (≾i )⟩ has a Nash equilibrium if: each Ai is a
nonempty compact convex Euclidean subset, and each ≾i is continuous
and quasiconcave on Ai .
Game theory: Nash’s theorem, pt. 2
xn → x ∈ X , yn → y ∈ X as n → ∞ =⇒ y ∈ f (x).
Corollary
Every finite strategic game ⟨I , (Ai ), (≾i )⟩ which allows mixed strategies has
a mixed-strategy Nash equilibrium.
Exercise
Prove the assertion that B has the closed graph property because ≾i is
continuous.
Aside: Coronavirus (COVID-19)
Figure: COVID-19 mortality rate by age (Gal 2020, sourced from Chinese Center
for Disease Control and Prevention; found in Bendix 2020, click for hyperlink)
Aside: How to protect yourself and others (COVID-19)
...Don’t smoke.
Eat a diet high in fruits, vegetables, and whole grains.
Take a multivitamin if you suspect that you may not be getting all the
nutrients you need through your diet.
Exercise regularly.
Maintain a healthy weight.
Control your stress level.
Control your blood pressure.
If you drink alcohol, drink only in moderation (no more than one to
two drinks a day for men, no more than one a day for women).
Get enough sleep.
Take steps to avoid infection, such as washing your hands frequently
and trying not to touch your hands to your face, since harmful germs
can enter through your eyes, nose, and mouth.
Aside: How to prepare for day-to-day life (COVID-19)
...For peace-of-mind, try to plan ahead for a possible outbreak.
For example, if there is an outbreak in your community, you may not
be able to get to a store, or stores may be out of supplies, so it will be
important for you to have extra supplies on hand.
Talk with family members and loved ones about how they would be
cared for if they got sick, or what would be needed to care for them in
your home.
Consider what you might do if your child’s school or daycare shuts
down, or if you need to or are asked to work from home.
Stay up-to-date with reliable news resources, such as the website of
your local health department. If your town or neighborhood has a
website or social media page, consider joining it to maintain access to
neighbors, information, and resources.”
(Harvard Health Publishing 2020; click for hyperlink)
Harvard University COVID-19 website (click for hyperlink)
Massachusetts COVID-19 website (click for hyperlink)
Game theory: Cooperation in the prisoners’ dilemma?
Zoom lectures will be recorded for the benefit of students who may
not be able to attend at the normal time, due to technological
difficulties or time-zone differences.
Please let me know in advance by email if you cannot attend one or
more of the remaining lectures.
“RECOMMENDATION: You should consider advising students that
Zoom provides an on-screen notification to meeting participants,
whenever a session is being recorded...
RULE: Instructors must not allow or enable students to record class
sessions, including by using Zoom...
RULE: Instructors may post links to class session recordings only in
the Zoom link of their Canvas course webpages...
RULE: Instructors must instruct students not to disclose any Zoom
recording URL — or any copies of the recording the student might
create or obtain — to anyone outside the class...
Aside: Zoom recording rules and recommendations, pt. 2
Suppose we have fixed a notion of the overall payoff, i.e., the utility
representation on sequences of payoffs.
e.g., Suppose preferences on payoff sequences are given by the
limit-of-means
i.e., agent i has overall payoff
T
X wt
(wit ) 7→ lim i
.
T →∞ T
t=1
(Note that this assumes the limit exists, which is not always true.)
Then, v is the overall payoff profile of (at ) precisely when each vi is
the limit-of-means of (ui (at )).
Game theory: When specifying the overall payoff is enough
Suppose that agents can only use memory-n strategies,
i.e., their actions in the T th stage game GT can only depend on the
n-history, the action profiles of GT −n , . . . , GT −1 .
Then, after fixing the agents’ memory-n strategies,
the set of I -tuples of agents’ n-histories can be thought of as a finite
state space, on which each passing stage game probabilistically acts as
a Markov chain.
Under certain conditions, this Markov chain has a unique invariant
distribution which is also its limiting distribution.
Thus, the limit of means
T
X wt i
lim .
T →∞ T
t=1
Proposition
Consider an infinitely repeated (T = ∞) strategic game ⟨I , (Ai ), (ui )⟩ in
which the players prioritize the limit of means. Every overall payoff profile
of a Nash equilibrium of the repeated game is an enforceable payoff profile
of G .
Theorem (Aumann–Shapley)
Every feasible strictly enforceable payoff profile w is the overall payoff
profile of a Nash equilibrium of the repeated game.
Game theory: Undiscounted ∞ly repeated games, pt. 3
for all i ∈ I .
Exercise
Prove that there exists a = (at ) such that the above is true. (Hint: Begin
by showing that any point in the convex hull of a finite set of points S can
be approximated by rational-coefficient linear sums of points in S.)
Game theory: Undiscounted ∞ly repeated games, pt. 4
t ) of
For each i ∈ I , there exists a minimaxing action profile (p−i
players −i.
Define player i’s strategy si = (sit ) of the repeated game by the one
in which she plays sit = ait until the first defection of a player j ̸= i
from a,
after which in all subsequent actions, player i plays to punish the first
t ) .
defector j, i.e., sit = (p−j i
The strategy profile s is a Nash equilibrium.
Indeed, if player i defects, then she receives at most her minimax
payoff vi ,
which is < wi by the definition of strict enforceability. Q.E.D.
Game theory: Finitely repeated games
Proposition
Suppose the payoff profile of each Nash equilibrium of G is precisely the
vector of minimax payoffs. Then, for any Nash equilibrium (at ) of GT ,
each at must be a Nash equilibrium of the constituent game G .
Game theory: Finitely repeated games, pt. 2
Theorem (Benoit–Krishna)
Suppose G has a Nash equilibrium â for which each player’s payoff is
strictly greater than her minimax payoff vi . Then, for any feasible strictly
enforceable action profile a∗ and ε > 0, there exists T ∗ such that if
T > T ∗ , a vector that only differs from ui (a∗ ) in each entry by at most ε
is the overall payoff profile of a Nash equilibrium of the repeated game GT .
Game theory: Finitely repeated games, pt. 4
Proof. Let L be arbitrary; we will set it later so that deviations are
unprofitable.
For each player i, choose a minimaxing action profile p−i of players −i.
Construct player i’s abstract machine Mi that determines her strategy:
The set of states of Mi is
In state Standardt , player i does the action ai∗ , and the state updates
to Standardt+1 if t < T − L and to Nash if t = T − L.
In the permanent state Nash, player i does the action âi .
At any point when a player j ̸= i plays an action ̸= a∗ during her
standard state Standardt , all other players’ states update to the
permanent state Punish(j),
and when acting in this state, they play their respective actions in the
minimaxing action profile p−j .
Game theory: Finitely repeated games, pt. 5
We choose L sufficiently large so that
∗
max ui (ai , a−i ) − ui (a∗ ) ≤ L(ui (â) − vi )
ai ∈Ai
for all i ∈ I ,
which guarantees that no player can profitably deviate in a standard
state.
Finally, choose T ∗ large enough so that
(T ∗ − L)ui (a∗ ) + Lui (â)
− ui (a∗ ) ≤ ε
T∗
i.e., so that the mean payoff of player i differs from ui (a∗ ) by at most
ε, for all players i. Q.E.D.
Exercise
Prove this result for the weaker hypothesis that G has Nash equilibria
aˆ1 , . . . , aˆI for which player i has payoff strictly greater than her minimax
payoff vi .
Game theory: Discounted ∞ly repeated games
With (limit of) mean preferences, a dichotomy:
Infinitely repeated games have full potential for cooperation,
since any profit from one-time deviation cannot matter in the grand
scheme of things.
Finitely repeated games have limited potential for cooperation,
because the above can matter.
An analogous dichotomy arises when we take into account
nonconstant temporal preferences,
i.e., agents who are impatient rather than patient.
Suppose now that preferences on payoff sequences are given by
exponential discounting, i.e., for a fixed δ ∈ (0, 1), the overall payoff is
∞
1 X t−1 t
(wit ) 7→ δ wi .
1−δ
t=1
Game theory: Discounted ∞ly repeated games, pt. 2
Nash folk theorem for infinitely repeated games with the overall payoff
given by exponential discounting:
Theorem (Fudenberg–Maskin)
Let w be a feasible strictly enforceable payoff profile of G = ⟨I , (Ai ), (ui )⟩.
For every ε > 0, there exists δ ∈ (0, 1) such that
Exercise
Prove the Fudenberg–Maskin theorem. Hint: A proof similar to that of
Aumann–Shapley works.
Game theory: Extensive-form games
An extensive game is comprised of the following:
a finite set I of players;
a history set H of sequences (finite or infinite) that satisfies three
properties,
the empty sequence ∅ is in H,
every subhistory (ak )k=1,...,L of (ak )k=1,...,K ∈ H is also in H, and
if every subhistory of an infinite sequence (ak )k=1,...,∞ is in H, then it
is also in H;
a player function P : H\ → I ∪ {c}, where c denotes “chance”;
for every h ∈ P −1 (c), a probability distribution fc (·|h) on A(h), the
action space of history h;
for each player i, a partition Ii of P −1 (i) into information sets, each
of which has shared action space A(h);
for each player i, a complete and transitive preference relation ≾i on
the set Z of terminal histories.
Game theory: Extensive-form games, pt. 2
A strategy of player i is a function that, to every nonterminal history
h ∈ p −1 (H \ Z ), assigns an action in A(h).
A Nash equilibrium is a strategy profile (si∗ )i∈I such that every player i
cannot deviate to another strategy to achieve a better outcome, i.e.,
∗ ∗
O(s−i , si ) ≾i O(s−i , si∗ ) for all strategies si of player i.
Exercise
Show that every strategic game (thought of as simultaneous actions by the
players) can be converted into an extensive game. Show that every
extensive game (thought of as: each player submits a strategy in the
beginning) can be converted into a strategic game.
Thus, our definitions and results for strategic games carry over to
extensive games (e.g., on Nash equilibria).
Game theory: The chain store’s dilemma
for every strategy si in the subgame Γ(h) that differs from si∗ |h only in the
entry of the action after the initial history of Γ(h). Hint: reverse induction.
Game theory: Punishing for a limited period of time
Theorem (Aumann-Shapley/Rubinstein)
Every feasible strictly enforceable payoff profile w is the overall payoff
profile of an SPE of the repeated game.
Game theory: Punishing for a limited period of time, pt. 2
Theorem (Fudenberg–Maskin)
For feasible strictly enforceable action profile a∗ , suppose there exists a
collection a(1), . . . , a(I ) of strictly enforceable action profiles such that for
every player i, we have a(i) ≺i a∗ and a(i) ≺i a(j) for all j ̸= i. Then,
there exists δ ∈ (0, 1) such that
Exercise
The above proof of the folk theorem implicitly requires that any deviation
from the minimax punishment (I − 1)-tuple would be detected, which
requires some restriction. Examples include the condition that the
comprising actions be pure strategies (whereas in general, they can be
mixed strategies; e.g., odds and evens) and the condition that each
punishing player’s random generator can be observed after her action.
Complete the proof of the folk theorem for the general setting.
Game theory: More folk theorems
i.e., the set of predicted outcomes equals the set of socially desired
outcomes.
One example of a social choice rule is f = Pareto,
which maps an I -tuple of preferences (≾i ) over C to the subset of C
comprised of outcomes that are not Pareto dominated by another
outcome, assuming the players have preferences (≾i ).
Mechanism design: S-implementation
Recall King Solomon’s environment ⟨{1, 2}, {a, b, d}, {(≾i ), (≾′i )}, G ⟩
f = TrueMother is defined by (≾i ) 7→ 1 and (≾′i ) 7→ 2.
His original strategic form: Announce d will happen.
Hope exactly one player will say “don’t cut the baby, give it to the
other player.”
Then, this player is the true mother, so do a or b depending on
whether she is player 1 or 2.
If both players say this, then...?
Despite his wisdom, strategic form is not completely defined; this trick
probably won’t work again when the dilemma reoccurs in the future.
Can King Solomon devise a strategic form which can DSE -implement
the social choice function f = TrueMother?
Can a strategic form Nash-implement f = TrueMother?
Mechanism design: DSE-implementation
It turns out that DSE-implementation is difficult.
A social choice rule f is dictatorial if and only if there is a player j ∈ I
such that for any (≾i ) ∈ P,
Theorem (Gibbard—Satterthwaite)
Suppose that in the environment ⟨I , C , P, G ⟩, the cardinality |C | is at least
three, P is the set of all strict preference profiles over C , and G is the set
of all strategic forms of the environment. If a social choice rule f is
DSE-implementable and satisfies
then f is dictatorial.
Mechanism design: DSE-implementation, pt. 2
Exercise
Show that if I is finite, every ultrafilter on I is principal.
Mechanism design: Proof of Impossibility Theorem
Lemma
A social choice function f : P → C is stable if and only if
Proof. We say that (x, y ) is optimal for a strict ordering ≾ if the top
choice is x and the second top choice is y .
We say that {x, y } is optimal for a strict ordering ≾ if the top two
choices are, in any order, x and y .
Mechanism design: Proof of Impossibility Theorem, pt. 4
φxy ′ ′ ′
j ((≾i )) = {(≾i ) :≾−j =≾−j , {x, y } is optimal for ≾j ,
and (x, y ) is optimal for ≾′j if j ∈ Vxy ((≾i ))},
Exercise
Prove the converse of this lemma.
Mechanism design: Proof of Impossibility Theorem, pt. 6
Due to the above lemma and the surjectivity of f , we obtain the weak
Pareto property: if every agent i prefers x to y , then y is not the
social choice.
Lemma
The preventing family Fxy is invariant under the choice of x ̸= y .
Proof. First, we show that Fxy ⊆ Fxz , thus showing that Fxy is
invariant under the choice of y .
Let S ∈ Fxy and T ∈ Fyz be arbitrary.
There exists a strict ordering (≾i ) such that x, y , and z are the top
three options of every agent, and
y ≾i z ≾i x for all i ∈ S \ T ,
x ≾i z ≾i y for all i ∈ T \ S,
z ≾i y ≾i x for all i ∈ S ∩ T , and,
x ≾i y ≾i z for all i ∈ I \ (S ∪ T ).
Mechanism design: Proof of Impossibility Theorem, pt. 7
Since every agent’s top three choices are x, y , z, the weak Pareto
property implies that f ((≾i )) is x, y , or z.
But we have constructed (≾i ) so that f ((≾i )) cannot be y or z.
Thus, it must be x.
Since Vxz ((≾i )) is equal to S,
it follows from the previous lemma that S ∈ Fxz .
(Because if f ((≾i )) = x, then for any other preference profile (≾′i )
that satisfies S ⊆ Vxz ((≾′i )), we have f ((≾′i )) ̸= z.)
A similar proof shows that Fxy ⊆ Fzy , thus showing that Fxy is
invariant under the choice of x. Q.E.D.
Exercise
Do the second case: write an analogous proof showing that Fxy ⊆ Fzy .
Mechanism design: Proof of Impossibility Theorem, pt. 8
We now show that F is an ultrafilter.
Condition (1) follows from the second last lemma and the weak Pareto
property.
Condition (2) follows from the weak Pareto property.
Condition (3): Suppose for the sake of a contradiction that
S, T , U ∈ F, but S ∩ T ∩ U = ∅.
We force a Cordocet cycle: there exists a strict ordering (≾i ) such
that x, y , and z are the top three choices for all agents,
x ≾i y if i ∈ S,
y ≾i z if i ∈ T , and
z ≾i x if i ∈ U.
By the weak Pareto property, f ((≾i )) must be one of x, y , and z.
But it cannot be any of them, by our construction. Contradiction.
Q.E.D. of G–S Impossibility Theorem, and thus, of the G–S Theorem
Mechanism design: Exercises
Exercise
What happens to the G–S Impossibility Theorem when |C | = 2?
Exercise
Use a similar argument to prove Arrow’s Impossibility Theorem: Let L
denote the set of strict orderings of C . If the cardinality |C | is at least
three, then it is impossible for a social aggregating function F : LI → L to
simultaneously satisfy the following axioms.
1 Unanimity: If (≾i ) satisfies that x ≾i y for all agents i, then xF (≾i )y .
2 Independence of irrelevant alternatives: For any two strict orderings
(≾i ) and (≾′i ) such that for all agents i, we have x ≾i y if and only if
x ≾′i y , it follows that xF ((≾i ))y if and only if xF ((≾′i ))y .
3 Non-dictatorship: No agent j enjoys the property that F ((≾i )) =≾j
for all (≾i ).
Mechanism design: When is DSE-implementation possible?
Theorem
The second-price, sealed-bid auction DSE-implements the above social
choice rule f .
Mechanism design: Generalized Vickrey auction
Before our proof, we generalize to the general Vickrey auction.
Let x̄ be a vector of auctioned objects.
Let vi (·) denote player i’s value function (true valuation of each
subvector of objects).
Each player self-reports a value function v̂i .
The auctioneer then chooses an allocation maximizing the summed
self-reported value functions,
Theorem (Vickrey)
Truthtelling is the unique weakly dominant bid for every player i.
= vi (xi∗ ) − pi∗ ,
.e., at most her utility from bidding her true valuation vi .
Exercise
Complete the proof: show that truthtelling is the only weakly dominant
strategy. This entails constructing a bid v̂−i such that a given non-truthful
bid v̂i gives strictly less utility than the truthful bid vi .
Mechanism design: Nash-implementation
Theorem (Maskin)
Let ⟨I , P, C , G ⟩ be an environment in which G is the set of all (relevant)
strategic game forms. If a social choice rule f : P ⊸ C is
Nash-implementable, then it is Maskin-monotonic.
Exercise
Construct an environment and a Maskin-monotonic social choice function
that shows that the converse of the above theorem is not true.
Theorem (Maskin)
Let ⟨I , P, C , G ⟩ be an environment in which G is the set of all (relevant)
strategic game forms. Suppose that a social choice rule f : P ⊸ C is
Maskin-monotonic, and furthermore, that it has the no-veto-power
property [c ∈ f ((≾i )) whenever at least I − 1 players maximally prefer c].
Then, f is Nash-implementable.
Figure 1, Sanchez and Dunning 2018 Figure 2, Sanchez and Dunning 2018
P(E | M)
P(M | E ) = P P(M)
m P(E | Mθ )P(Mθ )
Q
where P(E | M) = k P(ek | M).
Descriptive HDM: Bayesian DM hypothesis, pt. 2
Bayesian inference: Let e1 , . . . , en ∼ p(X | Mθ ).
Start with prior distribution π(Mθ ) and update it to
p(Mθ | e1 , . . . , en )π(Mθ )
,
m(e1 , . . . , en ) θ
Z
m(e1 , . . . , en ) = p(e1 , . . . , en | Mθ )π(Mθ )dθ.
The “relevant part” of the Bayesian CLT can fail in many ways,
but the failure we have in mind is the following:
“...we should be skeptical of a person’s ‘Bayesianness’ if we either
consistently observe that she repeatedly changes her mind a relatively
large amount without growing more confident, or, conversely,
consistently ends up very confident but with relatively little
fluctuation in beliefs.” (Augenblick–Rabin 2019; emphasis mine)
Suppose for the sake of a contradiction that one’s self-reported
confidence is a prior that is updated Bayesianly.
Persistent overconfidence, i.e., a large, positive difference between the
prior and the observations’ mean, is a contradiction.
Descriptive HDM: Non-monotonic confidence graphs
Figure 1, Sanchez and Dunning 2018 Figure 2, Sanchez and Dunning 2018
Exercise
If each fn is (Riemann) integrable, then the uniform limit f is integrable and
Z b Z b
lim fn (x)dx = f (x)dx
n→∞ a a
Exercise
If each fn is continuously differentiable such that fn′ uniformly converges to
g , then the uniform limit f is continuously differentiable such that f ′ = g .
Descriptive HDM: Guaranteeing uniform convergence
Theorem (Arzelà–Ascoli)
Let (fn ) be a sequence of continuous functions [a, b] → R that is uniformly
bounded
[There exists M such that |fn (x)| ≤ M for all x ∈ [a, b] and n]
[For every ε > 0, there exists δ > 0 such that for all x, y ∈ [a, b] and n,
|x − y | < δ =⇒ |fn (x) − fn (y )| < ε].
Exercise
Prove the Arzelà–Ascoli theorem.
Descriptive HDM: Guaranteeing uniform convergence, pt. 2
Theorem (Dini)
Let (fn ) be a sequence of continuous functions [a, b] → R that is monotone
Exercise
Prove Dini’s theorem. 6$PS7
Descriptive HDM: The learning model
The human decision-maker (DM) faces a task that requires an amount
a ∈ R>0 ∪ {∞} of knowledge to master.
(Information in the complement of payoff observations is not
straightforward to quantify—hence our use of more general model
parameters than previous papers on learning.)
She knows b ≤ a of this amount of knowledge.
The values of b and a, once fixed, deterministically fix the expected
marginal payoff fa (b) the DM can achieve.
By scaling, we suppose that fa : [0, a] → [0, 1], which we also assume
is continuous.
When fixing b, the payoff fa (b) is decreasing in a;
when fixing a, it is increasing in b.
We suppose that fa (a) = 1, the maximum payoff.
All payoffs are subject to time-discounting by an exponential factor
δ ∈ (0, 1).
Descriptive HDM: The learning model, pt. 2
The amount of knowledge that the DM knows, b, increases over time
in discrete jumps, each following the acquisition of a payoff in the
form of a high-uncertainty lottery, until it reaches a.
b increases as a discrete learning function
LS : R≥0 → R≥0 ,
Exercise
Prove a version of the monotone convergence theorem under the additional
assumption that fn are Riemann integrable. (Hint: Start from Dini’s
theorem).
Descriptive HDM: Convergence theorems, pt. 3
We say that the sequence of discrete learning models {(LSn , LS′n )}n>0
converges to a continuous learning model (L1 , L2 ) if the following
conditions hold.
The sequence of functions {LSn }n>0 monotonically converges to L1 in
a way such that L1 (tn,i ) = bn,i for all n and i.
The difficulty values of tasks of the first type are given by the same
distribution µ1 for L1 and all LSn .
The sequence of functions {LS′n }n>0 monotonically converges to L2 in
′ )=b
a way such that L2 (tn,i n,i for all n and i.
The difficulty values of tasks of the second type are given by the same
distribution µ2 for L2 and all LS′n .
The same payoff functions fa (·) are shared by (L1 , L2 ) and all
(LSn , LS′n ).
Descriptive HDM: Differences of the two task types
First, tasks of the first type have positive probability p on the event
that a = ∞,
i,e., that the task—due to its impossibility—will never learn to
completion and always give a marginal payoff of f∞ (b) = 0;
tasks of the second type do not.
Second, imitation learning is faster than innovation learning.
In the continuous learning model (L1 , L2 ), the latter difference
corresponds to the following condition.
Assumption C1 We have L1 (t) ≤ L2 (t) for all t ∈ R≥0 .
When a sequence of discrete learning models {(LSn , LS′n )}n>0
converges to the continuous learning model (L1 , L2 ),
the above condition on (L1 , L2 ) follows from an analogous one applied
to (LSn , LS′n ).
Assumption D1 We have ∆n,i ≥ ∆′n,i for all n and i.
Descriptive HDM: Information set
Recall our assumption that in the discrete learning model (LSn , LS′n ),
the knowledge jumps from bn,i−1 to bn,i that occur during innovation
learning are indistinguishable form those that occur during imitation
learning.
Thus, the only way to tell the two types of tasks apart are by observing
whether the amount of time spent in the ith learning period is ∆n,i
or the possibly smaller value ∆′n,i .
Furthermore, this can only be done when these two values, ∆n,i and
∆′n,i , are distinct.
Descriptive HDM: Information set, pt. 2
Second, define the quitting value function V (b) by the expected total
profit from quitting at level of knowledge b ∈ R≥β ∪ {∞} that has
not yet caught up to a.
i.e. the solution to
L−1
!
Z b Z
2 (a)
Z ∞
t t
V = δ fa (L2 (t)) + δ dt dµ(a)
0 0 L−1
2 (a)
L−1
!
Z ∞ Z
2 (b)
t L−1
2 (b)
+ δ fa (L2 (t))dt + δ V dµ(a).
b 0
Theorem
Suppose q = 1 and µ = µ2 is an exponential distribution of decay factor
η ∈ (0, 1). Under Assumption C3, the following are true.
1 The quitting value function V (b) is strictly increasing. In particular,
b ∗ = ∞ is the unique quitting point maximizing V (·).
2 The expected payoff function g (b) is strictly increasing.
Exercise
Show that every positive continuous random variable X with the
memoryless property is an exponential distribution.
Descriptive HDM: With tasks of infinite difficulty
L−1
!
1 (b)
Z Z
L−1
+ δ t fa (L1 (t))dt + δ 1 (b) V dµ(a).
a>b 0
as b → ∞.
Then, we have the following non-monotonicity:
Descriptive HDM: Non-monotonicity
Theorem
Suppose q = 0 and µ = µ1 is the distribution with probability p > 0 on the
event a = ∞ and the remaining probability distributed exponentially on
R>0 with decay factor η ∈ (0, 1).
1 Under Assumption C4, the derivative of the quitting value function,
V ′ (b), is negative for all sufficiently large b. In particular, the one or
more quitting points b ∗ that maximize V (·) are finite.
2 The expected payoff function g (b) converges to zero as b → ∞. In
particular, g (·) attains its maximum at one or more finite points b.
3 Under Assumption C5, g is C 1 such that g ′ (b) is negative for all
sufficiently large b.
Descriptive HDM: Non-monotonicity, pt. 2
To prove part (1), we show that the derivative
1
V ′ (b) = 2
−1
L1 (b) dµ(a)
R
1− a>b δ
−1
!
∞
δ L1 (b) fa (b)
Z Z Z
L−1
1 (b) t
· 1− δ dµ(a) µ(b) δ dt + dµ(a)
a>b L−1
1 (b) a>b L′1 (L−1
1 (b))
−1
!
δ L1 (b) log 1δ
Z
−1
− µ(b)δ L1 (b)
+ ′ −1
dµ(a)
a>b L1 (L1 (b))
L−1
!
Z b Z
1 (a)
Z ∞
· δ t fa (L1 (t)) + δ t dt dµ(a)
0 0 L−1
1 (a)
L−1
! !!
1 (b)
Z Z
+ δ t fa (L1 (t))dt dµ(a)
a>b 0
Corollary
In the setting of the previous theorem, fix all choices except that of p. For
every γ ≥ β ≥ 0, there exist a choice of p such that any optimal quitting
point b ∗ for which the event a = ∞ has probability p satisfies
b ∗ ≥ γ.
Descriptive HDM: DM does not quit soon, pt. 1
We now consider the mixed case (0 < q < 1) with both innovation
and imitation learning,
which we offer as an approximation of a realistic learning model in the
setting of unobserved payoffs.
We assume that µ = (1 − q)µ1 + qµ2 for q ∈ (0, 1), where µ1 and µ2
are of the form defined previously.
(We have abused notation here, because our probability distributions
µj were defined on the domains of difficulty values a: not on the
domains of tasks (j, a), which also include the data of the type j.)
We can repeat our previous analysis to show that the DM always
prefers being at a positive level of knowledge b (that has not yet
caught up to a) for a task of the second type,
to being at zero knowledge for a task of the second type,
which is—by Assumption C1—always preferred to being at zero
knowledge for a task of the first type.
Descriptive HDM: Mixed case, pt. 2
Define the conditional quitting value functions V1 (b) and V2 (b) by the
expected total payoff from quitting at level of knowledge b that has
not yet caught up to a,
conditional on the task being of type j = 1 or j = 2, respectively.
Also, define the unconditional quitting value function
Vu (b) = (1 − q)V1 (b) + qV2 (∞) by the expected total payoff from
employing quitting strategy b:
that of quitting tasks of the first type at level of knowledge b that has
not yet caught up to a.
Descriptive HDM: Mixed case, pt. 4
Define the conditional expected payoff functions g1 (b) and g2 (b) by
the expected marginal payoff at payoff of knowledge b that has not yet
caught up to a, conditional on the task being of type j = 1 or j = 2,
respectively.
Also, define the unconditional expected payoff function
gu (b) = (1 − q)g1 (b) + qg2 (b) by the expected marginal payoff at
level of knowledge b that has not yet caught up to a, unconditional on
the task type.
Distinguish the latter from the true expected payoff function gb∗ (b)
associated to quitting strategy b ∗ , defined by the piecewise function
given by gu (b) if b < b ∗ and g2 (b) if b ≥ b ∗ .
Theorem
Suppose 0 < q < 1 and µ = (1 − q)µ1 + qµ2 , where µ1 is the distribution
with probability p > 0 on the event a = ∞ and the remaining probability
distributed exponentially on R>0 with decay factor η ∈ (0, 1), and µ2 is the
exponential distribution of decay factor η...
Descriptive HDM: Desired non-monotonicity
Theorem
...Under Assumptions C1 and C2, the following are true.
1 The conditional quitting value function V2 (b) is strictly increasing
under Assumption C3, while under Assumption C4, Vu (b) has
negative derivative for all sufficiently large b. In particular, the one or
more quitting strategies b ∗ maximizing
max Vu (b),
b∈R≥β ∪{∞}
Exercise
Prove this theorem. (Note: The proof can be found in Park 2020, and you
can use it without citing it. But please write the solution in your own
words.)
Corollary
Retain the setting of the previous theorem and fix all choices except those
of p and q. For every γ ≥ β > 0, there exist choices of p and q such that
any optimal quitting point b ∗ for which tasks of the first type have
difficulty a = ∞ with probability p and tasks of the second type occur with
probability q satisfies
b ∗ ≥ γ.
Exercise
Prove this corollary. (Note: The proof can be found in Park 2020, and you
can use it without citing it. But please write the solution in your own
words.)
Descriptive HDM: Summary of CLM, pt. 2
max Vu (b)
b∈R≥β ∪{∞}
Theorem (continued)
1 When faced with a side opportunity of marginal value P, the DM
takes it if P is greater than
1
if bn,i has caught up to a,
gu (bn,i ) if bn,i has not caught up to a, and j is unknown,
g2 (bn,i ) if bn,i has not caught up to a, and j is known to be j = 2;
L−1
!
1 (b)
Z Z
t L−1
1 (b)
+ δ fa (LSn (t))dt + δ ((1 − q)V + qV2,n (∞))
a>b 0
Descriptive HDM: Proof, pt. 4
In the social sciences, there are a lot of questions that we do not know
how to satisfactorily answer yet.
An important bar to clear is the Lucas critique:
“Given that the structure of an econometric model consists of optimal
decision rules of economic agents, and that optimal decision rules vary
systematically with changes in the structure of series relevant to the
decision maker, it follows that any change in policy will systematically
alter the structure of econometric models.” (Lucas 1976)
Generally, in order to extrapolate from a model, it should not be just
based on statistical analysis;
it should necessarily take into account the “true” parameters that
determine the behavior of the model’s constituents (humans, particles,
etc.).
Concluding remarks, pt. 2
“It turns out that saying ‘high inflation has always been correlated with
low unemployment, so we can tackle unemployment by accepting
higher inflation’ is a bit like saying ‘Fort Knox has never been robbed,
so we can save money by sacking the guards.’ You can’t look just at
the empirical data, you need also to think about incentives. High
inflation has tended to mean lower unemployment because employers
and job seekers expected a particular rate of inflation and were
occasionally surprised by a surge in prices; employers mistook it for
increased demand and tried to hire workers, and workers thought that
they were being offered higher real wages. Both were mistaken: in
fact, what was really happening was that the economy was suffering
unexpected inflation and they’d been slow to notice. The problem was
that people wouldn’t keep on being surprised by inflation if
policymakers, beguiled by the Phillips curve, kept deliberately creating
inflation with the aim of suppressing unemployment. Nobody would be
fooled; they would see the inflation coming from a mile off. Inflation
would rise but unemployment would not fall.” (Harford 2013)
Concluding remarks, pt. 3
Accept uncertainty: There is a lot we don’t know yet, and that’s OK.
Treat overconfident claims with skepticism: The world is
uncertain, but the human mind is hard-wired to avoid uncertainty
(Berker et al. 2016).
As a result, there will be many claims made about subjects like the
economy,
and in the absence of a Lucas-robust, experiment-tested model (e.g.,
general relativity, germ theory),
these claims will often be overconfident.
When data has high variance, the human mind may not
meaningfully retain them (Park 2020): Record them while being
careful of statistical bias.
Goal: Find the “true” parameters of human decision-making
whose Lucas-robust model makes consistently accurate
predictions.