JO Ao Lopes Dias
JO Ao Lopes Dias
JO Ao Lopes Dias
Contents
1. Introduction 2
1.1. Preliminaries 2
1.2. Optimal points and values 2
1.3. The optimization problems 3
1.4. Existence of optimal points 4
2. Unconstrained local optimization 5
2.1. Critical points 5
2.2. Classification of critical points 6
3. Equality constraints 7
3.1. Lagrange theorem 7
3.2. Classification of critical points 8
4. Inequality constraints 11
4.1. Kuhn-Tucker conditions 11
4.2. Mixed constrains 13
5. Convex and concave optimizations 14
5.1. Optimization 15
References 17
Index 18
1. Introduction
Proof. This follows from the fact that the image of a compact set by a
scalar continuous function is a compact set in R. Compact sets in R
have always a maximum and a minimum. We can therefore apply this
to f (D) whose extreme points are the maximum and minimum of f on
D.
Remark 1.7. The Weierstrass theorem only states the existence of op-
timal points, it does not compute them. For that we need to introduce
other tools like differential calculus as is done in the next sections. This
restricts our study to differentiable functions. In the last sections we
will be dealing with non-differentiable functions which are convex or
concave.
Example 1.9. Let D =]0, 1[ (not closed) and f (x) = x. Then, f (D) =
]0, 1[ and f does not have optimal values on D.
Example 1.10. Let D = [0, 1] and
(
x, x ∈]0, 1[
f (x) = 1
2
, x ∈ {0, 1}
(not continuous). Then, f (D) =]0, 1[ and f does not have optimal
values on D.
(1) The converse of the above theorem is not true. There are exam-
ples of critical points which are not local optimal points. For
the function f (x) = x3 the point x∗ = 0 is critical but not
optimal.
(2) The above theorem is restricted to interior points of p D where
the function is differentiable. For example, f (x, y) = x2 + y 2
on D = {(x, y) ∈ Rs : x2 + y 2 ≤ 1} has a minimizer at the origin
(which is in the interior of D), but f is not differentiable there.
3. Equality constraints
Proof. First, reorder the coordinates such that the first m rows of
Dg(x∗ ) are linearly independent. Write the coordinates x = (w, z) ∈
Rm × Rn−m and suppose that x∗ = (w∗ , z ∗ ) ∈ D is a local maximizer
(we case use the same ideas for a minimizer).
8 JOÃO LOPES DIAS
∂g ∂g
Since rank ∂w (x∗ ) = m we have that det ∂w (x∗ ) 6= 0. By the implicit
function theorem, there is a C 1 function h : V → Rm defined on a
neighborhood V of z ∗ such that
• h(z ∗ ) = w∗ ,
• g(h(z), z) = 0, z ∈ V ,
∂g
• Dh(z ∗ ) = − ∂w (x∗ )−1 ∂g
∂z
(x∗ ).
∂f ∂g
Choose λ∗ T = − ∂w (x∗ ) ∂w (x∗ )−1 , so that
∂f ∗ ∂g ∗
(x ) + λ∗ T (x ) = 0.
∂w ∂w
Finally, let F : V → R, F (z) = f (h(z), z), which has a local max-
imum at z ∗ ∈ V . This is an unconstrained problem since V is open,
thus DF (z ∗ ) = 0. That is,
Dh(z ∗ )
∗ ∗
Df (h(z ), z ) = 0.
I
This yields, after simplification,
∂f ∗ ∂g
(x ) + λ∗ T (x∗ ) = 0.
∂z ∂z
Exercise 3.2. Let f (x, y) = x3 + y 3 and g(x, y) = x − y. Solve with
respect to (x, y) and λ the equations f (x, y) + λT g(x, y) = 0 and
g(x, y) = 0.
There is a criterion for the positive and negative definite cases. For
each i = m + 1, . . . , n define the (m + i) × (m + i) symmetric matrix
0 Bi
Ci = ,
BiT Ai
where Ai is the matrix consisting of the first i rows and i columns of
A, and Bi is made of the first i columns of B.
(1) A > 0 on Z iff (−1)m det Ci > 0 for every i = m + 1, . . . , n.
(2) A < 0 on Z iff (−1)i det Ci > 0 for every i = m + 1, . . . , n.
Example 3.3. Let
2 1
A= and B = 1 2 .
1 1
In this case m = 1 and n = 2. Thus,
0 1 2
C2 = 1 2 1
2 1 1
and (−1)1 det C2 = 5 > 0. Therefore, A > 0 on Z.
and y ∈ Z(x∗ ).
Consider the C 2 function L(x) = f (x) + λ∗ T g(x) on D by fixing
λ = λ∗ . Notice that DL(x∗ ) = 0 and Taylor’s formula around x∗ yields
1
L(xkn ) = L(x∗ ) + (xkn − x∗ )T D2 L(cn )(xkn − x∗ )
2
for some cn as before. Hence,
1 T 2 f (xkn ) − f (x∗ )
ykn D L(cn )ykn = ≤ 0.
2 kxkn − x∗ k2
For n → +∞ we get y T D2 L(x∗ )y ≤ 0.
Exercise 3.5. Determine the local optimal points of f on D for:
(1) f (x, y) = xy,
D = {(x, y) ∈ R2 : x2 + y 2 = 2a2 }
(2) f (x, y) = 1/x + 1/y,
D = {(x, y) ∈ R2 : (1/x)2 + (1/y)2 = (1/a)2 }
(3) f (x, y, z) = x + y + z,
D = {(x, y, z) ∈ R3 : (1/x) + (1/y) + (1/z) = 1}
(4) f (x, y, z) = x2 + 2y − z 2 ,
D = {(x, y, z) ∈ R3 : 2x − y = 0, x + z = 6}
(5) f (x, y) = x + y,
D = {(x, y) ∈ R2 : xy = 16}
(6) f (x, y, z) = xyz,
D = {(x, y, z) ∈ R3 : x + y + z = 5, xy + xz + yz = 8}
MATHEMATICAL ECONOMICS: OPTIMIZATION 11
4. Inequality constraints
D = {x ∈ U : h(x) ≥ 0}.
Moreover,
4.2. Mixed constrains. Using the previous results we can now mix
equality with inequality constraints in order to have a more general
optimization theorem.
Let U ⊂ Rn be open, g : U → Rk and h : U → R` be C 1 functions,
and
D = {x ∈ U : g(x) = 0, h(x) ≥ 0}.
Theorem 4.4. If x∗ is a local optimal point of f on D,
h1 (x∗ ) = · · · = hm (x∗ ) = 0, hm+1 (x∗ ), . . . , h` (x∗ ) > 0
and rank D(g1 , . . . , gk , h1 , . . . , hm )(x∗ ) = k + m, then there is λ∗ ∈ Rk+`
such that for j = k + 1, . . . , k + `,
(1) Df (x∗ ) + λ∗ T D(g, h) = 0,
(2) λ∗k+i hi (x∗ ) = 0, i = 1, . . . , `.
Moreover,
• if x∗ is a local minimizer, then (λ∗k+1 , . . . , λ∗k+` ) ≤ 0.
• if x∗ is a local maximizer, then (λ∗k+1 , . . . , λ∗k+` ) ≥ 0.
A set D ⊂ Rn is convex if
λx + (1 − λ)y ∈ D
for any λ ∈ [0, 1] and x, y ∈ D. That is, the line joining any two points
of D is still contained in D.
Given a convex set D ⊂ Rn , f : D → R is a convex function on D if
f (λx + (1 − λ)y) ≤ λf (x) + (1 − λ)f (y)
for any λ ∈ [0, 1] and x, y ∈ D. If the above inequality is strict, we say
that f is strictly convex on D.
On the other hand, f is a concave function on D if −f is convex In
addition, f is strictly concave on D if −f is strictly convex.
Notice that functions could be both convex and concave or neither
convex nor concave. Moreover, a function both convex and concave
can not be neither strictly convex nor strictly concave.
Example 5.1. Let f (x) = x3 on R and x = 1, y = −1. For λ = 1/4 we
have
f (λx + (1 − λ)y) < λf (x) + (1 − λ)f (y)
MATHEMATICAL ECONOMICS: OPTIMIZATION 15
Proof.
Theorem 5.4. Let f : D → R be C 2 on D ⊂ Rn open and convex.
(1) f is concave on D iff D2 f (x) ≤ 0 for any x ∈ D.
(2) f is convex on D iff D2 f (x) ≥ 0 for any x ∈ D.
(3) If D2 f (x) < 0 for any x ∈ D, then f is strictly concave.
(4) If D2 f (x) > 0 for any x ∈ D, then f is strictly convex.
Proof.
Example 5.5. Consider the function f (x) = log(xα1 . . . xαn ) for xi > 0
and α > 0. Hence,
(
∂ 2f − xα2 , i = j
(x) = i
∂xi ∂xj 0, i 6= 0.
So, D2 f (x) < 0 and f is strictly concave.
5.1. Optimization.
Theorem 5.6. Let D ⊂ Rn be convex and f : D → R convex (concave).
Then,
(1) any local minimizer (maximizer) of f on D is in fact global.
(2) the set of minimizers (maximizers) of f on D is either empty
or convex.
(3) if f is strictly convex (concave), then the set of minimizers
(maximizers) of f on D is either empty or it contains a single
point.
Proof.
16 JOÃO LOPES DIAS
Proof.
References
[1] R. K. Sundaram. A First Course in Optimization Theory. Cambridge Univer-
sity Press, 1996.
Index
concave nonlinear, 3
function, 14
strictly, 14 set
convex constraint, 3
function, 14 Theorem
strictly, 14 Weierstrass, 4
set, 14
function
concave, 3
convex, 3
objective, 3
functions
affine, 3
image
of a set, 2
maximizer, 2
local, 2
strict, 2
maximum, 2
local, 2
minimizer
local, 2
strict, 2
minimum, 2
local, 2
optimal
points, 3
local, 3
values, 3
local, 3
optimization
concave, 3
convex, 3
integer, 3
linear, 3
local
problem, 3
problem, 3
point
critical, 5
saddle, 6
stationary, 5
pre-image
of a set, 2
programming
integer, 3
linear, 3
18
MATHEMATICAL ECONOMICS: OPTIMIZATION 19