Optimization Theory
Optimization Theory
CASED
1
Chapter 1: Mathematical Background
A1 × A2 × . . . × Ap = {(a1 , a2 , . . . , ap ) : ai ∈ Ai , for i = 1, 2, . . . , p}
Qp
Very often people write i=1 Ai instead of A1 × A2 × . . . × Ap .
x ≥ y if xi ≥ yi , ∀i
x > y if xi ≥ yi , ∀i, and xi > yi for at least one i
x >> y if xi > yi , ∀i
kx + yk ≤ kxk + kyk
and
kx − zk ≤ kx − yk + ky − zk
|x · y| ≤ kxkkyk
2
• A sequence of real numbers {xk }k converges to zero (notation xk → 0) if:
for any ε > 0 there exists an integer N such that if k ≥ N then |xk | ≤ ε,
i.e. −ε ≤ xk ≤ ε.
Proof : Assume {xk }k in Rn has two limits a, b. Then, given ε > 0, one
can find Na , Nb such that, for any n ≥ Na , one has kxn − ak ≤ ε and any
n ≥ Nb , one has kxn − bk ≤ ε. In this case, if we take N ≥ max{Na , Nb },
we have
ka − bk = ka − xN + xN − bk ≤ ka − xN k + kxN − bk ≤ 2ε
Let ε go to 0. We obtain a = b.
3
Proposition 1 Let A be a nonempty set. The sup(A) is the unique ele-
ment (eventually +∞) such that :
(i) If x > sup(A) then x ∈
/A
(ii) if x < sup(A) then there exists a ∈ A such that x < a
(iii) There exists a sequence in A which converges to sup(A).
4
to +∞. Now, suppose sup(A) < +∞. Then for any ε > 0, there exists
ak ∈ A which satisfies sup(A) − ε < ak ≤ sup(A). But we also have
∀n > k, sup(A) − ε < ak ≤ an ≤ sup(A). That means the sequence
A converges to sup(A). The proof is similar when the sequence is non
increasing.
f −1 (U ) = {x ∈ S : f (x) = y, y ∈ U }
• Open and closed sets For x ∈ Rn , the open ball B(x, r) with centre x
and radius r is the set B(x, r) = {y ∈ Rn : ky − xk < r}.
5
closed. If not, S c is NOT open. Then there there exists y ∈ S c s.t.
for any k ∈ N, there exists a point xk ∈ S ∩ B(y, k1 ). The sequence
of elements in S, {xk }k , converges to y when k converges to infinity.
But y is not in S: a contradiction. Hence, S is closed.
S1 + S2 = {x ∈ Rn : x = x1 + x2 , x1 ∈ S1 , x2 ∈ S2 }
– The empty set is open. The whole space Rn is then closed. But this
one is also open. Hence the empty set is also closed.
– The union of an arbitrary collection of open sets is again open
– The intersection of an arbitrary collection of open sets is NOT AL-
WAYS open
– The intersection of a FINITE collection of open sets is open
– The sum of two open sets is open.
Proof : Let S1 , S2 be two open sets. If one of them is empty then
S1 + S2 is also empty, hence open. So, we assume that both of
them are non-empty. Let x1 ∈ S1 , x2 ∈ S2 . there exists ε > 0
s.t. B(x1 , ε) ⊂ S1 , B(x2 , ε) ⊂ S2 . We claim that the open ball
B(x1 + x2 , 2ε ) is in S1 + S2 . Indeed, let y1 satisfy ky1 − x1 k < 2ε .
This implies y1 ∈ B(x1 , ε) ⊂ S1 . Now, let y ∈ B(x1 + x2 , 2ε ) and
y2 = y − y1 . We have
ky2 − x2 k = ky − y1 − x1 + x1 − x2 k
ε ε
≤ ky − (x1 + x2 )k + ky1 − x1 k < + = ε.
2 2
Thus, y2 ∈ B(x2 , ε) ⊂ S2 . We then have: y = y1 + y2 with y1 ∈
S1 , y2 ∈ S2 . This ends the proof.
6
– A set S ∈ Rn is bounded if there exists M such that S ⊆ B(0, M ). In
other words, S is bounded if there exists M > 0 such that kxk ≤ M
for all x ∈ S.
– A set S is compact if it is bounded and closed.
Theorem 4 A set S is compact if, and only if, for all sequences
{xk } in S, there exists a subsequence {xkm }m which converges to a
point x ∈ S.
Proof : Assume S has the property that for all sequences {xk } in S,
there exists a subsequence {xkm }m which converges to a point x ∈ S.
Let us prove it is compact. First, S is bounded. If not there exists
a sequence {xk }k ⊂ S with limk→+∞ kxk k = +∞. There exists a
subsequence {xkn }n of {xk }k which converges to some x ∈ S, which
is impossible since limn kxkn k = +∞.
We now prove S is closed. Suppose {xk }k ⊂ S converges to x.
There exists a subsequence which converges in S. This limit must
be x. Hence, x ∈ S.
Assume that S is compact. We will assume S ⊂ R2 . One can
easily see that the proof can be carried on when the dimension of
the space is larger than 2. Let {xk } be a sequence of S. Write
xk = (xk1 , xk2 ), ∀k. Since S is bounded, there exists a subsequence
{xk1n }n which converges to some x1 ∈ R. But there exists also a
kn
subsequence {x2 l }l of {xk2n }n which converges to some x2 ∈ R.
kn kn
Since S is closed, (x1 , x2 ) ∈ S since it is the limit of {(x1 l , x2 l )}l .
To summarize, we have found a subsequence of {xk } which converges
to a point in S.
7
– Suppose f : R → Rn . Then we say that f (x) → l ∈ Rn as x → +∞,
if for every ε > 0 there exists M ∈ R such that kf (x) − lk < ε for
all x > M . We use the notation limx→+∞ f (x) = l. We use similar
definitions for x → −∞.
• Continuous mappings
8
is uniformly on S if for every ε > 0, there exists δ > 0 such that for
any x ∈ S, for any y in S and ky − xk < δ then kf (y) − f (x)k < ε.
Proposition 4 Let f : S → Rm , where S ⊆ Rn . If S is compact
and f is continuous on S, then f is uniformly continuous on S.
9
Corollary 1 Let A ⊆ R be a nonempty compact set. Then A has a
maximum and a minimum.
• Differentiability of mappings
• Differentiability of functions
10
Theorem 6 Let f : S → R, where S ⊆ Rn is an open set. The
function f is C 1 on S if, and only if, all partial derivatives of f exist
and are continuous on S. In that case we also have
∂f (x) ∂f (x) ∂f (x)
Df (x) = , ,...,
∂x1 ∂x2 ∂xn
• Second derivatives
n
f : S → R, where
– Let S ⊆ R . Then the derivative Df (x) =
∂f (x) ∂f (x) ∂f (x)
∂x1 , ∂x2 , . . . , ∂xn is also a mapping from S to Rn . If Df is
differentiable then f is called twice differentiable with second deriva-
tive D2 f (x). The partial derivatives of the partial derivatives of f
∂ 2 f (x) ∂ 2 f (x)
are denoted by ∂x i ∂xj
if i 6
= j, and by ∂x2i
if i = j. In these cases
we have 2
∂ f (x) ∂ 2 f (x)
2 . . . ∂x1 ∂xn
∂x1
. . . . .
D2 f (x) = .....
2 ..... 2
∂ f (x) ∂ f (x)
∂xn ∂x1 . . . ∂x2
n
• Taylor expansion
11
– positive definite if x0 Ax > 0 for all x ∈ Rn , x 6= 0
– positive semidefinite if x0 Ax ≥ 0 for all x ∈ Rn
– negative definite if x0 Ax < 0 for all x ∈ Rn , x 6= 0
– negative semidefinite if x0 Ax ≤ 0 for all x ∈ Rn
– We denote by Ak the k × k submatrix of A formed by taking just
the first k rows and columns of A.
Similarly, one can define a global minimum , local minimum , strict local
minimum , unconstrained local minimum
12
For any real number t, define xt = x∗ + ta. For t close enough to zero, we
have that xt ∈ B(x∗ , r) ∩ S. Then
13
x−x0
Assume D2 f (x0 ) positive definite. When x 6= x0 , let u = kx−x 0k
∈ S(0, 1).
0 2
We know that minx∈S(0,1) x D f (x0 )x = α > 0. Therefore,
1
f (x) = f (x0 ) + (x − x0 )0 D2 f (x0 )(x − x0 ) + R2 (x, x0 )kx − x0 k2
2
1
= f (x0 ) + kx − x0 k2 u0 D2 f (x0 )u + R2 (x, x0 )kx − x0 k2 , when x 6= x0
2
1
≥ f (x0 ) + kx − x0 k2 [α + 2R2 (x, x0 )] , when x 6= x0
2
When x is close to x0 but different from x0 , we have α+2R2 (x, x0 ) > 0 and
thus f (x) ≥ f (x0 ). We have that x0 is an unconstrained local minimum.
Similarly, when D2 f (x0 ) is negative definite then x0 is an unconstrained
local maximum.
1 Exercises
Exercise 1 Show that the ball B(x, r) = {y ∈ Rn : ky − xk ≤ r} is a closed
set, where r > 0.
Exercise 2 Let U ∈ Rn . Then, U is open if, and only if, for any x ∈ U there
exist r > 0 such that B(x, r) ⊆ S.
Exercise 3 For the sets A given below, show that EVERY function f : A → R
is continuous
(a) A ∈ Rn with just an element
(b) A ∈ Rn with a finite number of elements
(c) A = Z, considered as a subset of R.
Exercise 4 Determine which of the sets in (a)-(c) from above are compact.
14
Chapter 2: Convex Functions
This chapter is devoted to a class of functions from Rn into R called convex
functions and to give a first important property of such functions. Any convex
function is continuous.
15
The function f being convex, we have:
λ1 λp−1
f (λ1 x1 + . . . + λp xp ) ≤ sp−1 f x1 + . . . + xp−1 + (1 − sp−1 )f (xp ).
sp−1 sp−1
Thus:
f (λ1 x1 + . . . + λp xp ) ≤ λ1 f (x1 ) + . . . + λp f (xp ).
The converse is true by taking p = 2.
Proposition 7 (i) Let f , g be two convex functions from a convex set U, into
R. Then max(f, g) is convex. More generally, consider a collection of convex
functions {fi }i=1,...,I , from U into R. Then max{fi | i = 1, . . . , I} is convex
from U into R.
(ii) If (fi )i∈N is a sequence of convex functions from Rn into R which converges
pointwise, i.e. ∀x ∈ Rn , the sequence (fi (x))i∈N converges in R, then the func-
tion defined for all x ∈ Rn by f (x) = limi→+∞ fi (x) is convex from Rn into R.
(iii) If f : U → R, g : V → R, with f (U ) ⊂ V , are convex, if g is nondecreasing,
then g ◦ f is convex.
(iv) If f, g are convex from a convex set U into R and if λ is a nonnegative real
number, then f + g and (λf ) are convex.
Proof : Exercise
3 Continuity theorems
In this section, we want to prove that for any convex function f from an open
set U ⊂ Rn into R, it is continuous. Let T = {a1 , a2 , . . . , am } ⊂ Rn . The
convex hull of T denoted by coT is the set
m m
( )
X X
coT = x : x = λi ai , λi ≥ 0, ∀i, λi = 1
i=1 i=1
Proof : Since 0 ∈ U, one can choose α > 0 sufficiently small such that the
convex hull V = co{αe1 , . . . , αen , −αe1 , . . . , −αen }, where the ei are the vectors
of the canonical basis of Rn , is contained in U . V has a nonempty interior, 0
16
is in the interior of V . Thus, V contains a closed ball B̄(0, r). Any x in V
may be expressed as: x = α ni=1 λi ei − α ni=1 λ0i ei with λi ≥ 0, λ0i ≥ 0, ∀i,
P P
Pn 0
i=1 (λi + λi ) = 1. Since f is convex,
n
X
λi f (αei ) + λ0i f (−αei ) ≤ max f (αe1 ), . . . , f (αen ), f (−αe1 ), . . . , f (−αen ) .
f (x) ≤
i=1
Remark Actually the set V in the proof above is the ball B̄1 (0, α) of the norm
k.k1 . Indeed, if x ∈ V then
!
X X X X
0
x=α λi ei − λi ei = α (λi − λ0i )ei = xi ei
i i i i
X X X
|xi | = α |λi − λ0i | ≤ α (λi + λ0i ) = α
i i i
17
Proof : Let x0 ∈ U, V = U − {x0 }. Observe that V is convex, open and 0 ∈ V .
Consider the function x ∈ V → h(x) = f (x + x0 ) − f (x0 ). Obviously h is
convex, and f is continuous at x0 if, and only if, h is continuous at 0. Moreover
h(0) = 0. Since V is open and 0 ∈ V , there exists a closed ball B̄(0, r) ⊂ V . Let
{xn }n ⊂ V converge to 0. For n large enough, {xn }n ⊂ B̄(0, r). Let S denote
rxn n = − rxn .
the sphere of radius r: {x ∈ Rn : kxk = r}. Define y n = kx nk , z kxn k
n n
Then y ∈ S, z ∈ S, ∀n. One can check that
kxn ky n kxn k
r
xn = , 0= x n
+ zn
r r + kxn k r + kxn k
Since h is convex, we then get
kxn k kxn k kxn k
h(xn ) ≤ h(y n ) + 1 − h(0) = h(y n )
r r r
kxn k
≤ sup h(y)
r y∈S
kxn k
r n
0 = h(0) ≤ h(x ) + h(z n )
r + kxn k r + kxn k
kxn k
⇔0 ≤ h(xn ) + h(z n )
r
kxn k
≤ h(xn ) + sup h(y).
r y∈S
Let n → +∞. Then kxn k → 0 and
Summing up
4 Exercises
Exercise 8 1. Show that the segment [x, y] = {z : z = λx+(1−λ)y, λ ∈ [0, 1]}
is closed.
2. Let A be a non-empty set of Rn . Show that A is convex if, and only if,
for any x ∈ A, any y ∈ A, x2 + y2 ∈ A
18
Exercise 9 Let f : U → R be a continuous function on a convex, nonempty
set U of Rn . Show that f is convex on U if, and only if, ∀x ∈ U, ∀y ∈ U ,
f ((x + y)/2) ≤ (f (x) + f (y))/2.
d) Prove that:
Xn Xn n
X
( |xi + yi |p )1/p ≤ ( |xi |p )1/p + ( |yi |p )1/p
i=1 i=1 i=1
Hint:
Observe that:
n
X n
X n
X
p (p−1)
|xi + yi | ≤ |xi + yi | |xi | + |xi + yi |(p−1) |yi |.
i=1 i=1 i=1
Apply (1) to each sum of the second member of the previous inequality with p
and q = p/(p − 1).
e) Deduce from d) that the function:
Xn
x = (x1 , . . . , xn ) → ( |xi |p )1/p
i=1
is convex.
19
Exercise 12 Let f be a convex function from Rn into R. Let x, y be two
elements in Rn .
Show that the function g : t ∈ [0, 1] → f (tx + (1 − t)y) is convex and continuous.
20
is nonnegative, i.e., Ψ(h) ≥ 0, ∀h = (h1 , . . . , hn ).
Hint: Take a, a0 ∈ A. Consider the function ϕ(t) = f (ta+(1−t)a0 ) for t ∈ [0, 1].
21
Chapter 3: Convex Optimization
In this chapter we want to solve the problem min{f (x) | x ∈ C}, where f is
a convex function on Rn , and C is a convex, nonempty subset of Rn . A point
x∗ ∈ C is a global solution, or more simply a solution to this problem, or a
minimizer of f on C, if f (x∗ ) ≤ f (x), ∀x ∈ C.
The set C of constraints will be defined by a finite number of constraints
fi (x) ≤ 0, ∀i ∈ I. We give necessary and sufficient criteria to check that a point
is a minimizer of a convex function. These criteria are known as Kuhn-Tucker
Conditions.
Let I be a finite set. Then card(I) denotes the number of elements of I.
H is a hyperplane in Rn if there exists p ∈ Rn , p 6= 0 and α ∈ R such that
H = {x ∈ Rn : p · x = α}.
5 Separation theorems
Proposition 9 [First separation theorem] Let A and B be two nonempty dis-
joint convex subsets of Rn . There there exist α, β, α ≤ β and p ∈ Rn , p 6= 0,
such that p · a ≤ α ≤ β ≤ p · b, for all a ∈ A, all b ∈ B.
p · x ≤ 0 ≤ p · y, ∀x ∈ P, ∀y ∈ C
22
6 Kuhn-Tucker Conditions
6.1 Necessary and sufficient condition for optimality
The aim of this section is to give the necessary and sufficient conditions for a
point to be an optimal solution to Problem (P ):
fi (x) ≤ 0, ∀i ∈ I
(P ) Minimize f0 (x) under the constraints gi (x) ≤ 0, ∀i ∈ J
gi (x) = 0, ∀i ∈ K.
Proof : (i) Suppose f 6= 0. There exists x such that f (x) > 0. But we have a
contradiction 0 > −f (x) = f (−x) ≥ 0. Hence f = 0.
(ii) We can write ∀x, g(x) = f (x) + b where f is linear and b is a real
constant. Let x ∈ Rn . We have f (x) + b ≥ 0. Let λ > 0. We also have
λf (x) + b = f (λx) + b ≥ 0, ∀λ > 0. This is equivalent to f (x) + λb ≥ 0 for all
λ > 0. Let λ → +∞. We get f (x) ≥ 0. But x has been arbitrarily chosen.
That means f (x) ≥ 0, ∀x. Thus, f = 0 and g(x) = b, ∀x. And b ≥ 0.
Lemma 3 Let I, J and K be finite possibly empty sets, and for all i ∈ I, fi
is a convex, non-affine function from Rn into R, and for all i ∈ J ∪ K, gi is
non-null affine function. Assume there exists x0 such that
(
gi (x0 ) ≤ 0, ∀i ∈ J
gi (x0 ) = 0, ∀i ∈ K.
If the system:
fi (x) < 0, ∀i ∈ I
gi (x) ≤ 0, ∀i ∈ J
gi (x) = 0, ∀i ∈ K.
23
has no solution, then there exist nonnegative real scalars (λi )i∈I , (µi )i∈J , and
real numbers (µi )i∈K , at least one of the (λi )i∈I is not zero, which verify:
X X X
λi fi (x) + µi gi (x) + µi gi (x) ≥ 0, ∀x.
i∈I i∈J i∈K
µi ζi , ∀z ∈ Z, ∀ζ ∈ Rp+q
X X X X
λ i zi + µi zi ≥ λ i ζi + −
i∈I i∈J∪K i∈I i∈J
P P
and there exists z ∈ Z such that i∈I λi zi + i∈J∪K µi zi > 0.
If for some i, λi < 0, then letting zi tend to +∞, we get a contradiction. Thus
λi ≥ 0, ∀i ∈ I. If for some i ∈ J, one has µi < 0, then letting ζi tend to −∞,
we have another contradiction. Hence, µi ≥ 0, ∀i ∈ J. Let ε > 0 and x ∈ Rn .
Define z ∈ Z by (
zi = fi (x) + ε, ∀i ∈ I,
zi = gi (x), ∀i ∈ J ∪ K.
P P P
We have i∈I λi fi (x) + i∈J∪K µi gi (x) + ε i∈I λi ≥ 0. Let ε tend to zero.
We obtain that i∈I λi fi (x) + i∈J∪K µi gi (x) ≥ 0, ∀x ∈ Rn .
P P
To end the proof, it remains to show that at least one of the (λi )i∈I is strictly
P
positive. Assume the contrary. We then have i∈J∪K µi gi (x) ≥ 0, ∀x ∈
n
P P
R and hence i∈J∪K µi gi (x0 ) = 0. The affine function i∈J∪K µi gi has a
n
P
minimum in R . From Lemma 2, it must be equal to zero. But since i∈I λi zi +
P
i∈J∪K µi zi > 0 for some z ∈ Z, and since all the (λi )i∈I are equal to zero, there
P P
exists x such that i∈J∪K µi gi (x) > 0. That contradicts that i∈J∪K µi gi is
equal to zero.
The Problem (P ) satisfies Slater Condition (S) if there exists x0 such that:
fi (x0 ) < 0, ∀i ∈ I
(S) gi (x0 ) ≤ 0, ∀i ∈ J
gi (x0 ) = 0, ∀i ∈ K.
24
(µi )i∈J∪K such that:
(i) ∀i ∈ I, λi ≥ 0, λi fi (x) = 0,
∀i ∈ J, µi ≥ 0, µi gi (x) = 0.
(ii)
X X X X
f0 (x) + λi fi (x) + µi gi (x) ≤ f0 (x) + λi fi (x) + µi gi (x)
i∈I i∈J∪K i∈I i∈J∪K
for any x.
(iii) If f0 , (fi )i∈I , (gi )i∈J∪K are differentiable, then
X X
0 = Df0 (x) + λi Dfi (x) + µi Dgi (x)
i∈I i∈J∪K
The inequality (4) can be equivalently rewritten as h(x) ≥ α, ∀x. But h(x)−α =
P P P
i∈I λi fi (x) + i∈J∪K µi gi (x) ≤ 0. Hence, α = h(x). Thus i∈I λi fi (x) +
P
i∈J∪K µi gi (x) = 0 and since x is feasible, one has ∀i ∈ I, λi fi (x) = 0 and
∀i ∈ J, µi gi (x) = 0. We have proved Assertion (i). Now, h(x) ≥ α, ∀x is
equivalent to
X X X X
f0 (x) + λi fi (x) + µi gi (x) ≤ f0 (x) + λi fi (x) + µi gi (x)
i∈I i∈J∪K i∈I i∈J∪K
for any x. We have proved assertion (ii). Statement (iii) is obvious when
f0 , (fi )i∈I , (gi )i∈J∪K are differentiable. The proof is now complete.
25
The real numbers (λi )i∈I , (µi )i∈J∪K are called Lagrange parameters, La-
grange multipliers or Kuhn-Tucker coefficients or more simply multipliers of
Problem (P ).
The following conditions (i), (ii) and (iii) are called Kuhn-Tucker Conditions
for Problem (P ).
We say that x, (λi )i∈I , (µi )i∈J∪K satisfy Kuhn-Tucker Conditions of Problem
(P ) if they satisfy Conditions (i), (ii) and (iii).
(i) ∀i ∈ I, λi ≥ 0, fi (x) ≤ 0, λi fi (x) = 0,
∀i ∈ J, µi ≥ 0, gi (x) ≤ 0, µi gi (x) = 0.
(ii) ∀i ∈ K, gi (x) = 0.
P P P P
(iii) f0 (x)+ i∈I λi fi (x)+ i∈J∪K µi gi (x) ≤ f0 (x)+ i∈I λi fi (x)+ i∈J∪K µi gi (x)
for any x.
As particular case of Proposition 12, we get the following result when the
problem is without convex constraints.
Proof : Obvious
Remark 1 (i) The Slater condition is very important even in the one-dimensional
case. Consider the following example:
The problem has a unique solution x = 0. The Slater condition is not satisfied.
There exists no λ ≥ 0 such that 0 = x + λx2 ≤ x + λx for any x ∈ R.
(ii) It is important to notice that Slater Condition is not necessary to obtain
Kuhn-Tucker Conditions. In the previous example, Slater Condition is not
satisfied and there is no Kuhn-Tucker coefficient. Now replace this problem
by an equivalent problem which is min{f (x) = x | g(x) = |x| ≤ 0}. As before,
Slater Condition does not hold. The unique solution is always x = 0. Let λ = 1,
26
we have successively g(0) = 0, λg(0) = 0 and 0 = f (0) + λg(0) ≤ f (x) + λg(x) =
x + |x|, for any x ∈ R. In other words, Kuhn-Tucker Conditions hold.
(iii) In the previous example, one can check that Kuhn-Tucker Conditions are
sufficient for 0 to be a solution. This result is quite general as it will be proved
in the next proposition.
Proposition 13 Let x, (λi )i∈I , (µi )i∈J∪K verify Kuhn-Tucker Conditions for
Problem (P ). Then x is a solution to (P ).
Proof : Let ∀x ∈ Rn , h(x) = f0 (x) + i∈I (λi fi )(x) + i∈J∪K µi gi (x). Condi-
P P
tion (iii) is equivalent to h(x) ≥ h(x), ∀x. Combining conditions (i) and (ii), we
get h(x) ≥ f (x), ∀x. Let x satisfy the contraints of Problem (P ). We obtain
f0 (x) ≥ f0 (x), ∀x.
We now show how the Kuhn-Tucker coefficients and the optimal solutions
can be characterize in terms of the saddle-point of a certain concave-convex
function.
27
Proof : (i) Let (λ, µ, x) be a saddle-point of L. By definition of a saddle-point,
one has:
If for some i, we have fi (x) > 0, the second member of the inequality equals
+∞. Hence fi (x) ≤ 0, ∀i. If fi (x) < 0, the supremum is reached by λi = 0.
The same assertion holds for gi (x), i ∈ J. For i ∈ K, if gi (x) 6= 0 the second
member equals +∞. Hence gi (x) = 0, ∀i ∈ K.
One the other hand, we have:
28
Proof : (i) If (λ, µ, x) is a saddle-point of L, from Theorem 16, x is a solution to
(P ) and one has fi (x) ≤ 0, ∀i ∈ I, gi (x) ≤ 0, ∀i ∈ J, gi (x) = 0, ∀i ∈ I. Hence,
g(λ, µ) ≤ f0 (x), ∀(λ, µ) ∈ Rp+ × Rq+ × Rr−q and g(λ, µ) ≤ f0 (x). But from the
very definition of the saddle-point, one has g(λ, µ) ≥ L(λ, µ, x) = f0 (x). In
other words, (λ, µ) is a solution to (P ∗ ).
(ii) If g(λ, µ) = f0 (x) with x solution to (P ), one has:
X X
f0 (x) = g(λ, µ) ≤ f0 (x) + (λi fi )(x) + µi gi (x), ∀x ∈ Rn .
i∈I i∈J∪K
P P
Taking x = x, one gets 0 ≥ i∈I λi fi (x) + i∈J∪K µi gi (x) ≥ 0. Thus λi fi (x) =
0, ∀i ∈ I and µi gi (x) = 0, ∀i ∈ J. Since we have gi (x) = 0, ∀i ∈ K, Kuhn-
Tucker Conditions (i) and (ii) are fulfilled. On the other hand, we have
X X
g(λ, µ) = f0 (x) = f0 (x) + λi fi (x) + µi gi (x)
i∈I i∈J∪K
X X
≤ f0 (x) + λi fi (x) + µi gi (x), ∀x ∈ Rn .
i∈I i∈J∪K
The Kuhn-Tucker conditions (i), (ii) and (iii) are satisfied. From Assertion (ii)
of Theorem 16 (λ, µ, x) is a saddle-point of L.
7 Exercises
Exercise 15 Let f be a convex function from Rn into R. Consider the problem
(P ) min f (x)
x∈U
(i) Solve (P ).
(ii) Write its dual program (P ∗ ).
(iii) Solve (P ∗ ).
(iv) Check the duality theorem.
29
Exercise 17
fk (x) ≥ 0, ∀k = 1, . . . , p ⇒ g(x) ≥ 0
30
i) Prove that this problem has a solution.
ii) Show that this solution is in R2++ .
iii) Find a solution (x1 , x2 , λ) to the following system:
(α−1) β
αx1 x2 = λp1
α (β−1)
βx1 x2 = λp2
p 1 x1 + p 2 x2 = p · ω
x > 0, x > 0, λ > 0.
1 2
31
The following chapters will not be taught in the PreMaster program
32
Chapter4 : LinearProgramming
We say that f : Rn → R is a linear function (or a linear functional) if it is of
the form
f (x) = a1 x1 + a2 x2 + . . . , +an xn + b
where a1 , . . . , an , b are real constants.
f k (x) ≤ 0 , ∀k = 1, . . . , n ⇒ g(x) ≤ 0
p0 A ≤ 0, p ∈ Rm ⇒ p · b ≤ 0.
33
Proof : Apply Proposition 14 to the linear functionals f k , k = 1, . . . , q, −f k , k =
1, . . . , q, hk , k = 1, . . . , r and g.
holds true, it is necessary and sufficient that there exist x ∈ Rn+ such that
Ax ≥ b.
p0 [A, −I] ≤ 0 ⇒ p · b ≤ 0
where I is the identity matrix of Rm . Apply Corollary 3 to get that there exist
x ∈ Rn+ , z ∈ Rm + such that Ax − z = b. Hence Ax ≥ b. The converse is
immediate.
Proof : Let P = {p ∈ Rm 0
+ : p A ≤ 0}. This set is nonempty since 0 ∈ P . Either
there exists p ∈ P with p · b > 0 or ∀p ∈ P, p · b ≤ 0. The second statement is
equivalent to there exists x ∈ Rn+ such that Ax ≥ b.
holds true, it is necessary and sufficient that there exist x ∈ Rn such that
Ax ≥ b.
p0 [A, −A − I] ≤ 0 ⇒ p · b ≤ 0
where I is the identity matrix of Rm . Apply Corollary 3 to get that there exist
x1 ∈ Rn+ , x2 ∈ Rn+ , z ∈ Rm 1 2
+ such that Ax − Ax − z = b. Posit x = x − x .
1 2
34
Corollary 8 Let K be a skew-symmetric (n × n)-matrix (K 0 = −K). Then
the system of linear inequalities with a nonnegativity constraint Kx ≥ 0, x ≥ 0
has a solution verifying x + Kx 0.
Proof : Let i ∈ {1, . . . , n} and ei be the ith vector of the natural basis of Rn .
In view of Corollary 6 applied to the skew-symmetric matrix K, it follows that
one and only one of the two following systems:
Kx ≥ ei with x ≥ 0
minimize c · x = c1 x1 + c2 x2 + . . . + cn xn subject to
35
Obviously, at least one of the coefficients a1i must be different from zero. Assume
that is a1n . We can write
now becomes
(b1 − a11 x1 − a12 x2 − . . . − a1n−1 )
≤0
a1n
because xn ≥ 0, or equivalently
For the constraints x ∈ Rn , observe that any real number a can be written as
−
a = max(0, a) − max(0, −a). Hence, xi = x+ i i
i − xi , with x+ ≥ 0 and x− ≥ 0.
We get a new LP problem with a new variable.
The following problem will be called Problem P (P for primal)
minimize c · x = c1 x1 + c2 x2 + . . . + cn xn subject to
maximize b · p = b1 p1 + b2 p2 + . . . + bm pm subject to
36
a11 p1 + a21 p2 + . . . + am
1 pm − c1 ≤ 0
a12 p1 + a22 p2 + . . . + am
2 pm − c2 ≤ 0 (13)
...
a1n p1 + a2n p2 + ... + am
n pm − cn ≤ 0
xi (a1i p1 + . . . , +am
i pm − ci ) = 0, ∀i = 1, . . . , n (14)
ζi pi = 0, ∀i = 1, . . . , m (15)
n
X
−bi − xj aij − ζi = 0, ∀i = 1, . . . , m (16)
j=1
a1i p1 + . . . + am
i pm − ci ≤ 0, ∀i = 1, . . . , n (17)
pi ≥ 0, ∀i = 1, . . . , m (18)
Xm m
X
b·p=( pi ai1 )x1 + . . . + ( pi ain )xn
i=1 i=1
Xm m
X
c·x=( pi ai1 )x1 + . . . + ( pi ain )xn
i=1 i=1
Thus
b · p = c · x. (19)
Proposition 15 Problem P has a solution x̄ if, and only if, problem D has a
solution p̄. And we have c · x̄ = b · p̄.
Proof : Assume P has a solution x̄. Slater condition is satisfied since x̄ satisfies
the constraints. Therefore, the multipliers p̄, µ̄ exist. Define for i = 1, . . . , m,
ζ̄i = ai1 x̄1 + . . . + ain x̄n − bi . Then (x̄, ζ̄) are the multipliers for Problem D,
associated with p̄. From the Kuhn-Tucker theorem, p̄ solves D. The proof of
the converse claim is similar. From (19) we have c · x̄ = b · p̄.
am
1 . . . an
m
37
Its transposed A0 is
a11 . . . am
1
A0 = . . .
a1n . . . amn
Ax ≥ b
(P) x≥0
x ∈ Rn .
Maximize b · p subject to
A0 p ≤ c
(D) p≥0
p ∈ Rm .
Ax ≥ b
x≥0
x ∈ Rn .
A0 p ≤ c
p≥0
p ∈ Rm .
Lemma 4 If x and p are respectively feasible to problems (P) and (D), then
p · b ≤ c · x. If, moreover, p · b = c · x, then x and p are respectively optimal
solutions to problems (P) and (D).
Proof : Assume that x and p are feasible. Their transposed vectors will be
denoted by x0 , p0 . Then
p · b = p0 b ≤ p0 Ax = (p0 A) · x ≤ c0 x = c · x.
38
Lemma 5 The system of linear inequalities
Ax − tb ≥ 0
−A0 p + tc ≥ 0
p·b−c·x ≥ 0
Ax − tb + p 0
−A0 p + tc + x 0
p·b−c·x+t > 0.
b0 −c0 0
p ≥ 0, x ≥ 0, t ≥ 0.
39
deduces: p0 b ≤ p0 Ax0 ≤ 0 and c0 x ≥ p0 · Ax ≥ 0, which contradicts the relation
p · b > c · x in Lemma 5. We have thus proved that at least one feasible set is
empty. But in this case, at least one the two problems P, D has no solution.
From Proposition 15, the other problem has no solution too.
Proof : Let (x, p) be a couple of feasible elements to (P) and (D). If x and
p are optimal solutions, then the first alternative of the previous proposition
holds and c0 x = p0 b. Using the feasibility relations, one gets:
0 = p0 b − c0 x ≤ p0 Ax − c0 x = (p0 A − c0 )x ≤ 0
0 = p0 b − c0 x ≤ p0 b − p0 Ax = p0 (Ax − b) ≤ 0,
8 Exercises
Exercise 23 Consider problem P:
Minimize x1 + x2 − x4 subject to
x1 + x2 + 3x3 + 4x4 ≥ 18
3x1 + 2x2 + x3 + x4 = 10
(x1 , x2 , x3 , x4 ) ∈ R4
Maximize x + y subject to
40
x + 2y ≤ 3
−2x + y ≤ 10
5x ≤6
x ≥ 0, y ≥ 0
(a) Write this problem in its standard form, represent the feasible set and
use the complementary relations to get an optimal solution.
(b) Give the dual problem. Find an optimal solution to the dual problem
Maximize − x + 4y subject to
2x − y ≥ 4
x − 2y ≥ 3
x ≥ 0, y ≥ 0
41
Chapter 5: Non Convex Optimization
Definition 1 Let f be a continuously differentiable mapping from an open,
nonempty convex set U of Rn into Rn . Let a ∈ U . Then f (a) = (f1 (a), . . . , fn (a)).The
Jacobian matrix Jf (a) is
Df1 (a)
Jf (a) = Df2 (a)
. . . Dfn (a)
42
In particular, fx0 (0, 0, 0) · r ≥ 0, ∀r ∈ Rm such that r ≥ 0, fz0 (0, 0, 0) · t ≥ 0,
∀t ∈ Rq . Hence: fx0 (0, 0, 0) ≥ 0 and fz0 (0, 0, 0) = 0.
one gets: X X
Df0 (x) = − λi Dfi (x) + µi Dgi (x)
i∈I(x) i=1,...,K
43
Exercises
Exercise 27 Solve
√
min{(3 2x + 3y − 1)
y2
x2 + 2 = 1}
(i) Find the point of (E) which maximizes the Euclidean distance between (E)
and the axis 0y.
(ii) Find the point of (E) which minimizes the Euclidean distance between (E)
and the axis 0y.
44
Chapter 6: Dynamic Programming in Economics
9 Finite Horizon
We will present some economic models.
45
9.3 The human capital model
Consider problem P3 :
T
X
max β t u(ct )
t=0
For τ = 0, . . . , T
T
X
let Vτ (s) = max β t r(at , st )
t=τ
s.t. ∀t ≥ τ, st+1 = ψ(at , st ), at ∈ Φ(st ), sτ = s.
We obtain
and, for t ≤ T − 1
10 Infinite Horizon
The problem is:
∞
X
max β t r(at , st )
t=0
s.t. ∀t ≥ 0, st+1 = ψ(at , st ), at ∈ Φ(st ), s0 is given
46
Let
∞
X
V (s0 ) = max β t r(at , st )
t=0
s.t. ∀t ≥ 0, st+1 = ψ(at , st ), at ∈ Φ(st ), s0 is given
11 Optimal strategy
A strategy is a sequence (finite if finite horizon, infinite if infinite horizon)
(at , st+1 ) which satisfies: for any t, st+1 = ψ(at , st ), at ∈ Φ(st ). It is optimal if
it solves the problem
T
X
max β t r(at , st )
t=0
s.t. ∀t ≥ 0, st+1 = ψ(at , st ), at ∈ Φ(st ), s0 is given
where T is finite or infinite. Let (a∗t , s∗t+1 ) denote an optimal sequence. Then
we have:
a) for finite horizon: for t ≤ T − 1
47