Optimisation THM Proof
Optimisation THM Proof
Easter 2015
These notes are not endorsed by the lecturers, and I have modified them (often
significantly) after lectures. They are nowhere near accurate representations of what
was actually lectured, and in particular, all errors are almost surely mine.
Lagrangian methods
General formulation of constrained problems; the Lagrangian sufficiency theorem.
Interpretation of Lagrange multipliers as shadow prices. Examples. [2]
Network problems
The Ford-Fulkerson algorithm and the max-flow min-cut theorems in the rational case.
Network flows with costs, the transportation algorithm, relationship of dual variables
with nodes. Examples. Conditions for optimality in more general networks; *the
simplex-on-a-graph algorithm*. [3]
1
Contents IB Optimisation (Theorems with proof)
Contents
1 Introduction and preliminaries 3
1.1 Constrained optimization . . . . . . . . . . . . . . . . . . . . . . 3
1.2 Review of unconstrained optimization . . . . . . . . . . . . . . . 3
4 Non-cooperative games 9
4.1 Games and Solutions . . . . . . . . . . . . . . . . . . . . . . . . . 9
4.2 The minimax theorem . . . . . . . . . . . . . . . . . . . . . . . . 9
5 Network problems 11
5.1 Definitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
5.2 Minimum-cost flow problem . . . . . . . . . . . . . . . . . . . . . 11
5.3 The transportation problem . . . . . . . . . . . . . . . . . . . . . 11
5.4 The maximum flow problem . . . . . . . . . . . . . . . . . . . . . 12
2
1 Introduction and preliminaries IB Optimisation (Theorems with proof)
3
2 The method of Lagrange multipliersIB Optimisation (Theorems with proof)
g(λ) ≤ f (x).
In particular,
sup g(λ) ≤ inf f (x).
λ∈Y x∈X(b)
Proof.
g(λ) = inf
0
L(x0 , λ)
x ∈X
≤ L(x, λ)
= f (x) − λT (h(x) − b)
= f (x).
4
2 The method of Lagrange multipliersIB Optimisation (Theorems with proof)
φ(b) + λT (c − b) ≤ φ(c),
or
φ(b) ≤ φ(c) − λT (c − b),
This implies that
= g(λ)
φ(b) = g(λ)
= inf L(x, λ)
x∈X
≤ inf L(x, λ)
x∈X(c)
= φ(c) − λT (c − b)
5
2 The method of Lagrange multipliersIB Optimisation (Theorems with proof)
Proof. Consider b1 , b2 ∈ Rm such that φ(b1 ) and φ(b2 ) are defined. Let δ ∈ [0, 1]
and define b = δb1 + (1 − δ)b2 . We want to show that φ(b) ≤ δφ(b1 ) + (1 − δ)φ(b2 ).
Consider x1 ∈ X(b1 ), x2 ∈ X(b2 ), and let x = δx1 + (1 − δ)x2 . By convexity
of X, x ∈ X.
By convexity of h,
φ(b) ≤ f (x)
= f (δx1 + (1 − δ)x2 )
≤ δf (x1 ) + (1 − δ)f (x2 )
This holds for any x1 ∈ X(b1 ) and x2 ∈ X(b2 ). So by taking infimum of the
right hand side,
φ(b) ≤ δφ(b1 ) + (1 − δ)φ(b2 ).
So φ is convex.
Theorem. If a linear program is feasible and bounded, then it satisfies strong
duality.
6
3 Solutions of linear programs IB Optimisation (Theorems with proof)
x = δy + (1 − δ)z.
We will show there exists an optimal solution strictly fewer than r non-zero
entries. Then the result follows by induction.
By optimality of x, we have cT x ≥ cT y and cT x ≥ cT z.
Since cT x = δcT y + (1 − δ)cT z, we must have that cT x = cT y = cT z, i.e. y
and z are also optimal.
Since y ≥ 0 and z ≥ 0, x = δy + (1 − δ)z implies that yi = zi = 0 whenever
xi = 0.
So the non-zero entries of y and z is a subset of the non-zero entries of x. So
y and z have at most r non-zero entries, which must occur in rows where x is
also non-zero.
If y or z has strictly fewer than r non-zero entries, then we are done. Other-
wise, for any δ̂ (not necessarily in (0, 1)), let
7
3 Solutions of linear programs IB Optimisation (Theorems with proof)
Theorem. Let x and λ be feasible for the primal and the dual of the linear
program in general form. Then x and λ and optimal if and only if they satisfy
complementary slackness, i.e. if
cT x = λ T b
cT x = λT b
= inf
0
(cT x0 − λT (Ax0 − b))
x ∈X
≤ cT x − λT (Ax − b)
≤ cT x.
then
cT x = cT x − λT (Ax − b) = (cT − λT A)x + λT b = λT b.
Hence by weak duality, x and λ are optimal.
8
4 Non-cooperative games IB Optimisation (Theorems with proof)
4 Non-cooperative games
4.1 Games and Solutions
Theorem (Nash, 1961). Every bimatrix game has an equilibrium.
The left hand side is the worst payoff the row player can get if he employs the
minimax strategy. The right hand side is the worst payoff the column player
can get if he uses his minimax strategy.
The theorem then says that if both players employ the minimax strategy,
then this is an equilibrium.
Proof. Recall that the optimal value of max min p(x, y) is a solution to the linear
program
maximize v such that
m
X
xi pij ≥ v for all j = 1, · · · , n
i=1
m
X
xi = 1
i=1
x≥0
n
Adding slack variable z ∈ R with z ≥ 0, we obtain the Lagrangian
n m
! m
!
X X X
L(v, x, z, w, y) = v + yj xi pij − zj − v − w xi − 1 ,
j=1 i=1 i=1
y≥0
9
4 Non-cooperative games IB Optimisation (Theorems with proof)
This corresponds to the column player choosing a strategy (yi ) such that the
expected payoff is bounded above by w.
The optimum value of the dual is min max p(x, y). So the result follows from
y∈Y x∈X
strong duality.
max
0
p(x0 , y) = min
0
max
0
p(x0 , u0 )
x ∈X y ∈Y x ∈X
i.e. the x, y are optimizers for the max min and min max functions.
10
5 Network problems IB Optimisation (Theorems with proof)
5 Network problems
5.1 Definitions
5.2 Minimum-cost flow problem
5.3 The transportation problem
Theorem. Every minimum cost-flow problem with finite capacities or non-
negative costs has an equivalent transportation problem.
Proof. Consider a minimum-cost flow problem on network (V, E). It is wlog to
assume that mij = 0 for all (i, j) ∈ E. Otherwise, set mij to 0, mij to mij − mij ,
bi to bi − mij , bj to bj + mij , xij to xij − mij . Intuitively, we just secretly ship
the minimum amount without letting the network know.
Moreover, we can assume that all capacities are finite: if some edge has
infinite capacity but non-negativeP cost, then setting the capacity to a large
enough number, for example i∈V |bi | does not affect the optimal solutions.
This is since cost is non-negative,Pand the optimal solution will not want shipping
loops. So we will have at most |bi | shipments.
We will construct an instance of the transportation P problem as follows:
For every i ∈ V , add a consumer with demand k:(i,k)∈E mik − bi .
For every (i, j) ∈ E, add a supplier with supply mij , an edge to consumer i
with cost c(ij,i) = 0 and an edge to consumer j with cost c(ij,j) = cij .
P
i k:(i,k)∈E mik − bi
0
mij ij
cij P
j k:(j,k)∈E mjk − bj
The idea is that if the capacity of the edge (i, j) is, say, 5, in the original network,
and we want to transport 3 along this edge, then in the new network, we send 3
units from ij to j, and 2 units to i.
The tricky part of the proof is to show that we have the same constraints in
both graphs.
For any flow x in the original network, the corresponding flow on (ij, j) is
xij and the flow on (ij, i) is mij − xij . The total flow into i is then
X X
(mik − xik ) + xki
k:(i,k)∈E k:(k,i)∈E
11
5 Network problems IB Optimisation (Theorems with proof)
which is exactly the constraint for the node i in the original minimal-cost flow
problem. So done.
δ = min{C(S) : S ⊆ V, 1 ∈ S, n ∈ V \ S}
δ = fx (S, V \ S) − fx (V \ S, S).
δ = C(S).
12