0% found this document useful (0 votes)
4 views

Cone Programming

Uploaded by

carlosgg33
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views

Cone Programming

Uploaded by

carlosgg33
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 27

Agenda

1 Cone programming
2 Convex cones
3 Generalized inequalities
4 Linear programming (LP)
5 Second-order cone programming (SOCP)
6 Semidefinite programming (SDP)
7 Examples
Optimization problem in standard form

minimize f0 (x)
subject to fi (x) ≤ 0 i = 1, . . . , m
hi (x) = 0 i = 1, . . . , p

x ∈ Rn
f0 : Rn → R (objective or cost function)
fi : R n → R (inequality constraint functionals)
n
hi : R → R (equality constraint functionals)
Terminology

x is feasible if x obeys the constraints


feasible set C: set of all feasible points
optimal value: p? = inf{f0 (x), x ∈ C}
can be −∞; e.g. min log(x), x > 0.
by convention, p? = ∞ if C = ∅ (problem infeasible)
optimal solution: x? s.t. f (x? ) = p?
there may be no optimal solution: e.g. min log(x), x > 0
optimal set: {x : f (x) = p? }
Convex optimization problem

Convex optimization problem in standard form

minimize f0 (x)
subject to fi (x) ≤ 0 i = 1, . . . , m
aTi x = bi i = 1, . . . , p

f0 , f1 , . . . , fm convex
affine equality constraints Ax = b, A ∈ Rp×n
feasible set is convex

Abstract convex optimization problem

minimize f0 (x)
subject to x∈C

f0 convex
C convex
Why convexity?
A convex function has no local minimum that is not global

convex not convex

A convex set is connected and has feasible directions at any point

convex + not convex


feasible directions

A convex function is continuous and has some differentiability properties


Convex functions arise prominently in duality
Cone programming I

LP
minimize cT x
subject to Fx + g ≥ 0
Ax = b

Nonlinear programming → nonlinear constraints


Express nonlinearity via generalized inequalities
Orderings of Rn and convex cones

K is a convex cone if
(i) K is convex
(ii) K is a cone (i.e. x ∈ K =⇒ λx ∈ K ∀λ ≥ 0)
K is pointed if
(iii) x ∈ K and − x ∈ K =⇒ x = 0
(K does not contain a straight line through the origin)

Example: K = {x ∈ Rn : x ≥ 0} is a pointed convex cone

Two additional properties of Rn+


(iv) Rn+ is closed
(v) Rn+ has a nonempty interior
Implication: ordering
a K b ⇐⇒ a − b ∈ K

(i) - (iii) ensure that this is a good ordering


1 reflexivity: a  a follows from 0 ∈ K
2 antisymmetry: a  b, b  a =⇒ a = b (since K is pointed)
3 transitivity: a  b, b  c =⇒ a  c (since K is convex and a cone)

→ compatibility with linear operations


a  b & λ ≥ 0 =⇒ λa  λb
a  b & c  d =⇒ a + c  b + d

Good properties of LPs come from these properties

4 closedness: ai  bi , ai → a, bi → b =⇒ a  b
5 nonempty interior allows us to define strict inequalities:
a  b ⇐⇒ a − b ∈ int(K)
Examples of cones

Nonnegative orthant Rn+


{x ∈ Rn : x ≥ 0}

Second-order (or Lorentz or ice cream) cone


q
{x ∈ Rn+1 : x21 + . . . + x2n ≤ xn+1 }

Positive semidefinite cone

{X ∈ S n : X  0}
Cone programming II

minimize cT x
subject to Fx + g  0
Ax = b

K = Rn+ =⇒ linear programming


Minimize linear functional over an affine slice of a cone
Very fruitful point of view
useful theory (duality)
useful algorithms (interior point methods)
Linear programming (LP)

minimize cT x
subject to Fx + g ≥ 0
Ax = b

Linear objective
Linear equality and inequality constraints

x* (optimal)
Feasible set is a polyhedron
c

cTx = constant

Many problems can be formulated as LP’s


Example: Chebyshev approximation

A ∈ Rm×n
b ∈ Rm
minimize kAx − bk∞ ⇐⇒ minimize maxi=1,...,m |aTi x − bi |

Different from LS problem: minimizekAx − bk2


LP formulation (epigraph trick)

⇐⇒ minimize t
subject to |aTi x − bi | ≤ t ∀i

⇐⇒ minimize t
subject to − t ≤ aTi x − bi ≤ t ∀i
optimization variables (x, t) ∈ Rn+1
Example: basis pursuit

A ∈ Rm×n
b ∈ Rm
minimize kxk1
subject to Ax = b
LP formulations:
(a) P
minimize ti
subject to −ti ≤ xi ≤ ti
Ax = b
optimization variables (x, t) ∈ R2n
(b) P −
x+
P
minimize i + xi
subject to A(x+ − x− ) = b
x+ , x− ≥ 0
optimization variables (x+ , x− ) ∈ R2n
Second-order cone programming (SOCP)

minimize cT x
subject to kFi x + gi k2 ≤ cTi x + di i = 1, . . . , m
Ax = b

 
Fi x + gi
kFi x + gi k2 ≤ cTi x + di ⇐⇒ ∈ Li = {(yi , t) : kyi k ≤ t}
cTi x + di
(hence the name)

SOCP ⇐⇒ minimize cT x  


Fi gi
subject to x + ∈ Li
cTi di
Ax = b
affine mapping    
 Fi g
Fx + g = x+ i
cTi di i=1,...,m

cone product
K = L1 × L2 × . . . × Lm

   
Fi gi
x + ∈ Li ∀i ⇐⇒ F x + g ∈ K
cTi di

minimize cT x
∴ SOCP ⇐⇒ subject to Fx + g ∈ K
Ax = b
this is a cone program
Example: support vector machines

n pairs (xi , yi )
xi ∈ Rp : feature/explanatory variables
yi ∈ {−1, 1}: response/class label
Examples
xi : infrared blood absorption spectrum
yi : person is diabetic or not

SVM model: SVM as a penalized fitting


procedure
n
X
min [1 − yi f (xi )]+ + λkβk2 hinge loss
β
i=1
[1-yf(x)]+

f (x) = xT β
0 1
yf(x)
sometimes f (x) = xT β + β0 and same
minimum
SVM: formulation as an SOCP

Variables (β, t) ∈ Rp×n

ti + λkβk2 ti + λkβk2
P P
minimize ⇐⇒ minimize
subject to [1 − yi f (xi )]+ ≤ ti subject to yi f (xi ) ≥ 1 − ti
ti ≥ 0

this an SOCP, since SOCP’s are more general than QP’s and QCQP’s

Equivalence P
minimize ti + λu
subject to kβk2 ≤ u
yi f (xi ) ≥ 1 − ti
ti ≥ 0

 u + 1 2  u − 1 2  
β u+1
kβk2 ≤ u ⇐⇒ kβk2 ≤ − ⇐⇒ u−1 ≤
2 2 2 2
QP ⊂ SOCP ( =⇒ LP ⊂ SOCP)
QCQP
1 T T
minimize 2 x P0 x + q0 x + r0
1 T T
subject to 2 x Pi x + qi x + ri ≤ 0
P0 , Pi  0

QCQP ⊂ SOCP
quadratic convex inequalities are SOCP-representable
Example: total-variation denoising
Observe
bij = fij + σzij 0 < i, j < n

f is original image
b is a noisy version
Problem: recover original image (de-noise)
Min-TV solution
minimize kxkT V
subject to kx − bk ≤ δ
TV norm  
X xi+1,j − xi,j
kxkTV = kDij xk2 Dij x =
xi,j+1 − xi,j
Formulation as an SOCP
P
minimize tij
subject to kDij xk2 ≤ tij
kx − bk2 ≤ δ
Semidefinite programming (SDP)

minimize cT x
subject to F (x) = x1 F1 + . . . + xn Fn − F0  0

Fi ∈ S p (p × p symmetric matrices)
linear matrix inequality (LMI): F (x)  0
multiple LMI’s can be combined into one:
 
F1 (x)
Fi (x)  0 i = 1, . . . , m ⇐⇒ 
 .. 0

.
Fm (x)
SOCP ⊂ SDP (but the converse is not true!)
 
n+1 tIm x
(x, t) ∈ R : kxk ≤ t ⇐⇒ 0
xT t

SOCP constraints are LMI’s

Hierarchy: LP ⊂ SOCP ⊂ SDP

Many nonlinear convex problems can be cast as SDP’s


Example: minimum-norm problem

minimize kA(x)k
subject to A(x) = x1 A1 + . . . + xn An − B
with Ai ∈ Rp1 ×p2 , is equivalent to

minimize t
subject to kA(x)k ≤ t
 
tIp1 A(x)
kA(x)k ≤ t ⇐⇒ 0
AT (x) tIp2

Why? Eigenvalues of this matrix are {t ± σi (A(x))}


Example: nuclear-norm minimization

P
minimize kXk∗ = σi (X)
subject to Xij = Bij (i, j) ∈ Ω ⊂ [p1 ] × [p2 ]
This is an SDP (proof, later)
Stability analysis for dynamical systems

Linear system
dv
= v̇(t) = Qv(t) Q ∈ Rn×n
dt

Main question: is this system stable? i.e. do all trajectories tend to zero as
t → ∞?
Simple sufficient condition: existence of a quadratic Lyapunov function
(i) L(v) = v T Xv X  0
d
(ii) L̇ = dt L(v(t)) ≤ −αL(v(t)) (α > 0) for any trajectory

This condition gives L(v(t)) = v T (t)Xv(t) ≤ exp(−αt) L(v(0)) (Gronwall’s


inequality), whence

X  0 =⇒ v(t) → 0 as t→∞

Exsitence of X  0 and α > 0 provides a certificate of stability


dv
= v̇(t) = Qv(t), L(v) = v T Xv X  0
dt
d  T
v (t)Xv(t) = v̇ T Xv + v T X v̇ = v T (QT X + XQ)v

L̇ =
dt
i.e. L̇ ≤ −αL ⇐⇒ v T (QT X + XQ + αX)v <0 ∀v
⇐⇒ QT X + XQ + αX ≺0
Conclusion: to certify stability, it suffices to find X obeying

X  0, QT X + XQ ≺ 0

If the optimal value of SDP

minimize t 
X + tI 0
subject to 0
0 −(QT X + XQ) + tI

is negative, then the system is stable


Extension

v̇(t) = Q(t)v(t)
Q(t) ∈ conv{Q1 , . . . , Qn } time-varying
L(v) = v T Xv (X  0) s.t. L̇ ≤ −αL =⇒ stability

Similar calculations show that for all v


v T (QT (t)X + XQ(t) + αX)v ≤ 0
T
⇐⇒ Q (t)X + XQ(t) + αX ≺ 0, ∀Q(t) ∈ conv{Q1 , . . . , Qn }
⇐⇒ QTi X + XQi + αX ≺ 0, ∀i = 1, . . . , k

If we can find X such that

X0 & QTi X + XQi ≺ 0 ∀i = 1, . . . , k

then we have stability

This is an SDP!
References

1 A. Ben-Tal and A. Nemirovski, Lectures on Modern Convex Optimization:


Analysis, Algorithms, and Engineering Applications, MPS-SIAM Series on
Optimization

2 S. Boyd and L. Vandenberghe, Convex Optimization, Cambridge University


Press

You might also like