Optimality Conditions: Unconstrained Optimization: 1.1 Differentiable Problems
Optimality Conditions: Unconstrained Optimization: 1.1 Differentiable Problems
Optimality Conditions: Unconstrained Optimization: 1.1 Differentiable Problems
Optimality Conditions:
Unconstrained Optimization
1.1
Differentiable Problems
where lim
o(tkf (x)k)
f (x tf (x)) f (x)
.
= kf (x)k2 +
t
t
13
f (x + td) f (x)
1 T 2
o(t2 )
=
d
f
(x)d
+
t2
2
t2
f (x) f (x)
1 (x x)T 2
(x x)
o(kx xk2 )
=
f
(x)
+
.
kx xk2
2 kx xk
kx xk
kx xk2
o(kx xk2 )
kx xk2
4
whenever kx xk < . Then for all kx xk < we have from (1.1.2) and (1.1.3) that
f (x)f (x)
kxxk2
2)
1
+ 0(kxxk
2
kxxk2
1
.
4
1.2
15
Convex Problems
Observe that Theorem 1.1.1 establishes firstorder necessary conditions while Theorem 1.1.2
establishes both secondorder necessary and sufficient conditions. What about firstorder
sufficiency conditions? For this we introduce the following definitions.
Definition 1.2.1 [Convex Sets and Functions]
1. A subset C Rn is said to be convex is for every pair of points x and y taken from C,
the entire line segment connecting x and y is also contained in C, i.e.,
[x, y] C
where
[x, y] = {(1 )x + y : 0 1} .
Lemma 1.2.1 The function f : Rn R is convex if and only if for every two points
x1 , x2 dom (f ) and [0, 1] we have
f (x1 + (1 )x2 ) f (x1 ) + (1 )f (x2 ).
That is, the secant line connecting (x1 , f (x1 )) and (x2 , f (x2 )) lies above the graph of f .
Example: The following functions are examples of convex functions: cT x, kxk, ex , x2
The significance of convexity in optimization theory is illustrated in the following result.
f (x + td) f (x)
t
f 0 (x; d) := lim
t0
f (x + td) f (x)
.
t
Now if the difference quotient (1.2.4) is non-decreasing in t on (0, +), then the limit in
(1.2.6) is necessarily given by the infimum in (1.2.5). This infimum always exists and so
f 0 (x; d) always exists and is given by (1.2.5).
We now prove (1.2.4). Let x dom (f ) and d Rn . If x + td
/ dom (f ) for all t > 0,
then the result is obviously true. Thus, we may assume that
0 < t = sup{t : x + td dom (f )}.
17
Let 0 < t1 < t2 < t (we allow the possibility that t2 = t if t < +). Then
t1
td
h
t2 2
i
t1
f 1 t2 x + tt21 (x + t2 d)
1 tt12 f (x) + tt12 f (x + t2 d).
f (x + t1 d) = f x +
=
Hence
f (x + t1 d) f (x)
f (x + t2 d) f (x)
.
t1
t2
We now show Part 3 of this result. To see that f 0 (x; ) is positively homogeneous let
d Rn and > 0 and note that
f 0 (x; d) = lim
t0
f (x + (t)d) f (x)
= f 0 (x; d).
(t)
f 0 (x; u + v) = lim
t0
=
=
=
=
19
by (b).
2) Suppose f is convex and let x, d Rn , then by (b) of Part (1),
f (x + td) f (x) + tf (x)T d
for all t R. Replacing the left hand side of this inequality with its secondorder
Taylor expansion yields the inequality
f (x) + tf (x)T d +
t2 T 2
d f (x)d + o(t2 ) f (x) + tf (x)T d
2
or equivalently
1 t 2
o(t2 )
d f (x)d + 2 0.
2
t
Letting t 0 yields the inequality
dT 2 f (x)d 0.
Since d was arbitrary, 2 f (x) is positive semi-definite.
Conversely, if x, y Rn , then by the mean value theorem there is a (0, 1) such that
1
f (y) = f (x) + f (x)T (y x) + (y x)T 2 f (x )(y x)
2
where x = y + (1 )x. Hence
f (y) f (x) + f (x)T (y x)
since 2 f (x ) is positive semi-definite. Therefore, f is convex by (b) of Part (1).
We have established that f 0 (x; d) exists for all x dom (f ) and d Rn , but we have not
yet discussed to continuity properties of f . We give a partial result in this direction in the
next lemma.
Lemma 1.2.3 Let f : Rn R {+} be convex. Then f if bounded in a neighborhood of
a point x if and only if f is Lipschitz in a neighborhood of x.
g(x)
+
g(w).
1 + 1
1 + 1
1 + 1
1 + 1
!
Consequently,
g(y) g(x)
1
(g(w) g(x)) 2M 1 = 2M 1 kx yk.
1 + 1
1.3
Convex composite optimization is concerned with the minimization of functions of the form
f (x) := h(F (x)) where h : Rm R {+} is a proper convex function and F : Rn Rn is
continuously differentiable. Most problems from nonlinear programming can be cast in this
framework.
Examples:
(1) Let F : Rn Rm where m > n, and consider the equation F (x) = 0. Since m > n
it is highly unlikely that a solution to this equation exists. However, one might try
to obtain a best approximate solution by solving the problem min{kF (x)k : x Rn }.
This is a convex composite optimization problem since the norm is a convex function.
(2) Again let F : Rn Rm where m > n, and consider the inclusion F (x) C, where
C Rn is a non-empty closed convex set. One can pose this inclusion as the optimization problem min{dist(F (x)|C) : x Rn }. This is a convex composite optimization
problem since the distance function
dist(y | C) := inf ky zk
zC
is a convex function.
21
where
f (x; d) := h(F (x) + F 0 (x)d) h(F (x)).
From 1.3.8, one immediately obtains the following result.
Lemma 1.3.1 Let h : Rn R be convex and let F : Rn Rm be continuously differentiable.
Then the function f = h F is everywhere directional differentiable and one has
(1.3.9)
This result yields the following optimality condition for convex composite optimization
problems.
Theorem 1.3.1 Let h : Rm R be convex and F : Rn Rm be continuously differentiable.
If x is a local solution to the problem min{h(F (x))}, then d = 0 is a global solution to the
problem
(1.3.10)
x) + F 0 (
x)d).
minn h(F (
dR
There are various ways to test condition 1.3.8. A few of these are given below.
1.3.1
for 0 < 1 2 ,