Unconstrained Minimization
Unconstrained Minimization
minimize f (x)
where f : RN → R is convex and twice differentiable. We assume that the optimal
value p ? = inf x f (x) is attained (and finite).
Optimality condition
f is differentiable and convex, a point x ? is optimal iff ∇f (x ? ) = 0.
Sometimes, we can analytically solve the optimality equation but usually the problem
must be solved by an iterative algorithm.
Descent methods
We want to produce a minimizing sequence x (k) , k = 1, ..., where
x (k+1) = x (k) + t (k) ∆x (k) with f (x (k+1) ) < f (x (k) )
I ∆x (k) is the search direction
I t (k) > 0 is the step size
Descent direction
∆x (k) is a descent direction if ∇f (x (k) )> ∆x (k) < 0 (from convexity).
Algorithm 1: General descent method
Start at a point x ∈ dom f
repeat
Determine a descent direction ∆x
Choose a step size t > 0
Update x = x + t∆x
until stopping criterion is satisfied;
Choosing the step size - Line search
1
f (x + v ) ≈ f (x) + ∇f (x)> v + v > ∇2 f (x)v
2
I or equivalently solves the linearized optimality condition
∇f (x + v ) ≈ ∇f (x) + ∇2 f (x)v = 0
More
Explore further
I Steepest descent methods (generalizes gradient descent)
I Quasi-Newton methods (like BFGS, L-BFGS)
I Convergence analysis
I Self-concordance
I Gradient-free optimization