Lecture 1
Lecture 1
PPGEQ – UFRGS
Prof. Jorge Otávio Trierweiler
logo
Outline
1 Introduction
Introduction
Mathematical Background
Basic Concepts
2 Unconstrained Optimization
3 Constrained Optimization
4 Final Remarks
logo
Introduction
Introduction
logo
Introduction
Optimization Viewpoints
Introduction
Motivation
minimize f (x)
x
subject to h(x) = 0
g(x) ≤ 0
logo
Introduction
AOptimization
Aplicações
li õ de Otiin iProcess
d Otimizaçãoã emSystem
Process
P Engineering
S
System
t Engineering
E i i
MHENs (Redes) x x x x
R ti
Reactions x x x
Separation Systems x x x
Flowsheeting x x
Process Operations
Planning/ Scheduling x x x
Supply Chain x x x
Real Time Optimization x x x
Control
Predictive Control x x
Nonlinear Predictive Control x x logo
Hybrid Control x
Introduction
Process Optimization
subject to hi (x) = 0, i = 1, . . . , m
gj (x) ≤ 0, j = 1, . . . , r logo
Introduction
Some Remarks
x1 − x2 ≥ 0 ⇒ −x1 + x2 ≤ 0
ε-constrained method:
minimize f1 (x)
x
logo
subject to f2 (x) ≤ ε
Introduction
Class of Problems
Introduction
Modern Optimization
Modern Optimization
x2
Linear Programming Kantorovich (1939), Dantzig (1947)
x1
x2
Nonlinear Programming Karush (1939), Kuhn, A.W.Tucker (1951)
x1
y2
logo
y1
Introduction
Computational progress:
much faster algorithms/much faster computers
logo
logo
APPENDIX A. OPTIMIZATION OVERVIEW
LOGMIP
Generalized Disjunctive Programming (linear and non linear)
(GDP) Big-M/Convex hull Relaxation
Balas (1985)
Vecchietti and Grossmann (1999, 2000) LaGO
Lee and Grossmann (2000)
Genetic
OQNLP
Algorithms
Random
Search (RS) Simulated MSNLP
Annealing
. Chemical Processes
OTS
Scatter
Search
FilMINT
Constrained Optimization
of
Duran and Grossmann (1986) (dynamic optimization)
PTIMIZATION TOOLS
Viswanathan and Grossmann (1990)
Fletcher and Leyffer (1994)
Optimization
MINLP_BB
Generalized Benders Decomposition
(GBD)
Benders (1962), Geoffrion (1972)
SBB
Branch and Bound Method (B&B)
Gupta and Ravindran (1985)
Borchers and Mitchell (1994) LINDOGlobal
Leyffer (2001), Bussieck and Drud (2001)
α-BB
(underestimators)
1: OVERVIEW OF O
IPOPT
Interior Point Methods (IP)
LOQO
Forsgren, Gill, and Wright (2002)
Wächter (2002), Byrd et al. (1999)
Escobar
KNITRO
Augmented Langrangian Methods
Conn,Gould and Toint (1992)
LANCELOT
Reduced Gradient Methods
Unconstrained Optimization
FIGURE
Non Linear Programming
Murtagh and Saunders (1998)
(NLP) CONOPT
Optimization Overview
Prof. Marcelo
(SQP)
Gill, Murray and Saunders (1997)
MINOS
Byrd, Hribar and Nocedal (1999)
Successive linear programming (SLP)
Fletcher and Sainz de la Maza (1989) SNOPT
Byrd, Gould and Nocedal (2003)
NPSOL
Branch and Cut
LGO
Crowder et al. (1983)
Johnson et al. (2000)
COIN-CBC
Mixed Integer Linear Cutting Planes (CP)
Programming (MILP) Gomory (1960), Balas et al (1993)
COIN-GLPK
Branch and Bound (B&B)
Lang and Doing (1960), Dakin (1965)
MOSEK
XA
Interior Point
Karmarkar (1984)
Linear Programming (LP)
OSL
Simplex
Dantzig (1949)
XPRESS
CPLEX
Solution Modeling
Classes Major Codes Interfaces
Algorithms System
Introduction
Introduction
Introduction Unconstrained Optimization Constrained Optimization Final Remarks
Mathematical Background
Linear Algebra
Mathematical Background
Mathematical Background
Convex function
Planef (x) = k
I I
I I
I I
I I
I I
FIGURE 4.11
Illustration of a convex set formed by a plane f(x) = k cutting a convex function. logo
Mathematical Background
45
40
35
30
25
20
15
10
5 10 15 20 25 30 35 40 45
Mathematical Background
Mathematical Background
Multivariable Extension:
If λ > 0 of H(x) the curvature of the function is upwards;
If λ < 0 of H(x) the curvature of the function is downwards;
For indefinite Hessian: saddle point;
For ∇f (x) = 0: stationary point;
logo
Mathematical Background
1
f (x) ≈ f (xk ) + ∇f (xk )T d + d T H(xk )d
2
where d = x − xk is the direction vector joining the point x and xk .
logo
Mathematical Background
Solving for x:
xk+1 = xk − ∇c(xk )−1 c(xk )
Mathematical Background
Vectorial Algebra
Consider the matrices A and B and the vectors x and y the following
properties holds:
(A+B)T = AT + B T
(A×B)T = B T × AT
xT y = y T x
Basic Concepts
Nonlinear Programming
minimize f (x)
x
subject to h(x) = 0
g(x) ≤ 0
where:
f (x): objective function
h(x): equality constraints
g(x): inequality constraints
x: decision variables
Feasible Region:
logo
FR = x| h(x) = 0, g(x) ≤ 0
Basic Concepts
What is a solution?
Global Minimum:
A point x ∗ is a global minimizer if f (x ∗ ) < f (x) ∀ x ∈ F R
Local Minimum:
logo
A point x ∗ is a local minimizer if f (x ∗ ) < f (x) ∀ x ∈ N = x| kx − x ∗ k ≤ δ
Basic Concepts
A set is convex if it contains all the line segments connecting any pair of its
points.
Basic Concepts
logo
Basic Concepts
Unconstrained Optimization
Unconstrained Optimization
General form:
minimize f (x)
x
What conditions characterize an optimal solution
x2 Unconstrained Local Minim
Necessary Conditions
∇f (x*) = 0
pT∇2f (x*) p 0 for p∈ℜ
(positive semi-definite)
x*
Unconstrained Local Minim
Sufficient Conditions
∇f (x*) = 0
Contours of f(x) pT∇2f (x*) p > 0 for p∈ℜ
(positive definite)
x1
logo
For smooth functions, why are contours around optimum elliptical?
What conditions characterize
Taylor an optimal
Series in n dimensions aboutpoint?
x*:
Prof. Marcelo Escobar Optimization of Chemical Processes
1
Introduction Unconstrained Optimization Constrained Optimization Final Remarks
Unconstrained Optimization
Optimality Conditions
Unconstrained Optimization
Unconstrained Optimization
maximum saddle
λ1 = −2; λ2 = −2 λ1 = 2; λ2 = −2 logo
Unconstrained Optimization
logo
Unconstrained Optimization
logo
Unconstrained Optimization
f (x) = f (x ∗ ) + ∇f (x ∗ )T d
f (x) = f (x ∗ ) − ∇f (x ∗ )T ∇f (x ∗ )
∇f (x ∗ ) = 0
logo
Unconstrained Optimization
1
f (x) = f (x ∗ ) + ∇f (x ∗ )T d + d T H(xk )d
2
Since ∇f (x ∗ ) = 0.
The function increases (f (x) = f (x ∗ ) for any d 6= 0, only if
d T H(xk )d > 0
logo
Constrained Optimization
General form:
minimize f (x)
x
subject to hi (x) = 0, i = 1, . . . , m
gj (x) ≤ 0, j = 1, . . . , r
where:
f (x): objective function
h(x): equality constraints
g(x): inequality constraints
Constrained Optimization
minimize f (x)
x
subject to hi (x) = 0, i = 1, . . . , m
Feasibility:
h(x) = 0 logo
Constrained Optimization
logo
Constrained Optimization
subject to hi (x) = 0, i = 1, . . . , m
gj (x) = 0, j = 1, . . . , r
First Order Necessary Conditions:
Feasibility:
h(x) = 0; g(x) ≤ 0
Complementarity: logo
µj gj (x) = 0, µj ≥ 0
Constrained Optimization
Feasibility:
h(x) = 0; gA (x) = 0
logo
Constrained Optimization
If all g(x) are inactive, the first order conditions reduces to:
∇f (x ∗ ) = 0
On the other hand if one constraint is active. The Region R3 is empty only if
∇f (x ∗ ) and ∇g(x ∗ ) are parallel and have opposite directions.
∇f (x ∗ ) + µ∇g(x ∗ ) = 0 µ≥0
For multiple active constraints, the gradients must be linear dependent. logo
Constrained Optimization
Geometrical Interpretation
Optimal solution for inequality constrained problem
Min f(x)
s.t. g(x) 0∗ ∗
∇f (x ) + µ1 ∇g
Analogy: Ball rolling down 1 (x ) = 0
valley pinned by fence logo
≥ ∇g
Note: Balance of forcesµ(∇f, 0 1)
35
Constrained Optimization
Geometrical Interpretation
Optimal solution for general constrained problem
Constrained Optimization
Stationarity:
m
X r
X
∇f (x) + λi ∇hi (x) + µj ∇gj (x) = 0
i=1 j=1
Feasibility:
hi (x) = 0 ; gj (x) ≤ 0
Complementarity:
µj gj (x) = 0
µj ≥ 0 logo
Constrained Optimization
Stationarity:
Feasibility:
h(x) = 0 ; g(x) ≤ 0
Complementarity:
µ.g(x) = 0
µ≥0
logo
Constrained Optimization
Stationarity:
∇x L(x, λ, µ) = 0
Feasibility:
∇λ L(x, λ, µ) = 0 ; ∇µ L(x, λ, µ) ≤ 0
Complementarity:
µ.g(x) = 0
logo
µ≥0
Constrained Optimization
p T ∇2 L(x ∗ , λ∗ , µ∗ )p > 0
where,
∇h(x ∗ )T p = 0
∇g( x ∗ )T p ≤ 0
Constrained Optimization
∂f (x) ∂f (x)
λi = − µj = −
∂hi (x) ∂gj (x)
Also known as:
logo
Shadow Prices; Dual Variables; Lagrange Multipliers.
Constrained Optimization
hi (x) = 0 i = 1, . . . , m
gj (x) = 0 j ∈ J
3. If gj (x) ≤ 0 and µj ≥ 0 ∀j, then STOP. Solution found!
4. If any gj (x) > 0 (violated), or µj < 0 (wrong sign) then
Constrained Optimization
Then
−1 1 0
∇f (x) = x1 − 3 x2 − 1 ∇g1 (x) = ∇g2 = ∇g3 =
1 − 21 −1
Constrained Optimization
Iteration 2: µ1 = µ3 = 0, µ3 ≥ 0
KKT Conditions: ∇f (x) + µ2 ∇g2 (x) = 0; g2 (x) = 0
x1 − 3 −1
x2 − 1 + µ 2 x1 2.6
1
= 0 ⇒ x2 = 1.2
µ2 1.4
x1 − 21 x2 − 2
Final Remarks
Final Remarks
Final Remarks
Further Readings
Final Remarks
logo
Presented by: Course 22/08/2011 11:39
Marcelo Escobar Slide 23/38
quit