Engineering Optimization: Methods and Applications
Engineering Optimization: Methods and Applications
Engineering Optimization: Methods and Applications
Book Review
Page 1
Page 2
Constraints:
Good guys or bad guys?
Page 3
Constraints:
Good guys or bad guys?
reduces the region in which we search for optimum.
Page 4
Constraints:
Good guys or bad guys? makes optimization process very complicated
Page 5
( x 2)
x4
25 20 15
10
5 0
-2
-1
Page 6
Outline of Part 1
Equality-Constrained Problems Lagrange Multipliers Economic Interpretation of Lagrange Multipliers Kuhn-Tucker Conditions Kuhn-Tucker Theorem
Page 7
Outline of Part 1
Equality-Constrained Problems Lagrange Multipliers Economic Interpretation of Lagrange Multipliers Kuhn-Tucker Conditions Kuhn-Tucker Theorem
Page 8
Equality-Constrained Problems
GOAL solving the problem as an unconstrained problem by explicitly eliminating K independent variables using the equality constraints
Page 9
Example 5.1
Page 10
What if?
Page 11
Outline of Part 1
Equality-Constrained Problems Lagrange Multipliers Economic Interpretation of Lagrange Multipliers Kuhn-Tucker Conditions Kuhn-Tucker Theorem
Page 12
Lagrange Multipliers
Converting constrained problem to an unconstrained problem with help of certain unspecified parameters known as
Lagrange Multipliers
Page 13
Lagrange Multipliers
Lagrange function
Page 14
Lagrange Multipliers
Lagrange multiplier
Page 15
Example 5.2
Page 16
Page 17
positive definite
Page 18
Page 19
Example 5.3
Page 20
Page 21
Page 22
max
positive definite negative definite
Page 23
Outline of Part 1
Equality-Constrained Problems Lagrange Multipliers Economic Interpretation of Lagrange Multipliers Kuhn-Tucker Conditions Kuhn-Tucker Theorem
Page 24
The Lagrange multipliers have an important economic interpretation as shadow prices of the constraints, and their optimal values are very useful in sensitivity analysis.
Page 25
Outline of Part 1
Equality-Constrained Problems Lagrange Multipliers Economic Interpretation of Lagrange Multipliers Kuhn-Tucker Conditions Kuhn-Tucker Theorem
Page 26
Kuhn-Tucker Conditions
Page 27
NLP problem
Page 28
Page 29
Example 5.4
Page 30
Example 5.4
Page 31
Example 5.4
Page 32
Outline of Part 1
Equality-Constrained Problems Lagrange Multipliers Economic Interpretation of Lagrange Multipliers Kuhn-Tucker Conditions Kuhn-Tucker Theorem
Page 33
Kuhn-Tucker Theorems
Page 34
Let f, g, and h be differentiable functions x* be a feasible solution to the NLP problem. and for k=1,.,K are linearly independent
Page 35
Let f, g, andConstraint h be differentiable functions x* be a qualification feasible solution to the NLP problem.
and for k=1,.,K are linearly independent at the optimum If x* is an optimal solution to NLP problem, ! Hard to verify, since itthe requires that then there exists a (u*, v*) suchbe that (x*,u*, v*) the optimum solution known solves the KTP given by KTC.! beforehand
Page 36
For certain special NLP problems, the constraint qualification is satisfied: 1. When all the inequality and equality constraints are linear 2. When all the inequality constraints are concave functions and equality constraints are linear ! When the constraint qualification is not met at the optimum, there may not exist a solution to the KTP
Page 37
Example 5.5
x* = (1, 0)
Example 5.5
x* = (1, 0)
Page 39
not optimal
If it does not satisfy the KTCs
optimal
Example 5.6
Page 41
Let f(x) be convex the inequality constraints gj(x) for j=1,,J be all concave function the equality constraints hk(x) for k=1,,K be linear
If there exists a solution (x*,u*,v*) that satisfies KTCs, then x* is an optimal solution
Page 42
Example 5.4
f(x) be convex the inequality constraints gj(x) for j=1,,J be all concave function the equality constraints hk(x) for k=1,,K be linear
Page 43
Example 5.4
f(x) be convex
semi-definite
Page 44
Example 5.4
f(x) be convex v the inequality constraints gj(x) for j=1,,J be all concave function
Page 45
Example 5.4
f(x) be convex v the inequality constraints gj(x) for j=1,,J be all concave function the equality constraints hk(x) for k=1,,K be linear
Page 46
Remarks
For practical problems, the constraint qualification will generally hold. If the functions are differentiable, a KuhnTucker point is a possible candidate for the optimum. Hence, many of the NLP methods attempt to converge to a Kuhn Tucker point.
Page 47
Remarks
When the sufficiency conditions of Theorem 5.2 hold, a KuhnTucker point automatically becomes the global minimum. Unfortunately, the sufficiency conditions are difficult to verify, and often practical problems may not possess these nice properties. Note that the presence of one nonlinear equality constraint is enough to violate the assumptions of Theorem 5.2
Page 48
Remarks
The sufficiency conditions of Theorem 5.2 have been generalized further to nonconvex inequality constraints, nonconvex objectives, and nonlinear equality constraints. These use generalizations of convex functions such as quasi-convex and pseudoconvex functions
Page 49