Math 1013366
Math 1013366
---
Formula:
For a function f(x), if x* is a local extremum (maximum or minimum), then the first derivative of
f(x) at x* must be zero:
f'(x*) = 0
Example:
Let's say we want to find the maximum or minimum of the function f(x) = x^2 - 4x + 3.
Now, check whether this is a maximum or minimum by examining the second derivative: f''(x) =
2 Since f''(x) > 0, x = 2 is a minimum.
---
Formula:
Example:
---
Formula:
For an optimization problem with a constraint, the method of Lagrange multipliers states that
we need to find the stationary points of the Lagrangian function:
where:
The optimal point occurs when the gradient of the Lagrangian is zero:
∇L(x, λ) = 0
Example:
Taking partial derivatives and setting them equal to zero: ∂L/∂x = 1 + 2λx = 0 ∂L/∂y = 1 + 2λy = 0
∂L/∂λ = x^2 + y^2 - 1 = 0
---
For a constrained optimization problem with inequality constraints, the KKT conditions include:
2. Dual feasibility: λ ≥ 0
Example:
Primal feasibility: x - 1 ≥ 0
Dual feasibility: λ ≥ 0
Stationarity: 2x + λ = 0
---
Formula:
If f(x) is a convex function and X is a convex feasible region, then any local minimum is also a
global minimum.
Example:
Since f(x) = x^2 is convex (the second derivative f''(x) = 2 > 0), and the feasible region is R, the
global minimum occurs at x = 0.
---
Formula:
For a convex optimization problem, the dual problem provides a lower bound to the primal
problem. Strong duality means that the optimal value of the dual problem is equal to the optimal
value of the primal problem.
Example:
For a linear programming problem, the duality theorem states that solving the dual problem
gives the same optimal value as solving the primal problem.
---
7. Subdifferential Theorem (Subgradient)
Formula:
For a non-smooth function f(x), the subdifferential at a point x is the set of all vectors g such
that:
Example:
For the function f(x) = |x|, the subdifferential at x = 0 is {-1, 1}, because the function has a sharp
corner at x = 0.
---
Formula:
The Frank-Wolfe algorithm updates the current solution by solving a linear approximation of the
objective function at each step:
x(k+1) = argmin { x ∈ X } ∇f(x(k))ᵀ(x - x(k))
Example:
For a convex problem like minimizing f(x) = x^2 over the interval [0, 1], the Frank-Wolfe method
iteratively moves toward the optimal solution by solving linearized versions of the function.
---
Formula:
The gradient descent algorithm updates the solution by moving in the direction of the negative
gradient:
Example:
---
Formula:
The No-Free-Lunch Theorem states that no optimization algorithm can outperform others
across all possible problems. There’s no universal best algorithm.
Example:
For simple problems, methods like gradient descent work well, but for non-smooth or discrete
problems, algorithms like genetic algorithms or simulated annealing may perform better.
---
This content is now in plain text and can be easily used in any text editor or converted to a PDF
using online LaTeX tools.