Global Optimization Problems and Domain Reduction Strategies
Global Optimization Problems and Domain Reduction Strategies
A (2010) 125:123–137
DOI 10.1007/s10107-008-0263-4
Received: 21 May 2007 / Accepted: 25 November 2008 / Published online: 20 January 2009
© Springer-Verlag 2009
1 Introduction
A. Caprara
DEIS, Università di Bologna, Viale Risorgimento 2,
40136 Bologna, Italy
e-mail: [email protected]
M. Locatelli (B)
Dipartimento di Informatica, Università di Torino,
Corso Svizzera 185, 10149 Turin, Italy
e-mail: [email protected]
123
124 A. Caprara, M. Locatelli
is the box defined by the lower and upper bounds of the variables. These bounds might
be either explicitly given in the definition of the problem or can be derived by mini-
mizing and maximizing each variable in turn over the original feasible region, solving
the associated convex problem. In any case, we will assume throughout the paper that,
for each k ∈ N and t ∈ [lk , u k ], there exists (x1 , . . . , xk−1 , t, xk+1 , . . . , xn ) ∈ F,
observing that the bounds on xk can be tightened if this is not the case. For the sake
of simplicity, in the paper we will employ the following notation:
• x −k = (x1 , . . . , xk−1 , xk+1 , . . . , xn ) ∈ Rn−1 ;
• x (k) (t) = (x1 , . . . , xk−1 , t, xk+1 , . . . , xn ) ∈ Rn .
These problems are difficult ones. Even maximizing a quadratic function over the
unit simplex
⎧ ⎫
⎨ ⎬
= x ∈ Rn : x j = 1, 0 ≤ xi ≤ 1, i ∈ N ,
⎩ ⎭
j∈N
turns out to be strongly NP-hard (this follows from the continuous reformulation of the
MAX-CLIQUE problem proposed in [11]). We refer, e.g. to [4,19] for more complexity
results.
In spite of their difficulty, these global optimization problems often have enough
structure to be solved by deterministic approaches returning an optimal or, at least,
approximate solution in a finite time (we refer to [7,8,12] for a review of such
approaches and we mention software packages like BARON [15,17] that can solve
problems of this kind).
Among the solution approaches a prominent role is played by branch-and-bound
ones. These are based on: (a) branching, which is the successive subdivision of the
original problem (1) into smaller and smaller subproblems, generally corresponding
to the subdivision of the original box B into smaller and smaller subboxes; (b) the
computation of a lower bound, denoted in what follows with ∗ , on the optimal value
of (1) (the best, i.e. highest, function value observed within the feasible region); (c)
the computation of upper bounds for the subproblems into which the original one is
subdivided.
We do not give here the details of branching and lower bounding, for which we
refer, e.g. to [8], but we briefly discuss the upper bounding procedure. We restrict our
attention to the computation of an upper bound for the original problem (1) because
upper bounds over subboxes are obtained in a completely analogous way.
First, we need an overestimator of f over the box B, i.e. a function g such that
123
Global optimization problems and domain reduction strategies 125
The results of this paper apply to general overestimators, although the overestimators
of interest are generally concave functions, as discussed below. Note that it would be
enough to have an overestimator over the feasible region B ∩ F but, except in special
cases, this is much more difficult to obtain. As a result, the specific structure of F will
not play a relevant role in the strategies that we will discuss.
Remark 1 Function g usually depends on the box B, i.e. on the lower and upper bounds
of all variables.
Once we have an overestimator g, an upper bound for (1) is computed by solving the
following problem
Domain reduction strategies aim at reducing the current box B into a smaller one
B̃ ⊆ B. They can be classified into two broad classes:
• feasibility-based domain reductions, which reduce B without losing feasible solu-
tions, i.e.
B ∩ F = B̃ ∩ F;
B̃ ∩ {x ∈ F : f (x) ≥ ∗ } = B ∩ {x ∈ F : f (x) ≥ ∗ }.
123
126 A. Caprara, M. Locatelli
• problem dependent;
• general purpose.
The former exploit the special structure of the problem at hand and cannot be gen-
eralized to other problems, while the latter are applicable to all the problems we are
discussing.
Many domain reduction strategies have been proposed in the literature (see, e.g.
[5,6,10,14,21]). We also recall that in [18] a theory of domain reduction tools is
developed and many earlier results are re-derived from it.
The importance of domain reduction strategies is widely recognized in the liter-
ature. For instance, in [16, p. 23] it is stated that a preprocessing of the problem in
order to tighten the lower and upper bounds on the original problem variables can
significantly enhance algorithmic performance. This fact has also been confirmed by
personal experiences of the authors (see [2,9]).
What we will show in this paper is that the iterated version of a well known (prob-
ably the best known) optimality-based, general purpose domain reduction strategy
is equivalent to a strategy based on parameterizing the problem with respect to the
single variable whose domain we want to reduce. In this way, we will provide a new
interpretation of this widely employed reduction tool.
As a side result of this interpretation we will also observe that, for some subclasses
of problems with a special structure, parametric analysis allows one to perform the
domain reduction in a different way with respect to the iterated version (although the
final result is the same). In particular, we will observe that, while the iterated version is
just convergent to the final result after infinitely many iterations, parametric analysis
may return such a result after a finite time.
The paper is organized as follows. In Sect. 3 we will present the well known opti-
mality-based, general purpose domain reduction strategy together with its iterated
version, whereas in Sect. 4 we will introduce the new domain reduction strategy. In
Sect. 5 we will prove, under mild assumptions, the equivalence between these two
strategies, which, as pointed out above, allows a new interpretation of the former and
a different way to apply it. These mild assumptions are shown to be verified when the
overestimators used are the concave envelopes in Sect. 6. Finally, in Sect. 7 we discuss
the (in)dependency of the results on the order in which the (domains of the) variables
are processed, a fact that, to the authors’ knowledge, has not been explicitly stated in
the literature.
If a lower bound ∗ on the optimal value of (1) is available, then the best we can do
in order to reduce the domain of variable xk in order to exclude feasible points with
objective function lower than ∗ is to solve the two problems
min/max xk : f (x) ≥ ∗ , x ∈ F ∩ B .
123
Global optimization problems and domain reduction strategies 127
Unfortunately, these problems can be as difficult as the original problem (1); in partic-
ular they are nonconvex if f is not a concave function. What we can do is to substitute
f with its overestimator g. This leads to a quite popular domain reduction strategy,
based on moving the objective function of (3) into the constraints by imposing it does
not fall below ∗ , and then minimizing/maximizing in turn each variable xk . Formally,
the following two programs are solved for a given variable xk :
min/max xk : g(x) ≥ ∗ , x ∈ F ∩ B . (4)
In case g is concave, these are two convex programs. We call this standard domain
reduction (SDR) and denote the corresponding new lower and upper bound for xk
respectively by lkS D R and u kS D R .
An obvious way to improve the results of SDR is by iterating it (we stress that in
the procedure below both k and ∗ are fixed).
Iterated standard domain reduction (ISDR)
Step 0 Set lk0 = lk , u 0k = u k and i = 0.
Step 1 Solve the two programs
min/max xk : gi (x) ≥ ∗ , x ∈ F ∩ Bi , (5)
where
123
128 A. Caprara, M. Locatelli
one the dimension of the problem and eliminating all the nonlinearities due to variable
xk . The latter fact justifies the name nonlinearities removal domain reduction (NRDR)
of this domain reduction. By fixing xk = t we are led to the following problem:
where
is the (closed and nonempty) convex set given by the projection of the intersection
of F with the hyperplane {x ∈ Rn : xk = t} on the subspace of the variables
x1 , . . . , xk−1 , xk+1 , . . . , xn , and
f kt (x −k ) = f (x (k) (t))
Once function h k is available along with a lower bound ∗ on the optimal value of
(1), we compute the values
and
u NRDR
k = sup{t : h k (t) ≥ ∗ , t ∈ [lk , u k ]}. (11)
gkt (x −k ) ≤ g(x).
123
Global optimization problems and domain reduction strategies 129
The assumption simply states that, after fixing the value of variable xk to t, we are able
to find an overestimator gkt of f which is better (i.e. smaller or, at least, not larger)
than g.
Under this assumption we can prove that NRDR dominates SDR.
Observation 1 Under Assumption 1, it holds that
[lkNRDR , u NRDR
k ] ⊆ [lkSDR , u SDR
k ].
Proof We just prove that lkSDR ≤ lkNRDR . The other inequality is proved in a completely
analogous way. We notice that proving lkSDR ≤ lkNRDR is implied by proving that
For all feasible solutions of (1) with xk < lkSDR it holds, by definition, that g(x) < ∗ .
Moreover, also in view of Assumption 1, it holds that
which combined with g(x) < ∗ gives (12) and then the desired result.
Actually, strict dominance holds, as it can be seen through simple examples. The dom-
inance of NRDR with respect to SDR can also be extended to ISDR, if the following
extension of Assumption 1 holds
Assumption 2 For each i, for each t ∈ [lki , u ik ], and for each x ∈ B ∩ F with xk = t,
it holds that
gkt (x −k ) ≤ gi (x).
[lkNRDR , u NRDR
k ] ⊆ [lkISDR , u ISDR
k ].
But the above result can actually be strengthened: under mild assumptions we will
prove in the next section that equality holds. We will also show in Sect. 6 that these
assumptions are verified when gi and gki are the concave envelopes of f and f ki in the
respective domains.
This result is of theoretical interest because it allows a new interpretation of ISDR.
Indeed, it states the non obvious fact that iterating the standard domain reduction has
the same effect as removing all the nonlinearities involving variable xk in (1) and study-
ing the resulting relaxation. Moreover, the result also shows that we can substitute the
iterative solutions of problems performed by ISDR with the direct analysis of function
123
130 A. Caprara, M. Locatelli
Recall that at iteration i of the ISDR procedure we need to solve (5) to compute lki+1
and u i+1
k . We introduce the following two, quite reasonable, assumptions.
Assumption 3 The functions gkt are continuous both with respect to the variables
x1 , . . . , xk−1 , xk+1 , . . . , xn and with respect to the parameter t ∈ [lk , u k ].
Assumption 3 immediately implies the following continuity result for h k , whose proof
is reported below in order to make the paper self contained, but which could also be
derived from other results already existing in the literature [3,13,20].
Observation 3 Under Assumption 3, function h k is continuous over [lk , u k ].
Proof Given ε > 0 and t ∈ [lk , u k ], we have to prove that there exists δ > 0 such that
|h k (t) − h k (t)| ≤ ε
|gku (y −k ) − gkv (z −k )| ≤ ε
Note that the convexity and closedness of F guarantees the existence of such a
δ > 0. The proof now follows from
h k (t) − h k (t) = g t (x −k ) − g t (x −k ) ≤ g t (z −k ) + ε − g t (x −k ) ≤ ε,
123
Global optimization problems and domain reduction strategies 131
and
h k (t) − h k (t) = g t (x −k ) − g t (x −k ) ≤ g t (y −k ) + ε − g t (x −k ) ≤ ε.
Assumption 4 There exists some η∗ > 0 such that for each i, for each t ∈ [lki , u ik ],
and for each x −k ∈ B −k ∩ Fkt
Assumption 4 simply states that the gap between gi and gkt can not grow too fast
as we move away from the borders of the current domain [lki , u ik ] for xk . In partic-
li ui
ular, the assumption implies gkk (x −k ) = gi (x (k) (lki )) and gk k (x −k ) = gi (x (k) (u ik ))
for x −k ∈ B −k . Note that Assumptions 2 and 4 together give rise to the following
bracketing property
Under the above assumptions we can state the announced equivalence result.
Theorem 1 Under Assumptions 2, 3 and 4, it holds that
[lkNRDR , u NRDR
k ] ≡ [lkI S D R , u ISDR
k ].
Proof We just prove that lkNRDR = lkISDR . The proof that u NRDR
k = u ISDR
k is completely
analogous. Let us assume, by contradiction, that lk ISDR < lk NRDR . Then, in view of the
continuity of h k , established in Observation 3, and in view of the definition (10) of
lkNRDR , we must have
h k (lki ) ≤ ∗ − ρk ,
for some ρk > 0 small enough and for each i. In view of Assumption 4 we have that
for each x ∈ B ∩ F with xk = t. By definition we also have that there exists a feasible
(in fact, optimal) solution (x1 , . . . , xn ) of problem (5) with xk = lki+1 . Then
from which
ρk
lki+1 − lki ≥ > 0,
η∗
123
132 A. Caprara, M. Locatelli
In this section we prove that Assumptions 2, 3 and 4 hold if the original objective func-
tion f is Lipschitzian and the overestimators considered are the concave envelopes,
i.e. if, for each i,
Proof We first prove continuity of the gkt functions with respect to parameter t. Con-
sider t¯ ∈ [lk , u k ] and the associated gkt¯ . We show that for any ε > 0 and for any
t ∈ [lk , u k ] such that |t − t¯| ≤ ε/L, we have
gkt (x −k ) ≤ gkt¯ (x −k ) + ε.
Analogously,
implies
gkt¯ (x −k ) ≤ gkt (x −k ) + ε,
123
Global optimization problems and domain reduction strategies 133
Let gklk be the concave envelope of f klk over B −k , for which gklk (x −k ) ≤ g(x (k) (lk ))
since the map x −k → g(x (k) (lk )) is concave over B −k . We have, for each x ∈ B,
Proof First we note that, in view of Assumption 5, for each x ∈ Bi it holds that
In view of the same assumption it also holds that for each fixed t ∈ [lki , u ik ]
123
134 A. Caprara, M. Locatelli
[lkNRDR , u NRDR
k ] ≡ [lkISDR , u ISDR
k ].
In [1] three special classes of problems are considered for which Assumption 5 holds
and for which we can compute gi and gkt as in (13), (14), implying that the equivalence
result stated in the previous section holds. For each of these classes, function h k is
provided in [1].
A common practice is to fix some order P of the variables, to reduce the domain of
each variable in the given order, and to iteratively repeat this procedure until there are
reductions (in practical implementations, large enough reductions) from one iteration
to the next. The procedure is the following.
123
Global optimization problems and domain reduction strategies 135
[lii+1
k
, u ii+1
k
] = domain reduction(xik ; [lii+1
j
, u ii+1
j
],
j = 1, . . . , k − 1, [lii j , u ii j ], j = k, . . . , n).
Step 2 If l i+1
j = l ij and and u i+1
j = u ij for j ∈ N then stop. Otherwise, set i = i + 1
and repeat Step 1.
[l˜j , ũ j ] ⊆ [l j , u j ], j ∈ N ,
Note that the domain reduction procedures of the previous sections are monotone, as
well as all reasonable domain reductions we can think of.
Proposition 1 Under Assumption 6, the domains computed by Multiple domain
reduction satisfy:
l ij → l¯j and u ij → ū j as i → ∞, j ∈ N ,
and the limits l¯j and ū j do not depend on the order P of the variables.
123
136 A. Caprara, M. Locatelli
Therefore, after n iterations of the iterative procedure with variable ordering P
, the
domain of each variable is at least as tight as its domain after one iteration with variable
ordering P, i.e.
[l nj (P ), u nj (P )] ⊆ [l 1j (P), u 1j (P)], j ∈ N .
Next we can iterate the proof to show that for each ν = 1, 2, . . ., after at most νn
iterations with variable ordering P
, the domain of each variable is at least as good
as its domain after ν iterations with variable ordering P. Symmetrically we have that
after at most νn iterations with variable ordering P, the domain of each variable is at
least as good as its domain after ν iterations with variable ordering P
. This is only
possible if
8 Conclusions
In this paper we have discussed domain reduction strategies for nonconvex problems
over convex domains and box constraints. We proved that, under mild assumptions, the
iterated version of a well known domain reduction strategy is equivalent to a domain
reduction strategy based on fixing the value of the variable whose domain has to be
reduced, thus removing all the nonlinearities introduced by this variable. The result
is of theoretical interest because it allows a new interpretation of the results for the
iterated version of the standard domain reduction. Moreover, it also allows a different
way to compute such results by direct analysis of the function h k , on which the new
reduction strategy depends and whose definition is strictly related to the particular
problem at hand (see [1]).
Finally, we also proved that, under the natural monotonicity assumption, the result
of applying domain reduction strategies to all the variables in a given order and iterat-
ing until reductions are observed does not depend on the order in which the variables
are considered.
Acknowledgments We are grateful to Andrea Cassioli, Nick Sahinidis, Fabio Schoen and Fabio Tardella
for helpful discussions on the subject and to two anonymous referees for their comments. Michele Monaci
contributed to this work with some useful observations but still asked (or, better, pretended) not to be a
co-author; in absence of explicit laws that force people who deserve it to be included as co-authors we had
to remove his name eventually.
References
1. Caprara, A., Locatelli, M.: Domain reduction strategies: general theory and special cases. Technical
Report, Dipartimento di Informatica, Università di Torino (2008)
2. Caprara, A., Monaci, M.: Bidimensional packing by bilinear programming. Math. Program. (to appear)
3. Choquet, G.: Outlis topologiques et métriques de l’analyse mathématique. Technical Report, Centre
de documentation universitaire et SEDES, Paris (1969)
123
Global optimization problems and domain reduction strategies 137
4. de Klerk, E.: The complexity of optimizing over a simplex, hypercube or sphere: a short survey. CentER
Discussion Paper Series No. 2006-85 (2006)
5. Hamed, A.S.E., McCormick, G.P.: Calculation of bounds on variables satisfying nonlinear inequality
constraints. J. Global Optim. 3, 25–47 (1993)
6. Hansen, P., Jaumard, B., Lu, S.-H.: An analytical approach to global optimization. Math. Program.
52, 227–254 (1991)
7. Horst, R., Pardalos, P.M., Thoai, N.V.: Introduction to Global Optimization. Kluwer, Dordrecht (1995)
8. Horst, R., Tuy, H.: Global Optimization: Deterministic Approaches, 3rd edn. Springer, Berlin (1996)
9. Locatelli, M., Raber, U.: Packing equal circles into a square: a deterministic global optimization
approach. Discrete Appl. Math. 122, 139–166 (2002)
10. Maranas, C.D., Floudas, C.A.: Global optimization in generalized geometric programming. Comp.
Chem. Eng. 21, 351–370 (1997)
11. Motzkin, T.S., Straus, E.G.: Maxima for graphs and a new proof of a theorem of Turan. Can. J. Math.
17, 533–540 (1965)
12. Neumaier, A.: Complete search in continuous global optimization and constraint satisfaction. Acta
Numer. 13, 271–369 (2004)
13. Rockafellar, R.T.: Integrals which are convex functionals. Pacific J. Math. 39, 439–469 (1971)
14. Ryoo, H.S., Sahinidis, N.V.: A branch-and-reduce approach to global optimization. J. Global Optim.
8, 107–139 (1996)
15. Sahinidis, N.V.: BARON: A general purpose global optimization software package. J. Global Optim.
8, 201–205 (1996)
16. Sherali, H.D.: Tight relaxations for nonconvex optimization problems using the Reformulation-Lin-
earization/Convexification Technique (RLT). In: Pardalos, P.M., Romeijn, H.E. (eds.) Handbook of
global optimization, vol. 2, pp. 1–64. Kluwer, Dordrecht (2002)
17. Tawarmalani, M., Sahinidis, N.V.: BARON on the Web, http://archimedes.scs.uiuc.edu/cgi/run.pl
18. Tawarmalani, M., Sahinidis, N.V.: Global optimization of mixed-integer nonlinear programs: a theo-
retical and computational study. Math. Program. 99, 563–591 (2004)
19. Vavasis, S.A. Complexity issues in gobal optimization: a survey. In: Horst, R., Pardalos, P.M. (eds.)
Handbook of Global Optimization. Kluwer, Dordrecht (1995)
20. Wets, R.J.-B.: On inf-compact mathematical programs. Lect. Notes Comp. Sci. 5, 426–436 (1974)
21. Zamora, J.M., Grossmann, I.E.: A branch and contract algorithm for problems with concave univariate,
bilinear and linear fractional terms. J. Global Optim. 14, 217–249 (1999)
123