Optimization Structure and Applications
Optimization Structure and Applications
VOLUME 32
Managing Editor
Panos M. Pardalos (University of Florida)
Editor–Combinatorial Optimization
Ding-Zhu Du (University of Texas at Dallas)
Advisory Board
J. Birge (University of Chicago)
C.A. Floudas (Princeton University)
F. Giannessi (University of Pisa)
H.D. Sherali (Virginia Polytechnic and State University)
T. Terlaky (McMaster University)
Y. Ye (Stanford University)
Edited By
CHARLES PEARCE
School of Mathematical Sciences,
The University of Adelaide,
Adelaide, Australia
EMMA HUNT
School of Economics & School of Mathematical Sciences,
The University of Adelaide,
Adelaide, Australia
123
Editors
Charles Pearce Emma Hunt
Department of Applied Mathematics Department of Applied Mathematics
University of Adelaide University of Adelaide
70 North Terrace 70 North Terrace
Adelaide SA 5005 Adelaide SA 5005
Australia Australia
[email protected] [email protected]
ISSN 1931-6828
ISBN 978-0-387-98095-9 e-ISBN 978-0-387-98096-6
DOI 10.1007/978-0-387-98096-6
Springer Dordrecht Heidelberg London New York
List of Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
vii
viii Contents
3.1 P (f0 , f1 ). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
3.2 L(x; 52 ). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53
3.3 L+
s 1 (x; 1). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53
3
4.1 A closed-loop control system. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
12.1 A scheme of the Z-channel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 236
12.2 Algorithm for finding independent set partitions . . . . . . . . . . . . 240
13.1 Illustration of the performance of our method. . . . . . . . . . . . . . 256
13.2 Typical examples of a column reconstruction in the matrix
X (image “Lena”) after filtering and compression of the
observed noisy image (Figure 13.1b) by transforms H
(line with circles) and T 0 (solid line) of the same rank.
In both subfigures, the plot of the column (solid line)
virtually coincides with the plot of the estimate by the
transform T 0 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 257
18.1 Density of (a) polypropylene and (b) SBS in elastomer
compounds for different blending methods. . . . . . . . . . . . . . . . . 347
18.2 Comparison of tensile strength of the compounds obtained
in an injection-molding machine and in a Brabender mixer. . . 348
18.3 Influence of wood flour fractions and the modifier on the
tensile strength of injection-molded specimens of the (a) PP
and (b) SBS compounds. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 349
18.4 Relative cost of the (a) PP and (b) SBS compounds
depending on the content of wood flour
and maleated polymers. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 351
18.5 Photographs of the PP compounds containing 40% wood
flour of different fractions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 352
19.1 A Constrained Steiner Tree and some of its special cases. . . . . 357
19.2 A CSPI instance reduced from SET COVER (not all edges
shown). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 361
xv
xvi List of Figures
xvii
xviii List of Tables
xix
xx Preface
1.1 Introduction
Jonathan Borwein
Centre for Experimental and Constructive Mathematics,
Simon Fraser University, Burnaby, BC, CANADA V5A 1S6
e-mail: [email protected]
Rafal Goebel
Center for Control Engineering and Computation ECE, University of California, Santa
Barbara, CA 93106-9650 U. S. A.
e-mail: [email protected]
were first collected by Chabrillac and Crouzeix [5] and include measurability,
almost everywhere continuity, and almost everywhere Fréchet differentia-
bility. Note that nondecreasing functions, whether on the real line or IRn ,
are cone monotone with respect to the nonnegative cone, either [0, +∞) or
[0, +∞)n .
Recently, Borwein, Burke and Lewis [2] showed that functions on a sepa-
rable Banach space, monotone with respect to a convex cone with nonempty
interior, are differentiable except at points of an appropriately understood
null set. The main goal of the current chapter is to demonstrate how possible
extensions of this result, or other generalizations of finite-dimensional results
on regularity of cone-monotone functions, fail in a general Banach space.
Motivation for studying coordinate-wise nondecreasing functions in
Chabrillac and Crouzeix, and cone-monotone functions by Borwein, Burke
and Lewis, comes in part from the connections of such functions with Lip-
schitz, and more generally, directionally Lipschitz functions. Interest in the
latter stems from the work of Burke, Lewis and Overton [4] on approximation
of the Clarke subdifferential using gradients, an important idea in practical
optimization. It turns out that such approximations, like in the Lipschitz
case, are possible in the more general directionally Lipschitz setting.
Before summarizing the properties of nondecreasing functions in finite di-
mensions, we illustrate their connection with Lipschitz functions. Consider a
Lipschitz function l : IRn → IR and a K > 0 satisfying
f (y) − f (x) =
z, y − x + l(y) − l(x)
n
≥K (yi − xi ) − max (yi − xi ) ≥ 0.
i=1,2,...,n
i=1
For details and proofs consult Chabrillac and Crouzeix [5]. Statements (c)
and (f ) generalize the Lebesgue monotone differentiability theorem and in
fact can be deduced from the two-dimensional version given by S. Saks [9];
for details consult Borwein, Burke and Lewis [2]. The Banach space version
of (c) and (f ) of Theorem 1, proved by Borwein, Burke and Lewis, requires a
notion of a null set in a Banach space. We recall that a Banach space does not
admit a Haar measure unless it is finite-dimensional, and proceed to make the
following definitions – for details on these, and other measure-related notions
we use in this chapter, we refer the reader to Benyamini and Lindenstrauss
[1].
Let X be a separable Banach space. A probability measure μ on X is
called Gaussian if for every x∗ ∈ X ∗ , the measure μx∗ on the real line,
defined by μx∗ (A) = μ{y |
x∗ , y ∈ A}, has a Gaussian distribution. It is
additionally called nondegenerate if for every x∗ = 0 the distribution μx∗ is
nondegenerate. A Borel set C ⊂ X is called Gauss null if μ(C) = 0 for every
nondegenerate Gaussian measure on X. It is known that the set of points
where a given Lipschitz function f : X → IR is not Gâteaux differentiable is
Gauss null. This in fact holds for functions with values in a space with the
Radon–Nikodym property (Benyamini and Lindenstrauss [1] Theorem 6.42),
whereas it fails completely for the stronger notion of Fréchet differentiability.
In what follows, we show that all the assumptions in the above theorem
are necessary, and, more generally, demonstrate how the properties of cone-
monotone functions described in Theorem 1 fail to extend to a general Banach
space. Note that the results of Theorems 1 and 2 hold if the functions are
allowed to take on infinite values, as long as appropriate meaning is given
to the continuity or differentiability of such functions. Indeed, composing a
possibly infinite-valued function with, for example, an inverse tangent does
not change its monotonicity properties, while leading to finite values. We do
not address this further, and work with finite-valued functions. Moreover, we
only work with monotone functions which are in fact nondecreasing (homo-
tone) and note that nonincreasing (antitone) functions can be treated in a
symmetric fashion.
6 J. Borwein and R. Goebel
1.2 Examples
φ, l0 >
φ, K − K; see Holmes [6], page 21. Now let L = ker φ, l = IR l0 .
Define Pl (x) to be the projection of x onto l – the unique point of l such
that x ∈ Pl (x) + L. Given any function g : l → IR, let
The preceding example demonstrated the failure of (c) and (e) of Theorem
1 in a general Banach space. We note that the cone K in this example is a
variation on one given by Klee [7], of a pointed convex cone dense in X
(pointed means that K ∩ −K = 0). Such a set is obtained by considering all
elements of X for which the last nonzero coordinate in a given Hamel basis
is positive.
In the example below, we use a Schauder basis to construct a cone-
monotone function violating (c) and (e) of Theorem 1 similarly to Example
2, however continuous at every point in a dense set of directions.
We check that f is K-monotone. The only way this could fail is if for some
y ≥K x, f (y) = 0 and f (x) = 1. But f (y) = 0 states that y ∈ −K, y ≥K x
implies x = y − k for some k ∈ K, and since −K is a convex cone, x =
y + (−k) ∈ −K + (−K) = −K. Thus x ∈ −K and f (x) cannot equal 1. Thus
f is K-monotone. Additionally, as K is closed and convex, the function f is,
respectively, lower semicontinuous and quasiconvex (level sets {x | f (x) ≤
r} are closed and convex). Moreover, f is generically continuous by Fort’s
theorem (see Borwein, Fitzpatrick and Kenderov [3]).
However, for every x ∈ −K, there exists d ∈ X such that t → f (x +
td) is not continuous at t = 0. Indeed, suppose that this failed for some
x0 ∈ −K. Then for every d ∈ X there exists (d) > 0 so that |t| < (d)
implies f (x0 + td) = 0, and so x0 + td ∈ −K. Thus x0 is an absorbing point
of a closed convex set −K, and, as X is barelled, x0 ∈ int(−K). But the
latter set is empty.
To sum up, f is continuous on a dense set but it is not continuous (and
not differentiable) at any point of the given non Haar null set.
The closed convex set C ⊂ Y in the example above was chosen to be not
Haar null and has no interior. Such a set exists in any nonreflexive space,
and in fact can be chosen to contain a translate of every compact subset of
the space – see Benyamini and Lindenstrauss [1]. In c0 , the nonnegative cone
is such a set (this requires the cone to be not Haar null), whereas the Haar
null nonnegative cone in l1 is not. Still, in l1 , and in fact in any nonreflexive
space, a cone satisfying the mentioned conditions can be found.
Indeed, suppose the set C of Example 4 contains translates of all compact
subsets of Y . We show that the constructed cone K contains a translate of
every compact subset of X. Pick any compact D ⊂ X. Let g ∈ X ∗ be such
that Y = g −1 (1). Shift D by z1 so that mind∈D+z1 g(d) = 1, and moreover,
so that (D + z1 ) ∩ C = ∅. Pick any v ∈ (D + z1 ) ∩ C, and let E ⊂ Y be the
projection of D onto Y in the direction v. Then E is a compact subset of Y ,
and thus for some z2 ∈ ker g, E + z2 ⊂ C. Now note that E + z2 is exactly
the projection in the direction v onto Y of the set D + z1 + z2 , which implies
that the latter set is a subset of C + IR+ v. Now C + IR+ v ⊂ K, as C ⊂ K
and v ∈ C. In effect, K contains D + z1 + z2 .
We now address another question on regularity of cone-monotone func-
tions. Weak and strong notions of lower semicontinuity for convex functions
agree. One cannot expect the same to hold for monotone functions, as the
following example demonstrates.
x+
f (x) = .
x+ + (a − x)+
10 J. Borwein and R. Goebel
y + x+ x+
≥ + ≥ + ,
y + + (a − y)
+ x + (a − y)
+ x + (a − x)+
where the first inequality stems from the fact that for a fixed β ≥ 0, the
function α → α/(α + β) is nondecreasing. Thus f is monotone. Let {en }∞
n=1
be the standard unit vectors in X. Notice that for any fixed α > 0 and
large enough n, we have (x − αen )+ = x+ and (a − x + αen )+ =
max{ (a − x)+ , α}. In effect,
(x − αen )+
f (x − αen ) =
(x − αen )+ + (a − x + αen )+
x+
→ +
x + max{ (a − x)+ , α}
as n → ∞. Note that the last expression is less than f (x) whenever x+ > 0
and (a − x)+ < α. Similar analysis leads to
max{ x+ , α}
f (x + αen ) → ,
max{ x+ , α} + (a − x)+
with the limit greater than f (x) when (a − x)+ > 0 and x+ < α. For a
given α, the vectors αen converge weakly to 0. The constant α can be fixed
arbitrarily large, and thus the function f is not weakly lower semicontinuous
at any x with x+ > 0 (equivalently x ∈ −K), and not weakly upper
semicontinuous at any x with (a − x)+ > 0 (equivalently x ∈ a + K).
Consider any x with xn < 0 for all n. It is easy to verify that
f (x + th) − f (x)
lim =0
t→0 t h
for all h ∈ c00 , that is, for sequences h with finitely many nonzero entries (in
fact, the difference quotient is then 0 for all small enough t). As c00 is dense
in X, and f is Lipschitz continuous, f is Gâteaux differentiable at x, with
the derivative equal to 0 ∈ X ∗ . Similarly, f has Gâteaux derivative 0 ∈ X ∗
at every x such that xn > an for all n.
Theorem 2 of Burke, Borwein and Lewis states that functions monotone
with respect to a cone with interior are Gâteaux differentiable outside a Gauss
null set. In the example below, we show a failure of that conclusion for a cone
with empty interior, even in a Hilbert space.
1 On the nondifferentiability of cone-monotone functions in Banach spaces 11
We first recall that the nonnegative cone in c0 is not Haar null, and so
is not Gaussian null, whereas those in lp , 1 < p < ∞ are Haar null but not
Gauss null. To see that the nonnegative cone in l2 is not Gauss null, observe
for example that it contains the interval
J = {x ∈ l2 | 0 ≤ xn ≤ 1/8n , n = 1, 2, . . .}
and apply the fact that the closed convex hull of any norm-convergent to 0
sequence with dense span is not Gauss null (Theorem 6.24 in Benyamini and
Lindenstrauss). The non Gauss null interval J will be used in the example
below.
Example 6 (Holder continuity, lack of Gâteaux differentiability). We show
that the Holder continuous function
f (x) = x+ ,
−J = {x ∈ l2 | − 1/8n ≤ xn ≤ 0, n = 1, 2, . . .}.
Indeed, for any x ∈ −J, consider h with hn = 1/2n and tk = 1/2k . Then
√
f (x + tk h) − f (x) xk + tk hk −1/8k + 1/4k
≥ ≥ = 1 − 1/2k ,
tk tk 1/2k
∞
(x − ki )+
f (x) = .
i=1
2i
though monotone with respect to the nonnegative cone, is not Gâteaux dif-
ferentiable outside c0 , that is, at any x for which at least one of lim sup x+ n
and lim sup x− − −
n is positive. (Recall α = −α when α < 0 and α = 0 other-
wise.) Indeed, suppose that lim sup x+ n = α > 0, and choose a subsequence
nk so that lim xnk = α. Define h by hn2i = 1, hn2i+1 = −1, i = 1, 2, . . .,
and hn = 0 for n = nk , k = 1, 2, . . .. Notice that for t close enough to 0,
lim sup(x + th)+n = α + |t| (if t > 0 then (x + th)n2i = xn2i + t, i = 1, 2, . . .,
and these terms provide the lim sup(x + th)+ n ; if t < 0 one should consider
(x + th)n2i+1 = xn2i+1 − t). On the other hand, lim sup(x + th)− −
n = lim sup xn ,
and in effect, the limit of (f (x + th) − f (x))/t as t → 0 does not exist. The
case of lim sup x−
n = β > 0 can be treated in a symmetric fashion.
Borwein, Burke and Lewis [2] show that any directionally Lipschitz func-
tion decomposes (locally) into a linear function, and a cone-monotone one
(with respect to a cone with interior). Consequently, nondifferentiable
1 On the nondifferentiability of cone-monotone functions in Banach spaces 13
p(x + k) − p(x) ≥ − k ≥ −
e1 , k,
which translates to
p(x + k) +
e1 , x + k ≥ p(x) +
e1 , x,
Acknowledgments The first author’s research was partially supported by NSERC and
by the Canada Research Chair Programme. The second author performed this research at
the Centre for Experimental and Constructive Mathematics at Simon Fraser University
and at the Department of Mathematics at University of British Columbia.
References
Jean B. Lasserre
2.1 Introduction
Jean B. Lasserre
LAAS-CNRS, 7 Avenue du Colonel Roche, 31077 Toulouse Cédex 4, FRANCE
e-mail: [email protected]
version (defined below) of the integer program. Results for counting problems,
notably by Barvinok [4], Barvinok and Pommersheim [5], Khovanskii and
Pukhlikov [12], and in particular, Brion and Vergne’s counting formula [7],
will prove especially useful.
For this purpose, we will consider the four related problems P, Pd , I and
Id displayed in the diagram below, in which the integer program Pd appears
in the upper right corner.
Problem I (in which ds denotes the Lebesgue measure on the affine variety
{x ∈ Rn | Ax = b} that contains the convex polyhedron Ω(b)) is the inte-
gration version of the linear program P, whereas Problem Id is the counting
version of the (discrete) integer program Pd .
Why should these four problems help in analyzing Pd ? Because first, P
and I, as well as Pd and Id , are simply related, and in the same manner.
Next, as we will see, the nice and complete duality results available for P, I
and Id extend in a natural way to Pd .
2.1.1 Preliminaries
In fact, I and Id are the respective formal analogues in the algebra (+, ×) of
P and Pd in the algebra (⊕, ×), where in the latter, the addition a ⊕ b stands
for max(a, b); indeed, the “max” in P and Pd can be seen as an idempotent
integral (or Maslov integral) in this algebra (see, for example, Litvinov et al.
[17]). For a nice parallel between results in probability ((+, ×) algebra) and
optimization ((max, +) algebra), the reader is referred to Bacelli et al. [3,
Section 9].
Moreover, P and I, as well as Pd and Id , are simply related via
εf (b,c) = lim f(b, rc)1/r ; εfd (b,c) = lim fd (b, rc)1/r . (2.1)
r→∞ r→∞
2 Duality and a Farkas lemma for integer programs 17
has a natural analogue for integration, the Laplace transform, and thus the
inverse Laplace transform problem (that we call I∗ ) is the formal analogue of
P∗ and provides a nice duality for integration (although not usually presented
in these terms). Finally, the Z-transform is the obvious analogue for summa-
tion of the Laplace transform for integration. We will see that in the light of
recent results in counting problems, it is possible to establish a nice duality
for Id in the same vein as the duality for (continuous) integration and by
(2.2), it also provides a powerful tool for analyzing the integer program Pd .
(a) We first review the duality principles that are available for P, I and Id
and underline the parallels and connections between them. In particular, a
fundamental difference between the continuous and discrete cases is that in
the former, the data appear as coefficients of the dual variables whereas in the
latter, the same data appear as exponents of the dual variables. Consequently,
the (discrete) Z-transform has many more poles than the Laplace transform.
Whereas the Laplace transform has only real poles, the Z-transform has ad-
ditional complex poles associated with each real pole, which induces some
periodic behavior, a well-known phenomenon in number theory where the
Z-transform (or generating function) is a standard tool (see, for example, Io-
sevich [11], Mitrinovı́c et al. [18]). So, if the procedure of inverting the Laplace
transform or the Z-transform (that is, solving the dual problems I∗ and I∗d )
is basically of the same nature, that is, a complex integral, it is significantly
more complicated in the discrete case, due to the presence of these additional
complex poles.
(b) Then we use results from (a) to analyze the discrete optimization
problem Pd . Central to the analysis is Brion and Vergne’s inverse formula
[7] for counting problems. In particular, we provide a closed-form expression
for the optimal value fd (b, c) which highlights the special role played by the
so-called reduced costs of the linear program P and the complex poles of the
Z-transform associated with each basis of the linear program P. We also
show that each basis B of the linear program P provides exactly det(B)
18 J.B. Lasserre
It is well known that the standard duality for (2.5) is obtained from the
Legendre-Fenchel transform F (., c) : Rm → R of the value function f (b, c)
with respect to b, that is, here (as y → f (y, c) is concave)
λ → F (λ, c) := infm
λ, y − f (y, c), (2.7)
y∈R
P∗ → inf
λ, b − F (λ, c) = minm {b λ | A λ ≥ c}. (2.8)
λ∈Rm λ∈R
2 Duality and a Farkas lemma for integer programs 19
Similarly, the analogue for integration of the Fenchel transform is the two-
sided Laplace transform F(., c) : Cm → C of f(b, c), given by
λ → F(λ, c) := ε−λ,y f(y, c) dy. (2.9)
Rm
n
1
F(λ, c) = whenever Re(A λ − c) > 0, (2.10)
(A λ − c)k
k=1
(see for example [7, p. 798] or [13]). Thus F(λ, c) is well-defined provided
and f(b, c) can be computed by solving the inverse Laplace transform prob-
lem, which we call the (integration) dual problem I∗ of (2.12), that is,
γ+i∞
1
I∗ → f(b, c) := εb,λ F(λ, c) dλ
(2iπ)m γ−i∞
1 γ+i∞
εb,λ
= dλ, (2.12)
(2iπ)m γ−i∞
n
(A λ − c)k
k=1
One may compute f(b, c) directly using Cauchy residue techniques. That is,
one may compute the integral (2.12) by successive one-dimensional complex
integrals with respect to one variable λk at a time (for example starting
with λ1 , λ2 , . . .) and by repeated application of Cauchy’s Residue Theorem
20 J.B. Lasserre
[8]. This is possible because the integrand is a rational fraction, and after
application of Cauchy’s Residue Theorem at step k with respect to λk , the
ouput is still a rational fraction of the remaining variables λk+1 , . . . , λm . For
more details the reader is referred to Lasserre and Zeron [13]. It is not difficult
to see that the whole procedure is a summation of partial results, each of them
corresponding to a (multi-pole) vector λ ∈ Rm that annihilates m terms of
n products in the denominator of the integrand.
This is formalized in the nice formula of Brion and Vergne [7, Proposition
3.3 p. 820] that we describe below. For the interested reader, there are several
other nice closed-form formulae for f(b, c), notably by Barvinok [4], Barvinok
and Pommersheim [5], and Khovanskii and Pukhlikov [12].
The material in this section is taken from [7]. To explain the closed-form
formula of Brion and Vergne we need some notation.
Write the matrix A ∈ Rm×n as A = [A1 | . . . |An ] where Aj ∈ Rm denotes
the j-th column of A for all j = 1, . . . , n. With Δ := (A1 , . . . , An ) let C(Δ) ⊂
Rm be the closed convex cone generated by Δ. Let Λ ⊆ Zm be a lattice.
A subset σ of {1, . . . , n} is called a basis of Δ if the sequence {Aj }j∈σ is
a basis of Rm , and the set of bases of Δ is denoted by B(Δ). For σ ∈ B(Δ)
let C(σ) be the cone generated by {Aj }j∈σ . With any y ∈ C(Δ) associate
the intersection of all cones C(σ) which contain y. This defines a subdivision
of C(Δ) into polyhedral cones. The interiors of the maximal cones in this
subdivision are called chambers in Brion and Vergne [7]. For every y ∈ γ,
the convex polyhedron Ω(y) in (2.4) is simple. Next, for a chamber γ (whose
closure is denoted by γ), let B(Δ, γ) be the set of bases σ such that γ is
contained
in C(σ), and let μ(σ) denote the volume of the convex polytope
{ j∈σ tj Aj | 0 ≤ tj ≤ 1} (normalized so that m
vol(R /Λ) = 1). Observe that
for b ∈ γ and σ ∈ B(Δ, γ) we have b = j∈σ xj (σ)Aj for some xj (σ) ≥ 0.
Therefore the vector x(σ) ∈ Rn+ , with xj (σ) = 0 whenever j ∈ σ, is a vertex of
the polytope Ω(b). In linear programming terminology, the bases σ ∈ B(Δ, γ)
correspond to the feasible bases of the linear program P. Denote by V the
subspace {x ∈ Rn | Ax = 0}. Finally, given σ ∈ B(Δ), let π σ ∈ Rm be the
row vector that solves π σ Aj = cj for all j ∈ σ. A vector c ∈ Rn is said to
be regular if cj − π σ Aj = 0 for all σ ∈ B(Δ) and all j ∈ σ. Let c ∈ Rn
be regular with −c in the interior of the dual cone (Rn+ ∩ V )∗ (which is the
case if A u > c for some u ∈ Rm ). Then, with Λ = Zm , Brion and Vergne’s
formula [7, Proposition 3.3, p. 820] states that
εc,x(σ)
f(b, c) = σ
∀ b ∈ γ. (2.13)
μ(σ) k ∈σ (−ck + π Ak )
σ∈B(Δ,γ)
2 Duality and a Farkas lemma for integer programs 21
Thus f(b, c) is a weighted summation over the vertices of Ω(b) whereas f (b, c)
is a maximization over the vertices (or a summation with ⊕ ≡ max).
So, if c is replaced by rc and x(σ ∗ ) denotes the vertex of Ω(b) at which
c x is maximized, we obtain
⎡ ⎤ r1
⎢ εrc,x(σ)−x(σ ) ⎥ ∗
⎣ ⎦
n−m μ(σ) σ
x(σ):vertex of Ω(b)
r (−c k + π Ak )
k ∈σ
as indicated in (2.2).
with γr = rγ and we can see that (up to the constant (m − n) ln r) the loga-
rithm of the integrand is simply the well-known logarithmic barrier function
n
−1
λ → φμ (λ, b) = μ
b, λ − ln (A λ − c)j ,
j=1
22 J.B. Lasserre
with parameter μ := 1/r, of the dual problem P∗ (see for example den Hertog
[9]). This should not come as a surprise as a self-concordant barrier function
of a cone K ⊂ Rn is given by the logarithm of the Laplace transform
φK (x)−x,s
K∗
ε ds of its dual cone K ∗ (see for example Güler [10], Truong and
Tunçel [20]).
Thus, when r → ∞, minimizing the exponential logarithmic barrier func-
tion on its domain in Rm yields the same result as taking its residues.
2.2.6 Summary
Fenchel-duality Laplace-duality
f (b, c) := max c x
f (b, c) :=
εc x ds
Ax=b; x≥0 Ax=b; x≥0
F(λ, c) := ε−λ y f(y, c) dy
F (λ, c) := infm {λ y − f (y, c)}
y∈R Rm
1
=
n
(A λ − c)k
k=1
In the respective discrete analogues Pd and Id of (2.5) and (2.6) one replaces
the positive cone Rn+ by Nn (or Rn+ ∩ Zn ), that is, (2.5) becomes the integer
program
Pd : fd (b, c) := max {c x | Ax = b; x ∈ Nn } (2.15)
2 Duality and a Farkas lemma for integer programs 23
|z1A1k · · · zm
Amk
| (= |z Ak |) > εck ∀k = 1, . . . , n. (2.19)
24 J.B. Lasserre
F(λ, c) := ε−λ y f(y, c)dy Fd (z, c) := z −y fd (y, c)
Rm y∈Zm
n
1
n
1
= =
(A λ − c)k 1 − εck z −Ak
k=1 k=1
Observe that the dual problem I∗d in (2.20) is of the same nature as I∗ in
(2.12) because both reduce to computing a complex integral whose integrand
is a rational function. In particular, as I∗ , the problem I∗d can be solved by
Cauchy residue techniques (see for example [14]).
However, there is an important difference between I∗ and I∗d . Whereas the
data {Ajk } appears in I∗ as coefficients of the dual variables λk in F(λ, c),
it now appears as exponents of the dual variables zk in Fd (z, c). And an
immediate consequence of this fact is that the rational function Fd (., c) has
many more poles than F(., c) (by considering one variable at a time), and in
2 Duality and a Farkas lemma for integer programs 25
particular, many of them are complex, whereas F(., c) has only real poles. As
a result, the integration of Fd (z, c) is more complicated than that of F(λ, c),
which is reflected in the discrete (or periodic) Brion and Vergne formula
described below. However, we will see that the poles of Fd (z, c) are simply
related to those of F(λ, c).
εc x(σ)
fd (b, c) = Uσ (b, c) (2.21)
μ(σ)
σ∈B(Δ,γ)
εc x(σ)
fd (b, c) = Uσ (b, c), (2.22)
μ(σ)
x(σ): vertex of Ω(b)
where G(σ) := (⊕j∈σ ZAj )∗ /Λ∗ (where ∗ denotes the dual lattice); it is a
2iπb
finite abelian group of order μ(σ) and with (finitely many) characters ε
for all b ∈ Λ. In particular, writing Ak = j∈σ ujk Aj for all k ∈ σ,
ε2iπAk (g) = ε2iπ j∈σ ujk gj
k ∈ σ.
26 J.B. Lasserre
Moreover,
1 − ε−2iπAk (g)εck −π
σ
Ak
Vσ (g, c) = , (2.24)
k ∈σ
We are now in a position to see how I∗d provides some nice information about
the optimal value fd (b, c) of the discrete optimization problem Pd .
Remark 1. Of course, (2.25) is not easy to obtain but it shows that the optimal
value fd (b, c) of Pd is strongly related to the various complex poles of Fd (z, c).
It is also interesting to note the crucial role played by the reduced costs
ck − π σ Ak in linear programming. Indeed, from the proof of Theorem 1 the
optimal value fd (b, c) is the value of c x at some vertex x(σ) plus a sum of
certain reduced costs (see (2.50) and the form of the coefficients αj (σ, c)).
Thus, as for the LP problem P, the optimal value fd (b, c) of Pd can be found
by inspection of (certain sums of) reduced costs associated with each vertex
of Ω(b).
2 Duality and a Farkas lemma for integer programs 27
We next derive an asymptotic result that relates the respective optimal values
fd (b, c) and f (b, c) of Pd and P.
For a proof see Section 2.6.2. Thus, when b ∈ γ ∩ Λ is sufficiently large, say
b = tb0 with b0 ∈ Λ and t ∈ N, the “max” in (2.25) is attained at the unique
optimal basis σ ∗ of the LP (2.5) (see details in Section 2.6.2).
From Remark 1 it also follows that for sufficiently large t ∈ N, the optimal
∗
value fd (tb, c) is equal to f (tb, c) plus a certain sum of reduced costs ck −π σ Ak
∗ ∗
(with k ∈ σ ) with respect to the optimal basis σ .
Aσ θ ∈ Zm . (2.30)
Equivalently, θ belongs to (⊕j∈σ Aj Z)∗ , the dual lattice of ⊕j∈σ Aj Z.
Thus there is a one-to-one correspondence between the ρ(σ) solutions
{θ(k)} and the finite group G (σ) = (⊕j∈σ Aj Z)∗ /Zm , where G(σ) is a sub-
group of G (σ). Thus, with G(σ) = {g1 , . . . , gs } and s := μ(σ), we can write
(Aσ )−1 gk = θgk = θ(k), so that for every character ε2iπy of G(σ), y ∈ Λ, we
have
ε2iπy (g) = ε2iπy θg , y ∈ Λ, g ∈ G(σ) (2.31)
and
ε2iπAj (g) = ε2iπAj θg = 1, j ∈ σ. (2.32)
So, for every σ ∈ B(Δ), denote by {zg }g∈G(σ) these μ(σ) solutions of (2.28),
that is,
zg = ελ ε2iπθg ∈ Cm , g ∈ G(σ), (2.33)
with λ = (Aσ )−1 cσ , and where ελ ∈ Rm is the vector {ελi }m
i=1 .
So, in the linear program P we have a dual vector λ ∈ Rm associated with
each basis σ. In the integer program P, with each (same) basis σ there are now
associated μ(σ) “dual” (complex) vectors λ + 2iπθg , g ∈ G(σ). Hence, with a
basis σ in linear programming, the “dual variables” in integer programming
are obtained from (a), the corrresponding dual variables λ ∈ Rm in linear
programming, and (b), a periodic correction term 2iπθg ∈ Cm , g ∈ G(σ).
We next introduce what we call the vertex residue function.
Definition 1. Let b ∈ Λ and let c ∈ Rn be regular. Let σ ∈ B(Δ) be a
feasible basis of the linear program P and for every r ∈ N, let {zgr }g∈G(σ)
be as in (2.33), with rc in lieu of c, that is,
The vertex residue function associated with a basis σ of the linear program
P is the function Rσ (zg , .) : N → R defined by
1 b
zgr
r → Rσ (zg , r) := , (2.34)
μ(σ) −Ak rck
g∈G(σ) (1 − zgr ε )
j ∈σ
which is well defined because when c is regular, |zgr |Ak = εrck for all k ∈ σ.
The name vertex residue is now clear because in the integration (2.20),
Rσ (zg , r) is to be interpreted as a generalized Cauchy residue, with respect
to the μ(σ) “poles” {zgr } of the generating function Fd (z, rc).
Recall from Corollary 1 that when b ∈ γ ∩Λ is sufficiently large, say b = tb0
with b0 ∈ Λ and some large t ∈ N, the “max” in (2.25) is attained at the
unique optimal basis σ ∗ of the linear program P.
2 Duality and a Farkas lemma for integer programs 29
ln zg = λ∗ + 2iπθg
1 1
f (b, c) = lim ln Rσ∗ (|zg |, r) fd (b, c) = lim ln Rσ∗ (zg , r)
r→∞ r r→∞ r
The (continuous) Farkas lemma, which states that given A ∈ Rm×n and
b ∈ Rm ,
{x ∈ Rn | Ax = b, x ≥ 0} = ∅ ⇔ [A λ ≥ 0] ⇒ b λ ≥ 0, (2.37)
has no discrete analogue in an explicit form. For instance, the Gomory func-
tions used in Blair and Jeroslow [6] (see also Schrijver [19, Corollary 23.4b])
are implicitly and iteratively defined, and are not directly defined in terms of
the data A, b. On the other hand, for various characterizations of feasibility
of the linear diophantine equations Ax = b, where x ∈ Zn , the interested
reader is referred to Schrijver [19, Section 4].
Before proceeding to the general case when A ∈ Zm×n , we first consider
the case A ∈ Nm×n , where A (and thus b) has only nonnegative entries.
n
zb − 1 = Qj (z)(z Aj − 1), (2.38)
j=1
m
m
b∗ := bj − min Ajk . (2.39)
k
j=1 j=1
For a proof see Section 2.6.4. Hence Theorem 2 reduces the issue of existence
of a solution x ∈ Nn to a particular ideal membership problem, that is, Ax = b
has a solution x ∈ Nn if and only if the polynomial z b − 1 belongs to the
binomial ideal I =
z Aj − 1j=1,...,n ⊂ R[z1 , . . . , zm ] for some weights Qj with
nonnegative coefficients.
Interestingly, consider the ideal J ⊂ R[z1 , . . . , zm , y1 , . . . , yn ] generated by
the binomials z Aj − yj , j = 1, . . . , n, and let G be a Gröbner basis of J. Using
the algebraic approach described in Adams and Loustaunau [2, Section 2.8],
it is known that Ax = b has a solution x ∈ Nn if and only if the monomial
2 Duality and a Farkas lemma for integer programs 31
In this section we consider the general case where A ∈ Zm×n so that A may
have negative entries, and we assume that the convex polyhedron Ω := {x ∈
Rn+ | Ax = b} is compact.
The above arguments cannot be repeated because of the occurrence of
negative powers. However, let α ∈ Nn and β ∈ N be such that
jk := Ajk + αk ≥ 0,
A k = 1, . . . , n; bj := bj + β ≥ 0, (2.40)
or equivalently, ⎛ ⎞
n
+ ⎝β −
Ax αj xj ⎠ em = b,
j=1
n
and thus, as β ≥ ρ∗ (α) ≥ j=1 αj xj (see, for example, (2.41)), we see that
n
(x, u) with β − j=1 αj xj =: u ∈ N is a solution of (2.42). Conversely, let
and b, it
(x, u) ∈ Nn × N be a solution of (2.42). Using the definitions of A
then follows immediately that
n
n
Ax + em αj xj + uem = b + βem ; αj xj + u = β,
j=1 j=1
so that Ax = b. The system of linear equations (2.42) can be cast in the form
⎡ ⎤
A | em
x b
B = with B := ⎣ − − ⎦ , (2.43)
u β
α | 1
and as B only has entries in N, we are back to the case analyzed in Section
2.4.1.
for some real-valued polynomials {Qj }nj=0 in R[z1 , . . . , zm , y], all of which
have nonnegative coefficients.
The degree of the Qj in (2.44) is bounded by
⎡ ⎡ ⎤⎤
m m
(m + 1)β + bj − min ⎣m + 1, min ⎣(m + 1)αk + Ajk ⎦⎦ .
k=1,...,n
j=1 j=1
m
εb λ − 1 = Qj (eλ1 , . . . eλm , e− i λi
)(ε(A λ)j − 1). (2.45)
j=1
Therefore A λ ≥ 0 ⇒ ε(A λ)j − 1 ≥ 0 for all j = 1, . . . , n, and as the Qj have
nonnegative coefficients, we have eb λ − 1 ≥ 0, which in turn implies b λ ≥ 0.
Equivalently, evaluating the partial derivatives
n of both sides of (2.45) with
respect to λj , at the point λ = 0, yields bj = k=1 Ajk xk for all j = 1, . . . , n,
with xk := Qk (1, . . . , 1) ≥ 0. Thus Ax = b for some x ∈ Rn+ .
2.5 Conclusion
We have proposed what we think is a natural duality framework for the in-
teger program Pd . It essentially relies on the Z-transform of the associated
counting problem Id , for which the important Brion and Vergne inverse for-
mula appears to be an important tool for analyzing Pd . In particular, it
shows that the usual reduced costs in linear programming, combined with
the periodicities phenomena associated with the complex poles of Fd (z, c),
also play an essential role for analyzing Pd . Moreover, for the standard dual
34 J.B. Lasserre
2.6 Proofs
Next, from the expression of Vσ (b, c) in (2.24), and with rc in lieu of c, we see
that Vσ (g, rc) is a function of y := er , which in turn implies that Hσ (b, rc) is
also a function of εr , of the form
ε2iπb (g)
Hσ (b, rc) = (εr )c x(σ)
, (2.47)
j δj (σ, g, A) × (er )αj (σ,c)
g∈G(σ)
for finitely many coefficients {δj (σ, g, A), αj (σ, c)}. Note that the coefficients
αj (σ, c) are sums of some reduced costs ck − π σ Ak (with k ∈ σ). In addition,
the (complex) coefficients {δj (σ, g, A)} do not depend on b.
Let y := εr/q , where q is the l.c.m. of {μ(σ)}σ∈B(Δ,γ) . As q(ck −π σ Ak ) ∈ Z
for all k ∈ σ,
Pσb (y)
Hσ (b, rc) = y qc x(σ) × (2.48)
Qσb (y)
2 Duality and a Farkas lemma for integer programs 35
for some polynomials Pσb , Qσb ∈ R[y]. In view of (2.47), the degree of Pσb
and Qσb , which depends on b but not on the magnitude of b, is uniformly
bounded in b.
Therefore, as r → ∞,
Hσ (b, rc) ≈ (εr/q )qc x(σ)+deg(Pσb )−deg(Qσb ) , (2.49)
so that the limit in (2.46), which is given by max εc x(σ) lim Uσ (b, rc)1/r (as
σ r→∞
we have assumed unicity of the maximizer σ), is also
max εc x(σ)+(deg(Pσb )−deg(Qσb ))/q .
x(σ): vertex of Ω(b)
Proof. Let t ∈ N and note that f (tb, rc) = trf (b, c) = trc x∗ = trc x(σ ∗ ). As
in the proof of Theorem 1, and with tb in lieu of b, we have
⎡ % &t ⎤ r1
U (tb, rc) ε rc x(σ)
Uσ (tb, rc) ⎦
fd (tb, rc) = εtc x ⎣
1 ∗ σ∗
r +
μ(σ ∗ ) εrc x(σ∗ ) μ(σ)
vertex x(σ) =x∗
Observe that c x(σ ∗ )−c x(σ) > 0 whenever σ = σ ∗ because Ω(y) is simple
if y ∈ γ, and c is regular. Indeed, as x∗ is an optimal vertex of the LP problem
∗
P, the reduced costs ck − π σ Ak (k ∈ σ ∗ ) with respect to the optimal basis
σ ∗ are all nonpositive, and in fact, strictly negative because c is regular (see
Section 2.2.4). Therefore the term
Pσtb (y)
y −tqδσ
Qσtb (y)
vertex x(σ) =x∗
36 J.B. Lasserre
is negligible for t sufficiently large, when compared with Uσ∗ (tb, rc). This is
because the degrees of Pσtb and Qσtb depend on tb but not on the magnitude
of tb (see (2.47)–(2.48)), and they are uniformly bounded in tb. Hence taking
the limit as r → ∞ yields
% ∗
&1/r
fd (tb,c) εrtc x(σ ) ∗
ε = lim Uσ∗ (tb, rc) = εtc x(σ ) lim Uσ∗ (tb, rc)1/r ,
r→∞ μ(σ ∗ ) r→∞
Therefore
1 1 zgb
εc x(σ) Uσ∗ (b, c) =
∗
μ(σ ) μ(σ ∗ )
g∈G(σ ∗ )
(1 − zg−Ak εck )
= Rσ∗ (zg , 1),
and (2.35) follows from (2.25) because, with rc in lieu of c, zg becomes zgr =
∗
εrλ ε2iπθg (only the modulus changes).
∗
Next, as only the modulus of zg is involved in (2.36), we have |zgr | = εrλ
for all g ∈ G(σ ∗ ), so that
1 |zgr |b εrb λ
∗
= r(ck −Ak λ∗ ) )
,
μ(σ ∗ ) k ∈σ ∗ (1 − |zgr |
−A kε rc k)
k ∈σ ∗ (1 − ε
g∈G(σ ∗ )
and, as r → ∞ ,
∗
εrb λ ∗
r(ck −Ak λ∗ ) )
≈ εrb λ ,
k ∈σ ∗ (1 − ε
Proof. (ii) ⇒ (i). Assume that z b − 1 can be written as in (2.38) for some
polynomials {Qj } with
nonnegative coefficients {Qjα }, that is, {Qj(z) } =
α∈N m Q jα z α
= α∈N m Q jα z α1
1 · · · zmαm
, for finitely many nonzero (and
nonnegative) coefficients Qjα . Using the notation of Section 2.3, the function
fd (b, 0), which (as c = 0) counts the nonnegative integral solutions x ∈ Nn
to the equation Ax = b, is given by
1 z b−em
fd (b, 0) = · · · n −Ak )
dz,
j=1 (1 − z
(2πi)m
|z1 |=γ1 |zm |=γm
fd (b, 0) = B1 + B2 ,
with
1 z −em
B1 = ··· n −Ak )
dz
(2πi)m |z1 |=γ1 |zm |=γm j=1 (1 − z
and
1 z −em (z b − 1)
B2 := ··· n −Ak )
dz
(2πi)m |z1 |=γ1 |zm |=γm j=1 (1 − z
n
1 z Aj −em Qj (z)
= ··· −Ak )
dz
j=1
(2πi)m |z1 |=γ1 |zm |=γm k =j (1 − z
n
Qjα z Aj +α−em
= ··· −Ak )
dz.
j=1 α∈Nm
(2πi)m |z1 |=γ1 |zm |=γm k =j (1 − z
Qjα z Aj +α−em
Cjα := ··· −Ak )
dz
(2πi)m |z1 |=γ1 |zm |=γm k =j (1 − z
38 J.B. Lasserre
is equal to
and
' (
z Aj xj − 1 = (z Aj − 1) 1 + z Aj + · · · + z Aj (xj −1) , j = 1, . . . , n,
We immediately see that each Qj has all its coefficients nonnegative (and
even in {0, 1}).
Finally, the bound on the degree follows immediately from the proof for
(i) ⇒ (ii).
References
7. M. Brion and M. Vergne, Residue formulae, vector partition functions and lattice
points in rational polytopes, J. Amer. Math. Soc. 10 (1997), 797–833.
8. J. B. Conway, Functions of a Complex Variable I, 2nd ed. (Springer, New York, 1978).
9. D. den Hertog, Interior Point Approach to Linear, Quadratic and Convex Program-
ming (Kluwer Academic Publishers, Dordrecht, 1994).
10. O. Güler, Barrier functions in interior point methods, Math. Oper. Res. 21 (1996),
860–885.
11. A. Iosevich, Curvature, combinatorics, and the Fourier transform, Notices Amer.
Math. Soc. 48 (2001), 577–583.
12. A. Khovanskii and A. Pukhlikov, A Riemann-Roch theorem for integrals and sums of
quasipolynomials over virtual polytopes, St. Petersburg Math. J. 4 (1993), 789–812.
13. J. B. Lasserre and E. S. Zeron, A Laplace transform algorithm for the volume of a
convex polytope, JACM 48 (2001), 1126–1140.
14. J. B. Lasserre and E. S. Zeron, An alternative algorithm for counting integral points
in a convex polytope, Math. Oper. Res. 30 (2005), 597–614.
15. J. B. Lasserre, La valeur optimale des programmes entiers, C. R. Acad. Sci. Paris
Ser. I Math. 335 (2002), 863–866.
16. J. B. Lasserre, Generating functions and duality for integer programs, Discrete Optim.
1 (2004), 167–187.
17. G. L. Litvinov, V. P. Maslov and G. B. Shpiz, Linear functionals on idempotent
spaces: An algebraic approach, Dokl. Akad. Nauk. 58 (1998), 389–391.
18. D. S. Mitrinović, J. Sándor and B. Crstici, Handbook of Number Theory (Kluwer
Academic Publishers, Dordrecht, 1996).
19. A. Schrijver, Theory of Linear and Integer Programming (John Wiley & Sons, Chich-
ester, 1986).
20. V. A. Truong and L. Tunçel, Geometry of homogeneous convex cones, duality map-
ping, and optimal self-concordant barriers, Research report COOR #2002-15 (2002),
University of Waterloo, Waterloo, Canada.
21. L. A. Wolsey, Integer programming duality: Price functions and sensitivity analysis,
Math. Programming 20 (1981), 173–195.
Chapter 3
Some nonlinear Lagrange and penalty
functions for problems with a single
constraint
3.1 Introduction
where X ⊆ IRn and f0 (x), f1 (x) are real-valued, continuous functions. (We
shall assume that these functions are directionally differentiable in Section
3.4.) Note that a general mathematical programming problem:
J. S. Giri
School of Information Technology and Mathematical Sciences, University of Ballarat,
Victoria, AUSTRALIA
e-mail: [email protected]
A. M. Rubinov†
School of Information Technology and Mathematical Sciences, University of Ballarat,
Victoria, AUSTRALIA
e-mail: [email protected]
The function q(λ) = inf x∈X (f0 (x) + λf1 (x)) is called the dual function and
the problem
max q(λ) subject to λ > 0
is called the dual problem. The equality
is called the zero duality gap property. The number λ̄ > 0 such that
where p(u, v) = u + v. It has been shown in [4, 5] that for penalty func-
tions, increasing positively homogeneous (IPH) convolutions provide exact
penalization for a large class of objective functions. The question thus arises
“are there nonlinear convolution functions for which Lagrange multipliers
exist?” The most interesting example of a nonlinear IPH convolution func-
1/k
tion is the function sk (u, v) = uk + v k . These convolutions also of-
ten provide a smaller exact penalty parameter than does the traditional
linear convolution. (See Section 3.3 for the definition of an exact penalty
parameter.)
We will show in this chapter that for problems where a Lagrange multiplier
exists, an exact penalty parameter also exists, and the smallest exact penalty
parameter is equal to the smallest Lagrange multiplier.
We also show that whereas a generalized penalty function can often im-
prove the classical situation (for example, provide exact penalization with
a smaller parameter than that of the traditional function), this is not true
3 Some nonlinear Lagrange and penalty functions 43
3.2 Preliminaries
Let us present some results and definitions which we will make use of later in
this chapter. We will refer to the solution of the general problem, P (f0 , f1 ),
as M (f0 , f1 ).
We will also make use of the sets, X0 = {x ∈ X : f1 (x) ≤ 0} and X1 =
{x ∈ X : f1 (x) > 0}.
It will be convenient to talk about Increasing Positively Homogeneous
(IPH) functions. These are defined as functions which are increasing, that is,
if (δ, γ) ≥ (δ , γ ) then p(δ, γ) ≥ p(δ , γ ), and positively homogeneous of the
first degree, that is, p(α(δ, γ)) = αp(δ, γ), α > 0 .
We shall consider only continuous IPH functions defined on either the half-
plane {(u, v) : u ≥ 0} or on the quadrant IR2+ = {(u, v) ∈ IR2 : u ≥ 0, v ≥ 0}.
In the latter case we consider only IPH functions p : IR2++ → IR, which
possess the following properties:
We shall denote by P1 the class of all such functions. The simplest example
of functions from P1 is the following function sk (0 < k < +∞), defined on
IR2+ :
1
sk (u, v) = uk + v k k . (3.3)
2l + 1
If k = with k, l ∈ N then the function sk is well defined and IPH
2m + 1
on the half-plane {(u, v) : u ≥ 0}. (Here N is the set of positive integers.)
A perturbation function plays an important part in the study of extended
penalty functions and is defined on IR+ = {y ∈ IR : y ≥ 0} by
β(y) − β(0)
lim inf > −∞.
y→+0 yk
(Here d is a real number and df does not mean the differential of f .) The
dual function qp (d) with respect to p is defined by
L+
p (x, d) = p (f0 (x), df1 (x)), (x ∈ X, d ≥ 0),
+ +
Then
qp (d) = min(tp (d), rp+ (1, d)), (3.5)
where rp+ is defined by
(The function rp+ was introduced and studied in [4] and [5].)
If the restriction p+ of p on IR2+ belongs to P1 then rp+ (1, d) = qp+ (d) (see
([4, 5]), so
qp (d) = min(tp (d), qp+ (d)).
Note that the function tp is decreasing and
tp (d) ≤ tp (0) = M (f0 , f1 ), (d > 0).
The function qp+ (d) = rp+ (1, d) is increasing. It is known (see [4, 5]) that
qp+ (d) ≡ rp+ (1, d) ≤ lim rp (1, u) = M (f0 , f1 ). (3.7)
u→+∞
¯ + (x) ≥ df
Since p is an increasing function and df ¯ 1 (x) for all x ∈ X, we have
1
¯ = inf p+ (f0 (x), df
qp+ (d) ¯ + (x))
1
x∈X
¯ + (x))
= inf p(f0 (x), df 1
x∈X
¯ 1 (x)) = M (f0 , f1 ).
≥ inf p(f0 (x), df
x∈X
On the other hand, due to (3.7) we have qp+ (d) ≤ M (f0 , f1 ) for all d. Thus
¯ = M (f0 , f1 ), that is, d¯ is an exact penalty parameter of P (f0 , f1 ) with
qp+ (d)
respect to p+ .
Due to (3.5) we have
¯ rp+ (1, d))
min(tp (d), ¯ = M (f0 , f1 ).
Since tp (d) is decreasing and rp+ (1, d) is increasing, (3.9) implies the
equalities
tp (d) = M (f0 , f1 ), ¯
(0 ≤ d ≤ d),
rp+ (1, d) = M (f0 , f1 ), (d¯ ≤ d < +∞),
Note that for penalty parameters the following assertion (Apen ) holds:
a number which is greater than an exact penalty parameter is also an exact
penalty parameter.
The corresponding assertion, Alag :
a number, which is greater than a Lagrange multiplier is also a Lagrange
multiplier,
does not hold in general. Assume that a Lagrange multiplier exists. Then
according to Proposition 2 an exact penalty parameter also exists. It follows
from Remark 1 that (Alag ) holds if and only if Ds = [0, +∞), that is,
We now remove condition f1+ = f1 and consider very special IPH functions,
for which (3.10) holds without this condition. Namely, we consider a class P∗
of IPH functions defined on the half-plane IR2∗ = {(u, v) : u ≥ 0} such that
(3.10) holds for each problem (f0 , f1 ) ∈ CX .
The class P∗ consists of functions p : IR2∗ → IR, such that the restriction
of p on the cone IR2+ belongs to P1 and p(u, v) = u for (u, v) ∈ IR2∗ with
v ≤ 0.
It is clear that each p ∈ P∗ is positively homogeneous of the first degree. Let
us now describe some further properties of p. Let (u, v) ≥ (u , v ). Assuming
without loss of generality that v ≥ 0, v ≤ 0 we have
inf p(f0 (x), df1 (x)) = inf f0 (x) = M (f0 , f1 ) for all d ≥ 0.
x∈X0 x∈X0
In this section we consider problems P (f0 , f1 ) such that both f0 and f1 are
directionally differentiable functions defined on a set X ⊆ IRn . Recall that
a function f defined on X is called directionally differentiable at a point
x ∈ intX if for each z ∈ IRn there exists the derivative f (x, z) at the point
x in the direction z:
1
f (x, z) = lim (f (x + αz) − f (x)).
α→+0 α
Usually only directionally differentiable functions with a finite derivative are
considered. We also accept functions whose directional derivative can attain
the values ±∞. It is well known (see, for example, [1]) that the maximum of
two directionally differentiable functions is also directionally differentiable. In
particular the function f + is directionally differentiable, if f is directionally
differentiable. Let f (x) = 0. Then
Then the point x = 0 is a min-stationary point for f1 and f2 , but this point
is not min-stationary for f3 .
Proof. Indeed, for all z ∈ IRn and sufficiently small α > 0 we have
(1/α)(f (x + αu) − f (x)) ≥ 0. Thus the result follows.
Lsk (x, z; λ) = kf0 (x)k−1 (f0 ) (x, z) + λkf1 (x)k−1 (f1 ) (x, z). (3.12)
2) k = 1. Then
Lsk (x, z; λ) = (f0 ) (x, z). (3.14)
We have ⎧
⎨ +∞ if f1 (x, z) > 0,
A(z) = 0 if f1 (x, z) = 0,
⎩
−∞ if f1 (x, z) < 0.
Hence ⎧
⎨ +∞ if f1 (x, z) > 0,
Lsk (x, z; λ) = kf0 (x) k−1
(f0 ) (x, z) if f1 (x, z) = 0, (3.15)
⎩
−∞ if f1 (x, z) < 0.
Note that for problems P (f0 , f1 ) with (f0 , f1 ) ∈ CX a minimizer is located
on the boundary of the the set of feasible elements {x : f1 (x) ≤ 0}.
Proposition 2. Let k > 1. Let (f0 , f1 ) ∈ CX . Assume that the functions
f0 and f1 have finite directional derivatives at a point x̄ ∈ intX, which is a
minimizer of the problem P (f0 , f1 ). Assume that
(that is, x̄ is a not a min-stationary point for the function f0 over X). Then
the point x̄ is not a min-stationary point of the function Lk for each λ > 0.
Proof. Assume that x̄ is a min-stationary point of the function Lsk (x; λ) over
X. Then combining Proposition 1 and (3.13) we have
Since f0 (x̄) > 0 it follows that (f0 ) (x̄, z) ≥ 0 for all z, which contradicts
(3.16).
It follows from this proposition that the Lagrange multiplier with respect
to Lsk (k > 1) does not exist for a problem P (f0 , f1 ) if (3.16) holds. Condition
(3.16) means that the constraint f1 (x) ≤ 0 is essential, that is, a minimum
under this constraint does not remain a minimum without it.
Remark 2. Consider a problem P (f0 , f1 ) with (f0 , f1 ) ∈ CX . Then under
some mild assumptions there exists a number k > 1 such that the zero duality
50 J.S. Giri and A.M. Rubinov
gap property holds for the problem P (f0k , f1k ) with respect to the classical
Lagrange function (see [2]). This means that
that is, the zero duality gap property with respect to sk holds. It follows
from Proposition 2 that a Lagrange multiplier with respect to sk does not
exist. Hence there is no a Lagrange multiplier for P (f0 , f1 ) with respect to
the classical Lagrange function.
Remark 3. Let g(x) = f1+ (x). Then the penalty-type function for P (f0 , f1 )
with respect to sk coincides with the Lagrange-type function for P (f0 , g) with
respect to sk . Hence an exact penalty parameter with respect to this penalty
function does not exist if (3.16) holds.
Proposition 3. Let k < 1 and let (f0 , f1 ) ∈ CX . Assume that the functions
f0 and f1 have finite directional derivatives at a point x̄ ∈ intX, which is a
minimizer for the problem P (f0 , f1 ). Assume that
(that is, x̄ is not a min-stationary point for the function f0 over X). Then
the point x̄ is not a min-stationary point of the function Lsk for each λ > 0.
Proof. Assume that a min-stationary point exists. Then combining Proposi-
tion 1, (3.15) and (3.17) we get a contradiction.
It follows from this proposition that a Lagrange multiplier with respect
to Lsk , k < 1, does not exist if condition (3.17) holds. We now give the
simplest example, when (3.17) is valid. Let f1 be a differentiable function
and ∇f (x̄) = 0. Then (3.17) holds.
Consider now a more complicated and interesting example. Let f1 (x) =
maxi∈I gi (x), where gi are differentiable functions. Then f1 is a directionally
differentiable function and f (x̄, u) = maxi∈I(x̄) [∇gi (x), u], where I(x̄) = {i ∈
I : gi (x) = f1 (x)} and [x, y] stands for the inner product of vectors x and y.
Thus (3.17) holds in this case if and only if there exists a vector u such that
[∇gi (x̄), u] < 0 for all i ∈ I(x̄). To understand the essence of this result, let us
consider the following mathematical programming problem with m inequality
constraints:
We can present (3.18) as the problem P (f0 , f1 ) with f1 (x) = maxi∈I gi (x).
Recall the well-known Mangasarian–Fromovitz (MF) constraint qualification
3 Some nonlinear Lagrange and penalty functions 51
for (3.18) (see, for example, [3]): (MF) holds at a point x̄ if there exists a
vector u ∈ IRn such that [∇gi (x̄), u] < 0 for all i ∈ I such that gi (x̄) = 0. Thus
(3.17) for P (f, f1 ) is equivalent to (MF) constraint qualification for (3.18). In
other words, if (MF) constraint qualification holds then a Lagrange multiplier
for Lsk with k < 1 does not exist. (It is known that (MF) implies the existence
of a Lagrange multiplier with k = 1.)
Let (f, f1 ) ∈ CX , where f0 , f1 are functions with finite directional deriva-
tives. Let g = f1+ and x be a point such that f1 (x) = 0. Then g (x, z) =
max(f (x, z), 0) ≥ 0 for all z, hence (3.17) does not hold for the problem
P (f0 , g). This means that Proposition 3 could not be applied to a penalty
function for P (f, f1 ) with respect to sk .
Simple examples show that exact penalty parameters with respect to sk
with k < 1 can exist. We now present an example from [5]. We do not provide
any details. (These can be found in [5], Example 4.6.)
Example 2. Let 0 < b < c < a be real numbers and X = [0, c]. Let f (x) =
(a − x)2 , f1 (x) = x − b, so P (f, f1 ) coincides with the following problem:
Let k = 1. Then an exact penalty parameter exists and the least exact penalty
d¯1 is equal to 2(a − b). Let k = 1/2. Then an exact penalty parameter also
exists and the least exact penalty parameter d¯1/2 coincides with c − b. We
indicate the following two points:
1) d¯1 does not depend on the set X; d¯1/2 depends on this set.
2) d¯1 depends on the parameter a, that is on the turning point of the parabola;
d¯1/2 does not depend on this parameter.
3.5 Example
Consider the following one-dimensional optimization problem:
9x2 7x
min f0 (x) = x3 − + + 5, (3.19)
2 2
subject to f1 (x) = x − 2 ≤ 0, x ∈ X = [0, 4].
9x2 7x 1
Lsk (x, λ) = sk (f0 (x), λf1 (x)) = ((x3 − + + 5)k + λk (x − 2)k ) k ,
2 2
52 J.S. Giri and A.M. Rubinov
10
0 x
–6 –4 –2 0 2 4 6 8 10 12
2l + 1
recalling that k = .
2m + 1
Now consider
dL ∂L ∂L
= f0 (x̄) + λ f (x̄). (3.20)
dx̄ ∂f0 ∂f1 1
An easy calculation shows that
⎧
⎪
⎨−2,
5
k > 1,
dL
(x̄) = − + λ, k = 1,
5 (3.21)
dx ⎪
⎩ ∞,2 k < 1.
5
From this it is clear that an exact Lagrange multiplier λ̄ = 2 may exist
only for the case k = 1.
1
L+ k + k k
sk (x; λ) = (f0 + (λf1 ) )
2 1
((x3 − 9x2 + 7x2 + 5) + λ (x − 2) ) , for x ≥ 2,
k k k k
= 2
x3 − 9x2 + 7x2 + 5, for x ≤ 2.
3 Some nonlinear Lagrange and penalty functions 53
y
10
0 x
–8 –6 –4 –2 0 2 4 6 8 10 12
Fig. 3.3 L+
s 1 (x; 1).
3
From these results we have shown that whereas the adoption of extended
penalty functions of the form sk yields an improvement to the traditional
penalty function approach, this cannot be generalized to improve the La-
grange approach.
References
4.1 Introduction
Robert Wenczel
Department of Mathematics, Royal Melbourne University of Technology, Melbourne 3001,
AUSTRALIA
Andrew Eberhard
Department of Mathematics, Royal Melbourne University of Technology, Melbourne 3001,
AUSTRALIA
e-mail: [email protected]
Robin Hill
Department of Mathematics, Royal Melbourne University of Technology, Melbourne 3001,
AUSTRALIA
reasons for using error-sequences in the space l1 as the basic variable, and
using various measures of performance (see, for example, [15]) which includes
the l1 -norm (see, for example, [11]). The formulation of l1 -problems leads
to the computation of inf M f (and determination of optimal elements, if
any), where M is an affine subspace of a suitable product X of l1 with
itself. This subspace is generated by the YJBK (or Youla) parameterization
[24] for the set of all stabilizing feedback-controllers for a given linear, time-
invariant system. The objective f will typically be the l1 -norm, (see [6],
[11]). As is now standard in the literature on convex optimization ( [20],
[21]), we will use convex discontinuous extended-real-valued functions, of the
form f = · 1 + δC , where δC is the indicator function (identically zero on C,
identically +∞ elsewhere) of some closed convex set in l1 . This formalism is
very flexible, encompassing many problem formats, including the case of time-
domain-template constraints ([10, Chapter 14], [14]). These represent bounds
on signal size, of the form Bi ≤ ei ≤ Ai (for all i), where e = {ei }∞i=0 denotes
the error signal. Often there are also similar bounds on the control signal u. As
the Youla parameterization generates controllers having rational z-transform,
the variables in M should also be taken to be rational in this sense, if the
above infimum is to equal the performance limit [6] for physically realizable
controllers. (If this condition is relaxed, then inf M f provides merely a lower
bound for the physical performance limit.) The set M may be recharacterized
by a set of linear constraints and thus forms an affine subspace of X. The
approach in many of the above references is to evaluate inf M f by use of
classical Lagrangian-type duality theory for such minimization problems. The
common assumption is that the underlying space X is l1 , or a product thereof,
and that M is closed, forcing it to contain elements with non-rational z-
transform, counter to the “physical” model.
Consequently, inf M f may not coincide with the performance limit for
physically realizable (that is, rational) controllers. However, in the context
of most of the works cited above, such an equality is actually easily estab-
lished (as was first noted in [27]). Indeed, whenever it is assumed that C has
nonempty interior, on combining this with the density in M of the subset con-
sisting of its rational members [19], we may deduce (see Lemma 15 below)
that the rational members of C ∩ M are l1 -dense in C ∩ M . This yields the
claimed equality for any continuous objective function (such as the l1 -norm).
Use of the more modern results of conjugate duality permits the extension
of the above approach to a more general class of (−∞, +∞]-valued objective
functions f . They are applicable even when int C may vanish. In this case,
the question of whether inf M f equals the physical limit becomes nontrivial
(in contrast to the case when C has interior). Indeed, if inf M f is strictly less
than the physical limit, any result obtained by this use of duality is arguably
of questionable engineering significance.
It is therefore important to know when inf M f is precisely the performance
limit for physically realizable controllers, to ensure that results obtained via
the duality approaches described above are physically meaningful. Note that
4 Convergence of truncates in l1 optimal feedback control 57
this question is posed in the primal space, and may be analyzed purely in the
primal space. This question will be the concern of this chapter.
In this paper, we derive conditions on the system, and on the time-domain
template set C, that ensure inf M (f0 + δC ) = inf C∩M f0 is indeed the per-
formance limit for physically realizable controllers, for various convex lower-
semicontinuous performance measures f0 . Here we only treat a class of 1-
input/2-output problems (referred to as “two-block” in the control litera-
ture), although the success of these methods for this case strongly suggests
the possibility of a satisfactory extension to multivariable systems.
Existing results on this question (see, for example, [14, Theorem 5.4])
rely on Lagrangian duality theory, and thereby demand that the time-domain
template C has interior. Here, for the class of two-block problems treated, we
remove this interiority requirement. Our result will be obtained by demon-
strating the convergence of a sequence of truncated (primal) problems. More-
over, this procedure will allow the explicit calculation of convergence esti-
mates, unlike all prior works with the exception of [23]. (This latter paper
estimates bounds on the truncation length for a problem with H∞ -norm
constraints and uses state-space techniques, whereas our techniques are quite
distinct.) The approach followed in our chapter has two chief outcomes. First,
it validates the duality-based approach in an extended context, by ensuring
that the primal problem posed in the duality recipe truly represents the
limit-of-performance for realizable controllers. Secondly, it provides a com-
putational alternative to duality itself, by exhibiting a convergent sequence of
finite-dimensional primal approximations with explicit error estimates along
the sequence. (This contrasts with traditional “primal–dual” approximation
schemes, which generally do not yield explicit convergence rates.)
This will be achieved by the use of some recently developed tools in opti-
mization theory (the relevant results of which are catalogued in Section 4.5)—
namely, the notion of epi-distance (or Attouch–Wets) convergence for convex
sets and functions [2, 3, 4, 5]. Epi-distance convergence has the feature that
if fn converges to f in this sense, then subject to some mild conditions,
* *
* *
*inf fn − inf f * ≤ d(f, fn ),
X X
We let R stand for the extended reals [−∞, +∞]. For a Banach space X,
balls in X centered at 0 will be written as B(0, ρ) = { x ∈ X | x < ρ } and
B̄(0, ρ) = { x ∈ X | x ≤ ρ }. Corresponding balls in the dual space X ∗ will
be denoted B ∗ (0, ρ) and B̄ ∗ (0, ρ) respectively. The indicator function of a set
A ⊆ X will be denoted δA . We will use u.s.c. to denote upper-semicontinuity
and l.s.c. to denote lower-semicontinuity. Recall that a function f : X → R
is called proper if never equal to −∞ and not identically +∞, and proper
closed if it is also l.s.c. For a function f : X → R, the epigraph of f , denoted
epi f , is the set {(x, α) ∈ X × R | f (x) ≤ α}. The domain, denoted dom f,
is the set {x ∈ X | f (x) < +∞}. The (sub-)level set {x ∈ X | f (x) ≤ α}
(where α > inf X f ) will be given the abbreviation {f ≤ α}. For > 0, and
if inf X f is finite, -argmin f = {x ∈ X | f (x) ≤ inf X f + } is the set of -
approximate minimizers of f . Any product X×Y of normed spaces will always
be understood to be endowed with the box norm (x, y) = max{ x , y };
any balls in such product spaces will always be with respect to the box norm.
∞
∞the Banach space of all complex sequences a = {an }n=0
Here l1 (C) denotes
such that a 1 := n=0 |an | is finite;
l1 denotes the Banach space of all real sequences in l1 (C); and l∞ denotes
the Banach space of all real sequences a = {an }∞ n=0 such that a ∞ :=
supn |an | is finite.
4 Convergence of truncates in l1 optimal feedback control 59
Proof. We prove the last assertion only; the rest can be found in the cited
n − bn } ⊆ A − B be a bounded sequence and let λn ≥
reference. Let {a
∞
0 be such that n=1 λn = 1. Then, due to the assumed boundedness of
one of A or B, {an∞ } ⊆ A and {bn } ⊆ ∞B are both bounded, yielding the
∞
convergent
∞ sums ∞
n=1 λ n an ∈ A and n=1 λ n b n ∈ B. Thus n=1 λn (an −
bn ) = n=1 λn an − n=1 λn bn ∈ A − B.
Proof. Let ρ > inf A∩C · and let x̄ ∈ A ∩ C ∩ B(0, ρ). Then for x ∈ X, x =
λ(a−c) for some λ > 0, a ∈ A, c ∈ C since by assumption, cone (A−C) = X.
Then for any t ≥ 1 sufficiently large so that t−1 ( a + c ) + x̄ < ρ,
+ + , + , ,
1 1 1 1
x = tλ a+ 1− x̄ − c + 1 − x̄
t t t t
∈ tλ (A ∩ B(0, ρ) − C ∩ B(0, ρ)) ⊆ cone (A ∩ B(0, ρ) − C ∩ B(0, ρ)) .
and XO will, for simplicity, be taken to be the same space X. The system H
is said to be linear, if X is a linear space and H a linear operator thereon.
Our focus will be on SISO (single-input/single-output) linear discrete-time
systems. In this case, X will normally be contained in the space RZ of real-
valued sequences.
For n ∈ N define a time-shift τn on sequences in X by
If each τn commutes with H, then H is called shift- (or time-) invariant. Our
interest will be in linear time-invariant (or LTI) systems.
It is well known [12] that H is LTI if and only if it takes the form of
the convolution operator h∗ for some h ∈ X. This h is called the impulse-
response of H, since h = H(δ) = h∗δ, where δ is the (Dirac) delta-function in
continuous time, or is the unit pulse sequence (1, 0, 0, . . . ), in discrete time.
The discrete-time LTI system H = h∗ is causal if the support of h lies
in the positive time-axis N = {0, 1, 2, . . .}. The significance of this notion is
clarified after
observing the action of H on an input u, which takes the form
(Hu)n = k≤n hk un−k , so if h takes any nonzero values for negative time
then evaluation of the output at time n would require foreknowledge of input
behavior at later times.
Note that for LTI systems, a natural (but by no means the only) choice for
the “space of signals” X is the space of sequences for which the z-transform
exists. From now on we identify a LTI system H with its impulse-response
h := H(δ).
R∞ = l1 ∩ R(z) ,
a finite operating range, and it would not be desirable for, say, the control
signal generated by K to become too large and “blow up” the plant P .
Writing P = p∗ and K = k∗, with p̂ and k̂ in R(z) and noting that the
transfer function between any two points of the loop can be constructed by
addition or multiplication from the three transfer functions given in (4.1)
below, we obtain the following well-known result.
x̂n̂ + ŷ dˆ ≡ 1 in R(z) ,
This result has the following consequence for the stabilized closed-loop
mappings. Recall that 1/(1 + p̂ĉ) is the transfer function taking input w to e
(that is, ê = ŵ/(1 + p̂ĉ)) and ĉ/(1 + p̂ĉ) maps the input w to u. We now have
(see [24]) the following result.
64 R. Wenczel et al.
Corollary 2. The set of all closed-loop maps Φ taking w to (e, u), achieved
by some stabilizing compensator C ∈ S(P ), is
) + , + ,* .
*
Φ = dˆ ŷ − q̂ dˆ n̂ * q̂ ∈ R∞ , q̂ = ŷ/n̂ , (4.3)
x̂ dˆ *
where
c := {cos kθ}∞ ∞
k=0 and s := {sin kθ}k=0 .
We shall assume that the error signals (usually denoted by e here) are in l1 ,
so we are considering only those controllers that make the closed-loop system
track the input (in an l1 -sense), and we shall assume also that the associated
control signal remains bounded (and in fact resides in Xa ).
Definition 6. For any u ∈ Xa , let uc and us denote the (unique) real num-
bers such that u − uc c − us s is in l1 .
Let the plant P have transfer function P(z) with the coprime factorization
P(z) = n̂(z)/d(z)
ˆ where n̂ and dˆ are members of R[z] (the space of polynomi-
s
als, with real coefficients, in the complex variable z). Write n̂(z) = i=0 ni z i ,
ˆ = t di z i , n := (n0 , n1 , .., ns , 0, ..) and d := (d0 , .., dt , 0, ..).
d(z) i=0
Let x and y be finite-length sequences such that x̂n̂ + ŷ dˆ = 1 in R[z].
Their existence is a consequence of the coprimeness of n̂ and dˆ in R[z]
(see [24]).
If we aim to perform an optimization over a subset of the set S(P ) of
stabilizing controllers K for P , such that for each K in this subset, the
corresponding error signal φ(K) and control output u(K) are in l1 and Xa
respectively, the appropriate feasible set is
4 Convergence of truncates in l1 optimal feedback control 65
where the latter equality follows from the YJBK factorization for S(P ) (use
Property 3 or Corollary 2).
Henceforth we shall always assume that w ∈ Xa is rational (that is, has
rational z-transform), the plant P is rational and has no zero at a.
By the lemma to follow (whose proof is deferred to the Appendix) we ob-
serve that a single sinusoid suffices to characterize the asymptotic (or steady-
state) behavior of u for all feasible signal pairs (e, u) ∈ F0 .
Using this lemma, we can translate F0 in the u variable to obtain the set
F := F0 − (0, βc c + βs s), having the form (where we use the notation x ∗
(y, z) := (x ∗ y, x ∗ z) for sequences x, y, z)
{ u) ∈ l1 × l1 |
F =(e,
(e, u) = w ∗ d ∗ (y, x) − (0, βc c + βs s) − q ∗ w ∗ d ∗ (n, −d)
for some q such that q̂ ∈ R∞ \{ŷ/n̂}}.
M :=
⎧ * ⎫
⎪ * ê(p̄i ) = 0 ( p̄i pole of P : |p̄i | ≤ 1) i = 1,.., m1⎪
⎪ * ⎪
⎨ * ê(z̄ ) = ŵ(z̄j )(z̄j zero of P : |z̄j | ≤ 1) j = 1,.., m2⎬
(e, u) ∈ (l1 )2 ** j ,
⎪
⎪
⎩ * ê(v̄k ) = 0 (v̄k zero of ŵ : |v̄k | ≤ 1) k = 1,.., m3⎪ ⎪
⎭
d ∗ e*+ n ∗ u = w ∗ d − n ∗ (βc c + βs s)
with the understanding that in the above sets, whenever P and ŵ have
a common pole at a (and hence at ā), the constraint ê(a) = 0 is absent.
66 R. Wenczel et al.
Remark 1. We note the following relation between M (0) and M . Let (ē, ū) be
any element of M . Then M (0) − ē and M −(ē, ū) consist of elements satisfying
the corresponding constraints with right-hand sides set to zero. Assuming
that P (equivalently, n̂) has no zeros on the unit circle, and that all its zeros
in D are simple, then the map T on M (0) − ē taking e to −Z −1 (dê/n̂)
ˆ maps
into l (by Lemma 4 below) with T ≤ κ d 1 . Then (e, T e) ∈ M − (ē, ū),
1
The next two lemmas give simple sufficient conditions for the feasible
set F to be fully recharacterized as an affine subspace defined by a set of
linear constraints, and for when elements of finite length exist. Since they are
only minor modifications of standard arguments, we omit the proofs here,
relegating them to the Appendix.
Remark 2. In the above result, one can always choose that ē = 0, so that
(ē, ū) ∈ Mr , which coincides with F by Lemma 2. Thus the latter is nonempty.
Lemma 4. Let fˆ ∈ l1 and let p(·) be a polynomial with real coefficients with
no zeros on the unit circle. Assume also at each zero of p(·) in the open unit
4 Convergence of truncates in l1 optimal feedback control 67
and where in the above product, the zeros appear according to multiplicity,
and moreover, both members of any conjugate pair of zeros are included.
Proof. See the Appendix.
Remark 3. Such a result will not hold in general if p(·) has zeros on the unit
circle; a counterexample for |a| = 1 is as follows.
a−j
Let f0 = −1 and fj = j(j+1) for j ≥ 1. Then f ∈ l1 (C) and fˆ(a) = 0 since
∞ fˆ(z)
k=1 k(k+1) = 1. If q is the inverse z-transform of [z → z−a ], then for each
1
k,
* *
* 1 j*
|qk | = * ak+1 j>k j * = k+1
f a 1
so q ∈
/ l1 (C).
The hard bounds on the error and on input activity will be represented
by the closed convex sets C (e) and C (u) respectively, given by
(e) (e)
C (e) := {e ∈ l1 | Bi ≤ ei ≤ Ai for all i}, (4.4)
(e) (e)
where Bi ≤ 0 ≤ Ai eventually for large i, and
(u) (u)
C (u) := {u ∈ l1 | Bi ≤ ui ≤ Ai for all i}, where (4.5)
d(x, D) := inf x − d ,
d∈D
eρ (C, D) := sup d(x, D),
x∈C∩B(0,ρ)
haus ρ (C, D) := max{eρ (C, D), eρ (D, C)}, (the ρ-Hausdorff distance) .
lim dρ (fn , f ) = 0,
n→∞
where Δ(X) := {(x, x) | x ∈ X} and B(0, s)2 denotes a ball in the box norm
in X × X (that is, B(0, s)2 = B(0, s) × B(0, s)). Then for each ρ ≥ 2r + t
and all n ∈ N such that
we have
2r + s + ρ
dρ (fn + gn , f + g) ≤ [dρ+s (fn , f ) + dρ+s (gn , g)] .
s
In particular, fn + gn Attouch–Wets converges to f + g.
2r + s + ρ 2r + 2s + ρ
≤ dρ+s (fn , f ) + dρ+2s (Cn , C) .
s s
The estimates for dρ (ϕ, ϕn ) will now follow from two applications of Proposi-
tion 5. Indeed, from (4.9) and Proposition 5, whenever ρ ≥ 2r+t and n is such
that dρ+2s (Cn , C) < s, then dρ+s (Cn ∩M, C ∩M ) ≤ 2r+s+(ρ+s) s dρ+2s (Cn , C).
Taking n to be such that (4.13) holds, we find that dρ+s (fn , f ) + dρ+s (Cn ∩
M, C ∩ M ) < s, so from (4.10) and Proposition 5 again,
2r + s + ρ
dρ (ϕ, ϕn ) ≤ [dρ+s (fn , f ) + dρ+s (Cn ∩ M, C ∩ M )]
s
2r + s + ρ 2r + 2s + ρ
≤ dρ+s (fn , f ) + dρ+2s (Cn , C) .
s s
If we keep fn fixed in this process, then only one iteration of Proposition 5
is required, which will lead to a better coefficient for the rate of convergence
than that obtained from taking dρ (fn , f ) ≡ 0 in Corollary 3.
Then for any fixed ρ ≥ max{2r + t, ρ0 , α + 1}, and all n ≥ n0 for which
we have
* *
* *
* inf f − inf f * ≤ dρ (f + δC ∩M , f + δC∩M )
*Cn ∩M C∩M *
n
2r + s + ρ
≤ dρ+s (Cn , C) .
s
Proof. Similar to that of Corollary 3, but with only one appeal to
Proposition 5.
Our intention will be to apply the results of the preceding section with X =
l1 ×l1 (with the box norm), so the norm on X ×X will be the four-fold product
of the norm on l1 . From now on we will not notationally distinguish balls in
the various product spaces, the dimensionality being clear from context.
Before we can apply the convergence theory of Section 4.5, we need to check
that the constraint qualifications (4.9) and (4.10) (or (4.14)) are satisfied.
This will be the main concern in this section. First, we consider some more
readily verifiable sufficient conditions for (4.9) and (4.10) to hold.
Lemma 5. Suppose that for sets C and M we have
B(0, μ/2)2 ⊆ Δ(X) ∩ B(0, μ/2 + λ)2 − ({f ≤ ν} × (C ∩ M )) ∩ B(0, λ)2 (4.20)
and hence
from which (4.18) clearly follows (and is of the form (4.9) for suitable r, s
and t).
Next, suppose we have (4.19). If we define D := Δ(X) − ({f ≤ ν} × (C ∩
M ))∩B(0, λ)2 , then similarly P (D) = {f ≤ ν}∩B(0, λ)−(C ∩M )∩B(0, λ) ⊇
B(0, μ) so, proceeding as above, we obtain (4.20) (which is of the form (4.10)).
The last assertion follows from the first on substituting {f ≤ r} ∩ M for
M therein.
In this chapter, the objective functions f will always be chosen such that
(4.17) will imply (4.19) or (4.21). Hence in this section we focus on verification
of (4.17).
The constraints of M (0) have the form Ae = b, where m = 2m1 +2m2 +2m3
and
with A : l1 → Rm given by
4 Convergence of truncates in l1 optimal feedback control 73
Ae := (.., Re ê(z̄i ), Im ê(z̄i ),.., Re ê(p̄j ), Im ê(p̄j ),.., Re ê(v̄k ), Im ê(v̄k ),..)T ,
(4.23)
where z̄i , p̄j and v̄k denote distinct elements of the unit disk, and i, j and k
range over {1, . . . , m1 }, {1, . . . , m2 } and {1, . . . , m3 } respectively. Then A is
expressible as a matrix operator of the form
⎛ ⎞
1 Re z̄1 Re z̄12 · · ·
⎜ 0 Im z̄1 Im z̄12 · · · ⎟
⎜ ⎟
⎜ .. .. .. ⎟
⎜. . . ⎟
⎜ ⎟
⎜ 1 Re z̄m1 Re z̄m 2
· · · ⎟
⎜ 1 ⎟
⎜ 0 Im z̄m1 Im z̄m 2
· · · ⎟
⎜ 1 ⎟
⎜ 1 Re p̄1 Re p̄21 · · · ⎟
⎜ ⎟
⎜ 0 Im p̄1 Im p̄21 · · · ⎟
⎜ ⎟
⎜ ⎟
(aij )1≤i≤m; 0≤j<∞ = ⎜ ... ..
.
..
. ⎟,
⎜ ⎟
⎜ 1 Re p̄m Re p̄2 · · · ⎟
⎜ 2 m2 ⎟
⎜ 0 Im p̄m Im p̄2 · · · ⎟
⎜ 2 m2 ⎟
⎜ 1 Re v̄1 Re v̄ 2 · · · ⎟
⎜ 1 ⎟
⎜ 0 Im v̄1 Im v̄ 2 · · · ⎟
⎜ 1 ⎟
⎜. .. .. ⎟
⎜ .. . . ⎟
⎜ ⎟
⎝ 1 Re v̄m Re v̄ 2 · · · ⎠
3 m3
0 Im v̄m3 Im v̄m 2
3
···
where rows of imaginary parts of the matrix (aij ) and of b are omitted when-
ever the associated z̄i , p̄j or v̄k is real.
For integer K, define A(K) to be the truncated operator taking RK into
m
R given by the matrix
(aij )1≤i≤m,0≤j<K .
the norm being taken relative to the 1-norm on the range of the inverse
3 (m) 4−1
A and the ∞-norm on its domain. Note that αK satisfies
Proof. Place s := /αK and let η ∈ B(0, s). Set e := {ξ, 0, 0, ..} ∈ C (e) ∩M (0) .
As A(K) maps onto Rm , there exists p ∈ RK ⊆ l1 such that Ap = A(K) p =
Aη with
p + e 1 ≤ p 1 + e 1 < + ξ 1 ,
u 1 ≤ η 1 + T e 1
≤ η 1 + T ( ē 1 + ρ)
< s + T ( ē 1 + ρ) ≤ μ,
4 Convergence of truncates in l1 optimal feedback control 75
where the last inequality follows by Remark 1, and so u ∈ C (u) − ū from As-
sumption 7 and u+ ū ∈ C (u) ∩B(0, μ+ ū 1 ). Also, v + ē = v ∈ C (e) ∩B(0, σ)
and
Remark 4. The existence of (ē, ū) in M of finite length is ensured, for in-
(e) (e)
stance, under the conditions of Lemma 3. If the bounds Ai and Bi satisfy
(e) (e)
Bi < ēi < Ai for i ≤ l(ē) (where l(·) denotes length), then Condition 7
of Lemma 7 follows, for suitable constants, from Lemma 6. (Note again, that
(e) (e)
this holds for arbitrary Ai and Bi for i > l(ē), so by making these bounds
decay to zero sufficiently rapidly, as will be shown in Lemma 13 to come, we
can enforce compactness of C (e) , which will be essential for the Attouch–Wets
(e)
convergence of Cn to C (e) ).
(u) (u)
If, furthermore, the bounds Ai and Bi are chosen to envelop ū (∀i,
(u) (u)
Bi < ūi < Ai ), and to be bounded away from zero by sufficient distance,
for all i, Condition 7 of Lemma 7 is also satisfied.
since C (e) − ē and C (u) − ū are convex sets containing 0. Then (where T
denotes the mapping introduced in Remark 1)
Lemma 10. Suppose that C = C (e) × C (u) is locally compact in the sense of
Lemma 9 and n̂ has no zeros on the unit circle. Then
0∈
/ int (C − M ) .
χ = d∗ξ+n∗η
∈ d ∗ cone (C (e) ∩ B(0, ρ) − ē) + n ∗ cone (C (u) ∩ B(0, ρ) − ū)
−(d ∗ e + n ∗ T e)
= cone (d ∗ (C ∩ B(0, ρ) − ē)) + cone (n ∗ (C (u) ∩ B(0, ρ) − ū))
(e)
' (
⊆ cone d ∗ (C (e) ∩ B(0, ρ) − ē) + n ∗ (C (u) ∩ B(0, ρ) − ū) ,
78 R. Wenczel et al.
where the latter inclusion follows since both C (e) ∩ B(0, ρ) − ē and C (u) ∩
B(0, ρ) − ū are convex sets containing 0. Since χ ∈ l1 is arbitrary,
' (
cone d ∗ (C (e) ∩ B(0, ρ) − ē) + n ∗ (C (u) ∩ B(0, ρ) − ū) = l1 ,
so the compact convex set d ∗ (C (e) ∩ B̄(0, ρ) − ē) + n ∗ (C (u) ∩ B̄(0, ρ) − ū)
has a nonempty core (Definition 1) and hence by Proposition 1, a nonempty
interior. However, this latter property is forbidden for any compact subset of
an infinite-dimensional normed space. Thus we arrive at a contradiction.
We end this section with the promised verification that each truncation
(e)
Cn ∩ M = (Cn × C (u) ) ∩ M is indeed finite-dimensional.
Lemma 11. Under the assumptions of this section, Ck ∩ M is of finite di-
mension for each k.
Proof. Assume that C ∩ M has a member (ē, ū) with ē of finite length. Now
let (e, u) ∈ Ck ∩M , and let K := max{k, l(ē)}. Since d∗(e− ē)+n∗(u− ū) = 0
and e − ē ∈ M (0) − ē, then u − ū = T (e − ē). Since e − ē ∈ (M (0) − ē) ∩ RK ,
K
we can write e − ē = i=1 αi e(i) where the {e(i) }K i=0 is some spanning set for
(M (0) − ē) ∩ RK . Placing u(i) := T e(i) ∈ l1 , we obtain
K K
αi u(i) = T ( αi e(i) ) = T (e − ē) = u − ū
i=1 i=1
∞
(e) (e) (e) (e)
dρ (Cn , C) = max{|Ai |, |Bi |}for any ρ ≥ max{|Ai |, |Bi |}
i≥n i=0
(e)
Proof. Since Cn ⊆ C (e) for all n, we have
Lemma 13. The set C (e) is compact in l1 if and only if the bounds satisfy
∞
(e) (e)
max{|Ai |, |Bi |} < +∞ .
i=0
(e) (e)
Proof. Let i0 be such that Bi ≤ 0 ≤ Ai for all i ≥ i0 . If C (e) is compact,
(e) (e)
then xn := (A0 , .., An , 0, 0, ..) ∈ C (e) (n ≥ i0 ) must have a convergent
subsequence, along which we then have the uniform boundedness of the norms
nk (e) ∞ (e)
xnk = i=0 |Ai |, and since these increase with k, i=0 |Ai | is finite.
∞
Similarly, i=0 |Bi | is finite.
∞ (e) (e)
i=0 max{|Ai |, |Bi |} < +∞, the compactness of C
(e)
Conversely, if
(e)
follows from Lemma 9 since its truncations Cn are all compact and Attouch–
(e)
Wets converges to C by Lemma 12.
The next lemma shows that C ∩M is always bounded whenever the bounds
on e define sequences in l1 . This ensures that condition (4.12) will always be
satisfied for any objective f .
Lemma 14. Suppose (as usual here) that n̂ has no zeros on the unit circle
and that all its zeros in the unit disk are simple. Then C ∩ M ⊆ B(0, ρ0 ),
where
∞
6 7
(e) (e)
ρ0 = max max |Ai |, |Bi | ,
i=0
-
∞
6 7
(e) (e)
κ b 1 + d 1 max |Ai |, |Bi |
i=0
(e) (e)
and all n ≥ n0 for which i≥n max{|Ai |, |Bi |} < s and
2r + 2s + ρ (e) (e)
dρ+s (fn , f ) + max{|Ai |, |Bi |} < s ,
s
i≥n
it follows that
* * 6 7
* *
* inf fn − inf f * ≤ (2r + s + ρ)(2r + 2s + ρ) max |A
(e)
|, |B
(e)
|
* Cn ∩M C∩M * s2 i i
i≥n
2r + s + ρ
+ dρ+s (fn , f ) .
s
4 Convergence of truncates in l1 optimal feedback control 81
This, along with assumptions 1 and 1, may be inserted into Lemma 7 to yield
C ∩ B(0, σ) − {f ≤ r} ∩ M ∩ B(0, σ + s)
⊇ C ∩ B(0, σ) − M ∩ B(0, σ + s) ⊇ B(0, s) by (4.17)
we have
* * 6 7
* *
* inf f − inf f * ≤ 2r + s + ρ (e) (e)
max |Ai |, |Bi | .
*Cn ∩M C∩M * s
i≥n
Proof. This follows along similar lines to that for Theorem 1, but uses the
last displayed relation before Theorem 2 to obtain (4.21) from (4.17), so that
(4.14) is obtained for the above r, s and t and we may apply Corollary 4.
This argument in fact leads to a very quick proof of [14, Theorem 5.4]
(whose original proof contains a flaw, detailed in [27]) which asserts the
equality inf C∩M · 1 = inf C∩Mr · 1 under a slightly stronger assumption,
which in fact implies nonemptiness of (int C) ∩ M . To see the applicability of
Lemma 15 to the cited result from [14], we temporarily adopt the notation
thereof. Assign C and M as follows:
where btemp ∈ l∞ , bfeas ∈ Rcz × l1nz ×nw , Atemp : l1nz ×nw → l∞ and Afeas :
l1nz ×nw → Rcz × l1nz ×nw are bounded linear, and where the symbol ≤ stands
for the partial order on l∞ induced by its standard positive cone P + . The
assumption of [14, Theorem 5.4] is that btemp − Atemp Φ0 ∈ int P + for some
Φ0 ∈ M . However, the continuity of Atemp implies that Φ0 ∈ int C and hence
Φ0 ∈ (int C) ∩ M , which is the assumption of Lemma 15. This, coupled with
the density of M0 := Mr in M , gives the result.
As was discussed in Section 4.6.1, the approximating problems are con-
structed by truncating in the e-variable only, otherwise the method fails.
However, the truncated constraint-sets Ck ∩ M satisfy
Proof. Let (e, u) ∈ Ck ∩ M and let (e1 , u1 ) := (e, u) − (ē, ū). Then l(e1 ) ≤
max{l(e), l(ē)} and similarly for u1 . Also u1 = −d ∗ Z −1 (ê1 /n̂) from the
corresponding convolution relation defining M . Since ê1 is a polynomial hav-
ing zeros at each zero of n̂ in the unit disk and hence the whole plane (re-
call e1 ∈ M (0) − ē), we have that û1 is a polynomial of degree at most
l(d) − l(n) + l(e1 ) − 1 ≤ l(d) − l(n) + max{l(e), l(ē)} − 1, and so, as u = u1 + ū,
the result follows.
For the second assertion, suppose n̂(z0 ) = 0 for some |z0 | > 1. With (ē, ū)
as above, it follows from the closed–loop equation d∗ē+n∗ū = w∗d−n∗β =: b̃
(where β := βc c + βs s) that ŵ(z0 ) is finite and ēˆ(z0 ) = ŵ(z0 ). Let k exceed
both the number of constraints defining
M (0) and the length of ē, and let
ˆ ˆ
e ∈ M (0) ∩ Rk . If u := Z −1 b̃−dê
n̂ , the interpolation constraints on e ensure
that û has no poles in the closed unit disk (so u ∈ l1 ) and hence (e, u) ∈ M .
(e)
Also, since from the assumptions, cone (Ck − ē) ⊇ Rk and cone (C (u) − ū) =
1
l , we have
whence λ((e, u) − (ē, ū)) ∈ Ck ∩ M − (ē, ū) for some positive λ. Since now
λ(e, u) + (1 − λ)(ē, ū) ∈ Ck ∩ M , the hypothesized finiteness of length of
λu + (1 − λ)ū implies, via the equation
that (λe + (1 − λ)ē)(z0 ) = ŵ(z0 ) and hence ê(z0 ) = ŵ(z0 ). Since k is arbi-
trary, we have shown that every finite-length e ∈ M (0) satisfies an additional
interpolation constraint ê(z0 ) = ŵ(z0 ) at z0 , which yields a contradiction.
Remark 6. Under the assumptions of the preceding lemma, note that for k ≥
max{l(ē), l(ū) − l(d) + l(n)}, and (e, u) ∈ Ck ∩ M , we have l(u) ≤ k + l(d) −
l(n) := k + l, so for such k, Ck ∩ M consists precisely of those elements (e, u)
of C ∩ M with e of length k and u of length k + l.
If l ≤ 0, then
Cn ∩ M ⊆ Qn+l ⊆ Cn+l ∩ M,
4 Convergence of truncates in l1 optimal feedback control 85
inf fn → inf f
Qn C∩M
(e) (e)
1 Bi ≤ ēi ≤ Ai for all i ∈ N;
(e) (e)
2 i |Ai | and i |Bi | < ∞;
3 for a subsequence {ik }∞
(e) (e)
k=0 we have Bik = ēik = Aik , and for all i not in
(e) (e)
this subsequence, Bi < ēi < Ai .
Then 0 ∈ sqri (C (e) − M (0) ).
cone (C (e) − M (0) ) = cone P(C (e) − M (0) ) + (I − P)(M (0) − ē) . (4.34)
If we can show that cone P(C (e) − M (0) ) = Pl1 then (4.34) would give
which must be closed since Pl1 is closed and the linear subspace M (0) − ē
has finite codimension.
We now verify that cone P(C (e) −M (0) ) = Pl1 . (This part of the argument
parallels that of Lemma 6.) Let ξ ∈ Pl1 , with
−1 −1 (e) (e)
ξ 1 < αK := αK min {|Ai − ξi |, |Bi − ξi |} > 0 .
0≤i≤K−1; i∈{i
/ k}
ξ = Pη − (Pη − ξ)
∈ P(C (e) − ē) − P(M (0) − ē) = P(C (e) − M (0) ).
−1
This shows that B(0, αK ) ∩ Pl1 ⊆ P(C (e) − M (0) ) whence
cone P(C − M ) = Pl1 as required.
(e) (0)
Corollary 5. Under the conditions of Lemmas 17 and 18, and for any convex
closed real-valued f : l1 × l1 → R,
4 Convergence of truncates in l1 optimal feedback control 87
Proof. By the lemmas (and the cited result from [13]) the indicator functions
δCn ∩M Attouch–Wets converge to δC∩M , and (since dom f has an interior)
by any of the cited sum theorems, f + δCn ∩M → f + δC∩M also. The result
then follows from Proposition 4.
Our formulation of the control problem (see Section 4.4) was chosen to
permit its re-expression as an optimization over l1 × l1 . This choice resulted
from a wish to compare with the other methods we described in the in-
troduction, which used duality theory. Accordingly, we sought to formulate
minimizations over a space such as l1 , which has a nice dual (namely, l∞ ).
However, this is not the only problem formulation that can be treated by the
methods of this paper.
Recall from Section 4.4 that we considered only stabilizing controllers
K ∈ S(P ) for which the resulting control signal u was restricted to the
subspace Xa ⊆ l∞ . This requirement will now be relaxed to u ∈ l∞ . This
will be seen to entail only trivial changes to the main results of this section.
(Note that in this case the resulting optimization is taken over l1 × l∞ , so
duality methods would not be readily applicable, since the dual (l∞ )∗ has a
complicated characterization.)
The basic feasible set is now of the form
For simplicity, we consider only the required changes to Theorem 2, since this
result has a simpler statement than Theorem 1. In fact, with the assumptions
of Theorem 2, but with f : l1 ×l∞ → R given by f (e, u) = e 1 +ζ u ∞ , and
all references to ū 1 converted to ū ∞ , the form of Theorem 2 is unaltered,
88 R. Wenczel et al.
except that the ρ0 of Lemma 14 is not available, but another value may be
used (namely, ρ0 = (γ + 1) max{1, ζ −1 }).
we obtain
* * 6 7
* *
* inf f − inf f * ≤ 2r + s + ρ (e) (e)
max |Ai |, |Bi | .
*Cn ∩M C∩M * s
i≥n
Proof. Since the form of the convergence theory of Section 4.5 (and the as-
sociated CQs) is insensitive to the particular norm used, all we need to do
to obtain a proof in this context is to reinterpret all statements in u-space
as referring instead to the norm · ∞ . The only places where changes occur
are in Lemma 7, where all references to ū 1 become ū ∞ ; and instead of
using ρ0 from Lemma 14, we exploit the form of f to ensure the existence of
some ρ0 for which Condition (4.15) of Proposition 4 is valid.
4.8 Appendix
and so cannot have any poles in the closed disk D̄. Since
1 − z cos θ z sin θ
ĉ(z) = and ŝ(z) = ,
(z − a)(z − ā) (z − a)(z − ā)
it follows that
ˆ − uc n̂(z))(1 − z cos θ) + (ws d(z)
(wc d(z) ˆ − us n̂(z))z sin θ
z →
(z − a)(z − ā)
has no poles in D̄ and hence none at all in C, and must be a polynomial (so
that d ∗ (wc c + ws s) − n ∗ (uc c + us s) has finite length). If now θ = 0, so s = 0,
the above amounts to stating that
ˆ − uc n̂(z)
wc d(z)
z →
1−z
ˆ
is polynomial, from which we obtain uc = wc d(1)/n̂(1) = wc /P (1). Sim-
ˆ c n̂
wc d−u
ilarly, if θ = π (so again s = 0), we have 1+· polynomial, so that
ˆ
uc = wc d(−1)/n̂(−1) = wc /P (−1). For other values of θ, similar reason-
ing implies
ˆ − uc n̂(a))(1 − a cos θ) + (ws d(a)
(wc d(a) ˆ − us n̂(a))a sin θ = 0.
with the right-hand side vanishing if P has a pole at a. Solving this for uc
and us gives the desired relation.
90 R. Wenczel et al.
Lemma 19. Let fˆ ∈ l 1 (C) and let |a| < 1 be a zero of fˆ. Then fˆ(·)/(· − a) ∈
Lemma 20. If fˆ ∈ l
1 (C) and |a| > 1, then fˆ(·) 1 (C) and is in l1 if fˆ ∈ l1
∈ l
·−a
and a ∈ R.
Proof. As in the preceding lemma, let q be the inverse transform of the func-
tion in question. From the expression for qk given there, and an interchange
of summations,
∞ 1 ∞
1 f 1
q 1 ≤ |fj ||a|j = |fj ||a|j = 1 .
j=0
|a| k+1
j=0 |a|j 1 − 1 1 − |a|
k≥j |a|
d(z̄)
w̄(z̄) ˆ −d(z̄)(w
ˆ ˆ
c ĉ(z̄) + ws ŝ(z̄)) + n̂(z̄)(βc ĉ(z̄) + βs ŝ(z̄)) − ê(z̄)d(z̄)
d(z̄)
= w̄(z̄) ˆ − d(z̄)(w
ˆ ˆ
c ĉ(z̄) + ws ŝ(z̄)) − φ̂(z̄)d(z̄)
ˆ =0
= (ŵ(z̄) − ê(z̄))d(z̄)
References
Musa A. Mamedov
Abstract In this chapter we study the turnpike property for the nonconvex
optimal control problems described by the differential inclusion ẋ ∈ a(x). We
T
study the infinite horizon problem of maximizing the functional 0 u(x(t)) dt
as T grows to infinity. The purpose of this chapter is to avoid the convexity
conditions usually assumed in turnpike theory. A turnpike theorem is proved
in which the main conditions are imposed on the mapping a and the function
u. It is shown that these conditions may hold for mappings a with nonconvex
images and for nonconcave functions u.
Musa A. Mamedov
School of Information Technology and Mathematical Sciences, University of Ballarat,
Victoria 3353, AUSTRALIA
e-mail: musa [email protected]
We assume that the trajectories of system (5.1) are uniformly bounded, that
is, there exists a number L < +∞ such that
Note that in this work we focus our attention on the turnpike property of
optimal trajectories. So we do not consider the existence of bounded tra-
jectories defined on [0, ∞]. This issue has been studied for different control
problems by Leizarowitz [5], [6], Zaslavsky [19], [20] and others.
Definition 2. The trajectory x(·) is called optimal if J(x(·)) = JT∗ and is
called ξ-optimal (ξ > 0) if
5 Asymptotical stability of optimal paths in nonconvex problems 97
J(x(·)) ≥ JT∗ − ξ.
Definition 3. The point x is called a stationary point if 0 ∈ a(x).
Turnpike theorems for the problem (6.1), (5.2) have been proved in [14], [18]
and elsewhere, where it was assumed that the graph of the mapping a is a
compact convex set and the function u is concave. The main conditions are
imposed on the Hamiltonian. In this chapter a turnpike theorem is presented
in which the main conditions are imposed on the mapping a and the function
u. Here we present a relation between a and u which provides the turnpike
property without needing to impose conditions such as convexity of the graph
of a and of the function u. On the other hand this relation holds if the graph
of a is a convex set and the function u is concave.
Condition M There exists b < +∞ such that for every T > 0 there is a
trajectory x(·) ∈ XT satisfying the inequality
JT (x(·)) ≥ u∗ T − b.
Here the notation py means the scalar product of the vectors p and y. By |c|
we will denote the absolute value of c.
We also define the function
u(x)−u∗ u(y)−u∗
ϕ(x, y) = |c(x)| + c(y) .
Note that if Condition H is satisfied for any vector p then it is also satisfied
for all λp, (λ > 0). That is why we assume that ||p|| = 1.
Condition H1 means that derivatives of the system (6.1) are directed to
one side with respect to p, that is, if x ∈ B, x = x∗ , then py < 0 for all
y ∈ a(x). It is also clear that py ≤ 0 for all y ∈ a(x∗ ) and c(x∗ ) = 0.
Condition H2 means that there is a point x̃ on the plan {x ∈ Rn : p(x −
∗
x ) = 0} such that pỹ > 0 for some ỹ ∈ a(x̃). This is not a restrictive
assumption, but the turnpike property may be not true if this condition does
not hold.
The main condition here is H3. It can be considered as a relation between
the mapping a and the function u which provides the turnpike property.
Note that Conditions H1 and H3 hold if the graph of the mapping a is a
convex set (in Rn × Rn ) and the function u is strictly concave. In the next
example we show that Condition H can hold for mappings a without a convex
graph and for functions u that are not strictly concave (in this example the
function u is convex).
Example 1 Let x = (x1 , x2 ) ∈ R2 and the system (6.1) have the form
·
x1 = λ[x21 + (x22 + 1)2 + w], − 1 ≤ w ≤ +1,
·
x2 = f (x1 , x2 , v), v ∈ U ⊂ Rm .
Here λ > 0 is a positive number, the function f (x1 , x2 , v) is continuous and
f (0, 0, ṽ) = 0 for some ṽ ∈ U.
5 Asymptotical stability of optimal paths in nonconvex problems 99
Thus
u(x) − u∗ c(x2 + (1/2)ξ 2 ) c
< ≤ ,
|c(x)| λ(ξ 2 + x22 + 2x2 ) 2λ
u(y) − u∗ c(y2 + (1/2)ξ 2 ) c
< ≤− .
c(y) −λ(ξ 2 + y22 + 2y2 ) 2λ
100 M.A. Mamedov
From these inequalities we have ϕ(x, y) < 0, that is, the first part of H3
holds. The second part of Condition H3 may also be obtained from these
inequalities. Therefore Condition H holds.
We now formulate the main result of the current chapter.
Theorem 1. Suppose that Conditions M and H are satisfied and that the
optimal stationary point x∗ is unique. Then:
1. there exists C < +∞ such that
T
(u(x(t)) − u∗ )dt ≤ C
0
In this section a set D is introduced. This set will be used in all sections
below.
Denote
M∗ = {x ∈ Ω : c(x) ≥ 0}.
Clearly M ⊂ M∗ . We recall that B = {x ∈ Ω : u(x) ≥ u∗ }.
Consider a compact set D ⊂ Ω for which the following conditions hold:
a) x ∈ int D for all x ∈ B, x = x∗ ;
b) c(x) < 0 for all x ∈ D, x = x∗ ;
c) D ∩ M∗ = {x∗ } and B ⊂ D.
It is not difficult to see that there exists a set D with properties a), b) and
c). For example such a set can be constructed as follows. Let x ∈ B, x =
x∗ . Then c(x) < 0. Since the mapping a is continuous in the Hausdorff
metric the function c(x) is continuous too. Therefore there exists εx > 0
such that c(x ) < 0 for all x ∈ Vεx (x) ∩ Ω. Here Vε (x) represents the open
ε-neighborhood of the point x. In this case for the set
5 Asymptotical stability of optimal paths in nonconvex problems 101
⎧ ⎫
⎨ / ⎬
D = cl V 21 εx (x) ∩ Ω
⎩ ⎭
x∈B,x =x∗
u(x) ≤ u∗ − νε
Proof. Assume to the contrary that for any ε > 0 there exists a sequence xk
such that xk ∈ D, ||xk − x∗ || ≥ ε and c(xk ) → 0. Let x be a limit point of the
sequence xk . Then x ∈ D, x = x∗ and c(x ) = 0. This contradicts Property
b) of the set D.
We show that for every ε > 0 there exists δε > 0 such that
and
xn → x̄, yn → ȳ, ϕ(xn , yn ) → 0.
From Lemma 17.4.1 it follows that the sequence {(u(xn ) − u∗ )/|c(xn )|} is
bounded. Since ϕ(xn , yn ) → 0, the sequence {(u(yn ) − u∗ )/c(yn )} is also
bounded and therefore from Lemma 17.3.1 we obtain c(ȳ) > 0. We also
obtain c(x̄) < 0 from the inclusion x̄ ∈ D ∩ Xε . Thus the function ϕ(x, y)
is continuous at the point (x̄, ȳ). Then from ϕ(xn , yn ) → 0 it follows that
ϕ(x̄, ȳ) = 0, which contradicts (5.4).
We now consider the case where x ∈ D, x = x∗ . Assume that (5.5)
does not hold. Then there exist sequences (xn ) and (yn ), for which pxn =
pyn , c(xn ) < 0, c(yn ) > 0, xn ∈ D, yn ∈ M∗ ∩ Xε and xn → x̄, yn → ȳ,
ϕ(xn , yn ) → 0. If x̄ = x∗ , we have a contradiction similar to the first case. If
x̄ = x∗ , taking the inequality ȳ = x∗ into account we obtain a contradiction
to the second part of Condition H3.
Thus we have shown that (5.5) is true.
Define the function
u(x) − u∗ u(y) − u∗
ϕ(x, y)δ1 ,δ2 = + , (δ1 ≥ 0, δ2 ≥ 0).
|c(x)| + δ1 c(y) + δ2
u(x) − u∗
≤b for all x ∈ D, x = x∗ . (5.6)
|c(x)|
By using γ(ε) we divide the set M∗ ∩ Xε ∩ {y : c(y) > 0} into two parts:
1
Y1 = {y ∈ M∗ ∩ Xε ∩ {y : c(y) > 0} : c(y) ≥ γ(ε)},
2
1
Y2 = {y ∈ M∗ ∩ Xε ∩ {y : c(y) > 0} : c(y) < γ(ε)}.
2
5 Asymptotical stability of optimal paths in nonconvex problems 103
1
Consider the set Y2 . Denote δ̄2 = 2 γ(ε) and take any number δ̄1 > 0.
Then
u(x) − u∗
≤b for all x ∈ D, x = x∗ , δ1 ≤ δ̄1 (5.9)
|c(x)| + δ1
and
1 1
c(y) + δ2 ≤ c(y) + δ̄2 < γ(ε) + γ(ε) = γ(ε)
2 2
for all 0 ≤ δ2 ≤ δ̄2 , y ∈ Y2 . Using (5.8) we obtain
u(y) − u∗ νε νε
≤− ≤− for all y ∈ Y2 . (5.10)
c(y) + δ2 c(y) + δ2 γ(ε)
u(x) − u∗ u(x) − u∗
≤ for all x ∈ D, x = x∗ , u(x) ≥ u∗ , 0 ≤ δ1 ≤ δ̄1 .
|c(x)| + δ1 |c(x)|
Then if px = py we have
1 1 1
ϕ(x, y)δ1 ,δ2 ≤ ϕ(x, y) + δε ≤ − δ ε + δε = − δε , (5.12)
2 2 2
for all x ∈ D, x = x∗ , u(x) ≥ u∗ , y ∈ Y1 , px = py,
Now consider the case when x ∈ D and u(x) < u∗ . Since the function c(y)
is bounded on the set Y1 , then for ε > 0 there exist δ̄ˆ22 (ε) > 0 and δ̂ε > 0
such that
u(y) − u∗ 1
≤ − δ̂ε for all y ∈ Y1 , 0 ≤ δ2 ≤ δ̄ˆ22 (ε).
c(y) + δ2 2
Then
u(y) − u∗ 1
ϕ(x, y)δ1 ,δ2 ≤ ≤ − δ̂ε , (5.13)
c(y) + δ2 2
u(x) − u∗ u(x ) − u∗ 1
− ≤ δ̄ε
|c(x)| + δ1 |c(x )| + δ1 2
Here the functions η(x , y) and δ(y) are such that δ(y) is continuous and
∼ ∧ ∧
for every ε > 0, ε> 0 there exist δ ε > 0 and η ε,∼ε > 0 such that
∧ ∧
δ(y) ≥δ ε and η(x , y) ≥η ε,∼ε
∼
for all (x , y) for which ||x − x∗ || ≥ ε, ||y − x∗ || ≥ ε.
In this section, for a given trajectory x(t) ∈ XT we divide the interval [0, T ]
into two types of intervals and prove some integral inequalities. We assume
that x(t) is a given continuously differentiable trajectory.
5.5.1
Consider a set {t ∈ [0, T ] : x(t) ∈ int D}. This set is an open set and therefore
it can be presented as a union of a countable (or finite) number of open
intervals τk = (tk1 , tk2 ), k = 1, 2, 3, ..., where τk ∩ τl = ∅ if k = l. We denote
the set of intervals τk by G = {τk : k = 1, 2, ...}.
From the definition of the set D we have
d ·
px(t) = p x (t) ≤ c(x(t)) < 0 for all t ∈ τk , k = 1, 2, ....
dt
Then px(tk1 ) > px(tk2 ) for all k.
We introduce the notation piτk = px(tki ), i = 1, 2, ..., Pτk = [p2τk , p1τk ] and
Pτk = (p2τk , p1τk ).
0
106 M.A. Mamedov
5.5.2
We divide the set G into the sets gm (G = ∪m gm ), such that for every set
g = gm the following conditions hold:
a) The set g consists of a countable (or finite) number of intervals τk , for
which
Take some set g 1 ∈ G and consider all sets gm ∈ G, m = ±1, ±2, ±3, ... for
which
... < g−2 < g−1 < g 1 < g1 < g2 < ....
Proof. Consider case a). Take some gk . By Definition 5 it is clear that there
exists an interval τk ∈ gk and a point tk ∈ τk such that
x(tk ) ∈ D. (5.16)
Consider the interval [t2gk , t1gk+1 ]. Since gk < gk+1 by Definition 5 we have
px(t2gk ) < px(t1gk+1 ). Therefore there exists a point sk ∈ (t2gk , t1gk+1 ) such that
·
p x (sk ) ≥ 0, which implies
x(sk ) ∈ M∗ . (5.17)
5.5.3
Take the set Gi . We denote by t1i an exact upper bound of the points t2Gm sat-
isfying t2Gm ≤ t1Gi and by t2i an exact lower bound of the points t1Gm satisfying
t1Gm ≥ t2Gi .
Proposition 2. There exist points ti ∈ [t1i , t1Gi ] and ti ∈ [t2Gi , t2i ] such that
x(ti ) = x(ti ) = x∗ .
Proof. First we consider the interval [t1i , t1Gi ]. Two cases should be studied.
1. Assume that the exact upper bound t1i is not reached. In this case there
exists a sequence of intervals [t1Gm , t2Gm ] such that t2Gm → t1i and t1Gm →
t1i . Since the intervals (t1Gm , t2Gm ) are disjoint we obtain that x(t1i ) = x∗ .
Therefore ti = t1i .
108 M.A. Mamedov
Note We take ti = 0 (or ti = T ) if for the chosen set Gi there does not exist
Gm such that t2Gm ≤ t1Gi (or t1Gm ≥ t2Gi , respectively).
Lemma 4. The interval [0, T ] can be divided into an at most countable num-
ber of intervals [0, t1 ], [t1k , t2k ] and [t2 , T ], such that the interiors of these in-
tervals are disjoint and
a) [0, T ] = [0, t1 ] ∪ {∪k [t1k , t2k ]} ∪ [t2 , T ];
b) in each interval [0, t1 ], [t1k , t2k ] and [t2 , T ] there is only one set G0 , Gk
and GT , respectively, and
G = G0 ∪ {∪k Gk } ∪ GT ;
5.5.4
Proof: For the sake of definiteness, we assume that px(t) ∈ [p1 , p2 ], t ∈ [t1 , t2 ].
Otherwise we can consider an interval [t1 , t2 ] ⊂ [t1 , t2 ], for which px(t) ∈
[p1 , p2 ], t ∈ [t1 , t2 ] and pi = px(ti ), i = 1, 2.
We set t(q) = min{t ∈ [t1 , t2 ] : px(t) = q} for all q ∈ [p1 , p2 ] and then
define a set
m = {t(q) : q ∈ [p1 , p2 ]}.
Clearly m ⊂ [t1 , t2 ]. Consider a function a(t) = px(t) defined on the set m.
It is not difficult to see that for every q ∈ [p1 , p2 ] there is only one number
.
t(q), for which a(t(q)) = q, and also a (t(q)) ≥ 0, ∀q ∈ [p1 , p2 ].
We divide the interval [p1 , p2 ] into two parts as follows:
. .
P1 = {q : a (t(q)) = 0} and P2 = {q : a (t(q)) > 0}.
tk
2 βn
· · ·
p x (t)dt + p x (t)dt + p x (t)dt.
k n α
tk
1
n ∪n {βn }
110 M.A. Mamedov
· .
It is not difficult to observe that p x (t) = a (t) = 0, ∀t ∈ m(P1 ),
meas(∪n {βn }) = 0 and px(αn ) = px(βn ), n = 1, 2, ... (see [2]). Then we
obtain
p2 − p1 = (px(tk2 ) − px(tk1 )) = (pk2 − pk1 ).
k k
Therefore for the intervals [tk1 , tk2 ] all assertions of the lemma hold.
Lemma 6. Assume that on the intervals [t1 , t2 ] and [s2 , s1 ] the following con-
ditions hold:
1. px(ti ) = px(si ) = pi , i = 1, 2.
2. x(t) ∈ int D, ∀t ∈ (t1 , t2 ). In particular, from this condition it follows
·
that p x (t) < 0, ∀t ∈ (t1 , t2 ).
·
3. p x (s) > 0, ∀s ∈ (s2 , s1 ).
Then
t2
s1
s1
∗
u(x(t))dt + u(x(s))ds ≤ u [(t2 − t1 ) + (s1 − s2 )] − δ 2 (x(s))ds
t1 s2 s2
∼
ε= ρ (x∗ , {x(t) : t ∈ [t1 , t2 ]}) > 0 and ε = ρ (x∗ , {x(s) : s ∈ [s2 , s1 ]}) > 0.
∧ ∧
Now we use Lemma 17.5.1. We define δ =δ ε > 0 and η =η ε,∼ε for the chosen
∼
numbers ε and ε. We take any number N > 0 and divide the interval [p2 , p1 ]
into N equal parts [pk2 , pk1 ]. From Conditions 2 and 3 it follows that in this
case the intervals [t1 , t2 ] and [s2 , s1 ] are also divided into N parts, say [tk1 , tk2 ]
and [sk2 , sk1 ], respectively. Here px(tki ) = px(ski ) = pki , i = 1, 2, k = 1, ..., N.
Clearly
p 1 − p2
pk1 − pk2 = → 0 as N → ∞.
N
∼
Since x(t) ∈ D and ||x(t) − x∗ || ≥ ε > 0 for all t ∈ [t1 , t2 ] then from
Lemma 17.4.1 it follows that
·
p x (t) ≤ c(x(t)) < −η∼ε < 0.
That is why for every k we have tk2 − tk1 → 0 as N → ∞. Therefore for a
given η > 0 there exists a number N such that
5 Asymptotical stability of optimal paths in nonconvex problems 111
Suppose that (5.19) is not true. In this case there exists a sequence of inter-
vals [sk2N , sk1N ], such that ski N → si , i = 1, 2, and s2 < s1 . Since pk1N −pk2N → 0
as N → ∞ then px(s2 ) = px(s1 ) = p , and moreover px(s) = p for all
s ∈ [s2 , s1 ]. This is a contradiction. So (5.19) is true.
A. Now we take any number k and fix it. For the sake of simplicity we
denote the intervals [tk1 , tk2 ] and [sk2 , sk1 ] by [t1 , t2 ] and [s2 , s1 ], respectively. Let
pi = px(ti ) = px(si ), i = 1, 2.
Take any s ∈ (s2 , s1 ) and denote by t the point in the interval (t1 , t2 ) for
which px(t ) = px(s). From (6.8) it follows that
Therefore we can apply Lemma 17.5.1. We also note that the following
conditions hold:
·
• |cx(t))| ≤ |p x (t)| for all t ∈ (t1 , t2 );
·
• cx(s)) ≥ p x (s) for all s ∈ (s2 , s1 );
• u(x(s)) ≤ u∗ for all s ∈ [s2 , s1 ].
Then from Lemma 17.5.1 we obtain
u(x(t)) u(x(s)) ∗ 1 1
· + · ≤u · + ·
|p x (t)| + δ1 p x (s) + δ2 |p x (t)| + δ1 p x (s) + δ2
− δ(x(s)), (5.20)
∼
ξ = δ(x( s)).
s2 − s1
π = px(t) + ξ (t1 − t), t ∈ [t1 , t2 ],
t1 − t2
ω = px(s) + ξ(s − s1 ), s ∈ [s2 , s1 ].
112 M.A. Mamedov
Clearly
· ∼ ·
dπ = [p x (t)− ξ ]dt and dω = [p x (s) + ξ]ds,
∼
where ξ = ξ(s2 − s1 )/(t1 − t2 ).
· ∼ ·
Since p x (t)− ξ < 0 and p x (s) + ξ > 0 then there exist inverse functions
t = t(π) and s = s(ω). We also note that π1 = px(t1 ) = px(s1 ) = ω1 and
π2 = px(t2 ) + ξ(s2 − s1 ) = ω2 .
Therefore we have
t2
s1
A = u(x(t))dt + u(x(s))ds
t1 s2
2
π
ω1
u(x(t(π))) u(x(s(ω)))
= · ∼ dπ +
· dω
p x (s(ω)) + ξ
π1 p x (t(π))− ξ ω2
ω1
u(x(t(ω))) u(x(s(ω)))
= · ∼ + · dω.
ω2 |p x (t(ω))|+ ξ p x (s(ω)) + ξ
∼
Let δ¯1 > ξ . Since
ξ ≤ δ(x(t)) = δ(x(t(ω))), t(ω) ∈ [t1 , t2 ], s(ω) ∈ [s2 , s1 ],
⎛t ⎞
2
s1
s1
∗⎝ ·
=u ⎠
dt + ds − δ(x(s))[p x (s) + ξ]ds
t1 s2 s2
s1
∗
≤ u [(t2 − t1 ) + (s1 − s2 )] − ξδ(x(s)ds.
s2
∼
On the other hand δ(x(s)) ≥ ξ = δ(x( s)). Thus
∼
A ≤ u∗ [(t2 − t1 ) + (s1 − s2 )] − (s1 − s2 )δ 2 (x( s)).
t2
s1
∼
u(x(t))dt + u(x(s))ds ≤ u∗ [(tk2 − tk1 ) + (sk1 − sk2 )] − (sk1 − sk2 )δ 2 (x(sk )).
t1 s2
Therefore the lemma is proved taking into account (5.19) and passing to
the limit as N → ∞.
II. Now consider the case when p∗ = pi for some i = 1, 2. For the sake of
definiteness we assume that p∗ = p1 .
Take any number α > 0 and consider the interval [p2 , p1 − α]. Denote by
[t1 − t(α), t2 ] and [s2 , s1 − s(α)] the intervals which correspond to the interval
[p2 , p1 − α]. Clearly t(α) → 0 and s(α) → 0 as α → 0. We apply the result
proved in the first part of the lemma for the interval [p2 , p1 − α]. Then we
pass to the limit as α → 0. Thus the lemma is proved.
5.5.5
Lemma 7. Assume that π and ω are sets of 1st and 2nd type on the interval
[p2 , p1 ], respectively. Then
u(x(t))dt ≤ u∗ meas(π ∪ ω) − [u∗ − u(x(t))]dt − δ 2 (x(t))dt,
π∪ω Q E
where
a) Q ∪ E = ω ∪ π2 = {t ∈ π ∪ ω : x(t) ∈/ int D};
b) for every ε > 0 there exists a number δε > 0 such that
c) for every δ > 0 there exists a number K(δ) < ∞ such that
meas[(π ∪ ω) ∩ Zδ ] ≤ K(δ)meas[(Q ∪ E) ∩ Zδ ],
pn1 − pn2 = (pnm
1 − pnm
2 ).
m
(sm m
2 , s1 ), such that:
Now we take some interval d0k = (tk1 , tk2 ) and let pki = px(tki ), i = 1, 2. Denote
2 , p1 ] = [p2 , p1 ] ∩ [p2 , p1 ].
[pkm km k k m m
(5.22)
5 Asymptotical stability of optimal paths in nonconvex problems 115
·
Since p x (t) < 0, for all t ∈ d0k , from (5.21) it follows that there are two
intervals [tkm km km km
1 , t2 ] and [s2 , s1 ] corresponding to the nonempty interval
[pkm
2 , pkm
1 ], and
2 − t1 ) = t2 − t1 .
(tkm km k k
tk km
2 s1
u(x(t))dt + u(x(s))ds
k k,m km
tk
1 s2
⎡ ⎤ km
s1
≤ u∗ ⎣ (tk2 − tk1 ) + km ⎦
1 − s2 )
(skm − δ 2 (x(s))ds.
k k,m k,m km
s2
where Q = π2 ∪ (ω \ ω ) and E = ω .
Now we check Conditions a), b) and c) of the lemma.
Condition a) holds, because Q ∪ E = π2 ∪ (ω \ ω ) ∪ ω = π2 ∪ ω. Condition
b) follows from Lemma 17.5.1. We now check Condition c).
Take any number δ > 0 and denote Pδ = {l : |l − p∗ | ≥ δ}.
116 M.A. Mamedov
pk1 − pk2 = 1 − p2 ).
(pkm km
(5.23)
m
·
p x (t) ≤ c(x(t)) < −ηδ for all t ∈ π1 ∩ Zδ . (5.24)
On the other hand there exists a number K < +∞, for which
·
p x (t) ≤ K for all t ∈ [0, T ].
Therefore
·
meas [pk2 , pk1 ] ∩ Pδ = [−p x (t)]dt ≥ ηδ meas (dk ∩ Zδ ).
dk ∩Zδ
K K
meas (π1 ∩ Zδ ) ≤ meas (E ∩ Zδ ) ≤ meas [(Q ∪ E) ∩ Zδ ].
ηδ ηδ
But Q ∪ E = π2 ∪ ω and therefore
In this section we divide the sets G0 , Gk and GT (see Lemma 4) into sets
of 1st and 2nd type such that Lemma 7 can be applied. Note that x(t) is a
continuously differentiable trajectory.
5.6.1
Here πk and ωk are the sets of 1st and 2nd type in the interval [pk2 , pk1 ] and
the set F is either a set of 1st type in the interval [p2gN , p1g1 ], if p2gN ≤ p1g1 , or
is a set of 2nd type in the interval [p1g1 , p2gN ], if p2gN > p1g1 .
Proof. Take the set g1 and assume that [p2g1 , p1g1 ] and [t1g1 , t2g1 ] are the corre-
sponding intervals. Note that we are using the notation introduced in Section
5.5. Take the set g2 .
A. First we consider the case p1g2 < p1g1 . In this case there is a point
t ∈ [t1g1 , t2g1 ] such that px(t1 )) = p1g2 . Denote π1 = [t1 , t2g1 ] and ω1 = [t2g1 , t1g2 ].
1
Note that by Definition 6 we have p2g1 < p1g2 . It is clear that π1 and ω1 are
sets of 1st and 2nd type on the interval [p2g1 , p1g2 ], respectively. Therefore
2
tg2
u(x(t))dt = u(x(t))dt + u(x(t))dt,
t1g1 π1 ∪ω1 π11
where π11 = [t1g1 , t1 ] ∪ [t1g2 , t2g2 ] is a set of 1st type on the interval [p2g2 , p1g1 ].
B. Now we assume that p1g2 ≥ p1g1 . In this case there is a point t1 ∈ [t2g1 , t1g2 ]
such that px(t1 ) = p1g1 . Denote π1 = [t1g1 , t2g1 ] and ω1 = [t2g1 , t1 ]. Consider two
cases.
1. Let p2g2 ≥ p1g1 . Then there is a point t2 ∈ [t1 , t1g2 ] such that px(t2 )) = p2g2 .
In this case we denote
118 M.A. Mamedov
Therefore
2
tg2
u(x(t))dt = u(x(t))dt + u(x(t))dt,
i=1,2π ∪ω
t1g1 i i ω1
Therefore
2
tg2
u(x(t))dt = u(x(t))dt + u(x(t))dt,
i=1,2π ∪ω
t1g1 i i π1
t
2
Here πn and ωn are sets of 1st and 2nd type in the interval [p2n , p1n ] and the
set F is either a set of 1st type in the interval [p∗ , p1g1 ], if p∗ ≤ p1g1 , or is a
set of 2nd type in the interval [p1g1 , p∗ ], if p∗ > p1g1 .
Proof. We apply Lemma 8 for every n. From Proposition 1 we obtain that
x(t) → x∗ as t → t2 , and therefore p2gn → p∗ as n → ∞. This completes the
proof.
We can prove the following lemmas in a similar manner to that used for
proving Lemmas 8 and 9.
Lemma 10. Assume that the set Gi consists of a finite number of elements
gk , g1 > g2 > ... > gN . Then
2
tg1
u(x(t))dt = u(x(t))dt + u(x(t))dt.
t1g k π ∪ω F
k k
N
5 Asymptotical stability of optimal paths in nonconvex problems 119
Here πk and ωk are sets of 1st and 2nd type in the interval [p2k , p1k ] and the
set F is either a set of 1st type in the interval [p2g1 , p1gN ], if p1gN ≥ p2g1 , or is
a set of 2nd type in the interval [p1gN , p2g1 ], if p1gN < p2g1 .
Here πn and ωn are sets of 1st and 2nd type in the interval [p2n , p1n ] and the
set F is either a set of 1st type in the interval [p2g1 , p∗ ], if p∗ ≥ p2g1 , or is a
set of 2nd type in the interval [p∗ , p2g1 ], if p∗ < p2g1 .
In the next lemma we combine the results obtained by Lemmas 9 and 11.
Lemma 12. Assume that the set Gi consists of elements gn , n = ±1,
±2, . . ., where · · · < g−2 < g−1 < g1 < g2 < · · · , and where t1 =
limn→−∞ t1gn and t2 = limn→∞ t2gn . Then
t
2
u(x(t))dt = u(x(t))dt.
n π ∪ω
t1 n n
Here πn and ωn are sets of 1st and 2nd type in the interval [p2n , p1n ].
t
2
t2g
−1
u(x(t))dt = u(x(t))dt + u(x(t))dt. (5.28)
n ∪ω
t1 πn n F
We define π0 = F ∪ F and ω0 = [t2g−1 , t1g1 ]. Clearly they are sets of 1st and
2nd type in the interval [p2g−1 , p1g1 ] (note that p2g−1 < p1g1 by Definition 5).
Therefore the lemma is proved if we sum (5.27) and (5.28).
5.6.2
Now we use Lemma 4. We take any interval [t1k , t2k ] and let
120 M.A. Mamedov
We show that
2
tk
u(x(t))dt = u(x(t))dt + u(x(t))dt, (5.29)
n k ∪ω k
t1k πn n Ek
where πnk and ωnk are sets of 1st and 2nd type in the interval [p2nk , p1nk ] and
x(t) ∈ int D, ∀t ∈ E k .
If the conditions of Lemma 12 hold then (5.29) is true if we take
We have x(t1k ) = x(t2k ) = x∗ (see Lemma 4) and therefore π0k and ω0k are sets
of 1st and 2nd type in the interval [p2 , p1 ]. Thus (5.29) is true.
Now we apply Lemmas 8–12 to the intervals [0, t1 ] and [t2 , T ]. We have
t
1
T
u(x(t))dt = u(x(t))dt + u(x(t))dt + u(x(t))dt. (5.31)
n T ∪ω T
t2 πn n FT ET
Here
• F 0 and F T are sets of 1st type (they may be empty);
• [0, t1G0 ] ∪ [t2G0 , t1 ] ⊂ E 0 and [t2 , t1GT ] ∪ [t2GT , T ] ⊂ E T ;
• x(t) ∈ / int D for all t ∈ E 0 ∪ E T .
Thus, applying Lemma 4 and taking into account (5.29)–(5.31), we can
prove the following lemma.
Lemma 13. The interval [0, T ] can be divided into subintervals such that
5 Asymptotical stability of optimal paths in nonconvex problems 121
Here
1. The sets πn and ωn are sets of 1st and 2nd type, respectively, in the inter-
vals [p2n , p1n ], n = 1, 2, ....
2. The sets F1 and F2 are sets of 1st type in the intervals [p21 , p11 ] and [p22 , p12 ],
respectively, and
3. Also
x(t) ∈
/ int D, f or all t ∈ E. (5.36)
where the number C(δ) < ∞ does not depend on the trajectory x(t), on T
or on the intervals in (5.32).
Proof: We define
F1 = {t ∈ F 0 : x(t) ∈ int D}, F2 = {t ∈ F T : x(t) ∈ int D}
and E = ∪k E k ∪ E 0 ∪ E T . Then we replace π10 to π10 ∪ (F 0 \ F1 ) and π1T
to π1T ∪ (F T \ F2 ) in (5.30) and (5.31) (note that after these replacements
the conditions of Definition 6 still hold). We obtain (5.33) summing (5.29)–
(5.31). It is not difficult to see that all assertions of the lemma hold. Note that
(5.35) follows from the fact that the trajectory x(t) is uniformly bounded (see
(5.3)). The inequality (5.37) follows from Lemma 17.4.1, taking into account
Definition 6, and thus the lemma is proved.
Proof: From Condition H3 it follows that there exist a number ε > 0 and a
∼ ∼
trajectory x (·) to the system (6.1), defined on [ 0, Tε ], such that p x (0) =
∼
p∗ − ε, p x (Tε ) = p∗ + ε and
122 M.A. Mamedov
.
∼
p x (t) > 0 for almost all t ∈ [0, Tε ]. (5.39)
Define
∼
Rε = u(x (t))dt.
[0,Tε ]
Consider the set F1 and corresponding interval [p21 , p11 ]. Define a set
u(x(t)) dt ≤ Cε , (5.42)
F1 \F1ε
where the number Cε < +∞ does not depend on T or on the trajectory x(t).
Denote C = Tε u∗ + Cε − Rε . Then from (5.40)–(5.42) we obtain
u(x(t)) dt ≤ u∗ meas F1ε + C ≤ u∗ meas F1 + C
F1
and therefore
[u(x(t)) − u∗ ] dt ≤ C .
F1
5 Asymptotical stability of optimal paths in nonconvex problems 123
From Condition M it follows that for every T > 0 there exists a trajectory
xT (·) ∈ XT , for which
u(xT (t)) dt ≥ u∗ T − b. (5.43)
[0,T ]
5.7.1
that is,
JT (x(·)) − JT (xT (·)) ≤ − [u∗ − u(x(t))] dt − δ 2 (x(t)) dt + L + b.
Q A
(5.44)
Here
Q ∪ A = { t ∈ [0, T ] : x(t) ∈
/ int D}; (5.46)
b)
c) for every δ > 0 there exist K(δ) < +∞ and C(δ) < +∞ such that
The first assertion of the theorem follows from (5.44), (5.46) and (5.50) for
the case under consideration (that is, x(t) continuously differentiable). We
now prove the second assertion.
Let ε > 0 and δ > 0 be given numbers and x(·) a continuously differentiable
ξ-optimal trajectory. We denote
Assume that (5.51) is not true. In this case there exist sequences Tk → ∞,
k
Kε,ξ → ∞ and sequences of trajectories {xk (·)} (every xk (·) is a ξ-optimal
trajectory in the interval [0, Tk ]) and {xTk (·)} (satisfying (5.43) for every
T = Tk ) such that
meas Zδ ≤ Kδ,ξ
1
. (5.53)
We denote Xε/2 0
= {t ∈ [0, T ] : ||x(t) − x∗ || > ε/2.}. Clearly Xε/2
0
is an
open set and therefore can be presented as a union of an at most countable
∼
number of open intervals, say Xε/2 0
= ∪k τ k . Out of these intervals we
choose further intervals, which have a nonempty intersection with Xε , say
these are τk , k = 1, 2, .... Then we have
Xε ⊂ ∪k τk ⊂ Xε/2
0
. (5.54)
But the interval [0, T ] is bounded and therefore the number of intervals
τk is finite too. Let k = 1, 2, 3, ..., NT (ε). We divide every interval τk into
two parts:
∪k τk2 ⊂ (Q ∪ A) ∩ Xε/2
0
Define p1k = supt∈τk px(t) and p2k = inf t∈τk px(t). It is clear that
∼
p1k − p2k ≤ C , k = 1, 2, 3, ..., NT (ε), (5.58)
and
·
|p x (t)| ≤ K, for all t. (5.59)
∼
Here the numbers C and K do not depend on T > 0, x(·), ε or ξ. We divide
the interval τk into three parts:
· ·
τk− = {t ∈ τk : p x (t) < 0}, τk0 = {t ∈ τk : p x (t) = 0} and
·
τk+ = {t ∈ τk : p x (t) > 0}.
5 Asymptotical stability of optimal paths in nonconvex problems 127
Then we have
* * * *
*
* *
*
* * * *
· * · · *
p1k − p2k ≥ ** p x (t)dt ** = * p x (t)dt + p x (t)dt * .
* * * *
τk * τ− τ+
*
k k
· ·
We define α=− p x (t)dt and β = p x (t)dt. Clearly α > 0, β > 0
τk− τk+
and
)
−α + β, if α < β,
p1k − p2k ≥ (5.60)
α − β, if α ≥ β.
On the other hand, τk1 ⊂ τk− and therefore from (5.57) we have
or
∼
meas τk1 < Cε,ξ , where Cε,ξ = K· K ε/2,ξ /ηε/2 . (5.65)
and then
Now we show that for every ε > 0 and ξ > 0 there exists a number
Kε,ξ < +∞ such that
meas (∪k τk1 ) ≤ Kε,ξ . (5.67)
Assume that (5.67) is not true. Then from (5.66) it follows that NT (ε) → ∞
as T → ∞. Consider the intervals τk for which the following conditions hold:
1
meas τk1 ≥ σε and meas τk2 ≤ λ meas τk1 , (5.68)
2
where λ is any fixed number. Since NT (ε) → ∞, then from (5.55) and (5.56)
it follows that the number of intervals τk satisfying (5.68) increases infinitely
as T → ∞.
On the other hand, the number of intervals τk , for which the conditions
α < β,
meas τk2 > λ meas τk1 and λ = ηε/2 /K
hold, is finite. Therefore the number of of intervals τk for which the conditions
α ≤ β and (5.68) hold infinitely increases as T → ∞. We denote the number
of such intervals by NT and for the sake of definiteness assume that these are
intervals τk , k = 1, 2, ..., NT .
We set λ = ηε/2 /2K for every τk . Then from (5.63) and (5.68) we have
ηε/2 1
p1k − p2k ≥ ηε/2 meas τk1 − K· meas τk1 = ηε/2 meas τk1 .
2K 2
Taking (5.55) into account we obtain
where
1
eε = ηε/2 σε > 0 and NT → ∞ as T → ∞.
2
1
Let δ = 8 eε . From (5.69) it follows that for every τk there exists an interval
d
dk = [sk , sk ] ⊂ τk such that
1 2
NT
δ
meas Zδ ≥ meas ∪N
k=1 dk =
T
meas dk ≥ NT .
K
k=1
5.7.2
We now take any trajectory x(·) to the system (6.1). It is known (see, for
example, [3]) that for a given number δ > 0 (we take δ < ε/2) there exists a
∼
continuously differentiable trajectory x (·) to the system (6.1) such that
∼
|| x(t)− x (t)|| ≤ δ for all t ∈ [0, T ].
Since the function u is continuous then there exists η(δ) > 0 such that
∼
u(x (t)) ≥ u(x(t)) − η(δ) for all t ∈ [0, T ].
Therefore
∼
u(x (t)) dt ≥ u(x(t)) dt − T η(δ).
[0,T ] [0,T ]
Let ξ > 0 be a given number. For every T > 0 we choose a number δ such
that T η(δ) ≤ ξ. Then
∼ ∼
u(x(t)) dt ≤ u(x (t)) dt + T η(δ) ≤ u(x (t)) dt + ξ, (5.70)
[0,T ] [0,T ] [0,T ]
130 M.A. Mamedov
that is,
∗ ∼
[ u(x(t)) − u ] dt ≤ [ u(x (t)) − u∗ ] dt + ξ.
[0,T ] [0,T ]
∼
Since the function x (·) is continuously differentiable then the second integral
in this inequality is bounded (see the first part of the proof), and therefore
the first assertion of the theorem is proved.
Now we prove the second assertion of Theorem 13.6. We will use (5.70).
Take a number ε > 0 and assume that x(·) is a ξ-optimal trajectory, that is,
JT (x(·)) ≥ JT∗ − ξ.
∼
JT (x (·)) ≥ JT (x(·)) − ξ ≥ JT∗ − 2ξ.
∼
Thus x (·) is a continuously differentiable 2ξ-optimal trajectory. That is why
(see the first part of the proof) for the numbers ε/2 > 0 and 2ξ > 0 there
exists Kε,ξ < +∞ such that
∼
meas { t ∈ [0, T ] : || x (t) − x∗ || ≥ ε/2} ≤ Kε,ξ .
∼
{ t ∈ [0, T ] : || x(t) − x∗ || ≥ ε} ⊂ { t ∈ [0, T ] : || x (t) − x∗ || ≥ ε/2},
which implies that the proof of the second assertion of the theorem is com-
pleted, that is,
Assume that the third assertion of the theorem is not true, that is, there
is a point t ∈ (t1 , t2 ) such that ||x(t ) − x∗ || = c > 0.
Consider the function x(·). In [3] it is proved that there is a sequence of
continuously differentiable trajectories xn (·), t ∈ [t1 , T ], which is uniformly
convergent to x(·) on [t1 , T ] and for which xn (t1 ) = x(t1 ) = x∗ . That is, for
every δ > 0 there exists a number Nδ such that
On the other hand, for every δ > 0 there exists a number η(δ) > 0 such that
η(δ) → 0 as δ → 0 and
Then we have
u(x(t)) dt ≤ u(xn (t)) dt + T η(δ). (5.72)
[t1 ,T ] [t1 ,T ]
∼
We take a number δ < c/2. Then there exists a number β > 0 such that
∼
meas [∪k (Qnk ∪ Ekn )] ≥ β .
u(xn (t)) dt ≤ u∗ meas [∪k (πkn ∪ ωkn )] − β.
k π n ∪ω n
k k
+ αn − β
or
u(xn (t)) dt ≤ u∗ (tn − t1 ) + αn − β. (5.74)
[t1 ,tn ]
+ αn − β + λn + 2T η(δ)
= u(x∗ (t)) dt + αn − β + λn + 2T η(δ).
[t1 ,T ]
Here
λn = [ u(xn (t)) − u∗ ] dt → 0 as n → ∞,
[tn ,t2 ]
because tn → t2 . We choose the numbers δ > 0 and n such that the following
inequality holds:
αn + λn + 2T η(δ) < β.
In this case we have
u(x(t)) dt < u(x∗ (t)) dt
[t1 ,T ] [t1 ,T ]
and therefore
u(x(t)) dt < u(x∗ (t)) dt,
[0,T ] [0,T ]
References
1. D. Cass and K. Shell, The structure and stability of competitive dynamical systems,
J. Econom. Theory, 12 (1976), 31–70.
2. A. N. Kolmogorov and S. V. Fomin, Introductory Real Analysis (Moscow, Nauka,
1975).
134 M.A. Mamedov
B. D. Craven
6.1 Introduction
Pontryagin’s principle has been proved in at least four ways, for an optimal
control problem in continuous time with dynamics described by an ordinary
differential equation (ODE). One approach ([5], [6]) regards the control prob-
lem as a mathematical program, and uses the Karush–Kuhn–Tucker (KKT)
necessary conditions as the starting point (though with some different hy-
potheses) for deriving the Pontryagin theory. There are various results for
optimal control when the dynamics are described by a partial differential
equation (PDE), often derived (as, for example, by Lions and Bensoussan)
using variational inequalities, which are generally equivalent to mathemat-
ical programs in infinite dimensions. The results in [1]–[5], and others by
the same authors, obtain some versions of Pontryagin’s principle by quite
different methods to those used for ODEs. However, the Pontryagin theory
involving a PDE can also be derived from the mathematical programming
approach, using the KKT conditions, and replacing the time variable t by a
space variable z, say in R2 or R3 , or by (t, z) combined. Whatever approach
B. D. Craven
Department of Mathematics, University of Melbourne, Victoria 3010, AUSTRALIA
e-mail: [email protected]
Here x(.) is the state function, u(.) is the control function, the time interval
[0, T ] is fixed, f and m are differentiable functions. Other details such as
variable horizon T, an endpoint constraint on x(T ), and state constraints, can
readily be added to the problem. They are omitted here, since the purpose
is to show the method. The steps are as follows.
(a) The problem (6.1) is expressed as a mathematical program:
and
T
H(x, u, λ̂) := F (x, u) + λ̂M (x, u) = h(x(t), u(t), t, λ̄(t))dt.
0
with a similar requirement for M . Then minimality of (x̄, ū), namely that
F (x, u) − F (x̄, ū) ≥ 0, with (6.2), leads (see [7], Theorem 7.2.3) to
describing a quasimin (see [6]) of H(x̄, ., λ̂) over Γ (.) at ū. (Note that there
is no requirement of convexity on Γ (.).)
(e) Assuming that ū is a minimum in terms of the L1 norm, suppose if
possible that
h(x̄(t), u(t), t, λ(t)) < h(x̄(t), ū(t), t, λ(t))
for t in a set of positive measure. Then (see Note 3 in the Appendix) a set
of control functions {uβ (.) : β ≥ 0} ⊂ Γ is constructed (see [7], Theorem
7.2.6), for which
H(x̄, u, λ̂) − H(x̄, ū, λ̂) ≤ c u − ū
for some constant c > 0, thus contradicting (6.6). (A required chattering
property holds automatically for the considered control constraint.) This has
proved Pontryagin’s principle, in the following form.
Theorem 1. Let the control problem (6.1) reach a local minimum at (x, u) =
(x̄, ū) with respect to the L1 -norm for the control u. Assume that the differen-
tial equation Dx−M (x, u) determines x as a Lipschitz function of u, that the
differentiability property (6.5) (with respect to x) holds, and that −DM (x̄, ū)
is surjective, Then necessary conditions for the minimum are that the costate
138 B.D. Craven
λ(.) satisfies the adjoint equation (6.4), and that h(x̄(t), ., t, λ̄(t)) is minimized
over Γ (t) at ū(t), for almost all t.
subject to
The steps of Section 6.2 are now applied, but replacing t ∈ [0, T ] by z ∈ Ω. It
is observed that steps (a), (b), (d) and (e) remain valid – they do not depend
on t ∈ R. Step (c) requires a replacement for integration by parts. If D = ∇2 ,
it is appropriate to use Green’s theorem in the form
[λ∇ x − x∇ λ]dv =
2 2
[λ(∂x/∂n) − x(∂λ/∂n)]ds,
Ω ∂Ω
in which dv and ds denote elements of volume and surface. The right side of
(6.10) beomes the integrated part; the origin can be shifted, in the spaces of
6 Pontryagin principle with a PDE: a unified approach 139
with boundary conditions (6.11), where D∗ denotes the adjoint linear oper-
ator to D. Here, with D = ∇2 , (6.10) shows that D∗ = ∇2 also. Then (e),
with z ∈ Ω replacing t ∈ [0, T ], gives Pontryagin’s principle in the form:
h(x̄(z), ., t, λ̄(z)) is minimized over Γ (z) at ū(z), possibly except for a set of
z of zero measure.
If f and m happen to be linear in u, and if Γ (z) is a polyhedron with
vertices pi (or an interval if u(z) ∈ R), then Pontryagin’s principle may lead
to bang-bang control, namely u(z) = pi when z ∈ Ei , for some disjoint sets
Ei ⊂ Ω.
where ∇2z acts on the variable z. Here t (for the ODE) has been replaced by
3
(t, z) ∈ [0, T ] × Ω, for a closed bounded region Ω ⊂ R , and where m(.) is a
forcing function. Define the linear differential operator D := (∂/∂t) − c2 ∇2z .
The function spaces must be chosen so that D is a continuous linear mapping.
Define Ai ⊂ ∂Ω as in Section 6.3. The optimal control problem now becomes
(with a certain choice of boundary conditions)
T
M INx(.),u(.) J(u) := f (x(z, t), u(z, t), z, t)dtdz
0 Ω
subject to
Then steps (a), (b), (d) and (e) proceed as in Section 6.3, for an elliptic PDE.
The Hamiltonian is
6.5 Appendix
x := x ∗ + Dx ∗ ,
6 Pontryagin principle with a PDE: a unified approach 141
Acknowledgments The author thanks two referees for pointing out ambiguities and
omissions.
References
Alexander J. Zaslavski
7.1 Introduction
Let X be a metric space and let ρ(·, ·) be the metric on X. For the set X × X
we define a metric ρ1 (·, ·) by
Let Z be the set of all integers. Denote by M the set of all sequences
of functions v = {vi }∞
i=−∞ where vi : X → R is bounded from below for
1
∞
each i ∈ Z. Such a sequence of functions {vi }i=−∞ ∈ M will occasionally be
denoted by a boldface v (similarly {ui }∞
i=−∞ will be denoted by u, etc.)
The set M is equipped with the metric d defined by
˜ u) = sup{|vi (x, y) − ui (x, y)| : (x, y) ∈ X × X, i ∈ Z},
d(v, (1.1)
˜ u))−1 , u, v ∈ M.
˜ u)(1 + d(v,
d(v, u) = d(v,
In this paper we investigate the structure of “approximate” solutions of
the optimization problem
Alexander J. Zaslavski
Department of Mathematics, The Technion–Israel Institute of Technology, Haifa, Israel
2 −1
k
vi (xi , xi+1 ) → min, {xi }ki=k
2
1
⊂ X, xk1 = y, xk2 = z (P )
i=k1
σ(v, m1 , m2 , z1 , z2 ) =
m −1 -
2
2 −1
k
vi (xi , xi+1 ) ≤ σ(v, k1 , k2 , y, z) + δ.
i=k1
satisfying
m 2 −1
ρ(xi , x̄i ) ≤ , i = τ1 , . . . , τ2 .
of the problem (P) with y = x̄k1 and z = x̄k2 . In Section 4 we show that
under certain assumptions the turnpike property is equivalent to its weakened
version.
Theorem 1. Assume that v = {vi }∞ i=−∞ ∈ M has the turnpike property and
{x̄i }∞
i=−∞ ⊂ X is the turnpike for v. Then the following property holds:
For each > 0 there exist δ > 0, a natural number N and a neighborhood
U of v in M such that for each u ∈ U, each pair of integers m1 , m2 satisfying
m2 ≥ m1 + 2N and each sequence {xi }m i=m1 ⊂ X satisfying
2
m 2 −1
Proof. Let > 0. It follows from the property (TP) that there exist
δ0 ∈ (0, /4) (2.3)
m 2 −1
m 2 −1
Set
N = 4(N1 + N0 ) (2.7)
and
Assume that u ∈ U, m1 , m2 ∈ Z, m2 ≥ m1 + 2N ,
m 2 −1
{xi }m 2
i=m1 ⊂ X and ui (xi , xi+1 ) ≤ σ(u, m1 , m2 , xm1 , xm2 ) + δ. (2.9)
i=m1
Let
k ∈ {m1 , . . . , m2 }, m2 − k ≥ 2N. (2.10)
(2.9) implies that
−1
k+2N
ui (xi , xi+1 ) ≤ σ(u, k, k + 2N, xk , xk+2N ) + δ. (2.11)
i=k
−1
k+2N −1
k+2N
vi (xi , xi+1 ) ≤ ui (xi , xi+1 ) + δ/4
i=k i=k
≤ δ/4 + σ(u, k, k + 2N, xk , xk+2N ) + δ
≤ σ(v, k, k + 2N, xk , xk+2N ) + δ + δ/4 + δ/4. (2.12)
148 A.J. Zaslavski
We have that (2.12) holds for any k satisfying (2.10). This fact implies
that
2 −1
k
vi (xi , xi+1 ) ≤ σ(v, k1 , k2 , xk1 , xk2 ) + 2−1 · 3δ (2.13)
i=k1
k1 < k2 ≤ k1 + 2N.
It follows from (2.13), (2.7) and the property (P2) that for any integer k ∈
{m1 , . . . , m2 } satisfying m2 − k ≥ 2N0 + 2N1 there exists an integer q such
that q − k ∈ [2N0 , 2N0 + 2N1 ] and
ρ(xq , x̄q ) ≤ δ0 .
This fact implies that there exists a finite strictly increasing sequence of
integers {τj }sj=0 such that
m 2 −1
It follows from (2.19), (2.20), (2.18), (2.14) and the property (P1) that
m 2 −1
Proof. Assume the contrary. Then there exist a pair of integers m̄1 , m̄2 > m̄1 ,
a sequence {xi }m̄
i=m̄1 and a number Δ > 0 such that
2
m̄ 2 −1
m̄ 2 −1
There exists ∈ (0, Δ/4) such that the following property holds:
(P3) if i ∈ {m̄1 − 1, . . . , m̄2 + 1}, z1 , z2 ∈ X and
then
|vi (x̄i , x̄i+1 ) − vi (z1 , z2 )| ≤ Δ[64(m̄2 − m̄1 + 4)]−1 . (3.2)
By the property (TP) there exist δ ∈ (0, /4) and a natural number N such
that for each pair of integers n1 , n2 ≥ n1 + 2N and each sequence {yi }ni=n
2
1
⊂
X satisfying
2 −1
n
vi (yi , yi+1 ) ≤ σ(v, n1 , n2 , yn1 , yn2 ) + δ (3.3)
i=n1
and
m̄
2 +4N
vi (ȳi , ȳi+1 ) ≤ σ(v, m̄1 − 4N, m̄2 + 4N, ȳm̄1 −4N , ȳm̄2 +4N ) + δ/8.
i=m̄1 −4N
(3.6)
yi = xi , i ∈ {m̄1 , . . . , m̄2 }.
We will estimate
−1
m̄2 +4N −1
m̄2 +4N
vi (ȳi , ȳi+1 ) − vi (yi , yi+1 ).
i=m̄1 −4N i=m̄1 −4N
−1
m̄2 +4N −1
m̄2 +4N
vi (ȳi , ȳi+1 ) − vi (yi , yi+1 ) (3.9)
i=m̄1 −4N i=m̄1 −4N
m̄2
= [vi (ȳi , ȳi+1 ) − vi (yi , yi+1 )]
i=m̄1 −1
m̄ 2 −1
|vm̄1 −1 (ȳm̄1 −1 , ȳm̄1 ) − vm̄1 −1 (ȳm̄1 −1 , x̄m̄1 )| ≤ 2Δ[64(m̄2 − m̄1 + 4)]−1 , (3.10)
|vm̄2 (ȳm̄2 , ȳm̄2 +1 ) − vm̄2 (x̄m̄2 , ȳm̄2 +1 )| ≤ 2Δ[64(m̄2 − m̄1 + 4)]−1 , (3.11)
|vi (ȳi , ȳi+1 ) − vi (x̄i , x̄i+1 )| ≤ Δ[64(m̄2 − m̄1 + 4)]−1 , i = m̄1 , . . . , m̄2 − 1.
(3.12)
It follows from (3.9), (3.12), (3.11) and (3.1) that
−1
m̄2 +4N
[vi (ȳi , ȳi+1 ) − vi (yi , yi+1 )]
i=m̄1 −4N
Combined with (3.8) this fact contradicts (3.6). The contradiction we have
reached proves the theorem.
In this section we show that under certain assumptions the turnpike property
is equivalent to its weakened version.
m 2 −1
for each pair of integers m1 , m2 > m1 . Assume that the following two prop-
erties hold:
(i) For any > 0 there exists δ > 0 such that for each i ∈ Z, each
x1 , x2 , y1 , y2 ∈ X satisfying ρ(xj , yj ) ≤ δ, j = 1, 2,
(ii) for each > 0 there exist δ > 0 and a natural number N such that for
each pair of integers m1 , m2 ≥ m1 + 2N and each sequence {xi }m i=m1 ⊂ X
2
satisfying
152 A.J. Zaslavski
m 2 −1
the inequality
ρ(xi , x̄i ) ≤ , i = m1 + N, . . . , m2 − N
m+k−1
vi (xi , xi+1 ) ≤ σ(v, m, m + k, xm , xm+k ) + δ
i=m
the inequality
ρ(xi , x̄i ) ≤ , i = m, . . . , m + k (4.4)
holds.
Let > 0. There exists 0 ∈ (0, /2) and a natural number N such that for
each pair of integers m1 , m2 ≥ m1 + 2N and each sequence {xi }m i=m1 ⊂ X
2
satisfying
m 2 −1
the inequality
ρ(xi , x̄i ) ≤ , i = m1 + N, . . . , m2 − N (4.5)
holds.
By the property (i) there is δ ∈ (0, 0 /8) such that for each integer i
and each z1 , z2 , y1 , y2 ∈ X satisfying ρ(zj , yj ) ≤ δ, j = 1, 2 the following
inequality holds:
|vi (z1 , z2 ) − vi (y1 , y2 )| ≤ 0 /16. (4.6)
m+k−1
vi (xi , xi+1 )
i=m
m+k−1
≤ vi (zi , zi+1 ) + δ
i=m
m+k−1
≤δ+ vi (x̄i , x̄i+1 ) + |vm (x̄m , x̄m+1 ) − vm (xm , x̄m+1 )|
i=m
+ |vm+k−1 (x̄m+k−1 , x̄m+k ) − vm+k−1 (x̄m+k−1 , xm+k )|.
Combined with (4.3) and the definition of δ (see (4.6)) this inequality
implies that
m+k−1
m+k−1
vi (xi , xi+1 ) ≤ δ + 0 /8 + vi (x̄i , x̄i+1 ). (4.8)
i=m i=m
yi = xi , i ∈ {m, . . . , m + k}.
It follows from (4.9), (4.3) and the definition of δ (see (4.6)) that
and
|vm+k (x̄m+k , x̄m+k+1 ) − vm+k (ym+k , ym+k+1 )| ≤ 0 /16.
m+k
0
vi (yi , yi+1 ) ≤ vm−1 (x̄m−1 , x̄m ) + + vm+k (x̄m+k , x̄m+k+1 )
i=m−1
16
0
m+k−1
+ + vi (xi , xi+1 )
16 i=m
≤ 0 /8 + vm−1 (x̄m−1 , x̄m ) + vm+k (x̄m+k , x̄m+k+1 ) + δ
m+k−1
+ 0 /8 + vi (x̄i , x̄i+1 )
i=m
m+k
< 0 /2 + vi (x̄i , x̄i+1 ). (4.10)
i=m−1
154 A.J. Zaslavski
−1
m+k+2N
vi (yi , yi+1 )
i=m−2N
m−2
m+k −1
m+k+2N
= vi (yi , yi+1 ) + vi (yi , yi+1 ) + vi (yi , yi+1 )
i=m−2N i=m−1 i=m+k+1
m−2
m+k
≤ vi (x̄i , x̄i+1 ) + 0 /2 + vi (x̄i , x̄i+1 )
i=m−2N i=m−1
−1
m+k+2N
+ vi (x̄i , x̄i+1 )
i=m+k+1
−1
m+k+2N
= 0 /2 + vi (x̄i , x̄i+1 )
i=m−2N
Thus
−1
m+k+2N
vi (yi , yi+1 ) ≤ 0 /2 + σ(v, m − 2N, m + k + 2N, ym−2N , ym+k+2N ).
i=m−2N
ρ(yi , x̄i ) ≤ , i = m − N, . . . , m + k + N.
ρ(xi , x̄i ) ≤ , i = m, . . . , m + k.
Thus we have shown that the property (C) holds. Now we are ready to com-
plete the proof.
Let > 0. By the property (C) there exists δ0 ∈ (0, ) such that for each
pair of integers m1 , m2 > m1 and each sequence {xi }m
i=m1 ⊂ X satisfying
2
ρ(xi , x̄i ) ≤ δ0 , i = m1 , m2
m2 −1
There exist a number δ ∈ (0, δ0 ) and a natural number N such that for each
pair of integers m1 , m2 ≥ m1 + 2N and each sequence {xi }m i=m1 ⊂ X which
2
satisfies
m2 −1
(4.13). Then (4.14) is valid. Assume that ρ(xm1 , x̄m1 ) ≤ δ. Then by (4.14)
and (4.13),
ρ(xm1 +N , x̄m1 +N ) ≤ δ0
and
+N −1
m1
vi (xi , xi+1 ) ≤ σ(v, m1 , m1 + N, xm1 , xm1 +N ) + δ.
i=m1
It follows from these relations and the definition of δ0 (see (4.11) and (4.12))
that
ρ(xi , x̄i ) ≤ , i = m1 , . . . , m1 + N.
Analogously we can show that is ρ(xm2 , x̄m2 ) ≤ δ, then
ρ(xi , x̄i ) ≤ , i = m2 − N, . . . , m2 .
References
[Dzalilov et al. (2001)] Dzalilov, Z., Ivanov, A.F. and Rubinov, A.M. (2001) Difference
inclusions with delay of economic growth, Dynam. Systems Appl., Vol. 10, pp. 283–293.
[Dzalilov et al. (1998)] Dzalilov, Z., Rubinov, A.M. and Kloeden, P.E. (1998) Lyapunov
sequences and a turnpike theorem without convexity, Set-Valued Analysis, Vol. 6,
pp. 277–302.
[Leizarowitz (1985)] Leizarowitz, A. (1985) Infinite horizon autonomous systems with un-
bounded cost, Appl. Math. and Opt., Vol. 13, pp. 19–43.
[Leizarowitz (1989)] Leizarowitz, A. (1989) Optimal trajectories on infinite horizon de-
terministic control systems, Appl. Math. and Opt., Vol. 19, pp. 11–32.
[Leizarowitz and Mizel (1989)] Leizarowitz, A. and Mizel, V.J. (1989) One dimensional
infinite horizon variational problems arising in continuum mechanics, Arch. Rational
Mech. Anal., Vol. 106, pp. 161–194.
156 A.J. Zaslavski
[Makarov, Levin and Rubinov (1995)] Makarov, V.L, Levin, M.J. and Rubinov, A.M.
(1995) Mathematical economic theory: pure and mixed types of economic mechanisms,
North-Holland, Amsterdam.
[Makarov and Rubinov (1973)] Makarov, V.L. and Rubinov, A.M. (1973) Mathematical
theory of economic dynamics and equilibria, Nauka, Moscow, English trans. (1977):
Springer-Verlag, New York.
[Mamedov and Pehlivan (2000)] Mamedov, M.A. and Pehlivan, S. (2000) Statistical con-
vergence of optimal paths, Math. Japon., Vol. 52, pp. 51–55.
[Mamedov and Pehlivan (2001)] Mamedov, M.A. and Pehlivan, S. (2001) Statistical clus-
ter points and turnpike theorem in nonconvex problems, J. Math. Anal. Appl., Vol. 256,
pp. 686–693.
[Marcus and Zaslavski (1999)] Marcus, M. and Zaslavski, A.J. (1999) The structure of
extremals of a class of second order variational problems, Ann. Inst. H. Poincare, Anal.
non lineare, Vol. 16, pp. 593–629.
[McKenzie (1976)] McKenzie, L.W. (1976) Turnpike theory, Econometrica, Vol. 44,
pp. 841–866.
[Radner (1961)] Radner, R. (1961) Path of economic growth that are optimal with regard
only to final states; a turnpike theorem, Rev. Econom. Stud., Vol. 28, pp. 98–104.
[Rubinov (1980)] Rubinov, A.M. (1980) Superlinear multivalued mappings and their ap-
plications to problems of mathematical economics, Nauka, Leningrad.
[Rubinov (1984)] Rubinov, A.M. (1984) Economic dynamics, J. Soviet Math., Vol. 26,
pp. 1975–2012.
[Samuelson (1965)] Samuelson, P.A. (1965) A catenary turnpike theorem involving con-
sumption and the golden rule, American Economic Review, Vol. 55, pp. 486–496.
[Zaslavski (1995)] Zaslavski, A.J. (1995) Optimal programs on infinite horizon, 1 and 2,
SIAM Journal on Control and Optimization, Vol. 33, pp. 1643–1686.
[Zaslavski (1996)] Zaslavski, A.J. (1996) Dynamic properties of optimal solutions of
variational problems, Nonlinear Analysis: Theory, Methods and Applications, Vol. 27,
pp. 895–932.
[Zaslavski (2000)] Zaslavski, A.J. (2000) Turnpike theorem for nonautonomous infinite
dimensional discrete-time control systems, Optimization, Vol. 48, pp. 69–92.
Chapter 8
Mond–Weir Duality
B. Mond
8.1 Preliminaries
One of the most interesting and useful aspects of linear programming is du-
ality theory. Thus to the problem minimize ct x subject to Ax ≥ b, x ≥ 0
(when A is an m × n matrix) there corresponds the dual problem maximize
bt y subject to At y ≤ c, y ≥ 0. Duality theory says that for any feasible x
and y, ct x ≥ bt y; and, if x0 is optimal for the primal problem, there exists
an optimal y0 of the dual and ct x0 = bt y0 .
B. Mond
Department of Mathematics and Statistical Sciences, La Trobe University, Victoria 3086,
AUSTRALIA; Department of Mathematics and Statistics, University of Melbourne
Victoria 3052, AUSTRALIA
e-mail: [email protected] and [email protected]
matrix. Dorn [4] showed that this problem was dual to the quadratic problem
maximize − 21 ut Cu + bt y subject to At y ≤ Cu + p, y ≥ 0. Weak duality holds
and if x0 is optimal for the primal, there exists y0 such that (u = x0 , y0 )
is optimal for the dual with equality of objective functions. The require-
ment that C be positive semi-definite ensures that the objective function is
convex.
It can be shown [6] that if f is convex, ≥ 0 and g concave, > 0 (or f convex, g
linear > 0) then f /g is pseudo-convex. It follows from this that the objective
function in the (LFP) is pseudo-convex.
Mangasarian [7] points out that whereas some results (such as sufficiency
and converse duality) hold if, in (P), f is only pseudo-convex and g quasi-
convex, Wolfe duality does not hold for such functions. One example is the
following:
minimize x3 + x subject to x ≥ 1, which has the optimal value 2 at x = 1.
The Wolfe dual
maximize u3 + u + y(1 − u)
3u2 + 1 − y = 0, y ≥ 0
In order to weaken the convexity requirements, Mond and Weir [12] proposed
a different dual to (P).
(MWD) maximize f (u)
subject to ∇f (u) + ∇y t g(u) = 0,
y t g(u) ≥ 0, y ≥ 0.
It is easy to see that if also x0 is optimal for (P) and a constraint qualifi-
cation is satisfied, then there exists a y0 such that (u = x0 , y0 ) is optimal for
(MWD) with equality of objective functions.
Consider again the problem minimize x3 + x subject to x ≥ 1, to which
Wolfe duality does not apply. The corresponding Mond–Weir dual maximize
u3 + u subject to 3u2 + 1 − y = 0, y(1 − u) ≥ 0, y ≥ 0, has an optimal at
u = 1, y = 4 with optimum value equal to 2.
Although many variants of (MWD) are possible (see [12]), we give a dual
that can be regarded as a combination of (WD) and (MWD). Let M =
{1, 2, . . . , m} and I ⊆ M .
max f (u) + yi gi (u)
i∈I
∇f (u) + ∇y t g(u) = 0, y ≥ 0
yi gi (u) ≥ 0.
i∈M/I
Weak duality holds if f + yi gi is pseudo-convex and yi gi is quasi-
i∈I i∈M/I
convex.
8.5 Applications
max f (u)/g(u)
subject to ∇[f (u)/g(u) + y t h(u)] = 0
y t h(u) ≥ 0, y ≥ 0.
minimize f (x)/g(x)
h(x)
subject to ≤ 0.
g(x)
t
Here the Lagrangian is f (x)+y
g(x)
h(x)
and is psuedo-convex if f and h are
convex, g is concave > 0, f + y h ≥ 0 (unless g is linear). Thus his dual to
t
(FP) is
f (u) + y t h(u)
maximize
g(u)
+ ,
f (u) + y t h(u)
subject to ∇ =0
g(u)
f (u) + y t h(u) ≥ 0, y ≥ 0.
maximize λ
∇f (u) − λ∇g(u) + ∇y t h(u) = 0
f (u) − λg(u) + y t h(u) ≥ 0
y ≥ 0, λ ≥ 0.
maximize λ
∇f (u) − λ∇g(u) + ∇y t h(u) = 0
f (u) − λg(u) ≥ 0
y t h(u) ≥ 0, λ ≥ 0, y ≥ 0.
maximize λ
∇f (u) − λ∇g(u) + ∇y t h(u) = 0
f (u) − λg(u) + yi hi (u) ≥ 0
i∈I
yi hi (u) ≥ 0, λ ≥ 0, y ≥ 0
i∈M/I
Here yi hi (u) need only be quasi-convex for duality to hold.
i∈M/I
A fractional programming problem where Bector and Schaible duality do
not hold but the Mond–Weir fractional programming duals are applicable is
the following:
1
minimize − subject to x3 ≥ 1.
x>0 x
Here neither the Bector nor the Schaible dual is applicable. The Mond–
Weir Bector type dual is
1
maximize −
u>0 u
1
y = 4 , y(1 − u3 ) ≥ 0, y ≥ 0.
3u
The maximum value −1 is attained at u = 1, y = 13 .
The Mond–Weir Schaible type dual is
maximize λ
subject to − λ − 3yu2 = 0
−1 − λu ≥ 0
y(1 − u3 ) ≥ 0, y ≥ 0
u = 1, y = 1/3, λ = −1.
In [11] Mond established weak duality between (P) and (MD) under the
following conditions: If for all x, u, p
1
f (x) − f (u) ≥ (x − u)t ∇f (u) + (x − u)t ∇2 f (u)p − pt ∇2 f (u)p
2
(subsequently called second order convex by Mahajan [5]) and
1
gi (x) − gi (u) ≥ (x − u)t ∇gi (u) + (x − u)t ∇2 gi (u)p − pt ∇2 gi (u)p
2
i = 1, . . . , m
In [3], Dantzig, Eisenberg and Cottle formulated the following pair of sym-
metric dual problems:
Assuming that φ(·, y) and ψ(x, ·) are convex while φ(x, ·) and ψ(·, y) are
concave, then the objective function is pseudo-convex in x for fixed y and
pseudo-concave in y for fixed x. In this case weak duality holds, i.e., for
feasible (x, y) and (u, v)
Finally we point out that Mond–Weir duality has been found to be useful
and applicable in a great many different contexts. A recent check of Math
8 Mond–Weir Duality 165
Reviews showed 112 papers where the term Mond–Weir is used either in the
title or in the abstract. Seventy-eight of these papers are listed in [10].
References
Emma Hunt
9.1 Introduction
where each block is k × k, say. We restrict attention to the case where the
chain is irreducible but do not suppose positive recurrence. If the states are
partitioned conformably with the blocks, then the states corresponding to
block (≥ 0) are said to make up level and to constitute the phases of level
. The j-th phase of level will be denoted (, j).
In [18] Neuts noted a variety of special cases of the block–M/G/1 Markov
chain which occur as models in various applications in the literature, such as
Emma Hunt
School of Mathematical Sciences & School of Economics, The University of Adelaide,
Adelaide SA 5005, AUSTRALIA
e-mail: [email protected]
Bailey’s bulk queue (pp. 66–69) and the Odoom–Lloyd–Ali Khan–Gani dam
(pp. 69–71 and 348–353).
For applications, the most basic problem concerning the block–M/G/1
Markov chain is finding the invariant probability measure in the positive
recurrent case. We express this measure as π = (π0 , π1 , . . .), the components
πi being k-dimensional vectors so that π is partitioned conformably with the
structure of P . An efficient and stable method of determining π has been
devised by Ramaswami [20] based on a matrix version of Burke’s formula.
The key ingredient here is the fundamental matrix, G. This arises as follows.
Denote by Gr, (1 ≤ r, ≤ k) the probability, given the chain begins in
state (i + 1, r), that it subsequently reaches level i ≥ 0 and that it first
does so by entering the state (i, ). By the homogeneity of the transition
probabilities in levels one and above, plus the fact that trajectories are skip-
free downwards in levels, the probability Gr, is well defined and independent
of i. The fundamental matrix G is defined by G = (Gr, ).
A central property of G, of which we shall make repeated use, is that it is
the smallest nonnegative solution to the matrix equation
∞
G = A(G) := Ai Gi , (9.1)
i=0
∞
Xj+1 = Ai Xji , X0 = 0, (9.3)
i=0
or
∞
−1
Xj+1 = I− Ai Xji−1 A0 (9.5)
i=1
were introduced in Neuts [18], Latouche [12] in order to speed up the con-
vergence. However, the convergence of these numerical schemes still remains
linear. In Latouche [13] a Newton iteration was introduced in order to arrive
at a quadratic convergence, with an increase in the computational cost. In
Latouche and Stewart [15] the approximation of G was reduced to solving
nested finite systems of linear equations associated with the matrix P by
means of a doubling technique. In this way the solution for the matrix P
is approximated with the solution of the problem obtained by cutting the
infinite block matrix P to a suitable finite block size n.
In this chapter we present a probabilistic algorithm, Algorithm H, for the
determination of the fundamental matrix G in a structured M/G/1 Markov
chain. An account of the basic idea is given by the author in [11]. Algorithm
H is developed in the following three sections. In Section 5 we then consider
an alternative approach to the calculation of G. This turns out to be rather
more complicated than Algorithm H in general. We do not directly employ
this algorithm, Algorithm H, and we do not detail all its steps.
However Algorithm H serves several purposes. First, we show that it and
Algorithm H possess an interlacing property. This enables us to use it to
obtain (in Section 7) information about the convergence rate of Algorithm H.
Algorithm H reduces to Algorithm LR, the logarithmic-reduction algorithm
of Latouche and Ramaswami [14], in the quasi birth–death (QBD) case. Thus
for a QBD the interlacing property holds for Algorithms H and LR. This we
consider in Section 8.
In Section 9 we address the relation between Algorithm H and Bini and
Meini’s cyclic reduction algorithm, Algorithm CR. Algorithm CR was devel-
oped and refined in a chain of articles that provided a considerable improve-
ment over earlier work. See in particular [3]–[10]. We show that Algorithm H
becomes Algorithm CR under the conditions for which the latter has been
established. Algorithm H is seen to hold under more general conditions than
Algorithm CR. It follows from our discussion that, despite a statement to the
contrary in [8], Algorithm CR is different from Algorithm LR in the QBD
case.
170 E. Hunt
Thus Ur,s is the probability that, starting in (0, r), the process A0 revisits
level 0 at some subsequent time and does so with first entry into state (0, s).
The matrix U := (Ur,s ) may be regarded as the one-step transition matrix
of a Markov chain U on the finite state space K. The chain U is a censoring
of A0 in which the latter is observed only on visits to level zero. No state of
U is recurrent, for if r ∈ K were recurrent then the state (0, r) in A0 would
be recurrent, which is a contradiction. Since no state of U is recurrent, I − U
is invertible and
∞
U i = (I − U )−1 .
i=0
9 Computing the fundamental matrix of an M/G/1–type Markov chain 171
The matrix U is also strictly substochastic, that is, at least one row sum is
strictly less than unity.
any path whose probability contributes to Gr,s begins in (0, r), makes
In A,
some number n ≥ 0 of revisits to level 0 with −1 as a taboo level, and then
takes a final step to (−1, s). Suppose the final step to (−1, s) is taken from
(0, m). Allowing for all possible choices of m, we derive
∞
i
Gr,s = U (A0 )m,s ,
m∈K i=0 r,m
so that ∞
G= U i
A0 = (I − U )−1 A0 .
i=0
and
[I − U ()]−1 ↑ [I − U ]−1 as → ∞.
The probabilistic construction we are about to detail involves the exact
algorithmic determination (to machine precision) of U () for of the form
2N with N a nonnegative integer. This leads to an approximation
3 4−1
TN := I − U (2N ) A0
for G. We have
TN ↑ G as N → ∞.
The matrix TN may be interpreted as the contribution to G from those
trajectories from level 0 to level −1 in A that are restricted to pass through
only levels below 2N .
172 E. Hunt
(0)
that is, each chain Aj is of structured M/G/1 type. We have Bi = Ai for
(0)
i ≥ 1 and Ai = Ai for i ≥ 0.
In the previous section we saw that A0 contains no recurrent states, so
the same must be true also for the censorings A1 , A2 , ... . The substochastic
(j) (j)
matrices B1 , A1 , formed by censoring Aj to be observed only in levels 0
(j)
and 1 respectively, thus also contain no recurrent states. Hence I − B1 and
(j)
I − A1 are both invertible.
We now consider the question of deriving the block entries in P (j+1) from
(j) (j)
those in P (j) . First we extend our earlier notation and write Xt , Yt respec-
tively for the state and level of Aj at time t ∈ {0, 1, . . .}. For h a nonnegative
(j)
integer, define the event Ωs,t,h by
6 7
(j) (j) (j)
Ωs,t,h = Xt = (h, s), Yu(j) − Y0 is even (0 < u < t)
(j+1)
and for n ≥ 0, define the k × k matrix Ln by
% * &
/ (j) *
(j+1) * (j)
Ln :=P Ωs,t,2+2n * X0 = (2 + 1, r)
r,s *
t>0
(j)
n
(j) (j+1)
A(j+1)
n = A2n−1 + A2m Ln−m (n ≥ 1). (9.6)
m=0
(j)
n
(j) (j+1)
Bn(j+1) = B2n−1 + B2m Ln−m (n ≥ 1) (9.8)
m=1
for r, s ∈ K. Again the left-hand side is well defined. We may interpret this
as follows. Suppose Aj is initially in state (2 + 1, r). The (r, s) entry in
(j+1)
Kn is the probability that, at some subsequent time point, Aj is in state
(2 + 2n + 1, s) without in the meantime having been in any even–labelled
level.
(j+1)
Each path in Aj contributing to Ln consists of a sequence of steps each
of which involves even-sized changes of level, followed by a final step with an
odd-sized change of level. Conditioning on the final step yields
n
(j)
L(j+1)
n = (j+1)
Km A2(n−m) (n ≥ 0). (9.9)
m=0
(j+1)
Since Aj is skip-free to the left, trajectories contributing to K0 cannot
change level and so
∞
i −1
(j+1) (j) (j)
K0 = A1 = I − A1 . (9.10)
i=0
(j+1)
For n > 0, Kn involves at least one step in Aj with an increase in level.
Conditioning on the last such step yields the recursive relation
n−1
(j) (j+1)
Kn(j+1) = (j+1)
Km A2(n−m)+1 K0 (n ≥ 1). (9.11)
m=0
n−1
(j+1) (j)
Kn(j+1) := K0 (j+1)
A2(n−m)+1 Km (n ≥ 1). (9.12)
m=0
(N )
B1 = U (2N ) r,s
r,s
for r, s ∈ K, or
(N )
B1 = U (2N ).
Thus the recursive relations connecting the block entries in P (j+1) to those
in P (j) for j = 0, 1, . . . , N − 1 provide the means to determine U (2N ) exactly
and so approximate G.
9.4 Algorithm H
In the last section we considered the sequence of censored processes (Aj )j≥0 ,
(N )
each with the nonnegative integers as its levels. The determination of B1 re-
quires only a finite number of the matrix entries in each P (j) to be determined.
For the purpose of calculating TN , the relevant parts of the construction of
the previous section may be summarized as follows.
The algorithm requires initial input of A0 , A1 , . . . , A2N −1 . First we specify
Bn(0) = An (n = 1, . . . , 2N ),
A(0)
n = An (n = 0, 1, . . . , 2N − 1).
9 Computing the fundamental matrix of an M/G/1–type Markov chain 175
We then determine
(j) (j)
B1 , B2 , . . . , B2N −j ,
(j) (j)
A0 , A1 , . . . , A2N −j −1
n−1
(j) (j+1)
Kn(j+1) = (j+1)
Km A2(n−m)+1 K0 ,
m=0
for n = 0, 1, . . . , 2N −j−1 − 1.
Calculate (j+1) (j) (j+1)
A0 = A0 L0
and
(j)
n
(j) (j+1)
Bn(j+1) = B2n−1 + B2m Ln−m ,
m=1
(j)
n
(j) (j+1)
A(j+1)
n = A2n−1 + A2m Ln−m ,
m=0
for n = 1, 2, . . . , 2N −j−1 − 1.
(N )
The above suffices for the evaluation of B1 . We then compute
' (−1
(N )
TN = I − B1 A0 ,
that is,
These are set up recursively, beginning with M1 = A,
(1)
Ai := Ai (i ≥ 0). (9.13)
(j)
' (
(j) −1 (j)
Bi = I − A1 Ai . (9.14)
(j)
6 (j) (j) (j)
7
Ω s,t,h = X t = (k, s), Y u − Y 0 is even (0 < u < t) .
(j+1)
The matrices Ln are then defined for n ≥ 0 by
% * &
/ *
(j+1) (j) * (j)
Ln := P Ω s,t,2+2n−1 * X 0 = (2, r)
r,s *
t>0
(j+1) (j)
n
(j) (j+1)
An = B 2n−1 + B 2m Ln−m (n > 1) (9.15)
m=0
with
(j+1)
n
(j) (j+1)
An = B 2m Ln−m (n = 0, 1). (9.16)
m=0
The derivation is identical to that leading to (9.7) and (9.6). For n = 1 there is
no term corresponding to the first term on the right in (9.6) since the present
(j)
censoring requires B 1 := 0. The present censoring out of even-labeled, as
opposed to odd-labeled, levels means that no analogue to (9.8) is needed.
(j+1)
As before we now determine the matrices Ln in terms of matrices
(j+1)
Kn . We define
% * &
/ *
(j+1) (j) * (j)
Kn := P Ω s,t,2+2n * X 0 = (2, r)
r,s *
t>0
(j+1)
n
(j+1) (j)
Ln = Km B 2(n−m)
m=0
(j)
n
(j+1) (j)
= B 2n + Km B 2(n−m) (n ≥ 0). (9.17)
m=1
(j+1)
n−1
(j+1) (j)
Kn = Km B 2(n−m)+1
m=0
(j)
n−1
(j+1) (j)
= B 2n+1 + Km B 2(n−m)+1 (n ≥ 1), (9.18)
m=1
(j+1)
where again the empty sum for n = 1 is interpreted as zero. As with Ln ,
the leading term on the right-hand side corresponds to a single-step transition
in Nj while the sum incorporates paths involving more than one step in Nj .
178 E. Hunt
(N )
As with the computations involved in Algorithm H, B 0 can be calculated
in a finite number of steps. We may identify the relevant steps as follows.
(1)
We require initial input of A0 , A1 , ... , A2N −1 . First we specify An = An
for n = 0, 1, . . . , 2 − 1. We calculate
N
(j+1) (j)
n−1
(j+1) (j) (j+1)
Kn = B 2n+1 + Km A2(n−m)+1 K0 (9.20)
m=1
for n = 1, 2, . . . , 2N −j−1 − 1 and
(j+1) (j) n
(j+1) (j)
Ln = B 2n + K m B 2(n−m) (9.21)
m=1
(j+1)
n
(j) (j+1)
An = B 2m Ln−m (n = 0, 1), (9.22)
m=0
(j+1) (j)
n
(j) (j+1)
An = B 2n−1 + B 2m Ln−m (n = 2, . . . , 2N −j−1 − 1). (9.23)
m=0
(j)
For j ≥ 1, define M by
⎡ * ⎤
/ *
*
:= P ⎣ Φs,t ** X 0 = (0, r)⎦
(j) (j)
M (9.24)
r,s
t≥0 *
9 Computing the fundamental matrix of an M/G/1–type Markov chain 179
for r, s ∈ K, where
(j) 8 9
Φs,t = X t = (2j−1 , s), 0 ≤ Y u < 2j − 1, Y u = 2j−1 − 1 (0 < u < t) .
t>0
= B0 r,s
,
so that
(1)
V1 = B 0 . (9.27)
(j)
Proposition 9.6.1 For j ≥ 1, the matrices Vj , M are related by
(j) (j)
Vj = M B0 . (9.28)
where
(j) 8 9
Ψs,t = X t = (−1, s), 0 ≤ Y u < 2j − 1 (0 < u < t) ,
or
e − T 1 e ≥ e − T1 e ≥ e − T 2 e ≥ e − T2 e ≥ . . .
in the case of stochastic G.
The interlacing property need not carry over to other error measures such
as goodness of fit to the equation G = A(G). This will be further discussed
in the subsequent partner chapter in this volume.
Theorem 9.6.1 If A is transient and irreducible, then T N converges to G
quadratically as N → ∞.
Hence
2 2 2 2
2 2 2 2
2 N
2 2 ∞ 2 ∞
2G − V 2 ≤ 2 V 2≤ j−1
Kξ (2 )
2 j 2 2 j 2
2 j=1 2 2j+N +1 2 j=N +1
∞
N
= Kξ (2 )
ξ (2 )
=0
(2N )
< Kξ / 1 − ξ2 ,
Corollary 9.6.1 By the convergence result for T N and the interlacing prop-
erty, Algorithm H also converges to G quadratically when A is transient and
irreducible.
(j)
Remark 1. Similarly to the argument for B 0 , we have that
% * &
(j) / (j) **
B2 =P Λs,t * X 0 = (2j−1
− 1, r) , (9.30)
r,s *
t>0
where
(j) 8 9
Λs,t = X t = (2j − 1, s), 0 ≤ Y u < 3 · 2j−1 − 1 (0 ≤ u < t) .
(j) (j+1)
Because An = 0 for n > 2, (9.20) gives K n = 0 for n > 0. We have
(j+1)
already seen that K 0 = I. Relation (9.21) now provides
(j+1) (j) (j+1) (j) (j+1)
L0 = B0 , L1 = B 2 and Ln = 0 for n > 1.
(j+1)
(j) 2
An = Bn for n = 0, 2, (9.31)
(j)
' (
(j) −1 (j)
B n = I − A1 An for n = 0, 2. (9.33)
Equations (9.31)–(9.33) are simply the familiar defining relations for Al-
(j)
gorithm LR. We now turn our attention to the matrix M .
For a QBD with j > 1,
% * &
(j) / (j) **
M =P χs,t * X 0 = (0, r)
r,s *
t>0
where
(j) 8 9
χs,t = X t = (2j−1 − 1, s), 0 ≤ Y u < 2j−1 − 1, (0 < u < t)
and so
% * &
/ *
(j+1) j+1 *
M =P Ψs,t * X 0 = (0, r) .
r,s *
t>0
Thus in particular
% &
(2) /8 9 **
M =P X t = (1, s), Y u = 0, (0 < u < t) *X 0 = (0, r)
r,s
t>0
3 * 4
= P X 1 = (1, s)* X 0 = (0, r)
(1)
= B2 ,
r,s
so that
(2) (1)
M = B2 . (9.34)
For j ≥ 2, we derive by conditioning on the first passage of N1 to level
2j−1 − 1 that
9 Computing the fundamental matrix of an M/G/1–type Markov chain 183
(j+1)
M
r,s
% * & % * &
/ * / *
(j) * (j) *
= P χm,t * X 0 = (0, r) × P Υs,t,v * X t j−1
= (2 − 1, m)
* *
m∈K t>0 v>0
(j)
(j)
= M B2
r,m
m∈K m,s
(j) (j)
= M B2 ,
r,s
where
(j) 8 9
Υs,t,v = X t+v = (2j − 1, s), 0 ≤ Y u < 2j − 1 (t < u < t + v) .
Thus
(j+1) (j) (j)
M =M B2
and so for j > 2
(j+1) (2) (2) (j)
M =M B2 . . . B2
(1) (2) (j)
= B2 B2 . . . B2 .
0.5672 > 0.4800 > 0.3025 > 0.2619 > · · · > 4.9960e − 14 > 4.3854e − 14,
LR H
We now consider the relation between Algorithm H and Bini and Meini’s
Cyclic Reduction Algorithm CR. The latter is carried out in terms of formal
power series, so to make a connection we need to express Algorithm H in
these terms, too. For j ≥ 0, we define
∞
ψ (j) (z) := A(j) n
n z ,
n=0
∞
(j)
φ(j) (z) := Bn+1 z n .
n=0
Again, these power series are all absolutely convergent for |z| ≤ 1.
We introduce
∞
L(j+1) (z) := L(j+1)
n zn,
n=0
∞
K (j+1) (z) := Kn(j+1) z n .
n=0
∞
∞
∞
(j)
(j)
ψ (j+1) (z) = A2n−1 z n + A2m z m L(j+1)
n zn
n=1 m=0 n=0
= φ(j) (j)
o (z) + zφe (z)L
(j+1)
(z). (9.36)
186 E. Hunt
∞
n−1
(j+1) (j) (j+1)
K (j+1) (z) = K0 + zn (j+1)
Km A2(n−m)+1 K0
n=1 m=0
∞ ∞
(j+1) (j+1) (j) (j+1)
= K0 + Km z n A2(n−m)+1 K0
m=0 n=m+1
∞ ∞
(j+1) (j+1) m (j) (j+1)
= K0 + Km z z A2+1 K0
m=0 =1
∞
(j+1) (j) (j+1)
= K0 + K (j+1) (z) z A2+1 K0
=1
' (
(j+1) (j) (j+1)
= K0 + K (j+1) (z) ψo(j) (z) − A1 K0
' (
(j+1) (j+1)
= K0 + K (j+1) (z) ψo(j) (z) − I K0
' (
(j) (j+1)
+ K (j+1) (z) I − A1 K0 .
By (9.10) the last term on the right simplifies to K (j+1) (z). Hence we have
' (
(j+1) (j+1)
K0 = K (j+1) (z) I − ψo(j) (z) K0 .
(j) (j+1) −1
Postmultiplication by I − A1 = [K0 ]
yields
' (
I = K (j+1) (z) I − ψo(j) (z) ,
so that
' (−1
K (j+1) (z) = I − ψo(j) (z) .
and
' (−1
e (z) I − ψo (z)
ψ (j+1) (z) = zψo(j) (z) + φ(j) (j)
ψe(j) (z). (9.39)
and
∞
φ(0) (z) = An+1 z n . (9.41)
n=0
References
4. D. Bini and B. Meini, Exploiting the Toeplitz structure in certain queueing problems,
Calcolo 33 (1996), 289–305.
5. D. Bini and B. Meini, On the solution of a non–linear matrix equation arising in
queueing problems, SIAM J. Matrix Anal. Applic. 17 (1996), 906–926.
6. D. Bini and B. Meini, On cyclic reduction applied to a class of Toeplitz–like matrices
arising in queueing problems, in Computations with Markov Chains, Ed. W. J. Stewart,
Kluwer, Dordrecht (1996) 21–38.
7. D. Bini and B. Meini, Improved cyclic reduction for solving queueing problems, Nu-
merical Algorithms 15 (1997), 57–74.
8. D. A. Bini and B. Meini, Using displacement structure for solving non–skip–free
M/G/1 type Markov chains, in Advances in Matrix Analytic Methods for Stochas-
tic Models, Eds A. S. Alfa and S. R. Chakravarthy, Notable Publications, Neshanic
Station, NJ (1998), 17–37.
9. D. Bini and B. Meini, Solving certain queueing problems modelling by Toeplitz ma-
trices, Calcolo 30 (1999), 395–420.
10. D. Bini and B. Meini, Fast algorithms for structured problems with applications to
Markov chains and queueing models, Fast Reliable Methods for Matrices with Struc-
ture, Eds T. Kailath and A. Sayed, SIAM, Philadelphia (1999), 211–243.
11. E. Hunt, A probabilistic algorithm for determining the fundamental matrix of a block
M/G/1 Markov chain, Math. & Comput. Modelling 38 (2003), 1203–1209.
12. G. Latouche, Algorithms for evaluating the matrix G in Markov chains of P H/G/1
type, Bellcore Tech. Report (1992)
13. G. Latouche, Newton’s iteration for non–linear equations in Markov chains, IMA J.
Numer. Anal. 14 (1994), 583–598.
14. G. Latouche and V. Ramaswami, A logarithmic reduction algorithm for Quasi–Birth–
Death processes, J. Appl. Prob. 30 (1993), 650–674.
15. G. Latouche and G. W. Stewart, Numerical methods for M/G/1 type queues, in Proc.
Second Int. Workshop on Num. Solution of Markov Chains, Raleigh NC (1995), 571–
581.
16. B. Meini, Solving M/G/1 type Markov chains: recent advances and applications,
Comm. Statist.– Stoch. Models 14 (1998), 479–496.
17. B. Meini, Solving QBD problems: the cyclic reduction algorithm versus the invariant
subspace method, Adv. Performance Anal. 1 (1998), 215–225.
18. M. F. Neuts, Structured Stochastic Matrices of M/G/1 Type and Their Applications,
Marcel Dekker, New York (1989).
19. V. Ramaswami, Nonlinear matrix equations in applied probability – solution tech-
niques and open problems, SIAM Review 30 (1988), 256–263.
20. V. Ramaswami, A stable recursion for the steady state vector in Markov chains of
M/G/1 type, Stoch. Models 4 (1988), 183–188.
Chapter 10
A comparison of probabilistic
and invariant subspace methods
for the block M /G/1 Markov chain
Emma Hunt
10.1 Introduction
Emma Hunt
School of Mathematical Sciences & School of Economics, The University of Adelaide,
Adelaide SA 5005, AUSTRALIA
e-mail: [email protected]
G − G1 ∞ < G − G0 ∞ , (10.3)
with
r ≥ 1 and 0 < p < 1/r. (10.6)
This is an extension of the pure-birth/pure-death process of Latouche and
Ramaswami [7]. With these parameter choices, the QBD is irreducible. It is
null recurrent for r = 1 and positive recurrent for r > 1, with fundamental
matrix
10
G= .
10
we have
1 − p + py 0
A(GI ) = .
rpx + (1 − rp)xy 0
We have
G − G1 ∞ = 0.4 < 0.5 = G − G0 ∞
and so (10.3) holds. Also
0.75 0
A(G0 ) = , so that G0 − A(G0 ) ∞ = 0.25,
0.375 0
192 E. Hunt
and
0.95 0
A(G1 ) = , so that G1 − A(G1 ) ∞ = 0.35.
0.57 0
or
x − [1 − p + py] < [rpx + (1 − rp)xy] − y,
so that
Θ1 < −Θ2 ,
where
Θi := [GI − A(GI )]i,1 (i = 1, 2).
If Θ2 ≥ 0, then Θ1 < 0. Conversely, if Θ1 ≥ 0, then Θ2 < 0. In particular,
if Θ1 and Θ2 are both nonzero, then they are of opposite sign. This sort of
behavior does not appear to have been reported previously.
It is also worthy of note that, because of its use of rational functions, no
truncations need be involved in the computation of the error measure on the
left in (10.2) with the use of TELPACK.
the C program downloaded from Khosrow Sohraby’s website. All other code
has been implemented by us in MATLAB.
The following experiments illustrate a variety of issues.
10.3.1 Experiment G1
Our first experiment is drawn from the suite of TELPACK M/G/1 examples.
We use ⎡ ⎤
1 2 7
11i+1 11i+1 11i+1
⎢ ⎥
⎢ 18
10
i 1
10
i 1
10
i ⎥
⎢
Ai = ⎢ 21 21 ⎥
21 21 21 21 ⎥
⎣ ⎦
9
4
i+1 9
4
i+1 12
4
i+1
40 7 40 7 40 7
⎡ z −1
⎤⎡ ⎤
(1 − 11 ) 0 0 1 2 7
11 11 11
⎢ ⎥⎢ ⎥
⎢ ⎥ ⎢ 18 ⎥
=⎢
⎢ 0 (1 − 10z −1
21 ) 0 ⎥⎢
⎥ ⎢ 21
1 1
21 21
⎥
⎥
⎣ ⎦⎣ ⎦
4z −1
0 0 (1 − 7 )
9 9 12
70 70 70
⎡ 1 2 7
⎤
11 11 11
+ ,−1 ⎢ ⎥
1 10 4 ⎢ 18 ⎥
= I − z diag , , ⎢ 1 1 ⎥.
11 21 7 ⎢ 21 21 21 ⎥
⎣ ⎦
9 9 12
70 70 70
This example provides the simplest form of rational A(z) for which every
Ai has nonzero elements and can be expected to favor TELPACK.
Using the stopping criterion
TELPACK H
10.3.2 Experiment G2
As noted previously, TELPACK is designed for sequences (Ai ) for which the
generating function A(z) is, for |z| ≤ 1, a rational function of z. In this event
10 A comparison of probabilistic and invariant subspace methods 195
i=2
and −1
f
G(0) = 0, G(j + 1) = − Fi (G(j)) i−1
F0 .
i=1
196 E. Hunt
The most common choice of benchmark problem in the literature and the sub-
ject of our next three numerical experiments is a continuous-time teletraffic
example of Daigle and Lucantoni [5]. This involves matrices expressed in
terms of parameters K, ρd , a, r and M . The defining matrices Ai (i = 0, 1, 2)
10 A comparison of probabilistic and invariant subspace methods 197
M −j
(A1 )j,j+1 = ar (0 ≤ j ≤ K − 1), (A1 )j,j−1 = jr (1 ≤ j ≤ K).
M
10.3.3.1 Experiment G3
In this experiment the call holding rate is set at r = 100−1 s−1 and the calling
population size M is fixed at 512. In Tables 10.4 and 10.5 we compare the
number of iterations involved in estimating G with different algorithms for a
range of traffic parameter values from ρd = 0.01 to ρd = 0.29568. The latter
value was noted by Daigle and Lucantoni [5] to correspond to an instability
limit. The algorithms considered are the logarithmic-reduction algorithm of
Latouche and Ramaswami (LR), TELPACK and Algorithm H. We do not
give iteration counts for the original experiments of Daigle and Lucantoni.
These counts are not detailed in [5] but are mentioned as running to tens of
thousands.
Table 10.5 Iterations required with various traffic levels: Experiment G3 continued
It should be noted that in the references cited there is some slight variation
between authors as to the number of iterations required with a given method,
with larger differences at the instability limit. Akar et al. attribute this to
differences in the computing platforms used [4]. All computational results
given here are those obtained by us, either using our own MATLAB code or
by running TELPACK.
10.3.3.2 Experiment G4
Our fourth numerical experiment fixed the offered data traffic at 15%, the
call holding rate at r = 300−1 s−1 and then considered system behavior as a
function of the calling population size M (see Tables 10.6 and 10.7).
200 E. Hunt
In the light of its design versatility, Algorithm H compares quite well with the
above-mentioned more specialist algorithms. Its performance with respect to
CPU time and accuracy is comparable with that of the logarithmic-reduction
(LR) algorithm. Both the logarithmic-reduction algorithm and Algorithm
H require considerably less CPU time than does TELPACK (the difference
in times sometimes being as much as an order of magnitude) for superior
accuracy.
In Experiments 3 and 4 we employ the alternative error measure GI −
A(GI ) ∞ < suggested by Meini (see, for example, [8]). In terms of this
measure, the performance of TELPACK deteriorates steadily with an increase
in the size of M , whereas Algorithms H and LR are unaffected.
The last two TELPACK entries in Tables 10.5 and 10.7 are in small type-
face to indicate that TELPACK was unable to produce a result in these cases
and crashed, generating the error message ‘segmentation fault.’ Reducing
to 10−8 produced a result in both instances.
10.3.3.4 Experiment G5
We ran Algorithm H on the Daigle and Lucantoni problem with the call
holding rate fixed at r = 300−1 s−1 , the offered data traffic at 28% and the
calling population size M at 65,536, varying the size of the matrices from
24 × 24 to 500 × 500. In all cases we used (10.9) as a stopping criterion with
= 10−8 .
We found that although the iteration counts decreased as the size of the
matrices increased, CPU times increased substantially (see Table 10.8). This
held for all matrix sizes except for 24×24 (the first entry in Table 10.8) where
the computation required for the extra iterations outweighed the speed gain
due to smaller matrix size.
10.3.4 Experiment G6
We now turn our attention to the case of a null recurrent process where the
defining transition matrices for the system are given by
0.4 0 0 0.1 0.5 0
A0 = , A1 = and A2 = .
0 0.4 0.2 0.2 0 0.2
Results for this experiment are given in Table 10.9. The stopping criterion
used was (10.9) with = 10−8 . We note that this case is not covered by
202 E. Hunt
H
K Iterations I e − GI e∞ CPU Time (s)
23 29 9.4832e-09 0.110
24 19 5.1710e-11 0.080
25 18 8.0358e-11 0.080
26 17 2.6813e-08 0.090
27 17 9.2302e-11 0.100
28 17 4.4409e-16 0.100
29 16 2.5738e-08 0.110
39 15 2.3319e-11 0.200
49 14 1.18140e-09 0.260
59 14 2.2204e-15 0.600
69 13 3.6872e-08 1.130
79 13 4.5749e-10 2.250
89 13 4.5552e-12 4.170
99 13 5.3213e-13 7.670
149 12 3.0490e-09 76.400
299 12 9.7700e-15 853.990
499 12 5.8509e-14 3146.600
the Akar and Sohraby methodology and therefore that TELPACK cannot be
used for this experiment. The results for the H and LR Algorithms are several
orders more accurate than that for the Neuts Algorithm with significantly
lower CPU times.
10.3.5 Experiment G7
The numerical experiments above all involve matrix functions A(z) of rational
form. We could find no examples in the literature for which A(z) is not
rational. The following is an original example showing how Algorithm H
(and the Neuts Algorithm) perform when A(z) is not rational. We note that
these are the only two algorithms which can be applied here.
Suppose p, q are positive numbers with sum unity. We define two k × k
matrices Ω0 , Ω1 with ‘binomial’ forms. Let
10 A comparison of probabilistic and invariant subspace methods 203
⎡ ⎤
0 0 ... 0 0
⎢ .. .. .. .. .. ⎥
⎢ . . ⎥
Ω0 = ⎢ . . . ⎥,
⎣ 0 0 ... 0 0 ⎦
k−2
k−2 k−1
pk−1 k−1
1 p q k−1
. . . k−2 pq q
⎡ ⎤
p q 0 ... 0 0
⎢ p2 2pq q2 ... 0 0 ⎥
⎢ ⎥
⎢ .. .. .. .. .. .. ⎥
Ω1 = ⎢ . . . ⎥ .
⎢ k−1
. k−2 k−1
. k−3 2 k−1
. k−2 k−1 ⎥
⎣ pk−1 ⎦
1 p q 2 p q . . . k−2 pq q
0 0 0 ... 0 0
We now define
A0 := Ω0 e−r ,
rn rn−1
An := Ω0 e−r + Ω1 e−r (n ≥ 1),
n! (n − 1)!
∞
A(z) := Am z m
m=0
Ω := Ω0 + Ω1 = A(1).
ω [Ω1 + r(ω0 + Ω1 )] e ≤ 1,
that is,
3 4
ω (r + 1)e − (0, 0, . . . , ωk )T ≤ 1
or
r + 1 − ωk ≤ 1.
We deduce that G is stochastic if and only if
r ≤ ωk .
204 E. Hunt
The parameter choice r = 1 thus provides the new and interesting situation
of a transient chain. Results are given in Table 10.10 (with the size of the
matrices set to 5 × 5).
Since G is not stochastic we again revert to the use of
A(GI ) − GI ∞ <
as an error measure.
References
11.1 Introduction
S. S. Dragomir
School of Computer Science and Mathematics, Victoria University, Melbourne VIC 8001,
AUSTRALIA
e-mail: [email protected]
Emma Hunt
School of Mathematical Sciences & School of Economics, The University of Adelaide,
Adelaide SA 5005, AUSTRALIA
e-mail: [email protected]
C. E. M. Pearce
School of Mathematical Sciences, The University of Adelaide, Adelaide SA 5005,
AUSTRALIA
e-mail: [email protected]
Hadamard’s inequality are all means of f over the interval [a, b]. We denote
them respectively by mf (a, b), Mf (a, b), Mf (a, b) or simply by m, M, M
when f , a, b are understood. The bounds m and M for M are both tight.
More generally, the integral mean M is defined by
b
1
f (x)dx if a = b
Mf (a, b) = b−a a .
f (a) if a = b
and
+ , + , + ,
3a + b a + 3b 1 3a + b a + 3b
Mf , ≤ f +f
4 4 2 4 4
1
≤ Gf (t) dt
0
m+M
≤ . (11.3)
2
Inequality (11.2) was proved for differentiable convex functions. As this
class of functions is dense in the class of all convex functions defined on the
same interval with respect to the topology induced by uniform convergence,
(11.2) holds also for an arbitrary convex map.
Dragomir, Milošević and Sándor introduced a further map Lf : [0, 1] → R
given by
b
1
Lf (t) := [f (u) + f (v)] dx, (11.4)
2(b − a) a
where we define
Hf (t) + Hf (1 − t)
Hf (1 − t) ≤ Lf (t) and ≤ Lf (t). (11.5)
2
In this chapter we take these ideas further and introduce results involving
the modulus map. With the notation of (11.1) in mind, it is convenient to
employ
σ(x) := |x|, i(x) := x.
210 S.S. Dragomir et al.
so that + , + ,
1 a + w1 b + w2
Gf (t) = f +f . (11.6)
2 2 2
In Section 4 we derive some new results involving the identric mean I(a, b)
and in Section 5 introduce the univariate map Mf : [0, 1] → R given by
1
Mf (t) := [f (w1 ) + f (w2 )] (11.7)
2
and derive further results involving Lf . We remark that by convexity
so that
Mf (t) ≤ M. (11.8)
f (x) ≥ |g(x)| on I,
then
*
*
* *
f (x)dx ≥ * g(x)dx** .
*
I I
11 Interpolating maps, the modulus map and Hadamard’s inequality 211
*
*
* *
Proof. We have f (x)dx ≥ |g(x)|dx ≥ ** g(x)dx** .
I I I
and
Mf (a, b) − mf (a, b) ≥ |Mσ◦f s (a, b) − |mf (a, b)| | . (11.10)
we thus have
Inequality (11.9) now follows from evaluation of the integrals and a change
of variables.
Similarly we have for any α, β ∈ [a, b] that
+ , ** * * + ,**
α+β * * f (α) + f (β) * * **
f (α) + f (β)
−f ≥ ** ** * − *f α + β * * .
2 2 2 * * 2 **
212 S.S. Dragomir et al.
Set α = w1 , β = w2 . Then
α+β =a+b (11.14)
and by Lemma 2.1 we have
1
1
1
f (w1 ) dt + f (w2 ) dt − m
2 0 0
*
1* * *
* * f (w1 ) + f (w2 ) * *
≥* * * * dt − |m| * .
* 2 * *
0
|d| ≤ |d − c| + |c|.
and
M − m ≥ | Mσ◦f (a, b) − |m| | . (11.20)
A necessary and sufficient condition for strict inequality in (21.4) is that
there exists x ∈ [a, b] such that
(where L(a, b) refers to the logarithmic mean), which was first proved by
Ostle and Terwilliger [6] and Carlson [1], [2], and
Proof. For the convex function f (x) = − ln x (x > 0), the left-hand side of
(11.9) is
b
ln a + ln b 1
− + ln x dx
2 b−a a
1
= [b ln b − a ln a − (b − a)] − ln G(a, b)
b−a
I(a, b)
= ln .
G(a, b)
Since
ln b
(ln b)2 + (ln a)2
|x|dx =
ln a 2
and
b 3 4
| ln x|dx = ln aa bb e2−a−b ,
a
we have likewise for the same choice of f that the right-hand side of
(11.9) is
* ' (**
* (ln b)2 + (ln a)2
* − ln b b a 2−a−b 1/(b−a) *
a e
* ln((b/a)2 ) *,
(b − x) ln a + (x − a) ln b < 0.
Since the left-hand side is strictly increasing in x, this condition can be sat-
isfied if and only if the left-hand side is strictly negative for x = 1, that is,
we require
(b − 1) ln a + (1 − a) ln b < 0. (11.23)
By Corollary 2.3, a necessary and sufficient condition for the first inequality
in (21.7) to be strict is that there should exist x ∈ [a, b] such that
a+b
ln[x(a + b − x)] < 0 < ln ,
2
that is,
x(a + b − x) < 1 < (a + b)/2.
The leftmost term is minimized for x = a and x = b, so the condition reduces
to
ab < 1 < (a + b)/2 or 2 − b < a < 1/b.
Since 2 − b < 1/b for b = 1, there are always values of a for which this
condition is satisfied.
216 S.S. Dragomir et al.
Our first result in this section provides minorants for the difference between
the two sides of the first inequality and the difference between the outermost
quantities in (11.2).
and
Hf (t) − m ≥ |Mσ◦f s (a, b) − |m| | . (11.26)
Proof. We have
1
Gf (t) = [f (u1 ) + f (u2 )] = Mf (u1 , u2 ),
2
Hf (t) = Mf ◦yt (a, b) = Mf (u1 , u2 ),
mf (u1 , u2 ) = mf (a, b) = m,
so for t ∈ [0, 1] application of Theorem 2.2 to f on (u1 , u2 ) provides
so (11.27) yields (11.25) for t ∈ (0, 1]. As (11.25) also holds for t = 0, we have
the first part of the theorem. Using u1 + u2 = a + b, we derive the second
part similarly from Mσ◦f s (u1 , u2 ) = Mσ◦f s ◦yt (a, b).
Our next result provides a minorant for the difference between the two
sides of the second inequality in (11.3) and a corresponding result for the
third inequality.
11 Interpolating maps, the modulus map and Hadamard’s inequality 217
and
+ , + ,
1 3a + b a + 3b
M− f +f
2 4 4
*
1 * + , + ,* *
1 ** *
* 3a + b a + 3b ** *
*.
≥ * |Gf (t) + Gf (1 − t)| dt − *f +f * *
2 0 4 4
(11.30)
and * * + ,* *
* * 1 ** *
M − Gf (1/2) ≥ *Mσ◦Gf (0, 1) − **Gf
* *.
2 * *
s
where, as before, G(x, y) denotes the geometric mean of the positive numbers
x, y.
A(a, b) I(u1 , u2 )
1≤ ≤ . (11.33)
I(u1 , u2 ) γa,b (t)
H− ln (t) = − ln I(u1 , u2 ).
Since the map x :→ exp(−x) is order reversing, (b)–(e) follow from Theorem
B(ii),(iii). It remains only to establish (a).
Since dui /dt = (−1)i (b − a)/2 for i = 1, 2 and u2 − u1 = t(b − a), we have
from (11.32) that
and so
dγa,b (b − a)2 (b − a)2 ' 1/2 −3/2 1/2 −3/2
(
=− 1/2
− u2 u1 + u1 u2 < 0.
dt 8(u1 u2 ) 16
We now apply Theorems 3.1 and 3.2 to obtain further information about
the identric mean. For t ∈ [0, 1], put
Because
G− ln (t) = − ln γa,b (t), H− ln (t) = − ln ηa,b (t),
(11.25) provides
where
La,b (t) = Mσ (ln u1 , ln u2 ).
This yields
ηa,b (t)
≥ exp [La,b (t)] ≥ 1 for t ∈ [0, 1].
γa,b (t)
From (11.26) we derive
+ , * * + ,* *
* 1
b √ * a + b ** *
a+b * *
ln − ln ηa,b (t) ≥ * |ln yu2 | dx − **ln * * ≥ 0,
2 *b − a a 2 *
which gives
%* * + ,* *&
* 1
b √ * a + b ** *
A(a, b) * *
≥ exp * |ln yu2 | dx − **ln * *
ηa,b (t) *b − a a 2 *
≥1 for t ∈ [0, 1].
where *
*
* 1 *
Ka,b = **Mσ (ln A(a, b), ln G(a, b)) − |ln γa,b (t)| dt** .
0
Hence
I(a, b)
≥ exp [Ka,b ] ≥ 1.
G(A(a, b), G(a, b))
Finally, applying (11.30) to the convex mapping − ln provides
+ ,
a + 3b 3a + b
ln G , − ln I(a, b)
4 4
*
* + ,+ ,* *
1 ** 1 *
* 3a + b a + 3b ** *
*
≥ * |ln [γa,b (t)γa,b (1 − t)]| dt − *ln * *
2 0 4 4
= Ma,b ,
where
*
* * + ,*
* 1 * * 3a + b a + 3b **
Ma,b := ** * *
|ln G(γa,b (t), γa,b (1 − t))* dt − *ln G , *.
0 4 4
220 S.S. Dragomir et al.
Hence a+3b
G , 3a+b
4 4
≥ exp [Ma,b ] ≥ 1.
I(a, b)
Substituting Af (t), Bf (t) for the leftmost terms in the known inequalities
w1 + ,
1 1 f (a) + f (w1 ) a + w1 f (a) + f (w1 )
f (u)du ≤ +f ≤ ,
w1 − a a 2 2 2 2
b + ,
1 1 f (b) + f (w2 ) b + w2 f (b) + f (w2 )
f (v)dv ≤ +f ≤
b − w2 w2 2 2 2 2
11 Interpolating maps, the modulus map and Hadamard’s inequality 221
The first two inequalities in (11.34) follow from (11.6) and (11.7) and the
final inequality from (11.8).
Lf (t) − Hf (1 − t)
*
w1 * * *
* 1 * f (u) + f (u + t(b − a)) * *
≥ ** *
*
* du − Hσ◦f (1 − t)*
* *
(1 − t)(b − a) a 2
≥0 (11.35)
Proof. Put
Since
* *
1 * f (u) + f (v) *
b
* * dx
b−a * 2 *
a
w1 * *
1 * f (u) + f (u + t(b − a)) *
= * * du,
(1 − t)(b − a) a * 2 *
We now apply the above to obtain results for the identric mean. We may
compute Af , Bf for f = − ln to derive
w1
1
A− ln (t) = (− ln u)du = − ln I(w1 , a),
w1 − a a
b
1
B− ln (t) = (− ln u)du = − ln I(b, w2 ).
b − w2 w2
Thus
1
L− ln (t) = [A− ln (t) + B− ln (t)] = − ln ζa,b (t),
2
where the map ζa,b : [0, 1] → R is defined by
γa,b (t) ≥ ζa,b (t) ≥ [I(a, b)]1−t [G(a, b)]t ≥ G(a, b);
ηa,b (1 − t) ≥ ζa,b (t) and G(ηa,b (t), ηa,b (1 − t)) ≥ ζa,b (t).
Proof. Since
ζa,b (t) = exp [−L− ln (t)] for all t ∈ [0, 1]
and the map x :→ exp(−x) is order reversing, (a) and (b) follow from Theo-
rem C, parts 2 and 3.
Remark 5.5. Similar results may be obtained from Propositions 5.1 and
5.2.
11 Interpolating maps, the modulus map and Hadamard’s inequality 223
References
1. B. C. Carlson, Some inequalities for hypergeometric functions, Proc. Amer. Math. Soc.
17 (1966), 32–39.
2. B. C. Carlson, The logarithmic mean, Amer. Math. Monthly 79 (1972), 615–618.
3. S. S. Dragomir, D. S. Milošević and J. Sándor, On some refinements of Hadamard’s
inequalities and applications, Univ. Belgrad Publ. Elek. Fak. Sci. Math. 4 (1993),
21–24.
4. S. S. Dragomir and C. E. M. Pearce, Hermite–Hadamard Inequali-
ties, RGMIA Monographs, Victoria University, Melbourne (2000), online:
http://rgmia.vu.edu.au/monographs.
5. S. S. Dragomir and E. Pearce, A refinement of the second part of Hadamard’s in-
equality, with applications, in Sixth Symposium on Mathematics & its Applications,
Technical University of Timisoara (1996), 1–9.
6. B. Ostle and H. L. Terwilliger, A comparison of two means, Proc. Montana Acad. Sci.
17 (1957), 69–70.
7. K. B. Stolarsky, Generalizations of the logarithmic mean, Math. Mag. 48 (1975),
87–92.
8. K. B. Stolarsky, The power and generalized of logarithmic means, Amer. Math.
Monthly 87 (1980), 545–548.
Chapter 12
Estimating the size of correcting codes
using extremal graph problems
12.1 Introduction
• Two-deletion (e = 2d): F2d (u) ⊆ B l−2 and all elements of F2d (u) are
obtained by deletion of two of the components of u. For u = 0101 we have
F2d (u) = {00, 01, 10, 11}.
• Single transposition, excluding the end-around transposition (e = 1t):
F1t (u) ⊆ B l and all elements of F1t (u) are obtained by transposition of a
neighboring pair of components in u. For example, if l = 5 and u = 11100
then F1t (u) = {11100, 11010}.
• Single transposition, including the end-around transposition (e = 1et):
F1et (u) ⊆ B l and all elements of F1et (u) are obtained by transposition
of a neighboring pair of components in u, where the first and the last
components are also considered as neighbors. For l = 5 and u = 11100 we
obtain F1et (u) = {11100, 11010, 01101}.
• One error on the Z-channel (e = 1z): F1z (u) ⊆ B l and all elements
of F1z (u) are obtained by possibly changing one of the nonzero compo-
nents of u from 1 to 0. If l = 5 and u = 11100 then F1z (u) =
{11100, 01100, 10100, 11000}. The codes correcting one error on the Z-
channel represent the simplest case of asymmetric codes.
Our problem of interest here is to find the largest correcting codes. It appears
that this problem can be formulated in terms of extremal graph problems as
follows [24].
Consider a simple undirected graph G = (V, E), where V = {1, . . . , n}
is the set of vertices and E is the set of edges. The complement graph of G
is the graph Ḡ = (V, Ē), where Ē is the complement of E. Given a subset
W ⊆ V , we denote by G(W ) the subgraph induced by W on G. A subset
I ⊆ V is called an independent set (stable set, vertex packing) if the edge set
of the subgraph induced by I is empty. An independent set is maximal if it
is not a subset of any larger independent set and maximum if there are no
larger independent sets in the graph. The independence number α(G) (also
called the stability number) is the cardinality of a maximum independent set
in G. A subset C ⊆ V is called a clique if G(C) is a complete graph.
Consider a graph Gl having a vertex for every vector u ∈ B l , with an
edge joining
: the vertices corresponding to u, v ∈ B l , u = v if and only if
Fe (u) Fe (v) = ∅. Then a correcting code corresponds to an independent
set in Gl . Hence the largest e-correcting code can be found by solving the
maximum independent set problem in the considered graph. Note that this
problem could be equivalently formulated as the maximum clique problem in
the complement graph of G.
Another discrete optimization problem which we will use to obtain lower
bounds for asymmetric codes is the graph coloring problem, which is formu-
lated as follows. A legal (proper) coloring of G is an assignment of colors to
its vertices so that no pair of adjacent vertices has the same color. A color-
ing induces naturally a partition of the vertex set such that the elements of
each set in the partition are pairwise nonadjacent; these sets are precisely the
subsets of vertices being assigned the same color. If there exists a coloring
of G that uses no more than k colors, we say that G admits a k-coloring
12 Estimating the size of correcting codes using extremal graph problems 229
In this section we summarize the results obtained in [5, 21]. We start with
the following global optimization formulation for the maximum independent
set problem.
Theorem 1 ([1]). The independence number of G satisfies the following
equality:
n
α(G) = max n xi (1 − xj ). (12.1)
x∈[0,1]
i=1 (i,j)∈E
230 S. Butenko et al.
This formulation is valid if instead of [0, 1]n we use {0, 1}n as the feasi-
ble region, thus obtaining an integer 0–1 programming problem. In problem
(12.1), for each vertex i there is a corresponding Boolean expression:
⎡ ⎤
; ;
i ←→ ri = xi ⎣ xj ⎦ .
(i,j)∈E
To apply local search techniques to the above problem one needs to define
a proper neighborhood. We define the neighborhood on the set of all maximal
independent sets as follows.
For each jq ∈ I, q = 1, . . . , |I|,
⎧ ⎡ ⎤ ⎫
⎨ ; has exactly 2 literals ⎬
Listjq = / I : ri = ⎣xi
i∈ x̄k ⎦ with value 0, namely .
⎩ ⎭
(i,k)∈E xi = 0 and x̄jq = 0
jq ∈ I, q = 1, . . . |I|}.
We tested the proposed algorithm with the following graphs arising from
coding theory. These graphs are constructed as discussed in Section 12.1 and
can be downloaded from [24]:
• Graphs From Single-Deletion-Correcting Codes (1dc);
• Graphs From Two-Deletion-Correcting Codes (2dc);
• Graphs From Codes For Correcting a Single Transposition, Excluding the
End-Around Transposition (1tc);
• Graphs From Codes For Correcting a Single Transposition, Including the
End-Around Transposition (1et);
• Graphs From Codes For Correcting One Error on the Z-Channel (1zc).
The results of the experiments are summarized in Table 12.1. In this table,
the columns “Graph,” “n” and “|E|” represent the name of the graph, the
number of its vertices and its number of edges. This information is available
from [24]. The column “Solution found” contains the size of the largest inde-
pendent sets found by the algorithm over 10 runs. As one can see the results
are very encouraging. In fact, for all of the considered instances they were at
least as good as the best previously known estimates.
/
k
V = Ci ,
i=1
x ≥ 0, (12.5)
12 Estimating the size of correcting codes using extremal graph problems 233
y ≥ 0, (12.8)
where
1, if i ∈ Cj ,
aij =
0, otherwise.
m+n
min
m+n
ci yi , (12.9)
y∈R
i=1
s. t. Ay = e, (12.10)
yi ≥ 0, i = 1, . . . , m + n. (12.11)
234 S. Butenko et al.
m+n
m+n
cj yjk − cj yjk+1 < ε, where ε = 10−3 .
j=1 j=1
(b) Bounding: We use the approximate solution found as a lower bound and
the linear clique estimate OC (G) as an upper bound.
Tables 12.2 and 12.3 contain a summary of the numerical experiments with
the exact algorithm. In Table 12.2 Column “#” contains a number assigned to
Graph # 1 2 3
1tc128 1 5 4 4.0002
2 5 5 5.0002
3 5 5 5.0002
4 5 4 4.0001
1tc256 1 6 5 5.0002
2 10 9 9.2501
3 19 13 13.7501
4 10 9 9.5003
5 6 5 5.0003
1tc512 1 10 7 7.0003
2 18 14 14. 9221
3 29 22 23.6836
4 29 22 23.6811
5 18 14 14.9232
6 10 7 7.0002
1dc512 1 75 50 51.3167
2dc512 1 16 9 10.9674
1et128 1 3 2 3.0004
2 6 4 5.0003
3 9 7 7.0002
4 9 7 7.0002
5 6 4 5.0003
6 3 2 3.0004
1et256 1 3 2 3.0002
2 8 6 6.0006
3 14 10 12.0001
4 22 12 14.4002
5 14 10 12.0004
6 8 6 6.0005
7 3 2 3.0002
1et512 1 3 3 3.0000
2 10 7 8.2502
3 27 18 18.0006
4 29 21 23.0626
5 29 21 23.1029
6 27 18 18.0009
7 10 7 8.2501
8 3 3 3.0000
236 S. Butenko et al.
The error-correcting codes for the Z-channel have very important practi-
cal applications. The Z-channel shown in Fig. 12.1 is an asymmetric binary
channel, in which the probability of transformation of 1 into 0 is p, and the
probability of transformation of 0 into 1 is 0.
1 1
1−p
The problem we are interested in is that of finding good estimates for the
size of the largest codes correcting one error on the Z-channel.
Let us introduce some background information related to asymmetric
codes.
12 Estimating the size of correcting codes using extremal graph problems 237
It was shown in [20] that a code C with the minimum asymmetric distance
Δ can correct at most (Δ − 1) asymmetric errors (transitions of 1 to 0). In
this subsection we present new lower bounds for codes with the minimum
asymmetric distance Δ = 2.
Let us define the graph G = (V (l), E(l)), where the set of vertices V (l) =
B l consists of all binary vectors of length l, and (vi , vj ) ∈ E(l) if and only
if dA (vi , vj ) < Δ. Then the problem of finding the size of the code with
minimal asymmetric distance Δ is reduced to the maximum independent set
problem in this graph. Table 12.4 contains the lower bounds obtained using
the algorithm presented above in this section (some of which were mentioned
in Table 12.1).
Table 12.4 Lower bounds obtained in: a [27]; b [6]; c [7]; d [12]; e (this chapter)
l Lower Bound Upper Bound
4 4 4
5 6a 6
6 12b 12
7 18c 18
8 36c 36
9 62c 62
10 112d 117
11 198d 210
12 379e (378d) 410
The partitioning method [4, 12, 26] uses independent set partitions of the
vertices of graph G in order to obtain a lower bound for the code size. An
238 S. Butenko et al.
Recall that the problem of finding the smallest m for which a partition of
the vertices into m disjoint independent sets exists is the well-known graph
coloring problem.
The independent set partition (12.14) can be identified by the vector
Π(l) = (I1 , I2 , . . . , Im ).
which is called the index vector of partition Π(n), with Π(l). Its norm is
defined as
m
π(l) · π(l) = |Ii |2 .
i=1
The partitions obtained using the described partition algorithm are given in
Tables 12.5 and 12.6. These partitions, together with the facts that [11]
Π(l, 0) consists of one (zero) codeword,
Π(l, 1) consists of l codes of size 1,
Π(l, 2) consists of l − 1 codes of size l/2 for even l,
the index vectors of Π(l, w) and Π(l, l − w) are equal
240 S. Butenko et al.
10 1 112, 110, 110, 109, 105, 100, 99, 88, 75, 59, 37, 16, 4 97942 13
2 112, 110, 110, 109, 105, 101, 96, 87, 77, 60, 38, 15, 4 97850 13
3 112, 110, 110, 108, 106, 99, 95, 89, 76, 60, 43, 15, 1 97842 13
4 112, 110, 110, 108, 105, 100, 96, 88, 74, 65, 38, 17, 1 97828 13
5 112, 110, 110, 108, 106, 103, 95, 85, 76, 60, 40, 15, 4 97720 13
6 112, 110, 110, 108, 106, 101, 95, 87, 75, 61, 40, 17, 2 97678 13
7 112, 110, 109, 108, 105, 101, 96, 86, 78, 63, 36, 17, 3 97674 13
12 Estimating the size of correcting codes using extremal graph problems 241
Table 12.6 Partitions of constant weight codes obtained in: a (this chapter); b [4]; c [12]
l2 w # Partition Index-Vector Norm m
were used in (12.15), with = 0, to obtain new lower bounds for the asym-
metric codes presented in Table 12.7. To illustrate how the lower bounds
were computed, let us show how the code for l = 18 was constructed. We use
l1 = 8 and l2 = 10:
Table 12.7 New lower bounds. Previous lower bounds were found in: a [11]; b [12]
Lower Bound
l New Previous
18 15792 15762a
19 29478 29334b
20 56196 56144b
21 107862 107648b
22 202130 201508b
24 678860 678098b
12.4 Conclusions
In this chapter we have dealt with binary codes of given length correcting
certain types of errors. For such codes, a graph can be constructed in which
each vertex corresponds to a binary vector and the edges are built such
242 S. Butenko et al.
Acknowledgments We would like to thank two anonymous referees for their valuable
comments.
References
16. J. Håstad, Clique is hard to approximate within n1− , Acta Math. 182 (1999),
105–142.
17. D. S. Johnson and M. A. Trick (Eds), Cliques, Coloring, and Satisfiability: Second
DIMACS Implementation Challenge, Vol. 26 of DIMACS Series, (American Mathe-
matical Society, Providence, RI, 1996).
18. C. Lund and M. Yannakakis, On the hardness of approximating minimization prob-
lems, JACM 41 (1994), 960–981.
19. P. M. Pardalos, T. Mavridou and J. Xue, The graph coloring problem: a bibliographic
survey, in D.-Z. Du and P. M. Pardalos, Eds, Handbook of Combinatorial Optimization,
Vol. 2 (Kluwer Academic Publishers, Dordrecht, 1999), 331–395.
20. T. R. N. Rao and A. S. Chawla, Asymmetric error codes for some lsi semiconductor
memories, Proceedings of the 7th Southeastern Symposium on System Theory (1975)
(IEEE Computer Society Press, Los Alamitos, California, 1975), 170–171.
21. I. V. Sergienko, V. P. Shylo and P. I. Stetsyuk, Approximate algorithm for solving
the maximum independent set problem, in Computer Mathematics, (V.M. Glushkov
Institute of Cybernetics NAS of Ukraine, Kiev, 2001), 4–20 (in Russian).
22. V. Shylo, New lower bounds of the size of error–correcting codes for the Z–channel,
Cybernet. Systems Anal. 38 (2002), 13–16.
23. V. Shylo and D. Boyarchuk, An algorithm for construction of covering by indepen-
dent sets, in Computer Mathematics (V.M. Glushkov Institute of Cybernetics NAS of
Ukraine, Kiev, 2001), 151–157.
24. N. Sloane, Challenge problems: Independent sets in graphs, http://www.research.
att.com/∼njas/doc/graphs.html, 2001.
25. N. Sloane, On single–deletion–correcting codes, in K. T. Arasu and A. Suress, Eds,
Codes and Designs: Ray–Chaudhuri Festschrift (Walter de Gruyter, Berlin, 2002),
273–291.
26. C. L. M. van Pul and T. Etzion, New lower bounds for constant weight codes, IEEE
Trans. Inform. Theory 35 (1989), 1324–1329.
27. R. R. Varshamov, A class of codes for asymmetric channels and a problem from the
additive theory of numbers, IEEE Trans. Inform. Theory IT–19 (1973), 92–95.
Chapter 13
New perspectives on optimal
transforms of random vectors
P. G. Howlett
Centre for Industrial and Applicable Mathematics, The University of South Australia,
Mawson Lakes, SA 5095, AUSTRALIA
e-mail: [email protected]
C. E. M. Pearce
School of Mathematical Sciences, The University of Adelaide, Adelaide SA 5005,
AUSTRALIA
e-mail: [email protected]
A. P. Torokhti
Centre for Industrial and Applicable Mathematics, The University of South Australia,
Mawson Lakes, SA 5095, AUSTRALIA
e-mail: [email protected]
given by
n
T (y) = A + Bj zj , (13.2)
j=0
subject to
rank[A B0 B1 . . . Bn ] = r (13.4)
with r ≤ m. Here
⎡2 ⎛ ⎞22 ⎤
2 2
⎢2 2 ⎥
n
J(A, B0 , . . . , Bn ) = E ⎣2
2 x − ⎝A + B ⎠ 2
j j 2 ⎦,
z (13.5)
2 j=0 2
possesses a much smaller associated error than that of the GKLT. We then
proceed to address applications and simulations.
248 P.G. Howlett et al.
13.3 Preliminaries
2
Lemma 13.3.2 Let z = [z1T · · · znT ]T ∈ Rn ,
Then
GD† D = G.
Proof. Let
1
t= , S11 = 1 − S12 E[y],
y
S12 = −E[y T ]S22 , T
S21 = S12 , S22 = E †yy .
First we show that
† S11 S12
Ett = . (13.10)
S21 S22
13 New perspectives on optimal transforms of random vectors 249
We have
S11 S12 Q11 Q12
Ett Ett = ,
S21 S22 Q21 Q22
where
Q11 = S11 + E[y T ]S21 + S12 E[y] + E[y T ]S22 E[y] = 1,
Q12 = S11 E[y T ] + E[y T ]S21 E[y T ] + S12 Eyy + E[y T ]S22 Eyy = E[y T ],
Q21 = E[y]S11 + Eyy S21 + E[y]S12 E[y] + Eyy S22 E[y] = E[y]
and
Q22 = E[y]S11 E[y T ] + Eyy S21 E[y T ] + E[y]S12 Eyy + Eyy S22 Eyy = Eyy .
S11 S12
Hence Ett Ett = Ett , that is, the first condition for the Moore–
S21 S22
Penrose inverse of Ett to be given by (13.10) is satisfied. The remain-
†
ing Moore–Penrose conditions for Ett are also easily verified, and therefore
(13.10) is established.
Next, let
† †
R11 = Ett − R12 Ezt Ett , R12 = R21 T
, (13.11)
† †
R21 = −R22 Ezt Ett and R22 = Dzt , (13.12)
†
where Dzt = Ezz − Ezt Ett Etz = D.
Arguing much as above, we have by Lemmas 13.3.1 and 13.3.2 that
† R11 R12
Ess = . (13.13)
R21 R22
1/2 †
U ΣV T = Exs (Ess ) , (13.14)
where
U = (u1 , . . . , uq ) ∈ Rm×q and V = (v1 , . . . , vq ) ∈ Rq×q
and define
Θr = Θr(x,s) = Ur Σr VrT . (13.15)
Suppose
Φ = [A B0 . . . Bn ] ∈ Rm×q
and let Φ(:, η : τ ) be the matrix formed by the τ − η + 1 sequential columns
of Φ beginning with column η.
The optimal transform T 0 , introduced by (13.7), is defined by the following
theorem.
Theorem 13.4.1 The solution to problem (13.3) is given by
for j = 0, 1, . . . n, where
1/2 † 1/2 †
Φ0 = Θr (Ess ) + Mr [I − Ess
1/2
(Ess ) ],
with I the identity matrix and Mr ∈ Rm×q any matrix such that rank Φ 0
≤ r < m.
Proof. We have
J(A, B0 , . . . , Bn ) = E[ x − Φs 2 ].
By Lemma 13.3.1,
J(A, B0 , . . . , Bn )
8 †
9 8 †
9
= tr Exx − Exs Ess Esx + tr (Φ − Exs Ess )Ess (Φ − Exs Es† )T
8 †
9 22 †
22
1/2 2
= tr Exx − Exs Ess Esx + 2(Φ − Exs Ess )Ess 2 . (13.17)
The necessary and sufficient condition (see [2]) for (13.18) to have a solution
is readily verified to hold and provides the solution Φ = Φ0 . The theorem is
proved.
Remark 1. The proof above is based on Lemma 13.3.1. The first equation in
(13.8) has been presented in [8] but without proof.
Theorem 13.4.2 Let
†
Δ = (Exz − Exy Eyy Eyz )(D† )1/2 2 .
13 New perspectives on optimal transforms of random vectors 251
k
† 1/2 2
E[ x − T 0 (y) 2 ] = tr{Exx } + σi2 − Exy (Eyy ) − Δ.
i=r+1
where
k
U ΣV T − Θr 2 = σj2
j=r+1
k 2 22
l
2 † 2
σ̃j2 < 2(Exz − Exy Eyy Eyz )(D† )1/2 2 + ϑ2i , (13.19)
j=r+1 i=p+1
then the error associated with the transform T̃ 0 is less than that associated
with H, that is, 2(2
'2
2 2
E 2x − T̃ 0 (y)2 < E[ x − Hy ]2 .
Hence
'2 2(2
2 2
E[ x − Hy ]2 − E 2x − T̃ 0 (y)2
2 22
l
k
2 † † 1/2 2
= 2(Exz − Exy Eyy Eyz )(D ) 2 + ϑi −
2
σ̃j2 ,
i=p+1 j=r+1
We now address the solution of the minimization problem (13.3) without the
constraint (13.4). This is important in its own right. The solution is a special
form of the transform T 0 and represents a model of the optimal nonlinear
filter with x an actual signal and y the observed data.
Let
⎡ ⎤
D11 . . . D1n
⎢ D21 . . . D2n ⎥
P=⎢ ⎥
⎣ . . . . . . . . . ⎦ and G = [G1 G2 . . . Gn ],
Dn1 . . . Dnn
where
n
B00 = (E xy − Bk0 E zk y )E †yy + M1 [I − E 1/2 1/2 †
yy (E yy ) ], (13.21)
k=1
where Qij ∈ Rn×n for i, j = 1, . . . , n. The error associated with the transform
T (1) defined by
n
T (1) (y) = A0 + Bj0 zj ,
j=0
0
with A and Bj0 given by (13.20)–(13.22), is
n
E[ x − T (1) (y) 2 ] = tr{E xx } − E xy (E †yy )1/2 2 −
1/2
Gi Qii 2
i=1
− tr{Gj Qjk GTk }.
j,k=1,...,n
j =k
†
associated with the optimal linear filter H (1)) = Exy Eyy in [8].
Applications of our technique are abundant and include, for example, simul-
taneous filtering and compression of noisy stochastic signals, feature selection
in pattern recognition, blind channel equalization and the optimal rejection
of colored noise in some neural systems. For the background to these appli-
cations see [1, 4, 8, 18, 19].
The efficiency of a fixed-rank transform is mainly characterized by two
parameters; the compression ratio (see [8]) and the accuracy of signal restora-
tion. The signal compression is realized through the following device. Let p
be the rank of the transform H. Then H can be represented in the form
254 P.G. Howlett et al.
(13.22) in such a way that the matrices B10 , · · · , Bn0 in (13.22) are estimated
by a scheme similar to the Gaussian elimination scheme in linear algebra. A
rank restriction can then be imposed on the matrices B10 , · · · , Bn0 that will
bring about reduction of the computational work in finding certain pseudo-
inverse matrices.
Extensions of the technique can be made in the following directions. First,
the method can be combined with a special iterative procedure to improve the
associated accuracy of the signal estimation. Secondly, an attractive extension
may be based on the representation of the operator T (13.6) in the form
13.8 Simulations
(1)
with i = 1, . . . , 16 and j = 1, . . . , 8, where each Rij is a matrix with entries
(2)
uniformly distributed over the interval (0, 1) and each Rij is a matrix with
normally distributed entries with mean 0 and variance 1. The symbol .∗
signifies Hadamard matrix multiplication.
13 New perspectives on optimal transforms of random vectors 255
The transforms T 0 and H have been applied to each pair Xij , Yij with the
same rank r = 8, that is, with the same compression ratio. The correspond-
ing covariance matrices have been estimated from the samples Xij and Yij .
Special methods for their estimation can be found, for example, in [3, 9, 10]
and [17].
Table 13.1 represents the values of ratios
Table 13.1 Ratios ρij of the error associated with the GKLT H to that of the transform
T 0 with the same compression ratios
↓i j → 1 2 3 4 5 6 7 8
1 5268.3 3880.6 1864.5 1094.7 2605.4 2878.0 4591.6 1052.7
2 2168.4 995.1 1499.7 338.6 1015.1 3324.0 2440.5 336.1
3 2269.3 803.5 158.4 136.4 66.7 2545.4 1227.1 326.6
4 1394.3 716.2 173.7 62.9 451.4 721.6 227.8 691.6
5 3352.4 1970.1 98.9 192.8 390.0 92.8 680.4 3196.8
6 1781.5 758.6 93.6 79.3 59.8 223.2 110.5 2580.8
7 2077.4 1526.0 67.4 30.3 172.5 70.3 1024.4 4749.3
8 3137.2 901.2 27.1 38.5 475.3 445.6 1363.2 2917.5
9 2313.2 117.0 18.0 39.3 180.6 251.0 1500.4 2074.2
10 1476.0 31.5 35.7 119.3 859.3 883.5 2843.1 3270.6
11 1836.7 35.3 36.4 1015.5 460.6 487.0 2843.1 8902.3
12 1808.5 74.5 38.2 419.0 428.0 387.2 2616.9 8895.3
13 1849.1 17.6 30.3 492.4 1175.5 135.8 1441.9 1649.2
14 2123.6 54.9 38.6 302.0 1310.5 2193.8 2681.5 1347.9
15 1295.1 136.3 31.8 711.1 2561.7 5999.2 550.7 996.0
16 2125.5 114.9 31.5 732.3 2258.2 5999.2 550.7 427.1
Inspection of Table 13.1 shows that, for the same compression ratio, the
transform T 0 has associated error varying from one part in 17.6 to one part
in 8,895.3 to that of the transform H.
We also applied our filter T (1) (constructed from Theorem 13.6.1) and the
†
optimal linear filter H (1) = Exy Eyy to the same signals and data as above,
that is, to each pair Xij , Yij with i = 1, · · · , 16 and j = 1, · · · , 8.
The errors associated with filters T (1) and H (1) are
where the matrices XT and XH have been constructed from the submatrices
XT ij ∈ R16×32 and XHij ∈ R16×32 correspondingly, that is, XT = {XT ij } ∈
256 P.G. Howlett et al.
50 50
100 100
150 150
200 200
250 250
50 100 150 200 250 50 100 150 200 250
(a) Given signals. (b) Observed signals.
50 50
100 100
150 150
200 200
250 250
50 100 150 200 250 50 100 150 200 250
(c) Reconstruction after filtering and compression (d) Reconstruction after filtering and compression
by the GKLT. by our transform with the same rank as that of
the GKLT.
50 50
100 100
150 150
200
200
250
250
50 100 150 200 250
50 100 150 200 250
(e) Estimates by the filter H(1). (f) Estimates by the filter T(1).
200
180
160
140
120
100
80
0 50 100 150 200 250
(a) Estimates of the 18th column in the matrix X.
250
200
150
100
50
T 0 (which have been applied to each of the subimages Xij , Yij with the same
compression ratio.
Figures 13.1(e) and (f) represent estimates of the noisy image in Figure
13.1(b) by filters H (1) and T (1) , respectively.
To illustrate the simulation results in a different way, we present typical
examples of the plots of a column estimate in matrix X by transforms H
and T 0 . Note that differences of the estimate by T 0 from the column plot are
almost invisible.
Table 13.1 and Figures 13.1 and 13.2 demonstrate the advantages of our
technique.
13.9 Conclusion
References
P. K. Pollett
Abstract We consider the problem of how best to assign the service capacity
in a queueing network in order to minimize the expected delay under a cost
constraint. We study systems with several types of customers, general service
time distributions, stochastic or deterministic routing, and a variety of ser-
vice regimes. For such networks there are typically no analytical formulae for
the waiting-time distributions. Thus we shall approach the optimal alloca-
tion problem using an approximation technique: specifically, the residual-life
approximation for the distribution of queueing times. This work generalizes
results of Kleinrock, who studied networks with exponentially distributed
service times. We illustrate our results with reference to data networks.
14.1 Introduction
Since their inception, queueing network models have been used to study a
wide variety of complex stochastic systems involving the flow and interaction
of individual items: for example, “job shops,” where manufactured items are
fashioned by various machines in turn [7]; the provision of spare parts for
collections of machines [17]; mining operations, where coal faces are worked
in turn by a number of specialized machines [12]; and delay networks, where
packets of data are stored and then transmitted along the communications
links that make up the network [18, 1]. For some excellent recent expositions,
which describe these and other instances where queueing networks have been
applied, see [2, 6] and the important text by Serfozo [16].
P. K. Pollett
Department of Mathematics, University of Queensland, Queensland 4072, AUSTRALIA
e-mail: [email protected]
Under our assumption that service times have arbitrary distributions, the
model is rendered intractable. In particular, there are no analytical formu-
lae for the delay distributions. We shall therefore adopt one of the many
approximation techniques. Consider a particular queue j and let Qj (x) be
the distribution function of the queueing time, that is, the period of time a
customer spends at queue j before its service begins. The residual-life approx-
imation, developed by the author [14], provides an accurate approximation
for Qj (x):
∞
Qj (x) $ Pr(nj = n)Gjn (x) , (14.1)
n=0
φj x
where Gj (x) = μj 0 (1 − Fj (y)) dy and Gjn (x) denotes the n-fold convo-
lution of Gj (x). The distribution of the number of customers nj at queue j,
which appears in (14.1), is that of a corresponding quasi-reversible net-
work [10]: specifically, a network of symmetric queues obtained by imposing
a symmetry condition at each queue j. The term residual-life approximation
comes from renewal theory; Gj (x) is the residual-life distribution correspond-
ing to the (lifetime) distribution Fj (x/φj ).
One immediate consequence of (14.1) is that the expected queueing time
Qj is approximated by Qj $ n̄j (1 + μ2j σj2 )/(2μj φj ), where n̄j is the expected
number of customers at queue j in the corresponding quasi-reversible net-
work. Hence the expected delay at queue j is approximated as follows:
1 1 + μ2j σj2
Wj $ + n̄j .
μj φj 2μj φj
264 P.K. Pollett
Under the residual-life approximation, it is only n̄j which changes when the
service discipline is altered. In the current context, the FCFS discipline, which
is assumed to be in operation everywhere in the network, is replaced by a
preemptive-resume last-come–first-served discipline, giving n̄j = aj /(φj − aj )
with aj = αj /μj , for each j, and hence
+ ,
1 1 + μ2j σj2 αj
Wj $ + . (14.2)
μj φj 2μj φj μj φj − αj
We now turn our attention to the problem of how best to apportion resources
so that the expected network delay, or equivalently (by Little’s theorem)
the expected number of customers in the network, is minimized. We shall
suppose that there is some overall network budget F (dollars) which cannot
be exceeded, and that the cost of operating queue j is a function fj of its
capacity. Suppose that the cost of operating queue j is proportional to φj ,
that is, fj (φj ) = fj φj (the units of fj are dollars per unit of capacity, or
dollar–seconds per unit of service). Thus we should choose the capacities
subject to the cost constraint
J
fj φj = F . (14.3)
j=1
where cj = μ2j σj2 is the squared coefficient of variation of the service time
distribution Fj (x). We seek to minimize m̄ over φ1 , . . . , φJ subject to (14.3).
To this end, we introduce a Lagrange multiplier 1/λ2 ; our problem then
becomes one of minimizing
⎛ ⎞
1 J
L(φ1 , . . . , φJ ; λ−2 ) = m̄ + 2 ⎝ fj φj − F ⎠ .
λ j=1
2fj φ4j − 4aj fj φ3j + 2aj (aj fj − λ2 )φ2j − 2j a2j λ2 φj + j a3j λ2 = 0 ,
2φ4j − 4aj φ3j + 2aj (aj − λ2 )φ2j − 2j a2j λ2 φj + j a3j λ2 = 0 , (14.4)
φ1 + φ2 + · · · + φJ = 1 . (14.5)
J
λ = λ0 + λ1k k + O(2 ), (14.6)
k=1
J
φj = φ0j + φ1jk k + O(2 ) , j = 1, . . . , J, (14.7)
k=1
where O(2 ) denotes terms of order i k . The zero-th order terms come from
√
Kleinrock’s solution: specifically, φ0j = aj + λ0 aj , j = 1, . . . , J, where λ0 =
J J √
(1 − k=1 ak )/( k=1 ak ). On substituting (14.6) and (14.7) into (14.4) we
obtain an expression for φ1jk in terms of λ1k , which in turn is calculated
using the constraint (14.5) and by setting k = δkj (the Kronecker delta). We
find that the optimal allocation, to first order, is
√ √
√ aj aj
φj = aj + λ0 aj − J √ bk k + 1 − J √ bj j , (14.8)
k=1 ak k =j k=1 ak
3/2 √ √
where bk = 14 λ0 ak (ak + 2λ0 ak )/(ak + λ0 ak )2 . For most practical appli-
cations, higher-order solutions are required. To achieve this we can simplify
matters by using a single perturbation = max1≤j≤J |j |. For each j we
define a quantity βj = j / and write φj and λ as power series in :
∞
∞
λ= λn n , φj = φnj n , j = 1, . . . , J. (14.9)
n=0 n=0
Substituting as before into (14.4), and using (14.5), gives rise to an iterative
scheme, details of which can be found in [13]. The first-order approximation is
useful, nonetheless, in dealing with networks whose service-time distributions
are all ‘close’ to exponential in the sense that their coefficients of variation do
not differ significantly from 1. It is also useful in providing some insight into
how the allocation varies as j , for fixed j, varies. Let φi , i = 1, 2, . . . , J, be
the new optimal allocation obtained after incrementing j by a small quantity
δ > 0. We find that to first order in δ
√
aj
φj − φj = 1 − J √ bj δ > 0,
k=1 ak
√
ai
φi − φi = − J √ (φj − φj ) < 0 , i = j.
k=1 ak
14.5 Extensions
So far we have assumed that the capacity does not depend on the state of
the queue (as a consequence of the FCFS discipline) and that the cost of
operating a queue is a linear function of its capacity. Let us briefly consider
some other possibilities. Let φj (n) be the effort assigned to queue j when there
are n customers present. If, for example, φj (n) = nφj /(n + η − 1), where η
is a positive constant, the zero-th order allocation, optimal under (14.3), is
precisely the same as before (the case η = 1). For values of η greater than 1
the capacity increases as the number of customers at queue j increases and
levels off at a constant value φj as the number becomes large. If we allow η
to depend on j we get a similar allocation but with the factor
fj aj fj ηj aj
J √ replaced by J √
k=1 fk ak k=1 fk ηk ak
(see Exercise 4.1.6 of [10]). The higher-order analysis is very nearly the same
as before. The factor 1 + cj is replaced by ηj (1 + cj ); for the sake of brevity,
we shall omit the details.
As another example, suppose that the capacity function is linear, that is,
φj (n) = φj n, and that service times are exponentially distributed. In this
case, the total number
J of customers in the system has a Poisson distribu-
tion with mean j=1 (aj /φj ) and it is elementary to show that the optimal
allocation subject to (14.3) is given by
fj aj
φj = J √ F, j = 1, . . . , J.
fj k=1 fk ak
J
fj log(gj φj ) = F
j=1
268 P.K. Pollett
to account for ‘decreasing costs’: costs become less with each increase in
capacity. Under this constraint, the optimal allocation is φj = λaj /fj , where
=
J
J
log λ = F− fk log(gk ak /fk ) fk .
k=1 k=1
be the ordered sequence of links used by messages on that route; smn is the
number of links and rmn (s) is the link used at stage s. Let αj (m, n, s) = νmn
if rmn
(s) =j, and0smnotherwise, so that the arrival rate at link j is given by
αj = m n =m s=1 αj (m, n, s), and the demand (in bits per second) by
aj = αj /μ. Assume that the system is stable (αj < μφj for each j). The
optimal capacity allocation (φj , j = 1, 2, . . . , J) can now be obtained using
the results of Section
14.4. For unit costs, the optimal allocation of capacity
√
(constrained by j φj = 1) satisfies μφj = αj + λ αj , j = 1, . . . , J, where
J J √
λ = (μ − k=1 αk )/( k=1 αk ), in the case of exponential transmission
times. More generally, in the case where the transmission times have an ar-
bitrary distribution with mean 1/μ and variance σ 2 , the optimal allocation
satisfies (to first order in )
√
√ αj J
μφj = αj + λ αj + cj − J √ ck , (14.10)
k=1 αk k=1
3/2 √ √
where ck = 14 λαk (αk + 2λ αk )/(αk + λ αk )2 and = μ2 σ 2 − 1.
To illustrate this, consider a symmetric star network , in which a collection
of identical outer nodes communicate via a single central node. Suppose that
there are J outer nodes and thus J communications links. The corresponding
queueing network, where the nodes represent the communications links, is a
fully connected symmetric network. Clearly there are J(J −1) routes, a typical
one being R(m, n) = {m, n}, where m = n. Suppose that transmission times
270 P.K. Pollett
have a common mean 1/μ and variance σ 2 (for simplicity, set μ = 1), and,
to begin with, suppose that transmission times are exponentially distributed
and that all traffic is offered at the same rate ν. Clearly the optimal allocation
will be φj = 1/J, owing to the symmetry of the network. What happens to
the optimal allocation if we alter the traffic offered on one particular route
by a small quantity? Suppose that we alter ν12 by setting ν12 = ν + e. The
arrival rates at links 1 and 2 will then be altered by the same amount e. Since
μ = 1 we will have a1 = a2 = ν + e and aj = ν for j = 3, . . . , J. The optimal
allocation is easy to evaluate. We find that, for j = 1, 2,
√
(1 − Jν − 2e) ν + e 1 1 (Jν + 1)
φj = ν + e + √ √ = + (J − 2) e + O(e2 ),
(J − 2) ν + 2 ν + e J 2 J 2ν
and for j = 3, . . . , J,
√
(1 − Jν − 2e) ν 1 Jν + 1
φj = ν + √ √ = − e + O(e2 ).
(J − 2) ν + 2 ν + e J J 2ν
1 1 (J 2 ν 2 − Jν + 2)(J 2 ν 2 − 2Jν − 1)
φj = + e + O(e2 )
J 2 J 2ν
and for 3 ≤ j ≤ J by
1 (J − 2)(J 2 ν 2 − Jν + 2)(J 2 ν 2 − 2Jν − 1)
φj = − e + O(e2 ).
J 4J 2 ν
Thus, to first order in e, there is an O(J 3 ) decrease in the capacity at all
links in the network, except at links 1 and 2, where there is an O(J 2 ) increase
in capacity. Indeed, the latter is true whenever the squared coefficient of
variation c is not equal to 1, for it is easily checked that φj = 1/J + gJ (c)e +
O(e2 ), j = 1, 2, and φj = 1/J − (J/2 − 1)gJ (c)e + O(e2 ), j = 3, . . . , J, where
Jν(Jν − 1)3 c − (J 4 ν 4 − 3J 3 ν 3 + 3J 2 ν 2 + Jν + 2)
gJ (c) = .
2J 2 ν
Clearly gJ (c) is O(J 2 ). It is also an increasing function of c, and so this accords
with our previous general results on varying the coefficient of variation of the
service-time distribution.
14 Optimal capacity assignment in general queueing networks 271
14.7 Conclusions
Acknowledgments I am grateful to Tony Roberts for suggesting that I adopt the per-
turbation approach described in Section 14.4. I am also grateful to Erhan Kozan for helpful
comments on an earlier draft of this chapter and to the three referees, whose comments
and suggestions did much to improve the presentation of my results. The support of the
Australian Research Council is gratefully acknowledged.
References
15.1 Introduction
Julia Piantadosi
C.I.A.M., University of South Australia, Mawson Lakes SA 5095, AUSTRALIA
e-mail: [email protected]
Phil Howlett
C.I.A.M., University of South Australia, Mawson Lakes SA 5095, AUSTRALIA
e-mail: [email protected]
regular release from the second dam that reflects the known demand for
water. We wish to find a control policy that releases an optimal flow of water
from the first dam to the second dam. In the first instance we have restricted
our attention to a very simple class of control policies. To calculate the cost of
a particular policy it is necessary to find an invariant measure. This measure
is found as the eigenvector of a large transposed transition matrix. A key
finding is that for our simple class of control policies the eigenvector of the
large matrix can be found from the corresponding eigenvector of a small block
matrix. The cost of a particular policy will depend on the expected volume of
water that is wasted and on the pumping costs. An appropriate cost function
will assist in determining an optimal pumping policy for our system.
This work will be used to analyze water cycle management in a suburban
housing development at Mawson Lakes in South Australia. The intention is to
capture and treat all stormwater entering the estate. The reclaimed water will
be supplied to all residential and commercial sites for watering of parks and
gardens and other non-potable usage. Since this is a preliminary investigation
we have been mainly concerned with the calculation of steady-state solutions
for different levels of control in a class of practical management policies. The
cost of each policy is determined by the expected volume of water lost through
overflow. We have ignored pumping costs. A numerical example is used to
illustrate the theoretical solution presented in the chapter. For underlying
methodology see [1, 2].
Consider a system with two connected dams, D1 and D2 , each of finite ca-
pacity. The content of the first dam is denoted by Z1 ∈ {0, 1, . . . , h} and
the content of the second dam by Z2 ∈ {0, 1, . . . , k}. We assume a stochastic
supply of untreated stormwater to the first dam and a regular demand for
treated stormwater from the second dam. The system is controlled by pump-
ing water from the first dam into the second dam. The input to the first
dam is denoted by X1 and the input to the second dam by X2 . We have
formulated a discrete-state model in which the state of the system, at time t,
is an ordered pair (Z1,t , Z2,t ) specifying the content of the two dams before
pumping. We will consider a class of simple control policies. If the content of
the first dam is greater than or equal to a specified level U1 = m, then we
will pump precisely m units of water from the first dam to the second dam.
If the content of the first dam is below this level we do not pump any water
into the second dam. The parameter m is the control parameter for the class
of policies we wish to study. We assume a constant demand for treated water
from the second dam and pump a constant volume U2 = 1 unit from the
second dam provided the dam is not empty. The units of measurement are
chosen to be the daily level of demand.
15 Control policy for stormwater management in two connected dams 275
• for the state (z1 , z2 ) where z1 < m and 0 < z2 we do not pump water
from the first dam but we do pump from the second dam. If n units of
stormwater enter the first dam then
• for the state (z1 , 0) where z1 ≥ m we pump m units from the first dam
into the second dam. If n units of stormwater enter the system then
• for the state (z1 , z2 ) where z1 ≥ m and 0 < z2 we pump m units from the
first dam into the second dam and pump one unit from the second dam to
meet the regular demand. If n units of stormwater enter the system then
If we order the states (z1 , z2 ) by the rules that (z1 , z2 ) ≺ (ζ1 , ζ2 ) if z2 < ζ2
and (z1 , z2 ) ≺ (ζ1 , z2 ) if z1 < ζ1 then the transition matrix can be written in
the form
⎡ ⎤
A 0 ··· 0 B 0 ··· 0 0 0
⎢ ⎥
⎢A 0 ··· 0 B 0 ··· 0 0 0 ⎥
⎢ ⎥
⎢0 A ··· 0 0 B ··· 0 0 0 ⎥
⎢ ⎥
⎢ ⎥
⎢. .. .. .. .. .. .. .. ⎥
⎢ .. . ··· . . . ··· . . . ⎥
⎢ ⎥
⎢ ⎥
⎢0 0 ··· A 0 0 ··· 0 B 0 ⎥
H(A, B) = ⎢
⎢
⎥.
⎥
⎢0 0 ··· 0 A 0 ··· 0 0 B⎥
⎢ ⎥
⎢0 0 ··· 0 0 A ··· 0 0 B⎥
⎢ ⎥
⎢ ⎥
⎢ .. .. .. .. .. .. .. .. ⎥
⎢. . ··· . . . ··· . . . ⎥
⎢ ⎥
⎢ ⎥
⎣0 0 ··· 0 0 0 ··· A 0 B⎦
0 0 ··· 0 0 0 ··· 0 A B
276 J. Piantadosi and P. Howlett
The block matrices A = [ai,j ] and B = [bi,j ] for i, j ∈ {0, 1, . . . , h} are defined
by
⎧
⎪
⎪ 0 for 1 ≤ i ≤ m − 1, j < i
⎪
⎪
⎪ pj
⎪ for i = 0, 0 ≤ j ≤ h − 1
⎨
ai,j = pj−i for 1 ≤ i ≤ m − 1, 1 ≤ j ≤ h − 1 and i ≤ j
⎪
⎪
⎪
⎪ ph−i + for j = h, 0 ≤ i ≤ m − 1
⎪
⎪
⎩
0 for m ≤ i ≤ h, 0 ≤ j ≤ h
and
⎧
⎪
⎪ 0 for 0 ≤ i ≤ m − 1, 0 ≤ j ≤ h
⎪
⎪
⎪
⎪ p for i = m, 0 ≤ j ≤ h − 1
⎨ j
bi,j = pj−i+m for m + 1 ≤ i ≤ h, 0 ≤ j ≤ h − 1 and i − m ≤ j
⎪
⎪
⎪
⎪ ph−i+m + for j = h, m ≤ i ≤ h
⎪
⎪
⎩
0 for m + 1 ≤ i ≤ h, j < i − m,
⎡ ⎤
0 C C 0 ··· 0
1 ⎢0 ··· 0⎥
⎢ 0 C ⎥
⎢ ⎥
. ⎢ .. .. .. .. .. ⎥
F = .. ⎢ . . . . . ⎥
⎢ ⎥
k−1 ⎢0 ··· C ⎥
⎣ 0 0 ⎦
k 0 0 0 ··· 0
15 Control policy for stormwater management in two connected dams 277
and
⎡ ⎤
0 0 0 0 ··· 0 ··· 0
.. ⎢ . .. .. . . ⎥
⎢ .. · · · .. · · · .. ⎥
. ⎢ . . ⎥
⎢ ⎥
m−1 ⎢0 0 0 ··· 0 ··· 0 ⎥
⎢ ⎥
⎢D ··· 0 ··· 0 ⎥
Gm = m ⎢ D 0 ⎥.
⎢ ⎥
m+1 ⎢0 0 D ··· 0 ··· 0 ⎥
⎢ ⎥
.. ⎢ . .. .. . . .. .. ⎥
⎢ .. . . ··· . ⎥
. ⎣ . . ⎦
k 0 0 0 ··· D ··· D
[F + Gm ]x = x
Gm [I − F ]−1 y = y. (15.2)
(I − F )−1 = I + F + F 2 + · · · ,
we deduce that
⎡ ⎤
P P C P C 2 · · · P C k−1 P C k
⎢0 I C · · · C k−2 C k−1 ⎥
⎢ ⎥
⎢ ⎥
⎢0 0 I · · · C k−3 C k−2 ⎥
⎢ ⎥
(I − F )−1 =⎢ . . .. . . .. .. ⎥ ,
⎢ .. .. . . ⎥
⎢ . . ⎥
⎢ ⎥
⎣0 0 0 ··· I C ⎦
0 0 0 ··· 0 I
⎡ ⎤ ⎡ ⎤
DP C m−1 DP C m
⎢ DC m−2 ⎥ ⎢ DC m−1 ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ .. ⎥ ⎢ .. ⎥
⎢ ⎥ ⎢ . ⎥
⎢
.
⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ D ⎥ ⎢ DC ⎥
S0 = ⎢ ⎥, ⎢
S1 = ⎢ ⎥,···
⎢ ⎥ ⎥
⎢
0
⎥ ⎢ D ⎥
⎢ ⎥ ⎢ ⎥
⎢ .
.. ⎥ ⎢ .. ⎥
⎢ ⎥ ⎢ . ⎥
⎢ ⎥ ⎢ ⎥
⎣ ⎦ ⎢ 0 ⎥
0 ⎣ ⎦
0 0
⎡ ⎤ ⎡ ⎤
DP C k−m DP C k−m+1
⎢ DC k−m−1 ⎥ ⎢ DC k−m ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ .. ⎥ ⎢ .. ⎥
⎢ . ⎥ ⎢ . ⎥
⎢ ⎥ ⎢ ⎥
⎢ DC k−2m+1 ⎥ ⎢ DC k−2m+2 ⎥
⎢ ⎥ ⎢ ⎥
· · · Sk−2m+1 =⎢ ⎥, Sk−2m+2 =⎢ ⎥,···
⎢ DC k−2m ⎥ ⎢ DC k−2m+1 ⎥
⎢ ⎥ ⎢ ⎥
⎢ .. ⎥ ⎢ .. ⎥
⎢ ⎥ ⎢ ⎥
⎢ . ⎥ ⎢ . ⎥
⎢ ⎥ ⎢ ⎥
⎣ DC ⎦ ⎣ DC 2 ⎦
D D(I + C)
and finally ⎡ ⎤
DP C k−1
⎢ DC k−2 ⎥
⎢ ⎥
⎢ ⎥
⎢ .. ⎥
⎢ . ⎥
⎢ ⎥
⎢ ⎥
⎢ DC k−m ⎥
Sk−m =⎢
⎢
⎥.
⎥
⎢ DC k−m−1 ⎥
⎢ ⎥
⎢ .. ⎥
⎢ . ⎥
⎢ ⎥
⎢ DC m ⎥
⎣ ⎦
D(I + C + · · · + C m−1 )
it can be seen that u = 0 and that our original problem has reduced to solving
the matrix equation
(I − S)v = 0. (15.3)
15 Control policy for stormwater management in two connected dams 279
Thus we must find the eigenvector for S corresponding to the unit eigenvalue.
To properly describe the elimination process we need to establish some suit-
able notation. We will write S = [Si,j ] where i, j ∈ {0, 1, . . . , (k − m)} and
⎧
⎪ DP C m−1+j for i = 0
⎪
⎪
⎪ DC m−1−i+j
⎪ for 1 ≤ i ≤ k − m − 1
⎪
⎪
⎪
⎨ and i − m + 1 ≤ j
Si,j =
⎪
⎪ D(I + C + · · · + C j−k+2m−1 ) for i = k − m
⎪
⎪
⎪
⎪ and k − 2m + 1 ≤ j
⎪
⎪
⎩
0 for m + j ≤ i.
We note that
⎡ ⎤
S0,j
⎢ ⎥
⎢ S1,j ⎥
Sj = ⎢
⎢ ..
⎥
⎥
⎣ . ⎦
Sk−m,j
and that 1T Sj = 1T for each j = 0, 1, . . . , k − m. Hence S is a stochastic
matrix. One of our key findings is that we can use Gaussian elimination to
further reduce the problem from one of finding an eigenvector for the large
matrix S ∈ IR(h+1)(k−m+1)×(h+1)(k−m+1) to one of finding the corresponding
eigenvector for a small block matrix in IR(h+1)×(h+1) .
For the special case when m = 1 we have the block matrix G1 with the
following structure:
⎡ ⎤
0 0 0 0 ··· 0
1⎢ ⎢D D 0 ··· 0 ⎥
⎥
⎢ ⎥
⎢ ⎥
G1 = 2 ⎢ 0 0 D · · · 0 ⎥ .
.. ⎢ . . . .
. . . .. . ⎥
⎥
. ⎢
⎣ . . . . . ⎦
k 0 0 0 ··· D
Therefore % &
−1
0 0
Gk (I − F ) = ,
RS
(I − DP )v = 0.
We wish to find the eigenvector corresponding to the unit eigenvalue for the
matrix S. We use Gaussian elimination in a block matrix format. During the
elimination we will make repeated use of the following elementary formulae.
15.3.1 Stage 0
After stage 0 of the elimination we have a new matrix T (1) = I − S (1) where
⎧
⎪
⎪ 0 for j = 0
⎪
⎪
⎪
⎪ DP C m−1 (W0 C)C j−1 for i = 0, 1 ≤ j
⎪
⎪
⎪
⎪
⎪
⎪ DC m−1−i (W0 C)C j−1 for 1 ≤ i ≤ m − 1, 1 ≤ j
⎪
⎨
(1) DC m−1−i+j for m ≤ i ≤ k − m − 1
Si,j =
⎪
⎪ and i − m + 1 ≤ j
⎪
⎪
⎪
⎪
⎪
⎪ D(I + C + · · · + C j−k+2m−1 ) for i = k − m
⎪
⎪
⎪
⎪ and k − 2m + 1 ≤ j
⎪
⎩
0 for m + j ≤ i.
Note that column 0 is reduced to a zero column and that row 0 is fixed for all
subsequent stages. We therefore modify T (1) by dropping both column and
row 0.
After stage p−1 of the elimination, for 1 ≤ p ≤ m−2, we have T (p) = I −S (p)
where
282 J. Piantadosi and P. Howlett
⎧
⎪
⎪ 0 for j =p−1
⎪
⎪ p−1
⎪
⎪ DC m−p t=0 (Wt C)C j−p for i = p − 1, p ≤ j
⎪
⎪ p−1
⎪
⎪ DC m−1−i t=0 (Wt C)C j−p for p ≤ i ≤ m − 1, p ≤ j
⎪
⎨ p−1
(p) D t=i−m+1 (Wt C)C j−p for m ≤ i ≤ m + p − 2, p ≤ j
Si,j =
⎪
⎪ DC m−1−i+j for m+p−1≤i≤k−m−1
⎪
⎪
⎪
⎪ and i−m+1≤j
⎪
⎪ j−k+2m−1 t
⎪
⎪ D t=0 C for i = k − m, k − 2m + 1 ≤ j
⎪
⎩0 for m + j ≤ i.
Since column p is reduced to a zero column and row p is fixed for all subse-
quent stages we modify T (p+1) by dropping both column and row p.
15 Control policy for stormwater management in two connected dams 283
15.3.3 Stage m − 1
We modify T (p) by dropping both column and row p − 1 and consider a sub-
matrix M using the (p, p), (r, p), (m+p−1, p), (p, s), (r, s) and (m+p−1, s)th
elements, where p + 1 ≤ r ≤ m + p − 2 and p + 1 ≤ s. We have
⎡ p−1 p−1 ⎤
I − D t=p−m+1 (Wt C) −D t=p−m+1 (Wt C)C s−p
⎢ p−1 ⎥
M = ⎣ −D p−1 t=r−m+1 (Wt C) Iδr,s − D t=r−m+1 (Wt C)C s−p ⎦
−D Iδm+p−1,s − DC s−p
15.3.5 Stage k − 2m + 1
We modify T (p) by dropping both column and row p − 1 and consider a sub-
matrix M using the (p, p), (r, p), (k − m, p), (p, s), (r, s) and (k − m, s)th
elements, where p + 1 ≤ r ≤ k − m and p + 1 ≤ s. We have M given by
⎡ p−1 p−1 ⎤
I − D t=p−m+1 (Wt C) −D t=p−m+1 (Wt C)C s−p
⎢ −D p−1 p−1
⎣ t=r−m+1 (Wt C) Iδr,s − D 6t=r−m+1 (Wt C)C s−p 7 ⎥
⎦
s−p t
−DXp−1 Iδk−m,s − D t=0 C + X p C s−p
⎡ p ⎤
I −D t=p−m+1 (Wt C)C s−p−1
⎢0 p
M →⎣ Iδr,s −6D t=r−m+1 (Wt C)C s−p−1 7 ⎥
⎦.
s−p−1
0 Iδk−m,s − D t=0 C t + Xp+1 C s−p−1
⎧
⎪
⎪ 0 for j = p
⎪ p
⎪
⎪
⎪ D t=i−m+1 (Wt C)C j−p−1 for p + 1 ≤ i ≤ k − m − 1
⎪
⎨
(p+1) 6 and p + 1 ≤ j
Si,j = j−p−1 t
⎪
⎪ D C
⎪
⎪
t=0
9
⎪
⎪ +Xp+1 C j−p−1
for i = k − m, p + 1 ≤ j
⎪
⎩
0 for m + j ≤ i.
Hence
k−m−2 k−m−2
(k−m−1) I − D t=k−2m (Wt C) −D t=k−2m (Wt C)C
T = .
−DXk−m−1 I − D[I + (Xk−m−1 C)]
After the Gaussian elimination has been completed the system reduces to an
equation for v0 ,
k−m
m−1
v0 = DP C (W0 C) C j−1 vj ,
j=1
(I − DXk−m )vk−m = 0.
288 J. Piantadosi and P. Howlett
We begin by solving the final equation to find vk−m . The penultimate equa-
tion now shows us that
% k−m−1 &
vk−m−1 = D (Wt C) vk−m .
t=k−2m
k−m
C j−q−1 vj
j=q
k−m
= vq + C C j−q−2 vj
j=q+1
% & %k−m−1 &-
k−m−1
= D (Wt C) + C (Wt C) vk−m
t=q−m+1 t=q+1
% & - %k−m−1 &
q−1
= D (Wt C) Wq + I ×C (Wt C) vk−m
t=q−m+1 t=q+1
%k−m−1 &
= (Wt C) vk−m
t=q
and hence
% & k−m
q−1
vq−1 = D (Wt C) C j−q vj
t=q−m j=q
%k−m−1 &
=D (Wt C) vk−m .
t=q−m
k−m
C j−m+1 vj
j=m−1
k−m
= vm−1 + C C j−m vj
j=m
%k−m−1 & %k−m−1 &-
= D (Wt C) + C (Wt C) vk−m
t=0 t=m
%m−2 & - %k−m−1 &
= D (Wt C) Wm−1 + I × C (Wt C) vk−m
t=0 t=m
%k−m−1 &
= (Wt C) vk−m
t=m−1
k−m
C j−p−1 vj
j=p
k−m
= vp + C C j−p−2 vj
j=p+1
%p−1 & - %k−m−1 &
= DC m−p−1 (Wt C) Wp + I ×C (Wt C) vk−m
t=0 t=p+1
%k−m−1 &
= (Wt C) vk−m
t=p
290 J. Piantadosi and P. Howlett
and hence
%p−1 & k−m
vp−1 = DC m−p (Wt C) C j−q vj
t=0 j=p
%k−m−1 &
= DC m−p (Wt C) vk−m .
t=0
for p = 0, 1, . . . , m − 2 and
% &
k−m−1
vq = D (Wt C) vk−m (15.6)
t=q−m+1
(I − D)vk = 0.
We will show that (I − D)−1 is well defined and hence deduce that vk = 0.
Since
0 L2
D=
0 M2
it follows that (I − D)−1 is well defined if and only if (I − M2 )−1 is well
defined. We have the following result.
15 Control policy for stormwater management in two connected dams 291
1T M2 = [p1 + , 1, 1, . . . , 1]
1T M2 r = [α0 , α1 , . . . , αr−1 , 1, 1, . . . , 1]
A similar argument shows that the j th element is less than 1 for all j ≤ q
and indeed, for the critical case j = q, the j th element is given by
αq−1 p0 + p1 + = βq < 1.
The remaining elements for q < j ≤ k are easily seen to be equal to 1. Hence
the hypothesis is also true for r = q + 1. By induction it follows that
1T M2 k+1 < 1T .
292 J. Piantadosi and P. Howlett
Hence (I − M2 k+1 )−1 is well defined. The matrix (I − M2 )−1 is now defined
by the identity above. This completes the proof.
(I − DP )v0 = 0.
(I − DP )v0 = 0.
where ⎡1 1 1 1 1 1 1
⎤
2 4 8 16 32 64 64
⎢ ⎥
⎢0 1 1 1 1 1 1 ⎥
⎢ 2 4 8 16 32 32 ⎥
⎢ ⎥
⎢0 0 0 0 0 0 0 ⎥
⎢ ⎥
⎢ ⎥
A=⎢ ⎥
⎢0 0 0 0 0 0 0 ⎥
⎢ ⎥
⎢0 0 0 0 0 0 0 ⎥
⎢ ⎥
⎢ ⎥
⎢0 0 0 0 0 0 0 ⎥
⎣ ⎦
0 0 0 0 0 0 0
and ⎡ ⎤
0 0 0 0 0 0 0
⎢ ⎥
⎢0 0 0 0 0 0 0 ⎥
⎢ ⎥
⎢1 1 ⎥
⎢2 1 1 1 1 1
64 ⎥
⎢ 4 8 16 32 64 ⎥
⎢ 1 ⎥
B =⎢0 1 1 1 1 1
32 ⎥ .
⎢ 2 4 8 16 32 ⎥
⎢ 1 1 1 1 1 ⎥
⎢0 0 16 ⎥
⎢ 2 4 8 16 ⎥
⎢ 1 1 1 1 ⎥
⎣0 0 0 2 4 8 8 ⎦
1 1 1
0 0 0 0 2 4 4
(I − S)v = 0,
where S = [Si,j ] for i, j = {0, 1, 2}. Using the elimination process described
in Section 15.3 we find the reduced coefficient matrix
⎡ ⎤
I −W0 D(I − C)−1 C 2 −W0 D(I − C)−1 C 3
(I − S) → ⎣ 0 I −W1 DW0 C 2 ⎦,
0 0 I − D(I + W1 C)
where
⎡ 447 491 1
⎤
2296 2296 2 0 0 0 0
⎢ ⎥
⎢ 1021 1065 1 1
0⎥
⎢ 4592 4592 4 2 0 0 ⎥
⎢ ⎥
⎢ 1849 ⎥
⎢ 9184 1805 1 1 1
0 0⎥
⎢ 9184 8 4 2 ⎥
⎢ 2677 ⎥
D(I + W1 C) = ⎢ 0⎥
2545 1 1 1 1
⎢ 18368 18368 16 8 4 2 ⎥.
⎢ ⎥
⎢ 619 575 1 1 1 1 1⎥
⎢ 5248 5248 32 16 8 4 2⎥
⎢ ⎥
⎢ 619 575 1 1 1 1 1⎥
⎢ 10496 4⎥
⎣ 10496 64 32 16 8
⎦
619 575 1 1 1 1 1
10496 10496 64 32 16 8 4
294 J. Piantadosi and P. Howlett
(I − D(I + W1 C))v2 = 0
where f [z(s)|r] is the overflow from state z(s) when r units of stormwater
enter the first dam. We will consider the same pumping policy for four differ-
ent values m = 1, 2, 3, 4 of the control parameter. We obtain the steady-state
vector x = x[m] for each particular value of the control parameter and deter-
mine the expected total overflow in each case. Table 15.1 compares the four
parameter values by considering the overflow Ji = Ji [m] from the first and
second dams.
From the table it is clear that the first pumping policy results in less
overflow from the system. If pumping costs are ignored then it is clear that
the policy m = 1 is the best. Of course, in a real system there are likely to
15 Control policy for stormwater management in two connected dams 295
1 1 1 1
J1 7 25 25 17
9053 263 57
J2 0 62375 1350 272
1 11548 317 73
Total 7 62375 1350 272
and
296 J. Piantadosi and P. Howlett
⎡ ⎤
pm · · · p1 0 ··· 0
⎢ pm+1 · · · p p ··· 0 ⎥
⎢ 2 1 ⎥
⎢ .. . . .. ⎥
M2 = ⎢ . · · · .
. .
. ··· . ⎥.
⎢ ⎥
⎣ ph−1 · · · ph−m ph−m−1 · · · pm−1 ⎦
p+
h · · · p+ +
h−m+1 ph−m · · · p+m
Provided pj > 0 for all j ≤ h the matrix L1 is strictly sub-stochastic [3] with
1T L1 < 1T . It follows that (I − L1 )−1 is well defined and hence
(I − L1 )−1 0
P = (I − C)−1 =
M1 (I − L1 )−1 I
Proof.
1T D + 1T C = 1T ⇒ 1T D = 1T (I − C) ⇒ 1T D(I − C)−1 = 1T .
Hence
L1 m−1 0
1T DP C m−1 = 1T C m−1 = [1T , 1T ]
M1 L1 m−2 0
3 T m−1 4
= 1 L1 + 1T M1 Lm−2
1 , 0T
3 4
= (1T L1 + 1T M1 )Lm−2 1 , 0T
3 4
= 1T Lm−21 , 0T < 1T
We have
is well defined.
Lemma 4. With the above definitions the matrix Wp is well defined for each
p = 0, 1, . . . , m − 1 and for each such p we have
⎡ ⎤
m−1−p
p−1
T
1 D ⎣ C j⎦
(Wt C) ≤ 1T C p .
j=0 t=0
m−2
1T D (Wt C) ≤ 1T C m−1 .
t=0
and hence ⎡ ⎤
m−2
1T D ⎣ C j ⎦ = 1T [I − DP C m−1 ],
j=0
Thus the result is true for p = 1. Suppose the result is true for 1 ≤ s ≤ p − 1.
Then
p−2
1T DC m−p (Wt C) ≤ 1T C p−1 < 1T
t=0
and hence
Wp−1 = [I − DC m−p (W0 C) · · · (Wp−2 C)]−1
is well defined. Now we have
⎡ ⎤
m−1−p
p−2
p−2
1T D ⎣ Cj⎦ (Wt C) + (1T C p−1 )DC m−p (Wt C)
j=0 t=0 t=0
⎡ ⎤
m−p
p−2
≤ 1T D ⎣ Cj⎦ (Wt C)
j=0 t=0
≤ 1T C p−1
and hence
⎡ ⎤
m−1−p
p−2
p−2
1T D ⎣ Cj⎦ (Wt C) ≤ (1T C p−1 )[I − DC m−p (Wt C)],
j=0 t=0 t=0
Thus the result is also true for s = p. This completes the proof.
m−2
C j + C m−1 P = P
j=0
and hence ⎡ ⎤
m−2
1T D ⎣ C j ⎦ (W0 C) = 1T C,
j=0
Therefore the JP identity of the first kind is valid for p = 1. We will use
induction to establish the general identity. Let p > 1 and suppose the result
is true for s = p < m − 1. From
⎧ ⎡ ⎤ ⎡ ⎤ ⎫
⎨
p−1
p−1
m−p−1
p−1 ⎬
1T D I + ⎣ (Wt C)⎦ + ⎣ Cj⎦ (Wt C) = 1T
⎩ ⎭
j=1 t=p−j j=0 t=0
we deduce that
⎧ ⎡ ⎤ ⎡ ⎤ ⎫
⎨
p−1
p−1
m−p−2
p−1 ⎬
1T D I + ⎣ (Wt C)⎦ + ⎣ Cj⎦ (Wt C)
⎩ ⎭
j=1 t=p−j j=0 t=0
= 1 [I − DC
T m−p−1
(W0 C) · · · (Wp−1 C)]
300 J. Piantadosi and P. Howlett
= 1T C + 1T D = 1T .
Hence the result is also true for s = p + 1. This completes the proof.
Lemma 6. The matrix Wq exists for q = m, m + 1, . . . , k − m − 1 and for
each such q the JP identities of the second kind
⎧ ⎡ ⎤⎫
⎨ q−1
m−1 ⎬
1T D I + ⎣ (Wt C)⎦ = 1T (15.3)
⎩ ⎭
j=1 t=q−j
Therefore
⎧ ⎡ ⎤ ⎫
⎨
m−3 ⎬
1T D I + ⎣ C j ⎦ (W0 C) + 1T DC m−2 (W0 C) = 1T
⎩ ⎭
j=0
and hence
⎧ ⎡ ⎤ ⎫
⎨
m−3 ⎬
1T D I +⎣ C j ⎦ (W0 C) = 1T [I − DC m−2 (W0 C)],
⎩ ⎭
j=0
15 Control policy for stormwater management in two connected dams 301
then we have
⎧ ⎡ ⎤ ⎫
⎨
m−p−1 ⎬ p−2
p−2
1T D I + ⎣ C j ⎦ (W0 C) (Wt C) + (1T C p−2 )DC m−p (Wt C)
⎩ ⎭
j=0 t=1 t=0
≤1 CT p−2
and hence
⎧ ⎡ ⎤ ⎫
⎨
m−p−1 ⎬ p−2
1T D I +⎣ C j ⎦ (W0 C) (Wt C)
⎩ ⎭
j=0 t=1
p−2
≤ (1T C p−2 )[I − DC m−p (Wt C)],
t=0
m−1
1T D (Wt C) ≤ 1T C m−1 .
t=1
and hence
⎧ ⎡ ⎤⎫
⎨ q−1
m−2 ⎬
1T D I + ⎣ (Wt C)⎦ = 1T [I − D(Wq−m+1 C) · · · (Wq−1 C)].
⎩ ⎭
j=1 t=q−j
which we rewrite as
⎧ ⎡ ⎤⎫
⎨m−1
q ⎬
1T D ⎣ (Wt C)⎦ = 1T C
⎩ ⎭
j=1 t=q+1−j
Hence the JP identity of the second kind is valid for s = q + 1. To show that
Wq+1 is well defined, we must consider two cases. When q ≤ 2m − 2 we set
p = q − m + 2 in the JP identity of the first kind to give
⎡ ⎡ ⎤ ⎡ ⎤ ⎤
q+1−m
q+1−m
2m−q−1
q+1−m
1T D ⎣I + ⎣ (Wt C)⎦+⎣ C j⎦ (Wt C)⎦ = 1T .
j=1 t=q−m+2−j j=0 t=0
Therefore
⎧ ⎡ ⎤ ⎡ ⎤ ⎫
⎨
q+1−m
q+1−m
2m−q−2
q+1−m ⎬
1T D I + ⎣ (Wt C)⎦ + ⎣ Cj⎦ (Wt C)
⎩ ⎭
j=1 t=q−m+2−j j=0 t=0
= 1 [I − DC
T 2m−q−1
(W0 C) · · · (Wq+1−m C)]
and hence
⎧ ⎡ ⎤
⎨
q+1−m
q+1−m
1T D I+ ⎣ (Wt C)⎦
⎩
j=1 t=q−m+2−j
⎡ ⎤ ⎫
2m−q−2
q+1−m ⎬
+⎣ Cj⎦ (Wt C) (Wq−m+2 C) = 1T C.
⎭
j=0 t=0
q−m+2
+ (1T C)DC 2m−q−2 (Wt C) ≤ 1T C
t=0
304 J. Piantadosi and P. Howlett
If we continue this process the terms of the second sum on the left-hand side
will be eliminated after 2m − q steps, at which stage we have
⎧ ⎡ ⎤⎫
⎨
q+1−m
q+1−m ⎬ m−1
T
1 D I+ ⎣ (Wt C) ⎦ (Wt C) ≤ 1T C 2m−q .
⎩ ⎭
j=1 t=q−m+2−j t=q−m+2
The details change slightly but we continue the elimination process. Since
(1T C 2m−q ) < 1T we now have
⎧ ⎡ ⎤⎫
⎨
q−m
q+1−m ⎬ m−1
1T D I + ⎣ (Wt C)⎦ (Wt C)
⎩ ⎭
j=1 t=q−m+2−j t=q−m+2
m−1
+ (1T C 2m−q )D (Wt C) ≤ 1T C 2m−q
t=1
and hence
⎧ ⎡ ⎤⎫
⎨
q−m
q+1−m ⎬
m−1
1T D I+ ⎣ (Wt C)⎦ (Wt C)
⎩ ⎭
j=1 t=q−m+2−j t=q−m+2
The elimination continues in the same way until we eventually conclude that
q
T
1 D (Wt C) ≤ 1T C m−1 < 1T
t=q−m+2
15.9 Summary
References
1. D. R. Cox and H. D. Miller, The Theory of Stochastic Processes (Methuen & Co.,
London, 1965).
2. F. R. Gantmacher, The Theory of Matrices (Chelsea Publishing Company, New York,
1960).
3. D. L. Isaacson and R. W. Madsen, Markov Chains: Theory and Applications (John
Wiley & Sons, New York, 1976).
4. P. A. P. Moran, A probability theory of dams and storage systems, Austral. J. Appl.
Sci. 5 (1954), 116–124.
5. P. A. P. Moran, An Introduction to Probability Theory (Clarendon Press, Oxford,
1968).
6. P. A. P. Moran, A Probability Theory of Dams and Storage Systems (McGraw-Hill,
New York, 1974).
7. G. F. Yeo, A finite dam with exponential release, J. Appl. Probability 11 (1974),
122–133.
8. G. F. Yeo, A finite dam with variable release rate, J. Appl. Probability 12 (1975),
205–211.
Chapter 16
Optimal design of linear
consecutive–k–out–of–n systems
Malgorzata O’Reilly
16.1 Introduction
Malgorzata O’Reilly
School of Mathematics and Physics, University of Tasmania, Hobart TAS 7001,
AUSTRALIA
e-mail: [email protected]
More examples of these systems are in [2, 19, 34, 35, 38]. For a review of
the literature on consecutive–k–out–of–n systems the reader is referred to [8].
Also see [5] by Chang, Cui and Hwang.
Introducing more general assumptions and considering system topology
has led to some generalizations of consecutive–k–out–of–n systems. These
are listed below:
• consecutively connected systems [32];
• linearly connected systems [6, 7, 14];
• consecutive–k–out–of–m–from–n: F systems [36];
• consecutive–weighed–k–out–of–n: F systems [37];
• m–consecutive–k–out–of–n: F systems [15];
• 2–dimensional consecutive–k–out–of–n: F systems [31];
• connected–X–out–of–(m, n): F lattice systems [3];
• connected–(r, s)–out–of–(m, n): F lattice systems [27];
16 Optimal design of linear consecutive–k–out–of–n systems 309
k F System G System
k=1 ω ω
k=2 (1, n, 3, n − 2, . . . , −
n − 3, 4, n − 1, 2)
2 < k < n/2 − −
n/2 ≤ k < n − 2 − (1, 3, . . . , 2(n − k) − 1, ω,
2(n − k), . . . , 4, 2)
k =n−2 (1, 4, ω, 3, 2) (1, 3, . . . , 2(n − k) − 1, ω,
2(n − k), . . . , 4, 2)
k =n−1 (1, ω, 2) (1, 3, . . . , 2(n − k) − 1, ω,
2(n − k), . . . , 4, 2)
k=n ω ω
In the case when n ≥ 2k, a useful concept has been that of singularity,
which has been also applied in invariant optimal designs [13]. A design
X = (q1 , q2 , . . . , qn ) is singular if for symmetrical components qi and qn+1−i ,
1 ≤ i ≤ [n/2], either qi > qn+1−i or qi < qn+1−i for all i; otherwise the
design is nonsingular. According to Shen and Zuo [33] a necessary condi-
tion for the optimal design of a linear consecutive–k–out–of–n:G system with
16 Optimal design of linear consecutive–k–out–of–n systems 311
In this chapter we treat the case 2k + 2 ≤ n ≤ 3k and explore whether the re-
sults of Shen and Zuo [33] and O’Reilly [28] can be extended to this case. The
proofs included here are more complicated and the produced results do not
exactly mirror those when 2k ≤ n ≤ 2k + 1. We find that, although the nec-
essary conditions for the optimal design of linear consecutive–k–out–of–n:F
systems in the cases 2k ≤ n ≤ 2k +1 and 2k +2 ≤ n ≤ 3k are similar, the pro-
cedures to improve designs not satisfying this necessary condition differ in the
choice of interchanged components. Furthermore, the necessary conditions for
linear consecutive–k–out–of–n:G systems in these two cases are significantly
different. In the case when 2k + 2 ≤ n ≤ 3k, the requirement for the opti-
mal design of a linear consecutive–k–out–of–n:G system to be singular holds
only under certain limitations. Examples of nonsingular and singular optimal
designs are given. The theorems are built on three subsidiary propositions,
which are given in Sections 16.2 and 16.4. Proposition 16.4.1 itself requires
some supporting lemmas which are the substance of Section 16.3. The main
results for this case are presented in Section 16.5. The ideas are related to
those in the existing literature, though the detail is somewhat complicated.
The arguments are constructive and based on the following.
Suppose X ≡ (q1 , . . . , q2k+m ) is a design and {qi1 , . . . , qir } is an arbitrary
A(X) ≡ qs ,
s∈{i1 ,...,ir }
A (X) ≡ q2k+m+1−s ,
s∈{i1 ,...,ir }
Bl (X) ≡ qs and
s∈{l,...,k}\{i1 ,...,ir }
Bl (X) ≡ q2k+m+1−s ,
s∈{l,...,k}\{i1 ,...,ir }
Propositions 16.2.1 and 16.2.2 below contain results for Mt and RT , which
are later used to prove a result for W0 in Theorem 16.5.1 of Section 16.5. In
the proofs of Propositions 16.2.1, 16.2.2 and 16.4.1 and Theorem 16.5.1 we
assume q1 > q2k+m . Note that, by the symmetry of the formulas, reversing
the order of components in X and X ∗ would not change the values of Wt , Mt
and Rt for X and X ∗ . Therefore the assumption q1 > q2k+m can be made
without loss of generality.
for any X ∗ .
Proof. Without loss of generality we can assume q1 > q2k+m . Note that
m ≥ 2 and so t + 1 < m for all 0 ≤ t ≤ T . We have
Mt = ABt+1 ⊕ A∗ Bt+1 ,
Mt∗ = A∗ Bt+1 ⊕ ABt+1 ,
Mt − Mt∗ = (A − A∗ )(Bt+1 − Bt+1 ),
where A − A∗ > 0 and Bt+1 − Bt+1 > 0 by the singularity of X, and so
Mt − Mt∗ > 0,
RT > RT∗
for any X ∗ .
Proof. Without loss of generality we can assume q1 > q2k+m . Note that
m ≥ 2 and so T + 2 ≤ m. We have
RT = pk+T +1 qk+m−T qT +1 ABT +2 ⊕ A∗ BT +2
+ qk+T +1 pk+m−T q2k+m−T A∗ BT +2 ⊕ ABT +2 ,
RT∗ = pk+T +1 qk+m−T qT +1 A∗ BT +2 ⊕ ABT +2
+ qk+T +1 pk+m−T q2k+m−T ABT +2 ⊕ A∗ BT +2
16 Optimal design of linear consecutive–k–out–of–n systems 315
and
RT − RT∗ = qk+T +1 pk+m−T (BT +2 − q2k+m−T BT +2 )(A − A∗ )
− qk+m−T pk+T +1 (BT +2 − qT +1 BT +2 )(A − A∗ ),
and so
RT − RT∗ > 0,
proving the proposition.
Lemmas 16.3.1–16.3.3 below contain some preliminary results which are used
in the proof of a result for Rt in Proposition 16.4.1 of Section 16.4.
Lemma 16.3.1 Let X ≡ (q1 , . . . , q2k+m ) be singular, 2 < m ≤ k, with either
m = 2T + 1 for some T > 0 or m = 2T + 2 for some T > 0. Then for any
X ∗ and all 0 ≤ t ≤ T − 1 we have
k
m−1
qs = ABm qs , (16.1)
s=t+1 s=t+1
m+k−3t−3
F̄k−t−1 (qk+t+2 , . . . , qk+m−t−1 , qk+m+1 , . . . , q2k+m−t−1 )
2(m−2t−2)
= F̄m−2t−2 (qk+t+2 , . . . , qk+m−t−1 , q2k+t+2 , . . . , q2k+m−t−1 ) ·
2k+t+1
qs
s=k+m+1
2(m−2t−2)
= F̄m−2t−2 (qk+t+2 , . . . , qk+m−t−1 , q2k+t+2 , . . . , q2k+m−t−1 ) ·
2k+t+1
A∗ Bm qs , (16.2)
s=2k+2
316 M. O’Reilly
k
m+k−3t−3
F̄k−t−1 (qk+t+2 , . . . , qk+m−t−1 , qk+m+1 , . . . , q2k+m−t−1 ) qs
s=t+1
2(m−2t−2)
= F̄m−2t−2 (qk+t+2 , . . . , qk+m−t−1 , q2k+t+2 , . . . , q2k+m−t−1 ) ·
m−1 2k+t+1
∗
ABm A Bm qs qs , (16.3)
s=t+1 s=2k+2
m−1
k
qs∗ = qs A∗ Bm , (16.4)
s=t+1 s=t+1
2(m−2t−2) ∗ ∗ ∗ ∗
F̄m−2t−2 (qk+t+2 , . . . , qk+m−t−1 , qk+m−1 , . . . , q2k+m−t−1 )
2(m−2t−2)
= F̄m−2t−2 (qk+t+2 , . . . , qk+m−t−1 , q2k+t+2 , . . . , q2k+m−t−1 ) ·
2k+t+1
ABm qs , (16.5)
s=2k+2
k
m+k−3t−3 ∗ ∗ ∗ ∗
F̄k−t−1 (qk+t+2 , . . . , qk+m−t−1 , qk+m+1 , . . . , q2k+m−t−1 ) qs∗
s=t+1
2(m−2t−2)
= F̄m−2t−2 (qk+t+2 , . . . , qk+m−t−1 , q2k+t+2 , . . . , q2k+m−t−1 ) ·
m−1 2k+t+1
ABm A∗ Bm qs qs (16.6)
s=t+1 s=2k+2
and
k
m+k−3t−3
F̄k−t−1 (qk+t+2 , . . . , qk+m−t−1 , qk+m+1 , . . . , q2k+m−t−1 ) qs
s=t+1
k
m+k−3t−3 ∗ ∗ ∗ ∗
= F̄k−t−1 (qk+t+2 , . . . , qk+m−t−1 , qk+m+1 , . . . , q2k+m−t−1 ) qs∗ .
s=t+1
(16.7)
2k+m−t
2k+m−t
qs = A∗ Bm qs , (16.8)
s=k+m+1 s=2k+2
m+k−3t−3
F̄k−t−1 (qt+2 , . . . , qk , qk+t+2 , . . . , qk+m−t−1 )
2(m−2t−2)
k
= F̄m−2t−2 (qt+2 , . . . , qm−t−1 , qk+t+2 , . . . , qk+m−t−1 ) qs
s=m−t
2(m−2t−2)
m−1
= F̄m−2t−2 (qt+2 , . . . , qm−t−1 , qk+t+2 , . . . , qk+m−t−1 )ABm qs ,
s=m−t
(16.9)
2k+m−t
m+k−3t−3
F̄k−t−1 (qt+2 , . . . , qk , qk+t+2 , . . . , qk+m−t−1 ) qs
s=k+m+1
(qt+2 , . . . , qm−t−1 , qk+t+2 , . . . , qk+m−t−1 )ABm A∗ Bm ·
2(m−2t−2)
= F̄m−2t−2
2k+m−t m−1
qs qs , (16.10)
s=2k+2 s=m−t
2k+m−t
2k+m−t
qs∗ = ABm qs , (16.11)
s=k+m+1 s=2k+2
m+k−3t−3 ∗
F̄k−t−1 (qt+2 , . . . , qk∗ , qk+t+2
∗ ∗
, . . . , qk+m−t−1 )
m−1
(qt+2 , . . . , qm−t−1 , qk+t+2 , . . . , qk+m−t−1 )A∗ Bm
2(m−2t−2)
= F̄m−2t−2 qs ,
s=m−t
(16.12)
2k+m−t
m+k−3t−3 ∗
F̄k−t−1 (qt+2 , . . . , qk∗ , qk+t+2
∗ ∗
, . . . , qk+m−t−1 ) qs∗
s=k+m+1
F̄m−2t−2 (qt+2 , . . . , qm−t−1 , qk+t+2 , . . . , qk+m−t−1 )ABm A∗ Bm
2(m−2t−2)
= ·
2k+m−t
m−1
qs qs (16.13)
s=2k+2 s=m−t
318 M. O’Reilly
and
2k+m−t
m+k−3t−3
F̄k−t−1 (qt+2 , . . . , qk , qk+t+2 , . . . , qk+m−t−1 ) qs
s=k+m+1
2k+m−t
m+k−3t−3 ∗
= F̄k−t−1 (qt+2 , . . . , qk∗ , qk+t+2
∗ ∗
, . . . , qk+m−t−1 ) qs∗ .
s=k+m+1
(16.14)
Proof. Note that Lemma 16.3.1 is also true for designs reversed to X and
X ∗ , that is, for Xr = (q2k+m , . . . , q1 ) and Xr∗ = (q2k+m
∗
, . . . , q1∗ ). Therefore
all equalities of Lemma 16.3.2 are satisfied.
Proposition 16.4.1 below contains a result for Rt which is later used in the
proof of a result for W0 in Theorem 16.5.1 of Section 16.5.
Proposition 16.4.1 Let X ≡ (q1 , . . . , q2k+m ) be singular, 2 < m ≤ k, with
either m = 2T + 1 for some T > 0 or m = 2T + 2 for some T > 0. Then
for any X ∗ .
k
R̃t (X) ≡ pk+t+1 qk+m−t qs
s=t+1
m+k−3t−3
+ F̄k−t−1 (qk+t+2 , . . . , qk+m−t−1 , qk+m+1 , . . . , q2k+m−t−1 )
2k+m−t
+ qk+t+1 pk+m−t qs
s=k+m+1
m+k−3t−3
+ F̄k−t−1 (qt+2 , . . . , qk , qk+t+2 , . . . , qk+m−t−1 ) ,
with a similar formula for R̃t (X ∗ ). Put R̃t (X) = R̃t and R̃t (X ∗ ) = R̃t∗ .
Note that in the formulas for Rt and Rt∗ the following equalities are satisfied:
∗ ∗
qk+t+1 = qk+t+1 , qk+m−t = qk+m−t , (16.7) of Lemma 16.3.1 and (16.14) of
Lemma 16.3.2. Hence to prove that Rt > Rt∗ , it is sufficient to show R̃t > R̃t∗ .
Define
m−1
2k+m−t
T ≡ qs , T ≡ qs ,
s=t+1 s=2k+2
2(m−2t−2)
U1 ≡ F̄m−2t−2 (qk+t+2 , . . . , qk+m−t−1 , q2k+t+2 , . . . , q2k+m−t−1 ) and
2(m−2t−2)
U2 ≡ F̄m−2t−2 (qt+2 , . . . , qm−t−1 , qk+t+2 , . . . , qk+m−t−1 ).
Applying results (16.1), (16.2), (16.4) and (16.5) of Lemma 16.3.1, and
(16.8), (16.9), (16.11) and (16.13) of Lemma 16.3.2, we have
2k+t+1
∗
R̃t = pk+t+1 qk+m−t T ABm + A Bm U1 qs
s=2k+2
m−1
+ qk+t+1 pk+m−t A∗ Bm T + ABm U2 qs ,
s=m−t
2k+t+1
R̃t∗ = pk+t+1 qk+m−t T A∗ Bm + ABm U1 qs
s=2k+2
m−1
+ qk+t+1 pk+m−t ABm T + A∗ Bm U2 qs ,
s=m−t
320 M. O’Reilly
and so
m−1
R̃t − R̃t∗ = qk+t+1 pk+m−t Bm U2
qs − T Bm (A − A∗ )
s=m−t
2k+t+1
− qk+m−t pk+t+1 Bm U1 qs − T Bm (A − A∗ ). (16.15)
s=2k+2
m−1
= qs
s=t+2
2k+m−t−1
> qs
s=2k+2
>T . (16.21)
m−1
Bm U2 qs − T Bm > 0.
s=m−t
and since
qt+2 > q2k+m−t−1 , . . . , qm−t−1 > q2k+t+2 ,
16 Optimal design of linear consecutive–k–out–of–n systems 321
we have
2(m−2t−2)
F̄m−2t−2 (qt+2 , . . . , qm−t−1 , qk+m−t−1 , . . . , qk+t+2 )
2(m−2t−2)
> F̄m−2t−2 (q2k+m−t−1 , . . . , q2k+t+2 , qk+m−t−1 , . . . , qk+t+2 ).
(16.23)
that is,
U2 > U1 .
From (16.18)–(16.20) and (16.24) we have
m−1
2k+t+1
Bm U2 qs − T Bm > Bm U1 qs − T Bm .
s=m−t s=2k+2
16.5 Theorems
2(k−m)
WT +1 = F̄k−m (qm+1 , . . . , qk , qk+m+1 , . . . , q2k ),
with a similar formula for WT∗ +1 . By Theorem 16.5.1 (see also O’Reilly [28]) it
follows that WT +1 ≥ WT∗ +1 , with equality if and only if either {qm+1 , . . . , qk }
is a subset of {qi1 , . . . , qir } or the intersection of those sets is empty. Either
way we have
WT +1 ≥ WT∗ +1 .
STEP 2. Note that if m = 2T + 1, then
WT = pk+T +1 MT + qk+T +1 WT +1 ,
Wt > Wt∗ .
W0 > W0∗ ,
G2k+m
k (X) > G2k+m
k (X ∗ )
for any X ∗ .
Proof. This theorem for G systems can be proved in a manner similar to the
proof of Theorem 16.5.1 for F systems. That is, by giving similar definitions
for G systems, similar proofs for lemmas and propositions for G systems
can be given. Alternatively, the theorem can be proved by applying only
Theorem 16.5.1, as below.
Clearly, X ∗ is nonsingular. Define p̄i ≡ qi for all 1 ≤ i ≤ 2k + m. Then we
have
G2k+m
k (X) = F̄k2k+m (p̄1 , . . . , p̄2k+m ),
G2k+m
k (X ∗ ) = F̄k2k+m (p̄∗1 , . . . , p̄∗2k+m ),
G2k+m
k (Z) > G2k+m
k (Y ).
We have shown that a necessary condition for the optimal design of a lin-
ear consecutive–k–out–of–n:F system with 2k + 2 ≤ n ≤ 3k is for it to be
nonsingular (Corollary 16.5.1 of Section 16.5), which is similar to the case
2k ≤ n ≤ 2k + 1 treated in [28]. However, the procedures given in [28] can-
not be implemented in this case. This is due to the restriction placed on
the choice of interchanged symmetrical components ((3m − 2) components
excluded from the interchange).
The following procedure is a consequence of Theorem 16.5.1.
Procedure 16.6.1 In order to improve a singular design of a linear consec-
utive–k–out–of–(2k +m):F system with 2 ≤ m ≤ k, apply the following steps:
1. select an arbitrary nonempty set of pairs of symmetrical components so
that the first component in each pair is in a position from m to k; and
then
2. interchange the two components in each selected pair.
Note that the number of possible choices in Step 1 is 2(k−m+1) − 1. Conse-
quently, the best improvement can be chosen or, if the number of possible
choices is too large to consider all options, the procedure can be repeated as
required.
Because the result for systems with 2k +2 ≤ n ≤ 3k excludes some compo-
nents, it is not possible to derive from it, unlike the case when 2k ≤ n ≤ 2k+1,
that it is necessary for the optimal design of a linear consecutive–k–out–of–n:
G system to be singular. However, as stated in Corollary 16.5.2 of Section
16.5, if a subsystem composed of those excluded components is singular, then
the whole system has to be singular for it to be optimal. Consequently, the
following procedure can be applied. Note that, for a given nonsingular design,
the number of possible singular designs produced in this manner is 1.
Procedure 16.6.2 Suppose a design of a linear consecutive–k–out–of–(2k +
m): G system is nonsingular, with 2 ≤ m ≤ k. Consider its subsystem com-
posed of components in positions from 1 to (m − 1) , from (k + 1) to (k + m),
and from (2k + 2) to (2k + m), in order as in the design. If such a subsys-
tem is singular, then in order to improve the design, interchange all required
symmetrical components so that the design becomes singular.
The following examples, calculated using a program written in C ++ , are
given in order to illustrate the fact that both nonsingular and singular optimal
designs of linear consecutive–k–out–of–n:G systems exist.
Example 1. (q1 , q5 , q7 , q9 , q8 , q6 , q4 , q3 , q2 ) is a nonsingular optimal design of
a linear consecutive–3–out–of–9:G system. It is optimal for q1 = 0.151860,
q2 = 0.212439, q3 = 0.304657, q4 = 0.337662, q5 = 0.387477, q6 = 0.600855,
q7 = 0.608716, q8 = 0.643610 and q9 = 0.885895.
16 Optimal design of linear consecutive–k–out–of–n systems 325
References
Malgorzata O’Reilly
17.1 Introduction
For the description of the mathematical model of the system discussed here,
including nomenclature, assumptions and notation, the reader is referred to
[9], also appearing in this volume.
Zuo and Kuo [16] have proposed three methods for dealing with the variant
optimal design problem: a heuristic method, a randomization method and a
binary search method. The heuristic and randomization methods produce
Malgorzata O’Reilly
School of Mathematics and Physics, University of Tasmania, Hobart TAS 7001,
AUSTRALIA
e-mail: [email protected]
and so also
F̄kn (X 1;k+1 ) = qk+1 q1 F̄kn (1, q2 , . . . , qk , 1, qk+2 , . . . , qn )
+ pk+1 p1 F̄kn (0, q2 , . . . , qk , 0, qk+2 , . . . , qn )
+ (pk+1 q1 )qk+2 F̄kn−1 (q2 , . . . , qk , 1, 1, qk+3 , . . . , qn )
+ (pk+1 q1 )pk+2 F̄kn−1 (q2 , . . . , qk , 1, 0, qk+3 , . . . , qn )
+ (qk+1 p1 )qk+2 F̄kn (1, q2 , . . . , qk , 0, 1, qk+3 , . . . , qn )
+ (qk+1 p1 )pk+2 F̄kn (1, q2 , . . . , qk , 0, 0, qk+3 , . . . , qn ). (17.2)
Note that
F̄kn−1 (q2 , . . . , qk , 1, 0, qk+3 , . . . , qn )
= F̄kn (1, q2 , . . . , qk , 0, 0, qk+3 , . . . , qn )
k
k
= qs + F̄k n−k−2
(qk+3 , . . . , qn ) − qs F̄kn−k−2 (qk+3 , . . . , qn ),
s=2 s=2
(17.3)
and therefore
(p1 qk+1 )pk+2 F̄kn−1 (q2 , . . . , qk , 1, 0, qk+3 , . . . , qn )
+ (q1 pk+1 )pk+2 F̄kn (1, q2 , . . . , qk , 0, 0, qk+3 , . . . , qn )
= (pk+1 q1 )pk+2 F̄kn−1 (q2 , . . . , qk , 1, 0, qk+3 , . . . , qn )
+ (qk+1 p1 )pk+2 F̄kn (1, q2 , . . . , qk , 0, 0, qk+3 , . . . , qn ). (17.4)
Proof. This result follows from Proposition 17.3.1 and the fact that
where p̄i ≡ qi , 1 ≤ i ≤ n.
Lemma 17.3.1 Let X ≡ (q1 , . . . , qn ) be a design for a linear consecutive–
k–out–of–n:F system, n > 2k, k ≥ 2. If q1 < qk+1 , then X 1;k+1 is a better
design.
T
fs ≡ 0.
s=R
Define
if 1 ≤ i ≤ k − 2, then
and if i = k − 1, then
Wi = F̄kn−k+i (1, . . . , 1, 1, 1, qk+3 , . . . , qn )
@ A> ?
k−1
Since
k−2
k
k
pk + pk−i qs + qs = 1,
i=1 s=k−i+1 s=2
W ≡ Gn−1
k (p2 , . . . , pk , 1, 1, pk+3 , . . . , pn )
− n
Gk (1, p2 , . . . , pk , 0, 1, pk+3 , . . . , pn ). (17.9)
F̄k2k+1 (X) − F̄k2k+1 (X k;k+1 ) = (qk − qk+1 ) (q1 . . . qk−1 − qk+2 . . . q2k
+qk+2 . . . q2k+1 − q1 . . . qk−1 qk+2 . . . q2k+1 ) .
Proof. From
and consequently
we have
and
F̄k2k+1 (Y ) = F̄k2k (q2k , . . . , qk+2 , qk , qk+1 , qk−1 , . . . , q1 )
+ pk+1 qk−1 . . . q1 q2k+1 (1 − qk+2 . . . q2k qk ) . (17.14)
Note that
pk+1 qk+2 . . . q2k q2k+1 (1 − q1 . . . qk−1 qk )
− pk+1 qk−1 . . . q1 q2k+1 (1 − qk+2 . . . q2k qk )
= pk+1 q2k+1 (qk+2 . . . q2k − q1 . . . qk−1 ) . (17.15)
Also, we have
F̄k2k (q1 , . . . , q2k ) = pk pk+1 · 0
2(k−2)
+ qk qk+1 F̄k−2 (q2 , . . . , qk−1 , qk+2 , . . . , q2k−1 )
+ pk qk+1 (qk+2 . . . q2k ) + qk pk+1 (q1 . . . qk−1 ) (17.16)
and
F̄k2k (q2k , . . . , qk+2 , qk , qk+1 , qk−1 , . . . , q1 ) = pk pk+1 · 0
2(k−2)
+ qk qk+1 F̄k−2 (q2 , . . . , qk−1 , qk+2 , . . . , q2k−1 )
+ pk qk+1 (q1 . . . qk−1 ) + pk+1 qk (qk+2 . . . q2k ), (17.17)
and so
F̄k2k (q1 , . . . , q2k ) − F̄k2k (q2k , . . . , qk+2 , qk , qk+1 , qk−1 , . . . , q1 )
= (qk+1 − qk ) (qk+2 . . . q2k − q1 . . . qk−1 ) . (17.18)
Proof. Suppose q1 . . . qk−1 ≥ qk+2 . . . q2k . Then, since qk − qk+1 > 0 and
q1 > q2 > · · · > qk > q2k+1 > q2k > · · · > qk+2 , (17.23)
proving that
• taking the (2k + 1)-th component and putting it on the left-hand side of
the system, next to the first component (in position 0);
• interchanging components k and (k + 1); and then
• reversing the order of components.
Proof. From Corollary 17.3.1 we have min{q1 , q2k+1 } > qk+1 . If X is opti-
mal, then it satisfies the necessary conditions for the optimal design given by
Malon [7] and Kuo et al. [5], as stated in Section 17.2. From Lemma 17.4.1 it
follows that qk+1 > qk must be satisfied. Similarly, from Lemma
17.4.1 applied to the reversed design Xr ≡ (q2k+1 , . . . , q1 ), we have qk+1 >
qk+2 .
Propositions 17.5.1 and 17.5.2 below contain preliminary results for Lemma
17.5.1, followed by Corollary 17.5.1 which gives a necessary condition for the
optimal design of a linear consecutive–k–out–of–(2k + 2): F system.
Proposition 17.5.1 Let X ≡ (q1 , . . . , q2k+2 ), k > 2. Then
Proof. Since
2(k−2)
F̄k2k+2 (X) = qk+1 qk+2 F̄k−2 (q3 , . . . , qk , qk+3 , . . . , q2k )
+ pk+1 pk+2 F̄k2k+2 (q1 , . . . , qk , 0, 0, qk+3 , . . . , q2k )
+ pk+1 qk+2 F̄k2k+2 (q1 , . . . , qk , 0, 1, qk+3 , . . . , q2k+2 )
+ qk+1 pk+2 F̄k2k+2 (q1 , . . . , qk , 1, 0, qk+3 , . . . , q2k+2 ) (17.28)
338 M. O’Reilly
and
2(k−2)
F̄k2k+2 (X k+1;k+2 ) = qk+1 qk+2 F̄k−2 (q3 , . . . , qk , qk+3 , . . . , q2k )
+ pk+1 pk+2 F̄k2k+2 (q1 , . . . , qk , 0, 0, qk+3 , . . . , q2k )
+ pk+2 qk+1 F̄k2k+2 (q1 , . . . , qk , 0, 1, qk+3 , . . . , q2k+2 )
+ qk+2 pk+1 F̄k2k+2 (q1 , . . . , qk , 1, 0, qk+3 , . . . , q2k+2 ), (17.29)
we have
Proof. Since
and
we have
and so by Proposition 17.5.1 we have F̄k2k+2 (X) > F̄k2k+2 (X k+1;k+2 ), proving
that X k+1;k+2 is a better design.
340 M. O’Reilly
and from Proposition 17.5.2 it follows that F̄k2k+2 (X) > F̄k2k+2 (X 1;2k+2 ),
proving that X 1;2k+2 is a better design and completing the proof.
Proof. Without loss of generality we may assume q1 > q2k+2 . For q1 < q2k+2
we apply the reasoning below to the reversed design Xr ≡ (q2k+2 , . . . , q1 ).
Suppose (q1 , qk+1 , qk+2 , q2k+2 ) is nonsingular. Then qk+1 < qk+2 , and by
Lemma 17.5.1 we have that either X k+1;k+2 or X 1;2k+2 must be a bet-
ter design. Hence X is not optimal contrary to the assumption, and (i)
follows.
Suppose that (q1 , . . . , qk , qk+3 , . . . , q2k+2 ) is singular. Then, since from
above (q1 , qk+1 , qk+2 , q2k+2 ) must be singular, we have that X is singular,
contrary to the necessary condition of nonsingularity stated by O’Reilly in
([9], Corollary 1). Hence (ii) follows and this completes the proof.
The procedures below follow directly from the results of Lemmas 17.3.1,
17.3.2, 17.4.1, 17.5.1 respectively. Procedure 17.6.3 also applies the necessary
conditions for the optimal design given by Malon [7] and Kuo et al. [5], as
stated in Section 17.1.
• If qk+1 < qk ,
References
Abstract Despite the fact that wood flour has been known as an inexpen-
sive filler in plastics compounds for many years, commercial wood-filled plas-
tics are not widely used. One reason for this has been the poor mechanical
properties of wood-filled compounds. Recent publications report advances in
wood flour modification and compatibilization of polymer matrices, which
has led to an improvement in processability and the mechanical properties
of the blends. In most cases the compounds were obtained in Brabender-
type mixers. In this work the authors present the results for direct feeding of
mixtures of wood flour and thermoplastic materials (polypropylene and SBS
elastomer) in injection molding. The obtained blends were compared with
Brabender-mixed compounds from the point of view of physical and mechan-
ical properties and aesthetics. It was shown that polymer blends with rough
Pavel Spiridonov
Centre for Advanced Manufacturing Research, University of South Australia,
Mawson Lakes SA 5095, AUSTRALIA
e-mail: [email protected]
Jan Budin
Institute of Chemical Technology, Prague, CZECH REPUBLIC
e-mail: [email protected]
Stephen Clarke
Polymer Science Group, Ian Wark Research Institute, University of South Australia,
Mawson Lakes SA 5095, AUSTRALIA
e-mail: [email protected]
Jani Matisons
Polymer Science Group, Ian Wark Research Institute, University of South Australia,
Mawson Lakes SA 5095, AUSTRALIA
e-mail: [email protected]
Grier Lin
Centre for Advanced Manufacturing Research, University of South Australia,
Mawson Lakes SA 5095, AUSTRALIA
e-mail: [email protected]
grades of wood flour (particle size >300 microns) possess a better decorative
look and a lower density having, at the same time, poorer mechanical prop-
erties. Usage of compatibilizers allowed the authors to optimize the tensile
strength of these compounds.
18.1 Introduction
18.2 Methodology
18.2.1 Materials
Wood flour samples were pre-dried for 6 hours at 50–60o C in the electric
oven before blending. The polymer–wood flour blends were obtained in two
ways. For injection molding, the polymer and filler were mixed just before
molding. No additional pre-compounding was used. The specimens for ten-
sile test (Australian Standard AS 1145) were molded in a 22 (metric) tonne
injection-molding machine.
These blends were also pre-mixed in a Brabender mixer at 40 rpm at 180o C
to 190o C. The polymers were first introduced in the mixer; the wood flour was
then added when the polymers melted (a constant torque was reached). The
total mixing time was 6–8 min depending on the composition. Each blend
weighed 65–70 grams. While warm, the blended materials were formed into a
2-mm sheet in the laboratory vulcanization press under 10 MPa pressure at
180o C. The tensile specimens were punched from the sheets using a standard
cutting die. Tensile testing of the above specimens was conducted according
to AS 1145 on a horizontal tensile test machine. Five test samples of each
compound were tested.
Densities of the wood flour and polymer compounds were determined by
a volumetric method in either water or methylated spirits.
From Table 18.1 it can be seen that the density of wood flour depends on
the particle size in each fraction. The fractions consisting of smaller particles
have a higher density. This is because wood is a cellulose material, which has
a porous structure [4]. Larger wood particles retain this structure with con-
siderably less displacement of voids by the floatation medium. On the other
hand, for smaller particles a higher percentage of voids are filled by the flota-
tion liquid [14] resulting in a higher density. The difference in density should
influence the density of the polymer compounds that contain different wood
flour fractions. When the compounds were molded in the injection-molding
machine or pressed after mixing in the Brabender mixer, their density was
both measured and calculated. The calculations were based on the density
of the polymer matrix and the wood flour and their ratio in the compounds.
The results are presented in Figure 18.1.
A difference was observed between the two processing methods and the
calculated values. The results indicated that the densities of un-coupled PP
were lower when molded in the injection-molding machine. The most sta-
ble results were obtained when both PP and SBS compounds were mixed
in the Brabender mixer and then were formed in the press. Stark et al. [8]
have explained this observation by the compression of the compounds to
the maximum density that the wood cell walls can sustain. This correlates
with our results in regard to the Brabender method and the difference from
the calculated values. However, the pressure created in an injection mold
is comparable with the pressure developed in the press. Although an injec-
tion machine creates a bigger plasticizing effect, the total blending time is
shorter (1–1.5 min) than for the Babender mixer (6–8 min). Therefore, in
addition to the effect of compression, blending time is a very important
parameter.
The use of modifiers improved the quality of compounds without increas-
ing the blending time. Thus maleated polypropylene allowed us to obtain
compounds with close densities both in injection molding and the Braben-
der mixer (see Figure 18.1). This is because of the compatibilization impact
of maleic anhydride, which is achieved by improving the polymer matrix
impregnation, improving fiber dispersion, enhancing the interfacial adhesion
and other effects [9]–[13].
Fig. 18.1 Density of (a) polypropylene and (b) SBS in elastomer compounds for different
blending methods.
For the injection of SBS elastomer, the temperatures in the barrel were set
10–15o C higher. The observations of the injection molding of polymer–wood
flour mixtures showed that stability of this process depends on the filler /
polymer ratio and the temperature. When the content of wood flour was be-
low 40%, the process and the quality of the molded specimens were both quite
stable. When the content of wood flour exceeded 50%, its volume exceeded
the polymer volume and it became hard to obtain a consistent quality. In
addition, because wood flour is hard and does not melt during the process,
the friction between metal parts of the machine (screw, barrel) and the wood
flour particles is very high, which also prevents the polymer matrix from
forming a continuous phase. Therefore it was visually detected that the dis-
tribution of wood flour in the polymer was not regular (for example, particle
agglomerates were observed) when the wood filler content was greater than
50% weight. The maximum content of wood flour in the following experiments
was maintained at 40%.
During the injection-molding experiments it was noticed that higher tem-
peratures led to the formation of vapors in the compounds. It was accounted
for by decomposition of the wood flour, which is known to start at around
200o C [7]. In such cases it was difficult or even impossible to obtain good
specimens, despite high mold pressure. Thus we were unable to mold mix-
tures of nylon with wood flour, because nylon requires higher injection tem-
peratures (240–260o C). Therefore the direct feeding of polymer–wood flour
mixtures into an injection-molding machine can be done only when the
wood flour content is less than 40% and the polymer has a melting point
below 200o C.
18 Optimizing properties of plastics compounds containing wood flour 349
Fig. 18.3 Influence of wood flour fractions and the modifier on the tensile strength of
injection-molded specimens of the (a) PP and (b) SBS compounds.
350 P. Spiridonov et al.
Fig. 18.4 Relative cost of the (a) PP and (b) SBS compounds depending on the content
of wood flour and maleated polymers.
The calculations were based on the cost of raw materials and their content
in the compounds. Figure 18.4 shows that PP compounds are cheaper than
the control sample (with a relative cost of less than 1) where they contain
a considerable amount of wood flour and less than 17% PP–MA. For SBS
compounds, the equilibrium cost threshold was much higher. It follows from
Figure 18.4 that it is possible to introduce 50% SEBS–MA to the compound
containing 50% wood flour without increasing the cost of the modified com-
pound. The difference between the optimum cost of the PP and SBS com-
pounds is caused by the difference in the cost of raw materials. Thus the
cost of the SBS elastomer is ∼ 3 times higher than the cost of virgin PP and
the cost of maleated SEBS is much higher than the cost of PP–MA. Under
these conditions, the application of a cheap filler such as wood flour provides
an economic effect, allowing the manufacturers greater flexibility with the
composition.
It is necessary to say that in addition to financial savings, the use of wood
flour can provide plastics companies with the benefit of an aesthetically pleas-
ing, natural looking, wood-filled finish. Figure 18.5 provides an indication of
the decorative properties possible for PP compounds containing 40% of dif-
ferent grades of wood flour. It can be noticed that rough grades (fraction 1
and 2) provide a more natural wood look to the plastics than do fractions 3
and 4. Therefore wood flour with particle sizes in the range from 300 to 850
microns can be recommended for use in decorative plastics. Despite the fact
that those fractions decrease the mechanical properties of the compounds,
these properties may not necessarily be an essential criterion for decorative
parts and components. It should be possible to find an optimal balance be-
tween the properties and the cost of the compounds in the way described
above.
352 P. Spiridonov et al.
Fraction 1 Fraction 2
Fraction 3 Fraction 4
Fig. 18.5 Photographs of the PP compounds containing 40% wood flour of different
fractions.
18.4 Conclusions
grades with particles in the range between 300 to 850 microns can be used
in decorative plastics.
References
15. B. J. Lee, A. G. McDonald and B. James, Influence of fiber length on the mechani-
cal properties of wood-fiber/polypropylene prepreg sheets, Mat. Res. Innov. 4(2–3)
(2001), 97–103.
16. J. J. Balatinecz and M. M. Sain, The influence of recycling on the properties of wood
fibre plastic composites, Macro. Symp. 135 (1998), 167–173.
Chapter 19
Constrained spanning, Steiner trees
and the triangle inequality
Prabhu Manyem
Key words: Minimum spanning tree, maximum spanning tree, triangle in-
equality, Steiner tree, APX, approximation algorithm, asymptotic worst case
ratio
19.1 Introduction
Prabhu Manyem
Centre for Industrial and Applied Mathematics, University of South Australia
Mawson Lakes SA 5095, AUSTRALIA∗
e-mail: [email protected]
∗ Currently at The University of Ballarat.
The lower the value of RA , the better the heuristic A. A constant value of
RA is superior to a value that depends on the size of instances, for example,
19 Constrained spanning, Steiner trees and the triangle inequality 357
RA ∈ Θ(n) or RA ∈ Θ(log n). Lund and Yannakakis [7] show that the SET
COVER problem cannot be in the class APX, which is the class of prob-
lems for which it is possible to construct a polynomial time heuristic A that
guarantees a constant value on RA . Feige [3] showed that unless NP ⊂
DTIME(nlog log n ), the SET COVER problem cannot be approximated to
within Θ(log n). Manyem and Stallmann [9] have shown that an HCSP, and
hence a CSP too, cannot be in the complexity class APX. Results from [2]
indicate that a DCSP is unlikely to be in APX. Heuristics for the Steiner
tree version of the problem with general costs and weighted hops appear in
Manyem [8].
Marathe, Ravi et al. [10] consider networks with both cost and delay pa-
rameters on the edges. They provide an approximation algorithm that guar-
antees a diameter within O(log |V |) of the given threshold Δ and a total cost
within O(log |V |) of the optimum. A vast compendium of results on approx-
imability is provided in Ausiello et al. [1].
Figure 19.1 provides a road map of some of the optimization problems that
arise in telecommunication networks. Here S is the set of terminal nodes for
Steiner tree problems. In multicasting terminology, S is the set of conference
nodes. Problem 1, the Constrained Steiner Tree (CST), is the hardest in the
1 all
S=V delays
2
=1 5
3 4
Constr. CST with
Shortest CSP unit weight
HCST
Path NP-C edges
NP-C all
NP-C all delays NP-C
delays =1
all edge =1 S=V
all edge all edge
wts = 1
wts = 1 wts = 1 |S| = 2
CSP with
HCST with delay constr.
Problem unit weight HCSP
8 unit weight path with
#9 edges 7
NP-C edges min. # of edges
poly. time
edge wts NP-C poly. time
6 1 or 2
S=V |S| = 2 9
all all
delays delays
HCSP with height constr. =1
=1 10
edge weights path with
1 or 2 min. # of edges 11
NP-C poly. time
all edge
wts = 1
Fig. 19.1 A Constrained Steiner Tree and some of its special cases.
358 P. Manyem
figure – all other problems are special cases of CSTs. Given an instance of a
CST, if we set the multicast group to be all nodes in the network, we obtain
the Constrained Spanning Tree (Problem 3). Given an instance of a CST, if we
set the multicast group to be just two nodes which need to communicate with
each other, we obtain Problem 2, the Constrained Shortest Path. Problem 7
is a Hop-Constrained Spanning Tree, and Problem 4 is a Hop-Constrained
Steiner Tree.
All problems above the dotted line in Figure 19.1 are NP-complete, and
the ones below can be solved in polynomial time. Positive results from one
problem to another flow in the direction of the arrows, and negative results
flow in the direction against that shown by the arrows. For example, if we can
develop a heuristic for Problem 1 that guarantees an upper bound B on the
approximation ratio over all instances, this will also hold true for all problems
in the figure. On the other hand, if we can show (a negative result) that unless
NP ⊂ DTIME(nlog log n ), there can be no heuristic that guarantees an upper
bound of B for Problem 7, then this will also be true for Problems 1, 3 and
4. See Table 19.1 for further details.
The proof in [12] that Problem 10 in Figure 19.1 is NP-hard renders prob-
lems 1, 3, 4 and 7 NP-hard as well. Similarly, the proofs in [9] show that
unless NP ⊂ DTIME(nlog log n ), Problems 7 and 8 cannot be approximated
to better than Θ(log n). Hence this non-approximability result carries over
to Problems 1, 3, 4 and 5.
In this chapter, we consider special cases of CSPs and HCSPs where the
edge costs and delays obey the triangle inequality (we call these problems
CSPI s and HCSPI s respectively). First, in Section 19.2, we show that the
cost of spanning tree solutions for a CSPI and an HCSPI in a given network
G = (V, E) is at most |V | − 1 times the cost of any other spanning tree
solution for G. This implies that any solution is within a |V | − 1 factor of
the optimal solution.
Next, in Section 19.3, we prove that the lower bound for any approxima-
tion algorithm for a CSPI is Θ(log n). Unless NP ⊂ DTIME(nlog log n ), no
19 Constrained spanning, Steiner trees and the triangle inequality 359
We first show that for a given network G = (V, E) with non-negative costs
cij on undirected edges (i, j) ∈ E, the value of a spanning tree is at most
|V | − 1 times that of any other. We shall assume that the underlying graph
is complete without loss of generality (if the network is not complete, we can
add edges to the network with costs that obey the triangle inequality). We
start with a well-known result for such graphs.
Remark 1. For any two vertices i and j in G, where the edge costs of G obey
the triangle inequality, the edge (i, j) is also a least expensive path in G
between these two vertices.
We show here that Lmax , the cost of the most expensive edge in E, is at most
equal to Tmin , the cost of a MinSTG (minimum spanning tree for G). Let
the endpoints of the most expensive edge be s and t. Let L1 be the cost of
the s − t path using the edges in MinSTG . From Remark 1, it follows that
cst = Lmax ≤ L1 . Since L1 ≤ Tmin , we conclude that Lmax ≤ Tmin .
Remark 2. In a network G where the edge costs obey the triangle inequality,
the cost of the most expensive edge in G is at most the cost of a minimum
spanning tree of G.
Remark 4. For undirected networks where the edge costs obey the triangle
inequality, the performance ratio RA for any approximation algorithm A has
an upper bound of |V | for any version of the spanning tree problem that has
the objective of minimizing the sum of the edge costs in the feasible solution.
V (I, S)
= 1 + ε(I, S), (19.3)
opt(I)
For a CSPI , a spanning tree needs to be determined such that (1) its cost is
minimal and (2) the sum of the edge delays in the path from a specified vertex
s ∈ V (the source) to every other vertex in V is at most Δ, a non-negative
integer.
We create an instance of a CSPI as follows (see Figure 19.2). For each
xi ∈ X and yj ∈ Y in SET COVER, create a vertex. Create an additional
vertex s. Thus |V | = |X| + |Y | + 1. Since |V | = n, |X| = p and |Y | = q, we
s
n
n n
n+1
n
y1 1 y2 1 y3 ........ yq
1
1
1
1 1
1 1 1 1
x1 1 x2 1 x3 1 x4 . . . . . . . xp
Fig. 19.2 A CSPI instance reduced from SET COVER (not all edges shown).
362 P. Manyem
have n = p+q +1. The edges in E in the instance G = (V, E) of the CSPI are
assigned as in Table 19.2. The costs (delays) assigned to the edges are given
in Column 3 (Column 4) of the table respectively. The graph G is complete.
However, in the interests of clarity, not all edges are shown in Figure 19.2.
Only edge costs are shown in the figure, not edge delays.
Table 19.2 E-Reduction of a SET COVER to a CSPI : Costs and delays of edges in G
Edge Set Definition Cost Delay
E1 {(s, yj )| 1 ≤ j ≤ q} n 1
E2 {(s, xi )| 1 ≤ i ≤ p} n+1 2
E3 {(yj , xi )| xi ∈ yj , 1 ≤ i ≤ p, 1 ≤ j ≤ q} 1 1
E4 {(yj , xi )| xi ∈
/ yj , 1 ≤ i ≤ p, 1 ≤ j ≤ q} 1 2
E5 {(yi , yj )| 1 ≤ i < j ≤ q} 1 1
E6 {(xi , xj )| 1 ≤ i < j ≤ p} 1 1
Let Δ = the delay constraint at all vertices = 2. Note that both the edge
costs and the edge delays in G obey the triangle inequality. (In most cases,
both the cost and the transmission delay of an edge directly relate to its
length. Further, queueing and switching delays are usually minor. Hence the
cost cij and delay dij of an edge are closely related (they could be directly
proportional, for example). However, there may be instances where the in-
crease in edge delay is significantly faster than that of edge cost. For instance,
due to a high degree of congestion in the network, queueing and switching
delays could be far higher than normal. From Table 19.2, the total set of
edges of the graph G is given by E = ∪6i=1 Ei .
Recall that the delay constraint is equal to 2. It is possible for the (s, xi )
edges to be utilized in a feasible solution – if they are, they can be deleted
from the solution with no increase in cost. Note that there can be no paths
of the form s − xi − yj nor of the form s − yj − xi , where (xi , yj ) ∈ E4 ; that
is, when xi ∈/ yj in the SET COVER problem. This is due to the high delay
(2 units) of such edges. In either of the paths just mentioned, the leaf vertex
would experience a delay of 3.
Suppose for xi = x0 , the edge (s, x0 ) is in the feasible solution S0 re-
turned by a heuristic. This edge can be replaced as follows. For any yj ∈ Y ,
the edge (x0 , yj ) is not in the feasible solution S0 , otherwise the delay at
such a yj would be 3, violating the delay constraint. There are two possible
cases here:
• Suppose there exists a y0 such that x0 ∈ y0 in SET COVER, and edge
(s, y0 ) ∈ S0 . Then replace (s, x0 ) with (xi , y0 ) to obtain a new solution S1 .
Observe that cost[S0 ] − cost[S1 ] = n units (the cost decreases).
19 Constrained spanning, Steiner trees and the triangle inequality 363
• Alternatively, such a y0 may not exist. In any case, there is at least one
edge (x0 , yj ) in G for some yj ∈ Y , otherwise no feasible solution is ever
possible for the CSPI . This is due to the fact that x0 ∈ yj for at least one
yj ∈ Y in SET COVER. We name this yj , y1 . As per our assumption
for this case, the edge (s, y1 ) is not in S0 . This implies that y1 is a leaf
vertex in S0 , and has another yj (say y2 ) as its parent. To obtain a new
solution S1 , we can delete the edges (y1 , y2 ) and (s, x0 ), and replace them
with (s, y1 ) and (x0 , y1 ). In S1 , the vertex x0 remains a leaf, but y1 is
no longer a leaf. The delay constraints are still obeyed at all vertices.
We have
19.3.4.1 Structure of S1
The parents of the x’s in an FS have to be y’s — such y’s should in turn have s
as their parent. Also due to the delay constraint, a path such as s−yj −yr −xi
can also be ruled out for any 1 ≤ i ≤ p and 1 ≤ j < r ≤ q.
Not all y’s need to have s as their parent – some of the y’s can have another
y (say ya , for example) as their parent, as long as ya ’s parent is s (recall the
delay constraint of 2). Suppose we call a y such as ya a covering y and the
rest non-covering y’s. The covering y’s together form a cover to the x’s –
these y’s may or may not be leaves in an FS. The non-covering y’s will be
leaves. In other words, a yj is
• in the cover if s is yj ’s parent, and
• not in the cover otherwise. In such a case, the parent of yj would be a
covering y. The delay constraint forbids a non-covering yj to be the parent
of an xi in an FS.
It is sufficient for all the non-covering y’s to have a common parent. Let the
cover size (the number of covering y’s) be k. If in Figure 19.2, we move the
cover to the left (the y’s can be renumbered in such a way that y1 through
yk cover all elements in X), a feasible solution as described above will look
like the one in Figure 19.3.
364 P. Manyem
n
n n
1
y1 1
y2 ..... yk yk+1 ....... yq
1 1 1 1 1
x1 x2 x3 x4 . . . . . . . . . xp
Fig. 19.3 Feasible solution for our instance of CSPI (not all edges shown).
Note that it is cheaper for the non-covering y’s to have one of the covering
y’s as their parent, rather than s – cheaper by a factor of n. The spanning
tree (the feasible solution in Figure 19.3) includes the following:
• (s, yj ) edges: k in number, each with a cost of n,
• edges of the form (yi , yj ), where (s, yi ) is part of the FS, and (s, yj ) is not
(in other words, yi is in the cover and yj is not): q − k such edges, each
with unit cost,
• (xi , yj ) edges: p in number, each with unit cost.
Thus the cost of the spanning tree of Figure 19.3 equals
C(k) kn + n − k − 1 (n − 1)(k − l)
ε(J, T ) = −1 = −1 = .
C(l) ln + n − l − 1 nl + n − l − 1
(n − 1)(k − l) k−l
β ≥ ,
nl + n − l − 1 l
or
β ≥ 1 + l−1 . (19.5)
19.4 Conclusions
From Remark 4, it follows that certain versions of the minimum spanning tree
problem that are of interest in data networking (which need not necessarily
be single-source) have an approximation upper bound of |V |, the number of
nodes in the network, when the edge costs obey the triangle inequality. In
particular,
• the hop-constrained version HCSPI ,
• the delay-constrained version CSPI , and
• the diameter-constrained versions (weighted as well as unweighted)
have an upper bound of |V | on the performance ratio of any approximation
algorithm.
The result from Section 19.3 extends to the case of constrained Steiner
trees which satisfy the triangle inequality, since CSPI is a special case of such
problems for Steiner trees. Specifically, we can conclude that the following
theorem holds.
Theorem 2. The following single-source problems with edge costs obeying the
triangle inequality cannot have an approximation heuristic A that can guaran-
tee a performance ratio RA better than Θ(log n) unless NP ⊂ DTIME(nlog log n ),
and hence these problems cannot be in APX:
• the delay-constrained spanning tree problem CSPI , and
• the delay-constrained Steiner tree CSTI .
The CSTI is the triangle-inequality version of Problem 1 in Figure 19.1.
Both the edge costs and delays in the delay-constrained problem versions
mentioned in this section need to obey the triangle inequality.
Acknowledgments The author benefited from discussions with Matt Stallmann of North
Carolina State University. Support from the Sir Ross and Sir Keith Smith Foundation is
gratefully acknowledged. The comments from the referee were particularly helpful. Since
the early 1990s, the online compendium of Crescenzi and Kann, and more recently, their
book [1], has been a great help to the research community.
References
5. R. Hassin, Approximation schemes for the restricted shortest path problem, Math.
Oper. Res. 17 (1992), 36–42.
6. S. Khanna, R. Motwani, M. Sudan and U. Vazirani, On syntactic versus computational
views of approximability, SIAM J. Comput. 28 (1998), 164–191.
7. C. Lund and M. Yannakakis, On the hardness of approximating minimization prob-
lems, JACM 41 (1994), 960–981.
8. P. Manyem, Routing Problems in Multicast Networks, PhD thesis, North Carolina
State University, Raleigh, NC, USA, 1996.
9. P. Manyem and M.F.M. Stallmann, Approximation results in multicasting, Technical
Report 312, Operations Research, NC State University, Raleigh, NC, 27695–7913,
USA, 1996.
10. M. V. Marathe, R. Ravi, R. Sundaram, S. S. Ravi, D. J. Rosenkrantz and H. B. Hunt
III, Bicriteria network design problems, J. Alg. 28 (1998), 142–171.
11. C.H. Papadimitriou, Computational Complexity (Addison-Wesley, Reading, MA,
1994).
12. H. F. Salama, Y. Viniotis and D. S. Reeves, The delay–constrained minimum span-
ning tree problem, in Second IEEE Symposium on Computers and Communications
(ISCC’97), 1997.
Chapter 20
Parallel line search
A line search involves finding the minimal value of a real function f of a single
real variable x. We attempt to locate the minimizing argument to within a
“tolerance.” Formally, given an interval [a, b] ∈ IR, a function g : [a, b] → IR
T. C. Peachey
School of Computer Science and Software Engineering, Monash University, Clayton,
VIC 3800, AUSTRALIA
D. Abramson
School of Computer Science and Software Engineering, Monash University, Clayton,
VIC 3800, AUSTRALIA
A. Lewis
HPC Facility, Griffith University, Nathan, QLD 4111, AUSTRALIA
20.2 Nimrod/O
method simplex
starts 5
starting points random
tolerance 0.01
endstarts
endmethod
method bfgs
starts 5
starting points random
tolerance 0.01
line steps 8
endstarts
endmethod
Optimizations
Schedule
2 Dispatcher
Cluster
Controller 3
Cache
Results
This section presents a model for the execution time for a line search, in
terms of the number of steps used.
Suppose that each iteration of the line search uses k ≥ 3 steps; we assume
that the points are equally spaced. Let l be the length of the original search.
Each iteration reduces the length of the current domain to a proportion 2/k
of the previous (or 1/k if the minimum happens to fall at an end point). Let
r iterations be the most required to reduce the length to the tolerance d so r
is the least integer such that l(2/k)r ≤ d. Hence
+ ,
log(l/d)
r = ceil , (20.1)
log(k/2)
where ceil(x) signifies the least integer that is not less than x. We write Ti
for the evaluation time for the ith subdivision point and assume that all the
Ti have the same probability density function f (t) and distribution function
F (t). We write s for the number of evaluations required in an iteration. Note
that, after the first iteration, subsequent ones will not require evaluations at
the end points of the subinterval. Further, if k is even and the best point
in the previous interval was internal, then the objective at the midpoint of
the current interval will have been found in the previous iteration. So we
approximate s by k − 2 if k is even and k − 1 if k is odd. As these evaluations
are performed in parallel; the evaluation time for one iteration is B = maxi Ti .
For the scenario discussed above these times are much larger than the times
required for selection of the subdivision points and comparison of the values
there. So we assume that the time for each iteration is just B. We assume
also that the Ti are statistically independent. Under this condition, see for
example [3], the distribution function for B is F (t)s . Thus the mean time for
completion of a batch is approximately
∞
d
M= t [F (t)s ] dt. (20.2)
0 dt
where δ is the Dirac delta and a, x and y are constants with 0 < a < 1 and
x < y. Then (20.2) becomes
Graphs of these functions are shown in Figure 20.3. Figure 20.3(a) shows
how r decreases in a piecewise manner. Figure 20.3(b) gives M for the case
x = 1, y = 8, l/d = 1000 and a = 0.9. Figure 20.3(c) shows E, the product
of r and M . Since M increases and r is piecewise constant, E increases while
r is constant.
20
15
10
0
0 10 20 30 40 50 60 70
(a) Number of iterations, r
10
8
6
4
2
0
0 10 20 30 40 50 60 70
(b) Expected time per iteration, M
50
40
30
20
10
0
0 10 20 30 40 50 60 70
Fig. 20.3 Performance
with Bernoulli job times. (c) Expected time for line search, E
20 Parallel line search 375
0.5
–0.5
Figure 20.5(a) shows the mean total execution time, averaged over 10,000
runs, plotted against k for the exponential distribution. Each point is shown
with error bars enclosing three standard errors. Figure 20.5(b) does the same
for the rectangular distribution. Similar results were obtained for a wide
variety of tolerance values.
For some simulations the line search failed to locate the global minimum,
converging on a local minimum instead. Figure 20.5(c) shows the “effective-
ness,” the proportion of runs that achieved the global minimum. Here the
algorithm is deterministic so effectiveness for a given k is either 0 or 1. In
the next section the search will depend on the order of arrival of jobs and
effectiveness will be fractional.
20.3.4 Conclusions
The preceding results show that increasing the number of steps in a parallel
line search may be counter-productive; increases in k may produce consider-
able increases in E. For this to occur there must of course be variability in
the job times. Note that Figure 20.5(b) shows much less increase than does
Figure 20.5(a), although the mean and variance of the job times are simi-
lar. The significant factor is that the probability of job times is considerably
larger than the mean.
376 T.C. Peachey et al.
1
0.8
0.6
0.4
0.2
0
0 10 20 30 40 50 60 70
(c) Effectiveness versus k
A typical user of the line search algorithm will not have information on the
distribution of job times. However the total time E(k) has local minima at
points where r(k) decreases and these values can be predicted from knowledge
of just the initial interval length l and the tolerance
+ d.,Consideration of (20.1)
shows that r falls to a value ρ at k = ceil 2 ρ dl . This can be used to
compute the number of steps k for a desired number of iterations ρ.
Our analysis has assumed that evaluation times are independent. If these
are dependent, one may expect positive autocorrelation on the parameter
space. This would lead to reduced variation in the later iterations of the line
search which in turn would reduce growth in E between jumps. When the
objective function is continuous but not unimodal we expect a priori that
increasing the value of k makes attaining the global minimum more likely.
Figure 20.5(c) supports this.
20 Parallel line search 377
6
5
4
g(x)
3
2
1
0
0 1 2 3 4 5 6
(a) Strategy 1
6
5
4
g(x)
3
2
1
0
Fig. 20.6 Incomplete eval- 0 1 2 3 4 5 6
uation points. (b) Strategy 2
Strategy 1.
Suppose an iteration involves determination of objective values
g0 , g1 , . . . , gk . At any time suppose that S represents the set of the gi that
have been completed by parallel evaluation. When each new value gj arrives:
add it to the set S
determine gm , the least value in S
if 0 < m < k and gm−1 , gm+1 ∈ S return [xm−1 , xm+1 ]
else if m = 0 and g1 ∈ S return [x0 , x1 ]
else if m = k and gk−1 ∈ S return [xk−1 , xk ]
continue
378 T.C. Peachey et al.
The strategies were implemented for line searches on the test function of
Figure 20.4 with tolerance 0.001. For each k from 3 to 70, the search process
20 Parallel line search 379
8
6
4
2
0
0 10 20 30 40 50 60 70
steps in line search
(a) Execution times
1
0.8 "strategy_1"
effectiveness "strategy_1p"
0.6
"full_search"
0.4
0.2
0
0 10 20 30 40 50 60 70
steps in line search
(b) Effectiveness
was simulated 10,000 times with execution times selected randomly from
some probability distribution.
Strategies 1 and 1p were applied using exponential evaluation times with
a mean λ = 2. Figure 20.7(a) shows the mean execution times and Figure
20.7(b) the effectiveness. In each case the results for these strategies are
compared with those for a full search. Figure 20.8 shows times for the same
range of strategies but with evaluation times from a rectangular distribution
over the interval [0, 1].
These experiments were repeated with the other strategies. Figure 20.9
shows results for the same method as Figure 20.7 but with Strategies 1 and
1p replaced by 2 and 2p. Similarly Figure 20.10 shows results for Strategies
3 and 3p.
12 "strategy_1"
mean execution time
10 "strategy_1p"
8 "full_search"
6
4
2
Fig. 20.8 Strategy 1 with 0
rectangular distribution of 0 10 20 30 40 50 60 70
job times. steps in line search
380 T.C. Peachey et al.
6
4
2
0
0 10 20 30 40 50 60 70
steps in line search
(a) Execution time
1
0.8
effectiveness "strategy_2"
0.6 "strategy_2p"
"full_search"
0.4
0.2
0
0 10 20 30 40 50 60 70
steps in line search
(b) Effectiveness
Strategy 3.
10 "strategy_3p"
"full_search"
8
6
4
2
0
0 10 20 30 40 50 60 70
steps in line search
(a) Execution time
1
0.8 "strategy_3"
effectiveness
"strategy_3p"
0.6 "full_search"
0.4
0.2
0
0 10 20 30 40 50 60 70
steps in line search
(b) Effectiveness
20 Parallel line search 381
20.4.3 Conclusions
References
Ruhul A. Sarker
21.1 Introduction
Ruhul A. Sarker
School of Information Technology and Electrical Engineering, UNSW@ADFA,
Australian Defence Force Academy, Canberra, ACT 2600, AUSTRALIA
e-mail: [email protected]
absorb heat to evaporate and then superheat. The customers specify the
quality parameters (maximum percentage of ash and sulfur, and minimum
BTU/pound) for their coals.
The Coal Company considered in the present research currently operates
three mines. These mines differ greatly in their cost of production and coal
quality. Mine-3 is a relatively low cost mine, but its coal contains high sulfur
and does not have satisfactory metallurgical properties. On the other hand,
it contains reasonably low ash. Mine-1 is the highest cost mine, and the coal
contains relatively high ash (stone) and medium sulfur but it has excellent
metallurgical properties. Mine-2 is the largest and lowest cost mine. Its coal
contains higher ash and sulfur than mine-1 coal, but it has good metallurgical
properties. Because of the coal properties, only mine-1 and mine-2 coals are
used in the preparation of metallurgical coal.
Preparation and blending are the two coal upgrading and processing facili-
ties. Coal preparation (washing) is a process of removing physical impurities.
The process involves several different operations, including crushing (to cre-
ate a size distribution), screening (to separate sizes) and separators (mainly
cyclones, to remove the physical impurities). The objective of running a coal
preparation plant is to maximize the revenue from clean coal while removing
the undesirable impurities.
The processing of ROM coal from mine-3 in the preparation plant does not
improve the quality of coal with a reasonable yield. Therefore, the involve-
ment of the preparation plant with this low quality ROM coal means a lower
financial performance for the company. The customers do not accept these
high sulfur coals for their plant operations because of environmental pollution
restriction. The conversion of low quality ROM coals to a minimum accept-
able quality level will mean a better financial performance for the company.
A blending process provides an opportunity of quality improvement.
Blending is a common process in the Coal Industry. Blending allows up-
grading the low quality run of mine coals by mixing with good quality
coals. Furthermore, supplying the good quality ROM coals to the customers,
through blending, can reduce the cost of production, because it saves (i) the
cost of washing and (ii) lost BTU from the refuses of the preparation plant,
and (iii) it also eliminates the need for capital investment in washing facil-
ities. Most of the thermal coal customers accept blended products if they
satisfy their quality requirements.
In the blending process, the problem is to determine the quantity required
from each run-of-mine coals and washed products that maximizes the revenue
but satisfies the quality constraints of the customers.
A single period coal-blending problem can be formulated as a simple
linear programming model ([Gershon, 1986], [Hooban and Camozzo, 1981],
[Bott and Badiozamani, 1982], [Gunn, 1988], [Gunn et al., 1989], [Gunn and
Chwialkowska, 1989], and [Gunn and Rutherford, 1990]). For the multiperiod
case ([Sarker, 1990], [Sarker, 1991], [Sarker and Gunn, 1990], [Sarker, 1994],
[Sarker and Gunn, 1991], [Sarker and Gunn, 1997], [Sarker and Gunn, 1995],
21 Alternative Mathematical Models 385
[Sarker and Gunn, 1994], and [Sarker, 2003]), the modeling process depends
on the decision whether to carry inventory of run-of-mine (ROM) or of
blended coal (final product). The multiperiod coal-blending problem with
inventory of blended coal is a nonlinear program. On the other hand, the
multiperiod blending problem with inventory of ROM can be formulated as
a linear program. In this case, a number of alternative LP models can be
developed allowing the use of ROM inventory in n future periods.
A large-scale LP is solvable using any of the standard LP packages. How-
ever, a large-scale nonlinear program is complex and is not easy to solve. The
current model is an especially structured nonlinear program, and is solved
using a simple SLP (Successive Linear Programming) algorithm developed by
[Sarker and Gunn, 1997]. The solutions of some multiperiod LP models are
not practically feasible for several technical reasons. The quality of solutions,
complexity of formulation and solution approaches, and solution implemen-
tation difficulties for these models are compared and analyzed. A choice of
the most appropriate model is suggested.
The chapter is organized as follows. Following the introduction, we discuss
four alternative models for coal blending and unpgradation. The flexibility
of these models ares analyzed in Section 3. Section 4 discusses the problem
sizes and computational time required. The objective function values and the
nature of fluctuating situations for the test problems are presented in Section
5. The selection criteria for choosing the most appropriate model is discussed
in Section 6 and the conclusions are drawn in Section 7.
The details of the SPM are presented below for the readers. SPM considers
a period of 1 month long.
Variables
bpjl (metric) tonnes of blended product for customer j made at
location l
cmjl (metric) tonnes of run-of-mine coal from mine m used for blended
product j at location l
wck (metric) tonnes of washed product k produced
wbkjl (metric) tonnes of washed product k used for blended product j
at location l
mbkjl (metric) tonnes of middling product k used for blended product
j at location l
wpkc (metric) tonnes of washed product k sent to customer c
Data
J number of blended product customers
L(j) set of sites used for blended product for customer j
acm , scm , Bm
c
run-of-mine ash, sulfur and BTU/lb analysis for
mine m
aw w w
k , sk , Bk as received ash, sulfur and BTU/lb analysis for
washed product k
21 Alternative Mathematical Models 387
Data (continued)
am m m
k , sk , Bk as received ash, sulfur and BTU/lb analysis for
middling product k
a+ +
j , sj maximum allowable ash and sulfur analysis for
customer j
a− − −
j , sj , Bj minimum allowable ash, sulfur and BTU/lb analysis
for customer j
BT Uj+ , BT Uj− maximum and minimum BTU requirements for
customer j
+
SN S maximum allowable sulfur supplied to local customers
NS set of blended product customers who correspond to
local customers
ζj amount of SO2 per (metric) tonne of sulfur supplied to
customer j ∈ N S
I number of mines
c
rmk amount of run-of-mine coal from mine m used per
(metric) tonne of washed product k (this corresponds to
the washed product recipe)
k
rmid ratio of middling in product k produced
+ −
M Pm , M Pm maximum and minimum production from mine m
BCOSTjl blending cost (dollar/(metric) tonne) for customer j at
location l
M P ROm mining cost (dollar/(metric) tonne) for mine m
M
Bm BTU content (million BTU/(metric) tonne) for ROM
coal from mine m
BkW b BTU content (million BTU/(metric) tonne) for washed
product k
BkM b BTU content (million BTU/(metric) tonne) for
middlings of washed product k
P Bj price (dollar/million BTU) offered by the blended
customer j
P Pkc price (dollar/(metric) tonne) offered by customer c for
washed product k
T CBCjl transportation cost (dollar/(metric) tonne) to blended
product customer j from blending location l
T CM Lml transportation cost (dollar/(metric) tonne) from mine
m to blending location l
T CM Wm transportation cost (dollar/(metric) tonne) from mine m
to VJ plant
T CW Ll transportation cost (dollar/(metric) tonne) from VJ
plant to blending location l
T CW Cc transportation cost (dollar/(metric) tonne) from VJ
plant to washed customer c (may include banking and
pier costs)
388 R.A. Sarker
Z= [−(BCOSTjl + T CBCjl ) × bpjl ]
j l
+ M
[Bm × P Bj − T CM Ljl − M P ROm ] × cmjl
m j l
+ [−Wk − amk × (M CM Wm + M P ROm )] × wck
k m
+ [BkW b × P Bj − T CW Ll ] × wbkjl
k j l
+ [BkM b × P Bj − T CW Ll ] × mbkjl
k j l
+ [P Pkc − T CW Cc ] × wpkc
k c
Constraints
The constraints of SPM are presented below. Relations (21.1)–(21.8) all hold
for j = 1, J, l ∈ L(j).
1. Mass balance for blended products:
− bpjl + cmjl + wbkjl + mbkjl = 0 (21.1)
m k k
This model considers 12 periods where each period is 1 month long. The
variables of MNM are similar to those of SPM with an additional subscript t
to represent time period. To differentiate from SPM we use capital letters for
variables. Although the constraints of this model in each period are similar
to SPM, it has additional constraints to link one period to the next for the
entire planning horizon. The model allows the inventory of blended prod-
uct to be carried from one period to the next. The inventory variables and
inventory balance constraints maintain the links in this multiperiod model.
However, the quality (percentage of ash, sulfur and BTU content per pound)
parameters of blended coal inventories carried from one period to next are
unknown which introduce nonlinearity in the model. Although the details of
the mathematical model for MNM can be found in [Sarker and Gunn, 1997],
390 R.A. Sarker
the mass balance constraint is presented below to give an idea about the
nature of variables and constraints in MNM:
− BPjlt − Ijlt + Ijlt−1 + Cmjlt + W Bkjlt + M Bkjlt = 0 ∀j, l, t
m k k
(21.12)
where
BPjlt (metric) tonnes of blended product j, supplied to customer
j, made at location l in period t (jth product
corresponds to jth customer)
Cmjlt (metric) tonnes of run-of-mine coal from mine m used for
blended product j at location l in period t
W Bkjlt (metric) tonnes of washed product k used for blended
product j at location l in a period t
M Bkjlt (metric) tonnes of middling product k used for blended
product j at location l in period t
Ijlt inventory of blended product j at location l
at the end of period t
The above constraint indicates that the total amount of blended product
j (produced for customers and inventory) is equal to the sum of its con-
stituents of ROM coals, washed coals, middling products and blended coals
from inventories.
The MLM is similar to MNM except that it allows the transfer of the inven-
tory of ROM from one period to any or all future periods within the planning
horizon. This model forms the upper bound of the problem since it consid-
ers all possible savings from inventories and productions. The details of the
model can be found in [Sarker, 2003].
This model allows carrying the most attractive input(s) in terms of quality
and cost, for future periods. This is the upper bound of the planning problem
because:
1. this model considers all possible alternatives of supplying coals to cus-
tomers and
2. the solution to this model will give an objective value larger than or equal
to any feasible solution to the ”true problem.”
To have a feeling about the nature of variables and constraints in ULM,
we present the mass balance constraint below:
− BPjlt + Cmjlτ t + W Bkjlτ t + M Bkjlτ t = 0 (21.13)
m τ ≤t k τ ≤t k τ ≤t
21 Alternative Mathematical Models 391
where
Cmjlτ t ROM coal of mine m produced in period τ ,
used in blended product j at location l in a period t
W Bkjlτ t washed product k produced in period τ ,
used in blended product j at location l in a period t
M Bkjlτ t middling product k produced in period τ ,
used in blended product j at location l in a period t
The constraint represents that the total amount of blended product j
(produced for customers only) is equal to the sum of its constituents of ROM
coals, washed coals and middling products taken from current and previous
periods.
The MLM is similar to ULM except that it only permits the carrying of
inventory of ROM coals from one period to the next where the quality pa-
rameters of ROM coals are known. That means, the run-of-mine and washed
coals produced in period t (=τ ) will be carried for further use in period t+1
only. So the corresponding mass balance constraint will be as follows:
− BPjlt + Cmjl(t−1)t + W Bkjl(t−1)t + M Bkjl(t−1)t = 0 (21.14)
m k k
where
Cmjl(t−1)t ROM coal of mine m produced in period (t − 1),
used in blended product j at location l in a period t
W Bkjl(t−1)t washed product k produced in period (t − 1),
used in blended product j at location l in a period t
M Bkjl(t−1)t middling product k produced in period (t − 1),
used in blended product j at location l in a period t
A number of new models can be formulated between MLM and ULM by
varying n (the inventory of ROM and washed coals which can be carried from
one period to up to n period, where the maximum value of n is 11 in our 12 pe-
riod case). Please note that we intentionally ignore the mathematical details
of MNM, ULM and MLM in this chapter, as they are too long and the empha-
sis of the chapter is on comparisons, and refer [Sarker and Gunn, 1997] and
[Sarker, 2003] to interested readers. Alternatively they can be made available
by the author upon request.
These models differ in their capability of handling fluctuating situations,
the computational time required, the size of the problem, optimal objective
function values, number of coal banks required, etc. By a fluctuating situation
we mean a variable planning environment. These aspects are discussed in the
following sections.
392 R.A. Sarker
SPM, ULM and MLM have been solved using the XMP linear program-
ming codes. The MNM is a specially structured nonlinear program that is
solved using a simple SLP algorithm developed by [Sarker and Gunn, 1997].
The SPM is the least flexible model and ULM is the most flexible model. The
MNM is less flexible in choosing the inputs in comparison to MLM, but more
flexible in handling fluctuating situations. In our computational experience,
the objective function value of MNM is a little less than that of MLM when
there is a stable demand and production pattern. This is due to the fact that
the blended product inventory may not be an attractive input in the next
period. In the following, we discuss how the models work under fluctuating
demand and inputs.
Consider the following simplified situations for a three period problem:
Let X1 , X2 and X3 be the maximum level of inputs available in periods 1,
2 and 3, and Q1 , Q2 and Q3 are the respective demands in periods 1, 2 and 3.
The simple line diagrams for these three cases are shown in parts (a)–(e)
of Figure 21.1. The models treat each of the cases as follows:
21.3.1 Case-1
X1 X2 X3
Q1 Q2 Q3
to customers to customers to customers
(a) SPM
X1 X2 X3
inventory inventory
of blend of blend
IQ1 IQ2
Q1 Q2 Q3
to customers to customers to customers
(b) MNM
inputs X1 inputs X2 inputs X3
IX1 IX2
inventory inventory
of inputs of inputs
Q1 Q2 Q3
to customers to customers to customers
(c) MLM
inputs X1 inputs X2 inputs X3
IX2
IX13
inventory
of inputs IX1
inventory
of inputs
Q1 Q2 Q3
to customers to customers to customer
(d) ULM
X1 X2 X3
inventory inventory
of blend of blend
Q1 IQ1 Q2 IQ2 Q3
to customers to customers to customers
(e) A Variant of MNM
21.3.2 Case-2
21.3.3 Case-3
The ULM is a simple but large linear program. We can solve a reasonably
large linear program without much difficulty. The MLM model is also a linear
program. It is smaller than the upper bounding model. The MNM is smaller
than the ULM and close to MLM, but it takes the largest computational time.
In our study, we solved 36 test problems. In the test problems, we considered
the number of blended products up to 2, blending locations up to 3, number
of mines up to 3, coal washing facilities up to 3 and the time periods 3 to 12.
The arbitrary demand, capacity and quality data were randomly generated
for different test problems. The ranges for some monthly data are: blended
product demand, 200,000 to 250,000 (metric) tonnes, washed coal demand,
290,000 to 550,000 (metric) tonnes, production capacity of mine-1, 85,000
to 270,000 (metric) tonnes, capacity of mine-2, 95,000 to 300,000 (metric)
tonnes and capacity of mine-3, 70,000 to 240,000 (metric) tonnes. The relative
problem sizes of the models are shown in Table 21.1.
are required to solve the model using a SLP algorithm [Sarker and Gunn, 1997].
The ULM and MLM have a similar number of constraints and the MLM has
many fewer variables. For the largest problem, the ULM model contains 576
constraints and 4770 variables, whereas the MLM contains 576 constraints
and 1800 variables. With an increasing number of blended products, blending
locations, washed products, inputs and customers, the ULM becomes a very
large problem in comparison to the other two models.
The ULM is a possible candidate for the planning problem under consid-
eration. This model could be a very large linear program with a large number
of blended products, washed products, customers, mines and blending loca-
tions. If the problem size is too large, one could consider the following points
to reduce the size of the problem without losing the characteristics of the
model.
model. This model considers all possible savings from the carrying inventories
of inputs for the blending process. This model gives an upper bound of the
problem. The computational comparisons of these models are presented in
Table 21.2.
SMP is infeasible for both cases. Normally the MLM shows lower profit
than that of the ULM and higher than that of the SPM. However this model
shows infeasibility in highly fluctuating situations. The MNM also shows
lower profit than that of the ULM, higher than that of the SPM and close
to that of MLM for most cases. The MNM can handle a highly fluctuating
situation as well as the ULM.
The above analysis is used to suggest an appropriate model for the planning
problem. The ULM model can be proposed for use as a tactical model. The
analysis shows that
blending problems like food or crude oil blending. Since the algorithm devel-
oped for the MNM can be generalized for a class of nonlinear programs, its
contribution to the knowledge is justified.
21.7 Conclusions
References
[Bott and Badiozamani, 1982] Bott, D. L. and Badiozamani, K. (1982). Optimal blending
of coal to meeting quality compliance standards. In Proc. XVII APCOM Symposium,
(American Institute of Mining Engineers, New York, 1982) pages 15–23.
[Gershon, 1986] Gershon, M. (1986). A blending-based approach to mine planning and
production scheduling. In Proc. XIX APCOM Symposium, (American Institute of
Mining Engineers, New York, 1986) pages 120–126.
[Gunn, 1988] Gunn, E. A. (1988). Description of the coal blend and wash linear program-
ming model. In A Res. Report for Cape Breton Dev. Corp., Canada, (Cape Breton
Development Corporation, Cape Breton) pages 1–16.
[Gunn et al., 1989] Gunn, E. A., Allen, G., Campbell, J. C., Cunningham, B., and Ruther-
ford, R. (1989). One year of or: Models for operational and production planning in the
coal industry. In Won CORS Practice Prize-89, TIMS/ORSA/CORS Joint Meeting,
Vancouver, Canada.
[Gunn and Chwialkowska, 1989] Gunn, E. A. and Chwialkowska, E. (1989). Develop-
ments in production planning at a coal mining corporation. In Proc. Int. Ind. Eng.
Conference, (Institute of Industrial Engineers, Norcross, Georgia USA) pages 319–324.
21 Alternative Mathematical Models 399
[Gunn and Rutherford, 1990] Gunn, E. A. and Rutherford, P. (1990). Integration of an-
nual and operational planning in a coal mining enterprise. In Proc. of XXII APCOM
Intl. Symposium, Berlin, (Technische Universität, Berlin) pages 95–106.
[Hooban and Camozzo, 1981] Hooban, M. and Camozzo, R. (1981). Blending coal with
a small computer. Coal Age, 86:102–104.
[Sarker, 1990] Sarker, R. A. (1990). Slp algorithm for solving a nonlinear multiperiod coal
blending problem. In Honourable Mention Award, CORS National Annual Confer-
ence, Ottawa, Canada, (Canadian Operational Research Society, Ottawa) pages 1–27.
[Sarker, 1991] Sarker, R. A. (1991). A linear programming based algorithm for a specially
structured nonlinear program. In CORS Annual Conference, Quebec City, Canada.
[Sarker, 1994] Sarker, R. A. (1994). Solving a class of nonlinear programs via a sequence
of linear programs. In Eds Krishna, G. Reddy, R. Nadarajan, Stochastic Models,
Optimization Techniques and Computer Applications, pages 269–278. Wiley Eastern
Limited.
[Sarker, 2003] Sarker, R. A. (2003). Operations Research Applications in a Mining Com-
pany. (Dissertation. de Verlag, Berlin).
[Sarker and Gunn, 1990] Sarker, R. A. and Gunn, E. A. (1990). Linear programming
based tactical planning model for a coal industry. In CORS National Annual Confer-
ence, Ottawa, Canada.
[Sarker and Gunn, 1991] Sarker, R. A. and Gunn, E. A. (1991). A hierarchical production
planning framework for a coal mining company. In CORS Annual Conference, Quebec
City, Canada.
[Sarker and Gunn, 1994] Sarker, R. A. and Gunn, E. A. (1994). Coal bank scheduling
using a mathematical programming model. Applied Mathematical Modelling, 18:672–
678.
[Sarker and Gunn, 1995] Sarker, R. A. and Gunn, E. A. (1995). Determination of a coal
preparation strategy using a computer based enumeration method. Indian Journal of
Engineering and Material Sciences, 2:150–156.
[Sarker and Gunn, 1997] Sarker, R. A. and Gunn, E. A. (1997). A simple slp algorithm
for solving a class of nonlinear programs. European Journal of Operational Research,
101(1):140–154.
About the Editors