Calculus PDF
Calculus PDF
f ðxÞ ¼ 0
Mass balances of stirred tank reactors may lead
to ordinary differential equations:
it may be more difficult to find the x satisfying
dy
dt
¼ f ½yðtÞ this equation. These problems are compounded
when there are more unknowns, leading to
Radiative heat transfer may lead to integral simultaneous equations. If the unknowns
equations: appear in a linear fashion, then an important
Z1 consideration is the structure of the matrix
yðxÞ ¼ gðxÞ þ l Kðx; sÞf ðsÞd s representing the equations; special methods
0
are presented here for special structures.
Even when the model is a differential equa- They are useful because they increase the speed
tion or integral equation, the most basic step in of solution. If the unknowns appear in a
Mathematics in Chemical Engineering 3
nonlinear fashion, the problem is much more determinant are all zero, the value of the deter-
difficult. Iterative techniques must be used (i.e., minant is zero. If the elements of one row or
make a guess of the solution and try to improve column are multiplied by the same constant, the
the guess). An important question then is determinant is the previous value times that
whether such an iterative scheme converges. constant. If two adjacent rows (or columns)
Other important types of equations are linear are interchanged, the value of the new determi-
difference equations and eigenvalue problems, nant is the negative of the value of the original
which are also discussed. determinant. If two rows (or columns) are
identical, the determinant is zero. The value
of a determinant is not changed if one row (or
1.1. Matrix Properties column) is multiplied by a constant and added
to another row (or column).
A matrix is a set of real or complex numbers A matrix is symmetric if
arranged in a rectangular array.
2 3 aij ¼ aji
a11 a12 ... a1n
6 a21 a22 ... a2n 7
6 7
A ¼ 6 .. .. .. .. 7 and it is positive definite if
4 . . . . 5
am1 am2 . . . amn
n X
X n
xT Ax ¼ aij xi xj 0
The numbers aij are the elements of the matrix i¼1 j¼1
A0ij ¼ ð1Þiþj M ij
If AT ¼ A1 the matrix is orthogonal.
The value of jAj is given by Matrices are added and subtracted element
by element.
X
n X
n
jAj ¼ aij A0ij or aij A0ij A þ B is aij þ bij
j¼1 i¼1
A tridiagonal matrix is zero except for elements may be the size of a disturbance, and the output is
along the diagonal and one element to the right the gain [1]. If the rank is less than n, not all the
and left of the diagonal. variables are independent and they cannot all be
controlled. Furthermore, if the singular values
8
< 0 if j < i 1 are widely separated, the process is sensitive to
aij ¼ a otherwise small changes in the elements of the matrix and
: ij
0 if j > i þ 1
the process will be difficult to control.
Block diagonal and pentadiagonal matrices also
arise, especially when solving partial differen-
tial equations in two- and three-dimensions. 1.2. Linear Algebraic Equations
A ¼ USV The U is upper triangular; it has zero elements
below the main diagonal and possibly nonzero
where S is a k k diagonal matrix with diagonal values along the main diagonal and above it
elements dii ¼ s i > 0 for 1 i k. The eigen- (see Fig. 1). The L is lower triangular. It has
values of S S are s 2i . The vectors ui for k þ 1 i the value 1 in each element of the main
m and vi for k þ 1 i n are eigenvectors diagonal, nonzero values below the diagonal,
associated with the eigenvalue zero; the eigen- and zero values above the diagonal (see
values for 1 i k are s 2i . The values of s i are Fig. 1). The original problem can be solved
called the singular values of the matrix A. If A is in two steps:
real, then U and Vare real, and hence orthogonal
matrices. The value of the singular value decom- Ly ¼ f; Ux ¼ y solves Ax ¼ LUx ¼ f
position comes when a process is represented
by a linear transformation and the elements of A, Each of these steps is straightforward
aij , are the contribution to an output i for a because the matrices are upper triangular or
particular variable as input variable j. The input lower triangular.
Mathematics in Chemical Engineering 5
kðAÞ ¼k A k k A1 k
The determinant is given by the product of Generally, the inverse is not used in this way
the diagonal elements of U. This should be because it requires three times more operations
calculated as the LU decomposition is per- than solving with an LU decomposition. How-
formed. If the value of the determinant is a ever, if an inverse is desired, it is calculated
very large or very small number, it can be most efficiently by using the LU decomposition
divided or multiplied by 10 to retain accuracy and then solving
in the computer; the scale factor is then accu-
mulated separately. The condition number k AxðiÞ ¼ bðiÞ
can be defined in terms of the singular value ðiÞ
bj ¼
0 j 6¼ i
decomposition as the ratio of the largest dii to 1 j¼i
6 Mathematics in Chemical Engineering
If
requires multiplying a matrix by a vector, which dimensionless groups govern that phenomenon.
can be done very efficiently on parallel com- In chemical reaction engineering the chemical
puters: For sparse matrices this is a viable reaction stoichiometry can be written as
method. The original method was devised by
HESTENES and STIEFEL [8]; however, more recent X
n
aij Ci ¼ 0; j ¼ 1; 2; . . . ; m
implementations use a preconditioned conju- i¼1
gate gradient method because it converges
faster, provided a good “preconditioner” can where there are n species and m reactions. Then
be found. The system of n linear equations if a matrix is formed from the coefficients aij,
which is an n m matrix, and the rank of the
Ax ¼ f
matrix is r, there are r independent chemical
reactions. The other nr reactions can be
where A is symmetric and positive definite, is to deduced from those r reactions.
be solved. A preconditioning matrix M is
defined in such a way that the problem
is easy to solve exactly (M might be diagonal, Consider a single nonlinear equation in one
for example). Then the preconditioned conju- unknown,
gate gradient method is
f ðxÞ ¼ 0
Guess x0
Calculate r0 ¼ f Ax0
Solve M t0 ¼ r0 ; and set p0 ¼ t0
In Microsoft Excel, roots are found by using
for k ¼ 1; n ðor until convergenceÞ Goal Seek or Solver. Assign one cell to be x, put
r T tk the equation for f(x) in another cell, and let Goal
ak ¼ Tk
pk Apk
xkþ1 ¼ xk þ ak pk Seek or Solver find the value of x making the
rkþ1 ¼ rk ak Apk equation cell zero. In MATLAB, the process is
Solve Mtkþ1 ¼ rkþ1 similar except that a function (m-file) is defined
rT tkþ1
bk ¼ kþ1T and the command fzero(‘f’, x0) provides the
rk tk
pkþ1 ¼ tkþ1 þ bk pk solution x, starting from the initial guess x0.
test for convergence Iterative methods applied to single equations
enddo
include the successive substitution method
Note that the entire algorithm involves only xkþ1 ¼ xk þ bf ðxk Þ gðxk Þ
matrix multiplications. The generalized mini-
mal residual method (GMRES) is an iterative and the Newton–Raphson method.
method that can be used for nonsymmetric
systems and is based on a modified Gram– f ðxk Þ
xkþ1 ¼ xk
Schmidt orthonormalization. Additional infor- df =dxðxk Þ
mation, including software for a variety of
methods, is available [9–13]. The former method converges if the derivative
In dimensional analysis if the dimensions of of g(x) is bounded [3]. The latter method
each physical variable Pj (there are n of them)
are expressed in terms of fundamental measure- dg
ðxÞ m for jx aj < h
dx
ment units mj (such as time, length, mass; there
are m of them):
is based on a Taylor series of the equation about
a a
½Pj ¼ m1 1j m2 2j mammj the k-th iterate:
The second and higher-order terms are neglec- that is fast and guaranteed to converge, if the
ted and f (xkþ1) ¼ 0 to obtain the method. root can be bracketed initially [15, p. 251].
2
In the method of bisection, if a root lies
df 0
> 0; x1 x0 ¼ f ðx Þ b; and d f c between x1 and x2 because f(x1) < 0 and f(x2)
dx 0 df =dxðx0 Þ 0 dx2
x x > 0, then the function is evaluated at the center,
xc ¼ 0.5 (x1 þ x2). If f (xc) > 0, the root lies
Convergence of the Newton–Raphson method between x1 and xc. If f (xc) < 0, the root lies
depends on the properties of the first and second between xc and x2. The process is then repeated.
derivative of f(x) [3, 14]. In practice the method If f (xc) ¼ 0, the root is xc. If f (x1) > 0 and f (x2)
may not converge unless the initial guess is > 0, more than one root may exist between x1
good, or it may converge for some parameters and x2 (or no roots).
and not others. Unfortunately, when the method For systems of equations the Newton–Raph-
is nonconvergent the results look as though a son method is widely used, especially for equa-
mistake occurred in the computer program- tions arising from the solution of differential
ming; distinguishing between these situations equations.
is difficult, so careful programming and testing
are required. If the method converges the dif- f i ðfxj gÞ ¼ 0; 1 i; j n; where fxj g ¼ ðx1 ; x2 ; . . . ; xn Þ ¼ x
ference between successive iterates is some-
thing like 0.1, 0.01, 0.0001, 108. The error Then, an expansion in several variables
(when it is known) goes the same way; the occurs:
method is said to be quadratically convergent
when it converges. If the derivative is difficult X
n
@f i
to calculate a numerical approximation may be fi ðxkþ1 Þ ¼ f i ðxk Þ þ jxk ðxkþ1 xkj Þ þ
j¼1
@xj j
used.
The Jacobian matrix is defined as
df f ðxk þ eÞ f ðxk Þ
¼
dx xk e
@f
Jkij ¼ i
@x j xk
In the secant method the same formula is used
as for the Newton–Raphson method, except that
the derivative is approximated by using the and the Newton–Raphson method is
values from the last two iterates: X
n
Jkij ðxkþ1 xk Þ ¼ f i ðxk Þ
df f ðxk Þ f ðxk1 Þ j¼1
¼
dx k
x xk xk1
For convergence, the norm of the inverse of the
This is equivalent to drawing a straight line Jacobian must be bounded, the norm of
through the last two iterate values on a plot of f the function evaluated at the initial guess
(x) versus x. The Newton–Raphson method is must be bounded, and the second derivative
equivalent to drawing a straight line tangent to must be bounded [14, p. 115], [3, p. 12].
the curve at the last x. In the method of false A review of the usefulness of solution meth-
position (or regula falsi), the secant method is ods for nonlinear equations is available [16].
used to obtain xkþ1, but the previous value is This review concludes that the Newton–Raph-
taken as either xk1 or xk. The choice is made so son method may not be the most efficient.
that the function evaluated for that choice has Broyden’s method approximates the inverse
the opposite sign to f (xkþ1). This method is to the Jacobian and is a good all-purpose
slower than the secant method, but it is more method, but a good initial approximation to
robust and keeps the root between two points at the Jacobian matrix is required. Furthermore,
all times. In all these methods, appropriate the rate of convergence deteriorates for large
strategies are required for bounds on the func- problems, for which the Newton–Raphson
tion or when df/dx ¼ 0. Brent’s method com- method is better. Brown’s method [16] is
bines bracketing, bisection, and an inverse very attractive, whereas Brent’s is not worth
quadratic interpolation to provide a method the extra storage and computation. For large
Mathematics in Chemical Engineering 9
systems of equations, efficient software is avail- is used as the initial guess and the homotopy
able [11–13]. equation is solved for x1.
dx R 1
x1;0 ¼ x0 þ Dt ynþ1;i ¼ xn;i þ x0;i
dt Rþ1 Rþ1
10 Mathematics in Chemical Engineering
and the equilibrium equation gives where A and B are constants that must be
specified by boundary conditions of some kind.
yn;i ¼ K n;i xn;i When the equation is nonhomogeneous, the
solution is represented by the sum of a particu-
If these are combined, lar solution and a general solution to the homo-
geneous equation.
R 1
K nþ1;i xnþ1;i ¼ xn;i þ x0;i
Rþ1 Rþ1
xn ¼ xn;P þ xn;H
is obtained, which is a linear difference equa- The general solution is the one found
tion. This particular problem is quite compli- for the homogeneous equation, and the partic-
cated, and the interested reader is referred to ular solution is any solution to the non-
[19, Chap. 6]. However, the form of the differ- homogeneous difference equation. This can
ence equation is clear. Several examples are be found by methods analogous to those used
given here for solving difference equations. to solve differential equations: The method of
More complete information is available in [20]. undetermined coefficients and the method of
An equation in the form variation of parameters. The last method
applies to equations with variable coefficients,
xnþ1 xn ¼ f nþ1 too. For a problem such as
xn ¼ f n xnþ1 f n xn ¼ gn
Hamilton–Cayley theorem [19, p. 127] states 2. Experimental data must be fit with a math-
that the matrix A satisfies its own characteristic ematical model. The data have experimental
equation. error, so some uncertainty exists. The
parameters in the model as well as the
Pn ðAÞ ¼ ðAÞn þ a1 ðAÞn1 þ a2 ðAÞn2 uncertainty in the determination of those
þ þ an1 ðAÞ þ an I ¼ 0 parameters is desired.
A laborious way to find the eigenvalues of a These problems are addressed in this chap-
matrix is to solve the n-th order polynomial for ter. Section 2.2 gives the properties of polyno-
the li — far too time consuming. Instead the mials defined over the whole domain and
matrix is transformed into another form whose Section 2.3 of polynomials defined on segments
eigenvalues are easier to find. In the Givens of the domain. In Section 2.4, quadrature meth-
method and the Housholder method the matrix ods are given for evaluating an integral. Least-
is transformed into the tridiagonal form; then, squares methods for parameter estimation for
in a fixed number of calculations the eigenval- both linear and nonlinear models are given in
ues can be found [15]. The Givens method Sections 2.5. Fourier transforms to represent
requires 4 n3/3 operations to transform a real discrete data are described in Section 2.7. The
symmetric matrix to tridiagonal form, whereas chapter closes with extensions to two-dimen-
the Householder method requires half that num- sional representations.
ber [14]. Once the tridiagonal form is found, a
Sturm sequence is applied to determine the
eigenvalues. These methods are especially use-
ful when only a few eigenvalues of the matrix 2.2. Global Polynomial Approximation
are desired.
If all the eigenvalues are needed, the QR A global polynomial Pm (x) is defined over the
algorithm is preferred [21]. entire region of space
The eigenvalues of a certain tridiagonal
X
m
matrix can be found analytically. If A is a Pm ðxÞ ¼ cj xj
tridiagonal matrix with j¼0
1. A function is known exactly at a set of points Note that each coefficient of yj is a polynomial
and an interpolating function is desired. The of degree m that vanishes at the points {xj}
interpolant may be exact at the set of points, (except for one value of j) and takes the value of
or it may be a “best fit” in some sense. 1.0 at that point, i.e.,
Alternatively it may be desired to represent
a function in some other way. Pm ðxj Þ ¼ yj j ¼ 1; 2; . . . ; m þ 1
12 Mathematics in Chemical Engineering
If the function f (x) is known, the error in the to within multiplicative constants, which can be
approximation is [23] set either by requiring the leading coefficient to
be one or by requiring the norm to be one.
jxmþ1 x1 jmþ1
jerrorðxÞj
ðm þ 1Þ! Zb
ðmþ1Þ WðxÞP2m ðxÞdx ¼ 1
maxx1 xxmþ1 jf ðxÞj
a
The evaluation of Pm (x) at a point other than The polynomial Pm (x) has m roots in the closed
the defining points can be made with Neville’s interval a to b.
algorithm [15]. Let P1 be the value at x of The polynomial
the unique function passing through the point
(x1, y1); i.e., P1¼ y1. Let P12 be the value at x of pðxÞ ¼ c0 P0 ðxÞ þ c1 P1 ðxÞ þ cm Pm ðxÞ
the unique polynomial passing through the
points x1 and x2. Likewise, Pijk . . . r is the minimizes
unique polynomial passing through the points
xi, xj, xk, . . . , xr. The following scheme is used: Zb
I¼ WðxÞ½f ðxÞ pðxÞ2 dx
a
when
Zb
WðxÞ f ðxÞPj ðxÞdx
These entries are defined by using cj ¼ a
;
Wj
Piðiþ1ÞðiþmÞ ¼
Zb
ðx xiþm ÞPiðiþ1Þðiþm1Þ þ ðxi xÞPðiþ1Þðiþ2ÞðiþmÞ Wj ¼ WðxÞP2j ðxÞdx
xi xiþm a
Consider P1234: the terms on the right-hand side Note that each cj is independent of m, the
of the equation involve P123 and P234. The number of terms retained in the series. The
“parents,” P123 and P234, already agree at points minimum value of I is
2 and 3. Here i ¼ 1, m ¼ 3; thus, the parents
agree at xiþ1, . . . , xiþm1 already. The formula Zb X
m
makes Pi (iþ1) . . . (iþm) agree with the function I min ¼ WðxÞ f 2 ðxÞdx W j c2j
j¼0
at the additional points xiþm and xi. Thus, Pi a
Z1
The orthogonality includes a nonnegative Wðx2 ÞPk ðx2 ÞPm ðx2 Þxa1 dx ¼ 0
weight function, W (x) 0 for all a x b. 0
This procedure specifies the set of polynomials k ¼ 0; 1; . . .; m 1
Mathematics in Chemical Engineering 13
1 1 pffiffiffiffiffiffiffi
1 ffi
Chebyshev Tiþ1¼2xTiTi1
1x2
q1 pq
0 1 x (1x) Jacobi (p, q)
2
1 1 ex Hermite Hiþ1¼2xHi2 i Hi–1
c x
0 1 xe Laguerre (c) (iþ1) Liþ1c¼(xþ2 iþcþ1) L ci (iþc) Li1c
0 1 1 shifted Legendre
0 1 1 shifted Legendre, function of x2
where a ¼ 1 is for planar, a ¼ 2 for cylindrical, Rational polynomials are useful for approxi-
and a ¼ 3 for spherical geometry. These func- mating functions with poles and singularities,
tions are useful if the solution can be proved to which occur in Laplace transforms (see
be an even function of x. Section 4.2).
Fourier series are discussed in Section 4.1.
Rational Polynomials. Rational polynomials Representation by sums of exponentials is also
are ratios of polynomials. A rational polynomial possible [24].
Ri(iþ1) . . . (iþm) passing through m þ 1 points In summary, for discrete data, Legendre
polynomials and rational polynomials are
used. For continuous data a variety of orthogo-
yi ¼ f ðxi Þ; i ¼ 1; . . . ; m þ 1
nal polynomials and rational polynomials are
used. When the number of conditions (discrete
is
data points) exceeds the number of parameters,
Pm ðxÞ p0 þ p1 x þ þ pm xm
then see Section 2.5.
Riðiþ1ÞðiþmÞ ¼ ¼ ;
Qn ðxÞ q 0 þ q 1 x þ þ q n xn
mþ1¼ mþnþ1 2.3. Piecewise Approximation
An alternative condition is to make the rational Piecewise approximations can be developed
polynomial agree with the first m þ 1 terms in from difference formulas [3]. Consider a case
the power series, giving a Pade approximation, i. in which the data points are equally spaced
e.,
xnþ1 xn ¼ Dx
Keeping only the first two terms gives a straight Thus, the approximation is
line through (x0, y0) (x1, y1); keeping the first
three terms gives a quadratic function of posi- X
NT X
NT
yðxÞ ¼ ci N i ðxÞ ¼ yðxi ÞN i ðxÞ
tion going through those points plus (x2, y2). i¼1 i¼1
The value a ¼ 0 gives x ¼ x0; a ¼ 1 gives
x ¼ x1, etc. where ci ¼ y (xi). For convenience, the trial
Backward differences are defined by functions are defined within an element by
using new coordinates:
ryn ¼ yn yn1
x xi
r2 yn ¼ ryn ryn1 ¼ yn 2yn1 þ yn2 u¼
Dxi
The interpolation polynomial of order n The Dxi need not be the same from element to
through the points y0, y1, y2, . . . is element. The trial functions are defined as Ni (x)
(Fig. 3 A) in the global coordinate system and
aða þ 1Þ 2
ya ¼ y0 þ ary0 þ
2!
r y0 þ þ NI (u) (Fig. 3 B) in the local coordinate system
aða þ 1Þ ða þ n 1Þ n
(which also requires specification of the
n!
r y0 element). For xi < x < xiþ1
X
2
yðxÞ ¼ ceI N I ðuÞ ð2Þ
I¼1
with ceI ¼ yðxi Þ; i ¼ ðe 1Þ2 þ I The formulas for the cubic spline are derived
as follows for one region. Since the function is a
in the e th element
cubic function the third derivative is constant
and the second derivative is linear in x. This is
written as
Hermite cubic polynomials can also be used;
these are continuous and have continuous first x xi
C 00i ðxÞ ¼ C00i ðxi Þ þ ½C 00i ðxiþ1 Þ C 00i ðxi Þ
derivatives at the element boundaries [3]. Dxi
Splines. Splines are functions that match and integrated once to give
given values at the points x1, . . . , xNT, shown
in Figure 5, and have continuous derivatives up C0i ðxÞ ¼ C 0i ðxi Þ þ C 00i ðxi Þðx xi Þþ
to some order at the knots, or the points ðx xi Þ2
½C 00i ðxiþ1 Þ C00i ðxi Þ
x2, . . . , xNT1. Cubic splines are most com- 2Dxi
mon. In this case the function is represented by
a cubic polynomial within each interval and has and once more to give
continuous first and second derivatives at the
knots. Ci ðxÞ ¼ C i ðxi Þ þ C 0i ðxi Þðx xi Þ þ C 00i ðxi Þ
Consider the points shown in Figure 5A. The ðx xi Þ2 ðx xi Þ3
þ ½C 00i ðxiþ1 Þ C 00i ðxi Þ
notation for each interval is shown in Figure 5B. 2 6Dxi
Within each interval the function is represented
as a cubic polynomial. Now
C i ðxÞ ¼ a0i þ a1i x þ a2i x2 þ a3i x3 yi ¼ Ci ðxi Þ; y0i ¼ C 0i ðxi Þ; y00i ¼ C00i ðxi Þ
16 Mathematics in Chemical Engineering
xZ
0 þh Z1 X2 Z1
1 1
yðxÞdx ¼ ya hda ¼ Dxi ceI N I ðuÞdu ¼ Dxi ce1 þ ce2
I¼1
2 2
x0 0 0
h 1 Dxi
¼ ðy0 þ y1 Þ h3 y000 ðjÞ; x0 j x0 þ h ¼ ðy þ yiþ1 Þ
2 12 2 i
This corresponds to passing a straight line Since ce1 ¼ yi and ce2 ¼ yiþ1 , the result is the
through the points (x0, y0), (x1, y1) and integrat- same as the trapezoid rule. These formulas can
ing under the interpolant. For equally spaced be added together to give linear elements:
points at a ¼ x0, a þ Dx ¼ x1, a þ 2 Dx ¼
x2, . . . , a þ N Dx ¼ xN, a þ (N þ 1) Dx ¼ b ¼ Zb X Dxe
xnþ1, the trapezoid rule is obtained. yðxÞdx ¼ ðye1 þ ye2 Þ
e
2
a
Trapezoid Rule.
If the quadratic expansion is used (Eq. 3), the
Zb endpoints of the element are xi and xiþ2, and
h
yðxÞdx ¼ y þ 2y1 þ 2y2 þ þ 2yN þ yNþ1 Þ þ Oðh3 Þ xiþ1 is the midpoint, here assumed to be equally
2 0
a
spaced between the ends of the element:
x0Zþ2h Z2 X
3 Z1
h
yðxÞdx ¼ ya hda ¼ ðy0 þ 4y1 þ y2 Þ ¼ Dxi ceI N I ðuÞdu
3 I¼1
x0 0 0
5
h ðIVÞ 1 2 1
y ðjÞ; x0 j x0 þ 2h ¼ Dxe ce1 þ ce2 þ ce3
90 0 6 3 6
18 Mathematics in Chemical Engineering
For many elements, with different Dxe, qua- Table 2. Gaussian quadrature points and weights*
dratic elements: N xi Wi
Zb 1 0.5000000000 0.6666666667
X Dxe
yðxÞ ¼ ðye1 þ 4ye2 þ ye3 Þ 2 0.2113248654 0.5000000000
e
6 0.7886751346 0.5000000000
a
3 0.1127016654 0.2777777778
0.5000000000 0.4444444445
If the element sizes are all the same this gives 0.8872983346 0.2777777778
Simpson’s rule. 4 0.0694318442 0.1739274226
For cubic splines the quadrature rule within 0.3300094783 0.3260725774
0.6699905218 0.3260725774
one element is 0.9305681558 0.1739274226
5 0.0469100771 0.1184634425
xZiþ1 0.2307653450 0.2393143353
1
Ci ðxÞdx ¼ Dxi ðyi þ yiþ1 Þ 0.5000000000 0.2844444444
2
xi 0.7692346551 0.2393143353
0.9530899230 0.1184634425
1
Dx3 ðy00 þ y00iþ1 Þ
24 i i *
For a given N the quadrature points x2, x3, . . . , xNP1 are given
above. x1 ¼ 0, xNP ¼ 1. For N ¼ 1, W1 ¼ W3 ¼ 1/6 and for N 2,
For the entire interval the quadrature formula W1 ¼ WNP ¼ 0.
is
xZNT
X
1 NT1
yðxÞdx ¼ Dxi ðyi þ yiþ1 Þ polynomials give the quadrature formula
2 i¼1
x1
X
1 NT1 Z1
Dx3 ðy00 þ y00iþ1 Þ X
n
24 i¼1 i i ex yðxÞdx ¼ W i yðxi Þ
i¼1
0
The quadrature is exact when y is a polynomial (points and weights are available in mathemat-
of degree 2 m 1 in x. The m weights and m ical tables) [23].
Gauss points result in 2 m parameters, chosen
to exactly represent a polynomial of degree Romberg’s method uses extrapolation tech-
2 m 1, which has 2 m parameters. The Gauss niques to improve the answer [15]. If I1 is
points and weights are given in Table 2. The the value of the integral obtained by using
weights can be defined with W (x) in the inte- interval size h ¼ Dx, I2 the value of I obtained
grand as well. by using interval size h/2, and I0 the true value
For orthogonal collocation on finite ele- of I, then the error in a method is approximately
ments the quadrature formula is hm, or
Z1 X X
NP
yðxÞdx ¼ Dxe W j yðxeJ Þ I 1 I 0 þ chm
e j¼1
0
m
Each special polynomial has its own quad- I2 I0 þ c
h
rature formula. For example, Gauss–Legendre 2
Mathematics in Chemical Engineering 19
Replacing the by an equality (an approxima- sum of squares of the deviation between the
tion) and solving for c and I0 give experimental data and the theoretical equation
2m I 2 I 1 XN
I0 ¼ yi yðxi ; a1 ; a2 ; . . . ; aM Þ 2
2m 1 x2 ¼
i¼1
s i
I1 I2 I3 I4 XN
½yi yðxi ; a1 ; a2 ; . . . ; aM Þ2
J1 J2 J3 s2 ¼
i¼1
N
K1 K2
L1
likelihood.
In a least-squares parameter estimation, it is The parameters are easily determined using
desired to find parameters that minimize the computer software. In Microsoft Excel, the
20 Mathematics in Chemical Engineering
yk ¼ yðtk Þ; tk ¼ kD; k ¼ 0; 1; 2; . . . ; N 1
Nonlinear Regression. In nonlinear regres-
sion, the same procedure is followed except
that an optimization routine must be used to and the sampling rate is D; with only N values
find the minimum x2 (see Chap. 10). {yk} the complete Fourier transform Y (v)
cannot be determined. Calculate the value Y
(vn) at the discrete points
2.6. Fourier Transforms of Discrete
Data [15] vn ¼
2pn N
; n ¼ ; . . . ; 0; . . . ;
N
ND 2 2
Suppose a signal y (t) is sampled at equal X
N1
Yn ¼ yk e2pikn=N
intervals k¼0
Yðvn Þ ¼ DY n
yn ¼ yðnDÞ; n ¼ . . . ; 2; 1; 0; 1; 2; . . .
D ¼ sampling rate
The discrete inverse Fourier transform is
ðe:g:; number of samples per secondÞ
Z1
YðvÞ ¼ yðtÞeivt dt
The fast Fourier transform (FFT) is used to
1
calculate the Fourier transform as well as the
Z1
inverse Fourier transform. A discrete Fourier
1 transform of length N can be written as the sum
yðtÞ ¼ YðvÞeivt dv
2p
1 of two discrete Fourier transforms, each of
length N/2, and each of these transforms is
(For definition of i, see Chap. 3.) The separated into two halves, each half as long.
Nyquist critical frequency or critical angular This continues until only one component is
Mathematics in Chemical Engineering 21
left. For this reason, N is taken as a power of each of the grid points:
2, N ¼ 2p.
The vector {yj} is filled with zeroes, if need
dY 1 X N
be, to make N ¼ 2p for some p. For the dx
jn ¼ y0 e2ikpxn =L
L k¼N k
computer program, see [15, p. 381]. The stan-
dard Fourier transform takes N2 operations to
calculate, whereas the fast Fourier transform Any nonlinear term can be treated in the same
takes only N log2 N. For large N the difference way: Evaluate it in real space at N points and
is significant; at N ¼ 100 it is a factor of 15, take the Fourier transform. After processing
but for N ¼ 1000 it is a factor of 100. using this transform to get the transform of a
The discrete Fourier transform can also be new function, take the inverse transform to
used for differentiating a function; this is used obtain the new function at N points. This is
in the spectral method for solving differential what is done in direct numerical simulation of
equations. Consider a grid of equidistant turbulence (DNS).
points:
L
xn ¼ nDx; n ¼ 0; 1; 2; . . . ; 2 N 1; Dx ¼ 2.7. Two-Dimensional Interpolation
2N
and Quadrature
the solution is known at each of these grid
points {Y(xn)}. First, the Fourier transform is Bicubic splines can be used to interpolate a set
taken: of values on a regular array, f (xi, yj). Suppose
NX points occur in the x direction and NY points
1 2XN1 occur in the y direction. PRESS et al. [15] suggest
yk ¼ Yðxn Þe2ikpxn =L
2 N n¼0 computing NY different cubic splines of size NX
along lines of constant y, for example, and
The inverse transformation is storing the derivative information. To obtain
the value of f at some point x, y, evaluate each
1 X N of these splines for that x. Then do one spline of
YðxÞ ¼ y e2ikpx=L
L k¼N k size NY in the y direction, doing both the deter-
mination and the evaluation.
which is differentiated to obtain Multidimensional integrals can also be bro-
ken down into one-dimensional integrals. For
dY 1 X N
2pik 2ikpx=L
example,
¼ y e
dx L k¼N k L
Zb f Z2 ðxÞ Zb
zðx; yÞdxdy ¼ GðxÞdx;
Thus, at the grid points a f 1 ðxÞ a
fZ
2 ðxÞ
dY 1 X N
2pik 2ikpxn =L
jn ¼ y e GðxÞ ¼ zðx; yÞdx
dx L k¼N k L
f 1 ðxÞ
The variable i is the imaginary unit which has Since the arctangent repeats itself in multiples
the property of p rather than 2 p, the argument must be
defined carefully. For example, the u given
i2 ¼ 1 above could also be the argument of (x þi y).
The function z ¼ cos u þ i sin u obeys |z| ¼ |cos u
The real and imaginary parts of a complex þ i sin u| ¼ 1.
number are often referred to: The rules of equality, addition, and multipli-
cation are
Re ðzÞ ¼ x; Im ðzÞ ¼ y
z1 ¼ x1 þ i y1 ; z2 ¼ x2 þ i y2
A complex number can also be represented Equality:
graphically in the complex plane, where the
z1 ¼ z2 if and only if x1 ¼ x2 and y1 ¼ y2
real part of the complex number is the abscissa
and the imaginary part of the complex number Addition:
is the ordinate (see Fig. 7). z1 þ z2 ¼ ðx1 þ x2 Þ þ iðy1 þ y2 Þ
Another representation of a complex number Multiplication:
is the polar form, where r is the magnitude and
z1 z2 ¼ ðx1 x2 y1 y2 Þ þ iðx1 y2 þ x2 y1 Þ
u is the argument.
pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi The last rule can be remembered by
r ¼ jx þ i yj ¼ x2 þ y2 ; u ¼ argðx þ i yÞ
using the standard rules for multiplication,
Write keeping the imaginary parts separate, and
using i2 ¼ 1. In the complex plane, addition
z ¼ x þ i y ¼ rðcosu þ i sinuÞ
is illustrated in Figure 8. In polar form, multi-
plication is
so that
z1 z2 ¼ r1 r2 ½cosðu1 þ u2 Þ þ i sinðu1 þ u2 Þ
x ¼ r cos u; y ¼ r sin u
The magnitude of z1 þ z2 is bounded by
and
jz1
z2 j jz1 j þ jz2 j and jz1 j jz2 j jz1
z2 j
y
u ¼ arctan
x
as can be seen in Figure 8. The magnitude and
arguments in multiplication obey
Roots of a complex number are complicated by The second equation follows from the defini-
careful accounting of the argument tions
ez þ ez ez ez
cosh z ; sinh z
z ¼ w1=n with w ¼ Rðcos Q þ i sin QÞ; 0 Q 2p 2 2
sinðz1 þ z2 Þ ¼ sin z1 cos z2 þ cos z1 sin z2 Let f (z) be a single-valued continuous function
of z in a domain D. The function f (z) is
differentiable at the point z0 in D if
The same is true of hyperbolic functions. The
absolute boundaries of sin z and cos z are not f ðz0 þ hÞ f ðz0 Þ
lim
bounded for all z. h!0 h
Trigonometric identities can be defined by
using exists as a finite (complex) number and is
independent of the direction in which h tends
eiu ¼ cos u þ i sin u to zero. The limit is called the derivative, f 0 (z0).
The derivative now can be calculated with h
approaching zero in the complex plane, i.e.,
For example, anywhere in a circular region about z0. The
function f(z) is differentiable in D if it is dif-
eiðaþbÞ ¼ cosða þ bÞ þ i sinða þ bÞ ferentiable at all points of D; then f(z) is said to
¼ eia eib ¼ ðcos a þ i sin aÞ be an analytic function of z in D. Also, f(z) is
ðcos b þ i sin bÞ
analytic at z0 if it is analytic in some e neigh-
borhood of z0. The word analytic is sometimes
¼ cos a cos b sin a sin b replaced by holomorphic or regular.
þ i ðcos a sin b þ cos b sin aÞ The Cauchy–Riemann equations can be used
to decide if a function is analytic. Set
Equating real and imaginary parts gives
f ðzÞ ¼ f ðx þ i yÞ ¼ uðx; yÞ þ i vðx; yÞ
sinða þ bÞ ¼ cos a sin b þ cos b sin a Theorem [30, p. 51]. Suppose that f (z) is
defined and continuous in some neighborhood
The logarithm is defined as of z ¼ z0. A necessary condition for the exis-
tence of f 0 (z0) is that u (x, y) and v (x, y) have
ln z ¼ lnjzj þ i arg z first-order partials and that the Cauchy–Rie-
mann conditions (see below) hold.
and the various determinations differ by multi- @u @v @u @v
¼ and ¼ at z0
ples of 2 pi. Then, @x @y @y @x
eln z ¼ z
Theorem [30, p. 61]. The function f (z) is
lnðez Þ z 0ðmod 2 piÞ analytic in a domain D if and only if u and v
are continuously differentiable and satisfy the
Also, Cauchy–Riemann conditions there.
If f1(z) and f2(z) are analytic in domain D,
lnðz1 z2 Þ ln z1 ln z2 0ðmod 2 piÞ then a1f1(z) þ a2f2(z) is analytic in D for any
Mathematics in Chemical Engineering 25
Maximum Principle. If f (z) is analytic in a Also if s (t) is the arc length on C and l (C) is
domain D and continuous in the set consisting the length of C
of D and its boundary C, and if | f (z)| M on C,
then | f (z)| < M in D unless f (z) is a constant Z
f ðzÞdz max jf ðzÞjlðCÞ
[30, p. 134].
z2C
C
C : z ¼ zðtÞ; 0 t 1
Cauchy’s Theorem [25, 30, p. 111]. Suppose f
where z (t) is a continuous function of bounded (z) is an analytic function in a domain D and C
variation; C is oriented such that z1 ¼ z (t1) is a simple, closed, rectifiable curve in D such
precedes the point z2 ¼ z (t2) on C if and only if that f (z) is analytic inside and on C. Then
t1 < t2. Define I
f ðzÞdz ¼ 0 ð4Þ
C
Z Z1
f ðzÞdz ¼ f ½z ðtÞ dz ðtÞ
C 0 If D is simply connected, then Equation (4)
holds for every simple, closed, rectifiable curve
C in D. If D is simply connected and if a and b
The integral is linear with respect to the inte- are any two points in D, then
grand:
Zb
R
C ½a1 f 1 ðzÞ þ a2 f 2 ðzÞdz f ðzÞdz
R R a
¼ a1 C f 1 ðzÞdz þ a2 C f 2 ðzÞdz
Z Z1
f ðzÞdz ¼ f ½zðtÞz0 ðtÞdt Power Series. If f (z) is analytic interior to a
C 0 circle |z z0| < r0, then at each point inside the
Mathematics in Chemical Engineering 27
n ¼ 0;
1;
2; . . . ;
all m < n, then z0 is called a pole of order m. It is
a simple pole if m ¼ 1. In such a case,
and C is a closed curve counterclockwise in R.
A1 X
1
f ðzÞ ¼ þ An ðz z0 Þn
z z0 n¼0
Singular Points and Residues [33, p. 159,
30, p. 180]. If a function in analytic in every
If a function is not analytic at z0 but can be made
neighborhood of z0, but not at z0 itself, then z0 is
so by assigning a suitable value, then z0 is a
called an isolated singular point of the function.
removable singular point.
About an isolated singular point, the function
When f (z) has a pole of order m at z0,
can be represented by a Laurent series.
pðz0 Þ
A1 ¼
q0 ðz0 Þ
1
f 1 ðzÞ ¼
1z 3.5. Other Results
is the analytic continuation onto the entire z Theorem [32, p. 84]. Let P (z) be a polynomial
plane except for z ¼ 1. of degree n having the zeroes z1, z2, . . . , zn
Mathematics in Chemical Engineering 29
and let p be the least convex polygon contain- 4. Integral Transforms [34–39]
ing the zeroes. Then P0 (z) cannot vanish any-
where in the exterior of p. 4.1. Fourier Transforms
If a polynomial has real coefficients, the
roots are either real or form pairs of complex Fourier Series [40]. Let f (x) be a function that
conjugate numbers. is periodic on p < x < p. It can be expanded
The radius of convergence R of the Taylor in a Fourier series
series of f (z) about z0 is equal to the distance
from z0 to the nearest singularity of f (z). a0 X 1
f ðxÞ ¼ þ ðan cos n x þ bn sin n xÞ
2 n¼1
It is desired that Zp
1
bn ¼ f ðxÞsin n x dx
p
U ðj; hÞ ¼ u½^xðj; hÞ; ^yðj; hÞ p
be a harmonic function of j and h. The values {an} and {bn} are called the finite
cosine and sine transform of f, respectively.
Theorem [30, p. 284]. The transformation Because
z ¼ f ðzÞ ð6Þ 1
cos n x ¼ ðeinx þ einx Þ
2
1 inx
takes all harmonic functions of x and y into and sin n x ¼
2i
ðe einx Þ
harmonic functions of j and h if and only if
either f (z) or f (z) is an analytic function of
z ¼ j þi h. the Fourier series can be written as
Equation (6) is a restriction on the transfor-
mation which ensures that X
1
f ðxÞ ¼ cn einx
n¼1
@2 u @2 u @2 U @2 U
if þ ¼ 0 then þ 2 ¼0
@x2 @y2 @z2 @h where
8
Such a mapping with f (z) analytic and f 0 (z) 6¼ 0 >
>
1
< 2 ðan þ i bn Þ for n 0
is a conformal mapping. cn ¼
>
>
If Laplace’s equation is to be solved in the : 1 ðan i bn Þ for n < 0
2
region exterior to a closed curve, then the point
at infinity is in the domain D. For flow in a long
channel (governed by the Laplace equation) the and
inlet and outlet are at infinity. In both cases the
transformation Zp
1
cn ¼ f ðxÞeinx dx
2p
azþb p
z¼
z z0
If f is real
takes z0 into infinity and hence maps D into a
bounded domain D . cn ¼ cn :
30 Mathematics in Chemical Engineering
@T @ 2 T
If f (x) is continuous and piecewise continu-
¼ 2 ously differentiable,
@t @x
T ðx; 0Þ ¼ f ðxÞ
Z1
T ðp; tÞ ¼ T ðp; tÞ f ðxÞeivx dx
1
Let
converges for each v, and
X
1
T¼ cn ðtÞeinx lim f ðxÞ ¼ 0
1 x!
1
Then, then
X
1 X
1 df
dcn inx 2 inx F ¼ ivF ½f
e ¼ cn ðtÞðn Þe dx
1
dt 1
Z1
Fourier Transform [40]. When the function jf ðxÞj2 dx
f(x) is defined on the entire real line, the Fourier 1
Mathematics in Chemical Engineering 31
has a finite value. If f (x) and g (x) are square This is also the total power in a signal, which
integrable, the product f (x) g (x) is absolutely can be computed in either the time or the
integrable and satisfies the Schwarz inequality: frequency domain. Also
1 2 Z1 Z1
Z ^f ðvÞ^g ðvÞdv ¼ 2 p
f ðxÞg ðxÞdx
f ðxÞ g ðxÞdx
1 1
1
Z1 Z1 Z1
1
j^f ðvÞj2 dv ¼ 2 p jf ðxÞj2 dx ¼ eivx0 ^f ðvÞ^h ðvÞdv
2p
1 1 1
32 Mathematics in Chemical Engineering
Theorem. The product The finite Fourier sine and cosine transforms
are
^f ðvÞ^h ðvÞ
Zp
2
f s ðnÞ ¼ Fns ½f ¼ f ðxÞsin n xdx;
p
0
is the Fourier transform of the convolution
product f h. The convolution permits finding n ¼ 1; 2; . . . ;
n ¼ 0; 1; 2; . . .
@T @ 2 T
¼ 2 X
1
@t @x f ðxÞ ¼ f s ðnÞ sin n x;
T ðx; 0Þ ¼ f ðxÞ; 1 < x < 1 n¼1
1 X1
T bounded f ðxÞ ¼ f ð0Þ þ f c ðnÞ cos n x
2 c n¼1
The solution is
f, f 0 are continuous, f 00 is piecewise continuous
^ tÞ ¼ ^f ðvÞev t on 0 x p.
2
Tðv;
Z1
1
eivx ^f ðvÞev t dv
2
T ðx; tÞ ¼
2p
1
4pt
1 ðpxÞ
n p
ð1Þnþ1 x
n p
Also, 1ð1Þn
1
n
ðp cÞ x ðx cÞ
df
p
n2 sin n cð0 < c < pÞ
Fns ¼ nFnc ½f c ðp xÞ ðx cÞ
dx
x ðx < cÞ
p
n cos n cð0 < c < pÞ
df 2 2 px ðxÞ
Fnc ¼ nFns ½f f ð0Þ þ ð1Þn f ðpÞ
dx p p p2 ð1Þn1 n
n 2½1ð1Þ
n3 x2
2
pð1Þn n63 pn x3
When two functions F (x) and G (x) are defined
on the interval 2 p < x < 2 p, the function ½1 ð1Þn ec p ecx
n
n2 þc2
n sinh cðpxÞ
n2 þc2
Zp sinh c p
F ðxÞ G ðxÞ ¼ f ðx yÞ g ðyÞdy n
ðjkj 6¼ 0; 1; 2; . . .Þ sin kðpxÞ
n2 k2 sin k p
p
n cx
ð1Þ e 1
Z1 n2 þc2
1 cx
ce
Fvs ½f f ðxÞsin v x dx;
1 cosh cðpxÞ
0 n2 þc2 c sinh c p
it must be that
Thus, the techniques described herein can be
F 2 ¼ F 1 þ NðtÞ
applied only to linear problems. Generally, the
assumptions made below are that F (t) is at least ZT
where N ðtÞ dt ¼ 0 for every T
piecewise continuous, that it is continuous in
each finite interval within 0 < t < 1, and that it
0
may take a jump between intervals. It is also Laplace transforms can be inverted by using
of exponential order, meaning eat|F (t)| is Table 5, but knowledge of several rules is
bounded for all t > T, for some finite T. helpful.
The unit step function is
Substitution.
0 0t<k
Sk ðtÞ ¼ f
1 t>k f ðs aÞ ¼ L½eat FðtÞ
and its Laplace transform is This can be used with polynomials. Suppose
eks 1 1 2sþ3
L½Sk ðtÞ ¼ f ðsÞ ¼ þ ¼
s s s þ 3 s ðs þ 3Þ
1 1
L½1 ¼ L½1 ¼
s s
Mathematics in Chemical Engineering 35
L [F] F (t)
1
s 1
1
s2 t
1 tn1
sn ðn1Þ!
p1ffi p1ffiffiffiffiffi
s pt
pffiffiffiffiffiffiffi
s3/2 2 t=p
Figure 10. Square wave function
G ðkÞ
sk ðk > 0Þ tk1
1
sa ea t
has the Laplace transform
1
ðsaÞn
ðn ¼ 1; 2; . . .Þ 1
ðn1Þ! t
n1 a t
e
1 1
G ðkÞ L½S ðtÞ ¼
ðsaÞk
ðk > 0Þ tk1ea t s 1 ehs
1 1 at
ab ðe eb t Þ
ðsaÞðsbÞ
The Dirac delta function d (t t0) has the
s
ðsaÞðsbÞ
1
ab ðae
at
beb t Þ property
1 1
s2 þa2 a sin a t Z1
dðt t0 Þ F ðtÞ dt ¼ Fðt0 Þ
s
s2 þa2 cos a t
0
1 1
s2 a2 a sinh a t
Its Laplace transform is
s
s2 a2 cosh a t
L½d ðt t0 Þ ¼ est0 ; t0 0; s > 0
s t
ðs2 þa2 Þ2 2 a sin a t
s2 a2
t cos a t The square wave function illustrated in
ðs2 þa2 Þ2
Figure 10 has Laplace transform
1 1 at
ðsaÞ2 þb2 be sin b t
1 cs
sa
ea t cos b t L½F c ðtÞ ¼ tanh
ðsaÞ2 þb2 s 2
More generally, translation gives the following. Other Laplace transforms are listed in Table 5.
Translation.
t
b 1
f ða s bÞ ¼ f a s ¼ L ebt=a F ;a>0
a a a
8 1
>
> 0t<
>0
>
>
> h
>
<
1 2
S ðtÞ ¼ 1 t<
>
> h h
>
>
>
>
>
:2 2 3
t<
h h Figure 11. Triangular wave function
36 Mathematics in Chemical Engineering
Convolution properties are also satisfied: If the factor (s a) is repeated m times, then
Zt p ðsÞ Am Am1
F ðtÞ G ðtÞ ¼ F ðtÞ G ðt tÞdt f ðsÞ ¼ ¼ þ þ
q ðsÞ ðs aÞm ðs aÞm1
0
A1
þ þ h ðsÞ
sa
and
where
f ðsÞgðsÞ ¼ L½FðtÞ GðtÞ
ðs aÞm p ðsÞ
f ðsÞ
q ðsÞ
Derivatives of Laplace Transforms.The Laplace
integrals L [F (t)], L [t F (t)], L [t2 F (t)], . . . are 1 dmk f ðsÞ
Am ¼ f ðaÞ; Ak ¼ ja ;
ðm kÞ! dsmk
uniformly convergent for s1 a and
k ¼ 1; . . . ; m 1
1 1
All the factors are linear, none are repeated, f ðsÞ ¼ ; f0 ðsÞ ¼
s2 ðs 2Þ2
and the an are all distinct. If p (s) has a smaller
f ð1Þ ¼ 1; f0 ð1Þ ¼ 1
degree than q (s), the Heaviside expansion can
be used to evaluate the inverse transformation:
The inverse Laplace transform is then
X m
1 p ðsÞ p ðai Þ ai t
L ¼ e
q ðsÞ i¼1
q0 ðai Þ FðtÞ ¼ e2t þ ½1 t et
Mathematics in Chemical Engineering 37
Quadratic Factors. Let p (s) and q (s) have real To solve an integral equation:
coefficients, and q (s) have the factor
Zt
Y ðtÞ ¼ a þ 2 Y ðtÞcos ðt tÞdt
ðs aÞ þ b ; b > 0
2 2
0
p ðsÞ f ðsÞ
f ðsÞ ¼ ¼
q ðsÞ ðs aÞ2 þ b2 Then the Laplace transform is used to obtain
AsþB
¼ þ h ðsÞ a s
ðs aÞ2 þ b2 y ðsÞ ¼ þ 2 y ðsÞ 2
s s þ1
a ðs2 þ 1Þ
Let f1 and f2 be the real and imaginary parts of or y ðsÞ ¼
s ðs 1Þ2
the complex number f (a þi b).
1
ðs2 2 s þ 1Þ y ðsÞ ¼ The functions | f(s)| and |x f(s)| are bounded
s2
in the half-plane x x1 > x0 and f (s) ! 0 as
1
y ðsÞ ¼ |y| ! 1 for each fixed x. Thus,
ðs 2Þðs 1Þ2
jf ðx þ i yÞj < M; jx f ðx þ i yÞj < M;
lead to x x1 > x 0
lim f ðx þ i yÞ ¼ 0; x > x0
YðtÞ ¼ e2t ð1 þ tÞ et y!
1
38 Mathematics in Chemical Engineering
If F (t) is continuous, F0 (t) is piecewise contin- Also F (t) and its n derivatives are continuous
uous, and both functions are O [exp (x0t)], then functions of t of order O [exp (x0t)] and they
| f (s)| is O (1/s) in each half-plane x x1 > x0. vanish at t ¼ 0.
If F (t) and F0 (t) are continuous, F00 (t) is piece- Series of Residues [41]. Let f (s) be an analytic
wise continuous, and all three functions are O function except for a set of isolated singular
[exp (x0t)], then points. An isolated singular point is one for
which f (z) is analytic for 0 < |z z0| < % but z0
js2 f ðsÞ s Fð0Þj < M; x x1 > x0 is a singularity of f (z). An isolated singular
point is either a pole, a removable singularity, or
an essential singularity. If f (z) is not defined in
The additional constraint F (0) ¼ 0 is necessary the neighborhood of z0 but can be made analytic
and sufficient for | f (s)| to be O (1/s2). at z0 simply by defining it at some additional
points, then z0 is a removable singularity. The
Inversion Integral [41]. Cauchy’s integral for- function f (z) has a pole of order k 1 at z0 if
mula for f (s) analytic and O (sk) in a half-plane (z z0)k f (z) has a removable singularity at z0
x y, k > 0, is whereas (z z0)k1f (z) has an unremovable
isolated singularity at z0. Any isolated singu-
Z
gþib larity that is not a pole or a removable singu-
1 f ðzÞ
f ðsÞ ¼ lim dz; Re ðsÞ > g larity is an essential singularity.
2 pi b!1 sz
gib
Let the function f (z) be analytic except for
the isolated singular point s1, s2, . . . , sn. Let
%n (t) be the residue of ezt f (z) at z ¼ sn (for
Applying the inverse Laplace transformation on
definition of residue, see Section 3.4). Then
either side of this equation gives
X
1
Z
gþib
F ðtÞ ¼ %n ðtÞ
1 zt
F ðtÞ ¼ lim e f ðzÞdz n¼1
2 pi b!1
gib
Z
gþib If sn is a removable pole of f (s), of order m,
dn F 1
¼ lim ezt zn f ðzÞdz; then
dtn 2 pi b!1
gib
n ¼ 1; 2; . . . ; m fn ðzÞ ¼ ðz sn Þm f ðzÞ
Mathematics in Chemical Engineering 39
The solution is
Problem 1. Infinite domain, on 1 < x < 1.
pffiffiffiffiffiffiffiffi
Tðx; 0Þ ¼ f ðxÞ; initial conditions Tðx ; tÞ ¼ T 0 þ ½T 1 T 0 ½1 erf ðx= 4a tÞ
pffiffiffiffiffiffiffiffi
Tðx; tÞ bounded or Tðx; tÞ ¼ T 0 þ ðT 1 T 0 Þ erfc ðx= 4a tÞ
40 Mathematics in Chemical Engineering
Problem 3. Semi-infinite domain, boundary The solution for T1 can also be obtained by
condition of the first kind, on 0 x < 1 Laplace transforms.
Tðx; 0Þ ¼ f ðxÞ
t1 ¼ L½T 1
Tð0; tÞ ¼ g ðtÞ
where
and solving gives
T 1 ðx; 0Þ ¼ f ðxÞ; T 2 ðx; 0Þ ¼ 0
Thus,
Problem 4. Semi-infinite domain, boun-
2 dary conditions of the second kind, on
U 1 ðv; tÞ ¼ F vs ½f ev at
0 x < 1.
and [40, p. 322] Tðx; 0Þ ¼ 0
Z1
2 2
@T
T 1 ðx; tÞ ¼ F vs ½f ev at
sin vx dv k ð0; tÞ ¼ q0 ¼ constant
p @x
0
Solve for T2 by taking the sine transform Take the Laplace transform
@U 2 @2 t
¼ v2 a U 2 þ a gðtÞ v st ¼ a
@t @x2
U 2 ðv; 0Þ ¼ 0 @t q0
k ¼
@x s
Thus,
The solution is
Zt pffiffiffi
2 q0 a xpffiffiffiffiffi
U 2 ðv; tÞ ¼ ev aðttÞ
av g ðtÞdt t ðx; sÞ ¼ e s=a
k s3=2
0
tð0; sÞ ¼ tðL; sÞ ¼ 0
@T
k ð0; tÞ ¼ h T ð0; tÞ
@x
The solution is
Take the Laplace transform pffiffiffiffi pffiffi
T 0 sinh as x T 0 sinh as ðL xÞ T 0
t ðx; sÞ ¼ pffiffis pffiffi þ
s sinh a L s sinh as L s
@2 t
s t f ðxÞ ¼ a
@x2
@t The inverse transformation is [41, p. 220], [44,
k ð0; sÞ ¼ h t ð0; sÞ p. 96]
@x
Tð0; tÞ ¼ 0
Z1 Z1 TðL; 0Þ ¼ T 0 ¼ constant
2 2
T ðx; tÞ ¼ eb at
cos ½b x m ðbÞ f ðjÞ
p
0 0
Take the Laplace transform
cos ½b j m ðbÞ djdb
Problem 6. Finite domain, boundary condi- and the inverse transformation is [41, p. 201],
tion of the first kind [44, p. 313]
" #
Tðx; 0Þ ¼ T 0 ¼ constant x 2X 1
ð1Þn n2 p2 at=L2 n p x
T ðx; tÞ ¼ T 0 þ e sin
L p n¼1 n L
Tð0; tÞ ¼ T ðL; tÞ ¼ 0
42 Mathematics in Chemical Engineering
with magnitude
The transpose of A is
pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
jrj ¼ x2 þ y 2 þ z 2 ATij ¼ Aji
but
Because the dyadics may not be symmetric, the
u v 6¼ v u order of indices and which indices are summed
are important. The order is made clearer when
The Kronecker delta is defined as the dyadics are made from vectors.
u v ¼ jujjvjcos u; 0 u p c ¼ u v ¼ a j u j j v j sin u; 0 u p
where u is the angle between u and v. The scalar where a is a unit vector in the direction of uv.
product of two vectors is a scalar, not a vector. It The direction of c is perpendicular to the plane
is the magnitude of u multiplied by the projec- of u and v such that u, v, and c form a right-
tion of v on u, or vice versa. The scalar product handed system. If u ¼ v, or u is parallel to v,
of u with itself is just the square of the magni- then u ¼ 0 and uv ¼ 0. The following laws are
tude of u. valid for cross products.
u u ¼ ju2 j ¼ u2
uv¼vu Commutative law fails for vector
The following laws are valid for scalar products product
u(vw) 6¼ (uv)w Associative law fails for vector
product
u v ¼ v u Commutative law for scalar u(v þ w) ¼ uv þ uw Distributive law for vector
products product
u (v þ w) ¼ u v þ u Distributive law for scalar products ex ex¼ ey ey¼ ez ez ¼ 0
w
ex ey¼ ez, ey ez¼ ex, ez
ex ex ¼ ey ey¼ ez ez ¼ 1
ex¼ ey
ex ey¼ ex ez¼ ey ez ¼ 0
2 3
ex ey ez
If the two vectors u and v are written in com- 6
6
7
7
ponent notation, the scalar product is u v ¼ det6 ux uy uz 7
4 5
vx vy vz
u v ¼ u x vx þ u y vy þ u z vz
¼ ex ðuy vz vy uz Þ þ ey ðuz vz ux vz Þ
If u v ¼ 0 and u and v are not null vectors, then
þez ðux vy uy vx Þ
u and v are perpendicular to each other and u ¼
p/2.
This can also be written as
The single dot product of two dyadics is
XX X XX
AB¼ ei ej ð Aik Bkj Þ uv¼ ekij ui vj ek
i j k i j
Mathematics in Chemical Engineering 45
Thus e123 ¼ 1, e132 ¼1, e312 ¼ 1, e112 ¼ 0, for can be formed where l is an eigenvalue. This
example. expression is
The magnitude of uv is the same as the
area of a parallelogram with sides u and v. If
l3 I 1 l2 þ I 2 l I 3 ¼ 0
uv ¼ 0 and u and v are not null vectors, then u
and v are parallel. Certain triple products are
useful. An important theorem of HAMILTON and CAYLEY
[47] is that a second-order dyadic satisfies its
ðu vÞw 6¼ uðv wÞ own characteristic equation.
u ðv wÞ ¼ v ðw uÞ ¼ w ðu vÞ
Z t rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
dr dr
s ðtÞ ¼ dt
dt dt
a
This gives
Figure 15. Vector differentiation
2 2 2 2
ds dr dr dx dy dz
¼ ¼ þ þ
dt dt dt dt dt dt
is the velocity. The derivative operation obeys
the following laws.
Because
d du dv
ðu þ vÞ ¼ þ
dt dt dt dr ¼ dxex þ dyey þ dzez
d du dv
ðu vÞ ¼ vþu
dt dt dt then
d du dv
ðu vÞ ¼ vþu
dt dt dt ds2 ¼ dr dr ¼ dx2 þ dy2 þ dz2
d du du
ðuuÞ ¼ u þ u
dt dt dt
The derivative dr/dt is tangent to the curve in
If the vector u depends on more than one the direction of motion
variable, such as x, y and z, partial derivatives
are defined in the usual way. For example, dr
u ¼ dt
dr
if uðx; y; zÞ; then dt
@u uðx þ Dx; y; zÞ uðx; y; zÞ
¼ lim
@x Dt!0 Dx Also,
Rules for differentiation of scalar and vector dr
u¼
products are ds
@ @u @v
ðu vÞ ¼ vþu Differential Operators. The vector differential
@x @x @x
@ @u @v operator (del operator) b is defined in cartesian
ðu vÞ ¼ vþu coordinates by
@x @x @x
dðu vÞ ¼ du v þ u dv @f @f @f
rf ¼ ex þ ey þ ez
@u @u @u @x @y @z
du ¼ dx þ dy þ dz
@x @y @z
and is a vector. If f is height or elevation, the
gradient is a vector pointing in the uphill direc-
If a curve is given by r(t), the length of the tion. The steeper the hill, the larger is the
curve is [43] magnitude of the gradient.
The divergence of a vector is defined by
Z b rffiffiffiffiffiffiffiffiffiffiffiffi
dr dr
L¼ dt @ux @uy @uz
dt dt r u¼ þ þ
a @x @y @z
Mathematics in Chemical Engineering 47
and is a scalar. For a volume element DV, the net Useful formulas are [49]
outflow of a vector u over the surface of the r ðfuÞ ¼ rf u þ fr u
element is
r ðfuÞ ¼ rf u þ fr u
Z
r ðu vÞ ¼ v ðr uÞ u ðr vÞ
u ndS
DS r ðu vÞ ¼ v ru vðr uÞ u rv þ uðr vÞ
r ðr uÞ ¼ rðr uÞ r2 u
This is related to the divergence by [48, p. 411]
@2 f @2 f @2 f
r ðrfÞ ¼ r2 f ¼ þ þ ; where r2
Z @x2 @y2 @z2
1
r u ¼ lim u ndS
DV!0 DV is called the Laplacian operator. r(r
r w) ¼ 0.
DS
The curl of the gradient of f is zero.
Thus, the divergence is the net outflow per unit r ðr uÞ ¼ 0
volume.
The curl of a vector is defined by The divergence of the curl of u is zero. Formu-
las useful in fluid mechanics are
@ @ @ r ðrvÞT ¼ rðr vÞ
ru¼ þ ey þ ez
ex ðex ux þ ey uy þ ez uz Þ
@x @y @z
r ðt vÞ ¼ v ðr tÞ þ t : rv
@uz @uy @ux @uz
¼ ex þ ey 1
@y @z @z @x v rv ¼ rðv vÞ v ðr vÞ
2
@uy @ux
þez If a coordinate system is transformed by a rota-
@x @y
tion and translation, the coordinates in the new
and is a vector. It is related to the integral system (denoted by primes) are given by
0 1 0 1
x0 l11 l12 l13
Z Z B 0C B C
By C ¼ B l21 l22 l23 C
u ds ¼ us ds @ A @ A
C C z0 l31 l32 l33
0 1 0 1
x a1
which is called the circulation of u around path B C
ByC
B C
B a2 C
@ A þ @ A
C. This integral depends on the vector and the
contour C, in general. If the circulation does not z a3
depend on the contour C, the vector is said to be Any function that has the same value in all
irrotational; if it does, it is rotational. The coordinate systems is an invariant. The gradient
relationship with the curl is [48, p. 419] of an invariant scalar field is invariant; the same
Z is true for the divergence and curl of invariant
1
n ðr uÞ ¼ lim
D!0 DS
u ds vectors fields.
C The gradient of a vector field is required in
fluid mechanics because the velocity gradient is
Thus, the normal component of the curl equals used. It is defined as
the net circulation per unit area enclosed by the XX @vj
rv ¼ ei ej and
contour C. i j
@xi
The gradient, divergence, and curl obey a XX @vi
distributive law but not a commutative or asso- ðrvÞT ¼ ei ej
i j
@xj
ciative law.
The divergence of dyadics is defined
rðf þ cÞ ¼ rf þ rc !
P X @t ji
rt ¼ i ei and
r ðu þ vÞ ¼ r u þ r v j
@xj
" #
r ðu þ vÞ ¼ r u þ r v X X @
r ðfu vÞ ¼ ei ðfuj vi Þ
r f 6¼ fr i j
@xj
r ðftÞ ¼ rf t þ fr t
Vector Integration [48, pp. 206–212]. If u is a
nt : t ¼ ttn ¼ t : nt vector, then its integral is also a vector.
posed into
Theorems about this line integral can be written
@ in various forms.
v ¼ vII þ nvn ; r ¼ rII þ n
@n
Theorem [43]. If the functions appearing in the
The velocity gradient can be decomposed into line integral are continuous in a domain D, then
the line integral is independent of the path C if
@vII @vn
rv ¼ rII vII þ ðrII nÞvn þ nrII vn þ n þ nn and only if the line integral is zero on every
@n @n
simple closed path in D.
The surface gradient of the normal is the nega-
tive of the curvature dyadic of the surface. Theorem [43]. If u ¼ rf where f is single-
valued and has continuous derivatives in D,
rII n ¼ B then the line integral is independent of the
path C and the line integral is zero for any
The surface divergence is then closed curve in D.
rII v ¼ dII : rv ¼ rII vII 2 Hvn Theorem [43]. If f, g, and h are continuous
functions of x, y, and z, and have continuous
where H is the mean curvature. first derivatives in a simply connected domain
D, then the line integral
1 Z
H ¼ dII : B
2 ðf dx þ gdy þ hdzÞ
C
or if f, g, and h are regarded as the x, y, and z Then the divergence theorem in component
components of a vector v: form is
rv¼0 Z Z
@ux @uy @uz
þ þ dxdydz ¼ ½ux cos ðx; nÞ
@x @y @z
Consequently, the line integral is independent V S
of the path (and the value is zero for a closed þuy cos ðy; nÞ þ uz cos ðz; nÞdS
contour) if the three components in it are
regarded as the three components of a vector If the divergence theorem is written for an
and the vector is derivable from a potential (or incremental volume
zero curl). The conditions for a vector to be
Z
derivable from a potential are just those in the r u ¼ lim
1
un dS
DV!0 DV
third theorem. In two dimensions this reduces DS
to the more usual theorem.
the divergence of a vector can be called the
Theorem [48, p. 207]. If M and N are continu-
integral of that quantity over the area of a closed
ous functions of x and y that have continuous
volume, divided by the volume. If the vector
first partial derivatives in a simply connected
represents the flow of energy and the diver-
domain D, then the necessary and sufficient
gence is positive at a point P, then either a
condition for the line integral
source of energy is present at P or energy is
Z
leaving the region around P so that its temper-
ðMdx þ NdyÞ
C
ature is decreasing. If the vector represents the
flow of mass and the divergence is positive at a
to be zero around every closed curve C in D is point P, then either a source of mass exists at P
or the density is decreasing at the point P. For an
@M @N incompressible fluid the divergence is zero and
¼
@y @x
the rate at which fluid is introduced into a
If a vector is integrated over a surface with volume must equal the rate at which it is
incremental area d S and normal to the surface removed.
n, then the surface integral can be written as Various theorems follow from the diver-
gence theorem.
ZZ ZZ
u dS ¼ u ndS Theorem. If f is a solution to Laplace’s equa-
S S tion
Also if f satisfies the conditions of the theorem Theorem [48, p. 423]. The necessary and suf-
and is zero on S then f is zero throughout D. If ficient condition that the divergence of a vector
two functions f and c both satisfy the Laplace vanish identically is that the vector is the curl of
equation in domain D, and both take the same some other vector.
values on the bounding curve C, then f ¼ c;
i.e., the solution to the Laplace equation is Leibniz Formula. In fluid mechanics and
unique. transport phenomena, an important result is
The divergence theorem for dyadics is the derivative of an integral whose limits of
integration are moving. Suppose the region V
Z Z (t) is moving with velocity vs. Then Leibniz’s
r tdV ¼ n t dS
rule holds:
V S
ZZZ ZZZ ZZ
d @f
fdV ¼ dV þ fvs ndS
Stokes Theorem [48, 49]. Stokes theorem says dt @t
VðtÞ VðtÞ S
that if S is a surface bounded by a closed,
nonintersecting curve C, and if u has continu- Curvilinear Coordinates. Many of the rela-
ous derivatives then tions given above are proved most easily by
I ZZ ZZ using tensor analysis rather than dyadics. Once
u dr ¼ ðr uÞ ndS ¼ ðr uÞ dS proven, however, the relations are perfectly
C S S general in any coordinate system. Displayed
here are the specific results for cylindrical and
The integral around the curve is followed in the spherical geometries. Results are available for a
counterclockwise direction. In component few other geometries: Parabolic cylindrical,
notation, this is paraboloidal, elliptic cylindrical, prolate sphe-
roidal, oblate spheroidal, ellipsoidal, and
H
C ½ux cos ðx; sÞ þ uy cos ðy; sÞ þ uz cos ðz; sÞds bipolar coordinates [45, 50].
ZZ For cylindrical coordinates, the geometry is
@uz @uy @ux @uz
¼ cosðx; nÞ þ cosðy; nÞ shown in Figure 16. The coordinates are related
@y @z @z @x
S to cartesian coordinates by
@uy @ux
þ cosðz; nÞ dS
@x @y
@f eu @f @f
rf ¼ er þ þ ez
@r r @u @z
For spherical coordinates, the geometry is
21@ @f 1 @2 f @2 f shown in Figure 17. The coordinates are related
r f¼ r þ 2 2þ 2
r @r @r r @u @z to cartesian coordinates by
@ 2 f 1 @f 1 @ 2 f @ 2 f
¼ þ þ þ
@r2 r @r r2 @u2 @z2 pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
x ¼ r sin u cosf r ¼ x2 þ p y2ffiffiffiffiffiffiffiffiffiffiffiffiffiffi
þ z2 ffi
1@ 1 @vu @vz y ¼ r sin u sinf u ¼ arctanð
x2 þ y2 =zÞ
rv¼ ðr vr Þ þ þ z ¼ r cos u f ¼ arctan yx
r @r r @u @z
1 @vz @vu @vr @vz
r v ¼ er þ eu
r @u @z @z @r The unit vectors are related by
1@ 1 @vr
þ ez ðr vu Þ er ¼ sin u cos fex þ sin u sin fey þ cos uez
r @r r @u
@c 1 @c 1 @c
rc ¼ er
@r
þ eu
r @u
þ ef
r sin u @f
6. Ordinary Differential Equations
as Initial Value Problems
1 @ @c 1 @ @c
r2 c ¼ r2 þ 2 sin u þ
r @r
2 @r r sin u @u @u A differential equation for a function that
1 @2c
depends on only one variable (often time) is
þ 2 2 called an ordinary differential equation. The
r sin u @f2
general solution to the differential equation
1 @ 2 1 @ includes many possibilities; the boundary or
rv¼ ðr vr Þ þ ðvu sin uÞþ
r2 @r r sin u @u initial conditions are required to specify which
1 @vf of those are desired. If all conditions are at one
þ
r sin u @f point, the problem is an initial value problem
and can be integrated from that point on. If
1 @ 1 @vu some of the conditions are available at one point
r v ¼ er ðvf sin uÞ þ
r sin u @u r sin u @f and others at another point, the ordinary
1 @vr 1 @ differential equations become two-point bound-
þeu ðr vf Þ þ
r sin u @f r @r ary value problems, which are treated in
Chapter 7. Initial value problems as ordinary
1@ 1 @vr
þef ðr vu Þ
r @r r @u differential equations arise in control of
lumped-parameter models, transient models
1 @ 2 1 @ of stirred tank reactors, polymerization
r t ¼ er ðr t rr Þ þ ðt ur sin uÞ
r2 @r r sin u @u
reactions and plug-flow reactors, and generally
1 @tfr t uu þ t ff
þ
r sin u @f
r
þ in models where no spatial gradients occur in
the unknowns.
1 @ 1 @ 1 @tfu
þeu 3 ðr3 t ru Þ þ ðt uu sin uÞ þ
r @r r sin u @u r sin u @f
tur t ru t ff cot u
þ þ
r 6.1. Solution by Quadrature
1 @ 1 @ 1 @tff
þef 3 ðr 3 t rf Þ þ ðt uf sin uÞ þ
r @r r sin u @u r sin u @f
When only one equation exists, even if it is
tfr t rf þ t fu cot u
þ
r
nonlinear, solving it by quadrature may be
possible. For
@vr @vu @vf
rv ¼ er er þ er eu þ er ef þ
@r @r @r dy
¼ f ðyÞ
1 @vr vu 1 @vu vr dt
þeu er þ eu eu þ þ
r @u r r @u r y ð0Þ ¼ y0
1 @vf 1 @vr vf
þeu ef þ ef er þ
r @u r sinu @f r the problem can be separated
1 @vu vf
þef eu cot u þ
r sinu @f r dy
¼ dt
f ðyÞ
1 @vf vr vu
þef ef þ þ cot u
r sinu @f r r
and integrated:
@ 1 @ 2 1 @ @vr
r2 v ¼ er ðr vr Þ þ 2 sin u
@r r2 @r r sin u @u @u
Zy Zt
1 @ 2 vr 2 @ 2 @vf dy0
þ 2 2 ðvu sin uÞ 2 ¼ dt ¼ t
r sin u @f2 r2 sin u @u r sin u @f f ðy0 Þ
y0 0
1 @ @vu 1 @ 1 @
þeu r2 þ 2 ðvu sin uÞ
r @r
2 @r r @u sin u @u If the quadrature can be performed analytically
1 @ 2 vu 2 @vr 2 cot u @vf then the exact solution has been found.
þ 2 2 þ 2
r sin u @f 2 r @u r sin u @f
2
For example, consider the kinetics problem
1 @ @vf 1 @ 1 @
þef 2 r2 þ 2 ðvf sin uÞ with a second-order reaction.
r @r @r r @u sin u @u
1 @ 2 vf 2 @vr 2 cot u @vu
þ 2 2 þ þ 2 dc
r sin u @f 2 r sin u @f r sin u @f
2
¼ kc2 ; cð0Þ ¼ c0
dt
Mathematics in Chemical Engineering 53
To find the function of the concentration versus by using Frechet differentials [51]. In this case,
time, the variables can be separated and inte-
grated. exp
Ft dc F
þ c ¼
d
exp
Ft
c
V dt V dt V
dc
¼ kdt;
c2 Thus, the differential equation can be written as
1
¼ kt þ D
c d Ft Ft F
exp c ¼ exp cin
dt V V V
Application of the initial conditions gives the
solution: This can be integrated once to give
1 1 Zt 0
¼ kt þ Ft F Ft
c c0 exp c ¼ c ð0Þ þ exp cin ðt0 Þ dt0
V V V
0
can be converted into a set of first-order equa- Euler’s method is first order
tions. By using
ynþ1 ¼ yn þ Dtf ðyn Þ
ði1Þ
d y d ði2Þ dyi1
yi yði1Þ ¼ ¼ y ¼
dtði1Þ dt dt
Adams–Bashforth Methods. The second-order
Adams–Bashforth method is
the higher order equation can be written as a set
of first-order equations: Dt
ynþ1 ¼ yn þ ½3 f ðyn Þ f ðyn1 Þ
2
dy1
¼ y2
dt
The fourth-order Adams–Bashforth method is
dy2
¼ y3
dt
Dt
dy3 ynþ1 ¼ yn þ ½55 f ðyn Þ 59 f ðyn1 Þ þ 37 f ðyn2 Þ 9 f ðyn3 Þ
¼ y4 24
dt
...
Notice that the higher order explicit methods
dyn
¼ Fðyn1 ; yn2 ; . . . ; y2; y1 Þ require knowing the solution (or the right-hand
dt
side) evaluated at times in the past. Because
The initial conditions would have to be speci- these were calculated to get to the current time,
fied for variables y1(0), . . . , yn (0), or equiv- this presents no problem except for starting the
alently y (0), . . . , y(n1)(0). The set of evaluation. Then, Euler’s method may have to
equations is then written as be used with a very small step size for several
steps to generate starting values at a succession
dy of time points. The error terms, order of the
¼ fðy; tÞ
dt method, function evaluations per step, and
stability limitations are listed in Table 6.
All the methods in this chapter are described for The advantage of the fourth-order Adams–
a single equation; the methods apply to the Bashforth method is that it uses only one
multiple equations as well. Taking the single function evaluation per step and yet achieves
equation in the form high-order accuracy. The disadvantage is the
necessity of using another method to start.
dy
dt
¼ f ðyÞ Runge–Kutta Methods. Runge–Kutta methods
are explicit methods that use several function
evaluations for each time step. The general
multiplying by dt, and integrating once yields form of the methods is
Ztnþ1 X
v
dy 0 X
tnþ1
dt ¼ f y ðt0 Þ dt0 ynþ1 ¼ yn þ wi k i
dt0 tn i¼1
tn
with
This is
X
i1
ki ¼ Dt f tn þ ci Dt; yn þ aij kj
Ztnþ1 j¼1
dy 0
ynþ1 ¼ yn þ dt
dt0
tn
Runge–Kutta methods traditionally have been
writen for f (t, y) and that is done here, too. If
The last substitution gives a basis for the vari- these equations are expanded and compared
ous methods. Different interpolation schemes with a Taylor series, restrictions can be placed
for y (t) provide different integration schemes; on the parameters of the method to make it first
using low-order interpolation gives low-order order, second order, etc. Even so, additional
integration schemes [3]. parameters can be chosen. A second-order
Mathematics in Chemical Engineering 55
Method Error term Order Function evaluations per step Stability limit, l Dt
Explicit methods
h2 00
Euler 2 y 1 1 2.0
5 3 000
Second-order Adams–Bashforth 12 h y 2 1
251 5 ð5Þ
Fourth-order Adams–Bashforth 720 h y 4 1 0.3
Second-order Runge–Kutta
(midpoint) 2 2 2.0
Runge–Kutta–Gill 4 4 2.8
nþ1 nþ1
Runge–Kutta–Feldberg y z 5 6 3.0
Predictor–corrector methods
Second-order Runge–Kutta 2 2 2.0
Adams, fourth-order 2 2 1.3
Kutta–Gill method is popular because it is high For example, Euler’s method gives
order and does not require a starting method (as
does the fourth-order Adams–Bashforth ynþ1 ¼ yn l Dtyn or ynþ1 ¼ ð1 l DtÞyn
method). However, it requires four function or rmn ¼ 1 l Dt
evaluations per time step, or four times as
many as the Adams–Bashforth method. For The stability limit is then
problems in which the function evaluations
are a significant portion of the calculation l Dt 2
time this might be important. Given the speed
of computers and the widespread availability of The Euler method will not oscillate provided
desktop computers, the efficiency of a method
is most important only for very large problems l Dt 1
that are going to be solved many times. For
other problems the most important criterion for The stability limits listed in Table 6 are
choosing a method is probably the time the user obtained in this fashion. The limit for the Euler
spends setting up the problem. method is 2.0; for the Runge–Kutta–Gill
The stability of an integration method is best method it is 2.785; for the Runge–Kutta–
estimated by determining the rational polyno- Feldberg method it is 3.020. The rational poly-
mial corresponding to the method. Apply this nomials for the various explicit methods are
method to the equation illustrated in Figure 19. As can be seen, the
methods approximate the exact solution well as
dy
¼ ly; yð0Þ ¼ 1 l Dt approaches zero, and the higher order
dt
methods give a better approximation at high
values of l Dt.
and determine the formula for rmn: In solving sets of equations
dy
ykþ1 ¼ rmn ðl DtÞyk ¼ A y þ f; yð0Þ ¼ y0
dt
The rational polynomial is defined as all the eigenvalues of the matrix A must be
examined. FINLAYSON [3] and AMUNDSON [54,
pn ðzÞ p. 197–199] both show how to transform these
rmn ðzÞ ¼ ez
qm ðzÞ
jr mn ðzÞj 1
equations into an orthogonal form so that each error estimate, then Dt is reduced to meet that
equation becomes one equation in one criterion. Also, Dt is increased to as large a
unknown, for which single equation analysis value as possible, because this shortens com-
applies. For linear problems the eigenvalues do putation time. If the local truncation error has
not change, so the stability and oscillation been achieved (and estimated) by using a step
limits must be satisfied for every eigenvalue size Dt1
of the matrix A. When solving nonlinear prob-
lems the equations are linearized about the LTE ¼ c Dtp1
solution at the local time, and the analysis
applies for small changes in time, after which and the desired error is e, to be achieved using a
a new analysis about the new solution must be step size Dt2
made. Thus, for nonlinear problems the eigen-
values keep changing.
Richardson extrapolation can be used to e ¼ c D tp2
differential equations it is called the Crank– difference formulas. The formulas of various
Nicolson method. Adams methods exist as orders are [57]:
well, and the fourth-order Adams–Moulton
method is 1 : ynþ1 ¼ yn þ Dt f ðynþ1 Þ
4 n 1 n1 2
2 : ynþ1 ¼ y y þ Dt f ðynþ1 Þ
Dt 3 3 3
ynþ1 ¼ yn þ ½9 f ðynþ1 Þ þ 19 f ðyn Þ 5 f ðyn1 Þ þ f ðyn2 Þ
24
18 n 9 n1 2 n2
3 : ynþ1 ¼ y y þ y
11 11 11
The properties of these methods are given in 6
þ Dt f ðynþ1 Þ
Table 6. The implicit methods are stable for any 11
step size but do require the solution of a set of 48 n 36 n1 16 n2 3 n3
4 : ynþ1 ¼ y y þ y y
nonlinear equations, which must be solved 25 25 25 25
iteratively. An application to dynamic distilla- 12
tion problems is given in [56]. þ Dt f ðynþ1 Þ
25
All these methods can be written in the
300 n 300 n1 200 n2
form 5 : ynþ1 ¼ y y þ y
137 137 137
75 n3 12 n4
X
k X
k y þ y
137 137
y nþ1
¼ ai y nþ1i
þ Dt bi f ðy nþ1i
Þ
i¼1 i¼0
60
þ Dt f ðynþ1 Þ
137
maxi j Re ðli Þj
SR ¼ dc1
mini j Re ðli Þj ¼ f ðc1 ; c2 Þ
dt
dc2
SR ¼ 20 is not stiff, SR ¼ 103 is stiff, and SR dt
¼ k1 c1 k2 c22
¼ 106 is very stiff. If the problem is nonlinear,
the solution is expanded about the current
The first equation could be the equation for a
state:
stirred tank reactor, for example. Suppose both
X
k1 and k2 are large. The problem is then stiff,
dyi n
@f i
¼ f i ½yðtn Þ þ ½y yj ðtn Þ but the second equation could be taken at
dt j¼1
@yj j
equilibrium. If
The question of stiffness then depends on the
c1 @2c2
eigenvalue of the Jacobian at the current time.
Consequently, for nonlinear problems the prob-
lem can be stiff during one time period and not The equilibrium condition is then
stiff during another. Packages have been devel-
oped for problems such as these. Although the c22 k1
¼ K
c1 k 2
chemical engineer may not actually calculate
the eigenvalues, knowing that they determine
the stability and accuracy of the numerical Under these conditions the problem becomes
scheme, as well as the step size employed, is
useful. dc1
¼ f ðc1 ; c2 Þ
dt
0 ¼ kl c1 k2 c22
6.5. Differential–Algebraic Systems
Thus, a differential-algebraic system of equa-
Sometimes models involve ordinary differen- tions is obtained. In this case, the second equa-
tial equations subject to some algebraic con- tion can be solved and substituted into the first
straints. For example, the equations governing to obtain differential equations, but in the gen-
one equilibrium stage (as in a distillation eral case that is not possible.
Mathematics in Chemical Engineering 61
satisfy the tolerance with each of these methods Netlib web site, http://www.netlib.org/, choose
can be determined. Then the method and step “ode” to find packages which can be
size for the next step that achieves the biggest downloaded. Using Microsoft Excel to solve
step can be chosen, with appropriate adjustments ordinary differential equations is cumbersome,
due to the different work required for each order. except for the simplest problems.
The package generally starts with a very small
step size and a first-order method — the back-
ward Euler method. Then it integrates along, 6.7. Stability, Bifurcations, Limit
adjusting the order up (and later down) depend- Cycles
ing on the error estimates. The user is thus
assured that the local truncation error meets In this section, bifurcation theory is discussed
the tolerance. A further difficulty arises because in a general way. Some aspects of this subject
the set of nonlinear equations must be solved. involve the solution of nonlinear equations;
Usually a good guess of the solution is available, other aspects involve the integration of ordinary
because the solution is evolving in time and differential equations; applications include
past history can be extrapolated. Thus, the chaos and fractals as well as unusual operation
Newton–Raphson method will usually converge. of some chemical engineering equipment. An
The package protects itself, though, by only excellent introduction to the subject and details
doing a few (i.e., three) iterations. If convergence needed to apply the methods are given in [65].
is not reached within these iterations, the step For more details of the algorithms described
size is reduced and the calculation is redone for below and a concise survey with some chemical
that time step. The convergence theorem for the engineering examples, see [66] and [67]. Bifur-
Newton–Raphson method (Chap. 1) indicates cation results are closely connected with stabil-
that the method will converge if the step size ity of the steady states, which is essentially a
is small enough. Thus, the method is guaranteed transient phenomenon.
to work. Further economies are possible. The Consider the problem
Jacobian needed in the Newton–Raphson
method can be fixed over several time steps. @u
¼ F ðu; lÞ
Then if the iteration does not converge, the @t
Jacobian can be reevaluated at the current
time step. If the iteration still does not converge, The variable u can be a vector, which makes F a
then the step size is reduced and a new Jacobian vector, too. Here, F represents a set of equations
is evaluated. The successive substitution method that can be solved for the steady state:
can also be used — which is even faster, except
that it may not converge. However, it too will F ðu; lÞ ¼ 0
converge if the time step is small enough.
The Runge–Kutta methods give extremely If the Newton–Raphson method is applied,
good accuracy, especially when the step size is
F su d us ¼ F ðus ; lÞ
kept small for stability reasons. If the problem
is stiff, though, backward difference implicit usþ1 ¼ us þ dus
methods must be used. Many chemical reactor
problems are stiff, necessitating the use of is obtained, where
implicit methods. In the MATLAB suite of
ODE solvers, ode45 uses a revision of the @F s
F su ¼ ðu Þ
RKF45 program, while the ode15s program @u
uses an improved backward difference method.
Ref. [64] gives details of the programs in MAT- is the Jacobian. Look at some property of the
LAB. Fortunately, many packages are available. solution, perhaps the value at a certain point or
On the NIST web page, http://gams.nist.gov/ the maximum value or an integral of the solu-
choose “problem decision tree”, and then tion. This property is plotted versus the param-
“differential and integral equations” to find eter l; typical plots are shown in Figure 22. At
packages which can be downloaded. On the the point shown in Figure 22 A, the determinant
Mathematics in Chemical Engineering 63
det F u ¼ 0 sest X ¼ F ss st
ue X
For the limit point, The exponential term can be factored out and
@F ðF ss
u s dÞX ¼ 0
6¼ 0
@l
A solution exists for X if and only if
whereas for the bifurcation-limit point
det jF ss
u s dj ¼ 0
@F
¼0
@l
The s are the eigenvalues of the Jacobian.
The stability of the steady solutions is also of Now clearly if Re (s) > 0 then u0 grows with
interest. Suppose a steady solution uss; the time, and the steady solution uss is said to be
function u is written as the sum of the known unstable to small disturbances. If Im (s) ¼ 0 it
steady state and a perturbation u0 : is called stationary instability, and the distur-
bance would grow monotonically, as indicated
u ¼ uss þ u0 in Figure 23A. If Im (s) 6¼ 0 then the distur-
bance grows in an oscillatory fashion, as shown
This expression is substituted into the original in Figure 23B, and is called oscillatory
equation and linearized about the steady-state instability. The case in which Re (s) ¼ 0 is
value: the dividing point between stability and
instability. If Re (s) ¼ 0 and Im (s) ¼ 0 —
@uss @u0
þ ¼ F ðuss þ u0 ; lÞ the point governing the onset of stationary
@t @t
instability — then s ¼ 0. However, this means
@F
F ðuss ; lÞ þ ju u0 þ that s ¼ 0 is an eigenvalue of the Jacobian, and
@u ss
the determinant of the Jacobian is zero. Thus,
The result is the points at which the determinant of the
Jacobian is zero (for limit points and bifurca-
@u0 0
tion-limit points) are the points governing the
¼ F ss
u u
@t onset of stationary instability. When Re (s) ¼ 0
but Im (s) 6¼ 0, which is the onset of oscillatory
A solution of the form instability, an even number of eigenvalues pass
from the left-hand complex plane to the right-
u0 ðx; tÞ ¼ est XðxÞ hand complex plane. The eigenvalues are
64 Mathematics in Chemical Engineering
complex conjugates of each other (a result of the and apply Newton–Raphson with this initial
original equations being real, with no complex guess and the new value of l. This will be a
numbers), and this is called a Hopf bifurcation. much better guess of the new solution than just
Numerical methods to study Hopf bifurcation u0 by itself.
are very computationally intensive and are not Even this method has difficulties, however.
discussed here [65]. Near a limit point the determinant of the
To return to the problem of solving for the Jacobian may be zero and the Newton method
steady-state solution: near the limit point or may fail. Perhaps no solutions exist at all for the
bifurcation-limit point two solutions exist that chosen parameter l near a limit point. Also, the
are very close to each other. In solving sets of ability to switch from one solution path to
equations with thousands of unknowns, the another at a bifurcation-limit point is necessary.
difficulties in convergence are obvious. For Thus, other methods are needed as well:
some dependent variables the approximation arc-length continuation and pseudo-arc-length
may be converging to one solution, whereas continuation [66]. These are described in
for another set of dependent variables it may be Chapter 1.
converging to the other solution; or the two
solutions may all be mixed up. Thus, solution is
difficult near a bifurcation point, and special 6.8. Sensitivity Analysis
methods are required. These methods are dis-
cussed in [66]. Often, when solving differential equations, the
The first approach is to use natural contin- solution as well as the sensitivity of the solution
uation (also known as Euler–Newton contin- to the value of a parameter must be known. Such
uation). Suppose a solution exists for some information is useful in doing parameter estima-
parameter l. Call the value of the parameter l0 tion (to find the best set of parameters for a
and the corresponding solution u0. Then model) and in deciding whether a parameter
needs to be measured accurately. The differential
F ðu0 ; l0 Þ ¼ 0 equation for y (t, a) where a is a parameter, is
Exchanging the order of differentiation in the that depends upon the location of all the parti-
first term leads to the ordinary differential equa- cles (but not their velocities). Since the major
tion part of the calculation is in the evaluation of the
forces, or potentials, a method must be used that
d @y @ f @y @ f minimizes the number of times the forces are
¼ þ
dt @a @y @a @a calculated to move from one time to another
time. Rewrite this equation in the form of an
The initial conditions on @y/@a are obtained by acceleration.
differentiating the initial conditions
d 2 ri 1
¼ Fi ðfrgÞ ai
dt2 mi
@ @y
½y ð0; aÞ ¼ y0 ; or ð0Þ ¼ 0
@a @a
In the Verlot method, this equation is written
Next, let using central finite differences (Eq. 12). Note
that the accelerations do not depend upon the
@y velocities.
y1 ¼ y; y2 ¼
@a
ri ðt þ DtÞ ¼ 2ri ðtÞ ri ðt DtÞ þ ai ðtÞDt2
and solve the set of ordinary differential equa-
tions The calculations are straightforward, and no
explicit velocity is needed. The storage require-
dy1
¼ f ðy1 ; aÞ y1 ð0Þ ¼ y0 ment is modest, and the precision is modest (it
dt
is a second-order method). Note that one must
dy2 @ f @f
¼ ðy ; aÞ y2 þ y2 ð0Þ ¼ 0 start the calculation with values of {r} at time t
dt @y 1 @a
and tDt.
Thus, the solution y (t, a) and the derivative with In the Velocity Verlot method, an equation is
respect to a are obtained. To project the impact written for the velocity, too.
of a, the solution for a ¼ a1 can be used:
dvi
¼ ai
dt
@y
y ðt; aÞ ¼ y1 ðt; a1 Þ þ ðt; a1 Þða a1 Þ þ
@a
The trapezoid rule (see 2.4.) is applied to obtain
¼ y1 ðt; a1 Þ þ y2 ðt; a1 Þða a1 Þ þ
1
This is a convenient way to determine the vi ðt þ DtÞ ¼ vi ðtÞ þ ½ai ðtÞ þ ai ðt þ DtÞDt
2
sensitivity of the solution to parameters in the
problem. The position of the particles is expanded in a
Taylor series.
d 2 ri
7. Ordinary Differential Equations
mi ¼ Fi ðfrgÞ; Fi ðfrgÞ ¼ rV
dt2 as Boundary Value Problems
where mi is the mass of the i-th particle, ri is the Diffusion problems in one dimension lead to
position of the i-th particle, Fi is the force acting boundary value problems. The boundary con-
on the i-th particle, and V is the potential energy ditions are applied at two different spatial
66 Mathematics in Chemical Engineering
and integrated Dp r 2
hv ¼ þ c1 lnr þ c2
L 4
Zy Zt
dy0
¼ dt ¼ t Now the two unknowns must be specified from
f ðy0 Þ
y0 0 the boundary conditions. This problem is a
two-point boundary value problem because
If the quadrature can be performed analytically, one of the conditions is usually specified at
the exact solution has been found. r ¼ 0 and the other at r ¼ R, the tube radius.
As an example, consider the flow of a non- However, the technique of separating variables
Newtonian fluid in a pipe, as illustrated and integrating works quite well.
in Figure 24. The governing differential When the fluid is non-Newtonian, it
may not be possible to do the second step
analytically. For example, for the Bird–
Carreau fluid [74, p. 171], stress and velocity
are related by
h0
t¼h
dv
2 ið1nÞ=2
1þl dr
Putting this value into the equation for stress The model for a chemical reactor with axial
as a function of r gives diffusion is
h0 Dp r c1 1 dc2 dc
h
dv
2 ið1nÞ=2 ¼ L 2 þ r ¼ DaRðcÞ
1þl Pe dz2 dz
dr
1 dc dc
ð0Þ þ cð0Þ ¼ cin ; ð1Þ ¼ 0
This equation cannot be solved analytically for Pe dz dz
dv/dr, except for special values of n. For prob-
lems such as this, numerical methods must be where Pe is the Peclet number and Da the
used. Damk€ohler number.
The boundary conditions are due to
7.2. Initial Value Methods DANCKWERTS [75] and to WEHNER and WILHELM
[76]. This problem can be treated by using
An initial value method is one that utilizes the initial value methods also, but the method is
techniques for initial value problems but allows highly sensitive to the choice of the parameter s,
for an iterative calculation to satisfy all the as outlined above. Starting at z ¼ 0 and making
boundary conditions. Suppose the nonlinear small changes in s will cause large changes in
boundary value problem the solution at the exit, and the boundary con-
dition at the exit may be impossible to satisfy.
By starting at z ¼ 1, however, and integrating
d2 y dy
¼ f x; y;
dx 2 dx backwards, the process works and an iterative
scheme converges in many cases [77]. How-
with the boundary conditions ever, if the problem is extremely nonlinear the
iterations may not converge. In such cases, the
a0 yð0Þ a1
dy
ð0Þ ¼ a; ai 0 methods for boundary value problems
dx described below must be used.
dy Packages to solve boundary value problems
b0 yð1Þ b1 ð1Þ ¼ b; bi 0
dx
are available on the internet. On the NIST web
page, http://gams.nist.gov/ choose “problem
Convert this second-order equation into two
decision tree”, and then “differential and integral
first-order equations along with the boundary
equations”, then “ordinary differential equa-
conditions written to include a parameter s.
tions”, “multipoint boundary value problems”.
du
On the Netlib web site, http://www.netlib.org/,
¼v search on “boundary value problem”. Any
dx
dv spreadsheet that has an iteration capability can
¼ f ðx; u; vÞ
dx be used with the finite difference method. Some
uð0Þ ¼ a1 s c1 a packages for partial differential equations also
vð0Þ ¼ a0 s c0 a
have a capability for solving one-dimensional
boundary value problems [e.g., Comsol
The parameters c0 and c1 are specified by the Multiphysics (formerly Femlab)].
analyst such that
the transport coefficients at the last value and particularly for chemical reaction engineering.
then solve In the collocation method [3], the dependent
variable is expanded in a series.
Dðckiþ1=2 Þðckþ1 kþ1
iþ1 ci Þ Dðcki1=2 Þðcikþ1 ckþ1
i1 Þ
¼0
Dx2 X
Nþ2
yðxÞ ¼ ai yi ðxÞ ð18Þ
i¼1
The advantage of this approach is that it is
easier to program than a full Newton–Raphson
method. If the transport coefficients do not vary Suppose the differential equation is
radically, the method converges. If the method
does not converge, use of the full Newton– N½y ¼ 0
Raphson method may be necessary.
Three ways are commonly used to evaluate Then the expansion is put into the differential
the transport coefficient at the midpoint. The equation to form the residual:
first one employs the transport coefficient eval-
uated at the average value of the solutions on X
Nþ2
Residual ¼ N½ ai yi ðxÞ
either side: i¼1
1
Dðciþ1=2 Þ D ðciþ1 þ ci Þ
2 In the collocation method, the residual is set to
zero at a set of points called collocation
The second approach uses the average of the points:
transport coefficients on either side:
XN þ2
1 N½ ai yi ðxj Þ ¼ 0; j ¼ 2; . . . ; N þ 1
Dðciþ1=2 Þ ½Dðciþ1 Þ þ Dðci Þ ð17Þ i¼1
2
The orthogonal collocation method has found Furthermore, if Equation (18) is differentiated
widespread application in chemical engineering, once and evaluated at all collocation points, the
Mathematics in Chemical Engineering 71
first derivative can be written in terms of the Note that 1 is the first collocation point (x ¼ 0)
values at the collocation points: and N þ 2 is the last one (x ¼ 1). To apply the
method, the matrices Aij and Bij must be found
dy X
Nþ2
dy and the set of algebraic equations solved, per-
ðxj Þ ¼ ai i ðxj Þ; j ¼ 1; . . . ; N þ 2
dx i¼1
dx haps with the Newton–Raphson method. If
orthogonal polynomials are used and the col-
location points are the roots to one of the
This can be expressed as orthogonal polynomials, the orthogonal collo-
cation method results.
dy XN þ2
dy In the orthogonal collocation method the
ðxj Þ ¼ ½yi ðxk Þ1 yðxk Þ i ðxj Þ; j ¼ 1; . . . ; N þ 2
dx i;k¼1
dx solution is expanded in a series involving
orthogonal polynomials, where the polynomi-
als Pi1(x) are defined in Section 2.2.
or shortened to
X
Nþ2 X
N
dy y ¼ a þ bx þ xð1 xÞ ai Pi1 ðxÞ
ðxj Þ ¼ Ajk yðxk Þ;
dx k¼1 i¼1
ð22Þ
X
N þ2
dyi X
Nþ2
Ajk ¼ ½yi ðxk Þ1 ðxj Þ ¼ bi Pi1 ðxÞ
i¼1
dx i¼1
X
N þ2 y¼ di xi1
d2 y
ðxj Þ ¼ Bjk yðxk Þ; i¼1
dx2 k¼1
X
Nþ2
d2 yi
Bjk ¼ ½yi ðxk Þ1 ðxj Þ The collocation points are shown in Figure 26.
i¼1
dx2
There are N interior points plus one at each end,
This method is next applied to the differ- and the domain is always transformed to lie on
ential equation for reaction in a tubular 0 to 1. To define the matrices Aij and Bij this
reactor, after the equation has been made expression is evaluated at the collocation point-
nondimensional so that the dimensionless s; it is also differentiated and the result is
length is 1.0. evaluated at the collocation points.
X
Nþ2
1 d2 c dc yðxj Þ ¼ di xi1
¼ Da RðcÞ; j
Pe dx2 dx i¼1
ð19Þ
dc dc
ð0Þ ¼ Pe½cð0Þ cin ; ð1Þ ¼ 0 dy X
Nþ2
dx dx ðxj Þ ¼ di ði 1Þxji2
dx i¼1
1 XNþ2 X
Nþ2
Bjk cðxk Þ Ajk cðxk Þ ¼ Da Rðcj Þ ð20Þ
Pe k¼1 k¼1
X
Nþ2
Alk cðxk Þ ¼ Peðc1 cin Þ;
k¼1
ð21Þ
X
N þ2
ANþ2;k cðxk Þ ¼ 0
k¼1 Figure 26. Orthogonal collocation points
72 Mathematics in Chemical Engineering
These formulas are put in matrix notation, found; and once d is known, the solution for
where Q, C, and D are N þ 2 by N þ 2 matrices. any x can be found.
To use the orthogonal collocation method,
y ¼ Qd;
dy d2 y
¼ Cd; 2 ¼ Dd
the matrices are required. They can be calcu-
dx dx lated as shown above for small N (N < 8) and by
Qji ¼ xji1 ; Cji ¼ ði 1Þxji2 ; using more rigorous techniques, for higher N
Dji ¼ ði 1Þði 2Þxi3
(see Chap. 7). However, having the matrices
j
listed explicitly for N ¼ 1 and 2 is useful; this is
In solving the first equation for d, the first and shown in Table 7.
For some reaction diffusion problems, the
second derivatives can be written as
solution can be an even function of x. For
dy
example, for the problem
d ¼ Q1 y; ¼ CQ1 y ¼ Ay;
dx
ð23Þ
d2 y d2 c dc
¼ DQ1 y ¼ By ¼ kc; ð0Þ ¼ 0; cð1Þ ¼ 1 ð24Þ
dx2 dx2 dx
Thus the derivative at any collocation point the solution can be proved to involve only even
can be determined in terms of the solution at powers of x. In such cases, an orthogonal
the collocation points. The same property is collocation method, which takes this feature
enjoyed by the finite difference method (and into account, is convenient. This can easily
the finite element method described below), be done by using expansions that only involve
and this property accounts for some of even powers of x. Thus, the expansion
the popularity of the orthogonal collocation
X
N
method. In applying the method to yðx2 Þ ¼ yð1Þ þ ð1 x2 Þ ai Pi1 ðx2 Þ
Equation (19), the same result is obtained; i¼1
The polynomials are defined to be orthogonal orthogonal collocation is applied at the interior
with the weighting function W(x2). points
Z1 X
N þ1
Bji ci ¼ f2 Rðcj Þ; j ¼ 1; . . . ; N
Wðx2 ÞPk ðx2 ÞPm ðx2 Þxa1 dx ¼ 0 i¼1
ð25Þ
0
k m1
and the boundary condition solved for is
cNþ1 ¼ 1
where the power on xa1 defines the geometry as
planar or Cartesian (a ¼ 1), cylindrical (a ¼ 2),
and spherical (a ¼ 3). An analogous develop- The boundary condition at x ¼ 0 is satisfied
ment is used to obtain the (N þ 1)(N þ 1) automatically by the trial function. After the
matrices solution has been obtained, the effectiveness
factor h is obtained by calculating
X
Nþ1
yðxj Þ ¼ di x2i2 R1 P
Nþ1
j R½cðxÞxa1 dx W j Rðcj Þ
i¼1 0 i¼1
h ¼
R1 P
Nþ1
dy X
N þ1 R½cð1Þxa1 dx W j Rð1Þ
ðxj Þ ¼ di ð2i 2Þx2i3
j 0 i¼1
dx i¼1
X
Nþ1 Note that the effectiveness factor is the average
r2 yðxi Þ ¼ d i r2 ðx2i2 Þjxj reaction rate divided by the reaction rate eval-
i¼1
uated at the external conditions. Error bounds
dy have been given for linear problems [83, p.
y ¼ Qd; ¼ Cd; r2 y ¼ Dd
dx
356]. For planar geometry the error is
Qji ¼ x2i2
j ; Cji ¼ ð2i 2Þx2i3
j ;
w2ð2Nþ1Þ
Dji ¼ r2 ðx2i2 Þjxj Error in h ¼
ð2N þ 1Þ!ð2N þ 2Þ!
dy
d ¼ Q1 y; ¼ CQ1 y ¼ Ay; This method is very accurate for small N (and
dx
small f2); note that for finite difference meth-
r2 y ¼ DQ1 y ¼ By
ods the error goes as 1/N2, which does not
decrease as rapidly with N. If the solution is
In addition, the quadrature formula is desired at the center (a frequent situation
because the center concentration can be the
WQ ¼ f; W ¼ fQ1
most extreme one), it is given by
X
N þ1
cð0Þ ¼ d1 ½Q1 1i yi
where i¼1
Z1 X
Nþ1 The collocation points are listed in Table 8.
x2i2 xa1 dx ¼ W j x2i2
j For small N the results are usually more
j¼1
0
accurate when the weighting function in
¼
1
fi Equation (25) is 1 x2. The matrices for
2i 2 þ a
N ¼ 1 and N ¼ 2 are given in Table 9 for
the three geometries. Computer programs to
As an example, for the problem generate matrices and a program to solve
reaction diffusion problems, OCRXN, are
available [3, p. 325, p. 331].
1 d dc
xa1 ¼ f2 RðcÞ
xa1 dx dx Orthogonal collocation can be applied to
dc distillation problems. STEWART et al. [84, 85]
ð0Þ ¼ 0; cð1Þ ¼ 1
dx developed a method using Hahn polynomials
74 Mathematics in Chemical Engineering
Table 8. Collocation points for orthogonal collocation with The boundary conditions are typically
symmetric polynomials and W ¼ 1
dc dc
Geometry ð0Þ ¼ 0; ð1Þ ¼ Bim ½cð1Þ cB
dx dx
N Planar Cylindrical Spherical
1 0.5773502692 0.7071067812 0.7745966692 where Bim is the Biot number for mass transfer.
2 0.3399810436 0.4597008434 0.5384693101 These become
0.8611363116 0.8880738340 0.9061793459
3 0.2386191861 0.3357106870 0.4058451514 1 dc
0.6612093865 0.7071067812 0.7415311856 ðu ¼ 0Þ ¼ 0;
Dx1 du
0.9324695142 0.9419651451 0.9491079123
4 0.1834346425 0.2634992300 0.3242534234
0.5255324099 0.5744645143 0.6133714327 in the first element;
0.7966664774 0.8185294874 0.8360311073
0.9602898565 0.9646596062 0.9681602395
5 0.1488743390 0.2165873427 0.2695431560 1 dc
ðu ¼ 1Þ ¼ Bim ½cðu ¼ 1Þ cB ;
0.4333953941 0.4803804169 0.5190961292 DxNE du
0.6794095683 0.7071067812 0.7301520056
0.8650633667 0.8770602346 0.8870625998
0.9739065285 0.9762632447 0.9782286581 in the last element. The orthogonal collocation
method is applied at each interior collocation
point.
1 XNP
The reaction–diffusion equation is written as Dxk J¼1
A1;J cJ ¼ 0;
1 d dc d2 c a 1 dc
xa1 ¼ 2þ ¼ f2 RðcÞ in the first element;
xa1 dx dx dx x dx
Table 9. Matrices for orthogonal collocation with symmetric polynomials and W¼1x2
76 Mathematics in Chemical Engineering
These equations can be assembled into an the unknown solution is expanded in a series of
overall matrix problem known functions {bi(x)}, with unknown coef-
ficients {ai}.
A Ac ¼ f
X
NT
cðxÞ ¼ ai bi ðxÞ
The form of these equations is special and is i¼1
The error at the collocation points is more This process makes the method a Galerkin
accurate, giving what is known as superconver- method. The basis for the orthogonality condi-
gence. tion is that a function that is made orthogonal to
i
d each member of a complete set is then zero. The
ðy yexact Þ constantjDxj4
dxi
collocation points residual is being made orthogonal, and if the
basis functions are complete, and an infinite
number of them are used, then the residual is
7.6. Galerkin Finite Element Method zero. Once the residual is zero the problem is
solved. It is necessary also to allow for the
In the finite element method the domain is boundary conditions. This is done by integrat-
divided into elements and an expansion is ing the first term of Equation (26) by parts and
made for the solution on each finite element. then inserting the boundary conditions:
In the Galerkin finite element method an addi- Z1
1 d a1 dbi
tional idea is introduced: the Galerkin method bj ðxÞ
xa1 dx
x
dx
xa1 dx ¼
is used to solve the equation. The Galerkin 0
Combining this with Equation (26) gives in the e-th element. Then the Galerkin method
is
NTZ
1
X dbj dbi a1
N PZ
1
x dxai X 1 X dN J dN I
i¼1
dx dx ðxe þ uDxe Þa1 duceI
0
e
Dxe I¼1 du du
0
" #
XNT " #
Bim bj ð1Þ ai bi ð1Þ cB X X
NP
Bim N J ð1Þ ceI N I ð1Þ c1
i¼1 ð28Þ
e I¼1 ð29Þ
Z1 " # " #
XNT
X Z1 XNP
¼ f2 bj ðxÞ ai bi ðxÞ xa1 dx ¼ f2 Dxe N J ðuÞR ceI N I ðuÞ
i¼1 e I¼1
0 0
This equation defines the Galerkin method, The element integrals are defined as
and a solution that satisfies this equation (for
all j ¼ 1, . . . , 1) is called a weak solution. Z1
For an approximate solution the equation is 1 dN J dN I
BeJI ¼ ðxe þ uDxe Þa1 du;
Dxe du du
written once for each member of the trial 0
a1
cð1Þ ¼ cB ðxe þ uDxe Þ du
then the boundary condition is used (instead of whereas the boundary element integrals are
Eq. 29) for j ¼ NT,
B BeJI ¼ Bim N J ð1ÞN I ð1Þ;
X
NT
F F eJ ¼ Bim N J ð1Þc1
ai bi ð1Þ ¼ cB
i¼1
a1
CHRISTIANSEN [94]. For a second-order differen-
ðxe þ uk Dxe Þ tial equation and cubic trial functions on finite
elements, the error in the i-th element is given by
For an application of the finite element method
k Error ki ¼ cDx4i k uð4Þ ki
in fluid mechanics, see ! Fluid Mechanics.
Because cubic elements do not have a nonzero
7.7. Cubic B-Splines fourth derivative, the third derivative in adjacent
elements is used [3, p. 166]:
Cubic B-splines have cubic approximations
1 d 3 ci 1 d3 ciþ1
within each element, but first and second deriv- ai ¼ ; aiþ1 ¼ 3
Dx3i du3 Dxiþ1 du3
atives continuous between elements. The func- 2 3
tions are the same ones discussed in Chapter 2, 1 6 ai ai1 aiþ1 ai 7
k uð4Þ ki 4 þ 5
and they can be used to solve differential 2 1 1
ðxiþ1 xi1 Þ ðxiþ2 xi Þ
equations [92]. 2 2
Mathematics in Chemical Engineering 79
method of solving such problems is to refine the The Blasius equation becomes
mesh near the singularity, by relying on the
better approximation due to a smaller Dx. d3 f d2 f
2 z3 3 3z2 2 z
df
Another approach is to incorporate the singular dz dz dz
trial function into the approximation. Thus, if
d2 df
þf z2 2 þ z ¼ 0 for 0 z 1:
the solution approaches f(x) as x goes to zero, dz dz
and f(x) becomes infinite, an approximation
may be taken as The differential equation now has variable
coefficients, but these are no more difficult to
X
N
handle than the original nonlinearities.
yðxÞ ¼ f ðxÞ þ ai yi ðxÞ
i¼1 Another approach is to use a variable mesh,
perhaps with the same transformation. For
This function is substituted into the differential example, use z ¼ eh and a constant mesh
equation, which is solved for ai. Essentially, a size in z. Then with 101 points distributed
new differential equation is being solved for a uniformly from z ¼ 0 to z ¼ 1, the following
new variable: are the nodal points:
d3 d2
2 þf 2 ¼0
dh3 dh
8. Partial Differential Equations
df
f ¼ ¼ 0 at h ¼ 0
dh
Partial differential equations are differential
df
¼ 1 at h ! 1
equations in which the dependent variable is
dh a function of two or more independent varia-
bles. These can be time and one space dimen-
Because one boundary is at infinity using a sion, or time and two or more space dimensions,
mesh with a constant size is difficult! One or two or more space dimensions alone. Prob-
approach is to transform the domain. For exam- lems involving time are generally either hyper-
ple, let bolic or parabolic, whereas those involving
spatial dimensions only are often elliptic.
z ¼ eh Because the methods applied to each type of
equation are very different, the equation must
first be classified as to its type. Then the special
Then h ¼ 0 becomes z ¼ 1 and h ¼ 1 becomes methods applicable to each type of equation are
z ¼ 0. The derivatives are described. For a discussion of all methods, see
[100–103]; for a discussion oriented more
dz d2 z toward chemical engineering applications, see
¼ eh ¼ z; ¼ eh ¼ z
dh dh2
[104]. Examples of hyperbolic and parabolic
df df dz df equations include chemical reactors with radial
¼ ¼ z
dh dz dh dz
dispersion, pressure-swing adsorption, disper-
2 2
2
d f
¼
d f dz
þ
df d2 z d2 f
¼ z2 2 þ z
df sion of an effluent, and boundary value prob-
dh2 dz2 dh dz dh2 dz dz lems with transient terms added (heat transfer,
Mathematics in Chemical Engineering 81
mass transfer, chemical reaction). Examples of type. The surface is defined by this equation
elliptic problems include heat transfer and mass plus a normalization condition:
transfer in two and three spatial dimensions and
steady fluid flow. X
n
s 2k ¼ 1
k¼0
Thus,
where
s 21 þ s 22 ¼ 1 ðnormalizationÞ
X
n
a ¼ ða0 ; a1 ; . . . ; an Þ; jaj ¼ ai s 21 þ s 22 ¼ 0 ðequationÞ
i¼0
@ jaj
@a ¼
@ta0 @xa1 1 . . . @xan n These cannot both be satisfied so the problem is
elliptic. When the equation is
the characteristic equation for P is defined as
X @2 u @2 u
¼0
aa s a ¼ 0; s ¼ ðs 0 ; s 1 ; . . . ; s n Þ @t2 @x2
jaj¼m ð30Þ
s a ¼ s a0 0 s a1 1 . . . s an n
then
where s represents coordinates. Thus only the
highest derivatives are used to determine the j20 þ j21 ¼ 0
82 Mathematics in Chemical Engineering
Now real j0 can be solved and the equation is First-order quasi-linear problems are written
hyperbolic in the form
s 20 þ s 21 ¼ 1 ðnormalizationÞ Xn
@u
A1 ¼ f; x ¼ ðt; x1 . . . ; xn Þ
s 20 þ s 21 ¼ 0 ðequationÞ l¼0
@x1 ð31Þ
u ¼ ðu1 ; u2 ; . . . ; uk Þ
When the equation is
2
@c @ c @2c The matrix entries A1 is a kk matrix whose
¼D þ 2
@t @x 2 @y entries depend on u but not on derivatives of u.
Equation (31) is hyperbolic if
then
A ¼ Am
s 20 þ s 21 þ s 22 ¼ 1 ðnormalizationÞ
bs 20 þ s 21 ¼ 0 n ¼ 1; A0 ¼ 1; A1 ¼ FðuÞ
is required. Combining these gives Thus, the roots are real and the equation is
hyperbolic.
1 ð1 þ bÞs 20 ¼ 0 The final example is the heat conduction
problem written as
The roots are real and the equation is hyper-
bolic. When b ¼ 0 %C p
@T @q
¼ ; q ¼ k
@T
@t @x @x
j21 ¼ 0
In this formulation the constitutive equation for
and the equation is parabolic. heat flux is separated out; the resulting set of
Mathematics in Chemical Engineering 83
and for Equation (35): The concentration profile is steeper for the
dci ci ci1 df dci MacCormack method than for the upstream
f þ fui þ ð1 fÞ ji ¼0 derivatives, but oscillations can still be present.
dt Dx dc dt
The flux-corrected transport method can be
If the flow were from right to left, then the added to the MacCormack method. A solution
formula would be is obtained both with the upstream algorithm
dci Fðciþ1 Þ Fðci Þ ciþ1 2ci þ ci1
and the MacCormack method; then they are
þ ¼D combined to add just enough diffusion to elim-
dt Dx Dx2
inate the oscillations without smoothing the
If the flow could be in either direction, a local solution too much. The algorithm is compli-
determination must be made at each node i and cated and lengthy but well worth the effort
the appropriate formula used. The effect of [107–109].
using upstream derivatives is to add artificial If finite element methods are used, an
or numerical diffusion to the model. This can be explicit Taylor–Galerkin method is appropriate.
ascertained by taking the finite difference form For the convective diffusion equation the
of the convective diffusion equation method is
dci ci ci1 ciþ1 2ci þ ci1 1 nþ1 2 1
þu ¼D ðc cniþ1 Þ þ ðcnþ1 cni Þ þ ðcnþ1 cni1 Þ
dt Dx Dx2 6 iþ1 3 i 6 i1
uDt n DtD u2 Dt2
and rearranging ¼ ðciþ1 cni1 Þ þ þ ðcniþ1 2cni þ cni1 Þ
2Dx Dx2 2Dx2
ac
f ðcÞ ¼
1 þ Kc
Otherwise, if x is changed one side changes, The boundary conditions for u are
but the other cannot because it depends on t.
Call the constant l and write the separate uð0; tÞ ¼ 0
equations uðL; tÞ ¼ 0
TðtÞ ¼ Tð0ÞelDt
Separation of variables is now applied to this
equation by writing
and the second equation is written in the form
uðx; tÞ ¼ TðtÞXðxÞ
d2 X
þ lX ¼ 0
dx2
The same equation for T (t) and X (x) is
Next consider the boundary conditions. If obtained, but with X (0) ¼ X (L) ¼ 0.
they are written as
d2 X
þ lX ¼ 0
dx2
cðL; tÞ ¼ 1 ¼ TðtÞXðLÞ
Xð0Þ ¼ XðLÞ ¼ 0
cð0; tÞ ¼ 0 ¼ TðtÞXð0Þ
the boundary conditions are difficult to satisfy Next X (x) is solved for. The equation
because they are not homogeneous, i.e. with a is an eigenvalue problem. The general solution
zero right-hand side. Thus, the problem must is obtained by using emx pffiffiffi and finding that
be transformed to make the boundary condi- m2þ l ¼ 0; thus m ¼
i l. The exponential
tions homogeneous. The solution is written as term
the sum of two functions, one of which pffiffi
e
i lx
satisfies the nonhomogeneous boundary con-
ditions, whereas the other satisfies the homo-
geneous boundary conditions. is written in terms of sines and cosines, so that
the general solution is
cðx; tÞ ¼ f ðxÞ þ uðx; tÞ
pffiffiffi pffiffiffi
uð0; tÞ ¼ 0 X ¼ Bcos lx þ Esin lx
uðL; tÞ ¼ 0
The boundary conditions are
Thus, f (0) ¼ 1 and f (L) ¼ 0 are necessary. Now pffiffiffi pffiffiffi
the combined function satisfies the boundary XðLÞ ¼ Bcos lL þ Esin lL ¼ 0
conditions. In this case the function f (x) can be Xð0Þ ¼ B ¼ 0
taken as
If B ¼ 0, then E 6¼ 0 is required to have any
f ðxÞ ¼ L x solution at all. Thus, l must satisfy
pffiffiffi
The equation for u is found by substituting for c sin lL ¼ 0
in the original equation and noting that the f (x)
drops out for this case; it need not disappear in This is true for certain values of l, called
the general case: eigenvalues or characteristic values. Here,
they are
@u @2u
¼D 2
@t @x ln ¼ n2 p2 =L2
Mathematics in Chemical Engineering 87
Each eigenvalue has a corresponding eigen- number of terms are used, some error always
function occurs. For large times a single term is ade-
quate, whereas for small times many terms are
X n ðxÞ ¼ Esin n p x=L needed. For small times the Laplace transform
method is also useful, because it leads to
The composite solution is then solutions that converge with fewer terms. For
small times, the method of combination of
n p x ln Dt
X n ðxÞT n ðtÞ ¼ E A sin e variables may be used as well. For nonlinear
L
problems, the method of separation of variables
This function satisfies the boundary conditions fails and one of the other methods must be
and differential equation but not the initial used.
condition. To make the function satisfy the The method of combination of variables is
initial condition, several of these solutions useful, particularly when the problem is posed
are added up, each with a different eigenfunc- in a semi-infinite domain. Here, only one exam-
tion, and E A is replaced by An. ple is provided; more detail is given in [3, 113,
114]. The method is applied here to the non-
X
1
n p x n2 p2 Dt=L2
linear problem
uðx; tÞ ¼ An sin e
L
n¼1
@c @ @c @ 2 c d DðcÞ @c 2
¼ DðcÞ ¼ DðcÞ 2 þ
@t @x @x @x dc @x
The constants An are chosen by making u (x, t)
satisfy the initial condition.
with boundary and initial conditions
X
1
npx x
uðx; 0Þ ¼ An sin ¼ 1 cðx; 0Þ ¼ 0
n¼1
L L
cð0; tÞ ¼ 1; cð1; tÞ ¼ 0
Next, the Galerkin method is applied, and the The use of the 4 and D0 makes the analysis below
residual is made orthogonal to a complete set of simpler. The equation for c (x, t) is transformed
functions, which are the eigenfunctions. into an equation for f (h)
@c df @h @c df @h
ZL ¼ ; ¼
x mpx @t dh @t @x dh @x
1 sin dx
L L @ 2 c d2 f @h 2 df @ 2 h
0 ¼ þ
@x2 dh2 @x dh @x2
X1 Z L
mpx npx Am
¼ An sin sin dx ¼ @h x=2 @h 1 @2 h
L L 2 ¼ pffiffiffiffiffiffiffiffiffiffiffiffi ; ¼ pffiffiffiffiffiffiffiffiffiffi ; ¼0
n¼1
0 @t 4D0 t3 @x 4D0 t @x2
points. Thus, the boundary conditions can be where the matrix B is tridiagonal. The stability
combined to give of the integration of these equations is governed
by the largest eigenvalue of B. If Euler’s
f ð1Þ ¼ 0 method is used for integration,
D 2
The other boundary condition is for x ¼ 0 or Dt
Dx2 jljmax
h ¼ 0,
whereas the oscillation limit is given by boundary value problems by letting the solution
DDt 0:25
at the nodes depend on time. For the diffusion
equation the finite element method gives
Dx2 1 u
ADxðcnþ1
i cni Þ ¼ DtAðJ j1=2 J iþ1=2 Þ Now the matrix C C is not diagonal, so that a set
of equations must be solved for each time step,
where J is the flux due to convection and even when the right-hand side is evaluated
diffusion, positive in the þx direction. explicitly. This is not as time-consuming as it
seems, however. The explicit scheme is written
@c ci ci1=2
J ¼ uc D ; J i1=2 ¼ ui1=2 ci1=2 D as
@x Dx
cnþ1 cni
The concentration at the edge of the cell is CC ji i
¼ AAji cni
Dt
taken as
1 and rearranged to give
ci1=2 ¼ ðci þ ci1 Þ
2
CC ji ðcnþ1
i cni Þ ¼ DtAAji cni or
Rearrangement for the case when the veloc-
CCðcnþ1 cn Þ ¼ DtAAc
ity u is the same for all nodes gives
cnþ1
i cni uðciþ1 ci1 Þ D This is solved with an L U decomposition (see
þ ¼ 2 ðciþ1 2ci þ ci1 Þ
Dt 2Dx Dx Section 1.2) that retains the structure of the mass
matrix C C. Thus,
This is the same equation as obtained using
the finite difference method. This is not always CC ¼ LU
true, and the finite volume equations are easy to
derive. In two- and three-dimensions, the mesh At each step, calculate
need not be rectangular, as long as it is possible
to compute the velocity normal to an edge of the
cnþ1 cn ¼ DtU1 L1 AAcn
cell. The finite volume method is useful for
applications involving filling, such as injection
molding, when only part of the cell is filled with This is quick and easy to do because the inverse
fluid. Such applications do involve some of L and U are simple. Thus the problem is
approximations, since the interface is not reduced to solving one full matrix problem and
tracked precisely, but they are useful engineer- then evaluating the solution for multiple right-
ing approximations. hand sides. For implicit methods the same
The finite element method is handled in a approach can be used, and the LU decomposi-
similar fashion, as an extension of two-point tion remains fixed until the time step is changed.
90 Mathematics in Chemical Engineering
The method of orthogonal collocation uses and evaluated at each collocation point
a similar extension: the same polynomial of x
n
is used but now the coefficients depend on unþ1 unj @u
þ f ðunj Þ ¼ 0
j
time. Dt @x j
The prototype elliptic problem is steady- And this is converted to the Gauss–Seidel iter-
state heat conduction or diffusion, ative method.
2
@ T @2T Dx2
k þ 2 ¼Q 2 1 þ 2 T sþ1 s sþ1
i;j ¼ T iþ1;j þ T i1;j
@x 2 @y Dy
Dx2 s 2 Qi;j
þ ðT i;j1 Þ Dx
þ T sþ1
Dy2 i;jþ1 k
possibly with a heat generation term per unit
volume, Q. The boundary conditions can be Calculations proceed by setting a low i, com-
puting from low to high j, then increasing i and
Dirichlet or 1st kind: T ¼ T1 on boundary S1 repeating the procedure. The relaxation method
uses
Neumann or 2nd kind: k @T
@n ¼ q2 on boundary S2
k @T Dx2 Dx2
Robin, mixed, or 3rd kind: @n ¼ hðT T 3 Þ on boundary S3 2 1 þ 2 T i;j ¼ T siþ1;j þ T sþ1
i1;j þ
Dy Dy2
Qi;j
i;j1 Þ Dx
ðT si;jþ1 T sþ1 2
Illustrations are given for constant physical k
properties k, h, while T1, q2, T3 are known
functions on the boundary and Q is a known
function of position. For clarity, only a two- T sþ1 s
s
i;j ¼ T i;j þ bðT i;j T i;j Þ
dimensional problem is illustrated. The finite
difference formulation is given by using the If b ¼ 1, this is the Gauss–Seidel method. If
following nomenclature b > 1, it is overrelaxation; if b < 1, it is
underrelaxation. The value of b may be chosen
T i;j ¼ TðiDx; jDyÞ empirically, 0 < b < 2, but it can be selected
theoretically for simple problems like this [117,
The finite difference formulation is then p. 100], [3, p. 282]. In particular, the optimal
value of the iteration parameter is given by
T iþ1;j 2T i;j þ T i1;j T i;jþ1 2T i;j þ T i;j1
þ ¼ Qi;j =k ð40Þ
Dx2 Dy2 lnðbopt 1Þ R
If the boundary is parallel to a coordinate axis (when there are n points in both x and y direc-
the boundary slope is evaluated as in Chapter 7, tions). For Neumann boundary conditions, the
by using either a one-sided, centered difference value is
or a false boundary. If the boundary is more
irregular and not parallel to a coordinate line,
p2 1
more complicated expressions are needed and R¼
2n2 1 þ max½Dx2 =Dy2 ; Dy2 =Dx2
the finite element method may be the better
method.
Equation (40) is rewritten in the form Iterative methods can also be based on lines
(for 2D problems) or planes (for 3D problems).
Dx2 Dx2 Preconditioned conjugate gradient methods
2 1 þ 2 T i;j ¼ T iþ1;j þ T i1;j þ 2
Dy Dy have been developed (see Chap. 1). In these
Qi;j methods a series of matrix multiplications are
ðT i;jþ1 þ T i;j1 Þ Dx2
k done iteration by iteration; and the steps lend
92 Mathematics in Chemical Engineering
themselves to the efficiency available in paral- In these equations I and J refer to the nodes
lel computers. In the multigrid method the of the triangle forming element e and the
problem is solved on several grids, each summation is made over all elements. These
more refined than the previous one. In iterating equations represent a large set of linear equa-
between the solutions on different grids, one tions, which are solved using matrix techniques
converges to the solution of the algebraic equa- (Chap. 1).
tions. A chemical engineering application is If the problem is nonlinear, e.g., with k or Q a
given in [118]. Software for a variety of these function of temperature, the equations must be
methods is available, as described below. solved iteratively. The integrals are given for a
The Galerkin finite element method triangle with nodes I, J, and K in counter-
(FEM) is useful for solving elliptic problems clockwise order. Within an element,
and is particularly effective when the domain
or geometry is irregular [119–125]. As an T ¼ N I ðx; yÞT I þ N J ðx; yÞT J þ N K ðx; yÞT K
example, cover the domain with triangles
aI þ bI x þ cI y
and define a trial function on each triangle. NI ¼
2D
The trial function takes the value 1.0 at one aI ¼ xJ yK xK yJ
corner and 0.0 at the other corners, and is
bI ¼ yI yK
linear in between (see Fig. 30). These trial
functions on each triangle are pieced together cI ¼ xK x J
to give a trial function on the whole domain. plus permutation on I; K; J
For the heat conduction problem the method 2
1 xI yI
3
gives [3] 6 7
2D ¼ det6 7
4 1 xJ yJ 5 ¼ 2ðarea of triangleÞ
XX XX
AeIJ T eJ ¼ F eI ð41Þ 1 xK yK
e J e J
aI þ aJ þ aK ¼ 1
where bI þ bJ þ bK ¼ 0
Z Z cI þ c J þ cK ¼ 0
AeIJ ¼ krN I rN J dA h3 N I N J dC
k
C3 AeIJ ¼ ðbI bJ þ cI cJ Þ
4D
Z Z Z
F eI ¼ N I QdA þ N I q2 dC Q QD
N I h3 T 3 dC F eIJ ¼ ðaI þ bI x þ cI y Þ ¼
2 3
C2 C3
xI þ xJ þ x K y þ y J þ yK
x¼ ; y ¼ I
3 3
Also, a necessary condition is that
2
T i ¼ T 1 on C 1 aI þ bI x þ cI y ¼ D
3
element and the basis functions. In two dimen- In the finite difference method an explicit
sions, triangular elements are usually used technique would evaluate the right-hand side at
because it is easy to cover complicated geome- the n-th time level:
tries and refine parts of the mesh. However,
rectangular elements sometimes have advan- T nþ1 n
i;j T i;j
rC p
tages, particularly when some parts of the Dt
solution do not change very much and the ¼
k
ðT n 2T ni1;j þ T ni1;j Þ
elements can be long. In three dimensions Dx2 iþ1;j
the same considerations apply: tetrahedral ele- k
þ ðT n 2T ni;j þ T ni;j1 Þ Q
Dy2 i;jþ1
ments are frequently used, but brick elements
are also possible. While linear elements (in both
two and three dimensions) are usually used, When Q ¼ 0 and Dx ¼ Dy, the time step is
higher accuracy can be obtained by using qua- limited by
dratic or cubic basis functions within the ele-
Dx2 %C p Dx2
ment. The reason is that all methods converge Dt or
4k 4D
according to the mesh size to some power,
and the power is larger when higher order These time steps are smaller than for one-
elements are used. If the solution is dis- dimensional problems. For three dimensions,
continuous, or has discontinuous first deriva- the limit is
tives, then the lowest order basis functions are
used because the convergence is limited by the Dx2
Dt
properties of the solution, not the finite element 6D
approximation.
One nice feature of the finite element To avoid such small time steps, which must
method is the use of natural boundary condi- be smaller when Dx decreases, an implicit
tions. In this problem the natural boundary method could be used. This leads to large
conditions are the Neumann or Robin condi- sparse matrices, rather than convenient tri-
tions. When using Equation (41), the problem diagonal matrices.
can be solved on a domain that is shorter than
needed to reach some limiting condition (such
as at an outflow boundary). The externally
8.6. Special Methods for Fluid
applied flux is still applied at the shorter Mechanics
domain, and the solution inside the truncated
domain is still valid. Examples are given in The method of operator splitting is also
[107] and [131]. The effect of this is to allow useful when different terms in the equation
solutions in domains that are smaller, thus are best evaluated by using different methods
saving computation time and permitting the or as a technique for reducing a larger problem
solution in semi-infinite domains. to a series of smaller problems. Here the
method is illustrated by using the Navier–
Stokes equations. In vector notation the equa-
8.5. Parabolic Equations in Two or tions are
Three Dimensions
@u
% þ %u ru ¼ %f rp þ mr2 u
Computations become much more lengthy with @t
This can be done by using the finite difference here follows [142]. A lattice is defined, and one
[132, p. 162] or the finite element method [133– solves the following equation for f i ðx; tÞ, the
135]. probability of finding a molecule at the point x
In fluid flow problems solved with the with speed ci.
finite element method, the basis functions
for pressure and velocity are often different. @f i f f eq
þ ci rf i ¼ i i
This is required by the LBB condition (named @t t
after LADYSHENSKAYA, BREZZI, and BABUSKA)
[134, 135]. Sometimes a discontinuous basis The right-hand side represents a single time
function is used for pressure to meet this relaxation for molecular collisions, and t is
condition. Other times a penalty term is added, related to the kinematic viscosity. By means
or the quadrature is done using a small number of a simple stepping algorithm, the computa-
of quadrature points. Thus, one has to be tional algorithm is
careful how to apply the finite element method
to the Navier–Stokes equations. Fortunately, Dt
f i ðx þ ci Dt; t þ DtÞ ¼ f i ðx; tÞ ðf f eq
i Þ
software exists that has taken this into t i
account.
Consider the lattice shown in Figure 34. The
Level Set Methods. Multiphase problems are various velocities are
complicated because the terms in the
equations depend on which phase exists at a c1 ¼ ð1; 0Þ; c3 ¼ ð0; 1Þ; c5 ¼ ð1; 0Þ; c7 ¼ ð0; 1Þ
particular point, and the phase boundary may c2 ¼ ð1; 1Þ; c4 ¼ ð1; 1Þ; c6 ¼ ð1; 1Þ; c8 ¼ ð1; 1Þ
move or be unknown. It is desirable to compute
on a fixed grid, and the level set formulation c0 ¼ ð0; 0Þ
The physics governing the velocity of the where the weighting functions are
interface must be defined, and this equation is 4 1 1
solved along with the other equations repre- w0 ¼ ; w1 ¼ w3 ¼ w5 ¼ w7 ¼ ; w2 ¼ w4 ¼ w6 ¼ w8 ¼
9 9 36
senting the problem [130–137].
With these conditions, the solution for velocity
Lattice Boltzmann Methods. Another way to is a solution of the Navier—Stokes equation.
solve fluid flow problems is based on a These equations lead to a large computational
molecular viewpoint and is called the Lattice problem, but it can be solved by parallel proc-
Boltzmann method [138–141]. The treatment essing on multiple computers.
Mathematics in Chemical Engineering 95
then include approximations and correlations for evolutionary problems, whereas Fredholm
some features that would be difficult to solve for equations are analogous to boundary value
directly. Four widely used major packages are problems. The terms in the integral can be
Fluent (http://www.fluent.com/), CFX (now unbounded, but still yield bounded integrals,
part of ANSYS), Comsol Multiphysics and these equations are said to be weakly
(formerly FEMLAB) (http://www.comsol.com/), singular. A Volterra equation of the second
and ANSYS (http://www.ansys.com/). Of these, kind is
Comsol Multiphysics is particularly useful
because it has a convenient graphical user Zt
interface, permits easy mesh generation and re- yðtÞ ¼ gðtÞ þ l Kðt; sÞyðsÞds ð42Þ
a
finement (including adaptive mesh refinement),
allows the user to add in phenomena and
additional equations easily, permits solution whereas a Volterra equation of the first kind is
by continuation methods (thus enhancing
convergence), and has extensive graphical Zt
output capabilities. Other packages are also yðtÞ ¼ l Kðt; sÞyðsÞds
available (see http://cfd-online.com/), and a
@T
¼ GðT; tÞ; x ¼ 0; t > 0
@x
9.1. Classification
The solution to this problem is
Volterra integral equations have an integral with
a variable limit, whereas Fredholm integral Zt 1
1
equations have a fixed limit. Volterra equations G Tð0; sÞ; s pffiffiffiffiffiffiffiffiffiffi ex =4ðtsÞ ds
2
Tðx; tÞ ¼ pffiffiffi
p ts
are usually associated with initial value or 0
Mathematics in Chemical Engineering 97
If T (t) is used to represent T (0, t), then problems. An eigenvalue problem is a homoge-
neous equation of the second kind.
Zt 1
1
T ðtÞ ¼ pffiffiffi G TðsÞ; s pffiffiffiffiffiffiffiffiffiffi ds Zb
p ts
0 yðxÞ ¼ l Kðx; sÞyðsÞds ð44Þ
a
Thus the behavior of the solution at the bound-
ary is governed by an integral equation. NAGEL Solutions to this problem occur only for spe-
and KLUGE [156] use a similar approach to solve cific values of l, the eigenvalues. Usually the
for adsorption in a porous catalyst. Fredholm equation of the second or first kind is
The existence and uniqueness of the solution solved for values of l different from these,
can be proved [151, p. 30, 32]. which are called regular values.
Sometimes the kernel is of the form Nonlinear Volterra equations arise naturally
from initial value problems. For the initial value
Kðt; sÞ ¼ Kðt sÞ problem
dy
Equations of this form are called convolution ¼ F t; yðtÞ
dt
equations and can be solved by taking the
Laplace transform. For the integral equation
both sides can be integrated from 0 to t to obtain
Zt
YðtÞ ¼ GðtÞ þ l Kðt tÞYðtÞdt Zt
0 yðtÞ ¼ yð0Þ þ F s; yðsÞ ds
Zt 0
Solving this for y(s) gives Theorem [151, p. 55]. If g (t) is continuous, the
kernel K (t, s, y) is continuous in all variables
yðsÞ ¼
gðsÞ and satisfies a Lipschitz condition
1 kðsÞ
Zt
whereas a Fredholm equation of the first kind is ynþ1 ðtÞ ¼ gðtÞ þ K½t; s; yn ðsÞds
0
Zb
Kðx; sÞyðsÞds ¼ gðxÞ
Nonlinear Fredholm equations have special
a
names. The equation
Z1
The limits of integration are fixed, and these f ðxÞ ¼ K½x; y; f ðyÞdy
problems are analogous to boundary value 0
98 Mathematics in Chemical Engineering
is called the Urysohn equation [150, p. 208]. solve for y1, y2, . . . in succession. For a single
The special equation integral equation, at each step one must solve a
Z1
single nonlinear algebraic equation for yn. Typ-
f ðxÞ ¼ K½x; yF½y; f ðyÞdy ically, the error in the solution to the integral
0 equation is proportional to Dtm, and the power
m is the same as the power in the quadrature
is called the Hammerstein equation [150, p. error [151, p. 97].
209]. Iterative methods can be used to solve The stability of the method [151, p. 111] can
these equations, and these methods are closely be examined by considering the equation
tied to fixed point problems. A fixed point
problem is Zt
yðtÞ ¼ 1 l yðsÞds
x ¼ FðxÞ 0
Local convergence theorems prove the process Since the integral equation can be differentiated
convergent if the solution is close enough to the to obtain the initial value problem
answer, whereas global convergence theorems
are valid for any initial guess [150, p. 229–231]. dy
¼ ly; yð0Þ ¼ 1
The successive substitution method for non- dt
linear Fredholm equations is
the stability results are identical to those for
Z1 initial value methods. In particular, using the
ynþ1 ðxÞ ¼ K½x; s; yn ðsÞds trapezoid rule for integral equations is identical
0 to using this rule for initial value problems. The
method is A-stable.
Typical conditions for convergence include that Higher order integration methods can also
the function satisfies a Lipschitz condition. be used [151, p. 114, 124]. When the kernel is
infinite at certain points, i.e., when the problem
has a weak singularity, see [151, p. 71, 151].
9.2. Numerical Methods for Volterra
Equations of the Second Kind
Volterra equations of the second kind are ana- 9.3. Numerical Methods for Fredholm,
logous to initial value problems. An initial Urysohn, and Hammerstein Equations
value problem can be written as a Volterra of the Second Kind
equation of the second kind, although not all
Volterra equations can be written as initial value Whereas Volterra equations could be solved
problems [151, p. 7]. Here the general nonlinear from one position to the next, like initial value
Volterra equation of the second kind is treated differential equations, Fredholm equations
(Eq. 46). The simplest numerical method must be solved over the entire domain, like
involves replacing the integral by a quadrature boundary value differential equations. Thus,
using the trapezoid rule. large sets of equations will be solved and the
notation is designed to emphasize that.
yn yðtn Þ ¼ gðtn Þ
( ) The methods are also based on quadrature
1 X
n1
1
þ Dt Kðtn ; t0 ; y0 Þ þ Kðtn ; ti ; yi Þ þ Kðtn ; tn ; yn Þ formulas. For the integral
2 i¼1
2
Zb
This equation is a nonlinear algebraic equation IðwÞ ¼ fðyÞdy
for yn. Since y0 is known it can be applied to a
Mathematics in Chemical Engineering 99
is a known function. The integral equation is maxax;yb jF½x; y; f ðyÞ F½x; z; f ðzÞj Kjy zj
then replaced by
Theorem [150, p. 214]. If the constant K is < 1
X
n
f ðxÞ ¼ gðxÞ þ wi Kðx; yi Þ½f ðyi Þ f ðxÞ þ f ðxÞHðxÞ and certain other conditions hold, the succes-
i¼0 sive substitution method
X
n
converges to the solution of the integral
f ¼ ai fi ðxÞ
i¼0 equations.
100 Mathematics in Chemical Engineering
9.4. Numerical Methods for Eigenvalue where Green’s function is [50, p. 891]
Problems
1
Gðr; r0 Þ ¼ in three dimensions
r
Eigenvalue problems are treated similarly to
¼ 2 ln r in two dimensions
Fredholm equations, except that the final equa- qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
tion is a matrix eigenvalue problem instead of a where r ¼ ðx x0 Þ2 þ ðy y0 Þ2 þ ðz z0 Þ2
set of simultaneous equations. For example, in three dimensions
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
X
n and r ¼ ðx x0 Þ2 þ ðy x0 Þ2
wi Kðyi ; yi Þf ðyi Þ ¼ lf ðyj Þ;
i¼1 in two dimensions
i ¼ 0; 1; . . . ; n
For the problem
leads to the matrix eigenvalue problem @u
¼ Dr2 u; u ¼ 0 on S;
@t
K D f ¼ lf with a point source at x0 ; y0 ; z0
ZZZ
t
Z1 @u
þD fðx; y; z; tÞ dSdt
TðxÞ ¼
1
Gðx; yÞQðyÞdy @n
k 0
0
8
< xð1 yÞ xy When the problem is two-dimensional,
Gðx; yÞ ¼
: 1 2 2
yð1 xÞ yx u ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi e½ðxx0 Þ þðyy0 Þ =4DðttÞ
4pDðt tÞ
ZZ
Green’s functions for typical operators are c¼ ðuÞt¼0 f ðx; yÞdxdy
given below.
ZZt
For the Poisson equation with solution þD
@u
fðx; y; tÞ dCdt
decaying to zero at infinity @n
0
2
r c ¼ 4p% For the following differential equation and
boundary conditions
the formulation as an integral equation is
1 d dc
xa1 ¼ f ½x; cðxÞ;
Z xa1 dx dx
cðrÞ ¼ rðr0 ÞGðr; r0 ÞdV 0 dc 2 dc
ð0Þ ¼ 0; ð1Þ þ cð1Þ ¼ g
V dx Sh dx
Mathematics in Chemical Engineering 101
where Sh is the Sherwood number, the problem Thus, the problem ends up as one directly
can be written as a Hammerstein integral equa- formulated as a fixed point problem:
tion:
f ¼ Fðf Þ
Z1
cðxÞ ¼ g Gðx; y; ShÞf ½y; cðyÞya1 dy
When the problem is the diffusion–reaction
0
one, the form is
Green’s function for the differential operators Zx
are [163] cðxÞ ¼ g ½uðxÞvðyÞ uðyÞvðxÞ
0
a¼1
( 2
f ½y; cðyÞya1 dy avðxÞ
1 þ x; yx Z1
Sh a¼ uðyÞf ½y; cðyÞya1 dy
Gðx; y; ShÞ ¼ a ¼ 2Gðx; y; ShÞ
2
1 þ y; x<y 0
Sh
(2 1
9.6. Boundary Integral Equations and
þ 1; yx
Sh x Boundary Element Method
¼
2 1
þ 1; x<y
Sh y The boundary element method utilizes Green’s
theorem and integral equations. Here, the
Green’s functions for the reaction diffusion method is described briefly for the following
problem were used to provide computable error boundary value problem in two or three dimen-
bounds by FERGUSON and FINLAYSON [163]. sions
If Green’s function has the form
@f
( r2 f ¼ 0; f ¼ f 1 on S1 ; ¼ f 2 on S2
uðxÞvðyÞ 0yx @n
Kðx; yÞ ¼
uðyÞvðxÞ xy1
Green’s theorem (see chap. 5) says that for any
functions sufficiently smooth
the problem
Z Z
@c @f
Z1 ðfr2 c cr2 fÞdV ¼ f c dS
@n @n
f ðxÞ ¼ Kðx; yÞF½y; f ðyÞdy V S
where {x0, y0} or {x0, y0, z0} is a point in the This is an integral equation for f on the bound-
domain. The solution f also satisfies ary. Note that the order is one less than the
original differential equation. However, the
r2 f ¼ 0 integral equation leads to matrices that are
dense rather than banded or sparse, so some
so that of the advantage of lower dimension is lost.
Once this integral equation (Eq. 48) is solved to
Z
@c @f find f on the boundary, it can be substituted in
f c dS ¼ 0
@n @n Equation (47) to find f anywhere in the domain.
S
In the boundary finite element method, both
Consider the two-dimensional case. Since the function and its normal derivative along the
the function c is singular at a point, the inte- boundary are approximated.
grals must be carefully evaluated. For the
X N
region shown in Figure 31, the domain is N
@f X @f
f¼ fj N j ðjÞ; ¼ N j ðjÞ
S ¼ S1þ S2; a small circle of radius r0 is placed j¼1
@n j¼1
@n j
around the point P at x0, y0. Then the full
integral is One choice of trial functions can be the piece-
wise constant functions shown in Figure 32.
R @ ln r @f The integral equation then becomes
f ln r dS
S @n @n
2 3
Z
u¼2p Z Z
@ ln r0 @f XN
6 @ ln ri @fj 7
þ f ln r0 r0 du ¼ 0 pfi 4f j dS ln r i 5ds
@n @n @n @n
u¼0 j¼1
sj sj
and
Z
u¼2p
@ ln r0
lim f r0 du ¼ fðPÞ2p
r 0 !0 @n
u¼0
This set of equations is then solved for fi and mathematical optimization problems involving
@fi/@n along the boundary. This constitutes the continuous and discrete (or integer) variables.
boundary integral method applied to the Laplace This is followed by a review of solution meth-
equation. ods of the major types of optimization problems
If the problem is Poisson’s equation for continuous and discrete variable optimiza-
tion, particularly nonlinear and mixed-integer
r2 f ¼ gðx; yÞ nonlinear programming. In addition, we discuss
direct search methods that do not require deriv-
Green’s theorem gives ative information as well as global optimization
methods, before reviewing extensions of these
Z Z
@ ln r @f methods for the optimization of systems
f ln r dS þ g ln r dA ¼ 0
@n @n described by differential and algebraic equa-
S A
tions. These methods, which are designed for
Thus, for an internal point, single-objective optimization problems, form
the basis to tackle multicriteria optimization
R @ ln r @f problems and optimization under uncertainty,
2pfðPÞ ¼ f ln r dS
S @r @n ð49Þ which are reviewed in the two final parts of this
R chapter.
þ A g ln rdA
10. Optimization
We provide a survey of systematic methods for
a broad variety of optimization problems. The
survey begins with a general classification of Figure 33.
104 Mathematics in Chemical Engineering
x 2 n ; y 2 f0; 1gt
functions involved are nonlinear. If all functions Strict convexity requires that the inequalities
are linear it corresponds to a mixed-integer linear (53) and (54) be strict. Convex feasible
program (MILP). If there are no 0–1 variables, regions require g(x) to be a convex function
then Problem (51) reduces to a nonlinear pro- and h(x) to be linear. If Problem (52) is a
gram (52) or linear program (65) depending on convex problem, than any local solution is
whether or not the functions are linear. guaranteed to be a global solution to Problem
We first start with continuous variable opti- (52). Moreover, if the objective function is
mization and consider in the next section the strictly convex, then this solution x is unique.
solution of NLPs with differentiable objective On the other hand, nonconvex problems may
and constraint functions. If only local solutions have multiple local solutions, i.e., feasible
are required for the NLP, then very efficient solutions that minimize the objective function
large-scale methods can be considered. This is within some neighborhood about the solution.
followed by methods that are not based on local We first consider methods that find only
optimality criteria; we consider direct search local solutions to nonconvex problems, as
optimization methods that do not require deriv- more difficult (and expensive) search proce-
atives, as well as deterministic global optimi- dures are required to find a global solution.
zation methods. Following this, we consider the Local methods are currently very efficient and
solution of mixed-integer problems and outline have been developed to deal with very large
the main characteristics of algorithms for their NLPs. Moreover, by considering the structure
solution. Finally, we conclude with a discussion of convex NLPs (including LPs and QPs), even
of optimization modeling software and its more powerful methods can be applied. To
implementation in engineering models. study these methods, we first consider condi-
tions for local optimality.
fences [rgðx Þ] and rails [rhðx Þ] on the
ball are balanced by the force of “gravity”
[rf ðx Þ]. This condition can be stated by
the following KKT necessary conditions for
constrained optimality:
Stationarity Condition: It is convenient to
define the Lagrangian function L(x, l, n) ¼ f(x) þ
g(x)Tl þ h(x)Tn along with “weights” or multi-
pliers l and n for the constraints. These multipli-
ers are also known as “dual variables” and
“shadow prices”. The stationarity condition
(balance of forces acting on the ball) is then given
by:
(l, n) are guaranteed to be unique at the optimal Min ðx1 Þ2 4x1 þ 3=2ðx1 Þ2 7x2 þ x1 x2 þ 9 lnx1 lnx2
solution. The weaker Mangasarian–Fromovitz s:t: 4 x1 x 2 0 ð63Þ
constraint qualification (MFCQ) requires only 2x1 x2 ¼ 0
that rh(x ) have full column rank and that a
direction p exist that satisfies rh(x )Tp ¼ 0 and that corresponds to the plot in Figure 37. The
rgA(x )Tp > 0. With MFCQ, the KKT multi- optimal solution can be found by applying the
pliers (l, n) are guaranteed to be bounded (but first-order KKT Conditions (56–58):
not necessarily unique) at the optimal solution.
Additional discussion can be found in [172]. rLðx; l; nÞ ¼ rf ðxÞ þ rhðxÞl þ rgðxÞn
Second-Order Conditions: As with un-
2x1 4 þ x2 1=x1 2 x2
constrained optimization, nonnegative (posi- ¼ þ lþ n¼0
3x2 7 þ x1 1=x2 1 x1
tive) curvature is necessary (sufficient) in all gðxÞ ¼ 4 x1 x2 0; hðxÞ ¼ 2x1 x2 ¼ 0 ð64Þ
of the allowable (i.e., constrained) nonzero gðxÞn ¼ ð4 x1 x2 Þn; n 0
+
directions p. This condition can be stated in 1:4142
several ways. A typical necessary second-order x ¼ ; l ¼ 1:036; n ¼ 1:068
2:8284
condition requires a point x that satisfies LICQ
and first-order Conditions (56–58) with multi- and f(x ) ¼ 1.8421. Checking the second-
pliers (l, n) to satisfy the additional conditions order Conditions (60) leads to:
given by:
rxx Lðx ; l ; n Þ ¼ rxx ½f ðx Þ þ hðx Þl þ gðx Þn
pT rxx Lðx ; l; nÞp 0 for all p 6¼ 0; rhðx ÞT p ¼ 0; rgA ðx ÞT p ¼ 0 2
2:5 0:068
ð59Þ ¼ 2 þ 1=ðx1 Þ 1n
2 ¼
1n 3 þ 1=ðx2 Þ 0:068 3:125
2 2:8284
The corresponding sufficient conditions require ½rhðx ÞjrgA ðx Þp ¼ p ¼ 0; p 6¼ 0
1 1:4142
that the inequality in Condition (59) be strict.
Note that for the example in Figure 36, the Note that LICQ is satisfied. Moreover, because
allowable directions p span the entire space for
½rhðx ÞjrgA ðx Þ is square and nonsingular,
x while in Figure 37, there are no allowable there are no nonzero vectors p that satisfy the
directions p. allowable directions. Hence, the sufficient sec-
Example: To illustrate the KKT conditions,
ond-order conditions (pT rxx Lðx ; l ; n Þ > 0,
consider the following unconstrained NLP: for all allowable p) are vacuously satisfied
for this problem.
Min ðx1 Þ2 4x1 þ 3=2ðx2 Þ2 7x2 þ x1 x2 þ 9 lnx1 lnx2
ð60Þ Convex Cases of NLP. Linear programs and
quadratic programs are special cases of Problem
corresponding to the contour plot in Figure 36. (52) that allow for more efficient solution, based
The optimal solution can be found by solving on application of KKT Conditions (56–59).
108 Mathematics in Chemical Engineering
Because these are convex problems, any locally developed in the late 1940s [173], although,
optimal solution is a global solution. In particu- starting from KARMARKAR’s discovery in 1984,
lar, if the objective and constraint functions in interior point methods have become quite
Problem (52) are linear then the following linear advanced and competitive for large scale prob-
program (LP): lems [174]. The simplex method proceeds by
moving successively from vertex to vertex with
Min cT x improved objective function values. Methods to
s:t: Ax ¼ b ð65Þ solve Problem (66) are well implemented and
widely used, especially in planning and logisti-
Cx d
cal applications. They also form the basis for
MILP methods (see below). Currently, state-of-
can be solved in a finite number of steps, and the
the-art LP solvers can handle millions of vari-
optimal solution lies at a vertex of the polyhe-
ables and constraints and the application of
dron described by the linear constraints. This is
further decomposition methods leads to the
shown in Figure 39, and in so-called primal
solution of problems that are two or three orders
degenerate cases, multiple vertices can be alter-
of magnitude larger than this [175, 176]. Also,
nate optimal solutions with the same values of
the interior point method is described below
the objective function. The standard method to
from the perspective of more general NLPs.
solve Problem (65) is the simplex method,
Quadratic programs (QP) represent a slight
modification of Problem (65) and can be stated
as:
1
Min cT x þ xT Qx
2
s:t: Ax ¼ b ð66Þ
xd
1
Min rf ðxk ÞT p þ pT rxx Lðxk ; lk ; nk Þp
2
ð68Þ
s:t: hðxk Þ þ rhðxk ÞT p ¼ 0
gðxk Þ þ rgðxk ÞT p þ s ¼ 0; s 0
Figure 40. Contour plots of convex quadratic programs The KKT conditions of Equation (68) are given
by:
Often these strategies are heuristics built one-at-a-time search and methods based on
into particular NLP codes. While quite experimental designs [196]. At that time, these
effective, most of these heuristics do not direct search methods were the most popular
provide convergence guarantees for gen- optimization methods in chemical engineering.
eral NLPs. Methods that fall into this class include the
pattern search [197], the conjugate direction
Table 11 summarizes the characteristics of a method [198], simplex and complex searches
collection of widely used NLP codes. Much [199], and the adaptive random search methods
more information on widely available codes [200–202]. All of these methods require only
can also be found on the NEOS server objective function values for unconstrained
(www-neos.mcs.anl.gov) and the NEOS minimization. Associated with these methods
Software Guide. are numerous studies on a wide range of process
problems. Moreover, many of these methods
include heuristics that prevent premature ter-
10.3. Optimization Methods without mination (e.g., directional flexibility in the
Derivatives complex search as well as random restarts
and direction generation). To illustrate these
A broad class of optimization strategies does methods, Figure 41 illustrates the performance
not require derivative information. These of a pattern search method as well as a
methods have the advantage of easy imple- random search method on an unconstrained
mentation and little prior knowledge of the problem.
optimization problem. In particular, such
methods are well suited for “quick and dirty” Simulated Annealing. This strategy is related
optimization studies that explore the scope of to random search methods and derives from a
optimization for new problems, prior to class of heuristics with analogies to the motion
investing effort for more sophisticated model- of molecules in the cooling and solidification of
ing and solution strategies. Most of these metals [203]. Here a “temperature” parameter u
methods are derived from heuristics that nat- can be raised or lowered to influence the prob-
urally spawn numerous variations. As a result, ability of accepting points that do not improve
a very broad literature describes these meth- the objective function. The method starts with a
ods. Here we discuss only a few important base point x and objective value f(x). The next
trends in this area. point x’ is chosen at random from a distribution.
If f(x’) < f(x), the move is accepted with x’
Classical Direct Search Methods. Developed as the new point. Otherwise, x’ is accepted
in the 1960s and 1970s, these methods include with probability p(u, x’,x). Options include the
114 Mathematics in Chemical Engineering
large, complex simulation models. These variables and compares lower bound and upper
include the DAKOTA package at Sandia bound for fathoming each subregion. The one
National Lab [207], http://endo.sandia.gov/ that contains the optimal solution is found by
DAKOTA/software.html and FOCUS eliminating subregions that are proved not to
developed at Boeing Corporation [208]. contain the optimal solution.
Direct search methods are easy to apply to a For nonconvex NLP problems, QUESADA and
wide variety of problem types and optimization GROSSMANN [210] proposed a spatial branch-
models. Moreover, because their termination and-bound algorithm for concave separable,
criteria are not based on gradient information linear fractional, and bilinear programs using
and stationary points, they are more likely to linear and nonlinear underestimating functions
favor the search for global rather than locally [211]. For nonconvex MINLP, RYOO and SAHI-
optimal solutions. These methods can also be NIDIS [212] and later TAWARMALANI and SAHINIDIS
adapted easily to include integer variables. [213] developed BARON, which branches on
However, rigorous convergence properties to the continuous and discrete variables with
globally optimal solutions have not yet been bounds reduction method. ADJIMAN et al.
discovered. Also, these methods are best suited [214, 215] proposed the SMIN-aBB and
for unconstrained problems or for problems GMIN-aBB algorithms for twice-differentiable
with simple bounds. Otherwise, they may nonconvex MINLPs. Using a valid convex
have difficulties with constraints, as the only underestimation of general functions as well
options open for handling constraints are equal- as for special functions, ADJIMAN et al. [216]
ity constraint elimination and addition of pen- developed the aBB method, which branches on
alty functions for inequality constraints. Both both the continuous and discrete variables
approaches can be unreliable and may lead to according to specific options. The branch-
failure of the optimization algorithm. Finally, and-contract method [217] has bilinear, linear
the performance of direct search methods scales fractional, and concave separable functions in
poorly (and often exponentially) with the num- the continuous variables and binary variables,
ber of decision variables. While performance uses bound contraction, and applies the outer-
can be improved with the use of parallel com- approximation (OA) algorithm at each node of
puting, these methods are rarely applied to the tree. SMITH and PANTELIDES [218] proposed a
problems with more than a few dozen decision reformulation method combined with a spatial
variables. branch-and-bound algorithm for nonconvex
MINLP and NLP.
Because in global optimization one cannot
10.4. Global Optimization exploit optimality conditions like the KKT
conditions for a local optimum, these methods
Deterministic optimization methods are avail- work by first partitioning the problem domain
able for nonconvex nonlinear programming (i.e., containing the feasible region) into sub-
problems of the form Problem (52) that guar- regions (see Fig. 42). Upper bounds on the
antee convergence to the global optimum. objective function are computed over all sub-
More specifically, one can show under mild regions of the problem. In addition, lower
conditions that they converge to an e distance bounds can be derived from convex underesti-
to the global optimum on a finite number of mators of the objective function and constraints
steps. These methods are generally more for each subregion. The algorithm then pro-
expensive than local NLP methods, and they ceeds to eliminate all subregions that have
require the exploitation of the structure of the lower bounds that are greater than the least
nonlinear program. upper bound. After this, the remaining regions
Global optimization of nonconvex programs are further partitioned to create new subregions
has received increased attention due to their and the cycle continues until the upper and
practical importance. Most of the deterministic lower bounds converge. Below we illustrate
global optimization algorithms are based on the specific steps of the algorithm for nonlinear
spatial branch-and-bound algorithm [209], programs that involve bilinear, linear fractional,
which divides the feasible region of continuous and concave separable terms [210, 217].
116 Mathematics in Chemical Engineering
subject to
X X X
^f k ðx; y; zÞ ¼ aijk yij þ bijk zij þ ^gi;k ðxi Þ
ði;jÞ2BLk ði;jÞ2LF k i2C k
þ hk ðxÞ 0 k2K
ðx; y; zÞ 2 TðVÞ
R R Rn2 n n1
where aij, aijk, bij, bijk are scalars with i 2 I ¼ where ^gi ðxi Þ ¼ gi ðxi Þ at xi ¼ xLi , and
f1; 2; ; ng, j 2 J ¼ f1; 2; ; ng, and k 2 xi ¼ xUi ; likewise, ^ gi;k ðxi Þ ¼ gi;k ðxi Þ at
K ¼ f1; 2; ; mg. BL0 ; BLk ; LF 0 ; LF k are xi ¼ xLi , and xi ¼ xUi .
(i, j)-index sets, with i 6¼ j, that define the b. y ¼ fyij g is a vector of additional variables
bilinear and linear fractional terms present in for relaxing the bilinear terms in Problem
the problem. The functions hðxÞ; hk ðxÞ are (80), and is used in the following inequali-
convex, and twice continuously differentiable. ties which determine the convex and con-
C0 and Ck are index sets for the univariate twice cave envelopes of bilinear terms:
continuously differentiable concave functions
gi ðxi Þ; gi;k ðxi Þ. The set S
Rn is convex, and
v0Rn is an n-dimensional hyperrectangle yij xLj xi þ xLi xj xLi xLj ði; jÞ 2 BLþ
defined in terms of the initial variable bounds ð82Þ
yij xU
j xi þ xU
i xj xU U
i xj ði; jÞ 2 BLþ
xL,in and xU,in:
yij xLj xi þ xU U L
i xj xi x j ði; jÞ 2 BL
V0 ¼ fx 2 Rn : 0 xL;in x xU;in ; xL;in
j >0
ð83Þ
ifði; jÞ 2 LF 0 [ LF k ; i 2 I; j 2 J; k 2 Kg yij xU L L U
j xi þ xi xj xi x j ði; jÞ 2 BL
Mathematics in Chemical Engineering 117
s:t: x1 x x u
ð90Þ
For iteration k with a set of partitions Vk;j 3ðxu xÞ ðx xl Þ
and bounds in each subregion OLBj and OUBj: w ¼ ðxl Þ þ ðxu Þ3
ðxu xl Þ ðxu xl Þ
1. Bound: Define best upper bound: OUB ¼ In Figure 43 we also propose subregions that
Minj OUBj and delete (fathom) all subre- are created by simple bisection partitioning
gions j with lower bounds OLBj OUB. If rules, and we use a “loose” bounding tolerance
OLBj OUB e, stop. of e ¼ 0:2. In each partition the lower bound, fL
2. Refine: Divide the remaining active subre- is determined by Problem (90) and the upper
gions into partitiions Vk;j1 and Vk;j2 . (Several bound fU is determined by the local solution of
branching rules are available for this step.) the original problem in the subregion. Figure 44
3. Select: Solve the convex NLP (80) in the shows the progress of the spatial branch-
new partitions to obtain OLBj1 and OLBj2. and-bound algorithm as the partitions are
Delete partition if no feasible solution. refined and the bounds are updated. In
Mathematics in Chemical Engineering 119
Figure 43, note the definitions of the partitions Unlike local optimization methods, there are no
for the nodes, and the sequence numbers in each optimality conditions, like the KKT conditions,
node that show the order in which the partitions that can be applied directly.
are processed. The grayed partitions correspond
to the deleted subregions and at termination of Mixed Integer Linear Programming. If the
the algorithm we see that fLj fU e (i.e., objective and constraint functions are all linear
19.85 19.7–0.2), with the gray subregions in Problem (52), and we assume 0–1 binary
in Figure 43 still active. Further partitioning in variables for the discrete variables, then this
these subregions will allow the lower and upper gives rise to a mixed integer linear program-
bounds to converge to a tighter tolerance. ming (MILP) problem given by Equation (91).
A number of improvements can be made to
the bounding, refinement, and selection strate- Min Z ¼ aT x þ cT y
gies in the algorithm that accelerate the con- s:t: Ax þ By b ð91Þ
x 0; y 2 f0; 1gt
vergence of this method. A comprehensive
discussion of all of these options can be found
in [225–227]. Also, a number of efficient global As is well known, the (MILP) problem is NP-
optimization codes have recently been devel- hard. Nevertheless, an interesting theoretical
oped, including aBB, BARON, LGO, and result is that it is possible to transform it into
OQNLP. An interesting numerical comparison an LP with the convexification procedures pro-
of these and other codes can be found in [228]. posed by LOVACZ and SCHRIJVER [229], SHERALI
and ADAMS [230], and BALAS et al. [231]. These
procedures consist of sequentially lifting the
10.5. Mixed Integer Programming original relaxed xy space into higher dimen-
sion and projecting it back to the original space
Mixed integer programming deals with both so as to yield, after a finite number of steps, the
discrete and continuous decision variables. integer convex hull. Since the transformations
For simplicity in the presentation we consider have exponential complexity, they are only of
the most common case where the discrete deci- theoretical interest, although they can be used
sions are binary variables, i.e., yi ¼ 0 or 1, and as a basis for deriving cutting planes (e.g. lift
we consider the mixed integer Problem (52). and project method by [231]).
120 Mathematics in Chemical Engineering
method. The selection of binary variables for OSL, XPRESS. Recent trends in MILP include
branching, known as the branching rule, is the development of branch-and-price [241]
based on a number of different possible criteria; and branch-and-cut methods such as the lift-
for instance, choosing the fractional variable and-project method [231] in which cutting
closest to 0.5, or the one involving the largest of planes are generated as part of the branch-
the smallest pseudocosts for each fractional and-bound enumeration. See also [238] for a
variable. Branching strategies to navigate the recent review on MILP. A description of sev-
tree take a number of forms. More common eral MILP solvers can also be found in the
depth-first strategies expand the most recent NEOS Software Guide.
node to a leaf node or infeasible node and
then backtrack to other branches in the tree. Mixed Integer Nonlinear Programming The
These strategies are simple to program and most basic form of an MINLP problem when
require little storage of past nodes. On the other represented in algebraic form is as follows:
hand, breadth-first strategies expand all the
min z ¼ f ðx; yÞ
nodes at each level of the tree, select the s:t: gj ðx; yÞ 0 j 2 J ðP1Þ ð93Þ
node with the lowest objective function, and x 2 X; y 2 Y
then proceed until the leaf nodes are reached.
Here, more storage is required, but generally where f(), g() are convex, differentiable func-
fewer nodes are evaluated than in depth-first tions, J is the index set of inequalities, and x and
search. In practice, a combination of both y are the continuous and discrete variables,
strategies is commonly used: branch on the respectively. The set X is commonly assumed
dichotomy 0–1 at each node (i.e., like to be a convex compact set, e.g.,
breadth-first), but expand as in depth-first. X ¼ fxjx 2 Rn ; Dx d; xL x xU g; the
Additional description of these strategies can discrete set Y corresponds to a polyhedral set
be found in [240]. of integer points, Y ¼ fyjx 2 Z m ; Ay ag,
Example: To illustrate the branch-and-bound which in most applications is restricted to
approach, we consider the MILP: 0–1 values, y 2 {0,1}m. In most applications
of interest the objective and constraint func-
Min Z ¼ x þ y1 þ 2y2 þ 3y3 tions f(), g() are linear in y (e.g., fixed cost
s:t: x þ 3y1 þ y2 þ 2y3 0 charges and mixed-logic constraints):
ð92Þ
4y1 8y2 3y3 10
x 0; y1 ; y2 ; y3 ¼ f0; 1g
f ðx; yÞ ¼ cT y þ rðxÞ; gðx; yÞ ¼ By þ hðxÞ.
Methods that have addressed the solution of
Problem (93) include the branch-and-bound
The solution to Problem (92) is given by x ¼ 4, method (BB) [242–246], generalized Benders
y1 ¼ 1, y2 ¼ 1, y3 ¼ 0, and Z ¼ 7. Here we use a decomposition (GBD) [247], outer-approxima-
depth-first strategy and branch on the variables tion (OA) [248–250], LP/NLP based branch-
closest to zero or one. Figure 45 shows the and-bound [251], and extended cutting plane
progress of the branch-and-bound algorithm as method (ECP) [252].
the binary variables are selected and the bounds There are three basic NLP subproblems that
are updated. The sequence numbers for each can be considered for Problem (93):
node in Figure 45 show the order in which they
are processed. The grayed partitions correspond a. NLP relaxation
to the deleted nodes and at termination of the
algorithm we see that Z ¼ 7 and an integer
Min Z kLB ¼ f ðx; yÞ
solution is obtained at an intermediate node s:t: gj ðx:yÞ 0 j 2 J
where coincidentally y3 ¼ 0. x 2 X; y 2 Y R ðNLP1Þ ð94Þ
Mixed integer linear programming (MILP) yi aki i 2 I kFL
yi bki i 2 I kFU
methods and codes have been available and
applied to many practical problems for more
than twenty years (e.g., [240]. The LP-based where YR is the continuous relaxation of the
branch-and-bound method [234] has been set Y, and I kFL ; I kFU are index subsets of
implemented in powerful codes such as the integer variables yi, i 2 I which are
122 Mathematics in Chemical Engineering
x 2 X; y 2 Y ð98Þ
Given the assumption on convexity of the func-
tions f(x, y) and g(x, y), it can be proved that the
solution of Problem (98) Z kL corresponds to a
lower bound of the solution of Problem (93).
Note that this property can be verified in
Figure 46. Also, since function linearizations
are accumulated as iterations proceed, the mas-
ter Problems (98) yield a nondecreasing
sequence of lower bounds Z 1L . . . Z kL . . . Z kL
since linearizations are accumulated as itera-
tions k proceed.
The OA algorithm as proposed by DURAN
and GROSSMANN consists of performing a cycle
of major iterations, k ¼ 1,..K, in which Problem
Figure 47. Major steps in the different algorithms (95) is solved for the corresponding yK, and the
124 Mathematics in Chemical Engineering
relaxed MILP master Problem (98) is updated fjjgj ðxk ; yk Þ ¼ 0g and the set x 2 X is disre-
and solved with the corresponding function garded. In particular, consider an outer-approx-
linearizations at the point (xk, yk) for which imation given at a given point (xk, yk)
the corresponding subproblem NLP2 is solved.
If feasible, the solution to that problem is used 2 3
x xk
to construct the first MILP master problem; k T67
a f ðx ; y Þ þ rf ðx ; y Þ 4
k k k
5
otherwise a feasibility Problem (96) is solved y yk
to generate the corresponding continuous point 2 3 ð100Þ
x xk
[250]. The initial MILP master Problem (98) k k T6
gðx ; y Þ þ rg ðx ; y Þ 4
k k 7
50
then generates a new vector of discrete varia- y yk
bles. The Subproblems (95) yield an upper
bound that is used to define the best current
solution, UBk ¼ mink fZ kU g. The cycle of itera- where for a fixed yk the point xk corresponds to
tions is continued until this upper bound and the the optimal solution to Problem (95). Making
lower bound of the relaxed master problem Z kL use of the Karush–Kuhn–Tucker conditions and
are within a specified tolerance. One way to eliminating the continuous variables x, the
avoid solving the feasibility Problem (96) in the inequalities in Problem (100) can be reduced
OA algorithm when the discrete variables in as follows [251]:
Problem (93) are 0–1, is to introduce the fol-
lowing integer cut whose objective is to make a f ðxk ; yk Þ þ ry f ðxk ; yk ÞT ðy yk Þ þ ðmk ÞT ½gðxk ; yk Þ
infeasible the choice of the previous 0–1 values þ ry gðxk ; yk ÞT ðy yk Þ ð101Þ
generated at the K previous iterations [248]:
X X
yi yi jBk j 1 k ¼ 1; . . . K ð99Þ which is the Lagrangian cut projected in the y-
k
i2B i2N k
space. This can be interpreted as a surrogate
constraint of the equations in Problem (100),
where Bk¼fijyki ¼ 1g; N k ¼ fijyki ¼ 0g; k ¼ because it is obtained as a linear combination
1; . . . K. This cut becomes very weak as the of these.
dimensionality of the 0–1 variables increases. For the case when there is no feasible solu-
However, it has the useful feature of ensuring tion to Problem (95), then if the point xk is
that new 0–1 values are generated at each obtained from the feasibility subproblem
major iteration. In this way the algorithm (NLPF), the following feasibility cut projected
will not return to a previous integer point in y can be obtained using a similar procedure.
when convergence is achieved. Using the
above integer cut the termination takes place ðlk ÞT ½gðxk ; yk Þ þ ry gðxk ; yk ÞT ðy yk Þ 0 ð102Þ
as soon as Z KL UBK .
The OA algorithm trivially converges in
In this way, Problem (97) reduces to a problem
one iteration if f(x, y) and g(x, y) are linear.
projected in the y-space:
This property simply follows from the fact that
if f(x, y) and g(x, y) are linear in x and y the Min Z KL ¼ a
MILP master problem is identical to the origi-
st a f ðxk ; yk Þ þ ry f ðxk ; yk ÞT ðy yk Þ þ ðmk ÞT
nal Problem (93). It is also important to note
that the MILP master problem need not be ½gðxk ; yk Þ þ ry gðxk ; yk ÞT ðy yk Þ k 2 KFS
solved to optimality.
ðlk ÞT ½gðxk ; yk Þ þ ry gðxk ; yk ÞT ðy yk Þ 0 k 2 KIS
can be derived from master Problem (98), in the Convergence is achieved when the maximum
context of Problem (93), GBD can be regarded constraint violation lies within the specified
as a particular case of the outer-approximation tolerance. The optimal objective value of
algorithm. In fact one can prove that given the Problem (97) yields a nondecreasing sequence
same set of K subproblems, the lower bound of lower bounds. It is of course also possible to
predicted by the relaxed master problem is either add to problem linearizatons of all the
greater than or equal to that predicted by the violated constraints in the set Jk, or lineariza-
relaxed master Problem (103) [248]. This proof tions of all the nonlinear constraints j 2 J. In
follows from the fact that the Lagrangian and the ECP method the objective must be defined
feasibility cuts (Eqs. 101 and 102) are surro- as a linear function, which can easily be
gates of the outer-approximations (Problems accomplished by introducing a new variable
103). Given the fact that the lower bounds of to transfer nonlinearities in the objective as an
GBD are generally weaker, this method com- inequality.
monly requires a larger number of cycles or Note that since the discrete and continuous
major iterations. As the number of 0–1 variables variables are converged simultaneously, the
increases this difference becomes more pro- ECP method may require a large number of
nounced. This is to be expected since only one iterations. However, this method shares with the
new cut is generated per iteration. Therefore, OA method Property 2 for the limiting case
user-supplied constraints must often be added when all the functions are linear.
to the master problem to strengthen the bounds.
Also, it is sometimes possible to generate LP/NLP-Based Branch and Bound [251]. This
multiple cuts from the solution of an NLP method is similar in spirit to a branch-and-cut
subproblem in order to strengthen the lower method, and avoids the complete solution of the
bound [254]. As for the OA algorithm, the MILP master problem (M-OA) at each major
trade-off is that while it generally predicts iteration. The method starts by solving an initial
stronger lower bounds than GBD, the computa- NLP subproblem, which is linearized as in (M-
tional cost for solving the master problem (M- OA). The basic idea consists then of performing
OA) is greater, since the number of constraints an LP-based branch-and-bound method for (M-
added per iteration is equal to the number of OA) in which NLP Subproblems (95) are
nonlinear constraints plus the nonlinear solved at those nodes in which feasible integer
objective. solutions are found. By updating the represen-
If Problem (93) has zero integrality gap, the tation of the master problem in the current open
GBD algorithm converges in one iteration once nodes of the tree with the addition of the
the optimal (x , y ) is found [255]. This prop- corresponding linearizations, the need to restart
erty implies that the only case in which one can the tree search is avoided.
expect the GBD method to terminate in This method can also be applied to the
one iteration is that in which the initial discrete GBD and ECP methods. The LP/NLP method
vector is the optimum, and when the objective commonly reduces quite significantly the
value of the NLP relaxation of Problem (93) is number of nodes to be enumerated. The
the same as the objective of the optimal mixed- trade-off, however, is that the number of
integer solution. NLP subproblems may increase. Computa-
tional experience has indicated that often
Extended Cutting Plane (ECP) [252]. The
the number of NLP subproblems remains
ECP method, which is an extension of Kelley’s
unchanged. Therefore, this method is better
cutting-plane algorithm for convex NLP [256],
suited for problems in which the bottleneck
does not rely on the use of NLP subproblems
corresponds to the solution of the MILP mas-
and algorithms. It relies only on the iterative
ter problem. LEYFFER [257] has reported sub-
solution of problem by successively adding a
stantial savings with this method.
linearization of the most violated constraint at
Example: Consider the following MINLP
the predicted point (xk, yk):
problem, whose objective function and con-
J k ¼ f^j 2 argfmax gj ðxk ; yk Þgg straints contain nonlinear convex terms
j2J
126 Mathematics in Chemical Engineering
(e.g., M-MIPF, M-GBD, M-MIQP) do not Since the tree searches are not finite (except for
guarantee a valid lower bound Z KL or a valid e convergence), these methods can be compu-
bounding representation with which the global tationally expensive. However, their major
optimum may be cut off. advantage is that they can rigorously find the
Rigorous global optimization approaches for global optimum. Specific cases of nonconvex
addressing nonconvexities in MINLP problems MINLP problems have been handled. An exam-
can be developed when special structures are ple is the work of PO€ RN and WESTERLUND [264],
assumed in the continuous terms (e.g. bilinear, who addressed the solution of MINLP prob-
linear fractional, concave separable). Specifi- lems with pseudoconvex objective function and
cally, the idea is to use convex envelopes or convex inequalities through an extension of the
underestimators to formulate lower-bounding ECP method.
convex MINLP problems. These are then com- The other option for handling nonconvex-
bined with global optimization techniques for ities is to apply a heuristic strategy to try to
continuous variables [209, 210, 212, 217, 219, reduce as much as possible the effect of non-
225, 260], which usually take the form of convexities. While not being rigorous, this
spatial branch-and-bound methods. The lower requires much less computational effort. We
bounding MINLP problem has the general describe here an approach for reducing the
form, effect of nonconvexities at the level of the
MILP master problem.
Min Z ¼ f ðx; yÞ VISWANATHAN and GROSSMANN [265] proposed
s:t: g j ðx; yÞ 0 j 2 J ð105Þ
x 2 X; y 2 Y
to introduce slacks in the MILP master problem
to reduce the likelihood of cutting off feasible
solutions. This master problem (augmented pen-
where f ; g
, are valid convex underestimators alty/equality relaxation) has the form:
such that f ðx; yÞ f ðx; yÞ and the inequalities X
K
g ðx; yÞ 0 are satisfied if gðx; yÞ 0. A typi- min Z K ¼ a þ wkp pk þ wkq qk
cal example of convex underestimators are the k¼1 2 39
x xk>
>
convex envelopes for bilinear terms [211]. 6 7>
s:t: a f ðxk ; yk Þ þ rf ðxk ; yk ÞT 4 5>
>
>
>
>
Examples of global optimization methods yy >
k >
>
>
>
>
for MINLP problems include the branch-and- 2 3 >
>
>
x xk >
=
reduce method [212, 213], the a-BB method k T6
T k rhðxk ; y Þ 4
7
k ¼ 1; . . . K
5 pk
[215], the reformulation/spatial branch-and- >
>
yy k >
>
>
>
bound search method [261], the branch-and- 2 3 >
>
>
>
x xk >
>
cut method [262], and the disjunctive branch- k T6 7 >
>
>
gðx ; y Þ þ rgðx ; y Þ 4
k k k
5 qk >
>
and-bound method [263]. All these methods >
;
k
yy
rely on a branch-and-bound procedure. The P P
i2Bk yi i2N k yi jBk j 1 k ¼ 1; . . . K
difference lies in how to perform the branching
x 2 X; y 2 Y; a 2 R1 ; pk ; qk 0
on the discrete and continuous variables. Some ð106Þ
methods perform the spatial tree enumeration
on both the discrete and continuous variables of where wkp ; wkq
are weights that are chosen
Problem (105). Other methods perform a spa- sufficiently large (e.g., 1000 times the magni-
tial branch and bound on the continuous vari- tude of the Lagrange multiplier). Note that if the
ables and solve the corresponding MINLP functions are convex then the MILP master
Problem (105) at each node using any of the Problem (106) predicts rigorous lower bounds
methods reviewed above. Finally, other meth- to Problem (93) since all the slacks are set to
ods branch on the discrete variables of Problem zero.
(105), and switch to a spatial branch and bound
on nodes where a feasible value for the discrete Computer Codes for MINLP. Computer codes
variables is found. The methods also rely on for solving MINLP problems include the fol-
procedures for tightening the lower and upper lowing. The program DICOPT [265] is an
bounds of the variables, since these have a great MINLP solver that is available in the modeling
effect on the quality of the underestimators. system GAMS [266]. The code is based on the
128 Mathematics in Chemical Engineering
master Problem (106) and the NLP Subpro- review of logic-based optimization can be found
blems (95). This code also uses relaxed Sub- in [272, 273].
problem (94) to generate the first linearization Generalized disjunctive programming in
for the above master problem, with which the Problem (107) [269] is an extension of dis-
user need not specify an initial integer value. junctive programming [274] that provides an
Also, since bounding properties of Problem alternative way of modeling MILP and MINLP
(106) cannot be guaranteed, the search for problems. The general formulation of Problem
nonconvex problems is terminated when there (107) is as follows:
is no further improvement in the feasible NLP P
Min Z ¼ k2K ck þ f ðxÞ
subproblems. This is a heuristic that works
s:t: gðxÞ 0 ð107Þ
reasonably well in many problems. Codes 2 3
Y jk
that implement the branch-and-bound method _
j2J k 6 7
6 hjk ðxÞ 0 7; k 2 K
using Subproblems (94) include the code 4 5
ck ¼ g jk
MINLP_BB, which is based on an SQP algo-
rithm [246] and is available in AMPL, the code VðYÞ ¼ True
BARON [267], which also implements global
x 2 Rn ; c 2 Rm ; Y 2 ftrue; falsegm
optimization capabilities, and the code SBB,
which is available in GAMS [266]. The code a-
ECP implements the extended cutting-plane where Yjk are the Boolean variables that decide
method [252], including the extension by PO€ RN whether a term j in a disjunction k 2 K is true or
and WESTERLUND [264]. Finally, the code MIN- false, and x are continuous variables. The objec-
OPT [268] also implements the OA and GBD tive function involves the term f(x) for the
methods, and applies them to mixed-integer continuous variables and the charges ck that
dynamic optimization problems. It is difficult depend on the discrete choices in each dis-
to derive general conclusions on the efficiency junction k 2 K. The constraints g(x) 0 hold
and reliability of all these codes and their regardless of the discrete choice, and hjk(x) 0
corresponding methods, since no systematic are conditional constraints that hold when Yjk is
comparison has been made. However, one true in the j-th term of the k-th disjunction.
might anticipate that branch-and-bound codes The cost variables ck correspond to the fixed
are likely to perform better if the relaxation of charges, and are equal to gjk if the Boolean
the MINLP is tight. Decomposition methods variable Yjk is true. V(Y) are logical relations for
based on OA are likely to perform better if the the Boolean variables expressed as proposi-
NLP subproblems are relatively expensive to tional logic.
solve, while GBD can perform with some effi- Problem (107) can be reformulated as an
ciency if the MINLP is tight and there are many MINLP problem by replacing the Boolean var-
discrete variables. ECP methods tend to per- iables by binary variables yjk,
form well on mostly linear problems. P P
Min Z ¼ k2K j2J k g jk yjk þ f ðxÞ
s:t: gðxÞ 0
Logic-Based Optimization. Given difficulties
in the modeling and solution of mixed integer hjk ðxÞ M jk ð1 yjk Þ; j 2 J k ; k 2 K ðBMÞ
P ð108Þ
problems, the following major approaches based j2J k yjk ¼ 1; k 2 K
on logic-based techniques have emerged: gener-
Ay a
alized disjunctive programming (Problem 107)
[269], mixed-logic linear programming (MLLP) 0 x xU ; yjk 2 f0; 1g; j 2 J k ; k 2 K
[270], and constraint programming (CP) [271].
The motivations for this logic-based modeling where the disjunctions are replaced by “Big-M”
has been to facilitate the modeling, reduce the constraints which involve a parameter Mjk and
combinatorial search effort, and improve binary variables yjk. The propositional logic
the handling of nonlinearities. In this section statements V(Y) ¼ True are replaced by the
we mostly concentrate on generalized dis- linear constraints Ay A [275] and [276]. Here
junctive programming and provide a brief refer- we assume that x is a nonnegative variable with
ence to constraint programming. A general finite upper bound xU. An important issue in
Mathematics in Chemical Engineering 129
P P
Problem (108) is how to specify a valid value Min Z L ¼ k2K j2J k g jk ljk þ f ðxÞ
for the Big-M parameter Mjk. If the value is too s:t: gðxÞ 0
small, then feasible points may be cut off. If Mjk P P
x¼ vjk ; j2J k ljk ¼ 1; k 2 K ðCRPÞ
is too large, then the continuous relaxation j2J k
might be too loose and yield poor lower bounds. 0 x; vjk xU ; 0 ljk 1; j 2 J k ; k 2 K ð111Þ
Therefore, finding the smallest valid value for ljk hjk ðv =ljk Þ 0; j 2 J k ; k 2 K
jk
where hjk(x) are assumed to be convex and Constraint Programming. Constraint pro-
bounded over x. The convex hull relaxation of gramming (CP) [271, 272] is a relatively new
disjunction (Problem 109) [245] is given as modeling and solution paradigm that was orig-
follows: inally developed to solve feasibility problems,
but it has been extended to solve optimization
P P problems as well. Constraint programming is
x¼ j2J k vjk ; c¼ j2J ljk g jk
very expressive, as continuous, integer, and
jk ; Boolean variables are permitted and, moreover,
jk
0v ljk xU j 2 Jk
P variables can be indexed by other variables.
j2J k ljk ¼ 1; 0 ljk 1; j 2 J k ðCHÞ ð110Þ
Constraints can be expressed in algebraic form
ljk hjk ðvjk =ljk Þ 0; j 2 J k (e.g., h(x) 0), as disjunctions (e.g., [A1x b1]
_ A2x b2]), or as conditional logic statements
x; c; vjk 0; j 2 J k
(e.g., If g(x) 0 then r(x) 0). In addition, the
language can support special implicit functions
where vjk are disaggregated variables that are such as the all-different (x1, x2, . . . xn) con-
assigned to each term of the disjunction k 2 K, straint for assigning different values to the
and ljk are the weight factors that determine the integer variables x1, x2, . . . xn. The language
feasibility of the disjunctive term. Note that consists of Cþþ procedures, although the
when ljk is 1, then the j-th term in the k-th recent trend has been to provide higher level
disjunction is enforced and the other terms are languages such as OPL. Other commercial CP
ignored. The constraints ljk hjk ðvjk =ljk Þ are con- software packages include ILOG Solver [282],
vex if hjk(x) is convex [278, p. 160]. A formal CHIP [283], and ECLiPSe [284].
proof can be found in [245]. Note that the
convex hull (Eqs. 110) reduces to the result
by BALAS [279] if the constraints are linear. 10.6. Dynamic Optimization
Based on the convex hull relaxation Equations
(110), LEE and GROSSMANN [277] proposed Interest in dynamic simulation and optimiza-
the following convex relaxation program of tion of chemical processes has increased
Problem (107). significantly since the 1990s. Chemical
130 Mathematics in Chemical Engineering
processes are modeled dynamically using those that discretize the state and control pro-
differential-algebraic equations (DAEs), con- files (full discretization). Basically, the partially
sisting of differential equations that describe discretized problem can be solved either by
the dynamic behavior of the system, such as dynamic programming or by applying a non-
mass and energy balances, and algebraic equa- linear programming (NLP) strategy (direct-
tions that ensure physical and thermodynamic sequential). A basic characteristic of these
relations. Typical applications include control methods is that a feasible solution of the
and scheduling of batch processes; startup, DAE system, for given control values, is
upset, shut-down, and transient analysis; obtained by integration at every iteration of
safety studies; and the evaluation of control the NLP solver. The main advantage of these
schemes. We state a general differential- approaches is that, for the NLP solver, they
algebraic optimization Problem (112) as generate smaller discrete problems than full
follows: discretization methods.
Methods that fully discretize the continuous
time problem also apply NLP strategies to solve
Min F zðtf Þ; yðtf Þ; uðtf Þ; tf ; p
the discrete system and are known as direct-
s:t: F dz=dt; zðtÞ; uðtÞ; t; p ¼ 0; zð0Þ ¼ z0 simultaneous methods. These methods can use
different NLP and discretization techniques,
Gs ½zðts Þ; yðts Þ; uðts Þ; ts ; p ¼ 0 but the basic characteristic is that they solve
zL zðtÞ xU ð112Þ
the DAE system only once, at the optimum. In
addition, they have better stability properties
yL yðtÞ yU than partial discretization methods, especially
uL uðtÞ yU in the presence of unstable dynamic modes. On
pL p pU the other hand, the discretized optimization
problem is larger and requires large-scale
ttf tf tU
f
NLP solvers, such as SOCS, CONOPT, or
IPOPT.
where F is a scalar objective function at final With this classification we take into account
time tf, and F are DAE constraints, Gs addi- the degree of discretization used by the differ-
tional point conditions at times ts, z(t) differ- ent methods. Below we briefly present the
ential state profile vectors, y(t) algebraic state description of the variational methods, followed
profile vectors, u(t) control state profile vec- by methods that partially discretize the
tors, and p is a time-independent parameter dynamic optimization problem, and finally
vector. we consider full discretization methods for
We assume, without loss of generality, that Problem (112).
the index of the DAE system is one, consistent
initial conditions are available, and the objec- Variational Methods. These methods are
tive function is in the above Mayer form. based on the solution of the first-order neces-
Otherwise, it is easy to reformulate problems sary conditions for optimality that are obtained
to this form. Problem (112) can be solved either from Pontryagin’s maximum principle [285,
by the variational approach or by applying 286]. If we consider a version of Problem
some level of discretization that converts the (112) without bounds, the optimality conditions
original continuous time problem into a dis- are formulated as a set of DAEs:
crete problem. Early solution strategies, known
@Fðz; y; u; p; tÞ 0 @H @Fðz; y; u; p; tÞ
as indirect methods, were focused on solving l ¼ ¼ l (113a)
@z0 @z @z
the classical variational conditions for optimal-
ity. On the other hand, methods that discretize
Fðz; y; u; p; tÞ ¼ 0 (113b)
the original continuous time formulation can be
divided into two categories, according to the
Gf ðz; y; u; p; tf Þ ¼ 0 (113c)
level of discretization. Here we distinguish
between the methods that discretize only the
Gs ðz; y; u; p; ts Þ ¼ 0 (113d)
control profiles (partial discretization) and
Mathematics in Chemical Engineering 131
@H @Fðz; y; u; p; tÞ
¼ l¼0 (113e) control parametrization n. The sequential
@y @y
method is reliable when the system contains
only stable modes. If this is not the case, finding
@H @Fðz; y; u; p; tÞ
@u
¼
@u
l¼0 (113f) a feasible solution for a given set of control
parameters can be very difficult. The time
Ztf
horizon is divided into time stages and at
@Fðz; y; u; p; tÞ each stage the control variables are represented
l dt ¼ 0 (113g)
@p
0 with a piecewise constant, a piecewise linear, or
a polynomial approximation [287, 288]. A
where the Hamiltonian H is a scalar function of common practice is to represent the controls
the form H(t) ¼ F(z, y, u, p, y)Tl(t) and l(t) is a as a set of Lagrange interpolation polynomials.
vector of adjoint variables. Boundary and jump For the NLP solver, gradients of the objec-
conditions for the adjoint variables are given tive and constraint functions with respect to the
by: control parameters can be calculated with the
sensitivity equations of the DAE system, given
@F @F @Gf by:
lð t f Þ þ þ vf ¼ 0
@z0 @z @z
ð114Þ
@F
@Gs @F
þ
@F T @F @F @F T @zð0Þ
l ts þ vs ¼ l t sk 0 þ T sk þ T wk þ ¼ 0; sk ð0Þ ¼ k ¼ 1; . . . N q
@z0 @z @z0 s @z0 @z @y @qk @qk
ð115Þ
a solution to this boundary value problem. Equation (115), the cost of obtaining these sensi-
Normally, the state variables are given as initial tivities is directly proportional to Nq, the number
conditions, and the adjoint variables as final of decision variables in the NLP. Alternately,
conditions. This formulation leads to boundary gradients can be obtained by integration of the
value problems (BVPs) that can be solved by a adjoint Equations (113a, 113e, 113g) [282, 285,
number of standard methods including single 286] at a cost independent of the number of input
shooting, invariant embedding, multiple shoot- variables and proportional to the number of con-
ing, or some discretization method such as straints in the NLP.
collocation on finite elements or finite differ- Methods that are based on this approach
ences. Also the point conditions lead to an cannot treat directly the bounds on state vari-
additional calculation loop to determine the ables, because the state variables are not
multipliers vf and vs. On the other hand, included in the nonlinear programming prob-
when bound constraints are considered, the lem. Instead, most of the techniques for dealing
above conditions are augmented with addi- with inequality path constraints rely on defining
tional multipliers and associated complemen- a measure of the constraint violation over the
tarity conditions. Solving the resulting system entire horizon, and then penalizing it in the
leads to a combinatorial problem that is pro- objective function, or forcing it directly to zero
hibitively expensive except for small problems. through an end-point constraint [291]. Other
techniques approximate the constraint satisfac-
Partial Discretization. With partial discretiza- tion (constraint aggregation methods) by intro-
tion methods (also called sequential methods or ducing an exact penalty function [290, 292] or a
control vector parametrization), only the con- Kreisselmeier–Steinhauser function [292] into
trol variables are discretized. Given the initial the problem.
conditions and a given set of control parame- Finally, initial value solvers that handle path
ters, the DAE system is solved with a differen- constraints directly have been developed [288].
tial algebraic equation solver at each iteration. The main idea is to use an algorithm for con-
This produces the value of the objective func- strained dynamic simulation, so that any admis-
tion, which is used by a nonlinear programming sible combination of the control parameters
solver to find the optimal parameters in the produces an initial value problem that is
132 Mathematics in Chemical Engineering
feasible with respect to the path constraints. techniques to enforce feasibility, like the ones
The algorithm proceeds by detecting activation used in the sequential methods.
and deactivation of the constraints during the The resulting NLP is solved using SQP-type
solution, and solving the resulting high-index methods, as described above. At each SQP iter-
DAE system and their related sensitivities. ation, the DAEs are integrated in each stage and
objective and constraint gradients with respect to
Full Discretization. Full discretization meth- p, zi, and ni are obtained using sensitivity equa-
ods explicitly discretize all the variables of the tions, as in Problem (115). Compared to sequen-
DAE system and generate a large-scale non- tial methods, the NLP contains many more
linear programming problem that is usually variables, but efficient decompositions have
solved with a successive quadratic program- been proposed [294] and many of these calcula-
ming (SQP) algorithm. These methods follow a tions can be performed in parallel.
simultaneous approach (or infeasible path In collocation methods, the continuous time
approach); that is, the DAE system is not solved problem is transformed into an NLP by approx-
at each iteration; it is only solved at the opti- imating the profiles as a family of polynomials
mum point. Because of the size of the problem, on finite elements. Various polynomial repre-
special decomposition strategies are used to sentations are used in the literature, including
solve the NLP efficiently. Despite this charac- Lagrange interpolation polynomials for the
teristic, the simultaneous approach has advan- differential and algebraic profiles [295]. In
tages for problems with state variable (or path) [194] a Hermite–Simpson collocation form is
constraints and for systems where instabilities used, while CUTHRELL and BIEGLER [296] and
occur for a range of inputs. In addition, the TANARTKIT and BIEGLER [297] use a monomial
simultaneous approach can avoid intermediate basis for the differential profiles. All of these
solutions that may not exist, are difficult to representations stem from implicit Runge–
obtain, or require excessive computational Kutta formulae, and the monomial repre-
effort. There are mainly two different appro- sentation is recommended because of smaller
aches to discretize the state variables explicitly, condition numbers and smaller rounding
multiple shooting [293, 294] and collocation on errors. Control and algebraic profiles, on the
finite elements [184, 194, 295]. other hand, are approximated using Lagrange
With multiple shooting, time is discretized polynomials.
into P stages and control variables are parame- Discretizations of Problem (112) using col-
trized using a finite set of control parameters in location formulations lead to the largest NLP
each stage, as with partial discretization. The problems, but these can be solved efficiently
DAE system is solved on each stage, i ¼ using large-scale NLP solvers such as IPOPT
1, . . . P and the values of the state variables and by exploiting the structure of the colloca-
z(ti) are chosen as additional unknowns. In this tion equations. BIEGLER et al. [184] provide a
way a set of relaxed, decoupled initial value review of dynamic optimization methods using
problems (IVP) is obtained: simultaneous methods. These methods offer a
number of advantages for challenging dynamic
F dz=dt; zðtÞ; yðtÞ; ni ; p ¼ 0; optimization problems, which include:
t 2 ½ti1 ; ti ; zðti1 Þ ¼ zi ð116Þ Control variables can be discretized at the
ziþ1 zðti ; zi ; ni ; pÞ ¼ 0; i ¼ 1; . . . P 1 same level of accuracy as the differential
and algebraic state variables. The KKT
Note that continuity among stages is treated conditions of the discretized problem can
through equality constraints, so that the final be shown to be consistent with the varia-
solution satisfies the DAE system. With this tional conditions of Problem (116). Finite
approach, inequality constraints for states and elements allow for discontinuities in con-
controls can be imposed directly at the grid trol profiles.
points, but path constraints for the states may Collocation formulations allow problems
not be satisfied between grid points. This prob- with unstable modes to be handled in an
lem can be avoided by applying penalty efficient and well-conditioned manner.
Mathematics in Chemical Engineering 133
The NLP formulation inherits stability platforms are essential for the formulation task.
properties of boundary value solvers. These are classified into two broad areas: opti-
Moreover, an elementwise decomposition mization modeling platforms and simulation
has been developed that pins down platforms with optimization.
unstable modes in Problem (112).
Collocation formulations have been pro- Optimization modeling platforms provide gen-
posed with moving finite elements. This eral purpose interfaces for optimization algo-
allows the placement of elements both rithms and remove the need for the user to
for accurate breakpoint locations of con- interface to the solver directly. These plat-
trol profiles as well as accurate DAE forms allow the general formulation for all
solutions. problem classes discussed above with direct
interfaces to state of the art optimization
Dynamic optimization by collocation meth- codes. Three representative platforms are
ods has been used for a wide variety of process GAMS (General Algebraic Modeling Sys-
applications including batch process optimiza- tems), AMPL (A Mathematical Programming
tion, batch distillation, crystallization, dynamic Language), and AIMMS (Advanced Integrated
data reconciliation and parameter estimation, Multidimensional Modeling Software). All
nonlinear model predictive control, polymer three require problem-model input via a declar-
grade transitions and process changeovers, ative modeling language and provide exact
and reactor design and synthesis. A review of gradient and Hessian information through auto-
this approach can be found in [298]. matic differentiation strategies. Although pos-
sible, these platforms were not designed to
handle externally added procedural models.
10.7. Development of Optimization As a result, these platforms are best applied
Models on optimization models that can be developed
entirely within their modeling framework.
The most important aspect of a successful Nevertheless, these platforms are widely used
optimization study is the formulation of the for large-scale research and industrial applica-
optimization model. These models must reflect tions. In addition, the MATLAB platform
the real-world problem so that meaningful allows the flexible formulation of optimization
optimization results are obtained, and they models as well, although it currently has only
also must satisfy the properties of the problem limited capabilities for automatic differentia-
class. For instance, NLPs addressed by gradi- tion and limited optimization solvers. More
ent-based methods require functions that are information on these and other modeling
defined in the variable domain and have platforms can be found on the NEOS server
bounded and continuous first and second deriv- www-neos.mcs.anl.gov.
atives. In mixed integer problems, proper for-
mulations are also needed to yield good lower Simulation platforms with optimization are
bounds for efficient search. With increased often dedicated, application-specific modeling
understanding of optimization methods and tools to which optimization solvers have been
the development of efficient and reliable opti- interfaced. These lead to very useful optimiza-
mization codes, optimization practitioners now tion studies, but because they were not origi-
focus on the formulation of optimization nally designed for optimization models, they
models that are realistic, well-posed, and need to be used with some caution. In particu-
inexpensive to solve. Finally, convergence lar, most of these platforms do not provide exact
properties of NLP, MILP, and MINLP solvers derivatives to the optimization solver; often
require accurate first (and often second) deriv- they are approximated through finite differ-
atives from the optimization model. If these ence. In addition, the models themselves are
contain numerical errors (say, through finite constructed and calculated through numerical
difference approximations) then performance procedures, instead of through an open declar-
of these solvers can deteriorate considerably. ative language. Examples of these include
As a result of these characteristics, modeling widely used process simulators such as
134 Mathematics in Chemical Engineering
x2 f2
f
X f (X) Solutions dominated
x by f (x)
f(x)
x1 f1
Pareto optimal if there is no other solution y The convexity property, which is eligible for
such that f i ðxÞ < f i ðyÞ for all components i. optimization problems with one objective, is
Usually, weakly Pareto optimal solutions, also desirable for multicriteria problems. We
which are not Pareto optimal, are undesired, call a multicriteria optimization problem con-
because they are dominated, but some algo- vex if the feasible set X is convex and the
rithms for obtaining Pareto optimal solutions objective functions fi are convex. In that case
only guarantee weak Pareto optimality. But the Pareto set plus R kþ is convex. Otherwise, it
there might also be undesired Pareto optimal needs not to be convex and not even connected.
solutions. Assume two solutions x and y where x The concept of Pareto optimal solutions fits
is significantly worse than y with respect to fj the practical needs often better than breaking all
but at most only slightly better with respect to aspects down to one objective and several
any other objective, the deterioration may be constraints. Solving such a multicriteria opti-
not worth the improvement. So, the tradeoff mization problem is, however, much more
between these two components may be consid- involved than the solution of a single objective
ered undesired. Using a more general definition optimization problem: For an optimization
of dominance and Pareto optimality using so problem with a single objective, it is clear
called “ordering cones” one can exclude those that the problem is solved if an optimal solution
solutions. An equivalent formulation of “x is found. Finding all Pareto optimal solutions
dominates y” is the following: for a multicriteria optimization problem is in
general not possible nor usually necessary. The
f ðyÞ f ðxÞ 2 R kþ nf0g aim is to support the person or the group of
persons who have eventually to select a solution
The set R kþ :¼ fz 2 R k : z 0g is the usual by making the decision. That person or that
“ordering cone” for defining Pareto opti- group is referred to as decision-maker in the
mality. A point x is Pareto optimal if following. Therefore, two issues remain: 1)
f ðxÞ R kþ \ f ðxÞ ¼ u. By using a larger order- How to calculate Pareto optimal solutions;
ing cone such as: and 2) how to support the selection of a final
solution, that is, to make a decision.
R ke :¼ fz 2 R k : distðz; R kþ Þ e k z kg
f2 f2
Weight vector
Pareto points Pareto point
obtained by 1-norm
Pareto point
obtained by weighted
Tangent planes Euclidean norm
zref
f1 f1
X
gweighted sum ðxÞ :¼ wi f i ðxÞ (117)
for given barrieres ej with ji. Solving this
with nonnegative weights wi, see Figure 51. The problem for any i and any ej one obtains a
solution of weakly Pareto optimal solution and if the solu-
tion is unique it is Pareto optimal. Furthermore,
min gweighted sum ðxÞ s:t: x 2 X all Pareto optimal solutions can be obtained this
way.
is weakly Pareto optimal and Pareto optimal if
Another reformulation is based on norms
all weights are positive or if the solution is
(Fig. 53). Let zref 2k be a reference point, that
unique. If the original multicriteria optimiza-
is, a point in the objective space which is
tion problem is convex, then any Pareto optimal
desirable to be obtained but not necessarily
point is a solution of the weighted sum problem
possible to obtain. For any norm jj jj we can
for appropriately chosen weights. However,
define a new objective function ggoal ðxÞ :¼
otherwise this need not be the case.
jjf ðxÞ zref jj and solve:
Another approach is the e-constraint
method, see Figure 52. Here the new optimiza- min ggoal ðxÞ s:t: x 2 X: (118)
tion problem has the following shape
enumeration variable with few values attained, solutions. If this is combined with an intuitive
we can approximate the convex part for each navigation mechanism it gives the decision-
value of the enumeration variable and end up maker the feeling of having the complete Pareto
with a set of approximations. set at hand, even though only some point were
If the multicriteria optimization problem is actually calculated.
simply nonconvex, the Pareto optimal solu-
tions f ðxÞ can be approximated by hyperboxes. 10.9. Optimization under Uncertainty
This approach has the disadvantage that it
becomes very slow if the number of objectives Karl-Heinz K€ufer, Michael Bortz
increases.
10.9.1. Introduction
10.8.5. Other Approaches
In practical applications, models are affected by
A popular approach to obtain a set of Pareto
uncertainties of their parameters. These result
optimal solutions are population-based meta-
from the incomplete knowledge of model
heuristics. Here a set of solutions is modified
parameters, stemming from the model that is
iteratively according to some rules such that
only an approximation of the reality. Thus,
it gets closer to the set of Pareto optimal
model parameters (such as parameters in
solutions. Their main advantage is that they
reaction kinetics or equilibria) are often esti-
do not rely on differentiabity or continuity
mated from experiments. In addition, there are
assumptions, hence the objective evaluation
uncertainties in process control, affecting, for
can be a black box. One of the most well-
example, flow rates or feed compositions, but
known metaheuristics is the evaluationary
also in varying commodity or energy prices and
algorithm variant NSGA-II. In each iteration
in changing environmental conditions such as
the population is modified and extended and
atmospheric temperature and pressure. Since
afterwards a subset of solutions is selected as
these factors often randomly fluctuate, one
population of the next generation, respectively,
cannot simply fix them by one value but has
iteration. Possible criteria for the selection
to consider their “whole behavior”.
are non-dominance and the diversity or spread
The impact of uncertainties on optimal solu-
of the solutions to approximate the complete
tions is of special interest. Not only is it impor-
Pareto set.
tant to know how changes in the input affect the
It is also possible to use methods from
optimal settings, it is also desirable to compute
differential geometry to obtain Pareto optimal
settings that are robust to changes in the input
solutions: In the homotopy approach [303] the
(in this case, a local optimum can be better than
Pareto set is considered as a manifold which is
a global one, Fig. 55) or to compute settings that
iteratively sampled.
take into account the randomness of the input.
10.8.6. Decision Support The first aspect can be studied by using sensi-
tivity analysis, for the second, which is the
Having obtained a set of Pareto optimal solu- actual optimization under uncertainty, tech-
tions, it is still the question what to do with it. niques of robust and stochastic optimization
While it is sometimes enough to just calculate can be used. These tools are discussed in
one or more Pareto optimal solutions, often the more detail in the following.
decision-maker has to make a choice given a set In analogy to the general NLP Problem (52)
of solutions. And it is a challenge to support the we consider a parametric optimization problem:
decision in such a way that the decision-maker PðjÞ : minn f ðx; jÞs:t :gðx; jÞ 0;
x2R
can make use of all information available and (121)
not just select some solution after spending hðx; jÞ ¼ 0:
Interactions: Interactions occur when the Screening methods are sampling-based meth-
perturbation of two or more inputs simul- ods. Their objective is to identify input variables
taneously causes variation in the output which are contributing significantly to the output
greater than that of varying each of the variability in high-dimensional models, rather
inputs alone. than exactly quantifying sensitivity. They tend to
Given data: While in many cases the prac- have a relatively low computational cost com-
titioner has access to the model, in some pared to other approaches, and can be used in a
instances a sensitivity analysis must be preliminary analysis to weed out uninfluential
performed with “given data”, i.e. where variables before applying a more informative
the sample points (the values of the model analysis to the remaining set. One of the most
inputs for each run) cannot be chosen by the commonly used screening methods is the ele-
analyst. mentary effect method, where averaged finite
differences are calculated in order to estimate
There are a large number of approaches to the effect of uncertain model parameters.
perform a sensitivity analysis, many of which Variance-based methods are a class of prob-
have been developed to address one or more of abilistic approaches that quantify the input and
the constraints discussed above: output uncertainties as probability distributions
One of the simplest and most common and decompose the output variance into parts
approaches is that of changing one-factor-at- attributable to input variables and combinations
a-time (OFAT or OAT), to see what effect this of them. The sensitivity of the output to an input
produces on the output?. Thus, any change variable is therefore measured by the amount of
observed in the output will unambiguously be variance in the output caused by that input.
due to the single variable changed. The fixation Further approaches are emulators (also
of all other variables to their central or baseline known as meta- or surrogate models or response
values increases the comparability of the results surfaces) and Fourier amplitude sensitivity test
(all “effects“ are computed with reference to the (FAST).
same central point in space). However, this For a detailed introduction into sensitivity
approach does not fully explore the input space, analysis and its application to chemical engi-
since it does not take into account the simulta- neering amongst others see [304] and [305].
neous variation of input variables. Consequently, Many of the above mentioned methods, but
the OAT approach cannot detect the presence of also others are available in form of software.
interactions between input variables. Some references are:
Local methods involve taking the partial
derivative of the output with respect to an input SimLab: a free software for global sensi-
factor. Similar to OAT/OFAT, they do not tivity analysis [306]
attempt to fully explore the input space, since Sensitivity Analysis Excel Add-In: for sim-
they examine only small perturbations, typi- ple sample based sensitivity analysis runs
cally one variable at a time. in Excel [307]
Another simple but useful approach consists MUCM Project: an extensive resource for
in scatter plots: There the output variables are uncertainty and sensitivity analysis of
plotted against individual input variables, after computationally-demanding models [308]
(randomly) sampling the model over its input SALib: a sensitivity analysis library in
distributions. This approach can also deal with Python (Numpy), [309]
“given data” and gives a direct visual indication
of sensitivity.
Regression analysis, in the context of sensi- 10.9.3. Robust and Stochastic Optimization
tivity analysis, involves fitting a linear function
to the model response and using standardized Depending on the circumstances, there are
regression coefficients as measures of sensitiv- different approaches to treat optimization prob-
ity, which is only suitable when the model is lems depending on uncertain parameters:
linear. The advantages of regression analysis It is only known which values (discrete or
are its simplicity and low computational cost. continuous) a parameter can have.
Mathematics in Chemical Engineering 141
in this section to inequality constraint functions This formulation can be extended to a multi-
gðx; jÞ. stage setting by modeling the uncertainty as a
The simplest approach to deal with random- filtration process (successive realization of
ness in an optimization problem is to replace parameters). Under discrete distributions, this
the random variables or the whole (objective reduces to a scenario tree of parameter realiza-
or constraint) function by their means: tions. The compensation approach, however,
gðx; E½jÞ 0 or E½gðx; jÞ 0: But this leads requires that compensating actions exist at all
to the same problems as mentioned above. and can be reasonably modeled, which is not
Of the many concepts that have been dis- always possible.
cussed over the years, two concepts have been Two-stage linear stochastic programs can be
prevailed. efficiently solved for a discrete set of scenarios
in many cases. The solution methods build on
Two- and Multistage Stochastic Programs decomposition techniques of linear program-
with Recourse. The basic idea of recourse ming. Multi-stage integer and nonlinear sto-
(or compensation) relies on the possibility to chastic programs are often not or only with
adjust constraints in the system gðx; jÞ 0; great effort solvable for practically relevant
after observation of j by later compensating problem sizes.
actions. Accordingly in a two-stage program,
the set of variables splits into first stage deci-
Stochastic Programs with Chance/Probabilis-
sions x (to be fixed before realization of j) and
tic Constraints. If the emphasis is on the reli-
second stage or recourse decisions y (to be
ability of a system, it can be claimed that the
fixed after realization of j, see Fig. 56).
(system) constraints are not fulfilled for every
The adjustment of the constraint violation is
j 2 X but with high probability. That means
modeled by an inequality system ~ gðx; j; yÞ 0,
gðx; jÞ 0 is replaced by
connecting all three types of variables, and it
causes additional costs vðy; jÞ for the second
P½gðx; jÞ 0 a (126)
stage decisions. Of course, given x and j, y
should be chosen as to minimize second stage
costs among all feasible decisions. where a 2 ð0; 1 is some probability level. A
Summarizing, recourse models replace the constraint of this type is called chance or
original problem of minimizing the costs of first probabilistic constraint. Of course, the higher
stage decisions under stochastic inequalities by a the more reliable is the modeled system. On
a problem where the sum of first stage costs and the other hand, the set of feasible values is more
expected optimal second stage costs is mini- and more shrunken with a%, which increases
mized: the optimal value of the objective function, e.g.
the minimal costs, at the same time. The
min ff ðxÞ þ E½uðx; jÞg (125) extreme case a ¼ 1 is similar to the worst-
x2R n
case approach while the case a ¼ 1=2 relates
to the expected value approach. This makes
where
use of probabilistic constraints a good compro-
uðx; jÞ ¼ minn fvðy; jÞj~
gðx; j; yÞ 0g
mise between these two methods.
y2R
one, Equation (129) is nothing more than the The root-mean-square or quadratic mean is
scalarization of the (two) objective functions by
means of a weighted sum. By this, one can vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
u N
pffiffiffiffiffiffiffiffiffiffiffi uX
compare solutions for different l. Root mean square ¼ Eðy Þ ¼ t
2 y2i =N
In the same way, one can add the variability i¼1
The median is the middle value (or average of The value n 1 is used in the denominator
the two middle values) when the set of numbers because the deviations from the sample average
is arranged in increasing (or decreasing) order. must total zero:
The geometric mean is
X
n
ðyi y Þ ¼ 0
y G ¼ ðy1 y2 . . . yN Þ1=N i¼1
Mathematics in Chemical Engineering 145
Thus, knowing n1 values of yiy and the fact Table 12. Area under normal curve*
that there are n values automatically gives the n- Rz z2 =2
F ðzÞ ¼ p1ffiffiffiffi
2p
e dz
th value. Thus, only n1 degrees of freedom n 0
(y) is replaced by the sample mean y derived 0.0 0.0000 1.5 0.4332
from the data. 0.1 0.0398 1.6 0.4452
0.2 0.0793 1.7 0.4554
If data are taken consecutively, running
0.3 0.1179 1.8 0.4641
totals can be kept to permit calculation of the 0.4 0.1554 1.9 0.4713
mean and variance without retaining all the 0.5 0.1915 2.0 0.4772
data: 0.6 0.2257 2.1 0.4821
0.7 0.2580 2.2 0.4861
0.8 0.2881 2.3 0.4893
X
n X
n X
n
ðyi y Þ2 ¼ y21 2y yi þ ðy Þ2 0.9 0.3159 2.4 0.4918
i¼1 i¼1 i¼1 1.0 0.3413 2.5 0.4938
X
n 1.1 0.3643 2.7 0.4965
y ¼ yi =n 1.2 0.3849 3.0 0.4987
i¼1 1.3 0.4032 4.0 0.499968
1.4 0.4192 5.0 0.4999997
Thus, *
Table gives the probability F that a random variable will fall in the
shaded region of Figure 58. For a more complete table (in slightly
X
n X
n
different form), see [23, Table 26.1]. This table is obtained in
n; y21 ; and yi
i¼1 i¼1
Microsoft Excel with the function NORMDIST(z,0,1,1)-0.5.
Figure 57. Frequency of occurrence of different scores Figure 58. Area under normal curve
146 Mathematics in Chemical Engineering
0 FðxÞ 1
Fð1Þ ¼ 0; Fðþ1Þ ¼ 1
dFðxÞ
pðxÞ ¼
dx
where
Thus, variables yA and yC are distributed inde- Suppose Y, which is the sum or difference of
pendently. The joint distribution for two vari- two variables, is of interest:
ables is
Y ¼ y A
yB
pðyA ; yB Þ ¼ pðyA ÞpðyB jyA Þ
P
n s 2 ðYÞ ¼ a2i s 2 ðyi Þ
ðyAi y A ÞðyBi y B Þ i¼1
(131)
rðyA ; yB Þ ¼ i¼1 Xn X
n
ðn 1ÞsA sB þ2 ai aj Covðyi ; yj Þ
i¼1 j¼iþ1
If measurements are for independent, iden-
tically distributed observations, the errors are If the variables are uncorrelated and have the
independent and uncorrelated. Then y varies same variance, then
about E (y) with variance s 2/n, where n is the
!
number of observations in y . Thus if some- X
n
s 2 ðYÞ ¼ a2i s 2
thing is measured several times today and i¼1
every day, and the measurements have the
same distribution, the variance of the means
decreases with the number of samples in each This fact can be used to obtain more accurate
day’s measurement n. Of course, other factors cost estimates for the purchased cost of a
(weather, weekends) may cause the observa- chemical plant than is true for any one piece
tions on different days to be distributed of equipment. Suppose the plant is composed of
nonidentically. a number of heat exchangers, pumps, towers,
Mathematics in Chemical Engineering 149
pffiffiffi
and is then
ð40= nÞ%. Thus the standard
deviation of the cost for the entire plant is of significance, 0.95 ¼ 0.5 (for negative z) þ F
the standard deviation of each piece of equip- (for positive z). Thus, F ¼ 0.45 or z ¼ 1.645. In
ment divided by the square root of the number the two-sided test (see Fig. 61), if a single
of units. Under less restrictive conditions the sample is chosen and z < 1.96 or z > 1.96,
actual numbers change according to the above then this could happen with probability 0.05 if
equations, but the principle is the same. the hypothesis were true. This z would be
Suppose modifications are introduced into significantly different from the expected value
the manufacturing process. To determine if the (based on the chosen level of significance) and
modification causes a significant change, the the tendency would be to reject the hypothesis.
mean of some property could be measured If the value of z was between 1.96 and 1.96,
before and after the change; if these differ, the hypothesis would be accepted.
does it mean the process modification caused The same type of decisions can be made for
it, or could the change have happened by other distributions. Consider Student’s t-distri-
chance? This is a statistical decision. A hypoth- bution. At a 95 % level of confidence, with n ¼
esis H0 is defined; if it is true, action A must be 10 degrees of freedom, the t values are
2.228.
taken. The reverse hypothesis is H1; if this is Thus, the sample mean would be expected to be
true, action B must be taken. A correct decision between
is made if action A is taken when H0 is true or s
action B is taken when H1 is true. Taking action y
tc pffiffiffi
n
B when H0 is true is called a type I error,
whereas taking action A when H1 is true is with 95 % confidence. If the mean were outside
called a type II error. this interval, the hypothesis would be rejected.
The following test of hypothesis or test of The chi-square distribution is useful for
significance must be defined to determine if the examining the variance or standard deviation.
hypothesis is true. The level of significance is The statistic is defined as
the maximum probability that an error would be
accepted in the decision (i.e., rejecting the ns2
x2 ¼
hypothesis when it is actually true). Common s2
levels of significance are 0.05 and 0.01, and the ðy1 y Þ2 þ ðy2 y Þ2 þ . . . þ ðyn y Þ2
¼
test of significance can be either one or two s2
sided. If a sampled distribution is normal, then
the probability that the z score and the chi-square distribution is
Table 14. Percentage points of area under chi-square distribution with n degrees of freedom*
of freedom and a 95 % confidence level, the called y, with N2 samples. First, compute the
critical values of x2 are 0.025 and 0.975. Then standard error of the difference of the means:
pffiffiffi pffiffiffi vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
s n s n uP P
<s< u N1 N2
u ðxi xÞ2 þ ðyi y Þ2
x0:975 x0:025 ti¼1 i¼1 1 1
sD ¼ þ
N1 þ N2 2 N1 N2
or
pffiffiffi pffiffiffi
Next, compute the value of t
s n s n
<s<
20:5 3:25 x y
t¼
sD
with 95% confidence.
Tests are available to decide if two distribu- and evaluate the significance of t using
tions that have the same variance have different Student’s t-distribution for N1 þ N2 2 degrees
means [15, p. 465]. Let one distribution be of freedom.
called x, with N1 samples, and the other be If the samples have different variances, the
relevant statistic for the t-test is
x y
t ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
varðxÞ=N 1 þ varðyÞ=N 2
There is also an F-test to decide if two If each measured quantity has some variance,
distributions have significantly different vari- what is the variance in the thermal
ances. In this case, the ratio of variances is conductivity?
calculated: Suppose a model for Y depends on various
measurable quantities, y1, y2, . . . Suppose
varðxÞ
F¼
varðyÞ
several measurements are made of y1, y2, . . .
under seemingly identical conditions and sev-
where the variance of x is assumed to be larger. eral different values are obtained, with means E
Then, a table of values is used to determine (y1), E (y2), . . . and variances s 21 , s 22 , . . .
the significance of the ratio. The table Next suppose the errors are small and indepen-
[23, Table 26.9] is derived from the formula dent of one another. Then a change in Y is
[15, p. 169] related to changes in yi by
n
2 n1
QðFjn1 ; n2 Þ ¼ I n2 =ðn2 þn1 FÞ ; @Y @Y
2 2 dY ¼ dy þ dy þ . . .
@y1 1 @y2 2
where the right-hand side is an incomplete beta
function. The F table is given by the Microsoft If the changes are indeed small, the partial
Excel function FINV(fraction, nx, ny), where derivatives are constant among all the samples.
fraction is the fractional value ( 1) represent- Then the expected value of the change is
ing the upper percentage and nx and ny are the
N
X
degrees of freedom of the numerator and @Y
EðdYÞ ¼ Eðdyi Þ
denominator, respectively. i¼1
@yi
Example. For two sample variances with 8
degrees of freedom each, what limits will Naturally E (dyi) ¼ 0 by definition so that E
bracket their ratio with a midarea probability (dY) ¼ 0, too. However, since the errors are
of 90 %? FINV(0.95, 8, 8) 3.44. The 0.95 is independent of each other and the partial deriv-
used to get both sides to total 10%. Then atives are assumed constant because the errors
P½1=3:44 varðxÞ=varðyÞ 3:44 ¼ 0:90:
are small, the variances are given by Equation ()
[331, p. 550]
11.3. Error Analysis in Experiments N
X
@Y 2
s 2 ðdYÞ ¼ s 2i (132)
i¼1
@yi
Suppose a measurement of several quantities
is made and a formula or mathematical model Thus, the variance of the desired quantity Y can
is used to deduce some property of interest. be found. This gives an independent estimate of
For example, to measure the thermal conduc- the errors in measuring the quantity Y from the
tivity of a solid k, the heat flux q, the thickness errors in measuring each variable it depends
of the sample d, and the temperature differ- upon.
ence across the sample DT must be measured.
Each measurement has some error. The heat
flux q may be the rate of electrical heat input 11.4. Factorial Design of Experiments
Q_ divided by the area A, and both quantities and Analysis of Variance
are measured to some tolerance. The thickness
of the sample is measured with some accuracy, Statistically designed experiments consider, of
and the temperatures are probably measured course, the effect of primary variables, but
with a thermocouple, to some accuracy. These they also consider the effect of extraneous
measurements are combined, however, to variables, the interactions among variables,
obtain the thermal conductivity, and the error and a measure of the random error. Primary
in the thermal conductivity must be deter- variables are those whose effect must be deter-
mined. The formula is mined. These variables can be quantitative or
d qualitative. Quantitative variables are ones
k¼ Q
ADT that may be fit to a model to determine the
152 Mathematics in Chemical Engineering
model parameters. Curve fitting of this type is Table 15. Estimating the effect of four treatments
discused in Chapter 2. Qualitative variables Treatment 1 2 3 4
are ones whose effect needs to be known; no
attempt is made to quantify that effect other
than to assign possible errors or magnitudes.
Qualitative variables can be further subdivided
into type I variables, whose effect is deter-
mined directly, and type II variables, which
Treatment average, y t
contribute to performance variability, and Grand average, y
whose effect is averaged out. For example,
in studying the effect of several catalysts on
yield in a chemical reactor, each different type
that is, whether their means are different.
of catalyst would be a type I variable, because
The samples are assumed to have the same
its effect should be known. However, each
variance. The hypothesis is that the treatments
time the catalyst is prepared, the results are
are all the same, and the null hypothesis is that
slightly different, because of random varia-
they are different. Deducing the statistical
tions; thus, several batches may exist of what
validity of the hypothesis is done by an analysis
purports to be the same catalyst. The variabil-
of variance.
ity between batches is a type II variable.
The data for k ¼ 4 treatments are arranged
Because the ultimate use will require using
in Table 15. Each treatment has nt experi-
different batches, the overall effect including
ments, and the outcome of the i-th experiment
that variation should be known, because
with treatment t is called yti. The treatment
knowing the results from one batch of
average is
one catalyst precisely might not be represent-
ative of the results obtained from all batches of P
nt
the same catalyst. A randomized block design, i¼1
yti
incomplete block design, or Latin square y t ¼
nt
design, for example, all keep the effect of
experimental error in the blocked variables and the grand average is
from influencing the effect of the primary
variables. Other uncontrolled variables are P
k
nt y t X
k
accounted for by introducing randomization y ¼ t¼1
; N¼ nt
in parts of the experimental design. To study N t¼1
Table 17. Two-level factorial design with three variables chosen. CROPLEY [333] gives an example of
Run Variable 1 Variable 2 Variable 3
how to combine heuristics and statistical argu-
ments in application to kinetics mechanisms in
1 chemical engineering.
2 þ
3 þ
4 þ þ
5 þ References
6 þ þ
7 þ þ 1 D.E. Seborg, T.F. Edgar, D.A. Mellichamp: Process Dynamics
8 þ þ þ and Control, 2nd ed., John Wiley & Sons, New York 2004.
2 G. Forsyth, C.B. Moler: Computer Solution of Lineart Alge-
braic Systems, Prentice-Hall, Englewood Cliffs 1967.
3 B.A. Finlayson: Nonlinear Analysis in Chemical Engineering,
McGraw-Hill, New York 1980; reprinted, Ravenna Park, Seat-
and each effect is determined with the preci- tle 2003.
4 S.C. Eisenstat, M.H. Schultz, A.H. Sherman: “Algorithms and
sion of a fourfold replicated difference. The Data Structures for Sparse Symmetric Gaussian Elimination,”
advantage of a one-at-a-time experiment is the SIAM J. Sci. Stat. Comput. 2 (1981) 225–237.
gain in precision if the variables are additive 5 I.S. Duff: Direct Methods for Sparse Matrices, Charendon
and the measure of nonadditivity if it occurs Press, Oxford 1986.
6 H.S. Price, K.H. Coats, “Direct Methods in Reservoir Simu-
[338, p. 313]. lation,” Soc. Pet. Eng. J. 14 (1974) 295–308.
Interaction effects between variables 1 and 2 7 A. Bykat: “A Note on an Element Re-Ordering Scheme,” Int. J.
are obtained by comparing the difference Num. Methods Egn. 11 (1977) 194 –198.
between the results obtained with the high 8 M.R. Hesteness, E. Stiefel: “Methods of conjugate gradients
for solving linear systems,” J. Res. Nat. Bur. Stand 29 (1952)
and low value of 1 at the low value of 2 with 409–439.
the difference between the results obtained with 9 Y. Saad: Iterative Methods for Sparse Linear Systems, 2nd ed.,
the high and low value 1 at the high value of 2. Soc. Ind. Appl. Math., Philadelphia 2003.
The 12-interaction is 10 Y. Saad, M. Schultz: “GMRES: A Generalized Minimal Resid-
ual Algorithm for Solving Nonsymmetric Linear Systems.”
12 interaction ¼ ½ðy4 y3 þ y8 y7 Þ ½ðy2 y1 þ y6 y5 Þ=2 SIAM J. Sci. Statist. Comput. 7 (1986) 856–869.
11 http://mathworld.wolfram.com/
The key step is to determine the errors associ- GeneralizedMinimalResidualMethod.html.
ated with the effect of each variable and each 12 http://www.netlib.org/linalg/html_templates/Templates.html.
13 http://software.sandia.gov/.
interaction so that the significance can be deter- 14 E. Isaacson, H.B. Keller, Analysis of Numerical Methods,
mined. Thus, standard errors need to be J. Wiley and Sons, New York 1966.
assigned. This can be done by repeating the 15 W.H. Press, B.P. Flannery, S.A. Teukolsky, W.T. Vetterling:
experiments, but it can also be done by using Numerical Recipes, Cambridge University Press, Cambridge
1986.
higher order interactions (such as 123 interac- 16 R.W.H. Sargent: “A Review of Methods for Solving Non-linear
tions in a 24 factorial design). These are Algebraic Equations,” in R.S.H. Mah, W.D. Seider (eds.):
assumed negligible in their effect on the Foundations of Computer-Aided Chemical Process Design,
mean but can be used to estimate the standard American Institute of Chemical Engineers, New York 1981.
17 J.D. Seader: “Computer Modeling of Chemical Processes,”
error [338, pp. 319–328]. Then calculated AIChE Monogr. Ser. 81 (1985) no. 15.
effects that are large compared to the standard 18 http://software.sandia.gov/trilinos/packages/nox/loca_user.
error are considered important, whereas those html.
that are small compared to the standard error 19 N.R. Amundson: Mathematical Methods in Chemical Engi-
neering, Prentice-Hall, Englewood Cliffs, N.J. 1966.
are considered due to random variations and are 20 R.H. Perry, D. W Green: Perry’s Chemical Engineers’ Hand-
unimportant. book, 7th ed., McGraw-Hill, New York 1997.
In a fractional factorial design, only part of 21 D.S. Watkins: “Understanding the QR Algorithm,” SIAM Rev.
the possible experiments is performed. With k 24 (1982) 427–440.
22 G.F. Carey, K. Sepehrnoori: “Gershgorin Theory for Stiffness
variables, a factorial design requires 2k experi- and Stability of Evolution Systems and Convection-Diffusion,”
ments. When k is large, the number of experi- Comp. Meth. Appl. Mech. 22 (1980) 23–48.
ments can be large; for k ¼ 5, 25 ¼ 32. For k this 23 M. Abranowitz, I.A. Stegun: Handbook of Mathematical Func-
large, BOX et al. [331, p. 235] do a fractional tions, National Bureau of Standards, Washington, D.C. 1972.
24 J.C. Daubisse: “Some Results about Approximation of Func-
factorial design. In the fractional factorial tions of One or Two Variables by Sums of Exponentials, ” Int. J.
design with k ¼ 5, only 8 experiments are Num. Meth. Eng. 23 (1986) 1959–1967.
Mathematics in Chemical Engineering 155
25 O.C. McGehee: An Introduction to Complex Analysis, John 53 J.D. Lambert: Computational Methods in Ordinary Differential
Wiley & Sons, New York 2000. Equations, J. Wiley and Sons, New York 1973.
26 H.A. Priestley: Introduction to complex analysis, Oxford Uni- 54 N.R. Amundson: Mathematical Methods in Chemical Engi-
versity Press, New York 2003. neering, Prentice-Hall, Englewood Cliffs, N.J. 1966.
27 Y.K. Kwok: Applied complex variables for scientists 55 J.R. Rice: Numerical Methods, Software, and Analysis,
and engineers, Cambridge University Press, New York McGraw-Hill, New York 1983.
2002. 56 M.B. Bogacki, K. Alejski, J. Szymanowski: “The Fast Method
28 N. Asmar, G.C. Jones: Applied complex analysis with partial of the Solution of Reacting Distillation Problem,” Comput.
differential equations, Prentice Hall, Upper Saddle River, NJ, Chem. Eng. 13 (1989) 1081–1085.
2002. 57 C.W. Gear: Numerical Initial-Value Problems in Ordinary
29 M.J. Ablowitz, A.S. Fokas: Complex variables: Introduction Differential Equations, Prentice-Hall, Englewood Cliffs, N.J.
and applications, Cambridge University Press, New York 1971.
2003. 58 N.B. Ferguson, B.A. Finlayson: “Transient Modeling of a
30 J.W. Brown, R.V. Churchill: Complex variables and applica- Catalytic Converter to Reduce Nitric Oxide in Automobile
tions, 6th ed., McGraw-Hill, New York 1996; 7th ed. 2003. Exhaust,” AIChE J 20 (1974) 539–550.
31 W. Kaplan: Advanced calculus, 5th ed., Addison-Wesley, 59 W.F. Ramirez: Computational Methods for Process Simula-
Redwood City, Calif., 2003. tions, 2nd ed., Butterworth-Heinemann, Boston 1997.
32 E. Hille: Analytic Function Theory, Ginn and Co., Boston 60 U.M. Ascher, L.R. Petzold: Computer methods for ordinary
1959. differential equations and differential-algebraic equations,
33 R.V. Churchill: Operational Mathematics, McGraw-Hill, SIAM, Philadelphia 1998.
New York 1958. 61 K.E. Brenan, S.L. Campbell, L.R. Petzold: Numerical Solution
34 J.W. Brown, R.V. Churchill: Fourier Series and Boundary of Initial-Value Problems in Differential-Algebraic Equations,
Value Problems, 6th ed., McGraw-Hill, New York 2000. Elsevier, Amsterdam 1989.
35 R.V. Churchill: Operational Mathematics, 3rd ed., McGraw- 62 C.C. Pontelides, D. Gritsis, K.R. Morison, R.W.H. Sargent:
Hill, New York 1972. “The Mathematical Modelling of Transient Systems Using
36 B. Davies: Integral Transforms and Their Applications, 3rd ed., Differential-Algebraic Equations,” Comput. Chem. Eng. 12
Springer, Heidelberg 2002. (1988) 449–454.
37 D.G. Duffy: Transform Methods for Solving Partial Differen- 63 G.A. Byrne, P.R. Ponzi: “Differential-Algebraic Systems,
tial Equations, Chapman & Hall/CRC, New York 2004. Their Applications and Solutions,” Comput. Chem. Eng. 12
38 A. Varma, M. Morbidelli: Mathematical Methods in Chemical (1988) 377–382.
Engineering, Oxford, New York 1997. 64 L.F. Shampine, M.W. Reichelt: The MATLAB ODE Suite,
39 H. Bateman, Tables of Integral Transforms, vol. I, McGraw- SIAM J. Sci. Comp. 18 (1997) 122.
Hill, New York 1954. 65 M. Kubicek, M. Marek: Computational Methods in Bifurcation
40 H.F. Weinberger: A First Course in Partial Differential Equa- Theory and Dissipative Structures, Springer Verlag, Berlin
tions, Blaisdell, Waltham, Mass. 1965. 1983.
41 R.V. Churchill: Operational Mathematics, McGraw-Hill, New 66 T.F.C. Chan, H.B. Keller: “Arc-Length Continuation and
York 1958. Multi-Grid Techniques for Nonlinear Elliptic Eigenvalue Prob-
42 J.T. Hsu, J.S. Dranoff: “Numerical Inversion of Certain lems,” SIAM J. Sci. Stat. Comput. 3 (1982) 173–194.
Laplace Transforms by the Direct Application of Fast Fourier 67 M.F. Doherty, J.M. Ottino: “Chaos in Deterministic Systems:
Transform (FFT) Algorithm,” Comput. Chem. Eng. 11 (1987) Strange Attractors, Turbulence and Applications in Chemical
101–110. Engineering,” Chem. Eng. Sci. 43 (1988) 139–183.
43 E. Kreyzig: Advanced Engineering Mathematics, 9th ed., John 68 M.P. Allen, D.J. Tildesley: Computer Simulation of Liquids,
Wiley & Sons, New York 2006. Clarendon Press, Oxford 1989.
44 H.S. Carslaw, J.C. Jaeger: Conduction of Heat in Solids, 2nd 69 D. Frenkel, B. Smit: Understanding Molecular Simulation,
ed., Clarendon Press, Oxford–London 1959. Academic Press, San Diego 2002.
45 R.B. Bird, R.C. Armstrong, O. Hassager: Dynamics of Poly- 70 J.M. Haile: Molecular Dynamics Simulation, John Wiley &
meric Liquids, 2nd ed., Appendix A, Wiley-Interscience, New Sons, New York 1992.
York 1987. 71 A.R. Leach: Molecular Modelling: Principles and Applica-
46 R.S. Rivlin, J. Rat. Mech. Anal. 4 (1955) 681–702. tions, Prentice Hall, Englewood Cliffs, NJ, 2001.
47 N.R. Amundson: Mathematical Methods in Chemical Engi- 72 T. Schlick: Molecular Modeling and Simulations, Springer,
neering; Matrices and Their Application, Prentice-Hall, Engle- New York 2002.
wood Cliffs, N.J. 1966. 73 R.B. Bird, W.E. Stewart, E.N. Lightfoot: Transport Phe-
48 I.S. Sokolnikoff, E.S. Sokolnikoff: Higher Mathematics for nomena, 2nd ed., John Wiley & Sons, New York 2002.
Engineers and Physicists, McGraw-Hill, New York 1941. 74 R.B. Bird, R.C. Armstrong, O. Hassager: Dynamics of Poly-
49 R.C. Wrede, M.R. Spiegel: Schaum’s outline of theory and meric Liquids, 2nd ed., Wiley-Interscience, New York
problems of advanced calculus, 2nd ed, McGraw Hill, New 1987.
York 2006. 75 P.V. Danckwerts: “Continuous Flow Systems,” Chem. Eng. Sci.
50 P. M. Morse, H. Feshbach: Methods of Theoretical Physics, 2 (1953) 1–13.
McGraw-Hill, New York 1953. 76 J.F. Wehner, R. Wilhelm: “Boundary Conditions of Flow
51 B.A. Finlayson: The Method of Weighted Residuals and Vari- Reactor,” Chem. Eng. Sci. 6 (1956) 89–93.
ational Principles, Academic Press, New York 1972. 77 V. Hlavacek, H. Hofmann: “Modeling of Chemical Reactors-
52 G. Forsythe, M. Malcolm, C. Moler: Computer Methods for XVI-Steady-State Axial Heat and Mass Transfer in Tubular
Mathematical Computation, Prentice-Hall, Englewood Cliffs, Reactors. An Analysis of the Uniqueness of Solutions,” Chem.
N.J. 1977. Eng. Sci. 25 (1970) 173–185.
156 Mathematics in Chemical Engineering
78 B.A. Finlayson: Numerical Methods for Problems with Moving 100 R. Aris, N.R. Amundson: Mathematical Methods in Chemi-
Fronts, Ravenna Park Publishing Inc., Seattle 1990. cal Engineering, vol. 2, First-Order Partial Differential
79 E. Isaacson, H.B. Keller: Analysis of Numerical Methods, J. Equations with Applications, Prentice-Hall, Englewood
Wiley and Sons, New York 1966. Cliffs, NJ, 1973.
80 C. Lanczos: “Trigonometric Interpolation of Empirical and 101 R. Courant, D. Hilbert: Methods of Mathematical Physics, vol.
Analytical Functions,” J. Math. Phys. (Cambridge Mass.) 17 I and II, Intersicence, New York 1953, 1962.
(1938) 123–199. 102 P.M. Morse, H. Feshbach: Methods of Theoretical Physics, vol.
81 C. Lanczos: Applied Analysis, Prentice-Hall, Englewood I and II, McGraw-Hill, New York 1953.
Cliffs, N.J. 1956. 103 A.D. Polyanin: Handbook of Linear Partial Differential Equa-
82 J. Villadsen, W.E. Stewart: “Solution of Boundary-Value Prob- tions for Engineers and Scientists, Chapman and Hall/CRC,
lems by Orthogonal Collocation,” Chem. Eng. Sci. 22 (1967) Boca Raton, FL 2002.
1483–1501. 104 D. Ramkrishna, N.R. Amundson: Linear Operator Methods in
83 M.L. Michelsen, J. Villadsen: “Polynomial Solution of Differ- Chemical Engineering with Applications to Transport and
ential Equations” in R.S.H. Mah, W.D. Seider (eds.): Founda- Chemical Reaction Systems, Prentice Hall, Englewood Cliffs,
tions of Computer-Aided Chemical Process Design, NJ, 1985.
Engineering Foundation, New York 1981, pp. 341–368. 105 D.D. Joseph, M. Renardy, J.C. Saut: “Hyperbolicity and
84 W. E. Stewart, K. L. Levien, M. Morari: “Collocation Methods Change of Type in the Flow of Viscoelastic Fluids,” Arch.
in Distillation,” in A. W. Westerberg, H. H. Chien (eds.): Rational Mech. Anal. 87 (1985) 213–251.
Proceedings of the Second Int. Conf. on Foundations of Com- 106 H.K. Rhee, R. Aris, N.R. Amundson: First-Order Partial Differ-
puter-Aided Process Design, Computer Aids for Chemical ential Equations, Prentice-Hall, Englewood Cliffs, N.J. 1986.
Engineering Education (CACHE), Austin Texas, 1984, 107 B.A. Finlayson: Numerical Methods for Problems with Moving
pp. 535–569. Fronts, Ravenna Park Publishing Inc., Seattle 1990.
85 W.E. Stewart, K.L. Levien, M. Morari: “Simulation of Frac- 108 D.L. Book: Finite-Difference Techniques for Vectorized Fluid
tionation by Orthogonal Collocation,” Chem. Eng. Sci. 40 Dynamics Calculations, Springer Verlag, Berlin 1981.
(1985) 409–421. 109 G.A. Sod: Numerical Methods in Fluid Dynamics, Cambridge
86 C.L.E. Swartz, W.E. Stewart: “Finite-Element Steady State University Press, Cambridge 1985.
Simulation of Multiphase Distillation,” AIChE J. 33 (1987) 110 G.H. Xiu, J.L. Soares, P. Li, A.E. Rodriques: “Simulation of
1977–1985. five-step one-bed sorption-endanced reaction process,” AIChE
87 K. Alhumaizi, R. Henda, M. Soliman: “Numerical Analysis of J. 48 (2002) 2817–2832.
a reaction-diffusion-convection system,” Comp. Chem. Eng. 27 111 A. Malek, S. Farooq: “Study of a six-bed pressure swing
(2003) 579–594. adsorption process,” AIChE J. 43 (1997) 2509–2523.
88 E.F. Costa, P.L.C. Lage, E.C. Biscaia, Jr.: “On the numerical 112 R.J. LeVeque: Numerical Methods for Conservation Laws,
solution and optimization of styrene polymerization in tubular Birkh€auser, Basel 1992.
reactors,” Comp. Chem. Eng. 27 (2003) 1591–1604. 113 W.F. Ames: “Recent Developments in the Nonlinear Equations
89 J. Wang, R.G. Anthony, A. Akgerman: “Mathematical simula- of Transport Processes,” Ind. Eng. Chem. Fundam. 8 (1969)
tions of the performance of trickle bed and slurry reactors for 522–536.
methanol synthesis,” Comp. Chem. Eng. 29 (2005) 2474–2484. 114 W.F. Ames: Nonlinear Partial Differential Equations in Engi-
90 V.K.C. Lee, J.F. Porter, G. McKay, A.P. Mathews: “Application neering, Academic Press, New York 1965.
of solid-phase concentration-dependent HSDM to the acid dye 115 W.E. Schiesser: The Numerical Method of Lines, Academic
adsorption system,” AIChE J. 51 (2005) 323–332. Press, San Diego 1991.
91 C. deBoor, B. Swartz: “Collocation at Gaussian Points,” SIAM 116 D. Gottlieb, S.A. Orszag: Numerical Analysis of Spectral
J. Num. Anal. 10 (1973) 582–606. Methods: Theory and Applications, SIAM, Philadelphia, PA
92 R.F. Sincovec: “On the Solution of the Equations Arising From 1977.
Collocation With Cubic B-Splines,” Math. Comp. 26 (1972) 117 D.W. Peaceman: Fundamentals of Numerical Reservoir Simu-
893–895. lation, Elsevier, Amsterdam 1977.
93 U. Ascher, J. Christiansen, R.D. Russell: “A Collocation Solver 118 G. Juncu, R. Mihail: “Multigrid Solution of the Diffusion-
for Mixed-Order Systems of Boundary-Value Problems,” Convection-Reaction Equations which Describe the Mass and/
Math. Comp. 33 (1979) 659–679. or Heat Transfer from or to a Spherical Particle,” Comput.
94 R.D. Russell, J. Christiansen: “Adaptive Mesh Selection Strat- Chem. Eng. 13 (1989) 259–270.
egies for Solving Boundary-Value Problems,” SIAM J. Num. 119 G.R. Buchanan: Schaum’s Outline of Finite Element Analysis,”
Anal. 15 (1978) 59–80. McGraw-Hill, New York 1995.
95 P.G. Ciarlet, M.H. Schultz, R.S. Varga: “Nonlinear Boundary- 120 M.D. Gunzburger: Finite Element Methods for Viscous
Value Problems I. One Dimensional Problem,” Num. Math. 9 Incompressible Flows, Academic Press, San Diego 1989.
(1967) 394–430. 121 H. Kardestuncer, D.H. Norrie: Finite Element Handbook,
96 W.F. Ames: Numerical Methods for Partial Differential Equa- McGraw-Hill, New York 1987.
tions, 2nd ed., Academic Press, New York 1977. 122 J.N. Reddy, D.K. Gartling: The Finite Element Method in Heat
97 J.F. Botha, G.F. Pinder: Fundamental Concepts in The Numer- Transfer and Fluid Dynamics, 2nd ed., CRC Press, Boca Raton,
ical Solution of Differential Equations, Wiley-Interscience, FL 2000.
New York 1983. 123 O.C. Zienkiewicz, R.L. Taylor, J.Z. Zhu: The Finite Element
98 W.H. Press, B.P. Flanner, S.A. Teukolsky, W.T. Vetterling: Method: Its Basis & Fundamentals, vol. 1, 6th ed., Elsevier
Numerical Recipes, Cambridge Univ. Press, Cambridge Butterworth-Heinemann, Burlington, MA 2005.
1986. 124 O.C. Zienkiewicz, R.L. Taylor: The Finite Element Method,
99 H. Schlichting, Boundary Layer Theory, 4th ed. McGraw-Hill, Solid and Structural Mechanics, vol. 2, 5th ed., Butterworth-
New York 1960. Heinemann, Burlington, MA 2000.
Mathematics in Chemical Engineering 157
125 O.C. Zienkiewicz, R.L. Taylor: The Finite Element Method, 149 C.T.H. Baker: The Numerical Treatment of Integral Equations,
Fluid Dynamics, vol. 3, 5th ed., Butterworth-Heinemann, Clarendon Press, Oxford 1977.
Burlington, MA 2000. 150 L.M. Delves, J. Walsh (eds.): Numerical Solution of Integral
126 Z.C. Li: Combined Methods for Elliptic Equations with Sin- Equations, Clarendon Press, Oxford 1974.
gularities, Interfaces and Infinities, Kluwer Academic Pub- 151 P. Linz: Analytical and Numerical Methods for Volterra Equa-
lishers, Boston, MA, 1998. tions, SIAM Publications, Philadelphia 1985.
127 G.J. Fix, S. Gulati, G.I. Wakoff: “On the use of singular 152 M.A. Golberg (ed.): Numerical Solution of Integral Equations,
functions with finite element approximations,” J. Comp. Plenum Press, New York 1990.
Phys. 13 (1973) 209–238. 153 C. Corduneanu: Integral Equations and Applications, Cam-
128 N.M. Wigley: “On a method to subtract off a singularity at a bridge Univ. Press, Cambridge 1991.
corner for the Dirichlet or Neumann problem,” Math. Comp. 23 154 R. Kress: Linear Integral Equations, 2nd ed. Springer, Heidel-
(1968) 395–401. berg 1999.
129 H.Y. Hu, Z.C. Li, A.H.D. Cheng: “Radial Basis Collocation 155 P.K. Kythe P. Purl: Computational Methods for Linear Integral
Methods for Elliptic Boundary Value Problems,” Comp. Math. Equations, Birkh€auser, Basel 2002.
Applic. 50 (2005) 289–320. 156 G. Nagel, G. Kluge: “Non-Isothermal Multicomponent
130 S.J. Osher, R.P. Fedkiw:, Level Set Methods and Dynamic Adsorption Processes and Their Numerical Treatment by
Implicit Surfaces, Springer, New York 2002. Means of Integro-Differential Equations,” Comput. Chem.
131 M.W. Chang, B.A. Finlayson: “On the Proper Boundary Con- Eng. 13 (1989) 1025–1030.
dition for the Thermal Entry Problem,” Int. J. Num. Methods 157 P.L. Mills, S. Lai, M.P. Dudukovic, P.A. Ramachandran: “A
Eng. 15 (1980) 935–942. Numerical Study of Approximation Methods for Solution of
132 R. Peyret, T.D. Taylor: Computational Methods for Fluid Flow, Linear and Nonlinear Diffusion-Reaction Equations with Dis-
Springer Verlag, Berlin–Heidelberg–New York–Tokyo 1983. continuous Boundary Conditions,” Comput. Chem. Eng. 12
133 P.M. Gresho, S.T. Chan, C. Upson, R.L. Lee: “A Modified (1988) 37–53.
Finite Element Method for Solving the Time-Dependent, 158 D. Duffy: Green’s Functions with Applications, Chapman and
Incompressible Navier–Stokes Equations,” Int. J. Num. Hall/CRC, New York 2001.
Method. Fluids (1984) “Part 1. Theory”: 557–589; “Part 2: 159 I. Statgold: Green’s Functions and Boundary Value Problems,
Applications”: 619–640. 2nd ed., Interscience, New York 1997.
134 P.M. Gresho, R.L. Sani: Incompressible Flow and the Finite 160 B. Davies: Integral Transforms and Their Applications, 3rd ed.,
Element Method, Advection-Diffusion, vol. 1, John Wiley & Springer, New York 2002.
Sons, New York 1998. 161 J.M. Bownds, “Theory and performance of a subroutine for
135 P.M. Gresho, R.L. Sani: Incompressible Flow and the Finite solving Volterra integral equations,” J. Comput. 28 317–332
Element Method , Isothermal Laminar Flow, vol. 2, John Wiley (1982).
& Sons, New York 1998. 162 J.G. Blom, H. Brunner, “Algorithm 689; discretized collo-
136 J.A. Sethian, Level Set Methods, Cambridge University Press, cation and iterated collocation for nonlinear Volterra integral
Cambridge 1996. equations of the second kind,” ACM Trans. Math. Software
137 C.C. Lin, H. Lee, T. Lee, L.J. Weber: “A level set characteristic 17 (1991) 167–177.
Galerkin finite element method for free surface flows,” Int. J. 163 N.B. Ferguson, B.A. Finlayson: “Error Bounds for Approxi-
Num. Methods Fluids 49 (2005) 521–547. mate Solutions to Nonlinear Ordinary Differential Equations,”
138 S. Succi: The Lattice Boltzmann Equation for Fluid Dynamics AIChE J. 18 (1972) 1053–1059.
and Beyond, Oxford University Press, Oxford 2001. 164 R.S. Dixit, L.L. Taularidis: “Integral Method of Analysis of
139 M.C. Sukop, D.T. Thorne, Jr.: Lattice Boltzmann Modeling: An Fischer–Tropsch Synthesis Reactions in a Catalyst Pellet,”
Introduction for Geoscientists and Engineers, Springer, New Chem. Eng. Sci. 37 (1982) 539–544.
York 2006. 165 L. Lapidus, G.F. Pinder: Numerical Solution of Partial Differ-
140 S. Chen, S.G.D. Doolen: “Lattice Boltzmann method for fluid ential Equations in Science and Engineering, Wiley-Inter-
flows,” Annu. Rev. Fluid Mech. 30 (2001) 329–364. science, New York 1982.
141 H.T. Lau: Numerical Library in C for Scientists and Engineers, 166 C.K. Hsieh, H. Shang: “A Boundary Condition Dissection
CRC Press, Boca Raton, FL 1994. Method for the Solution of Boundary-Value Heat Conduction
142 Y.Y. Al-Jaymany, G. Brenner, P.O. Brum: “Comparative study Problems with Position-Dependent Convective Coefficients,”
of lattice-Boltzmann and finite volume methods for the simu- Num. Heat Trans. Part B: Fund. 16 (1989) 245–255.
lation of laminar flow through a 4:1 contraction,” Int. J. Num. 167 C.A. Brebbia, J. Dominguez: Boundary Elements–An Intro-
Methods Fluids 46 (2004) 903–920. ductory Course, 2nd ed., Computational Mechanics Publica-
143 R.L. Burden, R. L J. D. Faires, A.C. Reynolds: Numerical tions, Southhamtpon 1992.
Analysis, 8th ed., Brooks Cole, Pacific Grove, CA 2005. 168 J. Mackerle, C.A. Brebbia (eds.): Boundary Element Reference
144 S.C. Chapra, R.P. Canal: Numerical Methods for Engineers,” Book, Springer Verlag, Berlin 1988.
5th ed., McGraw-Hill, New York 2006. 169 L.T. Biegler, I.E. Grossmann, A.W. Westerberg: Systematic
145 H.T. Lau: Numerical Library in Java for Scientists and Engi- Methods for Chemical Process Design, Prentice-Hall Engle-
neers, CRC Press, Boca Raton, FL 2004. wood Cliffs, NJ 1997.
146 K.W. Morton, D.F. Mayers: Numerical Solution of Partial 170 I.E. Grossmann (ed.), “Global Optimization in Engineering
Differential Equations, Cambridge University Press, New Design”, Kluwer, Dordrecht 1996.
York 1994. 171 J. Kallrath, “Mixed Integer Optimization in the Chemical
147 A. Quarteroni, A. Valli: Numerical Approximation of Partial Process Industry: Experience, Potential and Future,” Trans.
Differential Equations, Springer, Heidelberg 1997. I. Chem. E., 78 (2000) Part A, 809–822.
148 F. Scheid: “Schaum’s Outline of Numerical Analysis,” 2nd ed., 172 O.L. Mangasarian: Nonlinear Programming, McGraw-Hill,
McGraw-Hill, New York 1989. New York 1969.
158 Mathematics in Chemical Engineering
173 G.B. Dantzig: Linear Programming and Extensions, Princeton 195 gPROMS User’s Guide, PSE Ltd., London 2002.
University Press, Princeton, N.J, 1963. 196 G.E.P. Box: “Evolutionary Operation: A Method for Increasing
174 S.J. Wright: Primal-Dual Interior Point Methods, SIAM, Industrial Productivity,” Applied Statistics 6 (1957) 81–101.
Philadelphia 1996. 197 R. Hooke, T.A. Jeeves: “Direct Search Solution of Numerical
175 F. Hillier, G.J. Lieberman: Introduction to Operations and Statistical Problems,” J. ACM 8 (1961) 212.
Research, Holden-Day, San Francisco, 1974. 198 M.J.D. Powell: “An Efficient Method for Finding the Mini-
176 T.F. Edgar, D.M. Himmelblau, L.S. Lasdon: Optimization of mum of a Function of Several Variables without Calculating
Chemical Processes, McGraw-Hill Inc., New York 2002. Derivatives, Comput. J. 7 (1964) 155.
177 K. Schittkowski: More Test Examples for Nonlinear Program- 199 J.A. Nelder, R. Mead: “A Simplex Method for Function
ming Codes, Lecture notes in economics and mathematical Minimization,” Computer Journal 7 (1965) 308.
systems no. 282, Springer-Verlag, Berlin 1987. 200 R. Luus, T.H.I. Jaakola: “Direct Search for Complex Systems,”
178 J. Nocedal, S.J. Wright: Numerical Optimization, Springer, AIChE J 19 (1973) 645–646.
New York 1999. 201 R. Goulcher, J.C. Long: “The Solution of Steady State Chemi-
179 R. Fletcher: Practical Optimization, John Wiley & Sons, Ltd., cal Engineering Optimization Problems using a Random
Chichester, UK 1987 Search Algorithm,” Comp. Chem. Eng. 2 (1978) 23.
180 A. W€achter, L.T. Biegler: “On the Implementation of an 202 J.R. Banga, W.D. Seider: in “Global Optimization of Chemical
Interior Point Filter Line Search Algorithm for Large-Scale Processes using Stochastic Algorithms,” C. Floudas, P. Parda-
Nonlinear Programming,” 106 Mathematical Programming los (eds.): State of the Art in Global Optimization, Kluwer,
(2006) no. 1, 25–57. Dordrecht 1996, p. 563.
181 C.V. Rao, J.B. Rawlings S. Wright: On the Application of 203 P.J.M. van Laarhoven, E.H.L. Aarts: Simulated Annealing:
Interior Point Methods to Model Predictive Control. J. Optim. Theory and Applications, Reidel Publishing, Dordrecht 1987.
Theory Appl. 99 (1998) 723. 204 J.H. Holland: Adaptations in Natural and Artificial Systems,
182 J. Albuquerque, V. Gopal, G. Staus, L.T. Biegler, B.E. Ydstie: University of Michigan Press, Ann Arbor 1975.
“Interior point SQP Strategies for Large-scale Structured Pro- 205 J.E. Dennis, V. Torczon: “Direct Search Methods on Parallel
cess Optimization Problems,” Comp. Chem. Eng. 23 (1997) Machines,” SIAM J. Opt. 1 (1991) 448.
283. 206 A.R. Conn, K. Scheinberg, P. Toint: “Recent Progress in
183 T. Jockenhoevel, L.T. Biegle, A. W€achter: “Dynamic Optimi- Unconstrained Nonlinear Optimization without Deriva-
zation of the Tennessee Eastman Process Using the tives,” Mathemtatical Programming, Series B, 79 (1997)
OptControlCentre,” Comp. Chem. Eng. 27 (2003) no. 11, no. 3, 397.
1513–1531. 207 M. Eldred: “DAKOTA: A Multilevel Parallel Object-Oriented
184 L.T. Biegler, A.M. Cervantes, A. W€achter: “Advances in Framework for Design Optimization, Parameter Estimation,
Simultaneous Strategies for Dynamic Process Optimization,” Uncertainty Quantification, and Sensitivity Analysis,” 2002,
Chemical Engineering Science 57 (2002) no. 4, 575–593. http://endo.sandia.gov/DAKOTA/software.html
185 C.D. Laird, L.T. Biegler, B. van Bloemen Waanders, R.A. 208 A.J. Booker et al.: “A Rigorous Framework for Optimization of
Bartlett: “Time Dependent Contaminant Source Determination Expensive Functions by Surrogates,” CRPC Technical Report
for Municipal Water Networks Using Large Scale Opti- 98739, Rice University, Huston TX, February 1998.
mization,” ASCE Journal of Water Resource Management 209 R. Horst, P.M. Tuy: “Global Optimization: Deterministic
and Planning 131 (2005) no. 2, 125. Approaches,” 3rd ed., Springer Verlag, Berlin 1996.
186 A.R. Conn, N. Gould, P. Toint: Trust Region Methods, SIAM, 210 I. Quesada, I.E. Grossmann: “A Global Optimization Algo-
Philadelphia 2000. rithm for Linear Fractional and Bilinear Programs,” Journal of
187 A. Drud: “CONOPT–A Large Scale GRG Code,” ORSA Global Optimization 6 (1995) no. 1, 39–76.
Journal on Computing 6 (1994) 207–216. 211 G.P. McCormick: “Computability of Global Solutions to Fac-
188 B.A. Murtagh, M.A. Saunders: MINOS 5.1 User’s Guide, torable Nonconvex Programs: Part I–Convex Underestimating
Technical Report SOL 83-20R, Stanford University, Palo Problems,” Mathematical Programming 10 (1976) 147–175.
Alto 1987. 212 H.S. Ryoo, N.V. Sahinidis: “Global Optimization of Noncon-
189 R.H. Byrd, M.E. Hribar, J. Nocedal: An Interior Point Algo- vex NLPs and MINLPs with Applications in Process Design,”
rithm for Large Scale Nonlinear Programming, SIAM J. Opt. 9 Comp. Chem. Eng. 19 (1995) no. 5, 551–566.
(1999) no. 4, 877. 213 M. Tawarmalani, N.V. Sahinidis: “Global Optimization of
190 R. Fletcher, N.I.M. Gould, S. Leyffer, Ph. L. Toint, A. Mixed Integer Nonlinear Programs: A Theoretical and Com-
Waechter: “Global Convergence of a Trust-region (SQP)-filter putational Study,” Mathematical Programming 99 (2004) no.
Algorithms for General Nonlinear Programming,” SIAM J. 3, 563–591.
Opt. 13 (2002) no. 3, 635–659. 214 C.S. Adjiman., I.P. Androulakis, C.A. Floudas: “Global Opti-
191 R.J. Vanderbei, D.F. Shanno: An Interior Point Algorithm for mization of MINLP Problems in Process Synthesis and Design.
Non-convex Nonlinear Programming. Technical Report SOR- Comp. Chem. Eng., 21 (1997) Suppl., S445–S450.
97-21, CEOR, Princeton University, Princeton, NJ, 1997. 215 C.S. Adjiman, I.P. Androulakis, C.A. Floudas: “Global Opti-
192 P.E. Gill, W. Murray, M. Wright: Practical Optimization, mization of Mixed-Integer Nonlinear Problems,” AIChE Jour-
Academic Press, New York 1981. nal, 46 (2000) no. 9, 1769–1797.
193 P.E. Gill, W. Murray, M.A. Saunders: User’s guide for SNOPT: 216 C.S. Adjiman, I.P. Androulakis, C.D. Maranas, C.A. Floudas:
A FORTRAN Package for Large-scale Nonlinear Program- “A Global Optimization Method, aBB, for Process Design,”
ming, Technical report, Department of Mathematics, Univer- Comp. Chem. Eng., 20 (1996) Suppl., S419–S424.
sity of California, San Diego 1998. 217 J.M. Zamora, I.E. Grossmann: “A Branch and Contract Algo-
194 J.T. Betts: Practical Methods for Optimal Control Using rithm for Problems with Concave Univariate, Bilinear and
Nonlinear Programming, Advances in Design and Control 3, Linear Fractional Terms,” Journal of Gobal Optimization 14
SIAM, Philadelphia 2001. (1999) no. 3, 217–249.
Mathematics in Chemical Engineering 159
218 E.M.B. Smith, C.C. Pantelides: “Global Optimization of 239 J.F. Benders: “Partitioning Procedures for Solving Mixed-
Nonconvex NLPs and MINLPs with Applications in Process variables Programming Problems,” Numeri. Math. 4 (1962)
Design,” Comp. Chem. Eng., 21 (1997) no. 1001, 238–252.
S791–S796. 240 G.L. Nemhauser, L.A. Wolsey: Integer and Combinatorial
219 J.E. Falk, R.M. Soland: “An Algorithm for Separable Non- Optimization, Wiley-Interscience, New York 1988.
convex Programming Problems,” Management Science 15 241 C. Barnhart, E.L. Johnson, G.L. Nemhauser, M.W.P. Savels-
(1969) 550–569. bergh, P.H. Vance: “Branch-and-price: Column Generation for
220 F.A. Al-Khayyal, J.E. Falk: “Jointly Constrained Biconvex Solving Huge Integer Programs,” Operations Research 46
Programming,” Mathematics of Operations Research 8 (1998) 316–329.
(1983) 273–286. 242 O.K. Gupta, V. Ravindran: “Branch and Bound Experiments in
221 F.A. Al-Khayyal: “Jointly Constrained Bilinear Programs and Convex Nonlinear Integer Programming,” Management Sci-
Related Problems: An Overview,” Computers and Mathemat- ence 31 (1985) no. 12, 1533–1546.
ics with Applications, 19 (1990) 53–62. 243 S. Nabar, L. Schrage: “Modeling and Solving Nonlinear
222 I. Quesada, I.E. Grossmann: “A Global Optimization Algo- Integer Programming Problems,” AIChE Meeting, Chicago
rithm for Heat Exchanger Networks,” Ind. Eng. Chem. Res. 32 1991.
(1993) 487–499. 244 B. Borchers, J.E. Mitchell: “An Improved Branch and Bound
223 H.D. Sherali, A. Alameddine: “A New Reformulation-Linear- Algorithm for Mixed Integer Nonlinear Programming,” Com-
ization Technique for Bilinear Programming Problems,” Jour- puters and Operations Research 21 (1994) 359–367.
nal of Global Optimization 2 (1992) 379–410. 245 R. Stubbs, S. Mehrotra: “A Branch-and-Cut Method for 0-1
224 H.D. Sherali, C.H. Tuncbilek: “A Reformulation-Convexi- Mixed Convex Programming,” Mathematical Programming 86
fication Approach for Solving Nonconvex Quadratic Pro- (1999) no. 3, 515–532.
gramming Problems,” Journal of Global Optimization 7 246 S. Leyffer: “Integrating SQP and Branch and Bound for Mixed
(1995) 1–31. Integer Noninear Programming,” Computational Optimization
225 C.A. Floudas: Deterministic Global Optimization: Theory, and Applications 18 (2001) 295–309.
Methods and Applications, Kluwer Academic Publishers, Dor- 247 A.M. Geoffrion: “Generalized Benders Decomposition,” Jour-
drecht, The Netherlands, 2000. nal of Optimization Theory and Applications 10 (1972) no. 4,
226 M. Tawarmalani, N. Sahinidis: Convexification and Global 237–260.
Optimization in Continuous and Mixed-Integer Nonlinear 248 M.A. Duran, I.E. Grossmann: “An Outer-Approximation Algo-
Programming: Theory, Algorithms, Software, and Applica- rithm for a Class of Mixed-integer Nonlinear Programs,” Math
tions, Kluwer Academic Publishers, Dordrecht 2002. Programming 36 (1986) 307.
227 R. Horst, H. Tuy: Global optimization: Deterministic 249 X. Yuan, S. Zhang, L. Piboleau, S. Domenech : “Une Methode
approaches. Springer-Verlag, Berlin 1993. d’optimisation Nonlineare en Variables Mixtes pour la Con-
228 A. Neumaier, O. Shcherbina, W. Huyer, T. Vinko: “A Compar- ception de Procedes,” RAIRO 22 (1988) 331.
ison of Complete Global Optimization Solvers,” Math. Pro- 250 R. Fletcher, S. Leyffer: “Solving Mixed Integer Nonlinear
gramming B 103 (2005) 335–356. Programs by Outer Approximation,” Math Programming 66
229 L. Lovasz A. Schrijver: “Cones of Matrices and Set-functions (1974) 327.
and 0-1 Optimization,” SIAM J. Opt. 1 (1991) 166–190. 251 I. Quesada, I.E. Grossmann: “An LP/NLP Based Branch and
230 H.D. Sherali, W.P. Adams: “A Hierarchy of Relaxations Bound Algorithm for Convex MINLP Optimization Prob-
Between the Continuous and Convex Hull Representations lems,” Comp. Chem. Eng. 16 (1992) 937–947.
for Zero-One Programming Problems,” SIAM J. Discrete 252 T. Westerlund, F. Pettersson: “A Cutting Plane Method for
Math. 3 (1990) no. 3, 411–430. Solving Convex MINLP Problems,” Comp. Chem. Eng. 19
231 E. Balas, S. Ceria, G. Cornuejols: “A Lift-and-Project Cutting (1995) S131–S136.
Plane Algorithm for Mixed 0-1 Programs,” Mathematical 253 O.E. Flippo, A.H.G.R. Kan: “Decomposition in General Math-
Programming 58 (1993) 295–324. ematical Programming,” Mathematical Programming 60
232 A.H. Land, A.G. Doig: “An Automatic Method for Solving (1993) 361–382.
Discrete Programming Problems,” Econometrica 28 (1960) 254 T.L. Magnanti, R.T. Wong: “Acclerated Benders Decomposi-
497–520. tion: Algorithm Enhancement and Model Selection Criteria,”
233 E. Balas: “An Additive Algorithm for Solving Linear Programs Operations Research 29 (1981) 464–484.
with Zero-One Variables,” Operations Research 13 (1965) 255 N.V. Sahinidis, I.E. Grossmann: “Convergence Properties of
517–546. Generalized Benders Decomposition,” Comp. Chem. Eng. 15
234 R.J. Dakin: “A Tree Search Algorithm for Mixed-Integer Pro- (1991) 481.
gramming Problems”, Computer Journal 8 (1965) 250–255. 256 J.E. Kelley Jr.: “The Cutting-Plane Method for Solving Convex
235 R.E. Gomory: “Outline of an Algorithm for Integer Solutions Programs,” J. SIAM 8 (1960) 703–712.
to Linear Programs,” Bulletin of the American Mathematics 257 S. Leyffer: “Deterministic Methods for Mixed-Integer Non-
Society 64 (1958) 275–278. linear Programming,” Ph.D. thesis, Department of Mathemat-
236 H.P. Crowder, E.L. Johnson, M.W. Padberg: “Solving Large- ics and Computer Science, University of Dundee, Dundee
Scale Zero-One Linear Programming Problems,” Operations 1993.
Research 31 (1983) 803–834. 258 M.S. Bazaraa, H.D. Sherali, C.M. Shetty: Nonlinear Program-
237 T.J. Van Roy, L.A. Wolsey: “Valid Inequalities for Mixed 0-1 ming, John Wiley & Sons, Inc., New York 1994.
Programs,” Discrete Applied Mathematics 14 (1986) 259 G.R. Kocis, I.E. Grossmann: “Relaxation Strategy for the
199–213. Structural Optimization of Process Flowsheets,” Ind. Eng.
238 E.L. Johnson, G.L. Nemhauser, M.W.P. Savelsbergh: “Progress Chem. Res. 26 (1987) 1869.
in Linear Programming Based Branch-and-Bound Algorithms: 260 I.E. Grossmann: “Mixed-Integer Optimization Techniques for
Exposition,” INFORMS Journal of Computing 12 (2000) 2–23. Algorithmic Process Synthesis”, Advances in Chemical
160 Mathematics in Chemical Engineering
Engineering, vol. 23, Process Synthesis, Academic Press, 282 ILOG, Gentilly Cedex, France 1999, www.ilog.com,
London 1996, pp. 171–246. 283 M. Dincbas, P. Van Hentenryck, H. Simonis, A. Aggoun, T.
261 E.M.B. Smith, C.C. Pantelides: “A Symbolic Reformulation/ Graf, F. Berthier: The Constraint Logic Programming Lan-
Spatial Branch and Bound Algorithm for the Global Optimi- guage CHIP, FGCS-88: Proceedings of International Con-
zation of Nonconvex MINLPs,” Comp. Chem. Eng. 23 (1999) ference on Fifth Generation Computer Systems, Tokyo,
457–478. 693–702.
262 P. Kesavan P.P.I. Barton: “Generalized Branch-and-cut Frame- 284 M. Wallace, S. Novello, J. Schimpf: “ECLiPSe: a Platform for
work for Mixed-integer Nonlinear Optimization Problems,” Constraint Logic Programming,” ICL Systems Journal 12
Comp. Chem. Eng. 24 (2000) 1361–1366. (1997) no.1, 159–200.
263 S. Lee I.E. Grossmann: “A Global Optimization Algorithm for 285 V. Pontryagin, V. Boltyanskii, R. Gamkrelidge, E. Mishchenko:
Nonconvex Generalized Disjunctive Programming and Appli- The Mathematical Theory of Optimal Processes, Interscience
cations to Process Systems,” Comp. Chem. Eng. 25 (2001) Publishers Inc., New York, NY, 1962.
1675–1697. 286 A.E. Bryson, Y.C. Ho: “Applied Optimal Control: Optimiza-
264 R. P€orn, T. Westerlund: “A Cutting Plane Method for Mini- tion, Estimation, and Control,” Ginn and Company, Waltham,
mizing Pseudo-convex Functions in the Mixed-integer Case,” MA, 1969.
Comp. Chem. Eng. 24 (2000) 2655–2665. 287 V. Vassiliadis, PhD Thesis, Imperial College, University of
265 J. Viswanathan, I.E. Grossmann: “A Combined Penalty Func- London 1993.
tion and Outer-Approximation Method for MINLP Opti- 288 W.F. Feehery, P.I. Barton: “Dynamic Optimization with State
mization,” Comp. Chem. Eng. 14 (1990) 769. Variable Path Constraints,” Comp. Chem. Eng., 22 (1998)
266 A. Brooke, D. Kendrick, A. Meeraus, R. Raman: GAMS–A 1241–1256.
User’s Guide, www.gams.com, 1998. 289 L. Hasdorff:. Gradient Optimization and Nonlinear Control,
267 N.V.A. Sahinidis, A. Baron: “A General Purpose Global Opti- Wiley-Interscience, New York, NY, 1976.
mization Software Package,” Journal of Global Optimization 8 290 R.W.H. Sargent, G.R. Sullivan: “Development of Feed
(1996) no.2, 201–205. Changeover Policies for Refinery Distillation Units,” Ind.
268 C.A. Schweiger, C.A. Floudas: “Process Synthesis, Design and Eng. Chem. Process Des. Dev. 18 (1979) 113–124.
Control: A Mixed Integer Optimal Control Framework,” Pro- 291 V. Vassiliadis, R.W.H. Sargent, C. Pantelides: “Solution of a
ceedings of DYCOPS-5 on Dynamics and Control of Process Class of Multistage Dynamic Optimization Problems,” I & EC
Systems, Corfu, Greece 1998, pp. 189–194. Research 33 (1994) 2123.
269 R. Raman, I.E. Grossmann: “Modelling and Computational 292 K.F. Bloss, L.T. Biegler, W.E. Schiesser: “Dynamic Process
Techniques for Logic Based Integer Programming,” Comp. Optimization through Adjoint Formulations and Constraint
Chem. Eng. 18 (1994) no.7, 563. Aggregation,”. Ind. Eng. Chem. Res. 38 (1999) 421–432.
270 J.N. Hooker, M.A. Osorio: “Mixed Logical.Linear Pro- 293 H.G. Bock, K.J. Plitt: A Multiple Shooting Algorithm for Direct
gramming,” Discrete Applied Mathematics 96-97 (1994) Solution of Optimal Control Problems, 9th IFAC World Con-
395–442. gress, Budapest 1984.
271 P.V. Hentenryck: Constraint Satisfaction in Logic Program- 294 D.B. Leineweber, H.G. Bock, J.P. Schl€oder, J.V. Gallitzend€or-
ming, MIT Press, Cambridge, MA, 1989. fer, A. Sch€afer, P. Jansohn: “A Boundary Value Problem
272 J.N. Hooker: Logic-Based Methods for Optimization: Combin- Approach to the Optimization of Chemical Processes
ing Optimization and Constraint Satisfaction, John Wiley & Described by DAE Models,” Computers & Chemical Engi-
Sons, New York 2000. neering, April 1997. (IWR-Preprint 97-14, Universit€at Heidel-
273 J.N. Hooker: Logic-Based Methods for Optimization, John berg, March 1997.
Wiley & Sons, New York 1999. 295 J.E. Cuthrell, L.T. Biegler: “On the Optimization of Differen-
274 E. Balas: “Disjunctive Programming,” Annals of Discrete tial- algebraic Process Systems,” AIChE Journal 33 (1987)
Mathematics 5 (1979) 3–51. 1257–1270.
275 H.P. Williams: Mathematical Building in Mathematical Pro- 296 A. Cervantes, L.T. Biegler: “Large-scale DAE Optimization
gramming, John Wiley, Chichester 1985. Using Simultaneous Nonlinear Programming Formulations.
276 R. Raman, I.E. Grossmann: “Relation between MILP Model- AIChE Journal 44 (1998) 1038.
ling and Logical Inference for Chemical Process Synthesis,” 297 P. Tanartkit, L.T. Biegler: “Stable Decomposition for Dynamic
Comp. Chem. Eng. 15 (1991) no. 2, 73. Optimization,” Ind. Eng. Chem. Res. 34 (1995) 1253–1266.
277 S. Lee, I.E. Grossmann: “New Algorithms for Nonlinear 298 S. Kameswaran, L.T. Biegler: “Simultaneous Dynamic Opti-
Generalized Disjunctive Programming,” Comp. Chem. Eng. mization Strategies: Recent Advances and Challenges,” Chem-
24 (2000) no. 9-10, 2125–2141. ical Process Control–7, to appear 2006.
278 J. Hiriart-Urruty, C. Lemarechal: Convex Analysis and Mini- 299 B. Bloemen Waanders, R. Bartlett, K. Long, P. Boggs, A.
mization Algorithms, Springer-Verlag, Berlin 1993. Salinger: “Large Scale Non-Linear Programming for PDE
279 E. Balas: “Disjunctive Programming and a Hierarchy of Constrained Optimization,” Sandia Technical Report
Relaxations for Discrete Optimization Problems,” SIAM J. SAND2002-3198, October 2002.
Alg. Disc. Meth. 6 (1985) 466–486. 300 I. Das, J.E. Dennis: “A closer look at drawbacks of minimizing
280 I.E. Grossmann, S. Lee: “Generalized Disjunctive Program- weighted sums of objectives for pareto set generation in multi-
ming: Nonlinear Convex Hull Relaxation and Algorithms,” criteria optimization problems”, Structural optimization 14
Computational Optimization and Applications 26 (2003) (1997) no. 1, 63–69.
83–100. 301 B. Schandl, K. Klamroth, M.M. Wiecek: “Norm-based approx-
281 S. Lee, I.E. Grossmann: “Logic-based Modeling and imation in bicriteria programming”, Computational Optimiza-
Solution of Nonlinear Discrete/Continuous Optimization tion and Applications 20 (2001) no. 1, 23–42.
Problems,” Annals of Operations Research 139 2005 302 M. Bortz, J. Burger, N. Asprion, S. Blagov, R. B€ottcher, U.
267–288. Nowak, A. Scheithauer, R. Welke, H. Hasse, K.-H. K€ufer:
Mathematics in Chemical Engineering 161
“Multi-criteria optimization in chemical process design and 321 A. Shapiro, D. Dentcheva, A. Ruszczy’nski: Lectures on
decision support by navigation on pareto sets”, Comp. Chem. Stochastic Programming–Modeling and Theory. 2nd ed.,
Eng. 60 (2014) 354–363. MOS-SIAM, Philadelphia 2014.
303 C. Hillermeier: “Nonlinear Multiobjective Optimization: A 322 N.V. Sahinidis: “Optimization under uncertainty: state-of-the-
Generalized Homotopy Approach”, International Series of art and opportunities”, Comp. & Chem. Eng. 28 (2004)
Numerical Mathematics, Springer, New York 2001. 971–983.
304 Saltelli, A., Ratto, M., Andres, T., Campolongo, F., Cariboni, 323 R. Henrion, P. Li, A. M€oller, M.C. Steinbach, M. Wendt, G.
J., Gatelli, D. Saisana, M., Tarantola, S.: Global Sensitivity Wozny: “Stochastic Optimization for Operating Chemical
Analysis - The Primer. Wiley, New York 2008. Processes under Uncertainty”, Online Optimization of Large
305 A. Saltelli, S. Tarantola, F. Campolongo, M. Ratto: Sensitivity Scale Systems, Springer-Verlag, Berlin Heidelberg 2001,
Analysis in Practice: A Guide to Assessing Scientific Models, pp. 457–478.
Wiley, New York 2004. 324 AIMMS Stochastic Programming Software: http://www
306 SimLab, http://ipsc.jrc.ec.europa.eu/?id=756 (accessed 1st .aimms.com/operations-research/mathematical-programming/
September 2015). stochastic-programming/ (accessed 1st September 2015).
307 Sensitivity Analysis Excel Add-In: http://www.life-cycle- 325 SLP-IOR: http://www.business.uzh.ch/professorships/qba/
costing.de/sensitivity_analysis/ (accessed 1st September 2015). publications/stochOpt.html (accessed 1st September 2015).
308 MUCM Project: http://mucm.ac.uk/index.html (accessed 1st 326 Stochastic Modeling Interface: http://www.coin-or.org/projects/
September 2015). Smi.xml (accessed 1st September 2015).
309 SALib: http://jdherman.github.io/SALib/ (accessed 1st Sep- 327 NEOS solver for stochastic linear programming: http://www
tember 2015). .neos-server.org/neos/solvers/index.html (accessed 1st Sep-
310 R. Hettich, K.O. Kortanek: “Semi-infinite programming– tember 2015).
Theory, methods, and applications”, SIAM Rev. 35 (1993) 328 AMPLDev SP Edition: http://www.optiriskindia.com/products/
380–429. ampl-dev-sp-edition-fort-sp (accessed 1st September 2015).
311 F. Guerra Vazquez, J.-J. R€uckmann, O. Stein, G. Still: 329 S. Zenios, A. Consiglio, S.S. Nielsen: “Mean-variance portfo-
“Generalized semi-infinite programming–A tutorial”, J. lio optimization”, Practical Financial Optimization–A Library
Comp. Appl. Math. 217 (2008) no. 2, 394–419. of GAMS Models, Wiley, New York 2009.
312 D. Bertsimas, D.B. Brown, C. Caramanis: “Theory and appli- 330 M. Suh, T. Lee: “Robust optimization method for the economic
cations of robust optimization”, SIAM Rev. 53 (2011) no. 3, term in chemical process design and planning”, Ind. Eng.
464–501. Chem. Res. 40 (2001) no. 25, 5950–5959.
313 A. Ben-Tal, L. El Ghaoui, A. Nemirovski: Robust Optimiza- 331 G.P. Box, J.S. Hunter, W.G. Hunter: Statistics for Experiment-
tion. Princeton University Press, 2009. ers: Design, Innovation, and Discovery, 2nd ed., John Wiley &
314 A. Ben-Tal, A. Nemirovski: “Robust optimization–method- Sons, New York 2005.
ology and applications”, Math. Prog. 92 (2002) no. 3, 332 D.C. Baird: Experimentation: An Introduction to Measurement
453–480. Theory and Experiment Design, 3rd ed., Prentice Hall, Engel-
315 AIMMS Robust Programming Software: http://www.aimms. wood Cliffs, NJ, 1995.
com/operations-research/mathematical-programming/robust- 333 J.B. Cropley: “Heuristic Approach to Comples Kinetics,”
optimization/ (accessed 1st September 2015). ACS Symp. Ser. 65 (1978) 292–302.
316 Robust Optimization with YALMIP: http://users.isy.liu.se/ 334 S. Lipschutz, J.J. Schiller, Jr: Schaum’s Outline of Theory and
johanl/yalmip/pmwiki.php?n=Tutorials.RobustOptimization Problems of Introduction to Probability and Statistics,
(accessed 1st September 2015). McGraw-Hill, New York 1988.
317 ROME - Robust Optimization Made Easy: http://www.robustopt 335 D.S. Moore, G.P. McCabe: Introduction to the Practice of
.com/ (accessed 1st September 2015). Statistics,” 4th ed., Freeman, New York 2003.
318 J.R. Birge, F. Louveaux: Introduction to Stochastic Program- 336 D.C. Montgomery, G.C. Runger: Applied Statistics and Prob-
ming. 2nd ed., Springer, Heidelberg 2011. ability for Engineers, 3 rd ed., John Wiley & Sons, New York
319 P. Kall, S.W. Wallace: Stochastic Programming–Numerical 2002.
Techniques and Engineering Applications, Springer, Heidel- 337 D.C. Montgomery, G.C. Runger, N.F. Hubele: Engineering
berg 1994. Statistics, 3rd ed., John Wiley & Sons, New York 2004.
320 A. Ruszczy’n ski, A. Shapiro (eds.): “Stochastic Pro- 338 G.E.P. Box, W.G. Hunter, J.S. Hunter: Statistics for Experi-
gramming”, Handbook in Operations Research and Manage- menters, John Wiley & Sons, New York 1978.
ment Science, Vol. 10, Elsevier, Amsterdam 2003.