Simulation Methods For Polymers PDF
Simulation Methods For Polymers PDF
Methods for
Polymers
edited by
Michael Ko telyanskii
Rudolph Technologies, Inc.
Flanders, New Jersey, U.S.A.
Doros N. Theodorou
National Technical University of Athens
Athens, Greece
m
MARCEL
MARCELDEKKER,
INC.
DEKKER
NEWYORK BASEL
0
Although great care has been taken to provide accurate and current information,
neither the author(s) nor the publisher, nor anyone else associated with this publication, shall be liable for any loss, damage, or liability directly or indirectly caused or
alleged to be caused by this book. The material contained herein is not intended to
provide specific advice or recommendations for any specific situation.
Trademark notice: Product or corporate names may be trademarks or registered
trademarks and are used only for identification and explanation without intent to
infringe.
Library of Congress Cataloging-in-Publication Data
A catalog record for this book is available from the Library of Congress.
ISBN: 0-8247-0247-6
This book is printed on acid-free paper.
Headquarters
Marcel Dekker, Inc., 270 Madison Avenue, New York, NY 10016, U.S.A.
tel: 212-696-9000; fax: 212-685-4540
Distribution and Customer Service
Marcel Dekker, Inc., Cimarron Road, Monticello, New York 12701, U.S.A.
tel: 800-228-1160; fax: 845-796-1772
Eastern Hemisphere Distribution
Marcel Dekker AG, Hutgasse 4, Postfach 812, CH-4001 Basel, Switzerland
tel: 41-61-260-6300; fax: 41-61-260-6333
World Wide Web
http://www.dekker.com
The publisher offers discounts on this book when ordered in bulk quantities. For
more information, write to Special Sales/Professional Marketing at the headquarters
address above.
Copyright 2004 by Marcel Dekker, Inc. All Rights Reserved.
Neither this book nor any part may be reproduced or transmitted in any form or
by any means, electronic or mechanical, including photocopying, microfilming, and
recording, or by any information storage and retrieval system, without permission in
writing from the publisher.
Current printing (last digit):
10 9 8 7 6 5 4 3 2 1
PRINTED IN THE UNITED STATES OF AMERICA
Preface
In the last 20 years, materials modeling and simulation has grown into an
essential component of research in the chemical and pharmaceutical industries. Because the field of process design is already quite mature, competitive
advances in chemical synthesis and separation operations occur primarily
through the development and use of material systems with tailored physicochemical characteristics (e.g., metallocene catalysts for polyolefin production, polymeric membranes offering a more favorable combination of
permeability and selectivity for separations, and environmentally acceptable
solvents with prescribed thermophysical properties). Furthermore, there is a
shift of emphasis from process to product design, which is intimately related
to materials behavior. With the current trend toward nanotechnology, the
scientist or engineer is often called on to develop new, often hierarchical
material structures with key characteristics in the 0.110-nm length scale, so
as to benefit from the unique mechanical, electronic, magnetic, optical, or
other properties that emerge at this scale. Materials that develop such structures through self-assembly processes, or modify their structure in response
to environmental conditions, are frequently sought.
Meeting these new design challenges calls for a quantitative understanding of structurepropertyprocessingperformance relations in materials. Precise development of this understanding is the main objective
of materials modeling and simulation. Along with novel experimental
techniques, which probe matter at an increasingly fine scale, and new
screening strategies, such as high-throughput experimentation, modeling has
become an indispensable tool in the development of new materials and
products. Synthetic polymers and biopolymers, either by themselves or in
iv
Preface
Preface
vi
Preface
Preface
vii
viii
Preface
Contents
Preface
Contributors
1.
Background
Michael Kotelyanskii and Doros N. Theodorou
I.
2.
3.
II.
4.
5.
III.
6.
Molecular Dynamics
Molecular Dynamics Simulations of Polymers
Vagelis A. Harmandaris and Vlasis G. Mavrantzas
Contents
IV.
7.
V.
8.
VI.
9.
Polymer Crystals
Contents
16.
17.
18.
xi
Contributors
xiv
Contributors
Engineering,
Polytechnic
Contributors
xv
Grant D. Smith
1
Background
MICHAEL KOTELYANSKII
Jersey, U.S.A.
DOROS N. THEODOROU
Athens, Greece
I.
Background
X
b
Vb
X
V
V
The chemical bonds are very stiff, deviations from the equilibrium bond
length usually being much less than a tenth of an A at room temperature.
Thus, the energy due to bond stretching is described by a harmonic
potential, proportional to the square of the deviation of the bond length
from its equilibrium value:
1
V b kb l l0 2
2
The bond angles are less stiff than the bond lengths; nevertheless, at usual
temperatures they normally do not deviate by more than a few degrees from
their equilibrium values. The bond angle potential is also fairly well
approximated by a harmonic potential
1
V k 0 2
2
FIG. 1 Bond angles (1, 2), and torsion angle () in the butane molecule. For
simplicity, only the carbon atoms are shown.
An cosn
*Full 360 degrees rotation around the bond does not change the energy. If some of the pendant
atoms happen to be the same, the period can be less than 360 degrees.
y
The reader should be careful about the conventions used in measuring torsion angles. The old
convention places 0 at trans, the newer convention places 0 in the energetically most
unfavorable conformation where all four carbon atoms are on the same plane, atoms 1 and 4
being on the same side of the middle bond between 2 and 3 [14]. In both conventions, is taken
as positive when it corresponds to rotation in a sense that would lead to unscrewing if the
middle bond were a right-handed screw.
Background
ij
12 6 !
V LJ rij 4
rij
rij
1 qi qj
4p0 rij
V q rij
organic compounds. Parameters for the same atoms may be quite different
between different force fields.
The way the potential energy function is split into various kinds of
interactions is quite arbitrary and may vary from one force field to another.
Two different force fields may give very similar energy for the same
molecule in the same conformation, but the individual contributions:
torsional, nonbonded, bond angles, etc. may be quite different in different
force fields. This is why it is generally dangerous to mix force fields, i.e.,
take, say, the Lennard-Jones parameters from one force field and bonded
parameters from another.
In polymer simulations energy is usually expressed in kcal/mol or kJ/mol,
distances in A and angles in radians or degrees, so the dimensions of
the constants are the followingkb: kcal/(mol A2), k: kcal/(mol rad2),
An: kcal/mol, : kcal/mol, rij, , l, and l0: A, charges qi: multiples of the
elementary charge e 1.6022 1019 C.
*These are the typical working frequency ranges for routine dielectric relaxation and rheological
experiments. More sophisticated techniques, such as NMR relaxation and inelastic neutron
scattering, are sensitive to higher frequencies up to 1012 Hz.
Background
Very large model systems, which are often necessary to track the
morphological characteristics responsible for the peculiar properties of
polymers, also present a great challenge in polymer simulations. Detailed
atomistic multichain polymer models used today seldom contain more than
a few thousands of atoms, although domain decomposition strategies on
parallel machines offer the possibility of going up to millions of atoms.
General computer simulation techniques useful for both low-molecular
weight molecules and polymers have been covered in some excellent
textbooks. We will briefly go over the fundamentals of these techniques in
this chapter. The major focus of the book, however, is on specific
approaches developed to handle the long time- and length scale challenges
of polymer problems. By necessity, these techniques must be hierarchical
ones, involving several levels of description, each designed to address a
specific window of time- and length scales. Going from one level to another
should entail a systematic coarse-graining procedure, wherein the detailed
information from more fundamental (shorter time- and length scale) levels
of modeling is built into some key parameters invoked by the more
macroscopic levels.
Analytical polymer theories usually address simplified models of polymer
chains, which capture universal features such as the chain topology,
flexibility, etc. Despite lacking many fine details, such models still manage to
predict, sometimes even quantitatively, many physical properties of polymer
networks, solutions, and melts. When a simplified general model turns out
to be capable of describing a particular polymer property or phenomenon,
this means that it successfully captures the relevant physics. Such results
provide valuable understanding of which particular features (e.g., chain
length, architecture, stiffness) are mainly responsible for a particular
property. Many polymer-specific effects and properties, such as rubber
elasticity, the viscoelastic rheological response of melts in the terminal
region, and overall molecular shapes in dilute solutions and melts in the
bulk and next to nonadsorbing surfaces, are similar for polymers of different
chemical structures. They can be rationalized with relatively simple arguments based on the decrease of entropy associated with chain extension and
on environment-dependent excluded volume interactions between segments.
Figure 2 shows the most popular simplified models of polymer chains
[18,19].
The freely-jointed chain model (Fig. 2a) is the simplest; it has been
described as the ideal gas of polymer physics, as all interactions between
the chain segments, except the chemical bonds connecting them, are
neglected. The model represents a chain as a sequence of steps, or rectilinear
statistical (Kuhn) segments, connected together at their ends. Each
statistical segment represents a sequence of several chemical bonds,
FIG. 2 The most common simplified models of polymer chains: freely-jointed chain
(a), bead-spring chain (b), rotational isomer model (c), and the typical torsional
potential V() for butane molecule (d).
depending on the conformational stiffness of the chain. The steps can point
in any direction, the directions of different steps being completely
independent, and the chain can intersect itself. Clearly, this model is
identical to the random flight model. It can easily be shown that the
probability to observe a certain value of the vector R connecting the two
ends of a sufficiently long freely-jointed chain obeys a Gaussian distribution
with zero average and mean squared value
hR2 i hR2 i NK b2K
for a chain of NK statistical segments, each of length bK. This is one of the
most fundamental results of polymer science; its derivation and corollaries
can be found in polymer textbooks [1820]. The radius of gyration tensor,
characterizing the average shape of the chain, and its properties can also
be calculated. Once the distribution is known, all the statistical and
thermodynamic properties of the ideal chain can be obtained.
A real chain whose conformation is governed only by local (effective)
interactions along its contour can be mapped onto a freely-jointed chain by
requiring that Eq. (8) is satisfied, and also that the contour length of the
freely-jointed chain, NKbK, matches that of the real chain at full extension.
Background
hR2 i C1 N2
10
N
X
!2 +
2
i1
N
1
X
N
X
hi j i
i1 ji1
N
X
hi i i
i1
N
1 X
N i
X
hi ik i N2
i1 k1
10
Background
11
N1
X
N kh1 k1 i=2
k1
N2 2N2
X
1 k=NCk
11
Here C(k) is a bond direction correlation function. As the number of monomers separating two bonds increases, their directions become uncorrelated,
and hence C(k) is a decreasing function of k. For a random walk, C(k) 0
for all k 1. For an unperturbed P
chain, C(k) falls exponentially with
increasing k at large k and C 1 2 kC(k).
The mean squared end-to-end distance of a linear unperturbed chain of N
monomers is proportional to N. As R is a sum of a large number of
independent random vectors, the probability density for the end-to-end
separation vector to have a certain value R is given by a Gaussian
distribution:
3
N R
2phR2 i
3=2
3 R2
exp
2 hR2 i
12
*
+
N
1 X
N
X
1
2
ri rj
N 12 i0 ji1
13
14
12
15
Background
13
A lnA
B lnB A B
NA
NB
RT
16
In the theory, A and B segments are envisioned as equal in volume, occupying sites on a lattice of coordination number Z. A is the volume fraction of
A-type segments in the system, B 1 A is the volume fraction of B-type
segments. (Z/kBT)("AB 1/2"AA 1/2"BB) is the Flory interaction
parameter, describing the interaction energy difference between unlikeand like- segment interactions occupying adjacent sites on the lattice. The
parameter is positive for most polymer mixtures. From Eq. (16) it is seen
that the entropy gain per mole of segments upon mixing polymers is smaller
than the corresponding gain for mixing monomers (NA NB 1) by a factor
equal to the inverse chain length. This explains the difficulty in mixing
different polymers. For NA NB N, Eq. (16) predicts an upper critical
solution temperature which scales proportionally with the chain length,
N assuming a value of 2 at the critical point. Traditionally, a strategy
14
for achieving miscible blends has been to use systems capable of developing
specific interactions, which can make negative.
As we discussed above, polymer chains are not rigid bodies, and there is
an entropy associated with the different polymer chain conformations.
Polymer chains usually contain very large numbers of monomers (thousands
or more), and, provided that environmental conditions ensure the same
excluded volume exponent , all polymers with the same chain topology are
similar at the length scales comparable to the chain radius of gyration.
Florys model is the simplest; it assumes that chain conformation does not
change upon mixing, and therefore there is no conformational entropy
contribution in Eq. (16). This contribution is, however, very important in
many cases and has to be taken into account to explain the thermodynamics
and phase behavior of mixtures and solutions where chain conformation is
different in different phases, for example in systems containing block
copolymers, chains next to surfaces or in restricted geometries, in the
swelling of polymer gels etc. [19,22,23].
Background
15
16
TA
T TV
17
Background
17
T To
T To C2
18
When the reference temperature To is chosen as the calorimetric glass temperature, Tg, the constants fall in relatively narrow ranges for all polymers:
C1 14 to 18 and C2 30 to 70K [26]. At the glass temperature, is on the
order of 1 min.
18
FIG. 3 State points at different moments in time t1<t2<t3 and trajectory in the
phase space {x, p} for two free particles traveling along the x axis with different
velocities x_ i vi and momenta pi mivi. Particle 2 goes to the left ( p2<0). The
crossed out trajectory is impossible.
Background
19
h2 2
^ @
rri V H
@t
2mi
19
y
The equivalence of these two averages is the essence of a very important system property called
ergodicity, which deserves a separate subsection further in this chapter.
zA hat is used here to emphasize that, in quantum mechanics, the Hamiltonian is an operator
acting on the function and not just a function of particle coordinates and momenta, as in
classical mechanics.
20
X p2
X p2
X1
X1
i
k
mi q_ 2i
mk v2k
2
2
2m
2m
i
k
i
i
k
k
Background
21
Hamilton equations
H q1 , q 2 , . . . , qN f , p1 , p2 , . . . , pN f K V
@qi @H q1 , q2 , . . . , qNf , p1 , p2 , . . . , pNf
@pi
@t
@H q1 , q2 , . . . , qNf , p1 , p2 , . . . , pNf
@pi
@t
@qi
20
Newton equations
dri
vi
dt
dvi
fi
mi
dt
@V r1 , r2 , . . . , rNf =3
fi
@ri
21
Lagrange equations
L q1 , q2 , . . . , qNf , q_ 1 , q_ 2 , . . . , q_ Nf K V
dqi
dt
d @L @L q1 , q2 , . . . , qN , q_ 1 , q_ 2 , . . . , q_N
dt @ q_ i
@qi
q_ i
22
In the above, Hq1 , q2 , . . . , qNf , p1 , p2 , . . . , pNf and L q1 , q2 , . . . , qNf , q_ 1 ,
q_ 2 , . . . , q_ Nf stand for the Hamiltonian and Lagrangian functions of the
system, respectively. As an illustration, let us consider these equations for
two examples. The first is a monatomic fluid, and the second is a harmonic
oscillator.
22
@V
rri V
@ri
23
dri
dt
N
X
1
i1
mi r_ 2i Vr1 , r2 , . . . , rN
d @L
d 2 ri @L
mi 2
rri V
dt @ r_ i
@ri
dt
24
N
X
p2i
Vr1 , r2 , . . . , rN
2mi
i1
@ri @H pi
@pi mi
@t
@pi
dvi @H
mi
rri V
@ri
@t
dt
25
In the special case where V(r1, r2 , . . . , rN) 0 for all configurations, the
above equations describe an ideal monatomic gas. The ideal gas particles
travel with their velocities not changing with time. This is because they do
not interact with each other or with anything else. The ideal gas is a limiting
case of a system with no interactions; it is a simple system, whose statistical
mechanics can be solved exactly. It is used as a reference to build more
complex systems, where interactions are non-negligible.
Background
23
V LJ
V0
V 0 ,
ku2
2
k
72
r20
26
@V
kx1 x2 r0
@x1
mx 2
@V
kx1 x2 r0
@x2
my i 0,
i 1, 2
mzi 0,
i 1, 2
27
24
Adding and subtracting the first and second equations gives the following
two equations:
mx 1 mx 2 0
mu 2ku
28
The first equation describes the motion of the dimer center of mass
xc (mx1 mx2)/2m. Its left hand side, mx 1 mx 2 , equals the total mass
2m times the center of mass acceleration. As long as there is no net external
force applied, the acceleration is zero. However, if the initial velocities of
the atoms v1(t 0) and v2(t 0) are such that the center of mass velocity
vc [mv1(t 0) mv2(t 0)]/2m is not zero, the value of vc does not change
with time, and the dimer travels as a whole with constant speed. The
second equation describes the motion of the atoms relative to each other,
as u measures the deviation of their separation from its equilibrium value
r0. The solution of the second equation,
ut A cos!t B sin!t
2k
!2
m
29
B! u_ t 0 v1 t 0 v2 t 0
30
Background
25
2
2
k
V jr1 r2 j r0 2
2
31
Here we used the harmonic approximation for potential energy (26) and set
the zero energy level at V 0 . Choosing the x axis along the line connecting the atoms, following the Lagrangian or Hamiltonian formalism, given
by the Eqs. (20) or (22), and remembering that, for simplicity, we consider
here the case m1 m2 m, one can arrive at the Newton equations of
motion (27). When Cartesian coordinates are used, Newton, Lagrange,
and Hamilton equations result in exactly the same differential equations
for the coordinates.
Newton equations, however, do require use of Cartesian coordinates,
while Lagrange and Hamilton equations do not. Let us see how the
Lagrange and Hamilton formulations can help us derive the more convenient set of Eqs. (28) from the very beginning. All we have to do is choose
the center of mass coordinate xc and the oscillation amplitude u as
generalized coordinates, describing the motion along the x direction in
which we are interested. We leave the Cartesian coordinates y1, z1 and y2, z2
for the other four degrees of freedom. The potential energy is simply
V ku2/2. The kinetic energy is a little more tricky:
K
2
2
2
m1 m2 vc 1 1
1 1 2
u_
2 m1 m2
2
32
Again, using the above equations in the Lagrange (20) or Hamilton (22)
formalism, working out through all the necessary derivatives, and using
m1 m2 m, one arrives directly at Eqs. (28) for u and xc. Figure 5 shows
the harmonic oscillators trajectories, corresponding to different initial
conditions.
The harmonic oscillator model is particularly useful in the study of
crystalline solids. Potential energy of the crystal can be expanded around the
state of mechanical equilibrium under given macroscopic dimensions
(quasiharmonic approximation). By diagonalizing the Hessian matrix
of second derivatives of the potential energy with respect to atomic
26
Background
27
33
LKV
NK
X
i gi
34
i1
*This technique is usually called the Method of Lagrange Multipliers in physical and mathematical literature.
28
Background
29
@V
1 @2 V 2
u
u Ou3
@q q0
2 @q2 q0
35
As q0 is an equilibrium point, the first derivative is zero and the leading term
is quadratic. Setting k @2 V=@q2 q0 , the equation of motion becomes
mu ku
ut A exp! t B exp! t
36
37
30
FIG. 6 Different potential energy extrema. Minimum (a), saddle point (b), and
maximum (c). Force f rV points along the direction of the fastest potential
energy decrease.
independent one-dimensional problems by a linear coordinate transformation u ! u . In terms of the new coordinates,
in the vicinity of the stationary
P
point, the potential energy V 12 i ki u 2i . When all ki are positive, the
potential energy has a minimum and the equilibrium is stable. When at least
one of ki is negative the equilibrium is unstable, as any small initial
perturbation along the direction u i will grow exponentially with time, taking
the system away from equilibrium. A maximum of the potential energy
occurs when all ki are negative.
Notice that, if we consider two trajectories starting from two points that
are close to each other in phase space in the vicinity of the stable equilibrium
point, the difference between the two trajectories will always remain of the
same order of magnitude as the initial difference; the two trajectories will
remain about as close to each other as their initial points are.
In contrast to this, if the two points are in the vicinity of an unstable
equilibrium point, the difference between the trajectories will grow
exponentially with time, and the trajectories will diverge. Thus, in the
vicinity of an unstable equilibrium point, the small uncertainty in the initial
condition will grow exponentially, with characteristic time of ! 1 . Any
attempt to predict the system motion for a time much longer than that will
fail. Notice, also, that the same argument applies to any two trajectories, as
soon as they are not confined to the immediate vicinity of the stable
equilibrium. If the system is unharmonic, as almost all systems are, and its
trajectories are not confined to the vicinity of a stable equilibrium, the
trajectories are exponentially divergent.
Background
31
lim
Aq , p d hAiobs
38
32
Background
33
Aq , p d
Z
hAji A d
0
39
H. Liouville Equation
Imagine that we start with a set of identical systems, whose states are
distributed in phase space according to a density distribution () at time
t 0, and let the systems move according to their equations of motion. The
ensemble constituted by the systems (points in phase space) evolves in time.
As the systems evolve, the density distribution () should, in general,
change with time. However, systems just move, no new systems are created,
and none of the systems is destroyed. Therefore, there should be a
conservation law for the probability density, similar to the continuity
equation (mass conservation) of hydrodynamics. The conservation law in
phase space is called the Liouville theorem. It states that the total time
34
@, t X
@
@
, t 0
p_ i
q_i
@t
@qi
@pi
i
40
@, t X @H @
@H @
, t 0
@t
@pi @qi @qi @pi
i
41
42
The (x) is a Dirac -function, which is nonzero only when its argument
equals zero. Here it selects those microscopic states that have total energy*
H equal to E0.
The above ensemble of systems with constant number of particles N,
occupying constant volume V, with the total energy E conserved, is called
the microcanonical, or (NVE) ensemble.
I.
Background
35
Q1
NVE H E0
X
H E0 A
hAi Q1
NVE
43
The summation over states here and in the previous equations is used for
the quantum case, where microscopic states are discrete and () has the
meaning of a probability. For one-component classical systems, on which
we have mainly focused so far, the sum should be replaced by the integral:
Z
X
1 1
d
!
N! h3N
d
N
Y
d 3 r i d 3 pi
44
i1
45
36
1 1
N! h3N
Z Y
N
K
V
d 3 pi exp
d 3 ri exp
kB T
kB T
i1
i1
Z Y
N
ex
Qig
NVT QNVT
47
Qig
NVT
48
V
d 3 ri exp
kB T
i1
Z Y
N
49
Background
37
ZNVT
Z Y
N
V
d ri exp
k
BT
i1
3
50
51
38
Z Y
N
PV
V
3
dV exp
d ri exp
kB T
kB T
i1
52
=kB T lnQ VT
53
Background
39
54
40
Background
41
A. Pressure
Pressure is one of the most important thermodynamic properties. According
to the previous subsection, to calculate pressure from a simulation we need a
function of the coordinates and momenta of the particles, whose average
equals the thermodynamic pressure.
In equilibrium thermodynamics, pressure is defined as a negative
derivative of the Helmholtz energy with respect to volume:
@A
1 @Q
1 @Z
kB T
kB T
@V N, T
Q @V
Z @V
Z
Z
@Z
@
. . . expV r1 , . . . , rN =kB T d 3 r1 d 3 rN
@V @V V
V
P
55
V
expV r1 , . . . , rN =kB T d s1 . . . d sN
@V @V
0
0
Z1
Z1
expV r1 , . . . , rN =kB T d 3 s1 . . . d 3 sN
NV N1
0
56
Z
Z1
VN 1
...
expV r1 , . . . , rN =kB T
kB T 0
0
"
#
N
X
@V
1 1 3
rk V
d s1 d 3 sN
@r
3
k
k1
N
P
Z
@V=@rk rk
N
1 k1
Z
VkB T
V
3
*The following argument is obviously valid for any complex container shape, but the calculations are much longer. In the case of two dimensions the factor 1/3 is to be replaced by 1/2.
42
57
58
The quantity @V=@ri rj can be taken as a definition of the force fkl on site
k due to site l, even in cases where V is not merely a sum of pairwise-additive
sitesite potentials [41]. Then,
*
+
1 X
N
X
1 N
f ij ri rj
PV NkB T
3 i1 ji1
59
Background
43
B. Chemical Potential
A very important theorem due to Widom [42] allows expression of the excess
chemical potential of a component in a fluid (i.e., the chemical potential of
the component minus the chemical potential it would have if it were an ideal
gas under the temperature and molar density with which it is present in the
fluid) as an ensemble average, computable through simulation. The Widom
theorem considers the virtual addition of a test particle (test molecule) of
the component of interest in the fluid at a random position, orientation, and
conformation. If V test is the potential energy change that would result from
this addition,
ex
i
V test
V ig
kB T ln exp
ln exp
kB T
kB T
60
where the first average is taken over all configurations of the fluid and all
positions, orientations, and conformations of the added test molecule, while
the second average is a weighted average of the intramolecular energy of one
molecule over all conformations adopted in the ideal gas state. Equation
(60) holds in the canonical ensemble. Analogous expressions for the NVE
and NPT ensembles have been derived [43]. Also, bias schemes that can deal
with the difficulties of randomly inserting long flexible molecules in the
simulated phase of interest have been developed [44].
The estimation of the chemical potential is also possible through virtual
deletion of a real particle (inverse Widom scheme), provided the bias
associated with creating a hole in the fluid as a result of the removal is
correctly accounted for [45].
C. Fluctuation Equations
The heat capacity, isothermal compressibility, and other thermodynamic
properties that correspond to second derivatives of the free energy, can be
calculated from the fluctuations in different ensembles [43,38,37]. Let us
consider the heat capacity in the NVT ensemble, as the simplest example.
With E hHi the internal energy, computed as an ensemble average of the
Hamiltonian, we have
@E
@hHi
cV
@T N, V
@T N, V
Z
1 1
H expH=kB T d
hHi
Q h3N N!
61
44
Z
@hHi
1 @Q 1
2
H expH=kB T d
@T
Q @T h3N N!
2
Z
1 1
H
expH=k
T
d
B
Q h3N N!
kB T 2
Z
@Q
1
H
1
hHiQ
expH=kB T
d
@T h3N N!
kB T 2
kB T 2
cV
hH2 hHi2 i H2
kB T 2
kB T 2
62
D. Structural Properties
One of the main goals of molecular simulations since their very early days
was to obtain information about structure together with the thermodynamic
properties, given the interaction potentials between the atoms.
The most important structural characteristics are the density distribution
and correlation functions. The n-particle distribution function g(n)(r1 , . . . , rn)
is defined so that ng(n)(r1 , . . . , rn), with being the mean density of particles
in a system, is the probability density for finding a particle at position r1, a
particle at position r2, . . . , a particle at position rn.
Particularly important is the pair distribution function, g(2)(r1, r2), or
simply g(r1, r2). In a homogeneous system this is only a function of the
interparticle distance vector r r2 r1. The quantity g(r) can be interpreted
as the mean local density of particles at position r relative to a particle that is
taken as reference. The quantity h(r) g(r) 1 is called the (total) pair
correlation function. If the system is also spherically symmetric, as in the
Background
45
63
Essentially, integration with the g(r) replaces sums over all pairs, and similar
expressions can be derived for averages of any pairwise function.
Calculating the pair distribution function in a simulation is straightforward; all we need to do is count the number of atom pairs separated by a
distance in the range from r to r r, and then normalize it. Usually, g(r) is
normalized by the number of pairs Nig(r, r r) that would be observed
in an ideal gas of the same density, so that, in the limit of large distances,
r ! 1, where correlations disappear, g(r) ! 1. A typical FORTRAN code
to calculate g(r) from 0 to rmax with the resolution deltar is given below.
The separation distance histogram is calculated for nbins rmax/
deltar bins, for a model of N particles in a box with sides boxx,
46
boxy, and boxz along the x, y, and z directions, respectively, with the
atomic coordinates stored in the arrays x(i),y(i),z(i).*
integer nbins, ibin, N
double precision hist(0:nbins)
double precision g(0:nbins)
double precision x(N),y(N),z(N)
double precision boxx,boxy,boxz,invbx,invby,invbz
double precision anorm, pi
double precision volume, rmax, deltar, dist
pidacos(1.0d0)
.........................................
volumeboxx*boxy*boxz
anorm N*(N1)/volume * 4.d0/3.d0 * pi
invbx1.d0/boxx
invby1.d0/boxy
invbz1.d0/boxz
nbinsrmax/deltar
do i1,N1
do ji1,N
dx x(i)x(j)
dx dx boxx*anint(dx*invbx)
dy y(i)y(j)
dy dy boxy*anint(dy*invby)
dz z(i)z(j)
dz dz boxz*anint(dz*invbz)
dist sqrt(dx*dxdy*dydz*dz)
ibin int(dist/deltar)
hist(ibin) hist(ibin)2.d0
enddo
enddo
do ibin1,nbins
r ibin*deltar
g(ibin)hist(ibin)/anorm/(r**3(r-deltar)**3)
enddo
.................
The above code fragment would be run for each configuration of the
simulation run, and the corresponding arrays would have to be averaged
over the configurations.
*This code is not optimized for performance. The periodic boundary conditions are taken into
account using the intrinsic function anint().
Background
47
48
64
65
The absolute value of this function is bounded between zero and one.
The self-diffusion coefficient D, for instance, is related to the velocity
autocorrelation function. In three dimensions
D
1
3
dthvi 0 vi ti
66
67
Here, , stand for the Cartesian components x, y, z of the corresponding quantityforce acting on the particle fi, particle coordinate ri, or
momentum pi.
Equations (66) and (67) are known as GreenKubo relations for the
transport coefficients.
Background
49
68
69
70
50
arrays a and b. To calculate the time correlation function CAB(t) one would
have to scan the array, picking up pairs of values a(l) and b(n),
calculating their product, and saving it in the appropriate location in the
array c(j), j|nl|, where the values of CAB( jt) are to be stored.
Afterwards, the accumulated quantities have to be normalized by the
number of a and b pairs used for each value of l. The following code
fragment implements the subroutine that takes the values of A(t) and B(t)
stored in the arrays a, b, and returns array c, containing the normalized
correlation function values for times from 0 to mcor t. It also calculates the
averages of A(t) and B(t) a_av and b_av and their root mean squared
deviations delta2a, delta2b, to obtain normalized correlation functions.
A and B are available for time slices 0 to m.
subroutine correl(a,b,c,a_av,b_av,delta2a,delta2b,mcor,m)
double precision a(0:m),b(0:m),c(0:mcor)
double precision a_av, b_av,a2av, b2av, delta2a,delta2b
double precision aux1,aux2, ai
integer mcor, count (0:mcor), maxt, m, i
a_av0.0d0
b_av0.0d0
a2av0.0d0
b2av0.0d0
c zero c and count
do i0,m
c(i)0.0d0
count(i)0.0d0
end do
c double loop through the arrays
do i0,m
aia(i)
a_ava_avai
b_avb_avb(i)
maxtmin(imcor,m)
do ji,maxt
1ji
count(1)count(1)1
c(1)c(1)ai*b(j)
end do
end do
c normalize averages
aux 1.0d0/dble(m1)
a_ava_av*aux
b_ava_av*aux
Background
51
a2ava2av*aux
b2avb2av*aux
delta2asqrt(a2av-a_av*a_av)
delta2bsqrt(a2av-b_av*b_av)
c normalize corr. function
aux1a_av*b_av
aux21.0d0/(delta2a*delta2b)
do i0,m
c(i)(c(i)/count(i) - a_av*b_av)*aux2
end do
return
end
For most of the modern high-speed CPUs, multiplication takes significantly fewer cycles than division. This is why division is replaced by
multiplication whenever possible. Also, the values unchanged inside the
loops are calculated in advance outside of the loop.
An alternative method for calculating the time correlation function,
especially useful when its spectrum is also required, involves the fast Fourier
transformation (FFT) algorithm and is based on the convolution theorem,
which is a general property of the Fourier transformation. According to the
convolution theorem, the Fourier transform of the correlation function CAB
equals the product of the Fourier transforms of the correlated functions:
^
B^
C^ AB A
71
52
states. It is this random nature of the technique which is responsible for its
name. The Monte Carlo method was developed in the Los Alamos National
Laboratory in New Mexico in the late 1940searly 1950s and its inventors
named it after Monte Carlo, capital of Monaco, a small country in Southern
Europe famous for its casinos.*
Let us first consider a very simple examplecalculation of the area of a
complex shape, such as the polygon in Fig. 7. The simplest way would be
to set up a mesh that is fine enough to provide the required accuracy, and to
calculate the number of mesh points inside the polygon. If the complex
shape happens to be a hole in the fence, this can be done by covering the
fence with chicken wire and counting the number of knots inside the hole
and the number of knots in a rectangular section ABCD of the fence which
completely surrounds the hole. The ratio of these two numbers times the
area of the rectangular section gives the area of the polygon.
Alternatively, one might draw a rectangle ABCD of known size
around the polygon and start throwing points inside ABCD at random
FIG. 7
Area calculations.
*Maybe, if the technique had been developed several years later, it could have been called Las
Vegas.
Background
53
M
1 X
Fx, yi
M i1
72
3
2
3=2
"
#
N
1
3 X
2
exp 2
ri ri1
b3
2b i2
73
r
i2 i ri1 can be thought of as an elastic energy
54
hR i
djrN r1 j2
74
The simplest Monte Carlo scheme for simulating this chain, as described
above, would generate random conformations , scattered with uniform
probability throughout configuration space; each time it would select the N
bead positions randomly within a three-dimensional domain whose
dimensions are much larger than b. Thus, it would calculate an average of
the product ()|rN r1|2 over all realizations of the chain.
Most of the randomly generated configurations, corresponding to
random positions of the monomers, however, will have a very small
weight, as the distances between connected beads will in most cases be much
longer than b. The probability density of each of these states () will then
be negligibly small, and so will be its contribution to the right hand side of
Eq. (74). This simplest Monte Carlo scheme will then be spending most of
the time calculating zeroes, almost never finding a configuration in which all
the distances between connected monomers are on the order of b, which
would contribute substantially to the average. This simple approach is very
inefficient.
There is a solution, though. Imagine that we have an algorithm that will
generate microscopic states using pseudorandom numbers, but with a
nonuniform probability density, visiting the relevant lower-energy states
more often and almost avoiding the high-energy states. The perfect choice
would be to sample the phase space with the probability density ()
directly. Then, ensemble averages can be calculated simply as averages over
the generated conformations:
M
1 X
R2
jrN r1 j2i
M i1
75
The core part of the Metropolis Monte Carlo simulation technique is the
algorithm to sample the microscopic states according to the required
ensemble probability distribution. The Monte Carlo algorithm simulates a
stochastic process (Markov chain) producing a sequence of points or states
in configuration space, in which the choice of the next state depends only on
the current state. The original Metropolis (or MR2T2) Monte Carlo
algorithm, proposed by Metropolis et al. [48], was for sampling the
Background
55
A. Microreversibility
Let us assume that the probability density for being at the phase space point
after completion of step k in the Metropolis algorithm is k
M . Let the
probability to go from one point to another 0 in one Monte Carlo step be
W( ! 0 ). This function W, describing how to proceed from one state to
the next, completely defines the algorithm. The following equation describes
the evolution of the probability density during our random process:*
X
0
0
kM k1
k1
M
M W !
0
0
k1
M W !
76
0
*This equation is very important in the theory of random processes; it has many dierent
names, of which master equation is most appropriate. We will not cover this subject in depth
but rather refer the interested reader to the excellent book by van Kampen [49].
y
Or an integral, if the phase space is continuous.
56
W ! 0 M 0
77
Equation (77) requires that the flux between any two points in phase space
be the same in both forward and reverse directions. We have to emphasize
that Eq. (77) is not equivalent to (76). It provides a sufficient but not necessary condition for the latter to hold. In simple terms, this means that,
if detailed balance is satisfied, Eq. (76) is always satisfied, too. However,
it may also be possible to make Eq. (76) true with M and W not
satisfying (77).
It is easy to check that detailed balance holds for the Metropolis
algorithm described above with M() equal to the canonical ensemble
distribution in configuration space. In the Metropolis algorithm,
W( ! 0 ) ( ! 0 )min(1, M(0 )/M()), where ( ! 0 ) is the probability of attempting a move from to 0 and Pacc( ! 0 )
min(1, M(0 )/M()) is the probability of accepting that move. In the
classical Metropolis algorithm, the attempt step is designed such that
( ! 0 ) (0 ! ). In the NVT ensemble, the acceptance probability is
min(1, exp[(V 0 V)/(kBT)]), as described above. One can easily verify that
the detailed balance equation (77) is satisfied under these conditions.
In recent years it has been realized that the sampling of configuration
space may be more efficient if a nonsymmetric attempt probability matrix
( ! 0 ) is utilized. In that case, the correct acceptance criterion for
ensuring microscopic reversibility is
0 ! M 0
Pacc ! 0 min 1,
! 0 M
78
Background
57
!
79
where J~ ! is the Jacobian of transformation from the set of coordinates ~ to the set of coordinates ; J~ ! times a differential volume in
~ -space gives the corresponding volume in -space. The presence of the
Jacobian takes care of the fact that equal volumes in ~ -space may not
correspond to equal volumes in -space.
Since both MC and molecular dynamics methods are available for
predicting the equilibrium properties of material systems, and since
molecular dynamics can provide kinetic in addition to thermodynamic
properties, one may ask why MC is used at all. The answer is that, by
appropriate selection of the moves employed, MC can achieve a sampling of
configurations many orders of magnitude more efficient than that provided
by MD. This means that the MC simulation can converge to wellequilibrated averages in a small fraction of the time that would be required
by MD. Another advantage of MC is that it can readily be adapted to
simulations with variable numbers of particles (e.g., grand canonical and
Gibbs ensemble simulations) and is therefore very convenient for phase
equilibrium calculations. Such simulation schemes will be described in the
remainder of this book.
58
r f=ms2 2_sr_ =s
80
The dots denote derivatives with respect to simulation time. The total
energy, including the kinetic and potential energy associated with the additional degree of freedom, is conserved. It can be shown that the original
system samples the canonical ensemble. Hoover [52] proposed a mathematically equivalent scheme in which the coupling parameter is a friction
Background
59
!
81
In the Hoover scheme, the distinction between real time and simulation time
disappears. Again here, f is the number of degrees of freedom and Qs
is a parameter that determines how strongly the system interacts with the
thermostat. It is chosen by trial and error for a particular system. If it
is too large, the system dynamics is strongly perturbed and is dominated by
the coupling to the thermostat. If Qs is too small, the coupling is weak,
and equilibration takes too long.
For the above techniques it is possible to prove that they reproduce the
canonical ensemble distribution.
There are, however, some not so rigorous but still very practical methods,
based on velocity rescaling. One obvious and the most crude method to
simulate constant temperature in MD would be to just rescale all the
velocities, so that the kinetic energy corresponds to the desired temperature
according to K 1/2 Nf kBT, where Nf is the number of degrees of freedom.
This, of course, is a caricature of the real NVT ensemble, where kinetic
energy is allowed to fluctuate and is not always constant. Even though this
approach is totally unsuitable for production runs, it is quite practical and is
often used to equilibrate the model at a required temperature, starting from
an initial configuration. Some gentler variations on this theme have been
suggested in [53]. There, the kinetic energy is not kept constant at every step.
Velocities are rescaled by a factor chi, which is dependent on the difference
between the current and desired kinetic energy values. Deviations of the
kinetic energy from the desired value relax to that value with a predefined
characteristic time:
1=2
t T
1
1
tT T
1 X
T
mi v2i
f =kB i
82
Here t is the time step and tT is a parameter that determines how strong the
perturbation is. Similar to the parameter Qs in the Nose and Hoover
60
*This is common for many stable integration algorithms. It is a well known fact that when the
most popular Verlet algorithm is used to numerically integrate the equations of motion, in very
long MD runs the system gets colder, due to the integration errors.
Background
61
undesired effects may occur [55]. One of the most common is the so-called
flying ice cube effect. The translational motion of the system as a whole
is often decoupled from other degrees of freedom, as potential energy does
not depend on it. There is no way for energy exchange between this degree
of freedom and the rest. Velocity rescaling can lead to the energy being
pumped into the overall translational motion, while being taken away
from other degrees of freedom. Nominally the kinetic energy of the system
is 1/2NkBT, but most of it is in the translational motion of the systems
center of mass. With no kinetic energy left in the other modes, the system
looks like a frozen cube flying in space at a rather high velocity. This can
be taken care of by correcting for the center of mass motion during the
run. If the initial velocities are chosen at random, there is always some
nonzero center of mass velocity component, which has to be subtracted in
the beginning of the run.* Overall center of mass motion can also
accumulate in very long MD runs due to the rounding errors.
With this said we see that, when using any one of the thermostats
described by Eqs. (80), (81), and (82), the time constant chosen should not
be too small to allow the energy to redistribute equally between the different
degrees of freedom after it is artificially pumped into the system or taken
from it by velocity rescaling.
83
62
particle and solvent system by projecting out the motion of the fast solvent
particles [56,57].
Multiplying both sides of Eq. (83) by p(0) and averaging over all
stochastic trajectories consistent with the same initial conditions, taking into
account the delta-correlation of the random force, we obtain:
d
hp0 pti hp0 pti
dt
hp0 pti hp2 i expt
84
1
m2
dt0
0
hp2 i
m2
2
hp i
2 2
m
2
dt0
0
t0
t
0
dt expt
0
t0
expt00 dt00
hp2 i t
hp2 i 1
2
1 expt
m2
m2 2
85
In the limit of long times (t 1), invoking the equipartition theorem for
kinetic energy,
lim hrt r02 i 2
t!1
hp2 i t
kB T t
6
6Dt
m2
m
86
[compare Eq. (68)]. So, the diffusion coefficient of the particle D is related
to the friction constant as
D
kB T
m
87
Background
63
Equation (87) is an Einstein relation for Brownian motion. There is a connection between the friction factor and the quantity Ro, measuring the
mean squared magnitude of the random force. Solving Eq. (83) just like a
deterministic ordinary differential equation gives
Z
expt0 f r t0 dt0
pt p0 expt expt
88
hp ti hp 0i exp2t 2 exp2t
expt0 f r t0 p0 dt0
dt0
exp2t
dt00 expt0 t00 f r t0 f r t00
89
f r t0 p0 0 and f r t0 f r t00 Ro t0 t00 ,
Ro
1 exp2t
2
90
As t ! 1, the particle will become thermally equilibrated with the bath and
so, by equipartition, limt!1 hp2 i 3mkB T. We have already
used
this in
deriving (86). On the other hand, Eq. (90) gives limt!1 p2 Ro =2.
Equating,
Ro 6mkB Thf r t f r t0 i 6mkB Tt t0
91
92
Let us try to derive an equation for the evolution in time of the phase
space probability density (p, r, t) for a random process generated by
64
numerical integration of the Langevin Eq. (92) with some time step t. As
in the Monte Carlo master Eq. (76), we should write down the balance of
the incoming and outgoing fluxes for the phase space point {p, r} after
the time increment t, taking into account that the coordinate increment
equals r (p/m)t [59] and the momentum increment equals p
p, with
p the momentum increment due to the random
p rVt
force. P(p) is the probability density of the momentum increment p
Z
p, r, t t p, r, t d 3 p p, r, t Pp
Z
d 3 p p p, r r, t Pp
93
The first two terms on the right handR side cancel each other, as (p, r, t) can
be pulled out from the integral and d3pP(p) 1. Expanding around
{p, r} under the integral of the third term on the right hand side,
Z
p, r, t t
d 3 pPpp, r, t
X
i
X @ Z
@pi
X
pi @
m @ri
d 3 pPp
d 3 pPppi
1
2
i, j
@2
@pi @pj
d 3 pPppi pj
Z
@ 2 pi pj
d 3 pPp
@ri @rj m2
Z
@ 2 pi
3
t
d pPppj
@ri @pj m
t2
94
The indices i, j run over all coordinate directions. Taking into account the
pi 0 and the lack
following relationships, which are a consequence of h
of correlation between the random force and momentum or position:
Z
Z
d 3 pPppi tpi ri V
p2 iij
d 3 pPppi pj t2 pi ri Vpj rj V 1=3h
95
Background
65
we obtain
2
p, r, t t p, r, t t4
X pi @
i
m @ri
X
i
3
X @
@
ri V
pi 5
@pi
@pi
i
1 X @2 2
h p i
6 i @p2i
X pi @ @
1
@ pj
pj rj V
t2
2
@rj m
m @ri @pj
ij
96
p2 i 6tmkB T fluctuationIn the limit of t ! 0, using the fact that h
dissipation theorem, compare Eq. (91), Eq. (96) reduces to:
X pi @ X
X @2
@p, r, t
@ X @
ri V
pi mkB T
@t
@pi
@pi
m @ri
@p2i
i
i
i
i
97
which is a FokkerPlanck equation [59]. By direct substitution one can
check that the Boltzmann distribution
2
p
V =kB T
p, r, t Const: exp
2m
is its stationary solution. This means that a Brownian dynamics method
solving the Langevin equation reproduces the canonical ensemble. Notice,
however, that such numerical schemes are approximate, with accuracy on
the order of (t2). This error is called discretization error. To obtain exact
canonical ensemble averages from a Brownian dynamics simulation it is
necessary to run simulations with different time steps and to extrapolate
the results to t ! 0 [58].
Another interesting and important fact is that it is really not necessary
for the random momentum increments to be drawn from a Gaussian
distribution. Any distribution with zero mean and appropriate mean square
p2 i 2mkB Tt along each coordinate
deviation satisfying the condition h
i
direction will be appropriate.
Brownian dynamics was first introduced by Ermak et al. [60,61] and is
commonly used to simulate polymers, ions, or colloidal particles in solution.
The monomer dynamics is modeled by the Langevin equation, with the
solvent modeled as the medium providing the frictional and random forces.
66
D
rV f r t
kB T
98
99
This equation describes diffusion in the external potential V(r). If the potential is constant (or zero), it reduces to the usual diffusion equation.
VII.
Background
67
FIG. 8 Free energy profile as a function of the reaction coordinate q showing two
states separated by a barrier. The lines in the upper part of the figure show different
ways in which dynamical trajectories can cross the barrier region.
68
governing the motion of the system in each of the states, but still very short
compared to the residence time of the system in either state (time scale
separation). Over this range of times, the probabilities of occupancy pA, pB
of the states evolve according to the master equations:
dpA
kA!B pA kB!A pB
dt
dpB
kB!A pB kA!B pA
dt
100
q_ 0qz q0 qt qz
qz q
101
where and stand for a Dirac delta function and a Heaviside function,
respectively. The numerator in Eq. (101) is an average rate of motion along
the reaction coordinate taken over all dynamical trajectories that cross the
barrier at time 0 and end up in state B after time t. Trajectories crossing the
barrier in the direction from A to B at time 0, such as (a) and (c) in Fig. 8,
contribute positively to the average in the numerator, while trajectories
crossing in the opposite direction, such as (b), contribute negatively. The
denominator is simply the equilibrium probability of occupancy of state A.
When time scale separation holds, the kA ! B calculated through Eq. (101)
turns out to exhibit a plateau value, independent of t, over a wide range of
(not too short) times. Eyrings Transition-State Theory (TST) rests on an
approximation: It assumes that any trajectory crossing the barrier in the
direction from A to B will eventually lead the system to state B. That is, it
ignores barrier recrossing trajectories along which the system ultimately
thermalizes in the state in which it originated [such as trajectory (b) in
Fig. 8]. In more formal terms, TST replaces q(t) in the numerator of
Eq. (101) with q(0), i.e., with its limit for t ! 0, at which the system is
bound to be in the state towards which q_ 0 is pointing. The TST estimate
of the rate constant thus reduces to the ratio of two equilibrium ensemble
averages:
kTST
A!B
q_ 0 q0 qz qz q0
qz q
q_ q_ qz q0
qz q
102
Background
69
kTST
A!B emerges essentially as an average velocity in the direction from A to B
along the reaction coordinate, which can readily be computed in most cases
from the momentum-space distribution of the system, times the ratio of
probabilities of being at the barrier and of residing in the origin state A.
The latter ratio can be calculated by MC or constraint MD methods
designed to sample the barrier region.
It is customary to express the rate constant kA ! B of Eq. (101) in terms
of its TST estimate, kTST
A!B , as
kA!B t kTST
A!B t
103
q_ 0 qz q0 qt qz
t
q_ 0 qz q0 q0 qz
104
70
between the particle and the medium (heat reservoir), as a result of which
the particle velocity q_ is distributed according to the requirements of
equilibrium at temperature T. By virtue of the separability of momentum
and configuration-space distributions, and realizing that positive and
negative values of q_ are equally probable, we can recast Eq. (102) for this
problem as
kTST
A!B
1 q_ qz q
2 qz q
kB T
2m
1=2
exp Vqz =kB T
R
q2A expVq=kB T dq
105
The integral in the denominator of Eq. (105) is taken over the entire state A,
to the left of qz. At low temperatures, the major contribution to this integral
comes from the immediate vicinity of the energy minimum qA, where the
function V(q) can be approximated by its Taylor expansion around qA
truncated at the second order term [compare Eq. (35)]. Setting
kA @2 V=@q2 jqA , we obtain
Z
Z 1
Vq
VqA
kA q qA 2
exp
exp
dq exp
dq
kB T
kB T
2kB T
q2A
1
2kB T
kA
1=2
VqA
exp
kB T
106
Using Eq. (106) in Eq. (105), we are led to the harmonic approximation for
HA
:
the transition rate constant, kTST,
A!B
1 kA 1=2
Vqz VqA
exp
2 m
kB T
Vqz VqA
A
exp
kB T
, HA
kTST
A!B
107
Background
71
origin state A times the Boltzmann factor of the barrier height measured
from the minimum of state A.
An ingenious analysis of the transition rate in systems such as the one
of Fig. 8, which takes explicitly into account energy exchange with the
degrees of freedom not participating in the reaction coordinate, was
carried out by Kramers [59]. In Kramers analysis, these degrees of
freedom are envisioned as comprising a Brownian bath which exerts
Langevin and frictional forces on the degree of freedom q. The motion of
q is described by a one-dimensional Langevin equation of the form (92).
This, of course, presupposes that the motion of the bath degrees of
freedom is fast relative to the evolution of q. The potential (or, more
generally, free energy) function V(q) is approximated by its Taylor
expansion truncated at the second order term both around the bottom of
the well of the origin state (qA) and around the top of the barrier qz.
Symbolizing by the friction coefficient, as in Eq. (92), by A the
curvature of the potential at qA, as above, and by kz>0 the opposite of
the curvature of the potential at the barrier, kz @2 V=@q2 jqz , Kramers
result for the rate constant is:
kA!B
p
Vqz VqA
2
1 exp
kB T
A
108
p
where A 1=2 A =m [natural frequency of oscillation
p in the harmonic
region of the well, as in Eq. (107), and =2 m=z is a dimensionless
parameter which increases with the friction coefficient and decreases with
the curvature of the potential at the barrier. When the coupling between
the reaction coordinate and the other degrees of freedom is weak,
1
and the Kramers result, Eq. (108), reduces to the harmonic approxiHA
. Note that the transmission coefficient
mation TST estimate, kTST,
pA!B
TST
kA!B =kA!B 2 1 predicted by Kramers theory is always
smaller than 1,
as it should be. In the case 1 (strong friction limit),
p
the quantity 2 1 is well approximated by 1/2 and the Kramers
result becomes
kA!B
A
r
kz
Vqz VqA
exp
kB T
m
109
In this limit (strong coupling between q and the bath and/or flat barrier)
kA ! B can be orders of magnitude lower than the TST estimate.
The exact location where qz is chosen affects the values of kTST
A!B and ,
but not their product kA ! B. Nevertheless, in practice every effort should be
72
made to define the reaction path so that the transition-state theory estimate
of the rate constant is minimized and the dynamical correction factor is
maximized.
How does one define the reaction path and the transition state given the
potential energy function of a system with many degrees of freedom? We
will assume here that the configuration is described in terms of the Nf massweighted Cartesian coordinates xi m1=2
i ri for all atoms. We will use the
symbol x to refer to all these degrees of freedom collectively. States, such
as A and B in the examples above, are constructed around local minima of
the potential energy, at which the gradient vector gx rVxx is 0 and
the Hessian matrix Hx @2 V=@x@xT jx is positive definite. Between two
neighboring local minima xA and xB there will be at least one saddle point
xz, at which g(xz) 0 and the Hessian H(xz) has one negative eigenvalue
with associated unit eigenvector nz. This saddle point is the highest energy
point on the lowest energy passage between xA and xB; it is usually a good
choice for the transition state z between A and B. The reaction path between
A and B, along which the reaction coordinate is measured, is a line in
Nf-dimensional space connecting xA and xB. To construct it, one initiates
two steepest descent trajectories at xz, one in the direction nz and the
other in the direction nz. Each such trajectory consists of small steps
dx (g(x)/|g(x)|) dq parallel to the direction of the local gradient vector
and terminates in one of the two minima connected through z (see Fig. 9).
The dividing surface between states A and B is defined as an (Nf 1)dimensional hypersurface with equation C(x) 0, with the following
properties: (a) it passes through the saddle point, i.e., C(xz) 0; (b) at the
saddle point it is normal to the eigenvector nz corresponding to the negative
eigenvalue of the Hessian (and, therefore, to the reaction path); (c) at all
points other than xz, it is tangent to the gradient vector. Conditions (b) and
(c) can be expressed mathematically as:
rCx
rCx
nz
110
xz
x 6 xz
rCx gx 0,
111
With this definition of the dividing surface, the TST estimate of the transition rate constant becomes [65]
kTST
A!B
d Nf x
x2A
d Nf p nx p Cx jrCxjNVT x, p
nxp>0
112
Background
73
74
kTST
A!B
R
kB T 1=2 x2A d Nf x Cx rCx expVx=kB T
R
Nf
2
x2A d x expVx=kB T
113
114
Vxz VxA
i1 i
exp
kTST
QNf z
A!B
kB T
i2 i
115
Background
75
coordinates x in the origin state A and, similarly, zi are the square roots
of the Nf 1 positive eigenvalues of the Hessian matrix at the saddle
point. Equation (115), first derived by Vineyard [67], is a useful generalz
ization of Eq. (107). In case the eigenfrequencies A
i or i are too high for
the corresponding vibrational motion to be describable satisfactorily by
means of classical statistical mechanics, a more appropriate form of
Eq. (115) is
kTST
A!B
QNf
kB T i1 1 exp hA
Vxz VxA
i =kB T
h
i
exp
h QNf 1 exp hz =k T
kB T
i2
116
(In the latter equation, classical partition functions for the harmonic
oscillators representing the modes have been replaced with quantum
mechanical ones. Zero point energy contributions are considered as being
part of the Vs.) The latter equation is a special case of the more general TST
expression
kTST
A!B
kB T
Gz GA
exp
h
kB T
117
where Gz is the Gibbs energy of the system confined to the dividing surface
and GA is the Gibbs energy of the system allowed to sample the entire origin
state.
The above discussion concerned ways of calculating the rate constant
for a transition between two states, A and B. The evolution of real
material systems (e.g., by diffusion or relaxation) often involves long
sequences of transitions between different states. Once the rate constants
are known between all pairs in a network of connected states, ensembles of
dynamical trajectories for the system can readily be generated by Kinetic
Monte Carlo simulation [64]. In such a simulation, the times for the next
transition to occur are chosen from the exponential distribution of waiting
times governing a Poisson process. If the rate constants are small, these
times are long; the evolution of the system can be tracked over time
periods many orders of magnitude longer than can be accessed by MD.
Thus, the long-time problem of MD is eliminated. The challenge in such
an approach is to identify the states and saddle points between them and
to include all relevant degrees of freedom in the transition rate constant
calculations.
76
Background
FIG. 10
77
78
XX X
uri rj nx Lx ex ny Ly ey nz Lz ex
118
j>i nx , ny , nz
nx 60, ny 60, nz 60
unx Lx ex ny Ly ey nz Lz ez
The sum over pairs of particles includes summation over the periodic images
of the second particle. The second term describes the contribution from a
particle interacting with its own periodic images.
If the interaction potential u is long range, all terms in the infinite sum
should be taken into account. The brute force approach here is to
truncate the summation at some large enough values of nx, ny, nz. Efficient
ways to do this for Coulombic interactionsthe Ewald summation, fast
multipole, and particlemesh methodswill be described in the following
chapters.
A much simpler technique is used for Lennard-Jones potentials or other
short-range interactions. The Lennard-Jones potential, Eq. (6), drops down
to only 0.016 at rc 2.5. If we just cut the potential off at this distance,
assuming that u 0 at r>rc, the error we make is less than 2%. We will see
later, how it can be accounted for. If such a cutoff is used, the sums in Eq.
(118) are limited only to the pairs of images that are closer to each other
than the cutoff distance rc. If rc is smaller than one half of the smallest of Lx,
Ly, or Lz, only the closest of images are taken into account in the first sum,
and there is no second summation over the self-images. This approximation
is called the minimum image convention. The intermolecular potential
energy is then given by:
V
XX
i
119
j>i
In this equation, n0x n0y n0z stands for the set of integers that correspond to the
pair of closest images of particles i and j. Notice that the circle centered
at particle A (Fig. 10) with radius rc may include a periodic image BN of
particle B, but not B itself.
The following fragment of FORTRAN code illustrates the implementation of the minimum image convention. It calculates the x, y, and z
Background
79
components of the separation vector ri rj, stored in the variables dx, dy,
dz, given that particle coordinates are stored in the arrays x, y, z, and box
sizes Lx, Ly, Lz are stored in the variables boxx, boxy, boxz.
..............................
invbx1.d0/boxx
invby1.d0/boxy
invbz1.d0/boxz
...............................
dx x(i)x(j)
dx dxboxx*anint(dx*invbx)
dy y(i)y(j)
dy dy boxy*anint(dy*invby)
dz z(i)z(j)
dz dz boxz*anint(dz*invbz)
............................
This piece of code is used throughout the modeling program with the
periodic boundary conditions whenever the distance between two particles is
needed: in the energy and force calculation, in the calculation of the pair
correlation function, etc. The code uses the standard Fortran function*
anint(a), which returns the nearest integer to the real or double
precision number a. Note that anint(0.51)1, anint(0.49)0,
anint(0.51)1, anint(0.49)0.
To fold all the particles inside the primary simulation cube,
Lx/2<xi<Lx/2, Ly/2<yi<Ly/2, Lz/2<zi<Lz/2, we can use similar
code:
..................................
invbx1.d0/boxx
invby1.d0/boxy
invbz1.d0/boxz
...............................
x(i) x(i) boxx*anint(x(i)*invbx)
y(i) y(i) boxy*anint(y(i)*invby)
z(i) z(i) boxz*anint(z(i)*invbz)
............................
When a chemical bond crosses the box boundary, its periodic image
reenters the box from the opposite side, as shown in Fig. 10. Unlike
intermolecular interactions, the intramolecular interactions should be
calculated between atoms of the same image of a molecule and not between
80
their periodic images. Nevertheless, it is also safe to use the minimum image
convention for the intramolecular interactions. As chemical bonds are
always shorter than half the box length, the convention will automatically
provide the right separation vectors, no matter whether the particle
coordinates are always folded back in the primary box, or whether the
particle trajectory in space is preserved.
When the potential energy is cut off at distance rc, interactions between
particles separated by a distance larger than rc are neglected. If these
interactions are described by spherically symmetric pairwise potentials, the
energy can be expressed in terms of the pair distribution function, as in
Eq. (63). If the rc is large enough, such that we can assume that there are
no structural correlations beyond this distance (gr 1), then the
long-distance part of the integration can be replaced by the integration of
the potential for r>rc:
Z
grur4r2 dr
V V sum N
120
r
Z c1
ur4r2 dr
Z1
1
dur
4r2 dr
PV Psum V N
gr r
6
dr
rc
Z1
1
dur
r4r2 dr
Psum V N
r
6
dr
rc
V sum N
rc
V sum and Psum are the values obtained from explicit summation over the
pairs within the cutoff radius rc. Similar equations can be derived for any
other property obtained via the summation over pairs of particles. The
integrals on the right hand sides of the above equations can easily be calculated for any pair potential u(r).*
Setting g(r) 1 corresponds to replacing the material at long distances
from the particle by a uniform background, with the average interaction
properties. Thus, on the one hand one would like the cutoff radius to be not
very large, to reduce the amount of calculations necessary for obtaining
the energy, on the other hand rc should be chosen large enough, so that
structural correlations can be negligible at distances r>rc. The latter
assumption has to be checked by calculating g(r).
Even though the interaction potential value u(rc) is quite small, the effect
of the cutoff can be quite substantial. For instance, using the Lennard-Jones
*They might be nonconverging for the long-range potentials or in fewer than three dimensions.
Background
81
(6-12) potential with cutoff rc 2.5 in the simple monatomic model of the
noble gas lowers the critical temperature by about 30% [54] compared to the
value obtained with a much longer cutoff or with the long-range corrections
given by Eq. (120).
If the interaction is not pairwise, similar expressions can be obtained using
the corresponding multiparticle correlation functions. If the interaction
potential u depends on the direction of the separation vector r as well as on
its length r, a multidimensional integration over the components of r is to be
performed instead of the spherical averaging leading to integration over
r |r|.
1=N
run
i1 Ai and
P
2
variance hA2 irun 1=N N
i1 Ai hAirun , the variance of the estimated
average hAi, 2(hAirun), equals h(A)2irun)/N. The probability to observe a
certain value of hAirun in a sample of N uncorrelated configurations is
described by a Gaussian distribution with average at the true mean and
variance h(A)2irun/N. The expected error in the simulation estimate of A is
1=2
.
thus hA2 i1=2
run =N
The configurations obtained in successive steps of a molecular dynamics,
Monte Carlo, or Brownian dynamics simulation are usually not very
different from each other. Hence, one should expect the energy, virial, or
any other function of the microscopic state obtained from the successive
configurations to be correlated. These correlations disappear after some
82
b
kX
hAiik hAirun 2
N i1
121
The correct block size k that ensures statistical independence of the hAik
i
can be determined through systematic analysis for different k [43].
To estimate errors in structural or other properties, expressed not just
as numbers but as functions of either distance (e.g., g(r)), or time (e.g.,
time correlation functions), a similar analysis should be carried out for
different values of the argument. For instance, when the time correlation
function is calculated, much better statistics is obtained for short times,
as there are much more data available. The time correlation function
values for times comparable to the length of the simulation run are
obtained with larger errors, as the average is taken over only a few pairs
of values.
When the property calculated is a single particle property, such as the
velocity autocorrelation function, or the particles mean squared displacement, averaging over the particles does help a lot. Averaging over even a
moderate number of particles of about 1000 decreases the error by more
than an order of magnitude. Unfortunately, this is not possible for all
properties. For instance, the calculation of viscosity calls for the stress
correlation function [Eq. (67)], which is not a single particle property. This
is why self-diffusion coefficients are usually estimated with much better
accuracy than viscosity.
A major challenge in polymer simulations is to produce a large number
of uncorrelated configurations, to sample various chain conformations. In
molecular dynamics, each time step involves small changes in the monomer
positions. Significantly changing the chain conformation involves large
Background
83
monomer displacements and requires many steps. The same thing happens
in Monte Carlo simulations involving only small monomer displacements
at each step. Global updates, involving large fragments of the polymer
molecules, which are used, e.g., in Configurational Bias Monte Carlo
techniques, help overcome this problem. Monte Carlo strategies for the
rapid equilibration of polymer systems will be discussed in subsequent
chapters of this book.
84
the crystal, which is far from the random coil conformation prevailing in
the polymer glass or melt. Therefore, special techniques are used to generate
realistic models of polymer melts or glasses [34,71].
REFERENCES
1. Hiller, A.; Wall, F.T.; Wheeler, D.J. Statistical computation of mean dimensions of macromolecules. J. Chem. Phys. 1954, 22, 1036.
2. Wall, W.A.; Seitz, F.T. Simulation of polymers by self-avoiding, non-intersecting random chains at various concentrations. J. Chem. Phys. 1977, 67,
37223726.
3. Alexandrowicz, Z.; Accad, Y. Monte Carlo of chains with excluded
volume: distribution of intersegmental distances. J. Chem. Phys. 1971, 54,
53385345.
4. Lal, M. Monte Carlo simulation of chain molecules I. Molec. Phys. 1969, 17, 57.
5. De Vos, E.; Bellemans, A. Concentration dependence of the mean dimension of
a polymer chain. Macromolecules 1974, 7, 812814.
6. Ryckaert, J.P.; Bellemans, A. Molecular dynamics of liquid butane near its
boiling point. Chem. Phys. Lett. 1975, 30, 123125.
7. Balabaev, N.K.; Grivtsov, A.G.; Shnol, E.E. Doklady Akad. Nauk SSSR
1975, 220, 10961098, in Russian.
8. Ceperley, H.L.; Frisch, H.L.; Bishop, M.; Kalos, M.H. Investigation of
static properties of model bulk polymer uids. J. Chem. Phys. 1980, 72,
32283235.
9. Weber, T.A.; Helfand, E. Molecular dynamics simulation of polymers. I.
Structure. J. Chem. Phys. 1979, 71, 47604762.
10. Vacatello, M.; Avitabile, G.; Corradini, P.; Tuzi, A. A computer model of
molecular arrangement in an n-paranic liquid. J. Chem. Phys. 1980, 73,
543552.
11. Fixman, M. Classical statistical mechanics of constraints. A theorem and application to polymers. Proc. Nat. Acad. Sci. USA 1974, 71, 30503053.
12. Go, N.; Scheraga, H. Analysis of the contribution of internal vibrations to the
statistical weights of equilibrium conformations of macromolecules. J. Chem.
Phys. 1969, 51, 47514767.
13. Go, N.; Scheraga, H. On the use of classical statistical mechanics in the treatment of polymer chain conformation. Macromolecules 1976, 9, 535542.
14. Mattice, W.; Suter, U.W. Conformational Theory of Large Molecules, the
Rotational Isomeric State Model in Macromolecular Systems; J. Wiley: New
York, 1994.
15. Insight II User Guide, version 2.3.0. Biosym Technologies: San Diego, 1993;
http://www.accelrys.com.
16. van Gunsteren, W.; Berendsen, H.J.C. Computer simulation of molecular
dynamics: methodology, applications and perspectives in chemistry. Angew.
Chem. Int. Ed. Engl. 1990, 29, 9921023.
Background
85
17. Weiner, S.J.; Kollman, P.A.; Nguyen, D.T.; Case, D.A. An all atom force eld
for simulations of proteins and nucleic acids. J. Comput. Chem. 1986, 7, 230,
http://www.amber.ucsf.edu/amber/amber.html.
18. Doi, M.; Edwards, S.F. The Theory of Polymer Dynamics; Clarendon Press:
Oxford, 1994.
19. Khokhlov, A.R.; Grosberg, A.Yu. Statistical Physics of Macromolecules; AIP
Press: New York, 1994.
20. Flory, P.J. Statistical Mechanics of Chain Molecules; Hanser Publishers:
Munich, 1989.
21. Ted Davis, H. Statistical Mechanics of Phases, Interfaces, and Thin Films; VCH:
New York, 1996.
22. deGennes, P.-G. Scaling Concepts in Polymer Physics; Cornell University Press:
Ithaca, 1979.
23. Fleer, G.J.; Cohen Stuart, M.A.; Scheutjens, J.M.H.M.; Cosgrove, T.; Vincent,
B. Polymers at Interfaces; Chapman and Hall: London, 1993.
24. Gotze, W.G. Recent tests of the mode-coupling theory for glassy dynamics.
J. Phys.: Condens. Matter 1999, 11, A1A45.
25. Cummins, H.Z. The liquid-glass transition: a mode coupling perspective.
J. Phys.: Condens. Matter 1999, 11, A95A117.
26. Strobl, G.R.; The Physics of Polymers: Concepts for Understanding Their
Structures and Behavior; Springer-Verlag: Berlin, 1997.
27. Car, R.; Parrinello, M. Unied approach for molecular dynamics and density
functional theory. Phys. Rev. Lett. 1985, 55, 24712474.
28. Reichl, L.E. A Modern Course in Statistical Physics; University of Texas Press:
Austin, 1980.
29. Arnold, V.I. Mathematical Methods of Classical Mechanics; Springer-Verlag:
2nd Ed., 1989.
30. Goldstein, H. Classical Mechanics; AddisonWesley: 2nd Ed., 1980.
31. Kneller, G.R.; Hinsen, K. Generalized Euler equations for linked rigid bodies.
Phys. Rev. 1994.
32. Hinsen, K.; Kneller, G.R. Inuence of constraints on the dynamics of polypeptide chains. Phys. Rev. 1995, E52, 6868.
33. Rice, L.M.; Brunger, A.T. Torsion angle dynamicsreduced variable conformational sampling enhances crystallographic structure renement. Proteins
Aug 1994, 19, 277290.
34. Theodorou, D.N.; Suter, U.W. Detailed molecular structure of a vinyl polymer
glass. Macromolecules 1985, 18, 14671478.
35. Ludovice, P.J.; Suter, U.W. Detailed molecular structure of a polar vinyl polymer glass. In Computational Modeling of Polymers; Bicerano, J., Ed.; Marcel
Dekker: New York, 1992; page 401.
36. Madras, N.; Sokal, A. Nonergodicity of local, length conserving MonteCarlo algorithms for the self-avoiding walk. J. Statist. Phys. 1987, 47(2),
573595.
37. McQuarrie, D.A. Statistical Mechanics; Harper and Row: New York,
1976.
86
Background
87
2
Rotational Isomeric State (RIS)
Calculations, with an Illustrative
Application to Head-to-Head,
Tail-to-Tail Polypropylene
E. DEMET AKTEN and WAYNE L. MATTICE
Akron, Ohio, U.S.A.
ULRICH W. SUTER
Zurich, Switzerland
I.
INTRODUCTION
The Rotational Isomeric State (RIS) model is excellent for the analysis of
conformation-dependent physical properties of chain molecules in their
unperturbed state. Its special strength is the incorporation of detailed
information about the covalent structure (bond lengths, bond angles,
torsion angles, and torsion potential energy functions) in a formalism that
can be evaluated quickly by the smallest computer. The answer is the exact
result for the specific model, as defined by the geometry and energies of the
short-range intramolecular interactions.
The most frequently calculated property is the mean square unperturbed
end-to-end distance, hr2i0. Other properties susceptible to rapid computation include the average of the end-to-end vector, hri0, and the mean square
unperturbed radius of gyration, hs2i0. The viscosity of a dilute solution in a
solvent can be estimated from hs2i0 or hr2i0 via the equivalent sphere
model for hydrodynamic properties. Several higher even moments, hr2pi0
and hs2pi0, p 2, 3, . . . , provide information about the shape (width,
skewness) of the distribution functions for r2 and s2. When combined with
information about the electronic charge distribution within individual rigid
units, the RIS model can calculate the mean square dipole moment, h
2i0.
Also accessible are optical properties that depend on the anisotropy of the
90
Akten et al.
polarizability tensor (optical anisotropy) in conjunction with r (stressoptical coefficient), k (molar Kerr constant), or the magnetic susceptibility
(CottonMouton constant).
The speed of the calculations derives from the use of efficient generator
matrix techniques that were introduced six decades ago [1]. These techniques
were adapted to the study of macromolecules a decade later [2]. Therefore
the description of conformation-dependent physical properties of macromolecules with the RIS model predates the widespread availability of
computers and simulation methods that require enormous computational
power for their implementation. A strong increase in the use of the RIS
model over the decade from 1959 to 1969 culminated in the publication of
Florys classic book on the subject [3]. Some of the subsequent developments are contained in two books published in the 1990s. One book has an
emphasis on self-instruction in the use of the RIS model [4]. The other book
is a compilation, in standard format, of RIS models for over 400 polymers
that have appeared in the literature through the mid-1990s [5].
There is one statistical weight matrix for each of the n bonds in the chain.
The dimensions of the statistical weight matrices depend on the number of
rotational isomers assigned to each bond. The continuous range for the
torsion angle, , is replaced by a finite set of torsion angles, i.e., a continuous conformational integral is replaced by a discrete summation. The sum
should be designed so that the terms include all regions of the local conformational space that have a significant population. Many unperturbed
chains have pairwise interdependence of their bonds, causing each Ui to
have a number of columns given by the number of rotational isomers at
bond i (denoted i), and a number of rows given by the number of rotational
isomers at the preceding bond. Each element in Ui is the product of a
statistical weight for a first-order interaction (which depends only on
the state at bond i) and the statistical weight for a second-order interaction (which depends simultaneously on the states at bonds i 1 and i).
91
Each statistical weight is formulated as a Boltzmann factor with the appropriate energy and temperature. Sometimes there is inclusion of a preexponential factor that accounts for the differences in conformational entropy of
the various rotational isomeric states.
The first equation incorporates all of the energetic information that will
be employed in the calculation. It incorporates structural information only
indirectly, via the assignment of the numbers of rotational isomers at the
various bonds, and the assignment of numerical values for the statistical
weights.
There is no concern here about the probability of observation of this conformation; we are ignoring, for the moment, the energetic information contained in Z. Each Fi includes the information about the manner in which
bond i affects f. The details, of course, depend on how we have chosen f. If f
is the squared end-to-end distance, we need to know the length of bond i, the
angle between bonds i and i 1, and the torsion angle at bond i. These three
properties are denoted li, i, and i, respectively. The square of the end-toend distance is written in terms of these local variables as
r2 2
n1 X
n
X
k1 jk1
n
X
lk2
k1
where li denotes the bond vector, expressed in a local coordinate system for
bond i,
2 3
li
li 4 0 5
0
92
Akten et al.
cos
Ti 4 sin cos
sin sin
3
sin
0
cos cos sin 5
cos sin cos i
The single and double sums in Eq. (3) are evaluated exactly via Eq. (2) using
F1 1 2lT1 T1 l12
3
2
1 2lTi Ti li2
7
6
Fi 6
Ti
li 7
5,
40
0
2
1<i<n
3
2
ln
6 7
7
Fn 6
4 ln 5
1
G1G2 . . . Gn
Z
G i Ui IF diagF , F , . . . , Fi
9
10
93
hr2 i0
nl 2
11
Can an analysis with the RIS model assist in the resolution of this
discrepancy in the experimental results? A three-state RIS model for the
head-to-head, tail-to-tail form of polypropylene, with a repeating sequence
given by -CH2-CH(CH3)-CH(CH3)-CH2-, is constructed and analyzed here.
The mean square unperturbed dimensions of exclusively head-to-head, tailto-tail polypropylene are determined for chains with various stereochemical
compositions and stereochemical sequences. The critical parameters in the
model are identified, and the predicted mean square unperturbed dimensions are compared with experiment.
94
Akten et al.
FIG. 1
95
13
VCC;l
1 1
6
41 !
1 !
3
!
7
!5
14
VCC;d QVCC;l Q
15
Here ! is the statistical weight for the second-order interaction of the terminal chain atoms when the conformation has two g placements of opposite
sign (the ! in the 2,3 and 3,2 elements), or the equivalent interaction of the
methyl group with the terminal CH (the remaining two !, which occur in the
1,3 and 2,2 elements of VCC;l, but in the 1,2 and 3,3 elements of VCC;d). The
full statistical weight matrices, incorporating both first- and second-order
interactions, are
2
6
UCC;l VCC;l DCC 4 1
1
!
7
! ! 5
!
UCC;d QUCC;l Q
16
17
96
Akten et al.
FIG. 2
Statistical Weight Matrices for the CH2CH Bond. This statistical weight
matrix, denoted by UCM;x, incorporates statistical weights for short-range
interactions that occur in the fragment -CH-CH2-CH2CH(CH3)-CH-,
where the longest dash denotes the bond that is now considered as bond i.
An example of this fragment is depicted in Fig. 2. Rotation about bond i
establishes the position of the bold CH and bold CH3 in -CH-CH2-CH2
CH(CH3)-CH- with respect to the remainder of the fragment. The firstorder interactions now occur between the bold CH2 and the other two bold
groups. Involvement of the methyl group demands two forms of the matrix
for the first-order interactions.
DCM;l diag1, 1,
18
DCM;d QDCM;l Q
19
6
VCM;l 4 !
1
7
!5
20
! !
VCM;d QVCM;l Q
21
UCM;l
1
6
4!
1
1
1
3
7
! 5
22
! !
UCM;d QUCM;l Q
23
FIG. 3
97
Statistical Weight Matrices for the CHCH Bond. This statistical weight
matrix incorporates short-range interactions in the fragment -CH2-CH2CH(CH3)CH(CH3)-CH2-, depicted in Fig. 3. The matrix is denoted by
UMM;xy, where x and y define the stereochemistry of the attachments of the
two methyl groups, in the order that they occur in the fragment. The firstorder interactions occur between pairs of bold groups in -CH2-CH2CH(CH3)CH(CH3)-CH2-, with a member of the pair being selected from
among the bold groups on either side of bond i. Involvement of two methyl
groups demands four forms of the diagonal matrix for the first-order
interactions. Two of these forms are identical.
DMM;dd DMM;ll diag1, ,
24
DMM;dl diag, 1,
25
DMM;ld QDMM;dl Q
26
Here is the statistical weight for three simultaneous first-order interactions, relative to the case where there are two first-order interactions.
The second-order interactions occur between the initial methylene and
the other two bold groups in -CH2-CH2-CH(CH3)CH(CH3)-CH2-.
Involvement of only one of the methyl groups means that there are only
two forms of the matrix of second-order interactions. These two matrices
were encountered in the consideration of the previous bond.
VMM;ll VMM;dl VCM;l
27
28
6
UMM;ll 6
4!
!
7
! 7
5
29
!
UMM;dd QUMM;ll Q
30
98
Akten et al.
6
UMM;dl 6
4 !
7
! 7
5
!
UMM;ld QUMM;dl Q
31
32
Statistical Weight Matrices for the CHCH2 Bond. This matrix incorporates statistical weights for short-range interactions in the fragment -CH2CH(CH3)-CH(CH3)CH2-CH2-, depicted in Fig. 4. It is denoted by
UMC;xy, where x and y define the stereochemistry of the attachments of the
two methyl groups, in the order that they occur in the fragment. The firstorder interactions involve the bold groups in -CH2-CH(CH3)-CH(CH3)
CH2-CH2-, and the second-order interactions involve the bold groups in
-CH2-CH(CH3)-CH(CH3)CH2-CH2-. All of the short-range interactions
have been encountered previously, as have all of the matrices of first- and
second-order interactions.
DMC;ll DMC;dl DCM;d
33
34
35
36
6
UMC;ll 6
41
1
7
! ! 7
5
! 1
UMC;dd QUMC;ll Q
FIG. 4
37
38
6
UMC;ld 6
41 !
1 !
99
7
! 7
5
39
UMC;dl QUMC;ld Q
40
41
6
6
U2 6 1
4
7
7
1 7D2
5
42
2 3
1
6 7
6 7
Un 6 1 7
4 5
43
1
Any ! that appears in U2 is replaced by 1, because the chain has been
initiated in a manner that eliminates any possibility of second-order interactions at bonds 1 and 2.
For long chains in which all of the pseudoasymmetric centers have the
same chirality (isotactic chains), Z is dominated by the terms
a
44
or
a
45
100
Akten et al.
a
a
UCC;d UCM;d UMM;dd UMC;dd Q UCC;l UCM;l UMM;ll UMC;ll Q
46
which demonstrates the equivalence of the Zs for these two chains (they are
enantiomers). The two chains with perfectly alternating (dl or ld) pseudoasymmetric centers also have the same Z [which, however, is different from
the Z in Eq. (46)] because the two structures are identical; they constitute
different representations of a mesoform, which is syndiotactic.
101
dd. . .dd), or where there is a strict alternation in chirality (either ld. . .ld or
dl. . .dl). The values of C1 can be calculated using the program in Appendix
C of Mattice and Suter [4]. This short program computes the mean square
unperturbed end-to-end distance, hr2i0, for a repeating unit of m bonds
embedded in a chain of n bonds, using a special case of Eq. (VI-66) of
Ref. [4], which for present purposes can be written as
2
1
r 0 G 1 G 2 G 3 . . . G m G 1 G 2 n=m1 G 3 . . . G m1 G m
Z
47
Z U1 U2 U3 . . . Um U1 U2 n=m1 U3 . . . Um1 Um
48
The internal bold italic G matrices in Eq. (47) are of dimensions 15 15,
formulated as described in Eq. (10) (see, e.g., pp. 122126 and 128130 of
Ref. [4]). The first and last matrices in Eqs. (47) and (48) take special forms
of a row and column, respectively [20]. The hr2i0 are discussed here using the
asymptotic limit for the dimensionless characteristic ratio, which is determined by linear extrapolation of Cn vs. 1/n (see pp. 1920 of Ref. [4]).
C1 lim Cn
49
n!1
Table 1 shows that identical results are obtained for a chain and its
mirror image, as expected, but different C1 can be obtained for pairs of
chains in which the stereochemical sequences are not mirror images. The
relationship between the two distinct results in Table 1 can be rationalized
qualitatively by identification of the preferred conformation of the sequence
of four rotatable bonds. Identification can be achieved by inspection of the
pattern of the appearance of , , , and ! in the U matrices, realizing that
each statistical weight has a value smaller than 1.
Equations (16), (22), (29), and (37), or, equivalently, Eqs. (17), (23), (30),
and (38), show that the preferred conformation is tttt when a single chirality
is propagated along the chain. The all-trans conformation has a statistical
weight of 1, and all other conformations have a statistical weight of less
Us (equation numbers)
C1
7.94
7.30
7.30
102
Akten et al.
than 1. The preferred conformation becomes either ttgt or ttgt when there
is a strict alternation in chirality of the pseudoasymmetric centers. The g
conformation is preferred at the CH2CH(CH3)CH(CH3)CH2 bond, due to
the pattern of the statistical weights denoted by 1 and in the D matrices in
Eqs. (24)(26), and hence in the U matrices in Eqs. (29)(32). Therefore, on
the basis of the preferred conformations, one would expect the first entry
in Table 1 to have a higher C1 than the last two entries.
@ ln x x0
2x
C1, x0
50
Here x is the statistical weight being varied, x0 is its value in the provisional
set, and all other parameters are held constant at their values in the provisional set. The values of this partial derivative are summarized in Table 2.
An increase in the value of any of the statistical weights produces a
decrease in C1 if all pseudoasymmetric centers are either d or l. Changes in
affect C1 relatively more than comparable changes in any of the other three
statistical weights. In contrast, a mixed pattern is seen with the chain
in which there is strict alternation in the pseudoasymmetric centers. C1
increases with an increase in or , or a decrease in or !, with the greatest
sensitivity seen upon a change in . Table 2 also shows how C1 is affected
by an increase in all bond angles, or by increasing only that half of the bond
Single chirality
0.14
0.069
0.22
0.11
3.8
0.65
0.028
0.106
0.055
0.067
4.6
1.3
103
104
Akten et al.
C1
(ld)
(ld)(dl)
(ll)(dd)
(ll)(ld)(dd)
(ll)(dd)(ld)
(ld)(dl)(dl)
7.30
6.03
6.97
6.94
6.74
6.32
a
Parentheses enclose symbols for two bonded carbon atoms
that each bear a methyl group.
105
is probably in the range 6.06.7. The predictions from the model are
definitely larger than the experimental estimate of 4.5 [17], but are consistent
with the experimental estimate of 6.16.4 [18].
FIG. 5 A fragment showing the defect arising from the 1,2 addition in the parent
polymer.
106
Akten et al.
a similar role. However, it does not seem likely that an increase in defects
from 3 to 6% could depress the characteristic ratio from 6.16.4 to 4.5.
D. Conclusion
The construction and analysis of a RIS model for a homopolymer have been
illustrated by consideration of head-to-head, tail-to-tail polypropylene.
Statistical weight matrices, incorporating first- and second-order interactions, are formulated for all bonds, and for all configurations of the
attachments of the methyl groups to the chain. Preliminary estimates of the
values of the statistical weights in these matrices could be obtained by
reference to prior work, which in turn is based on conformational energy
calculations for small hydrocarbons. Along with preliminary estimates of
the geometry in the three rotational isomeric states for each rotatable bond,
this information permits calculation of C1 for chains with simple repeating
sequences.
The influence of the individual energetic and structural parameters on
C1 is determined in the sensitivity tests.
The values of C1 specified by the model for an atactic homopolymer of
head-to-head, tail-to-tail polypropylene are in excellent agreement with
those reported by Graessley et al. [18]. Reasonable adjustments of the
parameters in the RIS model cannot reduce Cn into the range reported by
Arichi et al. [17].
ACKNOWLEDGMENT
The construction and analysis of the RIS model was supported by the
Edison Polymer Innovation Corporation.
REFERENCES
1. Kramers, H.A.; Wannier, G.H. Phys. Rev. 1941, 60, 252.
2. Volkenstein, M.V. Dokl. Akad. Nauk SSSR 1951, 78, 879.
3. Flory, P.J. Statistical Mechanics of Chain Molecules; Wiley-Interscience:
New York, 1969; reprinted with the same title by Hanser, Munchen (Munich),
1989.
4. Mattice, W.L.; Suter, U.W. Conformational theory of large molecules.
The Rotational Isomeric State Model in Macromolecular Systems; Wiley:
New York, 1994.
5. Rehahn, M.; Mattice, W.L.; Suter, U.W. Adv. Polym. Sci. 1997, 131/132.
6. Abe, Y.; Tonelli, A.E.; Flory, P.J. Macromolecules 1970, 3, 294.
7. Biskup, U.; Cantow, H.J. Macromolecules 1972, 5, 546.
107
3
Single Chain in Solution
REINHARD HENTSCHKE
I.
fb b bo 2
bonds
p
X X
f o 2
valence
angles
f#, k 1 cosmk # k
torsion k1
angles
X Aij Bij qi qj
6
rij
r12
rij
ij
i<j
!
1
The first three terms describe potential energy variations due to bond (b),
valence angle (), and bond torsion angle (#) deformations. The remaining
(nonbonding) terms are Lennard-Jones and Coulomb interactions between
interaction sites separated by a distance rij j~ri r~j j. Nonbonded interactions usually exclude pairs of sites belonging to the same bond or valence
110
Hentschke
FIG. 1 Snapshot taken during the initial stage of a molecular dynamics simulation
of oligo(vinylpyrrolidone) (20mer) in an ionic solution. The box indicates the cubic
simulation volume. Periodic boundary conditions are used to extend the system to
(quasi-)infinity. The arrows indicate (from left to right) Na, C6H5SO
3 , Oligomer,
and water.
111
112
Hentschke
sophisticated Monte Carlo techniques, are discussed in this book, and the
reader is referred to the pertinent chapters.
Force fields contain numerous parameters (here: fb, bo, f, o, f#, k , mk, k,
Aij, Bij, and qi). Parameterization procedures employ training sets, e.g.,
amino acids in the case of proteins or relevant monomer-analogous
molecules in the case of technical polymers. The parameters are adjusted
using thermodynamic, spectroscopic, or structural data, and increasingly
quantum chemical calculations (most importantly in the case of the torsion
potential in the valence part of a force field) available for the training set.
Probably the most difficult parameters are the partial charges, qi, usually
located on the nuclei, since they are influenced by a comparatively large
number of atoms in their vicinity. Let us consider the case of an oligomer
in explicit solution (cf. Fig. 1). One may start with the vacuum values for
the qi determined for the training molecules via an ESP procedure
(ElectroStatic Potential fitting) based on fast semiempirical methods like
AM1 (Austin Model 1) [8]. Polarization effects due to solvent are included
roughly in a mean field sense via a scaling factor multiplying the charges.
The scaling factor may be determined by comparison with suitable
experimental data (e.g., the density of a solution consisting of the solvent
of interest and a monomer-analogous solute as function of temperature and
solute concentration [9]). The same scaling factor is subsequently applied to
the charges on the oligomer, which also can be determined in vacuum using
AM1 (possibly including conformational averaging, which, if it is done
correctly, of course requires information not available at this point). A more
systematic and maybe in the long run more promising approach is the
fluctuating charge method [10,11]. Here the qi are dynamic variables just like
the positions, r~i . Their equations of motion follow from the Lagrangian
L
1XX
1XX
2
mim r~_im
mq q_ 2im Ufrg, fqg
2 m im
2 m im
The index m labels all molecules in the system, and im indicates the ith atom
or charge center in the mth molecule. The quantity mq is a mass parameter,
which, for simplicity, is the same for all charges. The potential energy is
Ufrg, fqg U non-Coulomb frg Efrg, fqg
Here {r} means all positions, and {q} means all partial charges. The first
term includes all non-Coulomb contributions [cf. Eq. (1)]. The second term,
the Coulomb part of the potential,
P is approximated
P in terms of a truncated
expansion in qi, E({r}, {q}) i Eio oi qi 12 i, j qi qj Jij . The quantities
113
Eio and oi are single atom parameters, and Jij Jij rij; i , j is a Coulomb
integral (note: Jij ! r1
ij ) computed from two charge distributions modrij !1
eled in terms of single Slater orbitals with the exponents i and j. The
equations of motion, i.e.,
~ r~ Ufrg, fqg
mim r~im r
im
and
mq q im
im
m
Rt
P
P
P
1
follow from t12 L m lm im qim dt 0, where
im
im
P the lm Nm
are Lagrange parameters, i.e., the conditions im qim 0 ensure that the
total charge on a molecule is zero. The quantity
im @E=@qim assumes
the role of the chemical potential of the charges, and Nm is the total
number of charge sites in the mth molecule.
The additional computational effort, compared to fixed partial charges, is
rather minor, i.e., about 15% for simulations of neat water [10,11] (the
authors consider the TIP4P and SPC/E water models). Figure 3 shows an
application of this technique to oligo(ethyleneoxide) in water [12]. This
simulation contained a 7mer immersed in 1000 water molecules at ambient
114
Hentschke
1
1 1
UM #i UMM #i , #j
UM #j
exp
kB T 2
2
The lower indices indicate neighboring monomers along the chain, and the
upper indices indicate RIS values of the corresponding torsion angles #i .
Here the total potential energy of the polymer is
U
X
i
UM #i
UMM #i , #i1
115
m
the general case of m distinct rotational isomeric states # 1 this allows
the following simple construction procedure for polymer conformations:
1.
2.
3.
4.
if
if
if
..
.
z P1jk
P1jk < z P1jk P2jk
P1jk P2jk < z P1jk P2jk P3jk
..
.
#m
if
It should be noted that in general the conformation of a polymer backbone is characterized by a periodic sequence of nonequivalent torsion
angles, #A , #B , #C , . . . , #A , #B , . . . , rather than by merely one (type of)
torsion angle, #, as in the present example. Nevertheless, the generalization
is straightforward.
A Mathematica implementation of the above algorithm for m 3
(polyethylene) is:
CP[i_,j_]:PP[[i,j]]/P[[j]];
k1;
b1{1,0,0};
b2{0,1,0};
R{0,0,0};
f112/180 Pi;
tab[n_]:Line[Table[zRandom[ ];
If[z<CP[1,k],l1,If[z<CP[1,k]CP[2,k],l2,l3]];
kl;
t(2 l-1) Pi/3;
b3N[-b2 Cos[f] CrossProduct[b2,b1] Sin[t] CrossProduct[CrossProduct[b2,b1],b2] Cos[t]];
b3b3/Sqrt[b3.b3];
RRb3;
b1b2;
b2b3;
R,{n}]];
Show[Graphics3D[tab[10000]]]
116
Hentschke
CP is the above conditional probability computed from the normal probabilities P and PP. The indices i and j may assume the values 1, 2, and 3,
here corresponding to the rotational isomeric states gauche(), trans, and
gauche().
An example for polyethylene, using the Jorgensen potential [15] to model
the potential in Eq. (7), is illustrated in the upper panel of Fig. 4. Here
n(r, r) is the average number of united atom carbon pairs, i.e., pairs of
effective methylene groups, divided by N, whose separation is between
r r/2 and r r/2. Here N is the total number of methylene groups in
the chain. For small r the distribution is discrete. The first pronounced peak
117
occurs at r 1 corresponding to the bond length in the units used here [note:
n(1, r) 2]. As r increases the distribution becomes continuous. An
important quantity based on n(r, r) is the scattering intensity
imax
X
I
sinqri
nri, r
N 1
Io
qri
i1
!
8
allowing the direct comparison with X-ray and neutron scattering experiments. Here Io is a constant independent of the polymer structure, r r i,
and q is the magnitude of the scattering vector. For simplicity we have set
the form factors for each scattering site equal to one. The scattering intensity
corresponding to the above n(r, r) is shown in the lower panel of Fig. 4. At
large q all curves come together, because the relevant lengths ( 2q1 ) are
small compared to any of the chain lengths considered. The effect of the
chain length is most pronounced for q values around 0.15, where a plateau
develops as N is increased. It is worth emphasizing that in general scattering
curves are more complex due to the more complex chemical architecture of
most polymers in comparison to polyethylene [16]. With some experience,
however, it is possible to relate many features of the scattering curve to
underlying conformational features, e.g., helical conformations. In the present very simple example there are two easily distinguishable q-regimes (for
large N), i.e., the linear increase beyond q 0.3, which reflects the simple
chemical structure of our model chain, and the initial increase with a
subsequent development of a plateau below q 0.3, which reflects the
statistical chain behavior. In this q-regime the scattering intensity of a
simple Gaussian chain is a good approximation to the shown scattering
intensity (for large N). The Gaussian chain in particular yields the plateau
q2 I=Io N ! 12=CN for large q (referring to the continuum limit, i.e., there
is no atomic structure no matter how large q is). The quantity
R2N
CN
Nb2
118
Hentschke
construction procedure to generate a large number of single chain conformations. The above value actually corresponds to polyethylene in a melt,
an example which we have chosen for its simplicity. For most dilute polymer
solutions one obtains larger values usually between 10 and 20. Notice that
C1 obtained in this fashion describes the shape of a polymer on an intermediate length scale, on which self-avoidance has no significant effect yet.
Because the above model includes short-range interactions only, a chain
may intersect itself. Polymers modeled with this method eventually obey
hR2N i / N 2v with v 1/2.
Powerful as the transfer matrix approach is, it suffers from certain
deficiencies. Including coupling beyond nearest neighbors is difficult.
Accounting for solvent effects also is difficult. In a crude approximation
solvent effects may be included via an effective dielectric constant in a
Coulomb term integrated into the coupling in Eq. (7) [17], which itself could
be improved to include local solventpolymer interactions in terms of
effective interactions (e.g., [18]). On the other hand, it is of course possible to
extract the probabilities p# and p##0 directly from oligomer simulations
in explicit solvent, and feed them into a construction procedure like the one
outlined above [19]. Applications of this general idea to poly(isobutene)
in vacuum (!) or to poly(vinylpyrrolidone) in water are discussed in [20]
and [9]. Thus, other than in the usual transfer matrix calculations of these
probabilities, molecular solvent effects are included on the length scale of
the oligomer. Some care must be taken to diminish end effects, however.
Moreover it is possible to include couplings beyond nearest neighbors, i.e.,
next-nearest neighbor effects may be introduced analogously via conditional
probabilities of the type p#j#0 #00 p##0 #00 =p#0 #00 . An example can be
found again in [9].
Before leaving the subject we briefly mention one common approximate
calculation of CN. The central assumption is the uncoupling of adjacent
monomers along the chain, i.e., UMM 0, which allows us to derive the
expression
2
1
1 2
N
CN I hTi I hTi hTi I hTi I hTi
10
N
1, 1
which is discussed in many polymer textbooks. Here I is a unit matrix, and
T, in cases where the polymer backbone contains one type of torsion angle
only (e.g., polyethylene), is given by
0
1
cos
sin
0
T @ cos # sin cos # cos sin # A
11
sin # sin sin # cos cos #
119
using the above convention. The index 1,1 indicates that CN is the 1,1element of the matrix. More complex polymer architectures may be accommodated by dividing the polymer into repeating sequences of effective bonds
(chemical units), again assumed to be independent, where T is the product
of the respective T-matrices for each effective bond. The conformation
averages (e.g., hcos i or hcos # sin i) again may be computed directly
from the corresponding oligomer simulations. Despite its crude approximations this method appears to yield reasonable first guesses for CN [21].
Frequently the molecular structure and dynamics of the polymersolvent
interface also are of interest. Structural information may be derived via pair
correlation functions of the general type considered in Fig. 3. For instance,
g2(r) may be calculated via g2(r) (4pr2r)1n(r, r), where n(r, r) is the
average number of solvent atoms of a certain type within a spherical shell of
radius r and thickness r centered on a polymer atom, and is the density
of the same solvent atoms in the bulk solvent. A dynamic quantity of
interest is the average residence time of a solvent atom or molecule in a
certain distance and/or orientation relative to a polymer atom or atom
group, e.g., hydrogen bonds between certain polar groups in a polymer and
water. Here n(t, te; r, r) is the average number of solvent atoms or molecules
again within a spherical shell (omitting the orientation dependence) at time
t, under the condition that the same solvent molecules were present in the
shell at t 0. The quantity te is an excursion time allowing a solvent
molecule to leave the shell and return within the excursion P
time without
max
n et= ,
being discarded. The result may be fitted via nt, te ; r, r 1
which is sensible for max 2 (cf. Ref. [9]). For large max, i.e., a broad
distribution of residence times, , the dynamics assumes glasslike behavior.
There are important systems, where one is interested in the intrinsic
stiffness of polymers, which in solution possess a regular superstructure
most commonly helical conformations. Here the above approaches are
less useful, and a method called the segment method, is more applicable.
A periodic section of the helix of interest is simulated in explicit solvent
using molecular dynamics. The top and bottom of the segment are
connected via suitable periodic boundary conditions. These are different
from the usual periodic boundary conditions, because they include a
covalent bond of the backbone. The simulation box should be large enough
(a) to ensure that the solvent attains its bulk structure at large perpendicular
distances from the helix, and (b) that the important undulatory wavelengths
along the helix backbone should be included. Provided the simulation run is
long enough one may extract sufficient independent segment (backbone)
conformations to Build up realistic and solvent specific helix conformations of a desired molecular weight. Build up means that an existing piece
of the helix may be extended by an extracted segment provided that a certain
120
Hentschke
u~0 u~s exps=P
12
FIG. 5 Four PBLG backbone conformations (IIV) constructed via the above
segment method. The inset illustrates the transformation of the explicit local
chemical structure into a persistent flexible rod.
121
where the u~ are unit vectors tangential to the backbone contour, and s is
their separation along the contour [23] (cf. the inset in Fig. 5). Note that P in
certain cases may be related to CN [24]. Via direct application of Eq. (12)
to contours like those shown in Fig. 5 one obtains P 1000 A for PBLG
in DMF at T 313K, in good accord with the experiments [22].
13
14
122
Hentschke
REFERENCES
1. Allen, M.P.; Tildesley, D.J. Computer Simulation of Liquids; Oxford: Clarendon
Press, 1990.
2. Lipkowitz, K.B.; Boyd, D.B. Reviews in Computational Chemistry; Weinheim:
Wiley/VCH, 1990.
3. Street, W.B.; Tildesley, D.J.; Saville, G. Mol. Phys. 1978, 35, 639648.
4. Tuckerman, M.E.; Berne, B.J. J. Chem. Phys. 1991, 95, 83628364.
5. Scully, J.L.; Hermans, J. Mol. Simul. 1993, 11, 6777.
6. Tschop, W.; Kremer, K.; Hahn, O.; Batoulis, J.; Burger, T. Acta Polymerica
1998, 49, 7579.
7. Tschop, W.; Kremer, K.; Batoulis, J.; Burger, T.; Hahn, O. Acta Polymerica
1998, 49, 6174.
8. Leach, A.R. Molecular Modeling; Harlow: Addison Wesley Longman Limited,
1996.
9. Flebbe, T.; Hentschke, R.; Hadicke, E.; Schade, C. Macromol Theory Simul
1998, 7, 567577.
10. Rick, S.W.; Stuart, S.J.; Berne, B.J. J. Chem. Phys. 1994, 101, 61416156.
11. Rick, S.W.; Berne, B.J. J. Am. Chem. Soc. 1996, 118, 672679.
123
4
Polymer Models on the Lattice
LLER
K. BINDER and M. MU
Mainz, Germany
J. BASCHNAGEL
I.
Johannes-Gutenberg-Universitat Mainz,
INTRODUCTION
R2 a2 N
and the number of configurations (i.e., the partition function of the chain
ZN) is (z is the coordination number of the lattice)
ZN zN
126
Binder et al.
FIG. 1 (a) Construction of a 22-step random walk (RW) on the square lattice.
Lattice sites are labeled in the order in which they are visited, starting out from the
origin (0). Each step consists in adding at random an elementary lattice vector
[(1, 0)a, (0, 1)a, where a is the lattice spacing] to the walk, as denoted by the
arrows. (b) Same as (a) but for a non-reversal random walk (NRRW), where
immediate reversals are forbidden. (c) Two examples of self-avoiding walks (SAWs),
where visiting any lattice site more than once is not allowed: trials where this
happens in a simple random sampling construction of the walk are discarded. From
Kremer and Binder [2].
!3=2
3 R2
exp
2
2 R
127
R2 2 hR2 i2
hR2 i2
2
3
128
Binder et al.
GibbsDi Marzio entropy theory of the glass transition [8,9], also based on a
lattice description, has been critically examined by a corresponding
simulation recently [10]. The key point is that not only has the
configurational entropy a well-defined meaning for a lattice model, but
also algorithms can be devised to accurately sample the entropy S of manychain systems [10], while in general the entropy is not a straightforward
output of computer simulations [3]. In addition, the simulation can then
extract precisely the quantities used in the theoretical formulation from the
simulated model as well, and hence a stringent test of the theory with no
adjustable parameters whatsoever becomes possible.
A third motivation is that with present day computers even for medium
chain length N of the polymers it is very difficult to study collective longwavelength phenomena (associated with phase transitions, phase coexistence, etc.), since huge linear dimensions L of the simulation box and a huge
number Nm of (effective) monomers in the simulation are required. For
example, for a finite size scaling [11] study of the nematicisotropic phase
transition in a model for a melt of semiflexible polymers a lattice with
L 130 was used, and with the bond fluctuation model [12,13] at volume
fraction of about 1/2 occupied sites this choice allowed for 6865 chains
of length N 20, i.e., 137,300 effective monomers, to be simulated [14].
Similarly, for a study of interfaces between unmixed phases of polymer
blends a lattice model with 512 512 64 sites was used, i.e., more than 16
million sites, and more than a million effective monomers (32,768 chains
of length N 32) [15]. Dealing with such large systems for more realistic
off-lattice models would be still very difficult.
Finally, we emphasize that lattice models are a very useful testing ground
for new algorithms that allow more efficient simulations of polymer
configurations under suitable circumstances. Many of the algorithms that
now are in standard use also for off-lattice models of polymerssuch as the
slithering snake algorithm [16,17], the pivot algorithm [18,19], the
configurational bias algorithm [20] (see also Chapter 7 by T.S. Jain and
J.J. de Pablo), and the chain breaking algorithm [21,22]were first
invented and validated for lattice models.
Why is there a need for these many different algorithms at all? Of course,
there would be no need for these algorithms if one wanted to simulate only
the simple random walk of Fig. 1a. But in fact, this is not a useful model for
polymer chains in most cases, since it allows arbitrarily many effective
monomers to sit on top of each other, ignoring excluded volume interactions
completely. A slightly better variant of the RW is the NRRW, Fig. 1b,
where immediate reversals are forbidden, but it is possible that loops are
formed where the NRRW crosses itself. The NRRW is still straightforward
to programat each step there are z 1 choices to proceed, from which one
129
R2 C1 a2 N,
N!1
0
R2 C1
a2 N 2 ,
N!1
where C1 is another constant, and the exponent differs from 1/2, 0.59
in d 3 dimensions and 3/4 in d 2 dimensions [24,5]. The behavior of
Eq. (6) is reproduced by the SAW model (Fig. 1c), and one can also predict
the corresponding chain partition function,
ZN / N 1 zN
eff ,
N!1
130
Binder et al.
or better of a NRRW (see Fig. 2 for pseudo-codes), and whenever the walk
intersects, thus violating the SAW constraint, the resulting configuration
has to be discarded, and a new trial configuration is generated (see Fig. 3 for
pseudo-code). Obviously, for large N the fraction of trials that is successful
becomes very small, since it is just given by the ratio of the respective
partition functions,
SAW
zeff N
ZN
1
/
N
NRRW
z 1
ZN
Since zeff<z 1 (e.g., zeff 2.6385 for the square lattice, zeff 4.6835 for
the simple cubic lattice, see Kremer and Binder [2] for various other lattices),
the fraction of successful constructions decreases exponentially with increasing N, namely as exp(
N) with
ln[zeff/(z 1)] 0.1284 (square lattice)
and 0.08539 (simple cubic lattice). This problem that the success rate
becomes exponentially small for large N is called the attrition problem,
131
FIG. 2 Pseudo-code for the simulation of a random walk (RW; code above the line)
and a non-reversal random walk (NRRW; code below the line) via simple sampling.
In both cases, the walks consist of N steps (innermost for loop), and the construction
of the walks is repeated M times (outer for loop) to improve statistics. Whereas these
repeated constructions always lead to a walk of N steps for the RW, they are not
always completed successfully for the NRRW because immediate back-folding can
occur. Then, the hitherto generated walk has to be discarded and the construction
must be resumed from the beginning (jump to start). In both cases, a new step is
appended to the walk by the subroutine add-next-step, in which ranf denotes a
random number uniformly distributed between 0 and 1 i.e., 0 ranf<1.
132
Binder et al.
equally probable walks, one has to correct for this bias by not counting each
chain with the weight WN 1 in the sampling, as one would do for simple
sampling, but rather with a weight,
W N fr i g
N
Y
ki =z0
i1
For large N, this weight varies over many orders of magnitude, however,
and hence the analysis of the accuracy reached is rather difficult in practice.
In fact, Batoulis and Kremer [27] demonstrated that the expected errors
increase exponentially with increasing N.
In principle, this method can also be applied to multichain systems, but
the problem of correcting for the bias becomes even more severe. In practice,
one therefore has to resort to the configurational bias method which is
an extension of the Rosenbluth sampling (see Chapter 7). But, inversely
restricted sampling is still one of the possible options for not too large
133
10
Thus the acceptance rate decreases only with a (small) power of N rather
than exponentially.
Still another useful technique is the enrichment method [30]. This early
work presented the first and compelling evidence that the exponent
[Eq. (6)] has the value 0.59 (d 3) and 0.75 (d 2), and hence is a
nice example of important discoveries made by Monte Carlo simulation.
The simple idea of the enrichment method is to overcome attrition by
using successfully generated short walks of length s not only once, but p
times: one tries to continue a chain of s steps to 2s steps by p different simple
sampling constructions. Then, the number N2s of 2s step chains is
N2s pNs exp
s
11
where
is the attrition constant. This process is continued for blocks of s
steps up to the desired chain length. One must fix p ahead of time to avoid a
bias. The best choice is p exp(
s), since taking p
exp(
s) does not reduce
the attrition problem enough, while p> exp(
s) would lead to an exploding
size of the sample and all walks are highly correlated. The fact that the
generated chains are not fully statistically independent and that therefore
the judgement of the accuracy of the procedure is a subtle problem is the
main disadvantage of the method.
There exist many extensions and variants for all these algorithms. For
example, for a star polymer with s arms one can use an enrichment
technique where one tries to add one step at each arm, and at each size of the
star this attempt is repeated p times [31]. It is also possible to combine
various methods suitably together. For example, Rapaport [32] combined
134
Binder et al.
12
135
FIG. 4 Various examples of dynamic Monte Carlo algorithms for SAWs: bonds
indicated as wavy lines are moved to new positions (broken lines), other bonds are
not moved. (a) Generalized VerdierStockmayer algorithms on the simple cubic
lattice showing three types of motion: end-bond motion, kink-jump motion,
crankshaft motion; (b) slithering snake (reptation) algorithm; (c) pivot
(wiggle) algorithm.
In the athermal case (the standard SAW problem) each configuration has
exactly the same weight (for single chains the normalized probability simply
is Peq(X) 1/ZN). Then, Eq. (12) says that the probability to select a motion
X1 ! X2 must be the same as the probabilty for the inverse motion,
X01 ! X1 . One has to be very careful to preserve this symmetry in the
actual realization of the algorithm, in particular if different types of move
are carried out in the simulation (e.g., Fig. 4). This can be achieved by
randomly choosing one of the possible moves and the new lattice sites to
which it should shift the monomer (or monomers), and by rejecting this
choice if it violates the excluded volume constraint. It would be completely
wrong if one first checked for empty sites and made a random choice only
between such moves that lead the monomers to these sites.
136
Binder et al.
13
Following Metropolis et al. [1] one can take the same transition probability
as in the athermal case if H 0, but multiplied by a factor exp(H/kBT)
if H > 0. Then, Eq. (13) is automatically satisfied at finite temperature,
if it was fulfilled in the athermal case.
Thus, at every step of the algorithm one performs a trial move X1 ! X2.
If W(X1 ! X2) is zero (excluded volume restriction being violated), the move
is not carried out, and the old configuration is counted once more in the
averaging. If W(X1 ! X2) is unity, the new configuration is accepted,
counted in the averaging, and becomes the old configuration for the next
step. If 0<W<1 we need a (pseudo-) random number x uniformly
distributed between zero and one. We compare x with W: If W x,
we accept the new configuration and count it, while we reject the trial
configuration and count the old configuration once more if W<x.
In the limit where the number of configurations M generated tends to
infinity, the distribution of states X obtained by this procedure is
proportional to the equilibrium distribution Peq(X), provided there is no
problem with the ergodicity of the algorithm (this point will be discussed
later). Then, the canonical average of any observable A(X) is approximated
by a simple arithmetic average,
hA i A
M
X
1
AX
M M0 M
14
where we have anticipated that the first M0 configurations, which are not yet
characteristic of the thermal equilibrium state that one wishes to simulate,
are eliminated from the average. Both the judgements of how large M0
should be taken and how large M needs to be chosen to reach some desired
statistical accuracy of the result, are hard to make in many cases. Some
guidance for this judgement comes from the dynamic interpretation of the
Metropolis algorithm [5,35], to which we turn next. This interpretation also
is very useful to evaluate the efficiency of algorithms.
e with the label of
We associate a (pseudo-) time variable t0 =N
e is the total number of
successively generated configurations, where N
137
At0 dt0
15
t0
16
138
Binder et al.
17
139
Note that the unit of time has been chosen such that each monomer on
average attempts to move once, and we have implied that the slowest modes
involve long distance properties of the order of the chains linear dimension.
Thus, the chain configuration should be fully
relaxed when the center of
p
mass has traveled a distance of the order of hR2 i. This argument implies
that the diffusion constant of the chain scales with chain length as
DN / N 1
18
140
Binder et al.
FIG. 6 (a) Example of a SAW on the square lattice that cannot move if the
slithering snake algorithm is used. (b) Example of a SAW on the square lattice
that cannot move if a combination of the slithering snake algorithm and the
VerdierStockmayer algorithm is used.
A0Bt hAihBi
AB t
hABi hAihBi
19
A0Bt A0Bt
1
tobs t t0
tobs t
At Bt t dt
t0
20
141
Even if one is not interested in the study of these correlations per se, they are
important because they control the statistical error of static observables.
Thus, if we estimate the error A of an average A [Eqs. (14), (15)] as follows
(there are n observations of A, i.e., A(t
) A
,
1, . . . , n)
!2
n
1X
A A
A
A
n
1
2
21
1
2
n
AA o
A hA i2 1 2
A2
,
n
t
for
tobs nt AA
22
AA
AA t dt
23
142
Binder et al.
been advocated in the literature. For a system of many short chains, where
the linear dimension of the box L is much larger than the length Rmax bN
of a chain fully stretched to its maximal extension, one can put all chains in
this fully stretched state parallel to each other in a lattice direction onto the
lattice. Excluded volume interactions can then be trivially respected. The
time t0 necessary to forget the initial state that is not characteristic of
equilibrium can then simply be measured via the decay of the orientational
order of the chains, which is initially present in the system. Although in
principle this is a good method, it is clear that for large N and close to
volume fractions that correspond to melt densities (i.e., 0 0.5) the
resulting time t0 is rather large. But the most important drawback is that it
cannot be used at all for systems wherepbN
exceeds L (note that in the bond
fluctuation model b is between 2 and 10 lattice spacings, and simulations
up to N 2048 for semidilute solutions using lattice linear dimensions
L 400 have been carried out [51]). In this case one has the following
alternatives: Either the chains in the initial state are put on the lattice with
the RosenbluthRosenbluth method or the configurational bias method, or
one uses NRRW chains, ignoring excluded volume initially. Then during the
initial equilibrium run only moves are accepted that do not increase the
number N of lattice sites for which the excluded volume constraint is
violated, and one has to run the simulation until N 0. There is no
guarantee, of course, that this latter algorithm works at all for every initial
configuration. An alternative would be to replace the hard excluded volume
constraint by a soft one, introducing an energy penalty U for every double
occupancy of a lattice site, and gradually increase U ! 1 during the
initial equilibration run. Despite the fundamental importance of producing
well-equilibrated initial configurations of many-chain (athermal) systems,
we are not aware of any systematic comparative evaluation of the efficiency
of the various approaches outlined above.
It should also be emphasized that the best method of choosing an initial
state for the simulation of a dense polymer system may depend on the
physical state of the system that one wishes to simulate. For example, the
configurational bias method is a good choice for creating initial configurations of long flexible chains in semidilute solutions [51] or melts [43], but not
for melts of stiff chains in a nematic state [14]: In this case, it is much better
to start with a monodomain sample of a densely filled lattice of fully
stretched chains, remove at random chains such that the desired volume
fraction of occupied lattice sites is reached, and then start a relaxation run
with slithering snake and local moves. Conversely, if we start from a
configuration with no initial nematic order, it already takes a very long time
to create small nematic domains out of chain configurations that are initially
random-walk like, and relaxing the system further so that the multidomain
143
144
Binder et al.
relaxation is provided, this diagonal bond model is also advocated for the
study of ordered phases in models of block copolymers [55].
ACKNOWLEDGMENTS
This chapter has benefited from many fruitful interactions with K. Kremer
and W. Paul.
REFERENCES
1. Metropolis, N.; Rosenbluth, A.W.; Rosenbluth, M.N.; Teller, A.H.; Teller, E.
J. Chem. Phys. 1953, 21, 1087.
2. Kremer, K.; Binder, K. Monte Carlo Simulations of Lattice Models for
Macromolecules. Comput. Phys. Rep. 1988, 7, 259.
3. Binder, K.; Heermann, D.W.; Monte Carlo Simulation in Statistical Physics. An
Introduction, Ed.; 3rd Ed, Springer: Berlin, 1997.
4. Milchev, A.; Binder, K.; Heermann, D.W. Z. Phys. 1986, B63, 521.
145
5. Sokal, A.D. Monte Carlo Methods for the Self-Avoiding Walk. In Monte Carlo
and Molecular Dynamics Simulations in Polymer Science; Binder, K., Ed.;
Oxford University Press: New York, 1995; 47 pp.
6. Flory, P.J. Principles of Polymer Chemistry; Cornell University Press: Ithaca,
New York, 1953.
7. Sariban, A.; Binder, K. Macromolecules 1988, 21, 711.
8. Gibbs, J.H.; Di Marzio. J. Chem. Phys. 1958, 28, 373.
9. Gibbs, J.H.; Di Marzio, J.D. J. Chem. Phys. 1958, 28, 807.
10. Wolfgardt, M.; Baschnagel, J.; Paul, W.; Binder, K. Phys. Rev. 1996, E54,
1535.
11. Binder, K. In Computational Methods in Field Theory; Gausterer, H., Lang,
C.B., Eds.; Springer: Berlin-New York, 1992; 59 pp.
12. Carmesin, I.; Kremer, K. Macromolecules 1988, 21, 2819.
13. Deutsch, H.-P.; Binder, K. J. Chem. Phys. 1991, 94, 2294.
14. Weber, H.; Paul, W.; Binder, K.; Phys. Rev. 1999, E59, 2168.
15. Muller, M.; Binder, K.; Oed, W. J. Chem. Soc. Faraday Trans. 1995, 91, 2369.
16. Kron, A.K. Polymer Sci. USSR. 1965, 7, 1361.
17. Wall, F.T.; Mandel, F. J. Chem. Phys. 1975, 63, 4592.
18. Lal, M. Mol. Phys. 1969, 17, 57.
19. Madras, N.; Sokal, A.D. J. Stat. Phys. 1988, 50, 109.
20. Siepmann, I. Mol. Phys. 1990, 70, 1145.
21. Olaj, O.F.; Lantschbauer, W. Makromol. Chem., Rapid Commun. 1982,
3, 847.
22. Manseld, M.L. J. Chem. Phys. 1982, 77, 1554.
23. Mattice, W.L.; Suter, U.W. Conformational Theory of Large Molecules, the
Rotational Isomeric State Model in Macromolecular Systems; Wiley: New
York, 1994.
24. de Gennes, P.G. Scaling Concepts in Polymer Physics; Cornell University Press:
Ithaca, 1979.
25. Caracciolo, S.; Serena Causo, M.; Pelissetto, A. Phys. Rev. 1998, E57, R1215.
26. Rosenbluth, M.N.; Rosenbluth, A.W. J. Chem. Phys. 1955, 23, 356.
27. Batoulis, J.; Kremer, K. J. Phys. 1988, A21, 127.
28. Meirovitch, H. J. Phys. 1982, A15, 735.
29. Alexandrovicz, Z. J. Chem. Phys. 1969, 51, 561.
30. Wall, F.T.; Erpenbeck, J.J. J. Chem. Phys. 1959, 30, 634.
31. Ohno, K.; Binder, K. J. Stat. Phys. 1991, 64, 781.
32. Rapaport, D.C. J. Phys. 1985, A18, 113.
33. Grassberger, P. Phys. Rev. 1997, E56, 3682.
34. des Cloizeaux J.; Jannink, G. Polymers in Solution: Their Modeling and
Structure; Oxford University Press: Oxford, 1990.
35. Muller-Krumbhaar, H.; Binder, K. J. Stat. Phys. 1973, 8, 1.
36. Binder, K.; Paul, W. J. Polym. Sci. 1997, B35, 1.
37. Rouse, P.E. J. Chem. Phys. 1953, 21, 1272.
38. Wittmann, H.-P.; Kremer, K. Comp. Phys. Comm. 1990, 61, 309.
39. Wittmann, H.-P.; Kremer, K. Comp. Phys. Comm. 1991, 71, 343.
146
Binder et al.
40. Paul, W.; Baschnagel, J. Monte Carlo Simulations of the Glass Transition of
Polymers. In: Monte Carlo and Molecular Dynamics Simulations in Polymer
Science; Binder, K., Ed.; Oxford Univeristy Press: New York, 1995; 307 pp.
41. Baschnagel, J.; Dynamic Properties of Polymer Melts above the Glass
Transition: Monte Carlo Simulation Results. In Structure and Properties of
Glassy Polymers, ACS Symposium Series No. 710; Tant, M.R., Hi, A.J.,
Eds.; American Chemical Society: Washington, DC, 1998; 53 pp.
42. Paul, W.; Binder, K.; Heermann, D.W.; Kremer, K. J. Phys. 1991, II (France)
1, 37.
43. Muller, M.; Wittmer, J.; Barrat, J.-L. Europhys. Lett. 2000, 52, 406412.
44. Kreer, T.; Baschnagel, J.; Muller, M.; Binder, K. Macromolecules 2001, 34,
11051117.
45. Verdier, P.H.; Stockmayer, W.H. J. Chem. Phys. 1962, 36, 227.
46. Verdier, P.H. J. Chem. Phys. 1966, 45, 2122.
47. Hilhorst, H.J.; Deutsch, J.M. J. Chem. Phys. 1975, 63, 5153.
48. Madras, N.; Sokal, A.D. J. Stat.Phys. 1987, 47, 573.
49. Wolfgardt, M.; Baschnagel, J.; Binder, K. J. Phys. 1995, II (France) 5, 1035.
50. Doi, M.; Edwards, S.F. Theory of Polymer Dynamics; Clarendon: Oxford 1986.
51. Muller, M.; Binder, K.; Schafer, L. Macromolecules 2000, 33, 4568.
52. Paul, W.; Binder, K.; Kremer, K.; Heermann, D.W. Macromolecules. 1991, 24,
6332.
53. Shaer, J.S. J. Chem. Phys. 1994, 101, 4205.
54. Shaer, J.S. J. Chem. Phys. 1995, 103, 761.
55. Dotera, T.; Hatano, A. J. Chem. Phys. 1996, 105, 8413.
56. Binder, K. Monte Carlo Studies of Polymer Blends and Block Copolymer
Thermodynamics. In Monte Carlo and Molecular Dynamics Simulations in
Polymer Science; Binder, K., Ed.; Oxford University Press: New York, 1995;
356 pp.
57. Binder, K.; Muller, M. Monte Carlo Simulations of Block Copolymers. Curr.
Opinion Colloid Interface Sci. 2000, 5, 315.
58. Dickman, R. J. Chem. Phys. 1987, 86, 2246.
59. Muller, M.; Paul, W. J. Chem. Phys. 1994, 100, 719.
60. Kumar, S.K.; Szleifer, I.; Panagiotopoulos, A. Phys. Rev. Lett. 1991, 66, 2935.
61. Lai, P.-Y.; Binder, K. J. Chem. Phys. 1993, 98, 2366.
62. Kinzel, W.; Reents, G. Physics by Computer; Springer: Berlin, 1998.
5
Simulations on the Completely
Occupied Lattice
TADEUSZ PAKULA Max-Planck-Institute for Polymer Research, Mainz,
Germany, and Technical University of Lodz, Lodz, Poland
I.
INTRODUCTION
There is a large number of different methods (algorithms) used for simulation of polymers on the coarse grained molecular scale [16]. Models of
polymers considered in this range usually disregard the details of the chemical constitution of macromolecules and represent them as assemblies of
beads connected by nonbreakable bonds. In order to speed up recognition
of neighboring beads, the simplified polymers are often considered to be on
lattices with beads occupying lattice sites and bonds coinciding with lines
connecting neighboring sites. The methods used for simulation of the lattice
polymers can be considered within two groups. The first group includes
algorithms that can operate only in systems with a relatively large fraction
of lattice sites left free and the second group includes algorithms suitable for
lattice systems in which all lattice sites are occupied by molecular elements.
Whereas, the systems considered within the first group should be regarded
as lattice gases, the systems treated within the second group of methods can
be considered as lattice liquids. This reflects the differences in the mechanisms of molecular rearrangements used within these two groups to move the
systems through the phase space in order to reach equilibrium. The latter
problem concerns the physical nature of molecular rearrangements in dense
polymer systems and is related to the microscopic mechanism of motion in
molecular liquids. Unfortunately, this is not yet solved entirely. The most
popular picture is that a molecule, or a molecular segment in the case of
polymers, needs a free space in its neighborhood in order to make a translational step beyond the position usually occupied for quite a long time. Most
simulation methods assume this picture and consequently a relatively large
portion of the space in the form of free lattice sites has to be left free to allow
148
Pakula
a reasonable mobility [15]. On the other hand, there is only one simulation
method that assumes a cooperative nature of molecular rearrangements on
the local scale and which does not require such a reserve space to allow the
molecular mobility. The method, which uses the mechanism of cooperative
rearrangements for polymer systems is the Cooperative Motion Algorithm
(CMA) suggested originally in [6] and presented in improved form in subsequent publications [710]. A mechanism of this kind has been formulated
recently also for low molecular weight liquids, based on assumptions taking
into account both a dense packing of molecules interacting strongly due to
excluded volume and a condition of preservation of local continuity of the
system. The later version of the microscopic mechanism, called the Dynamic
Lattice Liquid (DLL) model, has been described in detail [1114].
The aim of this chapter is to present a background to the simulations of
molecular lattice systems on a completely filled lattice (the DLL model) as
well as to show some examples of application of the CMA method for
simulation of static and dynamic properties of various polymers.
149
The model that has been proposed to answer this question is based on a
lattice structure and is called the Dynamic Lattice Liquid model [11,12]. The
positions of the molecules are regarded as coinciding with the lattice sites.
The lattice is, however, considered only as a coordination skeleton defining
the presence of nearest neighbors but not precisely the distances between
them. Under the condition of uniform and constant coordination (z), all
lattice sites are assumed to be occupied. It is assumed that the system has
some excess volume so that molecules have enough space to vibrate around
their positions defined by lattice sites but can hardly move over larger distances because all neighboring lattice sites are occupied. The vibrations are
assumed to take place with a mean frequency V 1/ V, related to the short
time scale ( V). Each large enough displacement of a molecule from the
mean position defined by the lattice is considered as an attempt to move
to the neighboring lattice site. For simplicity, the attempts are assumed to
take place only along the coordination lines but are independent and randomly distributed among z directions. Most of the attempts remain unsuccessful, because it is assumed that all the time the system remains quasicontinuous, which means that no holes of molecular sizes are generated
and multiple occupations of lattice sites are excluded (excluded volume
condition). The continuity condition of the system, for the vector field of
150
Pakula
@
"
@t
with
"!0
where J is the current of displacing molecules and is the density and should
not deviate considerably from 1 if all lattice sites remain occupied; " is
introduced as a small number allowing density fluctuations but excluding
generation of holes. Consequently, most of the attempted displacements
have to be compensated by a return to the initial position within the
period V. Only those attempts can be successful that coincide in such a
way, that along a path including more than two molecules the sum of the
displacements is close to zero. In the system considered, only the paths in a
form of closed loops can satisfy this condition (Fig. 1b). The probability
to find an element taking part in such coincidences will determine the
probability of longer range rearrangements and the slower time scale ( ).
A determination of this probability reduces to counting the number of
self-avoiding n-circuits on a given lattice, i.e., to a problem that has
already been extensively considered in the literature [23]. The generally
accepted result describing the probability to find the self-avoiding circuits
is given by
pn Bnh
n
151
this way, all vectors that do not contribute to correlated sequences (circuits)
satisfying the continuity condition Eq. (1) are set to 0. This concerns, for
example, such situations as attempts of displacements in opposite directions
(e.g., area 1 in Fig. 2) or attempts of motion starting from lattice sites,
towards which any other bead is not trying to move at the same time
(e.g., area 2 in Fig. 2). What remains after this operation are vectors contributing to a number of closed loop traces, which are considered as paths of
possible rearrangements (e.g., areas 3 in Fig. 2). If the system is considered
as athermal, all possible rearrangements found are performed by shifting
beads along closed loop traces, each bead to the neighboring lattice site.
The above procedure, consisting of (1) generation of motion attempts, (2)
elimination of unsuccessful attempts, and (3) displacing beads within closed
loop paths, is considered as a Monte Carlo step and assumed to take place
within the time scale V. The procedure is exactly repeated in subsequent
time steps always with a new set of randomly chosen directions of attempted
displacements.
The system treated in this way can be regarded as provided with the
dynamics consisting of local vibrations and occasional diffusional steps
resulting from the coincidence of attempts of neighboring elements to
displace beyond the occupied positions. Within a longer time interval, this
kind of dynamics leads to displacements of individual beads along random
walk trajectories with steps distributed randomly in time. Small displacements related to vibrations of beads could be considered explicitly but are
neglected here, for simplicity.
152
Pakula
153
Here, only the athermal version of the DLL model has been described. It
has been shown elsewhere [1113] that the model is able to represent the
temperature dependent dynamics of systems as well. This, however, requires
additional assumptions concerning free volume distribution and related
potential barriers for displacements of individual elements. It has been
demonstrated [1113] that in such a version the model can describe a broad
class of temperature dependencies of relaxation rates in supercooled liquids
with the variety of behavior ranging between the extremes described by the
free volume model on one side and the Arrhenius model on the other side.
154
Pakula
rearrangements, which has only been tested precisely for an athermal nonpolymeric system. The distribution of rearrangement sizes within the DLL
algorithm is well described by Eq. (2) with parameters
0.847, h 3, and
B 0.4, whereas the distribution of rearrangement sizes in the CMA is
considerably broader. It has been established that the two distributions
differ by the factor n2. This difference results in much higher efficiency of
simulation by means of CMA, achieved however at the expense of some
simplifications concerning short time dynamics of the system with respect to
that in the DLL model.
On the other hand, it has been proven for melts of linear polymers that
switching between DLL and CMA does not influence static properties of the
system. The CMA, similarly to the DLL algorithm, involves only conformational changes within polymers which all the time preserve their identities
given by the number of elements and by their sequences in individual chains.
In the CMA as in the DLL, a variety of conformational rearrangements
results from the algorithm and cannot be precisely specified. An important
advantage of both DLL and CMA methods with respect to other simulation algorithms of polymers on lattices, consists in allowed moves of chain
fragments both along contours of chains and transversally to the chain
contours. This probably makes both algorithms ergodic.
The CMA has been applied successfully for simulation of static and
dynamic properties of a variety of dense complex polymeric systems. Some
examples of such applications are presented in subsequent sections.
155
N
1 XX
bi tbi 0
Nn n i
156
2.
Pakula
3.
1X
R0 Rt
n n
with end-to-end vectors R(0) and R(t) at time t 0 and t, respectively; and
the mean squared displacements of monomers and of the centers of
mass of chains
1 XX
2
rm t rm 0
Nn n N
1X
hr2cm i
rcm t rcm 0 2
n n
hr2m i
5
6
where rm(t) and rcm(t) are monomer and chain center of mass coordinates at time t, respectively.
The first two quantities allow determination of relaxation times of
corresponding objects in the model (bonds and chains) and the last two
allow determination of diffusion constants. Chain length dependencies of
the self-diffusion constant and of the relaxation times of bonds and chains
are presented in Fig. 3. These results show that the dynamic behavior of
FIG. 3 Chain length dependencies of (a) self-diffusion constants of chains and (b)
relaxation times of bonds and end-to-end vectors in model systems of linear chain
melts.
157
158
Pakula
FIG. 4 Illustration of molecular objects of the same total chain length but different
topology and a comparison of intrachain segment densities (a) and center of mass
pair correlation functions (b) for three computer simulated melts with different
molecular objects: linear chains, multiarm stars, and microgels.
and interpenetrate each other, the stars and microgels constitute compact
objects with intramolecular density almost reaching the density of the
system. This means that the latter molecules do not interpenetrate each
other and remain at well correlated distances as indicated by the center
of mass pair correlation functions with distinct maxima at preferred
intermolecular distances (Fig. 4b).
The dynamics of the three systems is characterized in Fig. 5, where
various correlation functions are compared. For the linear chain system
the segment position (s) and the end-to-end vector (R) [see Eq. (4)]
159
FIG. 5 Dynamical properties of three simulated polymer melts with various molecules: (a) linear chains, (b) stars, and (c) microgels. Various correlation functions
are shown: s segment position autocorrelation, R end-to-end vector autocorrelation for linear chains, A arm end-to-end autocorrelation for stars, B
autocorrelation of microgel chain segments of length N 20, pos position
autocorrelation for all types of molecules.
pos t
N
1 XX
Ci 0Ci t
Nn n i
160
Pakula
C. Block Copolymers
In order to illustrate the application of the CMA to simulate heterogeneous
systems, we present here results concerning properties of a diblock copolymer melt considered in a broad temperature range including both the homogeneous and the microphase separated states. Only symmetric diblock
copolymers, of composition f 0.5 of repulsively interacting comonomers
A and B, are shown here. The simulation, in this case, allows information
about the structure, dynamics, and thermodynamic properties of systems
[9,3740] to be obtained.
In the case of copolymers, two types of partially incompatible monomers
(A and B) are considered, characterized by direct interaction parameters "ij,
and often it is assumed that the energy of mixing is given only by the
161
z
X
"kl i=z
i1
and (2) the specific heat calculated via the fluctuation-dissipation theorem
cV
hE 2 i hEi2
kB T 2
162
Pakula
FIG. 6 Temperature dependencies of: (a) the average interaction energy per
monomer, (b) the specific heat, and (c) concentration distributions in small volume
elements consisting of the nearest neighbors of each chain segment. Characteristic
structures corresponding to various temperature ranges are illustrated.
163
changed at temperatures close to TODT where at first a plateau and later two
maxima corresponding to two coexisting phases are detected. At TODT, a
sudden change transforms the system to a state with well defined microphases indicated by the most probable local concentrations corresponding
to pure components. These results indicate three characteristic ranges of
thermodynamic behavior of the system assigned as (1) disordered, (2)
weakly segregated, and (3) strongly segregated regimes appearing with
decreasing temperature. Structures of simulated systems corresponding to
these regimes are illustrated in Fig. 6 by assuming different colors for different copolymer constituents.
The structure of the simulated block copolymer systems has been characterized in detail [3840]. Temperature dependencies of various structural
parameters have shown that all of them change in a characteristic way in
correspondence to TODT. The microphase separation in the diblock copolymer system is accompanied by chain extension. The chains of the diblock
copolymer start to extend at a temperature well above that of the transition
to the strongly segregated regime. This extension of chains is related also to
an increase of local orientation correlations, which appear well above the
transition temperature. On the other hand, the global orientation correlation factor remains zero at temperature above the microphase separation
transition and jumps to a finite value at the transition.
In order to get information about dynamic properties of the system
various quantities have been monitored with time at equilibrium states corresponding to various temperatures [3840]: the mean squared displacement
of monomers, hr2m i, the mean squared displacement of the center of mass of
chains, hr2cm i, the bond autocorrelation function, b(t)), the end-to-end
vector autocorrelation function, R(t), and the autocorrelation of the endto-end vector of the block, bl(t). On the basis of these correlation functions,
various quantities characterizing the dynamic properties of the systems can
be determined, i.e., the diffusion constant of chains and various relaxation
times corresponding to considered correlations.
Examples of various correlation functions for the diblock copolymer
system at high and at low temperatures are shown in Fig. 7. It has been
observed that at high temperatures (T/N 1), the systems behave like a
homogeneous melt. All correlation functions show a single step relaxation.
The fastest is the bond relaxation and the slowest is the chain relaxation
described by the end-to-end vector autocorrelation function. The relaxation
of the block is faster than the whole chain relaxation by a factor of approximately two. Such relations between various relaxation times in the disordered state of the copolymer can be regarded as confirmed experimentally
for some real systems, in which the dielectric spectroscopy allows distinction
of the different relaxation modes [41]. At low temperatures, drastic changes
164
Pakula
165
FIG. 8 (a) Mean square displacements of monomers and mean square displacements of chain centers of mass vs. time for the diblock copolymer system at various
temperatures. Temperature dependence of the self-diffusion constant of block
copolymer chains is shown in the inset. (b) Orientation correlation factor between
the end-to-end vector and the center of mass displacement vector of copolymer
chains at various temperatures above and below the ODT.
166
Pakula
in the phase separated regime the chains diffuse easily in directions parallel
to the interface. This is indicated by negative values of the orientation
correlation factor and means that the limitations in mobility of diblock
chains imposed by the morphology concern only the diffusion through the
interface.
Besides the above presented example, the CMA has been applied for
simulation of various other copolymer systems with more complex topology
[42] of macromolecules and other distributions of comonomers along chains
[43,44]. In all cases, the high computational efficiency of the method has
enabled detailed information about the structure and dynamics to be
obtained, including also the microphase separated states, in which the
dynamics becomes considerably slower.
V. IMPLEMENTATION DETAILS
There are essentially several versions of algorithms that are based on similar
ideas of performing cooperative rearrangements in dense molecular systems.
All versions are equally suitable both for simulation of assemblies of nonbonded beads representing small molecules and for simulation of assemblies
of lattice structures, which mimic polymer skeletons with various complexities. Some details essential for implementation of these methods for complex polymers on the fcc lattice, taken as an example, will be described
here. The efficiency of an algorithm strongly depends on details concerning
methods used for description of system elements, methods of recognition
and description of systems states, methods of rearranging the elements,
and finally on the programming methods, which allow fast accessibility to
large data arrays.
167
168
Pakula
In a model system with all lattice sites occupied, each bead at a position
(x, y, z) is described by a sequence of numbers defining its position in the
chain, m(x, y, z), the chain to which it belongs, k(x, y, z), and the connectivity
with other beads by means of up to four vectors, a(x, y, z), b(x, y, z),
c(x, y, z), and d(x, y, z), which point towards the bonded neighbors. Vector
orientation is given by a code number, cd, assuming values from 1 to 13,
which describe the 12 possible orientations of the sitesite vectors in the 12
coordinated fcc lattice, as illustrated in Fig. 9b. The value d 13 is used as a
code for the vector of length equal to 0, which is used to describe chain ends
and single beads representing elements of simple liquid (e.g., solvent). An
example of a linear chain of length nw is illustrated in Fig. 9c. It is seen that
each bond is described by two vectors (a and b) pointing in opposite directions. This has the advantage that, at each position, (x, y, z), the connectivity
and consequently the local conformation can immediately be established.
Table 1 shows the conformation code, con(b,a), used for the description of
various angles between the neighboring bonds along the chain contour.
Positions of beads are described by the coordinates (x, y, z). It has appeared
useful to introduce matrices xnp, ynp, and znp by means of which coordinates (xn, yn, zn) of neighboring lattice sites in direction cd can be found as
xn xnp(x, cd ), yn ynp( y, cd ), and zn znp(z, cd ). These matrices can
consider system sizes and the type of boundary conditions (e.g., periodic).
TABLE 1
Angle between
neighboring
bonds
Illustration
Code (con)
Range of application
9
>
>
>
>
>
>
>
=
180
120
90
60
b 13, m 1
a 13, m nw
Chain interior
1<m<nw
a<13 and b<13
>
>
>
>
>
>
>
;
Ends of chains
169
170
Pakula
171
remaining chain structure and (ii) the searching path enters the chain contour and leaves it after a short walk along the chain (Fig. 11b). In the first
case, the motion can be performed without extending any bonds, whereas, in
the second case, in each step of the searching point, one and only one bond
is attempted to be extended but this extension must immediately be relaxed
by motion of other beads along the chain as long as the searching point
leaves the chain at places with suitable local conformations or through chain
ends. The latter forces the searching point to go along the chain contour for
some time. In fact, the first possibility can also be considered as a special
172
Pakula
case of the second possibility, in which the path of the searching point along
the chain is reduced to zero.
A very high efficiency of the implementation of the CMA for polymers
has been achieved, when it was recognized that there is a limited number of
situations in which the searching point can meet while making a walk
through a system containing beads bonded to polymers. A specific situation
code, pat cr(cd, a, b), is used in order to recognize these situations for
various orientations of bonds, various directions of chains, and various
local conformations. There are 15 situations distinguished; these are listed
in Table 2. From these 15 cases, one corresponds to the case, when propagation of motion of the searching point becomes impossible ( pat 1),
because the attempted movement would extend two bonds simultaneously.
Two cases, pat 2 and pat 3, describe the motion of the searching point
along a chain in two different directions. In cases pat 4, 5 and pat 10, 11,
the searching point can enter the chain contour through a chain end or
through a suitable local conformation, respectively. When pat 6, 7 or
pat 13, 14, the searching point leaves the chain contour. The other cases
correspond to situations of the type (i) illustrated in Fig. 11a. In all cases,
173
The Situation Code [pat cr(cd, a, b)] used in the CMA Procedures
Situation before ! after
Description
10
11
12
Searching point
contour
13
14
15
crosses
the
chain
174
Pakula
when the searching point is not forced to move along the chain, the direction
of motion is chosen randomly. Coding of situations and resulting rearrangements can considerably speed up the simulation. The described codes
should be considered as suitable examples. Other, maybe more effective,
solutions are certainly possible.
The various versions of the CMA are suitable for implementation and
usage on personal computers. With recently available technology, dense
polymeric systems of up to one million beads organized in various macromolecular architectures can be simulated in a reasonably inexpensive way.
REFERENCES
1.
2.
3.
4.
5.
6.
7.
175
8. Reiter, J.; Edling, T.; Pakula, T. J. Chem. Phys. 1990, 93, 837.
9. Gauger, A.; Weyersberg, A.; Pakula, T. Macromol. Chem. Theory Simul. 1993,
2, 531.
10. Gauger, A.; Pakula, T. Macromolecules 1995, 28, 190.
11. Pakula, T.; Teichmann, J. Mat. Res. Soc. Symp. Proc. 1997, 495, 211.
12. Teichmann, J. Ph.D. thesis, University of Mainz, 1996.
13. Pakula, T.J. Mol. Liquids 2000, 86, 109.
14. Polanowski, P.; Pakula, T. J. Chem. Phys. 2002, 117, 4022.
15. Barker, J.A.; Henderson, D. Rev. Mod. Phys. 1976, 48, 587.
16. Alder, B.J.; Wainwright, T.E. J. Chem. Phys. 1969, 31, 459.
17. Widom, A. Phys. Rev. A 1971, 3, 1394.
18. Kubo, R. Rept. Progr. Phys. 1966, 29, 255.
19. London, R.E. J. Chem. Phys. 1977, 66, 471.
20. Gotze, W.; Sjogren, L. Rep. Prog. Phys. 1992, 55, 241.
21. Cohen, M.H.; Grest, G.S. Phys. Rev. B 1979, 20, 1077.
22. Adam, G.; Gibbs, J.H. J. Chem. Phys. 1965, 43, 139.
23. Hughes, B.D. Random Walks and Random Environments; Clarendon Press:
Oxford, 1995.
24. Reiter, J. Macromolecules 1990, 23, 3811.
25. Reiter, J. Physica A 1993, 196, 149.
26. Pakula, T.; Geyler, S.; Edling, T.; Boese, D. Rheol. Acta 1996, 35, 631.
27. Pakula, T.; Harre, K. Comp. Theor. Polym. Sci. 2000, 10, 197.
28. Fox, T.G.; Flory, P.J. J. Appl. Phys. 1950, 21, 581.
29. Paul, W.; Binder, K.; Herrmann, D.W.; Kremer, K. J. Chem. Phys. 1991, 95,
7726.
30. Kolinski, A.; Skolnick, J.; Yaris, R. J. Chem. Phys. 1987, 86, 7164.
31. Fetters, L.J.; Kiss, A.D.; Pearson, D.S.; Quack, G.F.; Vitus, F.J.
Macromolecules 1993, 26, 647.
32. Roovers, J.; Lin-Lin Zbou, Toporowski, P.M.; van der Zwan, M.; Iatron, H.;
Hadjichristidis, N. Macromolecules 1993, 26, 4324.
33. Antonietti, M.; Pakula, T.; Bremser, W. Macromolecules 1995, 28, 4227.
34. Pakula, T. Comp. Theor. Polym. Sci. 1998, 8, 21.
35. Pakula, T.; Vlassopoulos, D.; Fytas, G.; Roovers, J. Macromolecules 1998, 31,
8931.
36. Grest, G.S.; Kremer, K.; Milner, S.T.; Witten, T.A. Macromolecules 1989, 22,
1904.
37. Pakula, T. J. Comp. Aided Mat. Des. 1996, 3, 329.
38. Pakula, T.; Karatasos, K.; Anastasiadis, S.H.; Fytas, G. Macromolecules 1997,
30, 8463.
39. Pakula, T.; Floudas, G. In Block Copolymers, Balta Calleja, F.J., Roslaniec, Z.,
Eds.; Marcel Dekker Inc.: New York, 2000; 123177.
40. Pakula, T. J. Macromol. Sci.Phys. 1998, B37, 181.
41. Karatasos, K.; Anastasiadis, S.H.; Floudas, G.; Fytas, G.; Pispas, S.;
Hadjichristidis, N.; Pakula, T. Macromolecules 1996, 29, 1326.
176
Pakula
42. Floudas, G.; Pispas, S.; Hadjichristidis, N.; Pakula, T.; Erukhimovich, I.
Macromolecules 1996, 29, 4142.
43. Vilesov, A.D.; Floudas, G.; Pakula, T.; Melenevskaya, E. Yu.; Birstein, T.M.;
Lyatskaya, Y.V. Macromol. Chem. Phys. 1994, 195, 2317.
44. Pakula, T.; Matyjaszewski, K. Macromol. Theory Simul. 1996, 5, 987.
45. Miller, P.J.; Matyjaszewski, K.; Shukla, N.; Immaraporn, B.; Gelman, A.;
Luokala, B.B.; Garo, S.; Siclovan, T.M.; Kickelbick, G.; Vallant, T.;
Homann, H.; Pakula, T. Macromolecules 1999, 32, 8716.
46. Polanowski, P. Ph.D. thesis, Technical University of Lodz, 2002.
6
Molecular Dynamics Simulations
of Polymers
VAGELIS A. HARMANDARIS and VLASIS G. MAVRANTZAS Institute
of Chemical Engineering and High-Temperature Chemical Processes, and
University of Patras, Patras, Greece
I.
2.
178
FIG. 1
3.
4.
5.
179
components: intermolecular (from atoms belonging to different polymer chains) and intramolecular (from atoms belonging to the same
chain).
The integration of the equations of motion follows with an appropriate method. The most popular of these will be described in detail
in the next section.
Actual measurements are performed (positions, velocities, energies,
etc., are stored) after the system has reached equilibration, periodically every Nk steps.
After completion of the central loop (N steps), averages of the measured quantities and of the desired properties are calculated and
printed.
@qk dt @q_ k
@L
@q_k
@H
,
@pk
p_k
@H
@qk
180
q_ k pk L
pi
,
mi
p_ i rri V
@V
Fi
@ri
hence
mi r i mi q_ i Fi
where Fi is the force acting on atom i. Solving the equations of motion then
involves the integration of the 3N second-order differential equations (6)
(Newtons equations).
The classical equations of motion possess some interesting properties, the
most important one being the conservation law. If we assume that K and V
do not depend explicitly on time, then it is straightforward to verify that
H_ dH=dt is zero, i.e., the Hamiltonian is a constant of the motion. In
actual calculations this conservation law is satisfied if there exist no
explicitly time- or velocity-dependent forces acting on the system.
A second important property is that Hamiltons equations of motion
are reversible in time. This means that, if we change the signs of all
the velocities, we will cause the molecules to retrace their trajectories
backwards. The computer-generated trajectories should also possess this
property.
There are many different methods for solving ordinary differential
equations of the form of Eq. (6). Criteria for the proper choice of an
algorithm include the following:
.
.
.
181
The algorithm should permit the use of a large time step dt.
The algorithm should be fast and require little memory.
7
8
9
10
11
182
12
vc t dt v p t dt c1 x
r c t dt r p t dt c2 x
13
14
r c t dt r p t dt c3 x
15
where ci, i 1, . . . , n are constants. The values of ci are such that they yield
an optimal compromise between desired level of accuracy and algorithm
stability [2].
The general scheme of an algorithm based on the predictor-corrector
method goes as follows:
1.
2.
3.
B. Verlet Methods
Algorithms in this family are simple, accurate, and, as we will see below,
time reversible. They are the most widely used methods for integrating the
classical equations of motion. The initial form of the Verlet equations [3] is
obtained by utilizing a Taylor expansion at times t dt and t dt
dt2
dt3
r t
rt Odt4
2
6
dt2
dt3
r t
rt dt rt dt vt
rt Odt4
2
6
rt dt rt dt vt
16
17
18
183
Two modifications of the Verlet scheme are of wide use. The first is the
leapfrog algorithm [3] where positions and velocities are not calculated at
the same time; velocities are evaluated at half-integer time steps:
dt
rt dt rt dt v t
2
dt
dt
v t
v t
dt r t
2
2
19
20
21
The problem of defining the positions and velocities at the same time can
be overcome by casting the Verlet algorithm in a different way. This is the
velocity-Verlet algorithm [3,6], according to which positions are obtained
through the usual Taylor expansion
rt dt rt dt vt
dt2
r t
2
22
dt
r t r t dt
2
23
with all accelerations computed from the forces at the configuration corresponding to the considered time. To see how the velocity-Verlet algorithm is
connected to the original Verlet method we note that, by Eq. (22),
rt 2dt rt dt dt vt dt
dt2
r t dt
2
24
dt2
r t
2
25
184
dt2
rt dt r t
2
26
27
which is indeed the coordinate version of the Verlet algorithm. The calculations involved in one step of the velocity algorithm are schematically shown
in [2, figure 3.2, page 80].
A sample code of the velocity-Verlet integrator is shown in Algorithm 1.
In this algorithm, N is the total number of atoms in the system and the
subroutine get_ forces calculates the total force on every atom within the
system.
Algorithm 1: Velocity-Verlet Integration Method
......
do i 1, N
r(i) r(i) dt * v(i) dt * dt/2 * F(i)
v(i) v(i) dt/2 * F(i)
! update positions at t dt
using velocities and forces at t
! update velocities at t dt using
forces at t
end do
call get_ forces (F)
do i 1, N
v(i) v(i) dt/2 * F(i)
! calculate forces at t dt
! update velocities at t dt
using forces at t dt
end do
......
185
28
2 s
2Q
29
30
31
ps Q
s_
s
32
186
HNose
Hoover
N
X
p2
p2
V rN gkB T ln s s
mi
2Q
i
i1
33
To construct MD simulations under constant P, an analogous reformulation of the Lagrangian was proposed by Andersen [9]. The constantpressure method of Andersen allows for isotropic changes in the volume
of the simulation box. Later, Hoover [8] combined this method with the
isothermal MD method described above to provide a set of equations for
MD simulations in the NPT ensemble. Parrinello and Rahman [10] extended
Andersens method to allow for changes not only in the size, but also in the
shape of the simulation box. This is particularly important in the simulation
of solids (e.g., crystalline polymers) since it allows for phase changes in the
simulation involving changes in the dimensions and angles of the unit cell.
34
and
dP
P Pext = P
dt
35
where Text and Pext are the desired temperature and pressure values and T
and P are time constants characterizing the frequency of the system coupling to temperature and pressure baths. T and P are the instantaneous
values of temperature and pressure, calculated from the momenta and configuration of the system [2]. The solution of these equations forces velocities
187
and positions to be scaled at every time step by factors T and xP, respectively, with
1=2
dt T
xT 1
1
T Text
dt
xP 1 T P Pext
P
36
37
FIG. 2
the
the
the
the
188
The NTLx yy zz ensemble can be viewed as a hybrid between the NVT
ensemble in the x direction and the isothermalisobaric (NPT) ensemble in
the y and z directions. The temperature T is kept fixed at a prescribed value
by employing the NoseHoover thermostat; the latter introduces an additional dynamical variable in the system, the parameter s, for which an
evolution equation is derived. Also kept constant during an MD simulation
in this ensemble is the box length Lx in the x direction; on the contrary, the
box lengths in the other two directions, Ly and Lz, although always kept
equal, are allowed to fluctuate. This is achieved by making use in the
simulation of an additional dynamical variable, the cross-sectional area
A( LyLz) of the simulation cell in the yz plane, which obeys an extra
equation of motion involving the instantaneous average normal stress
( yy zz)/2 in the two lateral directions y and z, respectively; ( yy zz)/2
remains constant on average and equal to Pext throughout the simulation.
The derivation of the equations of motion in the NTLx yy zz ensemble has
been carried out in detail by Yang et al. [12], and goes as follows: Consider a
system consisting of N atoms with rik being the position of atom i belonging
to polymer chain k. The bond lengths are kept fixed, with gik denoting the
constraint forces on atom i. The Lagrangian is written as a function of
~ k is the scaled (with respect
~ k , xik, A, s} where R
the extended variables {R
to the box edge lengths) position of the center of mass of every chain k, and
xik is the position of atom i in chain k measured relative to the chain center of
mass. This ensemble is extended in the sense that it invokes the additional
variables A and s, makes use of a scaled coordinate system, and is formulated
with respect to a virtual time t0 . The equations of motion are derived from
the extended Lagrangian by exactly the same procedure as for the other
statistical ensembles. The final equations are further recast in terms of real
coordinates and real time and have the following form:
s_
mi rxik Fxik gxik pxik
s
!
mi Ryk A_ 2
s_
A
mi ryik Fyik gyik pyik
2A
2A
s
!
mi Rzk A_ 2
s_
A
mi rzik Fzik gzik pzik
2A
2A
s
"
#
X X p2xik p2yik p2zik
s_2
g 1kB T
Qs Q s
s
mi
i
k
1
s_A_
2
38
39
40
41
42
189
where the forces with two indices indicate center of mass forces, while those
with three indices are forces on atoms within a particular polymer chain. Rk
denotes the center of mass of molecule k, while Q and W are inertial constants governing fluctuations in the temperature and the two normal stresses
yy and zz, respectively. The total Hamiltonian of the extended system,
derived from the Lagrangian, has the form:
!2
X p2
Q s_ 2
ln s W A_
i
V r
g 1
Pext Lx A
2 s
2 s
2mi
i
43
The first term on the right hand side represents the kinetic energy, the
second term is the potential energy, and the last four terms are the contributions due to the thermostat and the fluctuating box cross-sectional area
in the plane yz. Conservation of HNTL x yy zz is a good test for the
simulation.
For the solution of equations of motion, a modification of the velocityVerlet algorithm proposed by Palmer [14] can be followed.
iL
N
X
@
@
r_ i
Fi
@ri
@pi
i1
44
45
190
The next step is to decompose the evolution operator into two parts
such that
iL iL1 iL2
with
N
X
@
,
iL1
Fi
@pi
i1
N
X
@
_
iL2
ri
@ri
i1
46
expiL1 dt=2 expiL2 dt expiL1 dt=2P O t3 =P2
47
48
The evolution of the system at time t using the above factorization, Eq.
(48), is described through the following scheme [16]
rdt r0 dt v0
vdt v0
dt2
Fr0
2m
dt2
Fr0 Frdt
2m
49
50
191
N
N
N
X
X
X
@
@
@
r_ i
,
iL2
,
iL3
51
iL1
Ffi r
Fsi r
@pi
@ri
@pi
i1
i1
i1
The heart of the rRESPA algorithm is that the equations of motion are
integrated by using two different time steps, it is therefore a Multiple
Time Step (MTS) method: the slow modes (slow forces, iL3) are integrated
with a larger time step, t, whereas the fast modes (fast forces and
velocities, iL1, iL2) with a smaller time step, t(t t/n). In this case the
evolution operator becomes [16]
t
t
t n
t
exp iL3
expiLtexp iL3
exp iL1
expiL2 texp iL1
2
2
2
2
3
52
O t
The force calculated n times (fast force) is called the reference force. A
sample code of an MTS integrator is given in Algorithm 2.
Algorithm 2: rRESPA Integration
. . .. . ..
do i 1, N
v(i) v(i) t/2 * Fs(i)
end do
do j 1, n
do i 1, N
v(i) v(i) t/2 * Ff(( j1)t)
r(i) r(i) t * v(i)
end do
call fast_ forces (Ff )
do i 1, N
v(i) v(i) t/2 * F f( jt)
end do
end do
call slow_ forces (Fs)
do i 1, N
v(i) v(i) t/2 * Fs(i)
end do
. . .. . ..
192
53
with
N
X
@
f
iL1
F i r
,
@pi
i1
N
X
@
r_ i
iL2
,
@ri
i1
N
X
@
s
Fi r
iL3
@pi
i1
54
Also,
iLNH
N
X
i1
vi
@
@
@
G
@vi
@
@
55
where
G
N
X
p2
i
i1
mi
!
gkB T =Q,
56
t
t n
exp iL1
expiL2 t exp iL1
2
2
3
t
t
exp iL3
exp iLNH
O t
2
2
57
193
t
t
t
t n
exp iL1
expiL2 t exp iL1
exp iLNH
exp iLNH
2
2
2
2
3
t
t
t
exp iLNH
exp iL3
exp iLNH
O t
58
2
2
2
The resulting integrator is named XI-RESPA (eXtended system InsideREference System Propagator Algorithm).
Modifications of the RESPA method for MD simulations in the NPT
statistical ensemble have also been formulated in an analogous manner.
More details can be found in the original papers [15,16].
194
k k r
59
@k
F i gi
@ri
60
where the second term on the right hand side of Eq. (60) denotes the constraint forces. The question then is how to calculate the set of Lagrange
multipliers
k. Two methods that have widely been used in our MD simulations are discussed here: The EdbergEvansMorriss method and the
SHAKE method.
61
One then solves the following set of algebraic and differential equations
simultaneously:
mi r i Fi gi
X @k
gi
k
@ri
k
2
rij r ij r_ ij 0
62
63
64
195
dt2 X
@k t
k
k
@ri
mi
65
where rci are the constrained and rui the unconstrained positions. If the
constraints are satisfied at time t dt, then kc t dt 0. But if the
system moved along the unconstrained trajectory, the constraints would
not be satisfied at t dt. In this case, by performing a Taylor expansion
around the unconstrained positions, we get
kc t dt ku t dt
N
X
@k
i1
@ri
rui tdt
rci t dt rui t dt Odt4
66
dt
N
X
dt2 X
i1
mi
k0
k0
@k
@ri
@k0
@ri
67
68
By inverting the matrix, one can solve for the vector ,. However, since
the Taylor expansion in Eq. (66) has been truncated, the s should be
196
dt2
k @k t
mi @ri
69
N
X
1 @k t dt @k t
m
@ri
@ri
i
i1
70
from which
k dt2 PN
i1
ku t dt
1=mi @k t dt=@ri @k t=@ri
71
197
198
72
199
FIG. 3 Mean square displacement of the center of mass for the C156 (solid line),
C200 (dotted line), and C250 (dashed line) systems.
FIG. 4 Predicted and experimental [25] self-diffusion coefficients D vs. chain length
N in a loglog plot (T 450K, P 1 atm).
200
the figure are experimental data [25]. Three distinct regions appear in
the figure:
1. A region of a small-MW, alkane-like behavior (N<60), where D
follows a power-law dependence D Mb, with b>1. In this regime chain
end effects, which can be described through a free volume theory, dominate
system dynamics [21].
2. An intermediate, Rouse-like regime (from N 6070 up to N 156)
where b 1. System dynamics in this regime is found to obey the Rouse
model, at least for the overall relaxation of the chain [20].
3. A long chain-length, reptation-like regime (N>156), where chain
diffusivity exhibits a dramatic slow down, b 2.4. According to the original
formulation of reptation theory, the latter exponent should be 2.
Phenomena such as contour length fluctuations (CLF) and constraint
release (CR) typically accelerate the escape of the chain from the tube,
causing an increase in D and a decrease in 0 [26]. A recently proposed
theory that incorporates CLF and CR phenomena predicts a stronger exponent, between 2.2 and 2.3 [27]. These values agree with recent experimental results for concentrated polymer solutions and melts, which suggest
an exponent between 2.2 and 2.4 for a variety of polymer systems [25].
In contrast to D, the prediction of other viscoelastic properties, such as
the friction coefficient or the zero-shear rate viscosity 0, requires that the
atomistic MD data be mapped upon a mesoscopic theoretical model. For
unentangled polymer melts, such a model is the Rouse model, wherein a
chain is envisioned as a set of Brownian particles connected by harmonic
springs [25,28]. For entangled polymer melts, a better model that describes
more accurately their dynamics is the tube or reptation model [26].
According to this model, the motion of an individual chain is restricted by
the surrounding chains within a tube defined by the overall chain contour or
primitive path. During the lifetime of this tube, any lateral motion of the
chain is quenched.
The Rouse model is formulated in terms of three parameters: the number
of beads N, the length of the Kuhn segment b, and the monomeric friction
coefficient . The friction coefficient can be calculated directly from the
diffusion coefficient D through
kB T
73
ND
while the zero-shear rate viscosity 0 can be calculated from the density ,
the end-to-end distance hR2 i, and the diffusion coefficient D through
0
RThR2 i
36MD
74
201
kB T2
3NhR2 iD
75
and
0
76
77
To find the proper value of the projection parameter , one can follow a selfconsistent scheme based on the mean square displacement of the primitive
path points (s, s; t) [22]. (s, s; t) is defined as
s, s; t Rs, t Rs, 02
78
where R(s, t) is the position vector of the primitive segment at contour length
s at time t, and R(s, 0) is the position vector of the primitive segment at
contour length s at time 0. According to reptation theory
s, s; t 6Dt
1
X
4hR2 i
p1
p2 2
cos
ps
tp2
1 exp
L
d
79
202
where the sum is over all normal modes p and d denotes the longest relaxation or disengagement time. For small times (t< d), (s, s; t) is dominated
by the terms with large p and the above equation becomes
Z1
4La 1
tp2
dp 2 2 1 exp
s, s; t 6Dt
p 2
d
0
1=2
3
2
R D
6Dt 2
t1=2
80
Equation (80) offers a nice way of mapping atomistic MD trajectories
uniquely onto the reptation model, through a self-consistent calculation of
the parameter . First, a value of is chosen and the mapping from the
atomistic chain onto its primitive path is carried out by following the
procedure described by Kroger [30]. Then, Eq. (78) is used to calculate
(s, s; t) for the primitive path points, averaged over all s values. For times
t< d, the resulting curve is compared to that obtained from Eq. (80), using
the values of hR2 i and D (long-time diffusivity of the centers of mass)
calculated directly from the atomistic MD simulations. The procedure is
repeated until convergence is achieved, that is until a value is found for
which the two curves coincide. This mapping is performed self-consistently,
without any additional adjustable parameters or any experimental input. It
allows a reliable estimation of the tube diameter , by utilizing atomistically
collected MD data only for times shorter than d. Thus, the total duration of
the MD simulations required is governed solely by the time needed reliably
to calculate the center-of-mass diffusion coefficient D. With the values of
hR2 i, D, and , those of and 0 can be calculated using Eqs. (75) and (76).
With the above procedure the tube diameter was calculated to be
60 A for the longer-chain systems C200 and C250, whereas for the shorter
systems, N<200, no proper value of the parameter could be identified [22].
Figure 5 shows results for the monomeric friction factor as a function of
mean chain length N, over the entire range of molecular lengths studied.
Filled squares depict results obtained by mapping atomistic MD data onto
the Rouse model, whereas open circles depict results obtained from mapping the atomistic MD data onto the reptation model. According to its
definition, should be independent of chain length, its value determined
solely by the chemical constitution of the melt. The figure shows clearly that,
at around C156, a change in the mechanism of the dynamics takes place,
which cannot be accommodated by the Rouse model unless a chain-length
dependent is assumed. In contrast, in this regime (N>156), the reptation
model provides a consistent description of the system dynamics characterized by a constant (0.4 109 dyn s/cm) chain-length independent value
per methylene or methyl segment.
203
FIG. 5 Monomer friction coefficient vs. chain length N, obtained from mapping
the atomistic MD data onto the Rouse model (squares) or the reptation model
(circles).
204
FIG. 6 Zero-shear rate viscosity 0 vs. molecular weight M, obtained from the MD
simulation and the Rouse model for small M (circles) or the reptation model for high
M (squares). Also shown in the figure are experimentally obtained 0 values
(triangles).
RtRt
hR2 i0
81
In the above equation, R stands for the end-to-end vector of a macromolecule and hR2 i0 is the mean-squared magnitude of that vector in the equilibrium, quiescent state, where chains are unperturbed to an excellent
205
206
FIG. 7 Evolution of the diagonal components cxx, cyy and czz of the conformation
tensor c with time t for (a) the C24 and (b) the C78 PE melt systems. Results are
averaged over all NTLx yy zz trajectories (T 450K, Pext 1 atm).
207
FIG. 8 Evolution of the component xx of the stress tensor with time t for the C24
system. The results at every time t have been obtained either by applying the virial
theorem and averaging over all dynamical trajectories (broken line) or by using a
thermodynamic expression based on the free energy as a function of the
conformation tensor (thick solid line) (T 450K, Pext 1 atm).
The first (thin broken line) tracks the evolution of xx as obtained from
applying the molecular virial theorem on the relaxing configurations and
averaging over all dynamical trajectories. The second way (thick solid line)
uses the Helmholtz energy function and an affine deformation assumption
for chain ends [23,24]. According to the latter approach, the stress tensor at
every time t is calculated from the ensemble-averaged values of mass density
, conformation tensor cxx, and partial derivative of the Helmholtz energy
function with respect to cxx at time t, through
"
#
R
@A=Nch
xx t Pext 2 Ttcxx
M
@cxx T, , cjxxj
82
cxx cxx t
208
FIG. 9 Evolution of the components yy and zz of the stress tensor with time t for
the C24 system (T 450K, Pext 1 atm).
based on the current values of cxx and @A=Nch =@cxx , the latter taken from
the melt elasticity simulations presented in [23,24]. As expected, in the
estimates based on the virial theorem are subject to much higher statistical
uncertainty, owing to the fluctuations in the instantaneous configurations.
Clearly, averaging over many configurations is needed in order to improve
the statistical quality of the virial theorem results. Apart from highfrequency noise, the virial theorem results display an oscillatory character.
When the noise and oscillations are smoothed out, the ensemble-averaged
stress xx(t) from the virial theorem is in very good agreement with the
thermodynamic estimate obtained from cxx and @A=Nch =@cxx . This is an
important result, as it opens up the possibility of calculating stress with high
precision directly from ensemble average conformational properties, based
on a free energy function accumulated via efficient MC runs. The transverse
components yy and zz are displayed in Fig. 9. Both yy and zz fluctuate
continuously around the constant value Patm, as required by the macroscopic restrictions placed on the NTLx yy zz ensemble. The plots for the C78
system are similar. Relaxation times extracted from these stress relaxation
computer experiments are identical to those determined from equilibrium
MD simulations (Section VI.A).
VII.
209
PARALLEL MD SIMULATIONS OF
POLYMER SYSTEMS
A. Parallel MD Algorithms
Here we will focus on a discussion of algorithms for parallelizing MD
simulations with short-range interactions where the nonbonded forces
are truncated, so that each atom interacts only with other atoms within a
specified cutoff distance. More accurate MD models with long-range forces
are more expensive to deal with. Thus, special techniques are required for
parallelizing MD simulations with long-range interactions.
210
FIG. 10 Division of the force matrix among P processors in the atomdecomposition method.
developed based on this approach [34]. The key idea of this method is that
each processor is assigned a subset of atoms and updates their positions and
velocities during the entire simulation, regardless of where they move in the
physical domain.
Let us consider a polymeric system with a total number of atoms N,
where both intramolecular and intermolecular interactions are present. The
system is simulated in a parallel system of P processors. We first define x
and f as vectors of length N, which store the position and total force on
each atom, respectively. We also define the force matrix F, of dimensions
N N, with Fij denoting the force on atom i due to atom j. In the RD
method each processor is assigned a group of N/P atoms at the beginning of
the simulation. Each processor is also assigned a subblock of the force
matrix F which consists of N/P rows of the matrix, as shown in Fig. 10. If z
indexes the processors from 0 to P 1, then processor z computes forces in
the Fz subblock of rows. It also is assigned the corresponding position and
force subvectors of length N/P denoted as xz and fz.The computation of the
nonbonded force Fij requires only the two atom positions, xi and xj. But to
compute all forces in Fz, processor Pz will need the positions of many atoms
owned by other processors. In Fig. 10 this is represented by having the
horizontal vector x at the top of the figure span all the columns of F,
implying that each processor must store a copy of the entire x vector, hence
the name replicated-data. This also implies that, at each time step, each
processor must receive updated atom positions from all other processors
and send the positions of its own atoms to all other processors, an
operation called all-to-all communication.
211
2.
3.
4.
212
occurs naturally if the atom density is uniform across the simulation domain
(e.g., bulk simulations). In a different case (e.g., polymer chains at interfaces
or adsorbed polymers) load-balance can be achieved by randomizing the
order of the atoms at the start of the simulation or by adjusting the size of
the subset of each processor dynamically during the simulation to tune the
load-balance; these are called dynamical load-balancing techniques [36].
In summary, the RD algorithm divides the force computation evenly
across all processors. At the same time, its simplicity makes it easy to
implement in existing codes. However, the algorithm requires global
communication in steps (2) and (4), as each processor must acquire
information from all other processors. This communication scales as N,
independently of the number of processors P. Practically this limits the
number of processors that can be used effectively.
2. Force-Decomposition Method
The next parallel MD algorithm discussed here is based on a blockdecomposition of the force matrix rather than the row-wise decomposition
used in the RD algorithm. The partitioning of the force matrix F is shown in
Fig. 11 and the algorithm is called the force-decomposition (FD) algorithm
[37]. The method has its origin in block-decomposition of matrices, which is
commonly encountered in linear algebra algorithms for parallel machines.
The block-decomposition, shown in Fig. 11, is actually applied on a
permuted force matrix F0 , which is formed by rearranging the columns of the
original F in a particular way. The (ij) element of F is the force acting on
atom i in vector x due to atom j in the permuted vector x0 . Now the Fz0
subblock owned by each processor z is of size (N/P1/2) (N/P1/2). As shown
FIG. 11 The division of the force matrix among P processors in the forcedecomposition method.
213
in the figure, to compute the nonbonded forces in Fz0 , processor z must know
one N/P1/2-length piece of each of the x and x0 vectors, i.e., the subvectors xa
and x0b . As these elements are computed they will be accumulated into
the corresponding subblocks fa and fb0 . The subscripts a and b each run
from 0 to P1/2 and reference the row and the column position occupied by
processor z.
As in the RD algorithm, each processor has updated copies of the atom
positions xa and x0b needed at the beginning of the time step. A single time
step of the FD algorithm consists of the following steps:
1.
2.
3.
4.
The first step is the same as that of the RD algorithm. First the
nonbonded forces Fz0 are computed. The result is summed into
both fa and fb0 . Next, each processor computes a fraction N/P of
the bonded interactions. A critical point here is that in a preprocessing step of the run, we should guarantee that each processor knows
all the atom positions needed for the bonded (intramolecular) interactions. This step again scales as N/P.
Step (2) is also the same as that of the RD algorithm. The key
difference is that now the total force on atom i is the sum of the
elements in row i of the force matrix minus the sum of the elements in
column i0 , where i0 is the permuted position of column i. Thus this
step performs a fold of fa ( fb0 ) within each row (column) of processors
to sum-up these contributions. The important point is that now the
vector fa ( fb0 ) being folded is only of length (N/P1/2) and only the P1/2
elements in one row (column) are participating in the fold. Thus, this
operation scales as N/P1/2 instead of N as in the RD algorithm.
Finally, the two contributions are jointed to yield the total forces fz
( fz0 ) on the atoms owned by processor Pz.
The processor can now perform the integration for its own atoms,
as in the RD case. This operation scales as N/P.
Step (4) shares these updated positions with the processors that
need them for the next time step. As with the fold operation the
processors in each row (column) expand their xa (x0b ) subvectors
within the row (column) so that each acquires the entire xa (x0b ).
This operation scales again as N/P1/2 instead of N, as in the RD
algorithm.
214
3. Domain-Decomposition Method
The third parallel method discussed here for MD simulations of systems
with short-range interactions is domain-decomposition (DD) method
[31,3840]. In this method the physical simulation box is divided into
small 3D boxes, one for each processor. The partitioning of a simulation
box of length L in a DD algorithm is shown in Fig. 12 (2D projection). Each
processor z owns a sub-box labeled Bz with edge length Lz (Lz L/P) and
updates the positions of all atoms within its own box, xz, at each time step.
Atoms are reassigned to new processors as they move through the physical
domain. In order to compute the forces on its atoms, a processor must
know the positions of atoms in nearby boxes (processors), yz. Thus the
communication required is local in contrast to global in the AD and FD
algorithms. As it computes the force fz on its atoms, the processor also
computes the components of forces fzn on the nearby atoms.
A single time step of a DD algorithm consists of the following steps:
1.
2.
3.
For each processor Pz, the first step concerns the calculation of
bonded and nonbonded forces for atoms within box Bz. This step
scales with the numbers of atoms N/P per processor.
In step (2) the forces gz are shared with the processors owning neighboring boxes. The received forces are summed with the previously
computed fz to create the total force on the atoms owned by the
processor. The amount of data exchanged in this operation (and
consequently the scaling of this step) is a function of the force
cutoff distance and box length.
After computing fz, the atomic positions xz are updated. This
operation also scales as N/P.
FIG. 12
4.
5.
215
The scaling of steps (1) and (3) in the DD algorithm is again the optimal
N/P. The communication cost involved in steps (2), (4), and (5) is more
complicated. It proves to be dependent on the relative value of the cutoff
distance rc in comparison with the subdomain edge length Lz [31]. More
specifically, if Lz>rc the scaling goes as the surface-to-volume ratio
(N/P)2/3. If Lz rc, then the communication scales as N/P. In practice,
however, in the MD simulations of polymer systems there are several
obstacles to minimizing communication costs in the DD algorithm.
.
If Coulombic interactions are present then, because of the 1/r dependence, long cutoffs should be used. Thus Lz<rc and extra communication is needed in steps (2) and (4). Data from many neighboring
boxes must be exchanged and the communication operation scales as
r3c . Special techniques such as Ewald summation, particleparticle, or
particlemesh methods can be implemented in parallel [41].
As atoms move to new processors in step (5) the molecular connectivity information should be exchanged and updated between processors. This requires extra communication cost, depending on the type
of the bonded interactions.
If macromolecular systems are simulated uniformly in a simulation
domain, then all boxes have a roughly equal number of atoms (and
surrounding atoms); load-balance occurs. This will not be the case if
the physical domain is nonuniform (e.g., for polymers in vacuum or
with surrounding solvent). In this case it is not trivial to divide the
simulation domain so that each processor has an equal number of
atoms. Sophisticated load-balancing algorithms have been developed
[36] to partition an irregular or nonuniformly dense physical domain,
but the result is subdomains which are irregular in shape, or connected in an irregular fashion to their neighboring boxes. In both
cases the communication operation between processors becomes
more costly. If the physical atom density changes with time, then
the subject of load-balance becomes more problematic. Dynamical
load-balancing schemes are again needed, which require additional
computational time and data transfer.
216
B. EfficiencyExamples
Examples of applications of the above techniques can be found in various
articles that describe parallel MD applications for systems consisting of
many thousands (or even millions) of atoms performed on a large number of
processors [31,40]. Here we focus on the implementation of the parallel
algorithm for the united-atom polyethylene melt MD simulations described
in the previous section [44].
A critical point in any parallel implementation of an existing serial code
is the percentage of the code that can be parallelized. To measure the
performance of parallel implementations of existing sequential algorithms,
the speed-up parameter S is used. This is defined as the ratio of the execution
time of the serial (sequential) algorithm on a single processor to the
execution time of the parallel algorithm running on P processors
S P
s
p
83
where s and p denote the execution time of the algorithms on one and P
processors, respectively. For a fully parallel code running on P processors,
p s/P and S(P) P (linear speed-up).
A typical example where parallel implementation of atomistic MD
algorithms can substantially speed-up code execution is in NTLx yy zz
simulations described in Section VI.B (stress relaxation simulations). In this
case the simulated systems are independent and parallelism is straightforward to achieve by assigning each relaxing configuration to a different node
(processor). Since no data communication between different systems is
required in this case, excellent speed-up should be achieved.
This has been verified with trial runs on a Cray T3E 900 machine at the
Edinburgh Parallel Computing Center (EPCC) using the standard MPI
approach [45]. Figure 13 presents a graph of the speed-up of the (parallel)
code for a number of simulations, each one being executed with a model
system containing 40 chains of C78 PE melt. As was expected, the speed-up
is practically perfect (linear).
217
C. Parallel Tempering
A trend in modern MD (and MC) simulations is to enhance system
equilibration at low temperatures by the use of novel parallel techniques.
One such technique which has recently attracted considerable attention in
different variations is parallel tempering (PT) [47,48]. PT was introduced
in the context of spin glass simulations, but the real efficiency of the method
was demonstrated in a variety of cases, such as in the study of the
conformational properties of complex biological molecules [49,50]. Sugita
218
and Okamoto [51] used a similar method called replica exchange molecular
dynamics to simulate protein folding and overcome the multiple-minima
problem. They also used multiple-histogram reweighting techniques to
calculate thermodynamic quantities in a wide temperature range. Recently,
Yamamoto and Kob [52] used the same method to equilibrate a twocomponent Lennard-Jones mixture in its supercooled state. They found that
the replica-exchange MC method is 10100 times more efficient than the
usual canonical molecular dynamics simulation.
PT has also been used successfully to perform ergodic simulations with
Lennard-Jones clusters in the canonical and microcanonical ensembles
[53,54]. Using simulated tempering as well as PT, Irback and Sandelin
studied the phase behavior of single homopolymers in a simple hydrophobic/hydrophilic off-lattice model [55]. Yan and de Pablo [56] used
multidimensional PT in the context of an expanded grand canonical
ensemble to simulate polymer solutions and blends on a cubic lattice. They
indicated that the new algorithm, which results from the combination of a
biased, open ensemble and PT, performs more efficiently than previously
available techniques. In the context of atomistic simulations PT has been
employed in a recent study by Bedrov and Smith [57] who report parallel
219
n
Y
Qi
84
i1
where Qi denotes the individual partition function of the ith system in its
relevant statistical ensemble. The strategy of simultaneously equilibrating
not a single but a number of systems at different temperatures accelerates the
equilibration of the lowest temperature systems by accepting configurations
from the higher temperature systems, for which the rate of the system equilibration is higher. Thus, in the PT method, configurations are swapped
between systems being equilibrated at adjacent temperatures. The acceptance probability of a swapping move between system configurations
i and j i 1 within a parallel tempering series of NPT simulations is
given by:
acci, j ! j, i min1, expfi Uj Pj Vj j Ui Pi Vi
j Ui Pi Vi j Uj Pj Vj g
min1, expfU PV g
85
220
REFERENCES
1. Binder, K., Ed. Monte Carlo and Molecular Dynamics Simulations in Polymer
Science; Oxford University Press: New York, 1995.
2. Allen, M.P.; Tildesley, D.J. Computer Simulation of Liquids; Oxford University
Press: Oxford, 1987.
3. Frenkel, D.; Smith, B. Understanding Molecular Simulations: From Algorithms
to Applications; Academic Press: New York, 1996.
4. Ciccoti, G.; Hoover, W.G. In Molecular Dynamics Simulations of Statistical
Mechanics Systems, Proceedings of the 97th International Enrico Fermi
School of Physics, North Holland, Amsterdam, 1986.
5. Goldstein, H. Classical Mechanics; Addison-Wesley: Reading, MA, 1980.
6. Swope, W.C.; Andersen, H.C.; Berens, P.H.; Wilson, K.R. J. Chem. Phys.
1982, 76, 637649.
7. Nose, S. Mol. Phys. 1984, 52, 255268.
8. Hoover, W.G. Phys. Rev. A 1985, 31, 16951697.
9. Andersen, H.C. J. Chem. Phys. 1980, 72, 23842393.
10. Parrinello, M.; Rahman, A. Phys. Rev. Lett. 1980, 45, 11961199.
11. Berendsen, H.J.C.; Postma, J. P.M.; van Gunsteren, W.F.; DiNola, A.;
Haak, J.R. J. Chem. Phys. 1984, 81, 36843690.
12. Yang, L.; Srolovitz, D.J.; Yee, A.F. J. Chem. Phys. 1997, 107, 43964407.
13. Harmandaris, V.A.; Mavrantzas, V.G.; Theodorou, D.N. Macromolecules
2000, 33, 80628076.
14. Palmer, B.J. J. Comp. Phys. 1993, 104, 470472.
15. Tuckerman, M.; Berne, B.J.; Martyna, G.J. J. Chem. Phys. 1992, 97,
19902001.
16. Martyna, G.J.; Tuckerman, M.E.; Tobias, D.J.; Klein, M.L. Mol. Phys. 1996,
87, 11171157.
17. Edberg, R.; Evans, D.J.; Morriss, G.P. J. Chem. Phys. 1986, 84, 69336939.
18. Ryckaert, J.P.; Ciccoti, G.; Berendsen, H.J.C. J. Comp. Phys. 1977, 23,
327341.
19. Andersen, H.C. J. Comp. Phys. 1983, 52, 2434.
20. Harmandaris, V.A.; Mavrantzas, V.G.; Theodorou, D.N. Macromolecules
1998, 31, 79347943.
21. Harmandaris, V.A.; Doxastakis, M.; Mavrantzas, V.G.; Theodorou, D.N.
J. Chem. Phys. 2002, 116, 436446.
22. Harmandaris, V.A.; Mavrantzas, V.G.; Theodorou, D.N.; Kroger, M.;
Ram rez, J.; Ottinger, H.C.; Vlassopoulos, D. Macromolecules 2003,
36,13761387.
23. Mavrantzas, V.G.; Theodorou, D.N. Macromolecules 1998, 31, 63106332.
24. Mavrantzas, V.G.; Ottinger, H.C. Macromolecules 2002, 35, 960975.
221
25. Doi, M.; Edwards, S.F. The Theory of Polymer Dynamics; Clarendon: Oxford,
1986.
26. Milner, S.T.; McLeish, T.C.B. Phys. Rev. Lett. 1998, 81, 725728.
27. Lodge, T.P.; Phys. Rev. Lett. 1999, 83, 32183221.
28. Ferry, J.D. Viscoelastic Properties of Polymers; J. Wiley & Sons: New York,
1980.
29. Uhlherr, A.; Doxastakis, M.; Mavrantzas, V.G.; Theodorou, D.N.; Leak, S.J.;
Adam, N.E.; Nyberg, P.E. Atomic structure of a high polymer melt. Europhys.
Lett. 2002, 57, 506511.
30. Kroger, M.; Ram rez, J.; Ottinger, H.C. Polymer 2002, 43, 477487.
31. Plimpton, S. J. Comp. Phys. 1995, 117, 119.
32. Gupta, S. Comp. Phys. Comm. 1992, 70, 243270.
33. Smith, W. Comp. Phys. Comm. 1991, 62, 229248.
34. Smith, W.; Forester, T.R. Comp. Phys. Comm. 1994, 79, 5262.
35. Fox, G.C.; Johnson, M.A.; Lyzenga, G.A.; Otto, S.W.; Salmon, J.K.; Walker,
D.W. Solving Problems on Concurrent Processors; Prentice Hall: Englewood
Clis, NJ, 1988; Vol. 1.
36. Deng, Y.; Peiers, R.; Rivera, C. J. Comp. Phys. 2000, 161, 250263.
37. Hendrickson, B.; Plimpton, S. J. Parr. Distr. Comp. 1995, 27, 1525.
38. Clark, T.W.; Hanxleden, R.V.; McCammon, J.A.; Scott, L.R. In Proc.
Scalable High Performance Computing Conference94, IEEE Computer
Society Press: Los Alamitos, CA, 1994; 95102 pp.
39. Brown, D.; Clarke, J.H.R.; Okuda, M.; Yamazaki, T. Comp. Phys. Commun.
1994, 83, 113.
40. Putz, M.; Kolb, A. Comp. Phys. Commun. 1998, 113, 145167.
41. Plimpton, S.J.; Pollock, R.; Stevens, M. Particle-Mesh Ewald and rRESPA
for Parallel Molecular Dynamics Simulations, In Proceedings of the Eighth
SIAM Conference on Parallel Processing for Scientic Computing, SIAM,
Minneapolis, MN, 1997.
42. http://www.cs.sandia.gov/sjplimp/lammps.html
43. http://www.ks.uiuc.edu/Research/namd
44. Unpublished data collected at the Edinburgh Parallel Computing Centre
(EPCC) under the TRACS programme during the period 10 January
200020 February 2000.
45. http://www.mpi-forum.org/
46. http://www.openmp.org/
47. Newman, M.E.J.; Barkema, G.T. Monte Carlo in Statistical Physics;
Clarendon: Oxford, 1999.
48. Geyer, C.J. Computing Science and Statistics, Proceedings of the 23rd
Symposium on the Interface, Interface Foundation of North America: Fairfax
Station, VA, 1991, p. 156.
49. Hansmann, U.H.E. Chem. Phys. Lett. 1997, 281, 140150.
50. Wu, M.G.; Deem, M.W. Mol. Phys. 1999, 97, 559580.
51. Sugita, Y.; Okamoto, Y. Chem. Phys. Lett. 1999, 314, 141151.
52. Yamamoto, R.; Kob, W. Phys. Rev. E 2000, 61, 54735476.
222
53. Neirotti, J.P.; Calvo, F.; Freeman, D.L.; Doll, J.D. J. Chem. Phys. 2000, 112,
1034010349.
54. Calvo, F.; Neirotti, J.P.; Freeman, D.L.; Doll, J.D. J. Chem. Phys. 2000, 112,
1035010357.
55. Irback, R.; Sandelin, E. J. Chem. Phys. 1999, 110, 1225612262.
56. Yan, Q.; de Pablo, J.J. J. Chem. Phys. 2000, 113, 12761282.
57. Bedrov, D.; Smith, G.D. J. Chem. Phys. 2001, 115, 11211124.
58. Doxastakis, M.; Mavrantzas, V.G.; Theodorou, D.N. J. Chem. Phys. 2001,
115, 1135211361.
7
Configurational Bias Techniques for
Simulation of Complex Fluids
T. S. JAIN and J. J. DE PABLO
Madison, Wisconsin, U.S.A.
I.
University of WisconsinMadison,
INTRODUCTION
where U is the change in energy in going from one state of the system to
another, and is the inverse of kBT, with kB being Boltzmanns constant
and T the temperature.
224
225
order for the system to evolve to different states, it will have to be guided
out of local energy minima by biasing proposed trial moves. These biases
will in general be system specific; they can be designed in such a way
as to influence the position of a molecule, its orientation, or its internal
configuration. Whatever the bias is, however, it must be subsequently
removed in order to generate the desired, correct distribution of states
dictated by the ensemble of interest. The next section discusses how this can
be achieved.
where f(X) is the probability of being in state X and K(X|Y) is the probability of moving to Y given that the system is at X. The latter probability
can be further decomposed into the product of two factors: the probability
of proposing the transition, T(X|Y), and that of accepting the transition,
A(X|Y). With this separation, Eq. (2) becomes:
AXjYTXjY f X AYjXTYjX f Y
The way in which the moves are proposed is arbitrary; any function T(X|Y)
can be used as long as it is properly normalized, i.e.,
Z
TXjY dY 1
226
In order to satisfy detailed balance, the function A(X|Y) must have a specific
form. In the prescription of Metropolis et al. [1], moves are accepted with
probability
TYjX f Y
AXjY min 1,
TXjY f X
where min(x, y) is used to denote the smaller of its two arguments. It can
be shown that in the limit of a large number of trial moves, the sequence of
states generated by accepting trial moves according to Eq. (5) follows the
desired probability distribution function f.
In the particular case of a naive algorithm, i.e., the case in which
displacements are proposed at random, one can see that the transition
probabilities T(X|Y) and T(Y|X) are the same. For a canonical, NVT
ensemble, the function f takes the form
f X
1
expUX
Q
where Q is the partition function of the system. For the canonical ensemble
Eq. (5) therefore reduces to Eq. (1).
In a Configurational Bias algorithm, a bias is introduced in the forward
or reverse direction; as a result of the bias, the forward and reverse
transition probabilities are no longer equal to each other. This leads to
additional flexibility in the construction of a trial move, and can result in
considerable improvements of sampling efficiency. This bias is removed in
the acceptance criteria according to Eq. (5), thereby leading to sampling
according to the correct, original equilibrium probability distribution f.
A few illustrative examples on the use of Eq. (5) are discussed in the next
section. In each case, the correct acceptance criteria are arrived at by
establishing the form of the functions T(X|Y) and T(Y|X).
227
3.
4.
k
X
expU o l
l1
5.
Select one orientation j, from the set OF, according to the probability
distribution
p j
expU o j
RFW
k1
X
expU o l expU o k
10
l1
expU o k
RR
W
11
228
7.
AXjY min 1, W
new
old
RR
W
12
where subscripts old and new are used to denote quantities evaluated
in the original (old) and the trial (new) configurations, respectively.
The above algorithm can be applied to both lattice and continuum
simulations. One important difference, however, must be noted. In the
case of a lattice, all possible trial orientations of the molecule can be
considered explicitly; the sets OF and OR are therefore the same (on a
cubic lattice, a simple dimer or a bond can adopt an orientation along the
coordinate axes, and n 6). In a continuum, a molecule can adopt one of
an infinite number of trial orientations. By restricting the trial orientations
to the finite sets OF and OR, the probability of generating these sets also
enters the transition matrix along with the probability of choosing
a direction from the probability distribution (which itself is a function
of these sets [4]). The acceptance criteria are therefore also a function of
these two sets:
TYjXPOR f Y
AXjY, OF , OR min 1,
TXjYPOF f X
13
The a priori probabilities, P(OF) and P(OR), of generating these sets of trial
orientations are the same and drop out of the final expression. In order to
satisfy detailed balance, the condition of super-detailed balance is
imposed, i.e., detailed balance should be obeyed for every pair of sets
chosen. Hence the acceptance criteria of Eq. (12) generate the true equilibrium distribution.
229
230
FIG. 2 Swelling curve for two clay sheets of sodium montmorillonite from NPzzT
simulations.
clay were taken from Skipper et al. [6]. The interaction parameters for the
claywater system were based on the TIP4P and MCY models of water
[79]. Simulations suggest that the swelling of clays occurs through the
formation of discrete layers of water. The resulting basal spacings as a
function of number of water molecules (Fig. 2) agree with experiment. The
use of a grand canonical formalism, in which water is inserted into the clay
galleries via a configurational bias algorithm, permits calculation of the
water content of the clay (Fig. 3b) as a function of thermodynamic
conditions. The disjoining pressure can also be calculated as a function of
basal distance to provide reliable estimates of the mechanically stable basal
spacings (Fig. 3a), which is important in many applications of clays. In
addition to providing information on the swelling of the clay, these simulations reveal that, in general, the compensating cations (e.g., sodium or
calcium) are fully hydrated (Fig. 4). Arriving at such conclusions on the
basis of conventional Monte Carlo simulations would require considerable
computational resources; using an orientational bias increases efficiency
considerably and lends credence to the results.
231
severe constraints on the types of trial moves that can be used to generate
distinct configurations of the system. At liquid-like densities, it is clear that a
random displacement of the center of mass of a molecule (while maintaining
the internal configuration) would lead to overlaps with other molecules, and
a rejection of the trial move.
This problem can be partially alleviated by conducting highly localized
trial moves. On a simple cubic lattice, for example, the conformation of a
linear polymeric molecule can be explored by resorting to local trial
moves such as kink-jumps, end-rotations, or crankshafts[10]. These
moves typically displace one or two sites of the chain. The so-called
reptation [11] moves involve displacing an entire chain in a slithering,
snake-like motion along its contour. In a continuum, simple random moves
of individual sites can be used to sample the configuration space of the
232
FIG. 5
4.
233
eUi
pi j Pk
Uil
l1 e
5.
14
After all the sites have been grown calculate the Rosenbluth weight
of the chain as follows:
RFW
L
Y
k
X
eUi
15
iLn1 l1
L
Y
pi j
16
iLn1
6.
7.
Similarly, to construct the reverse move, a new set of k 1 orientations is generated about the old positions, and the Rosenbluth factor
of the original configuration of the molecule, RR
W , is calculated.
Substitution into Eq. (5) gives acceptance criteria of the form:
RF
AXjY min 1, W
RR
W
17
234
pli CeUi
l 1, 2, . . . , k
18
235
combination of an expanded grand canonical technique with configurational bias can lead to highly effective insertion schemes, even for long chain
molecules at elevated densities [18].
Formally, an expanded grand canonical ensemble can be defined through
the partition function:
1 X
M
X
QN, y, V, T exp
y exp N
19
N0 y1
1
exp
N expUN, y exp
20
QN, V, T, y
exp
N exp
p y
21
The goal of the algorithm is to sample fluctuations of the number of molecules effectively. If the molecules of interest consist of L sites, this can be
done by allowing a tagged chain to fluctuate in length; at the two extremes
of the spectrum these fluctuations can cause the tagged chain to comprise L
interaction sites and become a real, full chain, or they can cause the tagged
chain to have zero sites and disappear. In order for these sequential insertion
or deletion processes to be effective, the probability with which terminal
states of the tagged chain are visited should be appreciable. This can
be controlled by judicious assignment of the weighting factors appearing
in Eq. (19). The ratio of the probabilities for two states x and y of the system
is given by
px
exp
p y
QN, V, T, x
QN, V, T, y
22
The logarithm of the ratio of the partition functions, under identical values
of N, V, and T, is equal to the difference in the chemical potential for a
236
ex x ! y ex x, N ex y, N
23
3
Ny
y, N y
N y
ln
V
ex
ex
24
3.
4.
25
237
238
FIG. 6 Simulated adsorption isotherms for hard-sphere chains of 4, 16, and 100
sites in a slit-like pore. Reference refers to [24].
239
configurational bias moves are similar in spirit, but different in implementation, to concerted rotation (CONROT) moves [28,29], which are not
elaborated upon here.
In a topological configurational bias move, an arbitrary number of
inner sites of a chain are excised and regrown by taking into account
energetic biases, as well as geometric (topology) considerations that ensure
that connectivity is preserved. Detailed balance is satisfied by determining
the number of fixed-end random walks that exist between two sites of the
molecule. While different schemes may vary in the way in which the number
of random walks, N( j), is estimated, the general idea remains the same.
We begin by discussing the implementation on a lattice, and then describe
several more recent methods suitable for a continuum.
1. Lattice Case
An algorithm to perform such a move on a cubic lattice [30] could be:
1.
2.
3.
N l i, n 1eUi
4.
26
N j i, n 1eUi
pi j Pk
Uil
l
l1 N i, n 1e
5.
27
The modified Rosenbluth weight and Random Walk weight for the
chain are constructed as follows:
RFW
n X
k
Y
N l i, n 1eUi
i1 l1
n
Y
GFRW
N j i, n 1
i1
28
29
240
6.
To estimate the weight of the original configuration, the same procedure is repeated for the original positions of the sites that were
excised.
The modified acceptance criteria for this move then take the form
RF GR
RW
30
AXjY min 1, W
F
RR
W GRW
7.
N i,s n
NX
y X
N
x 0
Ns !
x !x x!y !y y!N x y !N x y !
y0
31
Ns x y z
2
Ns x y z
N
2
N
32
33
where x x x , y y y, z z z are the difference in the coordinates between site i and site n 1, Ns is the number of steps, n i 1, along
the chain between sites i and n 1, and x, y, z are the number of steps along
the positive coordinate axes and x , y , z are the steps along the negative
coordinate axes.
In order to assess the performance of this algorithm, we have studied the
decay of the bond autocorrelation function as a function of CPU time. The
bond autocorrelation function is defined by
bt
hB~t B~0i
hB~0 B~0i
34
where B~t denotes the vector along a bond of the molecule at time t.
Figure 8 shows the bond autocorrelation function corresponding to a simulation for cyclic chains consisting of 100 sites, in which only local moves
(kink-jump and crankshaft) were employed, and one in which topological
configurational bias moves were implemented. One can clearly see that the
decay in the latter case is much faster, thereby resulting in a much more
efficient simulation algorithm. In the case of ring molecules, care must
be exercised in the selection of the number of sites to be cut and regrown.
This is because the topological nature of the molecule can change (e.g., from
unknotted to knotted) if a large number of sites are selected. In the case
of linear chains this restriction is irrelevant and, in principle, any number
241
of sites can be used in an individual trial move. It turns out that an optimum
number of sites can be identified because the computational expense of the
trial move increases with the number of cut sites. This point is illustrated by
Fig. 9, which depicts the decay of the bond autocorrelation function for
linear chains of 500 sites on a lattice. For a fixed amount of CPU time, the
242
FIG. 8 Rate of decay of the bond autocorrelation function (b(t)) for cyclic (ring)
chains of 100 sites, using local moves and topological configurational bias moves, as
a function of CPU time.
2. Continuum Case
In the continuum case, various strategies can be used to achieve closure. In
the case of Extended continuum configurational bias (ECCB) Monte Carlo
[33], a simple geometric constraint is employed to ensure that the chain is
correctly closed. Since the geometric constraint does not take intramolecular
243
FIG. 9 Decay of the bond autocorrelation function (b(t)) for linear chains of
length 500 as a function of number of sites cut and regrown, for a given amount of
CPU time.
interactions into account, the ECCB method is efficient for molecules with
relatively flexible backbones. This method is described in some detail below.
1.
2.
3.
244
FIG. 10 Limiting cases for the bounds of the angle for the two regrowing sites
(1 and 2) involved in an extended configurational bias move (ECCB). (a) Neighbor
sites a and b in the configuration of maximum separation, (b) and (c) neighbor sites a
and b in the configuration of closest approach.
4.
5.
i1 and b, where b is the first uncut site on the side of the cut
opposite to the growing end (see Fig. 10).
Once the bounds for have been established, several orientations of
the bond vector bi, i1 are sampled according to the usual configurational bias method, and one is selected from the probability distribution given by Eq. (14).
The same procedure is repeated for the reverse move. There is a
temporary change in degrees of freedom used to generate the new
configuration because of the restricted sampling of (Fig. 11). In
order to correct for this, the Jacobian determinants must be incorporated into the acceptance criteria. The acceptance criteria for this
move become
RF J F
35
AXjY min 1, RW R
RW J
where JF and JR are the Jacobian determinants for the forward and
reverse moves, respectively. For a detailed explanation on the calculation of the bound of and the Jacobians readers are referred to [33].
245
FIG. 11 The entire spherical sampling space is not accessible for selection of the
trial orientation of site 1, because the angle must lie within the bounds [min, max]
(see Fig. 10). Note that there are no bounds on the spherical angle for the bond
vector.
246
FIG. 12 Illustration of the HOLE method. (a) Local configuration before the
move; (b) configuration after the move.
for a significant number of the proposed positions. The use of intramolecular potentials along with the probability distribution for distances
between sites has been successful in biasing the random walk and achieving
closure [37]. A preliminary run is necessary in order to generate a
probability distribution for site separations as a function of site spacing
along the chain. A brief outline of the implementation as proposed by
Escobedo and Chen is provided below:
1.
2.
3.
pli CeUi
l 1, 2, . . . , k
36
247
248
4.
pi j
37
38
l1
5.
wn1
j
eUn1 j eU n Jn1
wn1
kn1
X
ext
bend l
l
j
pn1
bend
39
40
l1
ii.
iii.
n
Y
wi
i1
41
249
FIG. 14 Illustration and notation of the reconstruction process of the last two sites
(n 1 and n) of an ECCB move for a chain that exhibits stiff intramolecular energy
functions [37].
iv.
b.
42
RCB2
i.
wn1, t
eU
bend
j
j
n,
tJ
n1, t
wn1, t
kn1
X
eU
bend
t 1, 2, . . . , m
l
l
n,
tJ
n1, t
43
44
l1
ii.
250
iii.
ext
j
pn1,
n
wn, n1
45
46
t1
iv.
After all the sites have been grown, the Rosenbluth weight
for the entire chain is constructed:
RFW wn, n1
n2
Y
wi
47
i1
251
FIG. 16 Half-chain end-to-end correlation function (h(t)) for RCB1 for an isolated
C30 molecule at 250K for 4 and 8 site moves. Three different forms of Poi rl, n1 are
used: the Continuous Unperturbed Chain (CUC) Poi rl, n1 (dash-dot line), the
uniform Poi rl, n1 implied by conventional ECCB moves (dashed line), and Poi rl, n1
for a fully flexible Guassian chain (solid line).
252
Nr
Y
Qi N, V, Ti
48
i1
49
where i, j is the difference in the inverse temperatures of the two replicas
and Ui, j is the difference in the potential energies. It can be seen from
Eq. (49) that the trial move will have a reasonable chance of being accepted
only if the energy distributions for the two replicas overlap to some extent.
FIG. 18
253
50
254
one can have Ni full molecules and one tagged molecule of length y (recall
Section IV.B.1). When two replicas exchange configuration, they also swap
a tagged chain. Within each of the replicas, however, the tagged chain is
allowed to grow or shrink according to a conventional expanded ensemble
formalism.
This technique has been applied to the study of phase behavior of long
polymeric molecules. Figure 20 compares the performance of various
simulation techniques to that of Hyperparallel Tempering Monte Carlo. It
can be seen that the decay of the end-to-end autocorrelation function for
HPTMC is considerably faster than that for conventional canonical or
grand canonical simulations. However, it is important to point out that
much of the performance of that method is due to the efficiency of the
underlying moves, namely the expanded ensemble configurational bias
gradual growth or deletion of sections of the molecules.
V. FUTURE DIRECTIONS
Configurational bias trial moves offer significant improvements in efficiency
over conventional, blind trial moves. As discussed throughout this chapter,
local random moves usually provide small displacements of the system,
whereas biased moves can efficiently displace several sites per move while
255
REFERENCES
1. Metropolis, N.; Rosenbluth, A.W.; Rosenbluth, M.N.; Teller, A.H.; Teller, E.
Equation of state calculations by fast computing machines. J. Chem. Phys. 1953,
21, 10871092.
256
2. Kalos, M.H.; Whitlock, P.A. Monte Carlo Methods; Wiley: New York, 1986.
3. Cracknell, R.F.; Nicholson, D.; Parsonage, N.G.; Evans, H. Rotational insertion bias: a novel method for simulating dense phases of structured particles,
with particular application to water. Mol. Phys. 1990, 71 (5), 931943.
4. Frenkel, D.; Smit, B. Understanding Molecular Simulation; Academic:
New York, 1996.
5. Chavez-Paez, M.; Van Workum, K.; de Pablo, L.; de Pablo, J.J. Monte Carlo
simulations of wyoming sodium montmorillonite hydrates. J. Chem. Phys.
2001, 114 (3), 14051413.
6. Skipper, N.T.; Chang, F.R.C.; Sposito, G. Monte-Carlo simulation of interlayer molecular-structure in swelling clay-minerals. 1. methodology. Clay Clay
Miner. 1995, 43 (3), 285293.
7. Boek, E.S.; Convey, P.V.; Skipper, N.T. Molecular modeling of clay hydration:
a study of hysteresis loops in the swelling curves of sodium montmorillonites.
Langmuir 1995, 11 (7), 46294631.
8. Jorgensen, W.L.; Chandrasekhar, J.; Madura, J.D. Comparison of simple
potential functions for simulating liquid water. J. Chem. Phys. 1983, 79 (2),
926935.
9. Lie, G.C.; Clementi, E.; Yoshimine, M. Study of the structure of molecular
complexes. XIII. Monte Carlo simulation of liquid water with a conguration
interaction pair potential. J. Chem. Phys. 1976, 64 (6), 23142323.
10. Lax, M.; Brender, K. Monte Carlo studies of lattice polymer dynamics.
J. Chem. Phys. 1977, 67 (4), 17851787.
11. Mandel, F. Macromolecular dimensions obtained by an ecient Monte Carlo
method: the mean square end-to-end separation. J. Chem. Phys. 1979, 70 (8),
39843988.
12. de Pablo, J.J.; Yan, Q.L.; Escobedo, F.A. Simulation of phase transitions in
uids. Annu. Rev. Phys. Chem. 1999, 50, 377411.
13. Rosenbluth, M.N.; Rosenbluth, A.W. Monte Carlo calculation of the average
extension of molecular chains. J. Chem. Phys. 1955, 23 (2), 356359.
14. de Pablo, J.J.; Laso, M.; Suter, U.W. Simulation of polyethylene above and
below the melting-point. J. Chem. Phys. 1992, 96 (3), 23952403.
15. Siepmann, J.I.; Frenkel, D. Congurational bias Monte-CarloA new sampling scheme for exible chains. Mol. Phys. 1992, 75 (1), 5970.
16. Smit, B.; Karaborni, S.; Siepmann, J.I. Computer simulations of vapour-liquid
phase equilibria of n-alkanes. J. Chem. Phys. 1995, 102 (5), 21262140.
17. Lyubartsev, A.P.; Martinovski, A.A.; Shevkunov, S.V.; VorontsovVelyaminov, P.N. New approach to Monte Carlo calculation of the free energy:
Method of expanded ensembles. J. Chem. Phys. 1991, 96 (3), 17761783.
18. Depablo, J.J.; Escobedo, F.A. Expanded grand canonical and Gibbs
ensemble Monte Carlo simulation of polymers. J. Chem. Phys. 1996, 105 (10),
43914394.
19. Kumar, S.K.; Szleifer, I.; Panagiotopoulos, A.Z. Determination of the
chemical-potentials of polymeric systems from Monte-Carlo simulations.
Phys. Rev. Lett. 1991, 66 (23), 23952398.
257
258
8
Molecular Simulations
of Charged Polymers
ANDREY V. DOBRYNIN
Connecticut, U.S.A.
I.
INTRODUCTION
Considerable theoretical and experimental work during the past half century
has been devoted to charged polymers [19]macromolecules with ionizable groups. Under appropriate conditions, such as in aqueous solutions,
these groups dissociate, leaving ions on chains and counterions in solutions.
If the charges on the polymers are all positive or all negative, these polymers
are called polyelectrolytes. Common polyelectrolytes are polyacrylic and
methacrylic acids and their salts, cellulose derivatives, sulfonated polystyrene, DNA and other polyacids and polybases. If after dissociation of the
charged groups the polymers carry both positive and negative charges, they
are called polyampholytes. Examples of polyampholytes include proteins,
for example gelatin, and synthetic copolymers made of monomers with
acidic and basic groups. If these groups are weak acids or bases, the net
charge of polyampholytes can be changed by varying the pH of aqueous
solutions and at high charge asymmetry these polymers demonstrate
polyelectrolyte-like behavior.
Despite these extensive efforts we are still far away from a complete
understanding of the behavior of polymeric systems with electrostatic
interactions. The main factor hindering our progress is the long-range
nature of the Coulombic forces between charged species. At this stage
computer simulations have proven to be a valuable tool for the elucidation
of structural and physical properties of charged systems as well as for
verification of old and creation of new theoretical models. The explosive
260
Dobrynin
growth of computer power over the last few years has led to development of
large scale simulation techniques whose goal is to reproduce and simulate
processes on the molecular level. It is possible now to simulate thousands of
atoms over nanosecond time scales. But such detailed molecular simulations
are still impractical for the simulation of real polymeric systems, for which
typical time scales to reach thermal equilibrium vary from microseconds to
seconds.
The coarse-grained models of polymers (see [10] for a review) allow the
extension of molecular simulations beyond nanosecond time scales by
leaving aside the atomistic details of the solvent and polymers and
concentrating on their macroscopic properties. In this approach the monomers are not chemical monomers, but rather groups of chemical units, and
solvent molecules are represented by a continuum with macroscopic physical properties. Figure 1 shows an example of such mapping by presenting
the snapshots of sodium polystyrene sulfonate in water and its representation as a coarse-grained chain with counterions in a continuum with
macroscopic physical properties. Of course, the challenging part is to relate
the parameters of the coarse-grained models with the real atomistic ones by
calculating the parameters of intermolecular and intramolecular coarsegrained potentials [10].
261
U bond fri g
bond
XX
i
i<j
!
qi qj
exp ri rj U sh ri rj
kB TlB
ri rj
where U bond ({ri}) is the bond potential describing the effect of the connectivity of monomers into the polymer chain, kB is the Boltzmann constant, T
is the absolute temperature, and lB is the Bjerrum length (lB e2/"kBT,y the
length scale at which the electrostatic interaction between two elementary
charges e in the medium with dielectric permittivity " is of the order of the
thermal energy kBT). The electrostatic interactions between free ions and
charged monomers are not explicitly included in the chain potential energy;
instead, their effect is treated through the dependence of the inverse
DebyeHuckel screening length on the electrolyte concentration
2 4lB
q2s cs
262
Dobrynin
U
accold ! new min 1, exp
kB T
263
FIG. 2 Variation of the total energy of the system Etot versus the number of MC
steps for different types of moves. (From Ref. [69].)
264
Dobrynin
1
265
is proportional to the number of monomers N with weak logarithmic correction. In Eq. (7) u is the ratio of the Bjerrum length lB to the bond length b.
The parameter uf 2 is sometimes called the coupling parameter. In fact,
polyelectrolytes are classified according to the strength of this parameter.
Polyelectrolyte chains with small values of the coupling parameter uf 2
1
are referred to as weakly charged polyelectrolytes, while ones with uf 2 01
are considered to be strongly charged. The larger the value of the coupling
parameter uf 2, the more elongated the polyelectrolyte chain is.
To better understand the different length scales involved in the problem,
it is useful to introduce the concept of the electrostatic blob [2631]. The
conformation of a chain inside the electrostatic blob is almost unperturbed
by electrostatic interactions, with the number of monomers in it being
ge (De/b)2 in a -solvent. The size of the electrostatic blob De containing ge
monomers can be found by comparison of the electrostatic energy of a blob
e2 g2e f 2 ="De with the thermal energy kBT. This leads to the electrostatic
blob size
1=3
De b uf 2
and the number of monomers in it
2=3
ge uf 2
266
Dobrynin
1=3
N
De bN uf 2
ge
10
However, for a polyelectrolyte chain to be elongated, the number of monomers in a chain N should be larger than the number of monomers in an
electrostatic blob ge N > uf 2 2=3 . Thus, short chains with number of
monomers N smaller than NG (uf 2)2/3 will still be Gaussian.
Of course, this simple scaling picture does not account for the
nonuniform stretching of a polyelectrolyte chain. In reality the chain is
more strongly stretched in the middle than at the ends. Logarithmic
corrections to the chain size may be obtained by allowing the blob size to
vary along the chain.
The generalization of this scaling approach to the case of the good
solvent is straightforward by replacing the relation between the blob size De
and the number of monomers in it ge from random to self-avoiding random
walk statistics [2631].
The conformational properties of an isolated polyelectrolyte chain in a
salt-free solution have been investigated by lattice Monte Carlo simulations
[3235], and by off-lattice Monte Carlo simulation of a chain made of hard
spheres connected by rigid bonds with fixed valance angles [36], of a freelyjointed chain [37], and of a bead-spring chain [3842]. These simulations
support the trivial scaling prediction that the chain crosses over from a coil
to a rod-like conformation with increasing the strength of the electrostatic
interactions (increasing value of the coupling parameter uf 2). This was done
either by adding ionized groups to a chain or by increasing the Bjerrum
length lB.
Quantitative analysis of the scaling model of a polyelectrolyte chain as a
chain of electrostatic blobs was done by Higgs and Orland [37] and by
Barrat and Boyer [38]. Figure 3 shows the mean square end-to-end distance
as a function of the degree of polymerization N for different values of the
coupling parameter uf 2 (in references [37,38] the parameter f was kept
constant and equal to unity). For each curve there is a crossover between
two regimes. For small N the electrostatic interactions between charges have
little or no effect at all on the conformation of the polyelectrolyte chain
and the chain remains Gaussian hR2e i N. For large N the electrostatic
interactions dominate and the chain adopts an elongated conformation with
size hR2e i N 2 .
267
FIG. 3
Mean square end-to-end distance hR2e i as a function of number of
monomers N for a polyelectrolyte chain in a -solvent. Curves are for three different
values of the parameter u 0.02, 0.05, and 0.2. The dasheddotted lines have
gradients 1 and 2. (From Ref. [37].)
268
Dobrynin
chain size grows faster than linearly, L0 Nln N1=3 , with the degree
of polymerization N. Figure 5 shows the p
dependence
of the normalized
269
p
FIG. 5
Root-mean-square end-to-end distance
hR2e i divided by N vs.
1=3
2 1=3
uf ln N=e , for chains with N 20, 80, 320, 1000, and 2000 (filled circles,
open squares, filled triangles, open circles, and filled squares, respectively). (From
Ref. [39].)
11
270
Dobrynin
number of sections 1/De by their length 1 one finds the following expression for the electrostatic persistence length [27,4547]
Le 2 =De
12
Indeed for salt concentrations such that 1>De the chain is stiffened at
the length scales larger than the Debye screening length Le>1. The size
of a chain with persistence length Le and a contour length of a chain of
electrostatic blobs L0 is
L2 Le L0 2 N=ge
13
occurs at 1 De N=ge bN 1=2 . For even higher salt concentrations the
number of the persistence segments increases, making it more probable for
interactions between segments separated by larger distances along the chain
contour to occur. The second virial coefficient for these interactions can
be estimated as that between two rods of length Le and thickness
1 L2e 1 . The excluded volume effect becomes important, when the
interaction parameter z of a polyelectrolyte chain
L0
z
Le
1=2
2 De bN 1=2
L3e
14
becomes
larger than unity. In this range of salt concentrations
p
(1 < De bN 1=2 ) the chain size scales as L / 3=5 N 3=5 . This regime continues until the Debye radius is larger than the electrostatic blob size De.
However, the predictions of this model are still challenged in the
literature. Computer simulations of weakly charged polyelectrolyte chains
[38,5861] and some experiments [6,4854] as well as analytical calculations
[5557] indicate that the exponent for the dependence of the electrostatic
persistence length Le on the Debye screening length is closer to 1 rather than
to 2 or even shows sublinear [62] dependence.
The most complete analysis to date of the electrostatic persistence length
dependence on the Debye screening length was performed by the Kremer
group [62]. Micka and Kremer have performed hybrid MC and MD
simulations of a polyelectrolyte chain with DebyeHuckel interaction
bonded by harmonic springs. For different bond lengths (b 2, 4, 8, 10, 16)
and chain lengths (N 16, 32, 64, 128, 256, 512) they investigated the
dependence of the chain dimensions and electrostatic persistence length on
the Debye screening length in the interval 0.001 1 0.48. The data for
271
the mean square end-to-end distance can be collapsed into universal curves
reasonably well by either assuming linear or quadratic dependence of the
electrostatic persistence length Le on the Debye radius 1. Both of these
plots show crossover from a rod-like chain hR2e i N 2 to a chain with
excluded volume interactions hR2e i N 6=5 as the value of the Debye radius
decreases. In order to differentiate between the models, the persistence
length of the chain was calculated from the bond angle correlation function
G(n). This function is defined by the scalar product of two normalized bond
vectors b(k) and b(k n)
Gn bk bk n
15
where the brackets hi denote the ensemble average over all chain conformations. The averaging procedure was improved by moving reference point k
along the chain. The persistence length was estimated from the exponential
decay of the function G(n). This method lead to sublinear dependence of the
electrostatic persistence length on the Debye radius (see Fig. 6). Similar
sublinear dependence was obtained by analyzing the chain structure
factor S(q).
The effect of internal chain stiffness on the dependence of the
electrostatic persistence length on the Debye radius was studied in [63,64].
FIG. 6 Persistence length Le dependence on the Debye screening length 1. The
two lines in the upper left corner indicate the slopes predicted by OSF (dashed) and
the variational method (solid). (From Ref. [62].)
272
Dobrynin
It was shown that with increasing internal chain stiffness the effective
exponent y for Le y crosses over from a value of one toward two as the
internal stiffness of a chain increases. The quadratic dependence of the
electrostatic persistence length on the Debye radius for the discrete
KratkyPorod model of the polyelectrolyte chain was recently obtained in
[65]. It seems that the concept of electrostatic persistence length works better
for intrinsically stiff chains rather than for flexible ones. Further computer
simulations are required to exactly pinpoint the reason for its failure for
weakly charged flexible polyelectrolytes.
273
FIG. 7 Reduced mean square radius of gyration hR2g i=N 2=3 as a function of reduced
valence fN1/2 for chains with degrees of polymerization N 16, 32, 64, 128, and 200.
274
Dobrynin
D. Conformational Properties of a
Polyampholyte Chain
Polyampholytes are charged polymers with both positively and negatively
charged monomers. In this case the Coulombic interactions between charges
are not only repulsive, as in the case of polyelectrolytes, but are also
275
3.
4.
N =N N .
p
Polyampholytes with charge imbalance, Q larger than N N , form
a necklace globule at low temperatures. There is a striking similarity
between
the instability of a polyampholyte globule with excess charge
p
Q N N and the necklace instability of a polyelectrolyte chain in a
poor solvent [68]. But for polyampholytes the factors responsible for the
shape of a globule are all electrostatic in nature. The main difference
between polyampholytes and uniformly charged polyelectrolytes is the
randomness in the charge sequence. For polyampholytes the structure of the
necklace is predetermined by the initial charge distribution. Monte Carlo
276
Dobrynin
studies by Kantor and Kardar [82,83] showed that the necklace may consist
of a few almost neutral globules connected by charged necks, or even of one
big neutral globule with a tail sticking out of it. The necklace instability was
also observed in MC simulation of circular polyampholytes by Lee and
Obukhov [88].
The effect of charge sequences on the collapse of polyampholytes was
studied by lattice MC simulations [89,90]. Polyampholytes with alternating
distribution of charges behave like polymers with short-range interactions
and can be characterized by an effective second virial coefficient [91]. The
collapse transition of these polyampholytes is similar to the coilglobule
transition of a neutral polymer [89]. A qualitatively different picture of the
collapse transition was discovered for diblock polyampholytes. The collapse
of these polymers happens in two stages. First, at high temperatures, a
zipping transition occurs that corresponds to the formation of dipolar pairs
between oppositely charged monomers. Second, at lower temperatures this
preassembled zipped structure undergoes an ordinary coilglobule transition. The possibility of a freezing transition in random polyampholyte
globules was investigated in computer simulations [92].
The influence of a uniform external electric field E on the conformation
and dynamics of polyampholyte chains has been studied by the MD
simulations of Soddemann et al. [87]. These simulations have shown that
the polyampholyte globule becomes unstable above the critical external
electric field Ec1 and breaks up forming a random necklace structure. A
polyampholyte collapses back into a globule as the strength of the external
electric field E is lowered below Ec2. The strength of the electric field Ec2 at
which a polyampholyte collapses back into a globule is weaker than that
required to break up a globule Ec1 (Ec2<Ec1). There is a hysteresis in
chain size dependence on the external electric field. This hysteresis is
a manifestation of the coexistence of two states of a chaincollapsed and
elongatedseparated by a barrier.
277
FIG. 9
278
Dobrynin
charge eqk located at point rk interacts with all other charges eqj at positions
rj (k 6 j ) within the cell as well as with all of their periodic images located at
rj n1a1 n2a2 n3a3 where n1, n2, n3 are all integer numbers between M
and M. The electrostatic energy of the unit cell interacting with an infinite
number of periodic images can be written as follows
X
Uel frj g
lB X
q q
k j
lim
rkj n
M!1 2
kB T
n k, j
16
where we introduced rkj rk rj. The first sum in Eq. (16) is over all vectors
n n1a1 n2a2 n3a3, and the asterisk indicates that the terms with k j
for n 0 are excluded from the summation.
The sum (16) is only slowly and conditionally convergentthe result
depends on the order in which the terms are added. The original method for
summation of the lattice sum was introduced by Ewald [98], who replaced
the sum in Eq. (16) by the sum of two absolutely convergent seriesa direct
sum in Cartesian space and a reciprocal sum in Fourier space. There is a
simple physical interpretation of this decomposition of the lattice sum
(see Fig. 10). In the direct sum each charge in the system qi is viewed as
being surrounded by a Gaussian charge distribution of the opposite sign
[r qi 3 exp2 r2 =3=2 , where the parameter determines the width
of the distribution], such that the net charge of this cloud exactly cancels qi.
The electrostatic interaction between screened charges is a rapidly decaying
function of distance, and is effectively reduced to short-range interactions in
the direct space. To counteract this diffuse cloud around each charge, a
second Gaussian charge distribution of the same sign as the original charge
is added for each point charge. This second distribution varies smoothly in
space and acts as a source term in a Poisson equation for the reciprocal
potential with periodic boundary conditions.
However, these most commonly used Ewald transformation formulas are
correct for centrosymmetric crystal structures but give results [98101] that
279
disagree with the explicit evaluation of sum (16) in the case where the unit
cell has a net dipole moment [104107]. The discrepancy is due to uncritical
use of the Fourier transformation converting the conditionally convergent
lattice sum in direct Cartesian space into a sum in reciprocal Fourier space.
The evaluation of the lattice sum (16) is cumbersome and details of
this calculation are given in the Appendix. The final expression for the
electrostatic energy of the cell interacting with all its periodic images (16)
can be represented as a sum of four terms: sums in the direct Udir ({rj}) and
reciprocal Urec ({rj}) space, self-energy term, and shape term that depends
on the net dipole moment of the cell
X
Uel frj g
erfc rkj n
lB X
qk qj
rkj n
2 n k, j
kB T
lB X exp 2 m2 =L2 2
SmSm
m2
2L m60
2
lB X 2 2lB X
p
qj
q
r
17
j
j
3L3 j
j
p R 1
error function
where erfcx 2= x dt expt2 is the complementary
P
which tends to zero as x ! 1 and S(m) qj exp i2m rj =L is the
j
charge structure factor.
The Coulombic forces on charge k are obtained by differentiating the
electrostatic energy of the system Uel({rj}) with respect to rk. The resulting
force
shape
rec
Fk Fdir
k Fk Fk
18
kB TlB qk
X
j
qj
X
n
!
2 2 erfc rkj n
2
p exp rkj n
rkj n
rkj n
rkj n2
19
2lB qk X im exp 2 m2 =L2 2
2im rk
kB T
exp
S m
m2
L2 m60
L
20
280
Dobrynin
4lB qk X
qj r j
3L3 j
21
N
optimal computational efficiency of this method one finds
that
M
B
p.
p
If the boxing is implemented with the optimum value Np the computer
time for lattice sum evaluation increases only as Np3=2 . This is a considerable
improvement over the N2 rate for traditional methods with fixed value of .
Before closing this section, let us comment on the parallel implementation of the Perram et al. algorithm [108]. In practice the computation of the
281
direct and reciprocal space contributions to the Ewald sum are done
separately. Since the direct sum represents short-ranged interactions
between particles, the parallelization of the summation in the direct sum
is handled by three-dimensional spacial domain decomposition based
upon blocks of chaining cells, considering the interactions only between
neighboring cells. The computation of the sum in the reciprocal (Fourier)
space is ideal for parallel implementation. The required charge structure
factor S(m) is rewritten as follows
X
XX
X
qj exp i2m rj =L
qj exp i2m rj =L
SP m
Sm
j
j2P
22
where P denotes processors. Particles are distributed to the processors and
the partial charge structure factor SP (m) is computed on each processor for
all m vectors. These partial charge structure factors SP (m) are then summed
across the processors to obtain the total charge structure factor S(m).
Computation of the reciprocal forces, where summation of factors
2imS(m) is involved, does not require further computer communications.
The factors 2imSP(m) can be summed on each processor for the whole
range of m vectors.
24
282
Dobrynin
are replaced by the linear combination of their values at the nearby grid
points. For example, let [u] denote the integer part of u. Then, using linear
interpolation, we can approximate the individual exponents in the r.h.s. of
Eq. (24) by their linear combination at the nearest grid points [u] and
[u] 1
2m u
2m
exp i
1 u u exp i
u
K
K
2m
u u exp i
u 1
K
25
Similar interpolation can be done for each term in the Eq. (24), with
1, 2, 3. Let us define W2(x), the linear hat function given by
W2(x) 1 |x| for |x|1, and W2(x) 0 for |x|>1. Then Eq. (25) can be
rewritten as
1
X
2m u
2m
k
W2 u k exp i
exp i
K
K
k1
26
The sum in the last equation is finite, because the function W2(x) is nonzero
only within the interval |x| 1. The accuracy of the approximation can be
improved by including more grid points in the interpolation scheme.
In the original implementation of the PME algorithm [109,111],
Lagrangian weight functions W2p(x) [116] were used for interpolation of
the complex exponents, by using values at 2p points in the interval |x| p.
Unfortunately, these functions W2p(x) are only piecewise differentiable,
so the approximate reciprocal lattice sum cannot be differentiated to arrive
at the reciprocal part of the Coulomb forces. Thus, the forces were
interpolated as well.
Instead of the Lagrangian weight functions W2p(x), the Cardinal
B-splines Mn(x) [117,118] were utilized in later versions of the PME
method [110,112,113]. These weight functions are continuously differentiable and allow the forces to be obtained from analytical differentiation of the
approximation of the reciprocal lattice sum. The Cardinal B-spline of
the second order M2(x) gives the linear hat function M2(x) 1 |x 1| in
the interval 0 x 2 and M2(x) 0 for x<0 and x>2. The nth order
B-spline satisfies the following properties:
1.
2.
3.
4.
5.
283
27
28
b m
Qk exp i m k
K
k2V
1
Sm
NP
X
qj
3
Y
b m
1
X
bm
b1 m1 b2 m2 b3 m3 Q
29
Qk
NP
X
j1
qj
3 X
Y
Mn uj, k s K
30
1 s
where the inner sum is over all integers s. Using this spline representation
of the charge structure factor S(m), the approximate reciprocal energy can
now be rewritten as
Urec fuj g
lB X exp 2 m2 =L2 2
bmQ
bm
BmQ
kB T
m2
2L m60
31
284
Dobrynin
where we introduce
Bm
3
Y
b m 2
32
1
33
where C is defined as
8
exp 2 m2 =L2 2
<1
, for m 6 0
34
Cm Bm
m2
:
0,
for m 0
0 0 0
with vector m, defined as m m1 , m2 , m3 , where m0 m for 0 m K/2
and m0 m K=2 otherwise. It is important to point out that
rec 1 , where
rec )1 is the inverse discrete Fourier transform defined
C d
P (d
3
b
as Am K
k2V Ak expi2=Km k. Using the property of the
bm K 3 Am, after some algebra, the reciproFourier transformation A
cal lattice sum reduces to
Urec fuj g
lB X
Qmrec
Qm
35
kB T
2L m2V
where rec
Q denotes the convolution of rec and Q defined as
rec
Qm
K
1 K
1 K
1
X
X
X
rec m1 k1 , m2 k2 , m3 k3 Qk1 , k2 , k3
k1 0 k2 0 k3 0
36
To obtain the reciprocal part of the Coulomb forces, Eq. (35) has to be
differentiated with respect to rj.
Frec
j
@Urec fuj g
kB TlB X @Qm
rec
Qm
@rj
L m2V @rj
37
285
3.
4.
5.
6.
286
Dobrynin
counter-ion cloud and is short ranged, while the reciprocal sum is due to the
Gaussian charge cloud and is a smooth function, with its Fourier transform
rapidly convergent. Below we outline the general steps of particlemesh
calculations of the long-range component.
In the P3M method the Gaussian charge clouds surrounding each charge
within the unit cells are substituted by finite range weight functions Wn(x)
[123,124] that interpolate the original charge density over the n grid points,
with spacing between them hL/K.
Qr
X qj 3
j
3=2
X
2
) Qg k
exp 2 r rj
qj W n uj k
38
where k is the radius vector of the grid point and vector k(k1, k2, k3) with
all 0 k K 1 being integers. As before, we rescale all lengths in the
system by the grid spacing h, such that uj rj/h. Due to periodic boundary
conditions, the rescaled coordinates u vary in the interval 0 u<K for
1, 2, 3.
The simplest one-dimensional weight function W1(x) [W1(x) 1, for
|x| 1/2 and W1(x) 0, for |x|>1/2] assigns the charge densities to the
nearest grid point. The full 3D weight function is a product of weight
functions in the x, y, and z directions. HigherR order weight functions are
1
convolutions of W1 x,Wn x W1
Wn1 x 1 W1 x x1 Wn1 x1 dx1 ,
and span n grid points in each direction. One can also use the Cardinal
B-splines, introduced for interpolation of the exponentials in the PME
method, as the weight functions for charge assignment. The higher order
Cardinal B-splines can also be generated by convolutions of the hat
function M1(x) [M1(x) 1, for 0 x 1 and M1(x) 0 otherwise]. In fact,
the charge assignment functions Wn(x) and Mn(x) are identical up to
translation.
Since the original charge density Q(u) has been replaced by charge
density over the grid points Qg(k), Hockney and Eastwood [119] suggested
to minimize the effect of such substitution by replacing the Coulomb
Greens function by function Gn(q), which will minimize the mean square
error in forces due to the new assignment function Wn and finite size grid
errors.
Consider two charges q1 and q2 placed at random positions u1 and
u2 in the unit cell. The mean square error in force due to charge assignment is
Z
du1
K3
2
du2 EW
F u2 , u1 FPM u2 , u1
3
K
39
287
where FEW (u2, u1) is the exact reciprocal force and FPM (u2, u1) is the
bn q that minimizes
approximated force. The optimal Greens function G
the functional (39) is [119,122124]
bn q lB
G
L
2
bqW
bn b q
q
A
b
2
2
P
b
W
q2
b
q
b n
40
Qg k expik q
41
288
2.
Dobrynin
where vector q is the vector with components (q1, q2, q3) with each
q 2m/K and 0 m K 1.
bn q and charge distribution
Using the modified Greens function G
b
g q over the grid points
Qg q solve for the electrostatic potential b
bn qQ
bg q
b
g q G
42
Using the inverse Fourier transform find the electric field at the grid
points Eg(k)
Eg k
4.
43
1 Xb
Eg q expik q
K3 q
44
45
289
calculating the electrostatic potential over the mesh points by applying the
inverse FFT to the electrostatic potential g(q). The electric field at the mesh
points is then obtained by numerical differentiation of the electrostatic
potential at the neighboring grid points. However, this implementation of
the P3M is less accurate than ones that use direct differentiation in the
Fourier space [125,126]. Extensive comparisons and analysis of both
methods can be found in [112,113,125,126].
1 X
n
X
Mnm m
Y ,
rn1 n
n0 mn
46
k
X
qj nj Ynm j , j
j1
FIG. 12
47
290
Dobrynin
where Ynm , is the spherical harmonic polynomial [131]. If we approximate this expansion by first p 1 terms, the truncation error is bounded
by (r/R)p [127,128]. Thus, fixing a precision , the number of terms in the
multipole expansion is estimated to be p log , where r/R.
There is a duality inherent in the situation depicted in Fig. 12; if the
location of test charge Q and a system of charges qj ( j 1, . . . , k) were
interchanged, we may write another expansion that represents the potential
due to particles outside the sphere of radius r and is correct as long as j>r,
r kB TlB
1 X
n
X
m
rn Lm
n Yn ,
48
n0 mn
where Lm
n are coefficients of the local expansion defined as
Lm
n
k
X
qj n1
Ynm j , j
j
49
j1
FIG. 13
Hierarchical subdivision of the full simulation space (2D case) into
children, grandchildren, etc.
291
neighboring cells might be within the cutoff distance 2Rn for this level.
Nearest neighbor cells are defined as cells sharing an edge or a corner
(in 3D also a face) with a given cell. The interactions between charges in cell
S with its nearest cells are taken by summing the electrostatic interactions
directly. We apply this approximation at each step to at least next nearest
neighbors and skip squares which lie in the regions that have been treated
at previous levels. Therefore, the boxes with which the particles in box S
interact at the present level, are (1) those that are not nearest neighbors of
S and (2) those whose parents were a nearest neighbor of the parent of S
at level n 1. Figure 14 shows the squares that are in the interaction list
of cell S.
The realization of the Fast Multipole Method involves six steps.
1.
2.
3.
4.
FIG. 14 Interaction list of cell S at level n. The squares at level n are separated by
thin lines, their parents at level n 1 by heavy lines. The cells in the interaction list
are labeled by I. The nearest neighbor cells are labeled by N.
292
5.
6.
Dobrynin
expansion describes the field due to all particles in the system that are
not contained in the current cubic box, its nearest neighbors and
second nearest neighbors.
Evaluate the potential and fields for each particle using the local
expansion coefficients on the finest level.
Add the contributions from other charges in the same cell and
near cells that are not included in the multipole expansion by
direct summation.
293
FIG. 15
Timing for complete force calculation of various size systems using
standard Ewald, fast multipole method, and P3M. (From Ref. [124].)
homogeneously in the solution with only a small fraction being near the
polyelectrolyte chains. The conformation of a chain is then determined by
the chain elasticity and by the electrostatic repulsion between charged
monomers (see Section II). As a result, chains adopt elongated conformations with size L bNuf 2 1=3 [see Eq. (10)]. With increasing polymer
concentration c, more and more counterions are drawn toward polyelectrolyte chains and, for sufficiently strong electrostatic interactions, can even
be localized on them [3,136]. This phenomenon is called counterion or
Manning condensation [3,136]. Localization of counterions leads to effective
renormalization of the total charge on a chain Qeff feffN<fN. A simple
estimate of the effective charge Qeff can be achieved by equating the
chemical potential of a counterion in a solution far away from a chain
kBT ln (cb3feff) with that of a counterion localized within the chain
1=3
2 1=3
volume LD2e ,
kB T lnf feff ufeff
kB Tu2=3 feff
[3,137140]. This
leads to the effective fraction of the charged monomers on the chain
!
2=3 1=3
cb3
exp u f
feff f 1
50
uf 2 1=3
being a decreasing function of polymer concentration c. The parameter
u2/3f 1/3 is the Manning parameter which determines the onset of counterion
2 1=3
condensation. Thus, the chains shrink, L bNufeff
, as polymer concentration increases. However, for small values of the Manning parameter
u2=3 f 1=3
1 the corrections to the chain size are small and can be
neglected.
294
Dobrynin
Chains overlap when the distance between them Rcm (N/c)1/3 becomes
of the order of their size L N. This happens at polymer concentration
c* N2. In semidilute solution, c>c*, polyelectrolytes form a temporary
network with mesh size . At length scales smaller than , the intrachain
electrostatic repulsion dominates and sections of the chain with g monomer
are strongly elongated such that g. Pure geometric arguments can be
used to obtain the concentration dependence of the correlation length , by
imposing the close packing condition for chain sections of size ,
c g/3 2. This leads to the correlation length c1/2 [26,27,29].
At length scales larger than the electrostatic interactions are screened by
the other chains and counterions and the statistics of a chain is Gaussian
with effective bond lengths of the order of the correlation length . Thus,
the chain
size
R in the semidilute salt-free polyelectrolyte solution is
p
2.
12 6 1
<
51
ULJ r 4"LJ r r 4 , for r rcut
:
0,
for r > rcut
The polymers were modeled as bead-spring chains. The attractive part of the
bond potential is described by the FENE potential
1 2
r2
UFENE r kr0 ln 1 2
52
2
r0
295
with the maximum extent r0 2 and the spring constant k 7"LJ/ 2. The
repulsive part of the bond potential was described by the shifted LennardJones potential. The strength of the electrostatic interactions in the simulations was controlled by the value of the Bjerrum length. In the simulations
of Stevens and Kremer [146148] most of the results were obtained for the
Bjerrum length lB equal to 0.833, which corresponds to the case of weakly
charged polyelectrolytes with the value of the coupling parameter u<1,
while in the Liao et al. [149] simulations the Bjerrum length lB was 3,
corresponding to strongly charged polyelectrolytes.
Figure 16 shows the dependence of the osmotic pressure of salt-free
polyelectrolyte solutions. Through almost the entire concentration range
considered, the osmotic pressure is proportional to the polymer concentration, supporting that it is controlled by the osmotic pressure of counterions,
both above and below the overlap concentration. There appears to be a
weak chain length dependence of the osmotic pressure for short chains.
However, this N-dependence is consistent with a 1/N correction to the
osmotic pressure due to chain translational entropy. The deviation from
linear dependence of the osmotic pressure occurs around polymer
concentration c 0.07 3, which is above the overlap concentration for all
samples. At very high polymer concentrations, where electrostatic interactions are almost completely screened by counterions and by charges on the
FIG. 16 The osmotic pressure is plotted as a function of the monomer density for
various chain lengths.
296
Dobrynin
FIG. 17 Plot of the correlation length as a function of density for different chain
lengths. The slopes are 0.5 and 0.33 for the solid and dashed lines, respectively.
chains, the osmotic pressure is proportional to c9/4, and the scaling law for
neutral polymers in a good solvent is recovered.
The correlation length of the solution (see Fig. 17) exhibits two regimes
as observed in experiment [1,6] and predicted by the theory [26,29,143145].
Above overlap concentrations the correlation length is chain length
independent and is inversely proportional to the square root of polymer
concentrations ( c1=2 ). At low polymer concentrations scales with
polymer concentration as c1/3. The crossover between these two regimes
occurs around the overlap concentration c*.
The effect of added salt on the conformation of polyelectrolyte chains in
dilute and semidilute solutions was investigated by Stevens and Plimpton
[151]. At high salt concentrations the electrostatic interactions between
charged monomers are screened and the chain conformations are similar
to those observed for good solvent, R N 0:588 . As the salt concentration
decreases, chains become more elongated, and finally at very low salt
concentrations the chain size saturates at its salt-free value R N [146,148].
297
298
Dobrynin
Limbach and Holm [157]. They have shown that the counterion distribution
around a quenched strongly charged polyelectrolyte chain shows an end
effect. The counterions are accumulated preferentially in the middle part of
the chain. Toward the end of the chain, the effective charge on the chain
increases, which is equivalent to a decrease in counterion concentration. The
reason for this behavior is the difference in the electrostatic potential created
by the charges of the polyelectrolyte chain. The electrostatic potential is
stronger in the middle of the chain than at its ends. Thus, an ion close to the
middle part of the chain is attracted more strongly than an ion sitting at the
chain end [158]. Besides the inhomogeneity of the local effective charge,
there is also inhomogeneity in the local chain conformations. The bonds are
stretched more in the middle of the chain. This nonuniform stretching of the
chain causes the polyelectrolyte to appear in the trumpet-like shape [158].
As one would expect, the end effects become less pronounced as salt is
added to the solution, due to the screening of the electrostatic interactions
over the Debye screening length.
For polyelectrolytes in a poor solvent, the situation is more peculiar. In
this case the end effects are smeared out over the end beads. There is a sharp
decrease in the effective charge in the region of the first strings. The
modulation in accumulation of the effective charge in the middle of the
chain repeats the necklace structure.
Condensation of counterions was studied by molecular dynamics
simulations [148,159]. These simulations have shown that in dilute
polyelectrolyte solutions at constant polymer concentrations the chain size
depends nonmonotonically on the Bjerrum length lB. First, the chain size
increases with increasing Bjerrum length lB, which is due to increasing
strength of the intrachain electrostatic repulsion between charged monomers. This will continue until the Bjerrum length lB reaches the crossover
value lB
. Above this value the counterion condensation kicks in, reducing
the effective charge on the chain, weakening intrachain electrostatic
repulsions, and chains begin to shrink [139,140,160].
299
approximation is only correct when the presence of polymer chains will only
slightly perturb the uniform distribution of the small ions in the system. This
approximation should also fail at high polymer and salt concentrations,
when excluded volume effects start to control the ionic atmosphere around
polyelectrolyte chains.
To investigate the effect of the DebyeHuckel approximation on the
solution properties, Stevens and Kremer [152] performed molecular
dynamics simulations of salt-free solutions of bead-spring polyelectrolyte
chains in which the presence of counterions was treated via a screened
Coulomb potential, and compared the results with their simulations with
explicit counterions [146,148]. To elucidate the effect of the DebyeHuckel
approximation, the dependence of the mean square end-to-end distance,
hR2e i, osmotic pressure, and chain structure factor on polymer concentration
was examined. Stevens and Kremer found that hR2e i tends to be larger at low
densities for DH simulations and is smaller at higher densities. However, the
difference in hR2e i between DH simulations and simulations with explicit
counterions is within 10%. This trend seems to be a generic feature for all N
in their simulations. The functional form and density dependence of the
chain structure factor are very close in both simulations. The most severe
DebyeHuckel approximation affects the dependence of the osmotic
pressure on polymer concentration. It appears that in the DH simulations
not only is the magnitude of the osmotic pressure incorrect, but also the
concentration dependence is wrong.
The effect of the Bjerrum length lB (strength of the Coulombic
interactions) on the applicability of the DebyeHuckel approximation was
investigated by Stevens and Plimpton [151]. They found that this
approximation works well for weak Coulomb interactions or for the
Bjerrum lengths lB smaller than (see Fig. 19). From this figure we see that
the DebyeHuckel approximation breaks down for lB>. In this range of
parameters the chain size monotonically increases in the DH simulations,
while for simulations with full Coulomb interactions the chain shrinks as the
strength of these interactions increases. This discrepancy should not be too
surprising, because, for lB>, the Coulomb attraction of counterions to a
polyelectrolyte chain in its vicinity becomes stronger than the thermal
energy kBT, leading to effective localization (condensation) of counterions
near the polymer chain [3,136,139,140,160]. This condensation reduces the
charge on the chain forcing it to contract [139,140,160].
The DebyeHuckel approximation can be improved by considering the
full PoissonBoltzmann approach to electrostatic interactions, in which
the counterion condensation phenomena are included implicitly. Tests of
the PoissonBoltzmann approach to a cell model of rigid polyelectrolytes
[161,162] were done in references [163165] by performing molecular
300
Dobrynin
301
302
Dobrynin
V. WHAT IS NEXT?
Our understanding of polyelectrolytes in solutions has significantly
progressed during the last ten years. The properties of polyelectrolyte
chains in dilute solution seem to be understood reasonably well. The only
problem that remains to be resolved here is the electrostatic persistence
length of flexible polyelectrolytes. Computer simulations of semidilute
polyelectrolyte solutions provide reasonable qualitative agreement with
existing experimental and theoretical works. We now better understand
counterion condensation phenomena and their relation to chain attraction
and clustering.
However, a lot of interesting work remains to be done as attention turns
to the dynamics of polyelectrolyte solutions. All current simulations of
polyelectrolytes neglect the hydrodynamic interactions [177]. The inclusion
of these interactions into molecular simulations will allow us to perform
simulations of realistic dilute and semidilute polyelectrolyte solutions. It
also will help us to answer the question of how hydrodynamic interactions
are really screened in polyelectrolyte solutions. The answer to this question
is extremely important for checking the assumptions of existing theoretical
models. Hydrodynamic interactions are also long range, thus the use of
Ewald summation [178180], PME, and P3M methods can significantly
speed up the simulations. Work in this direction has just begun.
Another area which has attracted significant interest over the last few
years is polyelectrolytes at surfaces and interfaces [8,181184]. The progress
in this area has implications for different areas of science and technology
such as colloids, biophysics, biochemistry, medical science, pharmacy, food
processing, water purification, etc. Some simulation work [185191] has
already appeared, suggesting a broad spectrum of interesting physical
phenomena.
Our list of future directions cannot be complete without mentioning
polyelectrolyte gels, polyelectrolyte brushes [192195] (polyelectrolytes
grafted to a surface), and complexation of polyelectrolytes with colloids
[196200], dendrimers [201], surfactants, and proteins. We anticipate serious
303
simulation studies of these systems in the near future that will have
tremendous practical applications.
APPENDIX
To perform direct evaluation of the lattice sum in Eq. (16) we first introduce
the integral representation of
1
2
p
jrj
dt exp jrj2 t2
dt exp jrj2 t2
53
With the help of this integral representation we can split the electrostatic
energy of the unit cell U ({rj}) into two parts
U rj Udir rj U2 rj
where we defined the direct sum
Z 1
X
2
Udir rj
lB X
lim p
qk qj
exp rkj n t2 dt
M!1
kB T
n
k, j
X
X
erfc rk rj n
lB
qk qj
rk rj n
2
n
54
k, j
and a sum
Z
X
2
U 2 rj
lB X
p
qk qj
dt exp rkj n t2
kB T
k, j
0
n
55
p R 1
In Eq. (54) erfcx 2= x dt exp t2 is the complementary error function. The direct sum Udir({rj}) is a rapidly convergent series and we may take
the limit M ! 1 without further ado. The second lattice sum in (55)
requires more careful attention. This sum is diffcult to evaluate as it is, so
we use the identity
2
exp rkj n t2
1
3=2 t3
2
u
d u exp 2 2irkj u 2in u
t
3
56
304
Dobrynin
X
d3u
exp u2 =2 2irkj u
exp2in u
2
u
n
In the last equation the second term in the r.h.s. is accounting for the selfenergy of the charges with n 0 and k j. The main contribution to the
sum comes from the regions where exp(2inu) 1 for which u G
m1b1 m2b2 m3b3 is a vector of the reciprocal lattice, b, 1, 2, 3 are
primitive vectors of the reciprocal lattice, and m1, m2, m3 are integer numbers. The vectors b have the property ba . [For the cubic lattice
these vectors are b1 (/L, 0, 0), b2 (0, /L, 0), and b3 (0, 0, /L).] This
suggests dividing the integration region over u into a sum of integrals over
the volume of a reciprocal lattice unit cell (the first Brillouin zone)
VG {/2L < </2L, for 1, 2, 3}. To do so we make the substitution u G v where v 2 VG . This leads to
!
X
Urec rj
lB X
2
2
lim
qk qj
exp G = 2irkj G IG rkj I0 rkj
M!1 2
kB T
G60
k, j
58
where we have defined
1
IG rkj 2
X
d 3 vgG v, rkj
exp2in v
VG
59
and
gG v, rij v G2 exp v2 =2 2v G 2irij v
The lattice sum in the integral IG(rkj) may be easily evaluated
X
n
exp2in v
3
M
X
Y
1 n M
exp2in L
3
Y
sin2M 1L
1
sinL
60
305
61
The first integral is equal to 3/L3 and the first term is equal to /L3G2. The
second integral may be estimated by integrating by parts with the use of
sin ((2M 1)L) ((2M 1)L)1 d cos ((2M 1)L)/d. The leading
terms are equal to zero because cos ((2M 1)/2) 0. Integration by parts
again gives terms that are O(((2M 1)L)2) and can be neglected in the limit
M ! 1. Collecting all terms together we obtain the final expression for the
sum in the reciprocal space with G 6 0
Urec rj
lB X exp G2 =2
S GS G
3
kB T
2L G60
G2
lB X exp 2 m2 =L2 2
SmSm
m2
2L m60
62
X
j
X
m rj
qj exp i2Grj
qj exp i2
L
j
63
where vector m (m1, m2, m3) with all integers m 2 1, 1 for 1, 2, 3.
Now let us evaluate I0(rkj). After some algebra it can be written as follows
I0 rkj
3
Z
exp 2irkj v 1 Y
1
sin2M 1L
3
I0 0 2
d v
2
v
VG
sinL
1
2 2
Z
exp v = 1
1
2
d 3v
VG
v2
3
Y
sin2M 1L
exp 2irij v 1
sinL
1
64
The function I0(0) is independent of rkj and its contribution to the lattice
sum cancels for electroneutral systems. The first integral can be estimated by
306
Dobrynin
expanding exp (2irij v) in powers of 2irkj v. All the odd powers in this series
do not contribute to the integral due to symmetry of the VG. Thus the
expansion starts with the term 2(rkj v)2. This is the only term in which
we are interested, because all other terms in the expansion are
O(((2M 1)L)2) and can be omitted in the limit M ! 1. The symmetry
of the integral allows us to change (rkj v)2 v2 by r2kj =3 and to perform the
integration over v obtaining 3/L3. The second integral is finite and can be
evaluated by integrating by parts leading to an asymptotic expansion in
inverse powers of (2M 1)L. This expansion starts with the term of the
order of O(((2M 1)L)2) and once again can be neglected in the limit
M ! 1. The final result for this part of the lattice sum is
2
Ushape rj
2lB X
2lB X
2
3
qk qj rkj
q
r
65
j j
kB T
3L k, j
3L3 j
.
Combining all terms together, we can write the final expression for the
electrostatic energy of the cell as a sum of four terms sums in the direct and
reciprocal space, self-energy term, and shape term that depends on the net
dipole moment of the cell
X
Uel rj
erfc rkj n
lB X
qk qj
rkj n
kB T
2 n k, j
lB X exp 2 m2 =L2 2
SmSm
m2
2L m60
2
lB X 2 2lB X
qj
q
r
66
p
j
j
3L3 j
j
REFERENCES
1. Polyelectrolytes; Hara, M., Ed.; Marcel Dekker: New York, 1993.
2. Tanford, C. Physical Chemistry of Macromolecules; Wiley: New York, 1961.
3. Oosawa, F. Polyelectrolytes; Marcel Dekker: New York, 1971.
4. Mandel, M. In Encyclopedia of Polymer Science and Engineering, Wiley:
New York, 1988.
5. Schmitz, K.S. Macroions in Solution and Colloidal Suspension; VCH: Weinheim,
1993.
6. Forster, S.; Schmidt, M. Adv. Polym. Sci. 1995, 120, 51133.
7. Barrat, J.L.; Joanny, J.F. Adv. Chem. Phys. 1996, XCIV, 166.
8. Colloid-Polymer Interactions: From Fundamentals to Practice; Dubin, P.L.,
Farinato, R.S., Eds.; Wiley-Interscience: New York, 1999.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
35.
36.
37.
307
308
38.
39.
40.
41.
42.
43.
44.
45.
46.
47.
48.
49.
50.
51.
52.
53.
54.
55.
56.
57.
58.
59.
60.
61.
62.
63.
64.
65.
66.
67.
68.
69.
70.
71.
Dobrynin
Barrat, J.L.; Boyer, D. J. Phys. II (France) 1993, 3, 343356.
Ullner, M.; Jonsson, B.; Widmark, P.O. J. Chem. Phys. 1994, 100, 33653366.
Jonsson, B.; Petersen, C.; Soderberg, B. J. Phys. Chem. 1995, 99, 12511266.
Peterson, C.; Sommelius, O.; Soderberg, B. J. Chem. Phys. 1996, 105,
52335241.
Migliorini, G.; Rostiashvili, V.G.; Vilgis, T.A. Eur. Phys. J. E. 2001, 4,
475487.
Odijk, T. J. Polym. Sci. Polym. Phys. Ed. 1977, 15, 477483.
Skolnick, J.; Fixman, M. Macromolecules 1977, 10, 944948.
Li, H.; Witten, T.A. Macromolecules 1995, 28, 59215927.
Netz, R.R.; Orland, H. Eur. Phys. J. B. 1999, 8, 8198.
Ha, B.Y.; Thirumalai, D. J. Chem. Phys. 1999, 110, 75337541.
Tricot, M. Macromolecules 1984, 17, 16981703.
Reed, C.E.; Reed, W.F. J. Chem. Phys. 1992, 97, 77667776.
Forster, S.; Schmidt, M.; Antonietti, M. J. Phys. Chem. 1992, 96,
40084014.
deNooy, A.E.J.; Besemer, A.C.; van Bekkum, H.; van Dijk, J.; Smit, J.A.M.
Macromolecules 1996, 29, 65416547.
Nishida, K.; Urakawa, H.; Kaji, K.; Gabrys, B.; Higgins, J.S. Polymer 1997,
38, 60836085.
Tanahatoe, J.J. J. Phys. Chem. B 1997, 101, 1044210445.
Beer, M.; Schmidt, M.; Muthukumar, M. Macromolecules 1997, 30,
83758385.
Barrat, J.L.; Joanny, J.F. Europhys. Lett. 1993, 24, 333338.
Bratko, D.; Dawson, K.A. J. Chem. Phys. 1993, 99, 53525364.
Ha, B.Y.; Thirumalai, D. Macromolecules 1995, 28, 577581.
Reed, C.E.; Reed, W.F. J. Chem. Phys. 1991, 94, 84798486.
Seidel, C. Ber. Bunse-Ges. Phys. Chem. 1996, 100, 757763.
Schafer, H.; Seidel, C. Macromolecules 1997, 30, 66586661.
Ullner, M.; Jonsson, B.; Peterson, C.; Sommelius, O.; Soderberg, B. J. Chem.
Phys. 1997, 107, 12791287.
Micka, U.; Kremer, K. Phys. Rev. E 1996, 54, 26532662.
Micka, U.; Kremer, K. J. Phys. Cond. Matter 1996, 8, 94639470.
Micka, U.; Kremer, K. Europhys. Lett. 1997, 38, 279284.
Cannavacciuolo, L.; Sommer, C.; Pedersen, J.S.; Schurtenberger, P. Phys. Rev.
E 2000, 62, 54095419.
Lord Rayleigh. Philos. Mag. 1882, 14, 182.
Hooper, H.H.; Beltran, S.; Sassi, A.P.; Blanch, H.W.; Prausnitz, J.M. J. Chem.
Phys. 1990, 93, 27152723.
Dobrynin, A.V.; Rubinstein, M.; Obukhov, S.P. Macromolecules 1996, 29,
29742979.
Chodanowski, P.; Stoll, S. J. Chem. Phys. 1999, 111, 60696081.
Lyulin, A.V.; Dunweg, B.; Borisov, O.V.; Darinskii, A.A. Macromolecules
1999, 32, 32643278.
Solis, F.J.; Olvera de la Cruz, M. Macromolecules 1998, 31, 55025506.
309
72. Balazs, A.C.; Singh, C.; Zhulina, E.; Pickett, G.; Chern, S.S.; Lyatskaya, Y.
Prog. Surface Sci. 1997, 55, 181269.
73. Edwards, S.F.; King, P.R.; Pincus, P. Ferroelectrics 1980, 30, 36.
74. Higgs, P.G.; Joanny, J-F. J. Chem. Phys. 1991, 94, 15431558.
75. Dobrynin, A.V.; Rubinstein, M. J. Phys. II (France) 1995, 5, 677695.
76. Gutin, A.; Shakhnovich, E. Phys. Rev. E 1994, 50, R3322R3325.
77. Quan, C.; Kholodenko, A. J. Chem. Phys. 1988, 89, 52735281.
78. Bratko, D.; Chakraborty, A.K. J. Phys. Chem. 1996, 100, 11641173.
79. McQuarrie, D.A. Statistical Mechanics; Harper & Row: New York, 1976;
Chapter 15.
80. Kantor, Y.; Kardar, M.; Li, H. Phys. Rev. Lett. 1992, 69, 6164.
81. Kantor, Y.; Kardar, M.; Li, H. Phys. Rev. E 1994, 49, 13831392.
82. Kantor, Y.; Kardar, M. Europhys. Lett. 1994, 27, 643648.
83. Kantor, Y.; Kardar, M. Phys. Rev. E 1995, 51, 12991312.
84. Kantor, Y.; Kardar, M. Phys. Rev. E 1995, 52, 835846.
85. Yamakov, V.; Milchev, A.; Limbach, H.J.; Dunweg, B.; Everaers, R. Phys.
Rev. Lett. 2000, 85, 43054308.
86. Tanaka, M.; Yu Grosberg, A.; Pande, V.S.; Tanaka, T. Phys. Rev. E 1997, 56,
57985808.
87. Soddemann, T.; Schiessel, H.; Bluman, A. Phys. Rev. E 1998, 57, 20812090.
88. Lee, N.; Obukhov, S.P. Eur. Phys. J. B 1998, 1, 371376.
89. Victor, J.M.; Imbert, J.B. Europhys. Lett. 1993, 24, 189195.
90. Imbert, J.B.; Victor, J.M.; Tsunekawa, N.; Hiwatary, Y. Phys. Lett. 1999, 258,
9298.
91. Wittmer, J.; Johner, A.; Joanny, J-F. Europhys. Lett. 1993, 24, 263269.
92. Pande, V.S.; Yu Grosberg, A.; Joerg, C.; Kardar, M.; Tanaka, T. Phys. Rev.
Lett. 1996, 77, 35653569.
93. Brender, C.; Lax, M.; Windwer, S. J. Chem. Phys. 1981, 74, 25762581.
94. Brender, C.; Lax, M. J. Chem. Phys. 1984, 80, 886892.
95. Christos, G.A.; Carnie, S.L. Chem. Phys. Lett. 1990, 172, 249253.
96. Woodward, C.; Jonsson, B. Chem. Phys. 1991, 155, 207213.
97. Valleau, J. Chem. Phys. 1989, 129, 163.
98. Ewald, P.P. Ann. Phys. 1921, 64, 253287.
99. Evjen, H.M. Phys. Rev. 1932, 39, 675.
100. Born, M.; Huang, K. Dynamical Theory of Crystal Lattices; Oxford University
Press: London, 1954.
101. Kittel, C. Introduction to Solid State Physics, 5th Ed.; John Wiley & Sons:
New York, 1976.
102. Toukmaji, A.Y.; Board, J.A. Jr. Comp. Phys. Comm. 1996, 95, 7392.
103. Sagui, C.; Dardin, T.A. Annu. Rev. Biophys. Biomol. Struct. 1999, 28,
155179.
104. Redlack, A.; Grinday, J. J. Phys. Chem. Solids 1975, 36, 7382.
105. De Leeuw, S.W.; Perram, J.W.; Smith, E.R. Proc. R. Soc. London 1980, A373,
2756.
106. Smith, E.R. Proc. R. Soc. London 1981, A375, 475505.
310
Dobrynin
107. Deem, M.W.; Newsam, J.M.; Sinha, S.K. J. Phys. Chem. 1990, 94, 83568359.
108. Perram, J.W.; Petersen, H.G.; De Leeuw, S.W. Molecular Physics 1988, 65,
875893.
109. Dardin, T.; York, D.; Pedersen, L. J. Chem. Phys. 1993, 98, 1008910092.
110. Essmann, U.; Perera, L.; Berkowitz, M.L.; Dardin, T.; Lee, H.; Pedersen, L.G.
J. Chem. Phys. 1995, 103, 85778593.
111. Petersen, H.G. J. Chem. Phys. 1995, 103, 36683680.
112. Saqui, C.; Dardin, T.A. In Simulation and Theory of Electrostatic Interactions
in Solutions, AIP Conference Proceedings, Prat, L.P., Hammer, G., Eds.; Vol.
492, 104113.
113. Dardin, T.A.; Toukmaji, A.; Pedersen, L.G. J. Chim. Phys. 1997, 94,
13461364.
114. Kholmurodov, K.; Smith, W.; Yasuoka, K.; Dardin, T.; Ebisazaki, T.
J. Comp. Chem. 2000, 21, 11871191.
115. Crowley, M.F.; Dardin, T.A.; Cheatham, T.E.; Deereld, D.W. J.
Supercomputing 1997, 11, 255278.
116. Press, W.H.; Teukolsky, S.A.; Vetterling, W.T.; Flannery, B.P. Numerical
Recipies in C. The Art of Scientic Computing; 2nd Ed.; Cambridge
University Press: New York, 1992.
117. Chui, C.K. An Introduction to Wavelets; Academic Press: San Diego, 1992.
118. Schoenberg, I.J. Cardinal Spline Interpolation; SIAM: Philadelphia, 1973.
119. Hockney, R.W.; Eastwood, J.W. Computer Simulations Using Particles;
McGraw-Hill: New York, 1981.
120. Luty, B.A.; Davis, M.E.; Tirony, I.G.; van Gunsteren, W.F. Mol. Simul. 1994,
14, 1120.
121. Shimada, J.; Kaneko, H.; Takada, T. J. Comp. Chem. 1993, 14, 867878.
122. Ferrell, R.; Bertschinger, E. Int. J. Mod. Phys. C 1994, 5, 933956.
123. Pollock, E.L.; Glosli, J. Comp. Phys. Comm. 1996, 95, 93110.
124. Pollock, E.L. In Simulation and Theory of Electrostatic Interactions in
Solutions, AIP Conferences Proceedings, Prat, L.P., Hammer, G., Eds.;
Vol. 492, 146158.
125. Deserno, M.; Holm, C. J. Chem. Phys. 1998, 109, 76787693.
126. Deserno, M.; Holm, C. J. Chem. Phys. 1998, 109, 76947701.
127. Greengard, L.; Rokhlin, V. J. Comp. Phys. 1987, 73, 325348.
128. Greengard, L.; Rokhlin, V. Lecture Notes in Mathematics. 1988, 1360,
121141.
129. Greengard, L.; Rokhlin, V. Chem. Scrip. 1989, 29A, 139144.
130. Greengard, L. Science 1994, 265, 909914.
131. Jackson, J.D. Classical Electrodynamics; 2nd Ed.; Wiley: New York, 1975.
132. Esselink, K. Comp. Phys. Com. 1995, 87, 375395.
133. Carrier, J.; Greengard, L.; Rokhlin, V. SIAM J. Sci. Stat. Comput. 1988,
9, 669686.
134. Schmidt, K.E.; Lee, M.A. J. Stat. Phys. 1991, 63, 12231235.
135. Schmidt, K.E.; Lee, M.A. J. Stat. Phys. 1997, 89, 411424.
136. Manning, G.S. J. Chem. Phys. 1969, 51, 924933.
311
312
Dobrynin
174. Ha, B.Y.; Liu, A.J. In Physical Chemistry of Polyelectrolytes; Radeva, T., Ed.;
Marcel Dekker: New York, 2001; 163181.
175. Stevens, M.J. Phys. Rev. Lett. 1999, 82, 101103.
176. Stevens, M.J. Biophys. J. 2001, 80, 130139.
177. Doi, M.; Edwards, S.F. The Theory of Polymer Dynamics; Oxford University
Press: New York, 1986.
178. Beenakker, C.W.J. J. Chem. Phys. 1986, 85, 1581.
179. Brady, J.F.; Phillips, R.J.; Lester, J.C.; Bossis, G. J. Fluid Mech. 1988, 195,
257.
180. Hase, K.R.; Powel, R.L. J. Phys. of Fluid 2001, 13, 3244.
181. Fleer, G.J.; Cohen Stuart, M.A.; Scheutjens, J.M.H.M.; Gasgove, T.; Vincent,
B. Polymers at Interfaces; Chapman and Hall: London, 1993.
182. Bajpai, A.K. Prog. Polym. Sci. 1997, 22, 523564.
183. Kawaguchi, M.; Takahashi, A. Adv. Coll. Inter. Sci. 1992, 37, 219317.
184. Decher, G.; Eckle, M.; Schmitt, J.; Struth, B. Curr. Opin. in Coll. Inter. Sci.
1998, 3, 3240.
185. Beltran, S.; Hooper, H.H.; Blanch, H.W.; Prausnitz, J.M. Macromolecules
1991, 24, 31783184.
186. Muthukumar, M. J. Chem. Phys. 1995, 103, 47234731.
187. Mashl, R.J.; Gronbech-Jensen, N. J. Chem. Phys. 1998, 109, 46174623.
188. Kong, C.Y.; Muthukumar, M. J. Chem. Phys. 1998, 109, 15221527.
189. Mashl, R.J.; Gronbech-Jensen, N.; Fitzsimmons, M.R.; Lutt, M.; Li, D.Q.
J. Chem. Phys. 1999, 110, 22192295.
190. Yamakov, V.; Milchev, A.; Borisov, O.; Dunweg, B. J. Phys. Cond. Matt.
1999, 11, 99079923.
191. Ellis, M.; Kong, C.Y.; Muthukumar, M. J. Chem. Phys. 2000, 112, 87238729.
192. Csajka, F.S.; van der Linden, C.C.; Seidel, C. Macromol. Symp. 1999, 146,
243249.
193. Csajka, F.S.; Seidel, C. Macromolecules 2000, 33, 27282739.
194. Ennis, J.; Sjostrom, L.; Akesson, T.; Jonsson, B. Langmuir 2000, 16,
71167125.
195. Csajka, F.S.; Netz, R.R.; Seidel, C.; Joanny, J.F. Eur. Phys. J. E 2001, 4,
505513.
196. von Goeler, F. J. Chem. Phys. 1994, 100, 77967803.
197. Wallin, T.; Linse, P. J. Chem. Phys. 1998, 109, 50895100.
198. Chodanowski, P.; Stoll, S. Macromolecules 2001, 34, 23202328.
199. Mateescu, E.M.; Jeppesen, C.; Pincus, P. Eur. Phys. Lett. 1999, 46, 493498.
200. Kunze, K.K.; Netz, R.R. Phys. Rev. Lett. 2000, 85, 43984392.
201. Welch, P.; Muthukumar, M. Macromolecules 2000, 33, 61596167.
9
Gibbs Ensemble and Histogram
Reweighting Grand Canonical
Monte Carlo Methods
ATHANASSIOS Z. PANAGIOTOPOULOS
Princeton, New Jersey, U.S.A.
I.
Princeton University,
INTRODUCTION
314
Panagiotopoulos
315
path to a state of known free energy [5], and the Widom test particle
insertion method [15]. The present chapter focuses on methods that were
specifically designed for phase equilibrium calculations. The relative
precision and accuracy of methods to obtain the chemical potential have
been examined in [16]. Applications of direct interfacial simulations, which
can be performed by either Monte Carlo or molecular dynamics
algorithms, have been reviewed by Rowlinson and Widom [17] and
Gubbins [18].
The plan of this chapter is as follows. Section II deals with the Gibbs
ensemble Monte Carlo method, which is based on simultaneous
calculations in two regions representing equilibrium phases, coupled
indirectly via particle transfers and volume changes. The method is now
commonly used for obtaining phase equilibria of fluids, because of its
simplicity and speed. A single Gibbs ensemble simulation gives a point on
the phase envelope of a multicomponent system. A number of other
methods designed for direct calculations of phase equilibria are described
in Section III. The NPT test particle method (Section III.A) is based
on chemical potential calculations. The method has roughly the same
range of applicability and limitations as the Gibbs ensemble, but requires
multiple simulations per coexistence point. GibbsDuhem integration
(Section III.B) does not require particle insertions and removals and is
applicable to transitions involving solids. It needs to start, however,
from a point on the phase envelope determined by one of the other
techniques. Pseudo-ensembles (Section III.C) provide significant flexibility
in determinations of phase equilibria under different external constraints
and can be implemented in combination with the Gibbs ensemble or
GibbsDuhem integrations. Histogram reweighting methods (Section IV)
provide the free energy and phase behavior with excellent accuracy and
can be used in the vicinity of critical points. The majority of simulation
methods for calculations of phase transitions rely on particle transfers,
which become impractical for dense systems or multisegment molecules. A
number of methods have been developed for improving the efficiency of
particle transfers and have been instrumental in enabling calculations for
realistic potential models. Configurational-bias sampling techniques that
perform smart insertions at favorable locations are described in Section
V.A. Expanded ensembles are based on gradual transfers of parts of
molecules and are described in Section V.B. The last part of this review
(Section VI) describes applications of simulations to calculations of the
phase behavior of polymeric systems. The chapter concludes with a
discussion of the relative strengths and weaknesses of the methods
discussed and provides some suggestions for possible future research
directions in the field.
316
Panagiotopoulos
317
ensemble was given by Smit et al. [21] and Smit and Frenkel [22], which we
follow here. A one-component system at constant temperature T, total
volume V, and total number of particles N is divided into two regions, with
volumes VI and VII V VI, and number of particles NI and NII N NI.
The probability density of states, }(NI,VI; N,V,T ), is proportional to
}NI , VI ; N, V, T
N!
/
expNI ln VI NII ln VII UI NI UII NII
NI !NII !
Smit et al. [21] used a free energy minimization procedure to show that
for a system with a first-order phase transition, the two regions in a Gibbs
ensemble simulation are expected to reach the correct equilibrium densities.
The acceptance criteria for the three types of moves can be immediately
obtained from Eq. (1). For a displacement step internal to one of the
regions, the probability of acceptance is the same as for conventional
constant-NVT simulations,
}move min1, expU
where U is the configurational energy change resulting from the displacement. For a volume change step during which the volume of region I is
increased by V with a corresponding decrease of the volume of region II,
"
VI V
VII V
NII ln
}volume min 1, exp UI UII NI ln
VI
VII
3
Equation (3) implies that sampling is performed uniformly in the volume
itself. The acceptance criterion for particle transfers, written here for
transfer from region II to region I is
"
}transfer
#
NII VI
min 1,
expUI UII
NI 1VII
318
Panagiotopoulos
number of particles of a species in one of the two regions becomes zero after
a successful transfer out of that region. Equation (4) in this case is taken to
imply that the probability of transfer out of an empty region is zero.
The acceptance rules to this point are for a simulation in which the total
system is at constant number of molecules, temperature, and volume. For
pure component systems, the phase rule requires that only one intensive
variable (in this case system temperature) can be independently specified
when two phases coexist. The vapor pressure is obtained from the simulation. By contrast, for multicomponent systems pressure can be specified
in advance, with the total system being considered at constant NPT. The
probability density for this case, }(NI, VI; N, P, T ) is proportional to
}NI ,VI ;N,P,T
h
i
N!
exp NI lnVI NII lnVII UI NI UII NII PVI VII
/
NI !NII !
5
and the only change necessary in the algorithm is that the volume changes in
the two regions are now made independently. The acceptance criterion
for a volume change step in which the volume of region I is changed by
V, while the other region remains unchanged is then
"
#
VI V
}volume min 1, exp UI NI ln
PV
6
VI
An interesting extension of the original methodology was proposed by
Lopes and Tildesley [23], to allow the study of more than two phases at
equilibrium. The extension is based on setting up a simulation with as many
boxes as the maximum number of phases expected to be present. Kristof and
Liszi [24,25] have proposed an implementation of the Gibbs ensemble in
which the total enthalpy, pressure, and number of particles in the total
system are kept constant. Molecular dynamics versions of the Gibbs
ensemble algorithm are also available [2628], and also in the following
Chapter 10 by M. Kotelyanskii and R. Hentschke.
The physical reason for the ability of the Gibbs ensemble to converge to a
state that contains phases at their equilibrium density in the corresponding
boxes, rather than a mixture of the two phases in each box, is the free energy
cost for creating and maintaining an interface. Essentially, in the Gibbs
ensemble, one trades off the directness of the interfacial simulation
approach with the (slight) inconvenience of setting up and maintaining
two or more boxes for the equilibrium phases. However, much smaller
system sizes can be used relative to interfacial simulations, and the
319
320
Panagiotopoulos
321
solvent and polymer for compositions different from the original simulation
conditions can be made using standard thermodynamic relations and the
number of required iterations is significantly reduced.
B. GibbsDuhem Integration
Most methods for determination of phase equilibria by simulation rely on
particle insertions to equilibrate or determine the chemical potentials of the
components. Methods that rely on insertions experience severe difficulties
for dense or highly structured phases. If a point on the coexistence curve is
known (e.g., from Gibbs ensemble simulations), the remarkable method of
Kofke [41,42] enables the calculation of a complete phase diagram from a
series of constant-pressure simulations that do not involve any transfers of
particles. For one-component systems, the method is based on integration
of the ClausiusClapeyron equation over temperature,
dP
d
sat
H
V
where sat indicates that the equation holds on the saturation line, and H
is the difference in enthalpy between the two coexisting phases. The right
hand side of Eq. (7) involves only mechanical quantities that can be simply
determined in the course of a standard Monte Carlo or molecular dynamics
simulation. From the known point on the coexistence curve, a change in
temperature is chosen, and the saturation pressure at the new temperature
is predicted from Eq. (7). Two independent simulations for the corresponding phases are performed at the new temperature, with gradual changes of the
pressure as the simulations proceed to take into account the enthalpies and
densities at the new temperature as they are being calculated.
Questions related to propagation of errors and numerical stability of the
method have been addressed in [42] and [43]. Errors in initial conditions
resulting from uncertainties in the coexistence densities can propagate and
increase with distance from the starting point when the integration path is
towards the critical point [43]. Near critical points, the method suffers from
instability of a different nature. Because of the small free energy barrier
for conversion of one phase into the other, even if the coexistence pressure is
set properly, the identity of each phase is hard to maintain and large
fluctuations in density are likely. The solution to this last problem is to
borrow an idea from the Gibbs ensemble and couple the volume changes of
the two regions [42]. Extensions of the method to calculations of three-phase
coexistence lines are presented in [44] and to multicomponent systems in [43].
Unfortunately, for multicomponent systems the GibbsDuhem integration
322
Panagiotopoulos
sat
u
,
where is the density (N/V) and u the energy per particle. This method
was applied to continuous-phase polymeric systems in [46] and to lattice
models in [50].
The GibbsDuhem integration method excels in calculations of solid
fluid coexistence [48,49], for which other methods described in this chapter
are not applicable. An extension of the method that assumes that the initial
free energy difference between the two phases is known in advance, rather
than requiring it to be zero, has been proposed by Meijer and El Azhar [51].
The procedure has been used in [51] to determine the coexistence lines
of a hard-core Yukawa model for charge-stabilized colloids.
C. Pseudo-Ensembles
The GibbsDuhem integration method represents a succesful combination
of numerical methods and molecular simulations. Taking this concept
even further, Mehta and Kofke [52] proposed a pseudo-grand canonical
ensemble method in which a system maintains a constant number of
particles and temperature, but has a fluctuating volume to ensure that, at
the final density, the imposed value of the chemical potential is matched.
The formalism still requires that estimates of the chemical potential be made
during the simulation. The main advantage of the approach over more
traditional grand canonical ensemble methods is that it provides additional
flexibility with respect to the method to be used for determination of
the chemical potential. For example, the chain increment method [40]
for chain molecules, which cannot be combined with grand canonical
simulations, can be used for the chemical potential evaluations in a pseudogrand canonical simulation (as in [38]).
The same pseudo-ensemble concept has been used by Camp and Allen
[53] to obtain a pseudoGibbs method in which particle transfers are
substituted by volume fluctuations of the two phases. The volume
323
A. One-Component Systems
A GCMC simulation for a one-component system is performed as follows.
The simulation cell has a fixed volume V, and is placed under periodic
boundary conditions. The inverse temperature, 1/kBT and the chemical
potential,
, are specified as input parameters to the simulation. Histogram
reweighting requires collection of data for the probability f(N,E) of
occurrence of N particles in the simulation cell with total configurational
energy in the vicinity of E. This probability distribution function follows the
relationship
f N, E
N, V, EexpE
N
, V,
324
Panagiotopoulos
10
QN, V, exp
N
, V,
11
12
325
FIG. 3 The function ln [f(N)]
N for the data of Fig. 2. The figure shows the raw
curves for
1 and
2 as well as a composite curve formed by shifting the data
by the amount indicated by the arrows. (From [14], 2000 IOP Publishing Ltd.)
Figure 3 shows the raw curves for
1 and
2 as well as a composite curve
formed by shifting data for the two runs by the amount indicated by the
arrows. The combined curve provides information over the combined range
of particle numbers, N, covered by the two runs. Note that by keeping onedimensional histograms for N we are restricted to combining runs of the
same temparature, while the more general form [Eq. (10)] allows combination of runs at different temperatures.
Simulation data are subject to statistical (sampling) uncertainties, which
are particularly pronounced near the extremes of particle numbers and
energies visited during a run. When data from multiple runs are combined as
shown in Fig. 3, the question arises of how to determine the optimal amount
by which to shift the raw data in order to obtain a global free energy
function. Ferrenberg and Swendsen [68] provided a solution to this problem
by minimizing the differences between predicted and observed histograms.
In this approach, it is assumed that multiple overlapping runs, i 1, 2, . . . , R
are available for a given system. The composite probability, }(N, E;
, ), of
observing N particles and energy E, if one takes into account all runs and
assumes that they have the same statistical efficiency, is
XR
fi N, E expE
N
}N, E;
, XRi1
K expi E i
i N Ci
i1 i
13
P
where Ki is the total number of observations [Ki N,E fi(N, E )] for run i.
The constants Ci (also known as weights) are obtained by iteration
326
Panagiotopoulos
XX
E
}N, E; i , i
14
Given an initial guess for the set of weights Ci, Eqs. (13) and (14) can be
iterated until convergence. When many histograms are to be combined this
convergence of the FerrenbergSwendsen weights can take a long time.
Once this has been achieved, however, all thermodynamic quantities for the
system over the range of densities and energies covered by the histograms
can be obtained. For example, the mean configurational energy U(
,) is
hUi
,
XX
E
}N, E; , E
15
1 XX
}N, E;
, N
V E N
16
2 , V, 2
2 P2 V 1 P1 V
1 , V, 1
17
327
328
Panagiotopoulos
FIG. 5 Schematic diagram of the mean number of particles, hNi, vs. chemical
potential,
, for a subcritical and a supercritical isotherm of a one-component fluid.
The curve for the supercritical isotherm has been shifted up for clarity. (From [14],
2000 IOP Publishing Ltd.)
state, the system remains on the high-density branch of the isotherm until
some value of the chemical potential that is lower than the chemical
potential of the jump from low- to high-density states.
The histogram reweighting method can be applied to systems with large
free energy barriers for transitions between states, provided that care is
taken to link all states of interest via reversible paths. One possibility is to
utilize umbrella or multicanonical sampling techniques [61,69] to artificially
enhance the frequency with which a simulation samples the intermediate
density region [62]. Multicanonical and umbrella sampling require as input
an estimate of the free energy in the intermediate density region, which has
to be obtained by trial and error. In addition, a significant fraction of
simulation time is spent sampling unphysical configurations of intermediate
density. An alternative approach is to link states by providing connections
through a supercritical path, in a process analogous to thermodynamic
integration [5]. This approach is illustrated schematically in Fig. 6. The filled
square represents the critical point for a transition, and open squares linked
by dashed lines represent tie-lines. Ellipses represent the range of particle
numbers and energies sampled by a single simulation. A near-critical
simulation samples states on both sides of the coexistence curve, while
subcritical simulations are likely to be trapped in (possibly metastable) states
on either side. However, as long as there is a continuous path linking all
states of interest, the free energies and pressures can be calculated correctly,
and an accurate phase envelope can be obtained.
An example of the application of histogram reweighting for determining
the phase behavior of a homopolymer model on the simple cubic lattice is
329
FIG. 6 Schematic diagram of the energy, E, vs. the number of particles, N, for a
one-component fluid with a phase transition. Squares linked by dashed lines are
coexisting phases joined by tie-lines and the filled square indicates the critical point
of the transition. Ellipses represent the range of particle numbers and energies
sampled during different GCMC runs. (From [14], 2000 IOP Publishing Ltd.)
illustrated in Fig. 7. The phase behavior and critical properties of the model
for a range of chain lengths have been studied in [63]. The system in this
example is for chain length r 8 and coordination number z 6. In this
example, we first performed a simulation at reduced temperature T * 11.5
and chemical potential
* 60.4, for which the raw histogram data are
shown in Fig. 4. The resulting average volume fraction for the run is
indicated on Fig. 7 by the filled circle at T * 11.5. The range of volume
fractions sampled during the simulation is indicated on Fig. 7 by the arrows
originating at the run point. Because this run is near the critical point, a very
broad range of particle numbers and thus volume fractions is sampled
during this single run. The histogram from this run was then reweighted to
lower temperatures and a preliminary phase diagram was obtained. The
estimated coexistence chemical potential at T * 9 was used as input to a
new simulation, which sampled states near the saturated liquid line. The
same procedure was repeated, now with combined histograms from the first
two runs, to obtain an estimate of the coexistence chemical potential at
T * 7. A new simulation was performed to sample the properties of the
liquid at that temperature. The total time for the three runs was 10 CPU min
on a Pentium III 300 MHz processor. The final result of these three
calculations was the phase coexistence lines shown by the thick continuous
lines on Fig. 7.
Two general observations can be made in relation to this example.
First, it should be pointed out that the histogram reweighting method works
much faster on smaller system sizes. As the system size increases, relative
330
Panagiotopoulos
331
and vapor peaks of the histograms. This overlap renders calculations of the
liquid and gas densities imprecise. Larger system sizes suffer less from this
effect and can be used to obtain coexistence densities near critical points. As
discussed in Section IV.C, a sequence of studies with increasing system size
is also required to obtain accurate estimates of critical points.
B. Multicomponent Systems
The histogram reweighting methodology for multicomponent systems
[6567] closely follows the one-component version described above. The
probability distribution function for observing N1 particles of component 1
and N2 particles of component 2 with configurational energy in the vicinity
of E for a GCMC simulation at imposed chemical potentials
1 and
2,
respectively, at inverse temperature in a box of volume V is
f N1 , N2 , E
N1 , N2 , V, EexpE
1 N1
2 N2
1 ,
2 , V,
18
332
Panagiotopoulos
and eventually exceeds the linear system size L of the simulation box.
Singularities and discontinuities that characterize critical behavior in the
thermodynamic limit are smeared out and shifted in finite systems. The
infinite-volume critical point of a system can, however, be extracted by
examining the size dependence of thermodynamic observables, through
finite-size scaling theory [7173]. The finite-size scaling approach proposed
by Bruce and Wilding [74,75] accounts for the lack of symmetry between
coexisting phases in most continuous-space fluids. Even though some recent
work has cast some shadow on its full reliability [76], the approach seems
to be quite robust and is easy to apply. For one-component systems, the
ordering operator, M, is proportional to a linear combination of the number
of particles N and total configurational energy U:
M / N sU
19
20
21
333
3 [81].
334
Panagiotopoulos
FIG. 9 Critical temperature (a) and density (b) scaling with linear system size for
the square-fluid of range
3. Solid lines represent a least-squares fit to the points.
(From [81], 1999 American Institute of Physics.)
335
B. Expanded Ensembles
Another approach for handling multisegment molecules is based on the
concept of expanded ensembles [9194]. Expanded ensembles for chain
molecules construct a series of intermediate states for the molecule of
interest, from a noninteracting (phantom) chain to the actual chain with all
segments and interactions in place. These intermediate states can be
semipenetrable chains of the full length [91,92] or shortened versions of the
actual chain [93,94]. Estimates of the free energy of the intermediate states
are required to ensure roughly uniform sampling, as for thermodynamic
and Hamiltonian scaling methods mentioned in the previous section. The
advantage of expanded ensembles over configurational-bias methods is that
arbitrarily complex long molecules can be sampled adequately, if sufficient
computational effort is invested in constructing good approximations of the
free energies of intermediate states.
336
Panagiotopoulos
337
FIG. 10 Phase diagrams of selected n-alkanes. The curves from bottom to top are
for ethane, propane, butane, pentane, hexane, octane, and dodecane. Circles
represent calculations for the model of [109]. Uncertainties are smaller than the size
of the symbols. A solid line is used for experimental data and a star for the
experimental critical point. (From [109], 1999 American Chemical Society.)
n-heptane between He and n-heptane liquid phases [110]. They have also
obtained partition coefficients of alkanes between water and octanol [111].
Delhommelle et al. [112,113] studied mixtures of n-alkanes using both a
united-atom model [96] and anisotropic united-atom models originated
by Toxvaerd [114,115]. Other recent studies of alkane mixtures include
[109,116,117]. The solubility of small molecules such N2 and methane and
their mixtures in polyethylene, including effects of polymer crystallinity,
was studied in [118]. Mixtures with -olephins were studied in [38,106]. In
general, excellent agreement between experiment and simulation results is
obtained for these nonpolar mixtures, provided that the pure component
potentials have been optimized to reproduce the phase envelope and vapor
338
Panagiotopoulos
FIG. 11 Vapor pressures of selected n-alkanes. The curves from right to left are for
ethane, propane, butane, pentane, hexane, octane, and dodecane. Symbols are the
same as for Fig. 10. (From [109], 1999 American Chemical Society.)
VII.
CONCLUDING REMARKS
339
340
Panagiotopoulos
ACKNOWLEDGEMENTS
Research in the authors group on which this chapter is based has been
supported by the Department of Energy (Office of Basic Energy Sciences),
the National Science Foundation, the American Chemical Society
Petroleum Research Fund, and the Camille and Henry Dreyfus Foundation.
REFERENCES
1. Rowlinson, J.S.; Swinton, F. Liquids and Liquid Mixtures, 3rd Ed.; Butterworth
Scientic: London, 1982.
2. Prausnitz, J.M.; Lichtenthaler, R.N.; de Azevedo, E.G. Molecular
Thermodynamics of Fluid Phase Equilibria, 3rd Ed.; Prentice-Hall: New York,
1999.
3. Hansen, J.-P.; Verlet, L. Phys. Rev. 1969, 184, 151161.
4. Allen, M.P.; Tildesley, D.J. Computer Simulation of Liquids; Oxford University
Press: Oxford, 1987.
5. Frenkel, D.; Smit, B. Understanding Molecular Simulation; Academic Press:
London, 1996.
6. Panagiotopoulos, A.Z. Molec. Simulation 1992, 9, 123.
7. Smit. B. In Computer Simulations in Chemical Physics, Allen, M.P., Tildesley,
D.J., Eds.; NATO ASI Ser C, 1993, 397, 173209.
8. Panagiotopoulos, A.Z. In Kiran, E., Levelt Sengers, J.M.H., Eds.; NATO ASI
Ser. E, 1994, 273, 411437.
9. Panagiotopoulos, A.Z. In Observation, Prediction and Simulation of Phase
Transitions in Complex Fluids. NATO ASI Ser. C 1995, 460, 463501.
10. Baus, M.; Rull, L.F.; Ryckaert, J.-P. Eds. Observation, Prediction and
Simulation of Phase Transitions in Complex Fluids; NATO ASI Ser. C,
1995, 460.
11. Binder, K. Rep. Prog. Phys. 1997, 60, 487559.
12. Deem, M.W. AIChE Journal 1998, 44, 25692596.
13. Siepmann, J.I. Adv. Chem. Phys. 1999, 105, 443460.
14. Panagiotopoulos, A.Z. J. Phys.: Condens Matter 2000, 12, R2552.
15. Widom, B. J. Chem. Phys. 1963, 39, 28082812.
16. Kofke, D.A.; Cummings, P.T. Molec. Phys. 1997, 92, 973996.
17. Rowlinson, J.S.; Widom, B. Molecular Theory of Capillarity; Clarendon Press:
Oxford, Ch. 6, 1982.
18. Gubbins, K.E. Molec. Simulation 1989, 2, 223252.
19. Panagiotopoulos, A.Z. Molec. Phys. 1987, 61, 813826.
341
20. Panagiotopoulos, A.Z.; Quirke, N.; Stapleton, M.; Tildesley, D.J. Molec. Phys.
1988, 63, 527545.
21. Smit, B.; de Smedt, P.; Frenkel, D. Molec. Phys. 1989, 68, 931950.
22. Smit, B.; Frenkel, D. Molec. Phys. 1989, 68, 951958.
23. Lopes, J.N.C.; Tildesley, D.J. Molec. Phys. 1997, 92, 187195.
24. Kristof, T.; Liszi, J. Molec. Phys. 1997, 90, 10311034.
25. Kristof, T.; Liszi, J. Molec. Phys. 1998, 94, 519525.
26. Hentschke, R.; Bast, T.; Aydt, E. et al. J. Molec. Model 1996, 2, 319326.
27. Kotelyanskii, M.J.; Hentschke, R. Molec. Simulation 1996, 17, 95112.
28. Baranyai, A.; Cummings, P.T. Molec. Simulation 1996, 17, 2125.
29. Valleau, J.P. J. Chem. Phys. 1998, 108, 29622966.
30. Bruce, A.D. Phys. Rev. E 1997, 55, 23152320.
31. Mon, K.K.; Binder, K. J. Chem. Phys. 1992, 96, 69896995.
32. Panagiotopoulos, A.Z. Int. J. Thermophys. 1989, 10, 447457.
33. Moller, D.; Fischer, J. Molec. Phys. 1990, 69, 463473; erratum in Molec. Phys.
1992, 75, 14611462.
34. Lot, A.; Vrabec, J.; Fischer, J. Molec. Phys. 1992, 76, 13191333.
35. Boda, D.; Liszi, J.; Szalai, I. Chem. Phys. Lett. 1995, 235, 140145.
36. Vrabec, J.; Fischer, J. Molec. Phys. 1995, 85, 781792.
37. Vrabec, J.; Lot, A.; Fischer, J. Fluid Phase Equil. 1995, 112, 173197.
38. Spyriouni, T.; Economou, I.G.; Theodorou, D.N. Phys. Rev. Lett. 1998, 80,
44664469.
39. Spyriouni, T.; Economou, I.G.; Theodorou, D.N. Macromolecules 1998, 31,
14301431.
40. Kumar, S.K.; Szleifer, I.; Panagiotopoulos, A.Z. Phys. Rev. Lett. 1991, 66,
29352938.
41. Kofke, D.A. Molec. Phys. 1993, 78, 13311336.
42. Kofke, D.A. J. Chem. Phys. 1993, 98, 41494162.
43. Mehta, M.; Kofke, D.A. Chem. Eng. Sci. 1994, 49, 26332645.
44. Agrawal, R.; Mehta, M.; Kofke, D.A. Int. J. Thermophys. 1994, 15,
10731083.
45. Kofke, D.A. Adv. Chem. Phys. 1999, 105, 405441.
46. Escobedo, F.A.; Pablo, J.J. J. Chem. Phys. 1997, 106, 98589868.
47. Escobedo, F.A.; Pablo, J.J. J. Chem. Phys. 1997, 106, 29112923.
48. Agrawal, R.; Kofke, D.A. Phys. Rev. Lett. 1995, 74, 122125.
49. Bolhuis, P.; Frenkel, D. J. Chem. Phys. 1997, 106, 666687.
50. Escobedo, F.A.; de Pablo, J.J. Europhys. Lett. 1997, 40, 111117.
51. Meijer, E.J.; El Azhar, F. J. Chem. Phys. 1997, 106, 46784683.
52. Mehta, M.; Kofke, D.A. Molec. Phys. 1995, 86, 139147.
53. Camp, P.J.; Allen, M.P. Molec. Phys. 1996, 88, 14591469.
54. Escobedo, F.A. J. Chem. Phys. 1998, 108, 87618772.
55. Escobedo, F.A. J. Chem. Phys. 1999, 110, 1199912010.
56. McDonald, I.R.; Singer, K. J. Chem. Phys. 1967, 47, 47664772.
57. Wood, E.E. J. Chem. Phys. 1968, 48, 415434.
58. Card, D.N.; Valleau, J.P. J. Chem. Phys. 1970, 52, 62326240.
342
59.
60.
61.
62.
63.
64.
65.
66.
67.
68.
69.
70.
71.
72.
73.
74.
75.
76.
77.
78.
79.
80.
81.
82.
83.
84.
85.
86.
87.
88.
89.
90.
91.
92.
93.
94.
Panagiotopoulos
Bennett, C.H. J. Comp. Phys. 1976, 22, 245268.
Ferrenberg, A.M.; Swendsen, R.H. Phys. Rev. Lett. 1988, 61, 26352638.
Berg, B.A.; Neuhaus, T. Phys. Rev. Lett. 1992, 68, 912.
Wilding, N.B. Phys. Rev. E 1995, 52, 602611.
Panagiotopoulos, A.Z.; Wong, V.; Floriano, M.A. Macromolecules 1998, 31,
912918.
Yan, Q.L.; de Pablo, J.J. J. Chem. Phys. 2000, 113, 59545957.
Wilding, N.B. Phys. Rev. E 1997, 55, 66246631.
Wilding, N.B.; Schmid, F.; Nielaba, P. Phys. Rev. E 1998, 58, 22012212.
Poto, J.J.; Panagiotopoulos, A.Z. J. Chem. Phys. 1998, 109, 1091410920.
Ferrenberg, A.M.; Swendsen, R.H. Phys. Rev. Lett. 1989, 63, 11951198.
Torrie, G.M.; Valleau, J.P. J. Comp. Phys. 1977, 23, 187199.
Wilding, N.B. J. Phys.: Condens Matter 1997, 9, 585612.
Ferdinand, A.E.; Fisher, M.E. Phys. Rev. 1969, 185, 832.
Fisher, M.E.; Barber, M.N. Phys. Rev. Lett. 1972, 28, 1516.
Privman, V. Ed. Finite Size Scaling and Numerical Simulation of Statistical
Mechanical Systems, World Scientic: Singapore, 1990.
Bruce, A.D.; Wilding, N.B. Phys. Rev. Lett. 1992, 68, 193196.
Wilding, N.B.; Bruce, A.D. J. Phys.: Condens Matter 1992, 4, 30873108.
Orkoulas, G.; Fisher, M.E.; Ustun, C. J. Chem. Phys. 2000, 113, 75307545.
Sengers, J.V.; Levelt-Sengers, J.M.H. Ann. Rev. Phys. Chem. 1986, 37,
189222.
Yan, Q.; Liu, H.; Hu, Y. Macromolecules 1996, 29, 40664071.
Frauenkron, H.; Grassberger, P. J. Chem. Phys. 1997, 107, 95999608.
Liu, A.J.; Fisher, M.E. Physica A 1989, 156, 35.
Orkoulas, G.; Panagiotopoulos, A.Z. J. Chem. Phys. 1999, 110, 15811590.
Valleau, J.P. Advances in Chemical Physics 1999, 105, 369404.
Kiyohara, K.; Spyriouni, T.; Gubbins, K.E. et al. Molec. Phys. 1996, 89,
965974.
Errington, J.R.; Panagiotopoulos, A.Z. J. Chem. Phys. 1998, 109, 10931100.
Rosenbluth, M.N.; Rosenbluth, A.W. J. Chem. Phys. 1955, 23, 356359.
Frenkel, D.; Mooij, G.C.A.M.; Smit, B. J. Phys.: Condens Matter 1992, 4,
30533076.
de Pablo, J.J.; Laso, M.; Siepmann, J.I.; Suter, U.W. Molec. Phys. 1993, 80,
5563.
Siepmann, J.I.; McDonald, I.R. Molec. Phys. 1993, 79, 457473.
Mooij, G.C.A.M.; Frenkel, D.; Smit, B. J. Phys.: Condens Matter 1992, 4,
L255L259.
Laso, M.; de Pablo, J.J.; Suter, U.W. J. Chem. Phys. 1992, 97, 28172819.
Lyubartsev, A.; Martsinovski, A.; Shevkunov, S. J. Chem. Phys. 1992, 96,
17761783.
Vorontsov-Velyaminov, P.; Broukhno, A.; Kuznetsova, T. J. Phys. Chem.
1996, 100, 11531158.
Wilding, N.B.; Muller, M. J. Chem. Phys. 1994, 101, 43244330.
Escobedo, F.; de Pablo. J.J. J. Chem. Phys. 1996, 105, 43914394.
343
95. Siepmann, J.I.; Karaborni, S.; Smit, B. J. Am. Chem. Soc. 1993, 115,
64546455.
96. Smit, B.; Karaborni, S.; Siepmann, J.I. J. Chem. Phys. 1995, 102, 21262140.
97. Siepmann, J.I.; Karaborni, S.; Smit, B. Nature 1993, 365, 330332.
98. Dodd, L.R.; Theodorou, D.N. Adv. Pol. Sci. 1994, 116, 249281.
99. Cui, S.T.; Cummings, P.T.; Cochran, H.D. Fluid Phase Equil. 1997, 141,
4561.
100. Siepmann, J.I.; Martin, M.G.; Mundy, C.J.; Klein, M.L. Mol. Phys. 1997, 90,
687693.
101. Martin, M.G.; Siepmann, J.I. J. Phys. Chem. B 1999, 103, 45084517.
102. Zhuravlev, N.D.; Siepmann, J.I. Fluid Phase Equil. 1997, 134, 5561.
103. Cui, S.T.; Siepmann, J.I.; Cochran, H.D.; Cummings, P.T. Fluid Phase Equil.
1998, 146, 5161.
104. Potter, S.C.; Tildesley, D.J.; Burgess, A.N. et al. Molec. Phys. 1997, 92,
825833.
105. Jedlovszky, P.; Mezei, M. J. Chem. Phys. 1999, 110, 29913002.
106. Spyriouni, T.; Economou, I.G.; Theodorou, D.N. J. Am. Chem. Soc. 1999,
121, 34073413.
107. Martin, M.G.; Siepmann, J.I. J. Phys. Chem. B 1998, 102, 25692577.
108. Nath, S.K.; Escobedo, F.A.; de Pablo, J.J. J. Chem. Phys. 1998, 108,
99059911.
109. Errington, J.R.; Panagiotopoulos, A.Z. J. Phys. Chem. B 1999, 103, 6314.
110. Chen, B.; Siepmann, J.I. J. Phys. Chem. B 1999, 103, 53705379.
111. Chen, B.; Siepmann, J.I. J. Am. Chem. Soc. 2000, 122, 64646467.
112. Delhommelle, J.; Boutin, A.; Tavitian, B.; Mackie, A.D.; Fuchs, A.H. Molec.
Phys. 1999, 96, 15171524.
113. Ungerer, P.; Beauvais, C.; Delhommelle, J.; Boutin, A.; Rousseau, B.; Fuchs,
A.H. J. Chem. Phys. 2000, 112, 54995510.
114. Padilla, P.; Toxvaerd, S. J. Chem. Phys. 1991, 95, 509519.
115. Toxvaerd, S. J. Chem. Phys. 1997, 107, 51975204.
116. Mackie, A.D.; Tavitian, B.; Boutin, A.; Fuchs, A.H. Mol. Simulation 1997,
19, 115.
117. Poto, J.J.; Errington, J.R.; Panagiotopoulos, A.Z. Molec. Phys. 1999, 97,
10731083.
118. Nath, S.K.; de Pablo, J.J. J. Phys. Chem. B 1999, 103, 35393544.
10
Gibbs Ensemble Molecular Dynamics
MICHAEL KOTELYANSKII
New Jersey, U.S.A.
REINHARD HENTSCHKE
I.
THE METHOD
346
The first term is the sum over atomatom pair potentials Uinter
P
2 i, 2 j,i>j,j 1,M(r) depending on the vectors r connecting atoms
and in the two molecules i and j. Uintra describes the interactions within
the molecule.
In order to simulate a variable number of molecules i in each of the two
boxes we introduce an extra degree of freedom i for every molecule. i can
vary between 1 and 0, where i 1 means that molecule i is in box one,
whereas i 0 means that it is in box two. For 1>i>0 it is in a transition
state, where it is felt in both boxes. Thus, we rewrite the intermolecular
potential energy of the system as a function of the coordinates and the i as
X
Uinter
r , V1
i
j
r , V2
1 i
1 j
where V1 and V2 are the volumes of the two boxes. The two terms represent
the intermolecular potential energies of the first and the second box, respectively. Consider, for instance, two particles with i 1 and j 1, i.e., both
particles belong to box one. In this case the product (1 i)(1 j) vanishes,
and only the first term in Eq. (2) will contribute to the nonbonded potential
energy. Notice also that as soon as we apply periodic boundary conditions,
and interparticle interactions are calculated involving the particles closest
images, the distance between them, and therefore the intermolecular
potential energy, is a function of the box dimensions (or of the volume if
the shape of the box is kept fixed).
347
gi
p_
@U
@Uinter @gi
@i
@i
@i
pV1 QP V_1
348
@U
@V1 "
X
@
r , V1
i
j
@V1 2i, 2j, i>j, j1, M
p_ V1
r , V2
1 i
1 j
Pe1 Pe2
Here p and pi are the momenta conjugate to the Cartesian coordinates r
and transfer variable i of the particle i. The first three equations describe
the evolution of a system coupled to an external heat bath with the temperature T [2,3]. is an additional degree of freedom, describing the coupling of the system to the external thermostat, which is necessary to simulate
constant temperature. QT describes the strength of the coupling, and X
equals the number of degrees of freedom, coupled to the thermostat.
The next two equations govern the evolution of the i, and thus the
transfer of the molecules between the boxes. Note that the intramolecular
part of the potential energy Uintra is independent of the i, and only
intermolecular interactions and the additional potential g(i) appear on the
right hand side.
The last two equations are the equations of motion of the box volume V1,
where pV1 is a momentum variable conjugate to the volume of the first box
V1, and QP is a parameter governing the volume relaxation. Again only the
intermolecular interactions Uinter depend on the box size, because periodic
boundary conditions are not applied to the intramolecular interactions.
Distances describing the intramolecular interactions Uintra are smaller than
the box sizes. They are of the order of a few chemical bond lengths, while the
box sizes are usually many times larger. As the sum of the volumes of the
two boxes is constant, pV1 pV2. Note that, similar to any constantpressure MD or MC algorithm, volume changes are controlled by the
difference between the instantaneous values of the external pressures [4,5].
349
time. In our case the phase space includes Cartesian coordinates and
momenta, thermostat coupling variable , volume V1, the transfer variables
i and their conjugate momenta pi and pV1.* The stationary (timeindependent) solution of this equation, describes the distribution function
for the ensemble reproduced.
X @
X @
@
@
@
r_
p_
p_ i
_ i
@t 2i, i1, M
@r
@p
@i
@pi
i1, M
@
@
@
p_ V1
_
V_1
@V1
@pV1
@
"
#
X @_i @V_ 1 @p_ V
X @_r @_p
@_
1
0
@r @p
@i @V1 @pV1 @
2i, i1, M
i1, M
By direct substitution and using Eqs. (4) to evaluate partial derivatives, one
can see that the following density function does indeed satisfy the Liouville
equation with @/@t 0. The solution is given by the following expression
GEMD fr g, fp g, fi g, fpi g, V1 , pV1 ,
"
1
/ exp
Utotal fr g, fi g, V1
kB T
X
X p2
p2
p2
Q T 2
V1
!#
6
Here, mi is a mass assigned to the particle for motion along the virtual
transfer direction i. Provided the system is ergodic, the averages over trajectories obtained by integrating Eqs. (4) are equivalent to the averages
calculated with the distribution function GEMD (6). Below we show that,
provided the number of particles in the transition state is small, averages for
particles in each box do in fact correspond to the ensemble with constant
temperature, pressure, and chemical potential, and that in fact the pressures
and chemical potentials are equal in both boxes.
The GEMD trajectory average of a property A, which depends on
the coordinates and momenta of the atoms 1, . . . , n, constituting the
350
Q0GEMD
hAiGEMD
Q0GEMD
"
Here we already integrated over , pi, pV1 and we have canceled the respective factors in the average and in the partition function, as indicated by the
primed Q0GEMD . By choosing a proper g(i) one can make the number of
particles in the transition state negligibly small. In this case we can assume
that i takes only two values 0 or 1, and therefore we can replace the
integration over i by the summation over all possible combinations of
the i values (0 or 1). The equation for the average becomes:
Z
M!
drn dpn dV1 Afr1 ,...,rn g,fp1 ,...,pn g
hAiGEMD 0
QGEMD m0 m!M m!
"
!#
n
X
1
p2
Utotal fr g,V1
exp
kB T
2m
2i,i1
"
!#
Z
M
2
X
1
p
drNn dpNn exp
Utotal fr g,V2
kB T
2m
2i,im1
1
M
X
8
Q0GEMD transforms analogously. M!/(m!(M m)!) accounts for particle distinguishability. Here we also write the potential energies in the two boxes as
Utotal({r},V1) and Utotal({r},V2), independent of the i, because we neglect
particles in the transfer state with 0<i<1, assuming i can only be equal
to 0 or 1. In this case the total potential energy is a sum of potential energies
due to the interactions inside each of the boxes, as seen from Eq. (2).
Now we are going to demonstrate, that the chemical potentials and the
pressures are indeed the same in the two simulation boxes. The second
integration over the coordinates and momenta of the N n atoms in the
second box equals QMm, V2 , T h3Mm M m!, where QM m,V2 ,T is the
351
P2 V2
C
1
QMm, P2 , T exp
1O
Mm
kB T N n1=2
2 n
exp
kB T
QMm, P2 , T
QM, P2 , T
10
where
2 is the chemical potential in the second box. Substituting all these in
the average hA i GEMD (8), and canceling terms independent of n between
numerator and denominator, we obtain in the thermodynamic limit:
hAiGEMD
1
2 m
exp
Q00GEMD m0 m!h3M
kB T
Z
drn dpn dV1 Afr1 , . . . , rn g, fp1 , . . . , pn g
"
!#
m
X
P2 V1
1
p2
Utotal fri g, V1
exp
exp
kB T
kB T
2m
2i, i1
1
M
X
11
and the corresponding expression for Q00GEMD , where the double prime is a
reminder that the common factors are canceled and that we keep only the
leading contribution in the limit of a large number of particles. The above
Eq. (11) coincides with the average over the constant- (
, P, T) ensemble [8].
Notice that the average for the first box is calculated with the pressure and
chemical potential values of the second box. This proves that these values
are indeed the same in both boxes.
III. IMPLEMENTATION
Two boxes in the GEMD are implemented in the following way: Cartesian
coordinates of the particles do not change when the particle is transferred
from one box to another. Its presence in either box is determined by the
352
353
354
The same changes as described above for the atomic system are to
be made.
2.
3.
355
All atoms in the same molecule are assigned the same transfer
mass mi.
After the forces fxi along the i direction are calculated for all
atoms, they are redistributed evenly between the atoms of the same
molecule, so that the total force along the i direction acting on the
molecule remains the same. Here is how it looks in a pseudo-code:
for all molecules
ftotal0
for all atoms in the molecule
ftotal ftotalfxi(atom)
end
ftotal ftotal/num_of_atoms
for all atoms in the molecule
fxi(atom)ftotal
end
end
4.
coordinate updates for the next time step, similar to those for
Cartesian coordinates are to be added.
356
FIG. 2 Liquidgas coexistence curve for hexane. Open circles: experimental data of
reference [12]; open triangles: Gibbs Ensemble Monte Carlo result with 13.8 A cutoff
with the long-range corrections from [13]. The large solid circle corresponds to the
experimental critical point. Open squares: GEMD using a cut and shifted LJ
potential, where the cutoff is at 10 A; solid squares: using a 15 A cutoff instead; small
solid circles at T 440K: liquid densities for a 12 A and 14 A cutoff, respectively;
pluses: result obtained for a 10 A cutoff including long-range corrections. As in
Fig. 1, the error bars indicate standard deviations. This is figure 3 taken from
reference [5].
357
FIG. 3 Mean squared displacements hr2i of the hexane carbon atoms vs. time t in
n-hexane calculated at coexistence in the liquid via GEMD (open symbols) and via
NVT-MD (solid symbols) under corresponding conditions for T 400K (circles) and
T 460K (squares). The lines (solid: GEMD; dashed: NVT-MD) are fits to the linear
range of hr2i. This is figure 6 taken from Reference [5].
IV. EXAMPLES
GEMD has been used to calculate phase diagrams for atomic as well as
molecular systems, such as Lennard-Jones spheres or n-hexane, as shown
in Figs. 1 and 2.
Being a molecular dynamics technique, GEMD provides dynamical
information such as diffusion coefficients in the coexisting phases [5]. Shown
in Fig. 3 are the mean squared displacements of the carbon atoms in
n-hexane, calculated in coexisting phases with GEMD, compared to the
results of the constant temperature molecular dynamics at the same
conditions. This shows that diffusion coefficients and other dynamical
information can be extracted from GEMD simulations, together with the
thermodynamic properties.
The Gibbs Ensemble Monte Carlo technique transfers particles between
the boxes in one stepas a whole. To be accepted, this insertion requires
a reasonable probability to find an empty space. In the dense phases this
is a problem. GEMD transfers particles continuously. Transferred atoms
gradually appear and disappear from the simulation box. This may be
a great advantage when modeling liquidliquid or liquidsolid phase
equilibria. More examples of GEMD applications and technical details can
be found in the original papers describing applications of GEMD to the
study of n-alkanes [5] and sorption in zeolites [14].
REFERENCES
1. Panagiotopoulos, A.Z. Direct Determination of Fluid Phase Equilibria by
Simulating the Gibbs Ensemble: A Review. Mol. Sim. 1992, 9, 123.
358
11
Modeling Polymer Crystals
GREGORY C. RUTLEDGE Massachusetts Institute of Technology,
Cambridge, Massachusetts, U.S.A.
I.
INTRODUCTION
360
Rutledge
Lt h, k, l
361
362
Rutledge
Rather than having a situation where a unit cell consists of one or more
molecules, as in small molecule crystals, the situation is reversed: a single
molecule participates in many unit cells. This has significant ramifications
for the correlations, both structural and dynamical, between unit cells.
Second, the periodicity of the crystal lattice usually implies a periodicity
for the conformation of the polymer chain itself. This condition places
a severe restriction on the conformation space available to the chain in
a crystalline solid. Polymer chains in crystals are more appropriately
described as helices rather than coils. A helix conformation consists
of a regular repetition of the torsion angles characteristic of a small subsection of the chain, the helix repeat unit [4,5]. Each helix has an axis and a
handedness of rotation (the direction in which the chain backbone winds
around the helix axis). The helix itself is characterized by the translational
displacement parallel to the helix axis, d, and the angle of rotation about the
helix axis, , executed by a single helix repeat unit. The coordinates of Eq. (1)
may be expressed in terms of these helical parameters as follows:
qh, Lh , n Rh, n hsh hLh Tn
where
2
cosn !h
6
Rh, n 4 sinn !h
0
Lh h0 , k0 , 0
sinn !h 0
7
05
cosn !h
0
and Tn 0, 0, nd
1
t
Here, the subscript h indexes different helices in the unit cell. In this convention, the fractional coordinates for each unique helix in the unit cell are
specified with the helix axis directed along the z-coordinate and passing
through the origin. Each helix may have a distinct setting angle, !h, and
h0 and k0 may take fractional values to indicate helices at nonprimitive lattice
positions. The lattice index l is replaced by the helix repeat index, n.
Third, nearly all known polymer crystal polymorphs exhibit chain helices
packed into the solid with axes parallel.* As a result, there exists a unique
chain continuation direction, which gives rise to a highly anisotropic
*The phase of isotactic polypropylene presents a notable exception. In this crystal form, the
polymer chains all have the same conformation, but are arranged in layers with an angle of
about 80 between the axes of chains in successive layers [6].
363
material on the molecular scale. In the chain direction the crystal is held
together by chemical bonds, while in the other two directions it is held
together by weaker, chain-packing forces, such as van der Waals forces or
hydrogen bonds. This combination of conformational periodicity, throughcrystal chemical bonding and directionality of strong and weak interactions
makes polymer crystals unique.
To specify a starting configuration for any polymer simulation, a reasonably good estimate of the unit cell geometry, and in many cases restrictions
on the atomic coordinates s, may be obtained by wide angle X-ray diffraction. Estimates of the crystal size can be obtained by wide or small angle
X-ray scattering, electron or optical microscopy. One of the unique features
of crystalline polymers is the chain folded lamellar morphology, common
among flexible polymers. These plate-like crystallites are the result of kinetic
factors that dominate during crystallization itself. The thicknesses of such
lamellae are also considerably less than the contour length of the chain
molecules. The ramification of this length mismatch is that the surfaces of
such lamellae involve reentry of the same molecule into the crystal, in the
form of loops or folds. The thinness of the lamellae and the molecular
connectivity of the loop or fold surfaces may have a significant effect on
the dynamical behavior of real polymer crystals. Detailed atomistic studies
of entire fold surfaces in polymer crystals are relatively few [7,8].
More recently, atomistic simulations of the crystal/amorphous interface of
a semicrystalline polymer matrix have been presented that provide insight
into the preferred topology for chain reentry along several relevant crystallographic facets [911]. The presence of disordered surfaces or interfaces
may affect the relative stability of finite-sized crystallites. Nevertheless, the
lowest energy forms of large polymer crystals are expected to be those with
fully extended chains (i.e., no fold or loop surfaces), and it is these that have
generally been the object of simulation.
Finally, a few comments about the uniqueness of polymer crystal structures and phase space localization are warranted. Almost all crystallizable
polymers exhibit polymorphism, the ability to form different crystal structures as a result of changes in thermodynamic conditions (e.g., temperature or pressure) or process history (e.g., crystallization conditions) [12].
Two or more polymorphs of a given polymer result when their crystal
structures are nearly iso-energetic, such that small changes in thermodynamic conditions or kinetic factors cause one or another, or both, to
form. Polymorphism may arise as a result of competitive conformations of
the chain, as in the case of syndiotactic polystyrene, or as a result of competitive packing modes of molecules with similar conformations, as in the
case of isotactic polypropylene. In some instances, the conformational
change may be quite subtle; isotactic polybutene, for example, exhibits
364
Rutledge
A. Optimization Methods
Consider an ideal case where the crystal exhibits none of the forms of disorder mentioned in the previous section. A given polymorph is then characterized by localization about an equilibrium point in phase space where
the potential energy surface is concave. The relevant part of the energy
365
U0
i!
i0
3b
366
Rutledge
displacement rules out any explicit thermal (e.g., entropy) or kinetic contributions to the structure and properties of the crystal. However, this does
not imply that the calculation is necessarily representative of a crystal at 0K.
Thermal contributions may creep into the calculation implicitly, through the
parameterization of the force field or the choice of lattice parameters h.
The relatively simple molecular mechanics calculation has proven very
useful in the study of polymer crystals. It is generally the method of first
resort for any new crystal study. The primary result of such a calculation
is structural information on a very detailed level, for purposes of gaining
insight into the balance of forces or obtaining structure factors to compare
with experiments. For very large problems, the calculations may be speeded
up through the use of one or more common simplifications invoked to
reduce the length of the vector q. First, the problem of determining the
helical conformation of the chain may be decoupled from that of the intermolecular packing structure. The chain conformation having the lowest
intramolecular energy is determined first using isolated chain calculations,
appropriately constrained to ensure helicity. (This calculation can usually be
accomplished simply by setting the lattice vectors a and b to be very large.)
Subsequently, the energy of the crystal is minimized with respect to the
intermolecular degrees of freedom and with the chains themselves treated
as rigid molecules having only three translational and one rotational degree
of freedom per chain [1517]. However, such rigid chain calculations do not
reflect the effects of strong intermolecular packing interactions on conformation. Packing interactions have been shown in some cases to distort
the conformation of the chain [18,19] or even to stabilize a crystal structure containing a completely different and less stable conformation [20].
Second, the simultaneous optimization of intramolecular and intermolecular
degrees of freedom may be retained at minimal cost by taking advantage of
known symmetries within the unit cell, usually deduced from X-ray results
[21,22]. In crystallography, the smallest group of atoms that contains no
internal symmetry is called the asymmetric unit and, in many instances,
comprises a fraction of the total number of atoms in the unit cell. In this
way, the number of degrees of freedom is reduced to the coordinates of the
asymmetric unit, the lattice vectors of the unit cell, and the symmetry operations defining the coordinates of the remaining atoms within the same
unit cell.
Thermal contributions to the properties of the crystal are introduced
by taking higher order terms in the Taylor series expansion [Eq. (3)] for
the energy surface in the vicinity of the equilibrium structure. The classical
theory of lattice dynamics was originally formulated for crystals by Born
and von Karman [23,24] and later presented in the comprehensive text of
Born and Huang [25]. There are several excellent texts that discuss lattice
367
dynamics in great detail [2528]; here, we will only summarize the main
points.
In the harmonic approximation, all contributions to the energy beyond
second order in atomic displacements are neglected. Since the gradient rU0
evaluated at equilibrium is zero, this yields:
1
Uq Uq Uq0 qt rrU0 q
2
3N X
3N 2
X
1
@ U
qi qj
2 i1 j1 @q@q
Differentiating both sides of Eq. (4) with respect to qi, one obtains an
equation for the force @U/@qi acting on coordinate qi, and hence the
equations of motion describing oscillatory displacement of the atoms
about their equilibrium positions:
mi
3N
X
d 2 qi t
@2 U
qj t
dt2
@qi @qj 0
j1
368
Rutledge
where the vector u0 consists of the 3Nu mass-weighted elements of one unit
p
cell, u0i ui mi , and D(k) is the 3Nu 3Nu mass-weighted dynamical
matrix, whose elements are defined as:
P @2 U
exp ik h sj si L
L @qi @qj 0
Dij k
p
mi mj
10
,(k) is the diagonal matrix whose elements are the 3Nu values !i2(k). D(k) is
Hermitian, i.e., D (D*)t, so values of !2i are always real and positive for
solution about a minimum in energy. It is apparent that the set of frequencies !i(k) is a function of the wavevector, k. However, the periodicity of the
lattice reduces the set of unique wavevectors to those located in the first
Brillouin Zone (BZ), the region of reciprocal space whose boundaries lie
halfway between reciprocal lattice points, eg., /a<k /a along the a*
dimension, and similarly for b* and c*; a* (b c)/V, b* (c a)/V and
c* (a b)/V are the reciprocal lattice vectors, while V (a b c) is the
unit cell volume. The k-dependence of the 3Nu values !i(k) is called a dispersion curve, of which there are 3Nu branches in the BZ. Of these 3Nu
branches, three are acoustic modes, which have the property that !i
approaches zero as the magnitude of k approaches zero; the other 3Nu 3
branches are called optic modes. Furthermore, most polymer crystal
structures have more than one chain participating in the unit cell. Each
chain has four molecular degrees of freedom (three translational, and one
.
369
rotational about the chain axis), giving rise to 4Nchain 3 molecular optic
modes, which tend to be lower in frequency than the remaining
3Nu 4Nchain atomic optic modes. For the finite but unbounded crystal
having hmax unit cells along a, kmax unit cells along b, and lmax unit cells along
c, and periodic boundary conditions, the wavevectors along each branch are
restricted to discrete values: k 2ha
=hmax kb
=kmax lc
=lmax ,
where h takes integer values running from hmax/2 to hmax/2, and similarly
for k and l, for a total of 3N vibration frequencies in the finite crystal. In the
limit that the crystal becomes infinitely large, the interval between allowed
wavevectors goes to zero, and both !i(k) and i(k) become continuous functions of k.
Each of the normal modes in the harmonic approximation is a traveling
wave in the crystal and represents a packet of energy; by analogy with the
wave/photon duality of electromagnetic radiation, the wave packets in crystals are called phonons. The energy of each phonon can be quantified either
classically or quantum mechanically. Since the normal modes are independent, the vibrational partition function for the crystal is simply the product
of the individual partition functions for each phonon. These are known
analytically for harmonic oscillators. The corresponding classical and
quantum mechanical partition functions for the crystal are:
classical :
quantum :
ln Qvib
cm
ln Qvib
qm
Z
dk
BZ
3Nu
X
h!
i k
ln
kB T
i1
11a
#
3Nu
X
h!
i k
h!i k
ln 1 exp
dk
2kB T
kB T
BZ
i1
Z
"
11b
RThe free energy is thus computed explicitly (Nk is the normalization factor
BZ dk):
kB T
ln Qvib
Ah, T U q0 Avib q0 , T U q0
Nk
12
370
Rutledge
371
FIG. 2 4Nchain lattice (acoustic and external optic) normal mode branches for
polyethylene along the [100] lattice direction. Each branch is symmetric about the
(k 0) point. 0K (dashed lines) and 300K (solid lines).
simultaneously [37]. This can be important for quantities that depend sensitively on internal strains. To date, only properties for polyethylenelike crystals described by relatively simple potentials have been obtained
by this method and evaluating the analytical derivatives at the static
geometry [38].
Although we have not discussed the specific form of the potential energy
function used to describe U(q), other than to suggest that it is an effective
potential described parametrically by the positions of atomic nuclei, it is
common to treat electrostatic forces in atomistic calculations using partial
atomic charges assigned to the various nuclei. However, this description for
electrostatics does not account for distortion of the electron distribution (the
electronic part of the BornOppenheimer separation) in the environment
of the crystal. The presence of dipoles in the anisotropic crystal can lead
to significant mutual induction, or polarization, effects. The lattice dynamical description presented up to this point can be easily extended to
include mutual induction through use of the shell model for atomic
polarizability [39]. This model introduces an additional 3N coordinates to
q, corresponding to the center of action of the electron distribution about
each atom, which may polarize the atom by situating off-center from the
nucleus itself. This results in an additional 3N normal modes, called plasmon modes, which are considerably higher in frequency than phonon
modes. The shell model has been used to study poly(vinylidene fluoride)
372
Rutledge
13
For mechanical work, p is the applied stress and e is the deformation strain,
defined with respect to a reference cell h0. For a cell deformed from h0 to h,
the (external) strain may be computed as:
e
1 t, 1 1
h Gh0 I
2 0
14
G is the metric tensor hth and V0 is the cell volume. A wide range of thermal
and mechanical properties may be estimated through optimization of
A(h,T ) or G(p,T ) at different temperatures and states of deformation.
Among these, the constant volume heat capacity may be computed using:
CV
kB
Nk
Z
dk
BZ
3Na
X
x2 ex ex 1 2
i1
where
h!
i k
kB T
15
1 t1 dG 1
h
a h0
2
dT 0
16
@r
@e
i:e:,
Cij
@i
@"j
17a
and
S C1
@e
@r
17b
373
or
C
1 @2 Ah, T
V0 @e@e
and
Cij
1 @2 Ah, T
V0 @"i @"j
17c
Here, p and e are the 6 1 vectors for stress and strain, respectively,
in Voigt notation (i.e., "i "ii for i 1,2,3; "4 2"23, "5 2"13, "6 2"12).
Lattice Gruneisen coefficients offer a measure of the thermal stress on a
crystal:
c
V
Ca
CV
18
374
Rutledge
PEa
iPPb
PVDFc
PPTAd
PETe
0.0196
28.3
10.7
2.4
1.2
0.97
2.3
6.0
7.3
280
2.7
1.7
2.9
3.4
8.7
9.5
0.00703
8.2
8.7
0.84
0.68
0.79
0.39
9.9
10.8
55.1
4.0
4.9
1.7
3.4
7.1
4.4
0.0306
6.56
5.13
0.38
0.79
0.66
0.040
24.5
26.1
276
na
na
na
1.2
7.6
9.1
0.1145
7.9
2.9
0.57
0.76
1.07
0.24
11.9
32.2
284
14.3
0.5
6.1
11.1
10.5
14.2
0.0470
11.4
4.12
1.07
1.13
1.02
0.69
14.4
17.3
178
6.6
1.4
1.2
6.4
3.4
9.5
375
FIG. 3 C33 axial modulus of polyethylene obtained using different force fields and
computational methods. All calculations are classical. Open symbols are from
optimizations (circles, triangles, and diamond are self-consistent QHA lattice
dynamics, all others are plotted at temperature of experimental lattice parameters;
potential energy minimizations are plotted at 0K). Filled symbols are Monte Carlo
(circles) or molecular dynamics (diamond, triangles). Force field of Ref. 55: open and
filled circles (from Ref. 46); force field of Ref. 30: open (from Ref. 33) and filled
(from Ref. 68) triangles; pcff force field of Biosym/MSI: open and filled diamonds
(from Ref. 66); (from Ref. 42); (from Ref. 43); open inverted triangle (from Ref.
44); open square (from Ref. 29). Curves are drawn as guides to the eye.
fi STi S exp 2iSt qi
19
376
Rutledge
X
exp 2iSt hL
fj STj S exp 2iSt hsj
20
For large, perfect crystals, the first summation contributes a constant prefactor Fc (hmaxkmaxlmax)2 and the scattering vector interferes constructively
(i.e., is nonzero) only for values corresponding to vertices of the reciprocal
lattice: S ha* kb* lc* h*L. The static structure factor then becomes
simply:
FS Fc
fj STj S exp 2iLt sj
21
1
Nk Nu mi
Z
dk
BZ
3Nu
X
Evib, j k
j1
!2j k
t
i, j k i,
j k
22
377
treated as isotropic and equivalent for all atoms; in this case the corresponding B-factor is simply 82(tr(Bi)), averaged over all the atoms of the
unit cell.
The principal limitations of lattice dynamics, of course, are the shape of
the potential energy surface implied by the harmonic approximation and
the restriction to localized phase spaces. When polymorphism is an issue,
each polymorph may be examined independently using lattice dynamics.
However, crystals with internal disorder at elevated temperatures include
significant contributions from configurations that are not accurately represented by the quadratic form of the Taylor series expansion for potential
energy. Studies of atomic crystals, as well as a recent study of polyethylene
[34,49], suggest that QHA is reasonably accurate up to about 2/3 of the
melting temperature, barring the presence of solidsolid transformations at
lower temperatures. Higher order anharmonic contributions may be treated
methodically through the use of perturbation theory [5053]. However, the
378
Rutledge
B. Sampling Methods
Both Monte Carlo and molecular dynamics methods sample directly the
phase space of a small but representative component of the crystal, the
former by performing stochastic moves through configuration space,
the latter by following a specified trajectory according to an equation of
motion and chosen initial condition. A typical Hamiltonian for molecular
dynamics simulation is [14]:
Hp, q
3N
X
p2i
Uq
2mi
i1
23
Fobs hFi
Fi eEi
lim Xi1
N
N!1
eEi
i1
1
lim
t!1 t
dt0 Ft0
24
0
@H l
dl
@l0
0
Al2 Al1
1
25
379
we will not go into these methods in as much detail as with lattice dynamics.
Nevertheless, a few comments are in order.
By sampling phase space directly, the methods of Monte Carlo and
molecular dynamics can avoid the restrictive approximations of lattice
dynamics. In principle, at least, polymorphism, anharmonicity, static and
dynamic disorder may all be rigorously captured by direct simulation.
However, other limitations arise which differ from those encountered in a
lattice dynamical approach. Comparison of both approaches can provide
insight into the significance of these limitations [46].
First, both Monte Carlo and molecular dynamics simulations must
approximate the macroscopic crystal using a very small number of atoms,
typically on the order of 103104 with current computers. With periodic
boundary conditions, atoms outside of the simulation box are restricted
to remain dynamically in phase with those inside the simulation box. For
a simulation cell consisting of hmax kmax lmax unit cells, only those wavevectors whose magnitudes are integer multiples of 2a*/hmax, 2b*/kmax,
and 2c*/lmax are sampled in the simulation. Thus, unless hmax, kmax, and
lmax are chosen suitably large, the dynamical resolution is rather poor. This
finite size effect may affect both lattice properties, such as thermal expansion, and atomic displacements, which as noted previously are most affected
by long wavelength acoustic modes. For polyethylene, finite size effects are
noticeable in small simulations containing 2 3 6 unit cells (432 atoms) at
temperatures around 100K, but become progressively worse at higher temperatures [54,55]. Furthermore, finite size effects in polyethylene were found
to be quite anisotropic, being largest along the b direction of the lattice, but
also strongly coupled to the lattice size in the chain direction. Free or rigid
boundary conditions [56] may be used instead of periodic boundaries, but
these incur surface as well as finite size effects, the implications of which are
less thoroughly quantified.
The anisotropy of polymer crystals has other implications for simulation,
as well. Along the chain direction, intramolecular degrees of freedom may
dominate the dynamical behavior, while dynamics lateral to the chain are
predominantly intermolecular in nature. As evidenced by the large separation between internal and external mode frequencies in lattice dynamics,
these two groups are largely decoupled. In molecular dynamics, this separation of dynamical time scales manifests itself as an inefficient transport of
energy between high frequency and low frequency motions. This may be
observed as poor equilibration of kinetic energy between intramolecular and
intermolecular degrees of freedom. To overcome this problem, Ryckaert
and Klein used massive stochastic collisions [57] in their molecular dynamics
simulation [58]. The collisions serve to redistribute the kinetic energy among
all the modes of motion. In Monte Carlo simulations, this effect is observed
380
Rutledge
E 2 h E i2
CV
kB T
26
381
where it is understood that the energy E includes both kinetic and potential
contributions. A similar useful relation exists for the elastic stiffness moduli
(in the NT ensemble) [65]:
C
kB T
t 1
""
hVi
or
Cij
kB T
1
"i "j
hVi
27
As indicated earlier, accurate calculation of the elastic moduli by this formula requires a simulation cell large enough to sample adequately the lattice
modes of vibration. For polyethylene and polypropylene, simulations consisting of 4 6 12 unit cells and 3 1 3 unit cells, respectively (corresponding to simulation cells approximately 20 A on a side in each case) have
been used with reasonable success [66]. Whereas simulations on the order of
104 to 105 Monte Carlo or molecular dynamics steps are typically required
to obtain good averages for energy, density, and many other properties,
fluctuation averages such as those used in Eqs. (27) typically require simulations one to two orders of magnitude longer in order to achieve accuracies
around 5% [67,68]. This has been attributed to slow convergence of high
frequency motions [68], as well as the difficulty in sampling external variables mentioned above. A formula that takes advantage of the correlated
fluctuations of stress and strain has been proposed to improve the convergence characteristics [67].
382
Rutledge
383
point in the potential energy surface, defined as a stationary point with one
and only one negative eigenvalue of the dynamical matrix. The problem
then centers on the characterization of the reactant and saddle point
(i.e., transition state) configurations. For simplicity, the present discussion
adheres to this conventional interpretation.
In conventional TST, the one-way flux, or reaction rate, from reactant to
product may be written as:
kTST
z
kB T Qvib
E
exp
h Qvib, 0
kB T
28
384
Rutledge
with the line defects of the latter; similarly, rotational mismatches are
termed disclinations. In polymers, the helical symmetry of the chain permits a combined translational/rotational mismatch that has been termed a
dispiration [70]. The identification of the structure of defects which satisfy
the constraints of a dislocation, disclination, or dispiration typically requires
considerable insight, inspiration, or brute force searching of conformation
space [7173]. For highly localized conformational defects, the process of
defect identification can be simplified by solving the conformational problem in torsion space alone, for defects up to six bonds long [74,75]. The
method involves solving a set of nonlinear equations, and also serves as
the basis of the Concerted Rotation Monte Carlo algorithm [76]. Once
the conformational problem is solved, one can construct a cell containing
a single reference chain containing the conformational defect, and thereby
compute the energy and distortion of the defect due to packing in the lattice.
Unlike perfect crystal simulations, simulations to date involving defects
have been performed almost exclusively using finite cells with rigid or free
boundary conditions so as to avoid defectdefect interactions.
Having once identified a stable defect in the crystal lattice and computed
its energy of formation, determining the kinetics of a process based on this
defect in TST requires the further identification of a saddle point along a
trajectory which displaces the defect by one lattice period; the periodicity of
the lattice permits a scenario where subsequent propagation of the defect
occurs through a sequence of similar hops in the lattice. Identification of
a suitable trajectory for defect propagation in polyethylene was performed
by introducing the defect into the lattice, as described above, and then
driving one or more coordinates (e.g., a torsion angle in the chain or setting
angle in the crystal) so that the defect is forced to move to the next unit cell
in the lattice. This type of calculation is essentially quasi-static; a constraint
on one or more rotation angles in the chain is incremented, and the potential
energy of the defect reminimized [73,77]. The process is repeated stepwise
until a potential energy barrier is crossed (i.e., the defect falls into the
product state upon minimization). However, this approach does not generally locate a first order saddle point on the potential energy surface, but
rather a maximum along the trajectory chosen by the coordinate-incrementing protocol employed. The resulting trajectories are thus sensitive to how
the defect is driven along the chain and provide upper bounds on the true
saddle point energy [77]. Alternative numerical approaches can be used to
refine the estimate of a true first order saddle point based on an adiabatic
reaction path connecting reactant to product. One particularly useful algorithm for this purpose is the conjugate peak refinement algorithm of Fischer
and Karplus [78], which requires knowledge only of the equilibrium conformation of the defect in order to identify a lowest energy path from
385
reactant to product. This saddle point search algorithm was used in a study
of poly(vinylidene fluoride) [79] and, more recently, polyethylene [80]. For
cases where a good initial guess of the saddle point configuration is available, more efficient gradient algorithms can be used to refine the estimate of
the saddle point configuration and energy [81]. Once the saddle point configuration is identified, Uz and Qz are readily computed using the methods
of lattice dynamics, and TST applied. Figure 5 illustrates two trajectories
computed for propagation of defects in poly(vinylidene fluoride) to account
for the -relaxation in the -crystalline form of that polymer. It is interesting
to note that, whereas heats of formation for defects may be fairly large, the
methods described above can be used to characterize the trajectory of the
defect through the lattice, regardless of whether it involves a hopping
mechanism between adjacent lattice sites, separated by large energy barriers,
or a relatively facile free-streaming motion involving only low energy barriers to propagation. Both hopping [70] and free-streaming [73] mechanisms
have also been proposed for motion of defects in polyethylene.
V. SUMMARY
There is currently a variety of methods for studying polymer crystals by
molecular simulation. As with other materials, the usual caveats regarding force field accuracy and proper execution of the simulation apply.
386
Rutledge
REFERENCES
1. Cullity, B.D. Elements of X-Ray Diraction, 2nd Ed.; Addison-Wesley:
Reading, 1978.
2. Alexander, L.E. X-Ray Diraction Methods in Polymer Science; Wiley
Interscience: New York, 1969.
3. Guinier, A. X-Ray Diraction in Crystals, Imperfect Crystals, and Amorphous
Bodies; W.H. Freeman and Co.: San Francisco, 1963.
4. Miyazawa, T. J. Polym. Sci. 1961, 55, 215.
5. Sugeta, H.; Miyazawa, T. Biopolymers 1965, 5, 673.
6. Meille, S.V.; Bruckner, S.; Porzio, W. Macromolecules 1990, 23, 41144121.
7. Dave, R.S.; Farmer, B.L. Polymer 1988, 29, 15441554.
8. Fisher, H.P.; Eby, R.K.; Cammarata, R.C. Polymer 1994, 35, 19231930.
9. Balijepalli, S.; Rutledge, G.C. J. Chem. Phys. 1998, 109, 65236526.
10. Balijepalli, S.; Rutledge, G.C. Comput. Theor. Polym. Sci. 2000, 10 (1,2),
103113.
11. Gautam, S.; Balijepalli, S.; Rutledge, G.C. Macromolecules 2000, 33 (24),
91369145.
12. Corradini, P.; Guerra, G. Advances in Polymer Science 1991, 100, 185.
13. Press, W.H.; Teukolsky, S.A.; Vetterling, W.T.; Flannery, B.P. Numerical
Recipes in C: The Art of Scientic Computing, 2nd Ed.; Cambridge
University Press: Cambridge, 1992.
14. Allen, M.P.; Tildesley, D.J. Computer Simulation of Liquids; Clarendon Press:
Oxford, 1987.
15. McCullough, R.L. J. Macromol. Sci. Phys. 1974, 9, 97.
16. Tripathy, S.K.; Hopnger, A.J.; Taylor, P.L. J. Phys. Chem. 1981, 85,
13711380.
17. Bacon, D.J.; Geary, N.A. J. Mater. Sci. 1983, 18, 853863.
18. Rutledge, G.C.; Suter, U.W. Macromolecules 1991, 24, 19211933.
19. Lacks, D.J.; Rutledge, G.C. Macromolecules 1994, 27, 71977204.
20. Zhang, R.; Taylor, P.L. J. Chem. Phys. 1991, 94, 32073212.
387
21. Tadokoro, H. Structure of Crystalline Polymers; Wiley and Sons: New York,
1979.
22. Tashiro, K. Prog. Polym. Sci. 1993, 18, 377435 and references therein.
23. Born, M.; von Karman, T. Phys. Zeit. 1912, 13, 297.
24. Born, M.; von Karman, T. Phys. Zeit. 1913, 14, 15.
25. Born, M.; Huang, K. Dynamical Theory of Crystal Lattices; Oxford University
Press: Oxford, 1954.
26. Willis, B.T.M.; Pryor, A.W. Thermal Vibrations in Crystallography; Cambridge
University Press: Cambridge, 1975.
27. Horton, G.K.; Maradudin, A.A. Eds. Dynamical Properties of Solids; NorthHolland: Amsterdam, 19741990; Vol. 17.
28. Dove, M.T. Introduction to Lattice Dynamics; Cambridge University Press:
Cambridge, 1993.
29. Sorensen, R.A.; Liau, W.B.; Kesner, L.; Boyd, R.H. Macromolecules 1988, 21,
200208.
30. Karasawa, N.; Dasgupta, S.; Goddard, W.A. J. Phys. Chem. 1991, 95,
22602272.
31. Boyd, R.H. In Advances in Polymer Science, Monnerie, L., Suter, U.W., Eds.;
Springer Verlag: Berlin, 1994; Vol. 116, 125.
32. Allan, N.L.; Barron, T.H.K.; Bruno, J.A.O. J. Chem. Phys. 1996, 105,
83008303.
33. Lacks, D.J.; Rutledge, G.C. J. Phys. Chem. 1994, 98, 12221231.
34. Lacks, D.J.; Rutledge, G.C. Macromolecules 1995, 28, 11151120.
35. Carbeck, J.D.; Lacks, D.J.; Rutledge, G.C. J. Chem. Phys. 1995, 103,
1034710355.
36. Rutledge, G.C. Macromolecules 1997, 30, 27852791.
37. Taylor, M.B.; Barrera, G.D.; Allan, N.L.; Barron, T.H.K. Phys. Rev. B 1997,
56 (22), 1438014390.
38. Bruno, J.A.O.; Allan, N.L.; Barron, T.H.K.; Turner, A.D. Phys. Rev. B 1998,
58 (13), 84168427.
39. Dick, B.; Overhauser, A. Phys. Rev. 1958, 112, 90.
40. Karasawa, N.; Goddard, W.A. Macromolecules 1992, 25, 72687281.
41. Carbeck, J.D.; Rutledge, G.C. Polymer 1996, 37, 50895097.
42. Anand, J.N. J. Macromol. Sci.-Phys. B1 1967, (3), 445.
43. Wobser, V.G.; Blasenbrey, S. Kolloid. Z. Z. Polym. 1970, 241. 985.
44. Tashiro, K.; Kobayashi, M.; Tadokoro, H. Macromolecules 1978, 11 (5), 914.
45. Martonak, R.; Paul, W.; Binder, K. Phys. Rev. E 1998, 57, 24252437.
46. Rutledge, G.C.; Lacks, D.J.; Martonak, R.; Binder, K. J. Chem. Phys. 1998,
108, 1027410280.
47. Tashiro, K.; Kobayashi, M.; Tadokoro, H.; Fukada, E. Macromolecules 1980,
13, 691.
48. Burnett, M.N.; Johnson, C.K. Ortep-III: Oak Ridge Thermal Ellipsoid Plot
Program for Cyrstal Stucture Illustration. Oak Ridge National Laboratory
Report ORNL-6895, 1996.
49. Lacks, D.J.; Rutledge, G.C. J. Chem. Phys. 1994, 101, 99619965.
388
50.
51.
52.
53.
54.
55.
56.
57.
58.
59.
60.
61.
62.
63.
64.
65.
66.
67.
68.
69.
70.
71.
72.
73.
74.
75.
76.
77.
78.
79.
80.
81.
Rutledge
Shukla, R.C.; Cowley, E.R. Phys. Rev. B 1985, 31, 372378.
Shukla, R.C.; Shanes, F. Phys. Rev. B 1985, 32, 25132521.
Shukla, R.C.; Bose, S.K.; Delogu, R.F. Phys. Rev. B 1992, 45, 1281212820.
Shukla, R.C.; Lacks, D.J. J. Chem. Phys. 1997, 107, 74097417.
Martonak, R.; Paul, W.; Binder, K. Comput. Phys. Commun. 1996, 99, 28.
Martonak, R.; Paul, W.; Binder, K. J. Chem. Phys. 1997, 106, 89188930.
Liang, G.L.; Noid, D.W.; Sumpter, B.G.; Wunderlich, B. Comput. Polym. Sci.
1993, 3, 101.
Andersen, H.C. J. Chem. Phys. 1980, 72, 2384.
Ryckaert, J.-P.; Klein, M.L. J. Chem. Phys. 1986, 85, 16131620.
Yamamoto, T. J. Chem. Phys. 1985, 82, 37903794.
Yamamoto, T. J. Chem. Phys. 1988, 89, 23562365.
Foulger, S.H.; Rutledge, G.C. Macromolecules 1995, 28, 70757084.
Foulger, S.H.; Rutledge, G.C. J. Polym. Sci.-Phys. 1998, 36, 727741.
Sumpter, B.G.; Noid, D.W.; Liang, G.L.; Wunderlich, B. In Advances in
Polymer Science; Monnerie, L., Suter, U.W., Eds.; Springer Verlag: Berlin,
1994; Vol. 116, 2772 and references therein.
Van Kampen, N.G. Stochastic Processes in Physics and Chemistry; NorthHolland: Amsterdam, 1992.
Parrinello, M.; Rahman, A. J. Chem. Phys. 1982, 76 (5), 26622666.
Zehnder, M.M.; Gusev, A.A.; Suter, U.W. Rev. Inst. Francais Petrole 1996,
51 (1), 131137.
Gusev, A.; Zehnder, M.M.; Suter, U.W. Phys. Rev. B. 1996, 54, 14.
Cagin, T.; Karasawa, N.; Dasgupta, S.; Goddard, W.A. Mat. Res. Soc. Symp.
Proc. 1992, 278, 61.
Glasstone, S.; Laidler, K.; Eyring, H. The Theory of Rate Processes; McGrawHill: New York, 1941.
Reneker, D.H. J. Mazur. Polymer 1988, 29, 313.
Reneker, D.H. J. Polym. Sci. 1962, 59 (168), S39.
Boyd, R.H. J. Polym. Sci.-Phys. 1975, 13, 2345.
Manseld, M.; Boyd, R.H. J. Polym. Sci.-Phys. 1978, 16, 1227.
McMahon, P.E.; McCullough, R.L.; Schlegel, A.A. J. Appl. Phys. 1967, 38 (11),
4123.
Go, N.; Scheraga, H.A. Macromolecules 1970, 3, 178.
Dodd, L.R.; Boone, T.D.; Theodorou, D.N. Mol. Phys. 1993, 78, 961996.
Reneker, D.H.; Fanconi, B.M.; Mazur, J. J. Appl. Phys. 1977, 48 (10), 4032.
Fischer, S.; Karplus, M. Chem. Phys. Lett. 1992, 194, 252261.
Carbeck, J.D.; Rutledge, G.C. Macromolecules 1996, 29 (15), 51905199.
Mowry, S.W.; Rutledge, G.C. Macromolecules 2002, 35 (11), 45394549.
Schlegel, H.B. In Ab Initio Methods in Quantum Chemistry; Lawley, K.P, Ed.;
Wiley and Sons: New York, 1987; 249 pp.
12
Plastic Deformation of
Bisphenol-A-Polycarbonate: Applying
an Atomistic-Continuum Model
JAE SHICK YANG and WON HO JO
Korea
I.
Eidgenossische Technische
INTRODUCTION
390
Yang et al.
II. MODEL
We briefly recapitulate the model, detailed previously [11], which is based on
the variable-metric total-energy approach by Gusev [12] and consists of a
system comprising an inclusion embedded in a continuous medium. The
inclusion behavior is described in atomistic detail whereas the continuum is
modeled by a displacement-based finite-element method. Since the atomistic
391
A. Continuum Model
The variable-metric total-energy approach [12] is adopted here for the
matrix. As illustrated in Fig. 1, the periodic system is described by the
scaling matrix [13,14] H [ABC], where A, B, and C are the overall systems
continuation vectors. Two kinds of nodal points are specified in the matrix:
One (xb) on the inclusion boundary and the other (xc) in the continuum
(throughout this text, vectors are written as column matrices). For convenience, the scaled coordinates [13,14] (sb and sc) are chosen as degrees of
freedom via xb Hsb and xc Hsc. These nodal points are used as vertices
of a periodic network of Delaunay tetrahedra. The Delaunay network
uniquely tessellates space without overlaps or fissures; the circumsphere of
the four nodal points of any tetrahedron does not contain another nodal
point of the system [15,16]. For each tetrahedron , the local scaling matrix
h [abc] is defined in the same manner as H for the overall system. The
local (Lagrange) strain e is assumed to be constant inside each tetrahedron
and defined as [13,14]
e
i
1 h T
G G I ,
2
1
G h h0
392
Yang et al.
where the subscript 0 represents the reference state before deformation and I
is the unit matrix of order 3. The system strain esys is calculated from Eq. (1)
by substituting H and H0 for h and h0 , respectively. The strain energy in the
continuum Econt can be evaluated from the sum over all tetrahedra:
1 X T
V e C e
E cont H, sb , sc
2
where V is the volume of the reference state and C the matrix of elastic
constants of tetrahedron .
B. Atomistic Model
A spatially periodic simulation box represents the contents of the atomistic
inclusion. Again, the shape of this box is expressed by a scaling matrix
h [abc] (cf. Fig. 1). The scaled coordinates of atom sa are used as degrees of
freedom via xa hsa. The Lagrange strain tensor of the atomistic box, eatom,
can be calculated from Eq. (1) by replacing h and h0 with h and h0,
respectively. The energy in the atomistic box can be expressed as Eatom
(h,{sa}) and can be obtained via any energy model for atomistic simulation,
e.g., a force-field approach.
C. Atomistic-Continuum Model
To connect the two models with different length scales, the displacement
of the nodal points on the inclusion boundary is associated with the
b
deformation of the atomistic box by xb hh1
0 x0 . This means that the
inclusion boundary is constrained to follow the homogeneous deformation
of the atomistic cell vectors. The degree of freedom DH is introduced to
relate h with H through
h HI DH H1
0 h0
There are four types of degrees of freedom in the system: H and sc for the
continuum model; h and sa for the atomistic model.
393
V inc
V atom
where Vinc and Vatom are the volumes of the inclusion and atomistic box,
respectively (evidently, the inclusion must have the same energy density as
the atomistic box). Note that V is kept constant during deformation
because xb deforms according to h.
The length scales in the model are given by the inclusion volume fraction
fI and V, e.g., in a cubic system box of edge length L with a spherical
inclusion of diameter d, defined by an atomistic box of edge length a,
d (6V/)1/3a and L (V/fI)1/3a.
While this two-scale model allows for a natural connection between
the two scales (i.e., the inclusion-boundary behavior and the atomistic cell
behavior are connected via the common strain transformation), it also is
limited by the required uniformity of strain in the atomistic inclusion
(through the periodic continuation conditions on the atomistic cell) and in
each of the tetrahedra of the Delaunay tessellation.
394
FIG. 2
Yang et al.
where the first four terms are valence terms (bond length, bond angle,
dihedral angle, and out-of-plane angle), the next two terms nonbonded
interaction terms (electrostatic and van der Waals), and the last eight
terms coupling valence terms. For efficiency, an in-house code was constructed that yields the same energy and partial derivatives with respect to
coordinates and components of the scaling matrix as the commercial program Discover [19]. The nonbonded energy cutoff and the spline width were
8 A and 2 A, respectively. The charged-group based method [19] was
adopted for the calculation of nonbonded energies to keep the total
charge inside the cutoff to zero.
0 0 0
B
0 0 0C
C
B
B
2
0 0 0 C
C
8
CB
B 0
0
0
0 0C
C
B
@ 0
0
0
0
0A
0
0
0
0 0
395
where
and
are the so-called Lame constants. The Youngs modulus E,
the shear modulus G, the bulk modulus B, and the Poissons ratio are
obtained by
3
2
,
2
B
,
3
B
esys B
@0
0
0
2
B
B
@0
esys
1 q
C
C 103 for uniaxial extension
A
10
1 q
1
0
0
C
3
2 r 0 C
A 10 for pure shear
0
11
where q and r are small values adjusted to keep the system volume constant.
Thus, for both, uniaxial extension and pure shear, the system volume Vsys
was held constant while the atomistic box was free to change its volume
Vatom [through the coupling degree of freedom DH, Eq. (3)].
The work-equivalent tensile strains in the system esys
eq and in the atomistic
are
given
by
[2]
box eatom
eq
2 sys 2 1=2
1
Tr e^
where e^ sys esys Tresys I
3
3
1=2
2
2
1
Tr e^ atom
where e^ atom eatom Tr eatom I
3
3
esys
eq
12
eatom
eq
13
396
Yang et al.
The internal stress tensors in the system psys and in the atomistic box patom
are calculated by [13,14]
psys
sys T
1
@E
H
,
sys
V
@H
patom
h
atom
atom T
@E
@h
14
sys
The von Mises-equivalent tensile stresses of the system eq
and of the atoatom
mistic box eq are given by [2]
sys
eq
atom
eq
3
Trr^ sys 2
2
1=2
where
3 atom 2 1=2
Tr r^
2
1
r^ sys rsys Trrsys I
3
1
where r^ atom ratom Tr ratom I
3
15
16
Catom
7:46
B 4:63
B
B 4:52
B
B 0:00
B
@ 0:11
0:16
4:63
7:08
4:38
0:02
0:15
0:01
4:52
0:00
4:38 0:02
7:03 0:06
0:06 1:47
0:10 0:01
0:03 0:03
0:11
0:15
0:10
0:01
1:51
0:06
1
0:16
0:01 C
C
0:03 C
CGPa
0:03 C
C
0:06 A
1:55
17
Comparing Eq. (17) with Eq. (8) leads us to accept the BPA-PC models as
isotropic glasses with Lame constants and moduli as listed in Table 1.
Comparison with experimental values [23] indicates that the agreement is
satisfactory.
B. Plastic Deformation
Figure 3 shows stressstrain curves from the simulation runs. Because more
than 80% of the system volume is assumed to behave perfectly elastically, the
system stress increases almost linearly with the system strain [cf. Eq. (5)].
397
Property
Calculated
Experimental [23]
Lame constant
(GPa)
Lame constant
(GPa)
Youngs modulus E (GPa)
Shear modulus G (GPa)
Bulk modulus B (GPa)
Poissons ratio
4.35 0.10
1.43 0.06
3.93 0.14
1.43 0.06
5.30 0.10
0.38 0.01
4.275.55
0.81.1
2.32.5
0.81.1
5.06.1
0.420.43
FIG. 3 Stressstrain curves of the system under uniaxial extension (filled squares)
and pure shear (open circles).
However, the lines are not smooth but contain numerous small discontinuities where the system stress decreases precipitously after a small strain
increment owing to a plastic response of the inclusion; arrows mark the
particularly conspicuous drops. The two curves for uniaxial and pureshear deformations would superpose perfectly if the system behaved totally
elastically; deviations are evident after a sufficient number of plastic unit
events, which are irreversible and dissipative and probably different for
398
Yang et al.
FIG. 4 Stressstrain curves of the atomistic box under uniaxial extension (filled
squares) and pure shear (open circles).
399
sys
FIG. 5 The ratio of the von Mises-equivalent tensile stress of the system, eq
, to
atom
that in the atomistic box, eq , against the system strain under uniaxial extension
(filled squares) and pure shear (open circles).
400
Yang et al.
FIG. 7 The volume of the atomistic box, Vatom/V0atom , as a function of the atomistic
strain, "atom
eq , under uniaxial extension (filled squares) and pure shear (open circles).
about 15% strain (and thereafter becomes greater than the latter). The cause
(cf.
of this is evident when the atomistic box volume is plotted against "atom
eq
Fig. 7, note that the system volume Vsys is kept constant during deformation
under both uniaxial extension and pure shear). The volume of the atomistic
0:07 and then
box Vatom remains at its initial value V0atom until "atom
eq
gradually increases with strain while there are sharp increases at the points
where the sudden drops in the stress are observed. Hence, our simulations
indicate that the plastic unit event is a dilatational process, in which the
transforming domain assumes a reduced density while the surrounding
matrix takes up the concomitant elastic deformation. This is in apparent
contrast with the deductions by Hutnik et al. [3] who found densification
during plastic deformation for BPA-PC (experimental results show that the
plastic deformation of polymer glasses at constant pressure is a nearly
constant-volume process, with a small dilatation for vinyl polymers and a
small volume contraction for BPA-PC) [2428]. Further studies are needed
to gain a satisfactory understanding of these observations.
Because the matrix is modeled as a perfectly elastic medium, the matrix
energy is quadratic with respect to the system strain and, consequently, the
energy of the continuum portion of the model against strain must exhibit a
401
FIG. 8 Energy change in the continuum matrix, iEcont, and in the atomistic
inclusion, ViEatom, as a function of the corresponding work-equivalent tensile
strains, under uniaxial extension (filled squares) and pure shear (open circles).
slope of 2, plotted doubly logarithmically (cf. Fig. 8). The inclusion energy
(VEatom) shows the same tendency at low strain, but deviates as plastic
events begin to soften the material.
The backbone dihedral angles in the constitutional repeat unit can be
categorized into three groups [29]: (i) 1 and 2, which determine the
conformation of the phenylene rings with respect to the isopropylidene
group; (ii) 3 and 6, which specify the conformation of the phenylene rings
with respect to the carbonate group; (iii) 4 and 5, which give the
conformation of the carbonate group. We now analyze the distribution of
the angles in these groups and their changes during deformation.
1 and 2 are strongly interdependent and prefer propeller-like
conformations. Figure 9 contains a plot of all pairs of values of 1 and 2
angles concurrently occurring at one isopropylidene moiety. The intramolecular ground states are located at (45 , 45 ) and (135 , 45 ) [29]. Upon
deformation to "atom
0:256, these distributions do not change perceptibly
eq
402
Yang et al.
FIG. 9 Distribution of the dihedral angles 1 and 2. The large plot contains the
values for the pairs (1, 2) before deformation ("atom
0:0, open circles) and after
eq
deformation under pure shear ("atom
0:256,
filled
circles).
The polar plot indicates
eq
the changes in 1 and 2, where the angle on the perimeter denotes the direction of
the change and the distance from the center denotes the magnitude of the change.
403
(the pairs of angles are marked with open circles for the unstrained and with
filled circles for the deformed state). An analysis of the changes in 1 and 2
during this massive deformation experiment reveals that the conformations
change mostly from one propeller-like state to a symmetry-related one and
that the magnitude of this change often exceeds 180 ; the polar plot included
in Fig. 9 indicates the direction of the change in the (1, 2) plane on
the circumference and its magnitude as the distance from the center. The
orientation of this plot is such that the direction of change conforms to
the large plot in Fig. 9.
Angles 3 and 6 assume values of 45 at conformations of minimal
energy, while 4 and 5 are preferably at 0 (trans) and 180 (cis). Figure 10
FIG. 10 Distribution of the dihedral angles 3 (or 6) and 4 (or 5). The large plot
contains the values for the pairs [3 (or 6), 4 (or 5)] before deformation
0:0, open circles) and after deformation under pure shear ("atom
0:256,
("atom
eq
eq
filled circles). The polar plot indicates the changes in 3 (or 6) and 4 (or 5), where
the angle on the perimeter denotes the direction of the change and the distance from
the center denotes the magnitude of the change.
404
Yang et al.
V. CONCLUSIONS
The elastic constants of BPA-PC, calculated by atomistic modeling, show
relatively good agreement with experimental values. The atomisticcontinuum model is successfully applied to the large deformation behavior
of a BPA-PC system where an inclusion described in an atomistically
detailed manner deforms plastically in an elastic medium. In the atomistic
box, the stressstrain curves show typical microstructural plastic behavior
with sharp drops in the stress. Because the matrix is characterized by elastic
constants of BPA-PC calculated by atomistic modeling, the strain of the
atomistic box is equal to the system strain until a certain point. At large
deformations, the strain in the atomistic box becomes higher than the
system strain.
The volume of the atomistic box initially remains constant because the
volume of the system is kept constant during deformation under both
uniaxial extension and pure shear. At strains greater than "atom
0:07, the
eq
volume of the atomistic box grows gradually with the strain and there are
sharp increases at several points where sudden drops in the stress in the
atomistic box are observed. No considerable change in the distribution of
the backbone dihedral angles is observed during deformation up to the
system strain 0.2.
The emerging picture of plastic unit events in the polymeric glass is that
of a sudden dilatation followed by melt-like flow.
405
REFERENCES
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
27.
28.
29.
30.
31.
13
Polymer Melt Dynamics
WOLFGANG PAUL
Germany
MARK D. EDIGER
Wisconsin, U.S.A.
GRANT D. SMITH
DO Y. YOON
I.
INTRODUCTION
408
Paul et al.
given as
V b constraint
1
V b kb b b0 2
2
1
or V k0 cos cos0 2
2
or
1
V k 0 2
2
6
X
V
An cosn
2
3
4
n0
6 #
r0
V E6 rij expfrij =g
rij
409
where and r0 are the length scales of the repulsive and dispersive parts,
respectively, and a Coulomb term
V q rij
1 qi qj
40 rij
410
Paul et al.
treat the bonded interactions is to use topology files for the 2-body
(bond lengths), 3-body (bond angles), and 4-body (dihedral angles)
interactions. These files describe the chain structure in specifying which
atom is bonded to which others, which three atoms participate in a given
bond angle degree of freedom, and which four atoms participate in a given
dihedral angle degree of freedom. The bonded interactions can then be
calculated by working through these lists. In this way a bead-spring model
and an explicit atom model just differ in the topology files specifying the
dynamic degrees of freedom, and the potential form and parameter set to
use with one specific degree of freedom.
411
way is to estimate the target density and grow the chains into a box of the
corresponding volume without considering the nonbonded interactions.
In the first part of the simulation the repulsive part of the nonbonded
interaction is then regularized to a finite value at distance zero,
which is in turn increased in a stepwise fashion until all overlaps have
disappeared [5].
When we have generated the system at the desired temperature and
pressure/density we still have to ensure that it is equilibrated with respect
to the other observables of the system. In general, equilibration of a
system can be judged by looking at the quantity with the longest
relaxation time, which in an amorphous polymer melt is the end-to-end
distance of the chains. When we are looking at chains which are long
enough to obey random walk statistics for the end-to-end distance, this
equilibration can be quantified by looking at the Gaussianity of the endto-end vector distribution P(R) [6] or at the decay of the end-to-end vector
autocorrelation function hRt R0i. Typically, properties which are
sensitive to shorter distance behavior equilibrate faster than those living
on larger scales. When there is no coupling between the length scales, it
may be sufficient to equilibrate the scale one wants to study. For instance,
when looking at the structural relaxation in a polymer melt as a function
of temperature, it may be sufficient to ensure equilibration on the scale of
one (or a few) statistical segments of the chain, but this can only be
decided a posteriori, if one has independent measurements to compare
with. If there is a coupling to larger scale behavior one has a quenched
structure and static as well as dynamic quantities will depend on the
quenching rate.
412
Paul et al.
FIG. 2 Single chain structure factor for a united atom model of polyethylene
calculated for neutron scattering [7].
Debye function, which is the analytical expression for the structure factor of
a random walk,
2
8
SDeb qRg 4 4 expfq2 R2g g 1 q2 R2g
q Rg
which is the best experimental way to determine the radius of gyration of the
chains. In the simulation it can be obtained in the same way and also by
direct calculation
R2g
K
1X
ri Rcm 2
K i1
where K is the number of atoms per chain and Rcm is the center of mass of
the chain. For larger q the structure factor will be sensitive to the actual
positions of the true scattering centers in the melt, i.e., in this case the
positions of the carbons and deuteriums and their respective coherent scattering lengths. This is also true for the global structure factor of the melt,
which therefore has to be calculated as a superposition of the partial structure factors of all the atom types
Scoh q 1
M X
M
X
coh
q
x x
coh
coh h^
q
1 1
10
413
Here and
label the M different atom types, the x are their number
fractions, the coh are their coherent scattering cross sections, is the density
of scattering centers, and the average coherent scattering cross section is
defined as
coh
M
X
x coh
11
1
12
For the calculation of the melt structure factor the indices i and j in the
above equation run over all atoms in the simulation volume, whereas for the
calculation of the single chain structure factor they only run over the atoms
of that chain. Equation (10) is written in a notation suggestive of neutron
scattering, when we only have nuclear scattering of the neutrons and the
scattering amplitude can be expressed as a scattering length. When we want
to compare a simulated structure with X-ray scattering experiments on that
structure, we have to take into p
account
the wavelength dependence of the
atomic form factor and replace coh by f q.
The above discussed information is, of course, contained in an explicit
atom simulation, and to a certain degree also contained quantitatively in a
united atom simulation. This information is only qualitatively contained in
a bead-spring type simulation. For the united atom model we can go beyond
calculating the scattering of the united atom centerswhich would be
similar to the bead-spring systemand reinsert the hydrogen atoms into
the simulated configurations [8]. For this, the hydrogen atoms are placed
into their T 0 mechanical equilibrium positions (C-H bond length,
C-C-H bond angle, and H-C-H bond angle are fixed to their mechanical
equilibrium values) so that all thermal fluctuations of these degrees of
freedom are neglected. Nevertheless, the distribution of scattering lengths in
the simulated volume resembles closely that of an explicit atom simulation.
Figure 3 shows the X-ray structure factor of the alkane C44H90 from a
united atom [8] and an explicit atom simulation [9]. The intensities are not
normalized as in Eq. (10). For both simulations the position and width of
the amorphous halo (first peak) correspond well with experimental results.
The amorphous halo is the most prominent feature of the structure factor
414
Paul et al.
FIG. 3 X-ray structure factor of C44H90 compared between a united atom and an
explicit atom simulation [8].
V. DYNAMIC PROPERTIES
In the discussion of the dynamic behavior of polymer melts we will start
with the local reorientational motion of chain segments. Experimentally this
motion can for example be observed in Nuclear Magnetic Resonance
(NMR) experiments measuring the 13C spin lattice relaxation time. Owing
to the different chemical shift for different environments of these nuclei, this
measurement can be very specific for selected positions in the polymer chain
[810]. It turns out that this reorientation dynamics is mainly determined
by torsional transitions of dihedral angles adjacent to the relaxing CH
vector. We discussed above that these transitions are exponentially sensitive
to the barrier heights in the torsion potentials. These in turn can only be
determined with an uncertainty of about 100K even through high-level
quantum chemical calculations. Furthermore, we introduce some error in
the torsional dynamics when we use a united atom representation of the
chain. Although both errors may be small, they get exponentially amplified in the torsional transition rate and therefore a comparison with
415
1
3he^CH t e^CH 02 i 1
2
13
For an isotropic system the spin lattice relaxation time T1 and a related
quantity, the Nuclear Overhauser Enhancement (NOE), can be determined
from the power spectrum of this relaxation function.
J!
1
2
P2 tei!t dt
14
1
1
K J!H !C 3J!C 6J!H !C
nT1
NOE 1
H
6J !H !C J !H !C
C J !H !C 3J !C 6J !H !C
15
16
416
Paul et al.
FIG. 4 Temperature dependence of the spin lattice relaxation time for two different
alkanes as seen in experiment (performed at 75 MHz carbon resonance frequency)
and MD simulation [11].
17
For short PE chains the relaxation behavior of the four terminal carbons at
the chain end can be resolved in the NMR experiment and the inner part of
the chain gives rise to a collective resonance. In Fig. 4 we show the behavior
of the spin lattice relaxation time for these inner carbons as a function of
temperature for two alkanes of different chain length, n-C44H90 and
n-C100H202. This seemingly local quantity depends strongly on the chain
length even for these center carbons [11]. Inspecting directly the P2 function
obtained in the simulation we can see that this is due to the fact that a few
percent of the correlations measured by this function decay not through
local motion but through a coupling to the much slower overall conformational relaxation of the chains. This effect is also nicely captured in the
simulations. For the united atom simulations of C44 we can furthermore
see two sets of data points. The one agreeing perfectly with the experimental
results stems from an improved force field, where the torsional barriers in
417
Fig. 1 were slightly raised to improve the agreement with the experimental
data. Other experimental observables such as dielectric relaxation [12] and
the coherent dynamic structure factor for neutron spin echo [13] have also
been calculated from simulations and used to judge the realism of the local
relaxation processes in the simulations.
Simulations of local relaxation in polymers have provided several
important qualitative insights about the mechanisms of molecular motion.
When a conformational transition occurs, nearby atoms along the chain
backbone must adjust their positions to accommodate the transition. This
cooperativity length scale is not accessible experimentally and identifies the
fundamental kinetic unit (which may be the relevant motion for dynamics
near the glass transition temperature). Simulations on polyisoprene [14],
polyethylene [8,15], and other flexible polymers indicate that 46 carbon
atoms along the chain backbone are involved in localizing a conformational
transition. Simulations have also shown significant spatial heterogeneity in
conformational transition rates [15] and the importance of motions other
than conformational transitions in local relaxation processes [16,17].
For an analysis of the overall conformational relaxation of a polymer
chain we focus here on the chain dynamics on length scales which show the
self-similar random walk behavior statically. Imagine looking at a given
polymer chain in a stroboscopic fashion and from a certain distance. The
center of mass of the chain will move (at long times it has to diffuse) and the
conformation of the chain will change from one random walk realization to
a different one. The simplest model for polymer melt dynamics which
captures these effects is the Rouse model [18], in which the polymers are
modeled by a chain of phantom particles connected by entropic springs (this
generates the Gaussian random walk conformations) moving under the
impact of stochastic forces in an unstructured background characterized by
a friction coefficient.
dri t
p
3kB T
ri1 t 2ri t ri1 tdt 2kB T dWi t
2
bK
18
The statistical segment length or Kuhn length bK is the step length of the
random walk with the same large scale structure as the chain under study
and the segmental friction and the strength of the stochastic force are
connected through the fluctuation-dissipation theorem
hdWi tdWj t0 i 2kB Tdtij t t0
19
418
Paul et al.
derived. Since there is only a random force acting on the center of mass of a
chain moving according to the Rouse model, the center of mass motion is
purely diffusive with a diffusion coefficient D kBT/K. Of interest to us is
an analytical prediction for the single chain coherent intermediate dynamic
structure factor. This is the dynamic counterpart to the static single chain
structure factor we discussed earlier on and this quantity is measured in
Neutron Spin Echo (NSE) experiments. The Rouse model yields [18]
Sq; t
1
exp q2 Dt
K
2 2
K
X
q bK
2Kq2 b2K
exp
jn mj
6
32
n;m1
K
X
p1
cos
pn
pm
2
cos
1 ep t= R
K
K
20
)
where R K 2 b2K =32 kB T is the so-called Rouse time. For small momentum transfers q, one is only sensitive to the motion of the polymer chain as a
whole and the above equation simplifies to the scattering law for diffusive
motion S(q, t) S(q, 0) exp(Dq2t). In this way one can determine the center
of mass self-diffusion coefficient from the NSE data and of course also from
the computer simulation. The computer simulation furthermore also has
access to this property by measuring directly the average mean square displacement of the center of mass of a chain and using Einsteins relation
hR2cm ti
t!1
6t
D lim
21
419
2
q
hexp iq ri t rj 0 i exp h ri t rj 0 i
6
22
In the simulation one can prove that the above assumption is not fulfilled by
calculating the scattering function correctly averaging over the phase factors
[left hand side of Eq. (22)] or by employing the dynamic Gaussian assumption [right hand side of Eq. (22)]. Figure 5 shows this comparison for C100 at
509K. One can clearly see, that the dynamic Gaussian assumption leads to
an overestimation of the decay of correlations and to a wrong prediction of
the scattering.
A quantitative comparison with experiments, however, can only be performed for a chemically realistic simulation. For this the positions of the
scatterers in the time dependent generalization of Eqs. (10) and (12) run over
420
Paul et al.
the carbon and deuterium atoms of a single chain and are evaluated with a
time displacement t. Figure 6 shows a comparison of the experimental result
for the single chain intermediate dynamic structure factor of polyethylene at
T 509K with simulations of a united atom as well as an explicit atom
model. For this comparison actually only the deuterium scattering has been
evaluated. When plotting this quantity against scaled time to account for a
20% difference in the observed center of mass self-diffusion coefficients,
the relaxation behavior of the chains on all length scales agrees perfectly
between simulation and experiment [10,21].
421
M
1 X
expfiq rj t rj 0g
M j1
23
where M is the number of hydrogens in the system and where we dropped the
scattering amplitude which only adjusts the absolute scale. For an isotropic
amorphous melt one can perform a spherical average of this equation to get
Sinc q; t
M
1 X
sinqjrj t rj 0j
M j1 qjrj t rj 0j
24
422
Paul et al.
423
FIG. 8 Master curve for the mean square displacements of the beads and of the
centers of mass of the chains obtained by plotting the common part of the mean
square displacement curves for all temperatures as a function of time scaled by the
respective diffusion coefficient of that temperature [27].
VII.
OUTLOOK
424
Paul et al.
REFERENCES
1. Smith, G.D.; Jae, R.L. J. Phys. Chem. 1996, 100, 18718.
2. Smith, G.D.; Paul, W. J. Phys. Chem. A 1998, 102, 1200.
3. Lucretius MD code and sample data les for PE: http://www.che.utah.
edu/gdsmith/mdcode/main.html.
4. Nose, S. J. Chem. Phys. 1984, 81, 511.
5. Kremer, K.; Grest, G.S. J. Chem. Phys. 1990, 92, 5057.
6. Yoon, Do, Y. unpublished.
7. Paul, W.; Smith, G.D.; Yoon, Do.Y. Macromolecules 1997, 30, 7772.
8. Paul, W.; Yoon, Do.Y.; Smith, G.D. J. Chem. Phys. 1995, 103, 1702.
9. Smith, G.D.; Yoon, Do.Y.; Zhu, W.; Ediger, M.D. Macromolecules 1994, 27,
5563.
10. Smith, G.D.; Paul, W.; Monkenbusch, M.; Willner, L.; Richter, D.; Qiu, X.H.;
Ediger, M.D. Macromolecules 1999, 32, 8857.
11. Qiu, X.H.; Ediger, M.D. Macromolecules 2000, 33, 490.
12. Smith, G.D.; Yoon, Do.Y.; Wade, C.G.; OLeary, D.; Chen, A.; Jae, R.L.
J. Chem. Phys. 1997, 106, 3798.
13. Moe, N.E.; Ediger, M.D. Phys. Rev. E 1999, 59, 623.
14. Moe, N.E.; Ediger, M.D. Polymer 1996, 37, 1787.
15. Boyd, R.H.; Gee, R.H.; Han, J.; Hin, Y. J. Chem. Phys. 1994, 101, 788.
16. Moe, N.E.; Ediger, M.D. Macromolecules 1995, 28, 2329.
17. Moe, N.E.; Ediger, M.D. Macromolecules 1996, 29, 5484.
18. Doi, M.; Edwards, S. The Theory of Polymer Dynamics; Clarendon Press:
Oxford, 1986.
19. Paul, W.; Heermann, D.W.; Kremer, K.; Binder, K. J. Chem. Phys. 1991, 95,
7726.
20. Kopf, A.; Dunweg, B.; Paul, W. J. Chem. Phys. 1997, 107, 6945.
21. Paul, W.; Smith, G.D.; Yoon, Do.Y.; Farago, B.; Rathgeber, S.; Zirkel, A.;
Willner, L.; Richter, D. Phys. Rev. Lett. 1998, 80, 2346.
22. Smith, G.D.; Paul, W.; Monkenbusch, M.; Richter, D. J. Chem. Phys., 2001,
114, 4285.
23. Proceedings of the Third International Discussion Meeting on Relaxation in
Complex Systems, J. Non-Cryst Solids, 1998, 235237.
24. Han, J.; Ge, R.H.; Boyd, R.H. Macromolecules 1994, 27, 7781.
25. Gotze, W.; Sjogren, L. In Transport Theory and Statistical Physics; Yip, S.,
Nelson, P., Eds.; Marcel Dekker: New York, 1995; 801853.
26. Bennemann, C.; Baschnagel, J.; Paul, W. Eur. Phys. J. B 1999, 10, 323.
27. Bennemann, C.; Baschnagel, J.; Paul, W.; Binder, K. Comp. Theor. Polym. Sci.
1999, 9, 217.
14
Sorption and Diffusion of
Small Molecules Using
Transition-State Theory
MICHAEL L. GREENFIELD
Rhode Island, U.S.A.
I.
INTRODUCTION
426
Greenfield
p
1=2
6Dt 6 109 cm2 s1 108 s 0:75 A
This distance is 30100 times smaller than the typical simulation cell sizes
used. Furthermore, since diffusion through a glassy polymer is thought to
occur by a hopping mechanism [79], such an average displacement more
likely would result only from statistics: one molecule with a 7.5 A jump
and nine molecules with net displacements of zero, for example. Direct
molecular dynamics simulations are thus prone to large errors in predicted
diffusivity in a glassy polymer, due to (1) mismatches between the simulation time and the expected waiting time for a jump to occur and (2) poor
sampling of the many jump rates that would be found in a macroscopic
sample.
One method for predicting diffusion rates in glassy polymers has been
proposed by Boyd and coworkers [49,52]. As in glassy polymers, a hopping
mechanism has been found in melt polymers not far above their glass
427
*Explicit descriptions of the hopping mechanism can also be found in most of the recent MD
papers. References [80,81] focus specically on the jump mechanism.
428
Greenfield
in X-ray scattering. The jump path was again confined to the three degrees
of freedom of the penetrant, but now within the free energy field
characteristic of the vibrating polymer. This method has led to reasonable
agreement with experiment in particular cases [24,63,97100] and has been
implemented in a popular molecular simulation software package [101].*
However, the predictions are sensitive to the value of the DebyeWaller
parameter, and hence some judgment is required when using this technique.
In the most recent method [88,91,102], which we refer to as the explicit
polymer method, the dimensionality of the jump path was increased to
include explicitly both penetrant displacements and polymer chain motions.
Each resulting rate constant then captures the details of the particular chain
motions that accompany the jump, rather than mean-field average motions,
but at the expense of a much higher computational cost. This method has
been applied to a few penetrants in polypropylene [88,89,91], with reasonable agreement compared to experiment and correlation.
In this chapter we will begin by formulating the transition-state theory
method within the context of small-molecule diffusion in a polymer matrix,
and then show how the transition-state theory equations are applied to a
single jump in each of these three approaches. We then compare different
methods for establishing the network of possible penetrant jumps and
explain how diffusion can be simulated on that network. We close with
a summary and a discussion of the outlook for other systems. Several
appendices describe aspects of particular approaches in more detail.
429
A. Transition State
For transition-state theory to be applicable, the physical system must
include some high-energy region through which the system must pass in
order to convert from reactant to product. Numerous such regions
exist for a small molecule diffusing through a glassy polymer, and the socalled transition states lie within them. Mathematically, a transition state is
usually taken as a first-order saddle point on the 3N-dimensional potential
energy hypersurface, meaning that all elements of the potential energy
gradient equal zero and all eigenvalues of the Hessian matrix of second
derivatives but one are positive. In Cartesian coordinates,
@V
0 for all positions X i
@X i
@xi
@xi @q j
j1
*While the dividing surface technically describes a hypothetical equilibrium, it can be sampled
using equilibrium statistical mechanics.
430
Greenfield
X
i
qi rq V
qi
i
s
431
432
Greenfield
From the transition state, the IRC is directed along the eigenvector corresponding to the negative eigenvalue [115,116]. Motions along all other
eigenvectors locally define a multidimensional dividing surface that delineates the reactant and product regions. The IRC is orthogonal to this dividing surface. In other words, the IRC step through the transition state
(over the pass in a mountain crossing analogy) is directed along the
single direction of decreasing local curvature and orthogonal to all directions
of increasing local curvature.
Away from the transition state, the IRC is formally defined by a steepest
descent trajectory* in mass-weighted Cartesian coordinates [112],
x rx V
*This usage of steepest descent refers to a path constructed by taking a sequence of small
steps, each directed opposite to the local gradient. This usage diers from that in the optimization literature; there [109], the magnitude of each step is chosen such that the objective function
is minimized with respect to motion along the step vector.
433
10
X @xk @xk
k
@qi @q j
11
X energy
kg m2
qj 2
Hqq q i
q i qj
s qi
j
a qi
X kg1=2 m2
kg m2
qj
qi qj
qi
j
and
has units of s2, or frequency squared, as expected.
Canned routines that solve the generalized eigenvalue problem are
available in standard numerical libraries.* Because a J 0 TJ 0 , the covariant
metric tensor a must be symmetric and positive definite, which facilitates
finding the eigenvalues and eigenvectors, compared to a more general case
in which the matrix characteristics of a are unknown.
*The routines REDUC, TRED2, and TQL2 of EISPACK, available at [120], are examples of
canned routines. Matrix routines from the LAPACK [121] and LINPACK [122] packages are
also available.
434
Greenfield
Away from the transition state, the IRC equation is written as [88,102]
a q rq V
12
Taking a step thus requires calculating the gradient and covariant metric
tensor, and then solving a set of linear equations.
13
meaning that the potential energy gradient with respect to each flexible
coordinate equals zero.
435
14
The step to take away from the transition state is the eigenvector qf corresponding to the negative eigenvalue of the generalized eigenvalue problem
Eq. (14). The first step along the IRC (from the transition state) is found
by the following steps:
1.
2.
3.
A small step along this eigenvector qf defines the first step along the IRC.
Along the step, changes occur in only the flexible coordinates; infinitely stiff
coordinates remain unchanged.
Away from the transition state, the remaining steps along the diffusion
path are calculated from
a0 qf rqf V
15
Tracking all but the first step along the IRC thus requires these steps:
1.
2.
3.
4.
The elements of the Jacobian matrices (and thus the elements of a0) vary
with chain conformation and must be recalculated after each step. The
IRC ends when it reaches a local minimum (the gradient rqfV approaches
zero).
D. Final State(s)
The two IRCs that begin at a transition state each terminate at a local
potential energy minimum, where the gradient equals zero. Near the
436
Greenfield
minimum, the gradient becomes sufficiently small (a step size of order 1010,
for example [88]) that each IRC step [Eq. (15)] is excessively short. It would
then take a large number of steps to travel the small remaining distance to
the local minimum configuration. Instead, a quasi-Newton scheme [109]
can be used at this point to reach the local minimum. Relatively few steps
should be required, and the resulting potential energy decreases should be
negligible. If large decreases in energy do occur, that indicates larger-scale
structural relaxation of the polymer matrix, and the system configuration
prior to quasi-Newton steps should be restored [i.e., the last configuration
found via Eq. (15)]. In such a case, additional regular IRC steps should be
taken despite the low gradient, and the step size should soon increase again.
This approach maintains the validity of the calculated IRC. Such behavior
could occur along a near-flat plateau on the potential energy landscape that
is surrounded by more steep regions of similar orientation. After continued
energy decreases, a small gradient should recur, and quasi-Newton steps can
be attempted again.
E. Rate Constant
The pair of rate constants for each jump path (IRC) can be calculated in
one of several ways. A formal, microscopic formulation [83] was discussed
in Chapter 1 for the rate constant,
q_ 0 qz q0 qt qz
z
kA!B t
q q
1:101
437
kB T Qz
Vz V0
exp
h Q
kB T
16
kB T
Az A0
exp
h
kB T
(from Chapter 1). The exponential in Eq. (16) arises from a difference in the
zero of energy between Qz and Q.
Several approximations to Eq. (16) can then be made. The simplest
approximation is to assume a constant prefactor and to use the potential
energy difference between the transition state and the reactant sorption
state,
Vz V0
kTST k0 exp
kB T
17
This approach can be interpreted via Eq. (1.117) as assuming that the free
energy difference equals the potential energy difference (entropy change
equals zero) and that the temperature dependence in the exponential
dominates that in the prefactor.
A less drastic approximation is to assume that (1) potential energy
increases relative to the origin or transition state energies are adequately
described locally by a harmonic approximation, and (2) potential energies
outside these harmonic regions are sufficiently large such that the
Boltzmann factor approaches zero. With this combination of assumptions,
the origin state and dividing surface boundaries can be extended to 1, and
the partition functions can be calculated from the vibrational frequencies at
the transition state and the reactant state, with respect to the flexible degrees
of freedom. The frequencies are obtained from the eigenvalues of Eq. (14) by
i
i
2
i 2, . . . , f transition state
i 1, . . . , f reactant state
18
438
Greenfield
@Y
f
2
00
a1
Vz V0
A det
00 z exp
kB T
z det a1
1
19
00
lim
!1
kB T
0
hi
but instead
hi
lim 1 exp
1
!1
kB T
Consequently the contributions from infinitely stiff modes become unity and
the rate constant can be written as
h
1 exp
1
kB T
Vz V0
kB T
exp
h Yf
kB T
hz
1
exp
2
kB T
Yf
kTST
20
The jump rate constants reported in our previous work [88] were calculated
using Eq. (20). Within the harmonic approximation, the difference in
*Despite this zero limit, the ratio of the products of these frequencies, calculated via
00z
00
(det a1 = det a1 0 ), is well-dened with a value near unity. See [88,123] for details.
439
1
1 00 z
S
B h 1 det a
C
ln@
Yf
A1
z det a1 00 min
kB
kB T
2
Yf
21
(classical mechanics-based) or
1
0Y
f
hz
1
exp
B 2
S
k T C
B
C
1
lnB
@
Y
f
h A
kB
1 exp
1
kB T
22
440
Greenfield
441
442
Greenfield
atom (x1, y1, z1) and the orientation of the bond vector in spherical
coordinates,
x2 x1 cos sin
23
y2 y1 sin sin
24
z2 z1 cos
25
and are the projection of the bond on the x-y plane (0 < 2) and its
azimuthal angle relative to the z-axis (0 ). The mass-weighted bond
length is
m1=2 X2 X1 2 Y2 Y1 2 Z2 Z1 2 1=2
26
443
0
2
0
I1 sin2 sin2
6
4 I2 sin2 cos2
C
C
C
C
0
0
C
C
C
C
I3 cos I1 I2 sin sin cos C
C
C
C
C
C
I3
0
A
0
3
7
5
I3 cos2
I3 cos
I1 cos2
I2 sin2
29
This metric tensor can then be used in Eqs. (14) and (15). Its block-diagonal
character separates translation and rotation steps, an advantage for deciphering diffusion mechanisms. One complication, however, is that the
Eulerian angles must be carefully monitored and redefined in case any
pass through zero [138].
In principle, the transition-state theory equations developed above [Eqs.
(13), (14), (15), (16), (1.117), (20)] can be used for each jump in the frozen
polymer method. While the individual components of the saddle-point search
have a reputation of being slow (calculating the Hessian matrix, evaluating all
eigenvalues and eigenvectors), in practice these steps are fast because the
dimensionality is small. Standard matrix routines [120122,140] are recommended for the matrix calculations. Typically following the IRC will require
significantly more computational effort than finding the transition state.
A good estimate for the transition state location will speed up the search
process and help to ensure a successful search. Geometric interpretations
of the simulation cell are often useful for this purpose. The voxel method
(useful throughout the frozen polymer method) is discussed below. Another
approach is to subdivide (or tessellate) the periodic simulation cell into
tetrahedra [87,93,141145] or cubes [146,147], with a hard-sphere representation of the atoms replacing an energy calculation. Sorption states can
be assigned to the center of tetrahedra or cubes not fully occupied by atoms
of the polymer chain, with transition states assigned to the center of surface
triangles [93]. Alternatively, tetrahedra or cubes can be lumped into sorption
states, and initial guesses for transition state locations can be taken at the
center of tetrahedra/cubes that form the intersection of sorption states
*Equation 12 in [138] [the equivalent of Eq. (29) substituted into Eq. (15)] contains a typographical error. This form of a0 is correct.
444
Greenfield
[88,102]. The important point is using such criteria to recognize the threedimensional regions in a polymer structure that most closely resemble the
necks between (presumably) larger sorption state regions.
An alternative approach typically used within the frozen polymer method
and the average fluctuating polymer method is the so-called voxel
approach [85,94]. A finely spaced cubic grid (lattice spacing of 0.1 A, for
example) is overlaid onto the simulation cell, and the energy required to insert
a penetrant at the center of each cube is calculated. For each voxel (i.e., cube)
i, the lowest-energy neighboring cube j is identified, and the connectivity i ! j
is stored. Alternatively [85], a steepest descent calculation can be conducted
from each voxel. After looping over all voxels, the net result is a connectivity
map for the particular polymer structure chosen. Voxels with no neighbors of
lower energy correspond to local minima. Each voxel can then be labeled as
leading (or belonging) to the state defined by the local minimum in which
a connectivity path terminates. (If i ! j ! k ! , then voxels i, j, k, and
belong to state .) The dividing surface between states is defined by the
common faces of voxels in different states, and the center of the face between
the lowest-energy voxels on the dividing surface approximates the position of
the transition state. This position is not an exact solution to Eq. (13), but it is
close to one. Similarly, the connectivity path defined by adjoining, lowest-inenergy voxels approximates the IRC defined by Eqs. (14) and (15).
For a spherical penetrant, the energy at each voxel can be used to
calculate each rate constant through numerical integration of Eq. (16).
The kinetic energy contributions to each partition function are integrated
analytically in mass-weighted Cartesian coordinates,
2kB T
z
Q
ZA!B
h2
2kB T 3=2
Q
ZA
h2
z
30
31
Zj rVj m3=2
32
voxel j in
state A
Vj
exp
Vj m3=2
kB T
voxel j in
X
state A
33
445
A!B
boundary
The voxel energy V j is assumed constant within the volume Vj and across
the surface area Aj.
The dividing surface is a curved region, which voxels approximate as a
zig-zag of cubic surfaces. To calculate each Aj (the area of the dividing
plane that cuts through voxel j ), Gusev and Suter [97] proposed
Aj KL2v
35
where Lv is the voxel edge length and K (see Table 1) depends on the number
of neighboring voxels assigned to state B (the destination state). The fewer
neighboring voxels in the destination state, the better this approximation.
Gusev et al. [94] estimated area errors A of under 7% based on a voxel edge
of Lv 0.2 A. From Eqs. (16), (30), and (34), these errors propagate directly
into the rate constant. Smaller voxels would lead to smaller area errors,
according to
jAj L2v jKj
36
where K is the difference between the value listed in Table 1 and the true
ratio of dividing surface size to voxel face area. Consequently decreasing
TABLE 1 Coefficients Used when Estimating the
Contribution of Each Voxel to the Dividing Surface
Area (Equation 35 from [97]). A Voxel can Have a
Maximum of 6 Nearest Neighbors
Number of neighboring
voxels in state B
1
2
3
4
5
6
K
1p
2
1.41
2.0
0
0
446
Greenfield
voxel size can significantly reduce dividing surface area errors. The final
voxel-based expression for the rate constant is thus
Vj
Kj L2v
exp
k
T
B
voxel j on
boundary
X
Vj
exp
L3
kB T v
voxel j
X
TST
kB T 1=2
2m
37
in state
2h2 i
exp
dxi hxi i
2h2 i
i
38
447
where hxi is the vector of all polymer atom positions in the penetrant-free
equilibrated (i.e., preaveraged) polymer sample. The three-dimensional position of a single atom is written xi. The mean-squared fluctuation along each
average position (x, y, or z) equals h2 i; hence the parameter notation as
an RMS value. The unnormalized probability density for all penetrant
positions rp and polymer atom positions x is then written as
Prp , xdx hxi
Y
V ip rp , xi
jxi hxi ij2
2 3=2
2h i
exp
exp
dxi hxi i
kB T
2h2 i
i
39
1=2
Zj rp
Vrp , x
Px hxidx hxi
kB T
40
The Boltzmann factor depends on the positions of all atoms that interact
with the penetrant. For a pairwise additive penetrantpolymer force field,
such as the Lennard-Jones expression
Vrp
"
4ip
ip
jrp xi j
12
6 #
ip
jrp xi j
41
Y
i
Y
3=2
V ip rp , xi
Zji rp
2h i
exp
kB T
i
jxi hxi ij2
dxi hxi i
exp
2h2 i
2
42
448
Greenfield
449
1=2
p
6Dt
43
1=2
The value of h2 i at 1ns will be of order 10 times larger than the value at
1=2
10 ps. Even in the reptation regime [113,114], in which hr2i t1/2, h2 i
would appear to increase by about 3 over this time scale, which based on
1=2
Fig. 1 would have a big effect on diffusion. Clearly the value of h2 i
depends on the time scale of interest.
1=2
Several methods have been proposed for choosing the parameter h2 i .
Its value is crucial, because the magnitude and characteristics of meansquared displacement (normal vs. anomalous diffusion) are sensitive
1=2
functions of h2 i . Regrettably, several practitioners of this method
2 1=2
values from their publications. Arizzi [93] noted that an
omit h i
analytic form is available from the X-ray scattering literature [148],
h2 i
Z D =T
3h2 T
T
d
D
e 1 4T
42 mkB D 2 D 0
44
450
Greenfield
They reasoned that since the jump significantly changes the list of neighboring polymer chain atoms (and especially the distances from the atoms that
remain neighbors), it is appropriate for the elastic effects to be considered
afresh. This process [97] then entails self-consistency:
1.
2.
3.
4.
5.
1=2
guess h2 i ,
calculate the escape time distribution,
1=2
rechoose h2 i based on the most likely escape time,
recalculate the escape time distribution,
iterate to convergence.
This method implicitly requires the peak and breadth of the escape time
1=2
distribution to be equally sensitive to h2 i , since the average and breadth
of a waiting time distribution are governed by only one parameter [103].
However, the sensitivities are not necessarily the same, and Gusev et al. [24]
1=2
noted that the h2 i value so-calculated is only approximate. They next
suggested [24] conducting both MD and TST simulations, using the mod1=2
erate time behavior (t<10 ns) to provide a target for choosing h2 i :
1.
2.
3.
4.
Such a combination of MD and TST has promise for systems in which the
assumption of uncoupled elastic fluctuations is valid.
451
The general approach for applying the explicit polymer method involves
1.
2.
3.
4.
A molecular scale network spanning the original simulation cell, with its
jump paths and rate constants, can then be used when calculating the diffusivity. Alternatively, a larger network can be constructed [91] (see Section
IX) that incorporates molecular-level correlations.
The task of searching for a penetrant jump transition state was
discussed in Section II. Each multidimensional transition state is a solution
to Eq. (13), which is solved most efficiently in stages [88]. First a lowdimensional transition state is found for the penetrant by keeping the
polymer frozen (as above). Next, polymer degrees of freedom can be
introduced into the transition state search, using the transition state found
in the lower dimensionality as an initial guess. The appropriate degrees of
freedom to include are those defined by chain segments near the penetrant.
All of the selected degrees of freedom (including the penetrant) are allowed
to evolve during the transition state search in the higher dimension. The
resulting transition state has two elements of unstable equilibrium. Forces
on the penetrant are balanced such that it senses neither a force towards
the origin state nor one towards the destination state. Simultaneously,
neighboring polymer chains have withdrawn in response to penetrant
polymer forces until they are balanced by the force required to compress
the polymer matrix locally. The result is a more open channel for
penetrant diffusion.
To some extent, as the number of coordinates involved in calculating
the transition state increases, the transition state energy decreases. As an
example, Fig. 2 shows the decrease in transition state energy for a methane
penetrant in atactic polypropylene at three different transition states, as a
function of transition state dimensionality. The greater the number of
degrees of freedom, the larger the chain region able to take up this
compressive force and the smaller the transition state energy. However,
the decrease in energy differs from transition state to transition state; no
simple, quantitative correlation for the magnitude of the energy decrease
exists. The overall energy decrease relative to that in the initial state
(about an order of magnitude smaller than for the frozen polymer case)
corresponds to a rate increase by a factor of about 106. A penetrant
jump assisted by polymer motions is thus much more likely than an
452
Greenfield
FIG. 2 Penetrant diffusion transition state energy relative to the local minimum
energy of pure polymer, as a function of the number of coordinates considered
flexible during the transition state search. Different symbols correspond to different
transition states. The transition state energy at three flexible coordinates would be
used in the frozen polymer method.
unassisted jump, as found for the fluctuating vs. frozen polymer methods.
As seen in the figure, the transition state energy ultimately reaches a
plateau value, after which it is unaffected by including additional degrees
of freedom.
In order to reach such a plateau, the initial polymer structure (no
penetrant present) must be in a local potential energy minimum. Otherwise,
each additional polymer coordinate included will contribute a new
opportunity for chain relaxation, leading to a transition state lower in
energy than the original structure but higher than a local potential energy
minimum.
For conducting numerous penetrant diffusion simulations, it isnt
reasonable to inspect Fig. 2 by eye to choose when to stop adding
degrees of freedom; an automated procedure is required instead. One
approach [88] is to continue until all degrees of freedom defined by atoms
within the penetrant neighbor list have been added. The remaining atoms
are affected only indirectly: the penetrant interacts with atom i, which then
453
interacts with atom j. This approach led to about 350 degrees of freedom
(penetrant position, bond angles, torsion angles, sometimes chain start
positions and orientations) for methane in glassy atactic polypropylene.
A second approach is to identify all atoms that the penetrant is likely to
encounter during its jump between two sorption states, and then to include
all of them in the transition state search. Such a list could be created by
following the IRC in only the penetrant dimensionality, in the spirit of the
frozen polymer method. An advantage of this approach would be that all
degrees of freedom that would participate in the IRC would be identified
in advance and would move towards a zero-force conformation at the
transition state. A disadvantage is the large number of degrees of freedom
involved (of order 500 for methane in atactic polypropylene). In either case,
we found [88] it was worthwhile to increase the included number of degrees
of freedom slowly (5, 15, . . . , etc.), creating a sequence of higher dimensional
initial guesses, rather than to attempt the complete search initially. While
this would seem to require more calculation, in practice the sequential initial
guesses were good enough such that in most cases no single search required
a long time.
Next, the multidimensional IRC must be determined for this jump.
The Jacobian and the covariant metric tensor include contributions from
all f flexible coordinates, and the IRC is found by following a single step
away from the transition state [Eq. (14)] and then multiple steps along the
IRC [Eq. (15)]. Qualitative descriptions of polymer contributions to the
diffusion mechanism can be taken directly from how each flexible
coordinate changes along the IRC. For example, Fig. 3 shows different
chain conformations along one diffusion jump. Dashed and solid triangles
indicate original and instantaneous conformations for selected chain
segments. Due to coupled changes in polymer bond and torsion angles,
chains move back from the penetrant path as the jump occurs, creating a
wider channel.
Next the rate constant is calculated for each jump. In prior work [88] we
used a harmonic approximation in the flexible degrees of freedom, with the
quantum mechanics-based partition function [Eq. (20)]. Such a rate constant
calculation is fast in comparison to tracking the IRC. Using a free energy
method (Eq. (1.117) or sampling as in Ref. [111]) would yield more accurate
rate constants at a cost of a larger, more time consuming calculation.
Dynamical corrections [Eqs. (1.103) and (1.104)] would increase accuracy
but add additional computational requirements. If the software and run
time are available, these more advanced methods are recommended.
Finally, the rate constants can be used to calculate the diffusivity.
Methods based on jump network structure and kinetic Monte Carlo
simulation are discussed below.
454
Greenfield
VII.
455
Numerical artifacts can result from curvature along the IRC. For example,
too large a step along the gradient can lead to a departure from the true
diffusion path. The gradient at the end of the step will direct the system back
towards the IRC, but due to components orthogonal to the IRC it is likely
that so-called overshoot will occur again. The net result is a sequence of
steps that repeatedly cross the IRC [126]. This leads to two disadvantages:
1.
2.
FIG. 4 Cartoon of a zig-zag approximation of the IRC (bold dotdash line). The
thin solid path that repeatedly crosses the IRC (dotdash path) is likely to result
when no Hessian information is used, such as with Eqs. (14) and (15). Dashed and
dotted lines indicate contours of constant potential energy.
456
Greenfield
Path-based methods [110,111,118,127,128,151153] differ from the approaches discussed above in that they attempt to find an entire reaction/
diffusion path (the IRC) between local minima, without first performing a
transition state search. Instead, the IRC is represented by a sequence of
order Nst 1050 state points in configuration space. Each state point
includes all f flexible mass-weighted Cartesian or generalized coordinates.
The different methods correspond to different algorithms, typically
[110,118,152] some sort of minimization, for solving for the resulting Nstf
coordinates. On a multidimensional potential energy surface, the solution
corresponds to a discretized approximation to a low-energy pathway
traversing as directly as possible between local minima. The f 2 analog
inspires descriptive names for these approaches: chain of states [118],
nudged elastic band [110,152], throwing ropes over mountain passes in
the dark [111]. In model problems, the local minima have not even had to
neighbor one another; the path obtained could pass through multiple
transition states and intermediate local minima. However, the high
dimensionality has prevented path-based methods from being applied to
diffusion in polymers with a flexible matrix. For example, 20 state points
along an IRC in [88] (about 350 flexible coordinates) would require a
minimization with respect to 7000 variables. Evaluating the Hessian matrix
is required in some methods.
It is also possible that via thermal motions a penetrantpolymer system
would sample different sets of pathways, each with its own transition state
structure, for a penetrant jump between two associated local minima.
The methods in [111] are particularly suited to calculating the rate in
that case.
VIII. SORPTION
Penetrant sorption has three impacts on small-molecule permeability.
A direct impact is the proportionality between permeability P and solubility
in solution-diffusion transport [1,154]
P DS
45
46
The higher the sorption coefficient, the larger the number of penetrant
molecules dissolved in the polymer sample and able to diffuse across it.
457
SB kA!B
47
=P
p!0 hVpolym i
kB T
kB T
kB T
kB T
48
The ratio hNpeniA/hVpolymi equals the number of penetrant molecules dissolved in a particular sorption state A per total volume of polymer, and P
equals the total pressure. The ensemble average brackets h i imply averaging over all penetrant positions within a sorption state A while the polymer conformation fluctuates within the canonical ensemble according to
V polym, the total potential energy in the absence of a penetrant molecule.
458
Greenfield
ex
A equals the excess chemical potential for a penetrant in sorption state A.
While each sorption coefficient SAP
depends inversely on the system volume,
the total sorption coefficient S ASA is an intensive quantity, since the
number of sorption states varies linearly with the total polymer volume.
Equation (48) applies to the low-concentration limit because it neglects
penetrantpenetrant interactions.
Several methods are available for calculating the excess chemical
potential of a penetrant molecule using molecular simulation. Widom
particle insertion [156,157] is the most straightforward. The potential energy
change that would occur if a penetrant were inserted at a randomly chosen
position is calculated, and the excess chemical potential is proportional to
the average of the resulting Boltzmann factor,
V pen
ex
kB T ln exp
49
kB T
This approach has been applied several times to small-molecule sorption in
polymers [46,77,158,159]. However, a difficulty arises for moderate-sized or
large penetrants: there would be insufficient space for a penetrant to be
inserted; hence the increase in potential energy is large and the predicted
sorption is low. Several alternative approaches exist, and here we list just a
few that have been applied to penetrant sorption. In mixed canonical/grand
canonical simulations, the penetrant chemical potential is specified and its
concentration fluctuates, but the amount of polymer remains constant
[155,160162]. In Gibbs ensemble Monte Carlo [163,164], molecule
exchanges between two simulation cells are used to achieve phase equilibrium. Thermodynamic integration [22,165] is based on using a coupling
parameter to relate a more easily simulated system to the system of interest,
and several recent penetrantpolymer implementations show promise
[166168]. Extended ensembles treat these coupling parameters as variables
in the simulation [168171]. For chemically similar small and large molecules, chain scission/regrowth or connectivity reassignment can be used to
estimate the chemical potential [172,173]. Combinations, such as using
an extended ensemble within a Gibbs ensemble framework [45,174], are
also useful. Other approaches [161,171,175] lead to segment-level incremental chemical potentials, which are useful for calculating relative solubility.
Finally, it was recently proposed that a polymer membrane in contact with a
penetrant phase can be simulated directly [176].
Techniques that can improve statistical sampling are often crucial in
these approaches. An excluded-volume map can improve efficiency by
focusing particle insertions into free volume voids [159,177179]. Specially
designed Monte Carlo moves allow polymer chain fluctuations within
459
a densely packed system [180182] (see also Chapter 7). Recent reviews
[183,184] and Chapters 9 and 10 provide in-depth descriptions of methods for calculating chemical potentials in polymer systems, and their
applications.
Another approach to calculating penetrant sorption [185] is to calculate
the penetrant energy as in the Widom insertion method, but at points along
a regularly spaced lattice, using the results to calculate the Henrys Law
constant. A related, popular technique is Gusev and Suters theory [186],
which couples particle insertion concepts and the voxel method described
above for finding transition states in three dimensions. Their approach is
particularly convenient when applying the frozen polymer method or the
fluctuating polymer method, in part because the voxel calculations are
performed anyway and in part because it has been implemented in a popular
software package [101]. It differs from the prior approach by applying a
local partition function to each sorption state. The resulting functional form
is similar to the popular dual-mode sorption model [1], with a different
Langmuir parameter bi for each state.
One aspect of a typical sorption calculation may require some
modification in order to use the results in a diffusion simulation. As
suggested by Eq. (47) and illustrated below, the sorption coefficient
distribution is required for formulating the jump network and thereby
obtaining the penetrant diffusion coefficient. Thus the sorption calculation
must be formulated such that individual results are obtained for each
sorption state, rather than a single sorption result for an entire simulation
cell. If the network generation method described below will be used (rather
than the molecular-scale jump network that emerges directly from simulated
molecular structures), it is convenient to accumulate sorption correlations
during molecular-level simulations. For example, mutual distributions of
sorption magnitude and sorption state connectivity (number of neighbors
accessible via low-energy paths from each sorption state) can be later
incorporated into a large-scale network.
460
Greenfield
461
50
462
Greenfield
51
XX
i
Ri!j t
52
463
and
Ri!j Ni tki!j
53
equals the flux from state i to state j. The average waiting time until the next
jump will occur anywhere on the network equals
Z
1
W
Rt expRt dt
1
eRt
1
1
Rt
te
0 0 0
R
R
R
0
tp t dt
0
54
The amount of time that elapses between penetrant jumps thus depends
explicitly on the instantaneous sorption state population and on the distribution of jump rate constants out of occupied states.
The sorption state populations evolve according to a master equation.
For a system in which each sorption state can be represented by a single
local minimum,
X
X
dpi
ki!j pi
kj!i pj
dt
j
j
55
!
pA
NB X
NA
X X
B
j1 i1
pj
kj!i
pB
!
pB
56
A and B represent different overall sorption states and i and j are individual
regions of states A and B, respectively. The ratio pi/pA equals the fraction of
sorption in state A that occurs within region i. After some time, the sorption
eq
state probabilities stabilize at their equilibrium values peq
i or pA that satisfy
microscopic reversibility.
The master equations can be solved numerically using kinetic Monte
Carlo simulation [85,86,155]. To begin, a number of independent,
noninteracting ghost penetrants are placed in the sorption states
P of
each network, according to the equilibrium distribution pAeq SA = SA .
464
Greenfield
These initial positions are stored in a vector r0. The distribution can be
sampled using the same method as described above when assigning the
sorption coefficients to each sorption state:
.
n
X
Si =
i1
Ns
X
Si
57
i1
Ri!j t
Rt
58
are calculated. Next, the expected waiting time until the next event, , is
found as follows. The instantaneous waiting time distribution [Eq. (51)]
leads to a cumulative waiting time distribution by
Zt
W
P t
pW t dt 1 expRt
59
0
1
ln1
R
60
The jump that occurred in this event is chosen from the probabilities qi!j t
[from Eqs. (53) and (58) a larger rate constant implies a larger probability],
and the particular ghost penetrant leaving state i (if more than one was
present) is chosen uniformly. Selecting from the qi!j t distribution is performed in the same manner as above, using the cumulative probability
density distribution. The state label for the jumping penetrant is updated,
the occupancies Ni(t) and Nj(t) are reset to account for the jump, and the
process is then repeated many times. Each jump changes the average meansquared displacement (it can increase or decrease) and increases the elapsed
time by . In practice, only fluxes out of the sorption states i and j that
465
involve the most recent jump have to be recalculated according to Eq. (53);
this decreases the number of computations significantly for a large system.
The kinetic Monte Carlo simulation should begin by conducting many
steps in order to equilibrate the penetrant sorption state positions, after which
r0 is reset to zero. A production run can then begin. In the production run, the
net result of many steps (103104 per particle) is a list of penetrant sorption
state occupancies as a function of time. The mean-squared displacement can
be calculated (during or after the simulation) at each time as
N
D
X
E
rt r0 2 1
xti x0 2 yti y0 2 zti z0 2
N i1
61
1
@hr2 i
1
lim
m
2d t!1 @t
2d
62
where m is the long-time slope on a linear plot of hr2i vs. t, and d is the
dimensionality.* Over shorter times, diffusion may be anomalous, meaning
hr2 i / t n
63
with n<1. (Note that this is opposite to so-called case II diffusion [188],
in which penetrant mean-squared displacement increases with time squared.
Here diffusion is slower than in the Fickian case.) Features of interest are
the value of the exponent n and the times and distances over which diffusion
is anomalous. Anomalous diffusion can occur due to a distribution of rate
constants [189191] or network connectivity [24,91,192]; it can also depend
on the penetrant size [42]. The essential feature is a limitation of the diffusion paths such that they cannot span the full dimensionality over the simulation time scales available.
An example of the change in mean-squared displacement with time is
shown in Fig. 6 for methane in atactic polypropylene [91]. These
calculations were performed using a network generated via the methods
described in Section IX. At times of less than 1 ns, little information is
*Even for a network whose connectivity can be described using a fractal dimension, d is the
dimensionality of the space in which the network is represented; i.e., d 2 for surface diusion
and d 3 for bulk diusion.
466
Greenfield
467
468
Greenfield
3N X
3N
3N
X
X
@qk @ql @2 V
@2 ql @V
i
j
k
l
@x @x @q @q
@xi @xj @ql
l1
k1 l1
or
0
B
1
B
@q @q2
@q3N B
B
Hij
,
,...,
@xi @xi
@xi B
B
@
@2 V
@q1 @q1
..
.
...
@2 V
@q1 @q3N
..
.
@2 V
@2 V
...
3N
1
3N
@q @q
@q @q3N
1
0
@V
B @q1 C
C
2 1
B
C
@ q
@ 2 q2
@2 q3N B
B ... C
,
,
.
.
.
,
C
B
i
j
i
j
i
j
@x @x @x @x
@x @x B
C
@ @V A
@q3N
10
@q1
@xj
..
.
CB
C
CB
C
CB
C
CB
C
CB
C
CB
C
A@ @q3N A
@xj
64
469
for each term Hij of the Hessian. Different rows and columns of H result
from taking all i 1, . . . , 3N and j 1, . . . ,3N components, respectively.
The resulting matrix equation is
2 T
@ q
T
65
H J Hqq J
rq V
@xi @xj
Hqq is the Hessian of second derivatives with respect to generalized coordinates and @2 qT =@xi @xj , a third-rank tensor, may be thought of as a 3N 3N
matrix, in which each component is itself a vector; the ijth vector contains
the coordinate transformation second derivatives of q with respect to xi and
x j. The resulting rightmost term is a matrix for which the ijth element is the
dot product of the ijth vector with rqV. This tensor complicates the use
of an analytic Hessian matrix in generalized coordinates, since additional
derivatives are required compared to typical simulation needs.
Next, the proper equation for determining eigenvalues and eigenvectors
must be reformulated. The Hessian H can be replaced using Eq. (65), and
x (renormalized to a small size x rather than unity) is replaced using
xi
3N
X
@xi
j1
@q j
q j
or
x J 0 q
66
470
Greenfield
69
@x1
@q1
..
.
B
B
B
J B
B
@ @x3N
0
@q1
1
@x1 @x1
@x1
@q f @q f 1
@q3N C
C
.. C 0 0
..
..
..
..
Jf Js
. C
.
.
.
.
C
3N
3N
3N A
@x
@x
@x
f
f
1
@q @q
@q3N
70
471
a0
a @ -a0T
---
--
71
The f f submatrix a0 in the upper left is the covariant metric tensor in only
the flexible coordinates. Each element
a
X @xk @xk
k
@q @q
72
using the Jacobian subsets introduced above. Each element of the submatrix
a0 and its transpose a0 T involve one flexible coordinate and one stiff coordinate. All derivatives in a00 are with respect to stiff coordinates.
The contravariant metric tensor a1 in the full space can be subdivided
in a similar manner. However, a subset of a1 in the reduced space is not
simply a single submatrix, since the partial derivatives that define each term
a1
X @q @q
k
@xk @xk
are taken with all other mass-weighted positions held constant, not all other
generalized coordinates. Instead, the reduced-space contravariant metric
tensor equals the matrix inverse in the reduced dimensionality,
0
a1 a0 1
The matrix a1 0 is sometimes denoted G [105].
73
472
Greenfield
1
a0
a0T
a0
a00
a1
0T
a1
a1
00
a1
I
If
0
I 3Nf
74
a0 a1 a0 a1
0 1 0
aa
0T
0 1 00
aa
If
75
76
00
00
1
77
a1
0T
a1
a1
00
a1
a0
a0T
a0
a00
I
If
0
I 3Nf
78
multiplying the second row of a1 and the first column of a leads to
0T
00
79
473
0T
1
80
which is used below. Finally, each side of Eq. (74) can be postmultiplied by
a0
a0T
0
I
a1
0T
00
a1
00
a0
a0T
0
I
81
Substituting using the transpose of Eq. (75) and Eq. (79) leads to
a1
00
a1
If
a
0
a0
a0T
0
I
82
83
This interrelationship among the magnitude of different metric tensor determinants was first derived by Fixman [195] and is useful in discussions of the
rigid and flexible models of polymer conformation (see Appendix C).
Next, the saddle point and Hessian calculations can be narrowed to a
subset of all coordinates. At a saddle point in the full dimensionality, the
potential energy gradient equals zero with respect to all generalized
coordinates. The same criterion holds for the flexible coordinates in a
subset of the full dimensionality. For constrained degrees of freedom
A f 1, . . . ,3N, the stiff potential maintaining the constraints is of the
form
V
X1
A
2
kstiff A qA qA
0
474
Greenfield
1
A 2
2 kstiff A q
H
@2 V
@q @q
is positive except for the lowest, which must be negative. The form of the
eigenvalue equation to solve will be discussed next.
As in the full dimensionality, the first step of the IRC is directed along
the eigenvector corresponding to the negative eigenvalue. The eigenvalue
equation to solve in the subset of flexible coordinates may be found from the
submatrix partitions of a, H, and q. In order to distinguish between the
flexible and stiff coordinates, it is convenient to premultiply Eq. (8) by a1,
leading to
0
@
a1
10
00
A@
a1
0T
a1
a1
H0qq
H0qq
H0T
qq
H00qq
1
A
qf
qs
qf
!
84
qs
Expanding the left hand side leads to separate equations for the flexible and
stiff coordinates,
0
0
0
0
1
H0qq a1 H00qq qs
qf
a1 H0qq a1 H0T
qq qf a
0T
0T
00
1
0
1 00 00
a
H
a
H
q
a1 H0qq a1 H0T
f
qq
qq
qq qs
qs
85
86
As described above, qA
s ! 0 for each stiff coordinate A in the infinite
stiffness limit. Even with the force constants approaching infinity, the
right hand side of Eq. (86) approaches zero, as does the second term on
the left hand side of the same equation. Since the step qf in the flexible
475
0T
1
H0qq
a1 H0T
qq a
87
88
89
H0T
qq qf
90
H00qq qs
0T
00
a qf
a qs
Setting qs ! 0 in the first leads directly to Eq. (14) . However, the same
limit for the second leads to the seemingly contradictory
0T
H0T
qq qf
a qf
91
0T
1
a1 H0T
a0 qf
qq qf
a
92
The left hand side can then be replaced using Eq. (87), leading to
0T
0T
a1 H0qq qf
a1 a0 qf
93
0T
Removing the common matrix a1 from the leftmost side of each side of
this equation yields Eq. (14).
Theodorou (D. N. Theodorou, personal communication, 1993) has
obtained the same results with a derivation that remains in the Cartesian
coordinate system. First, he partitioned the eigenvectors into flexible and
476
Greenfield
stiff components, finding that deviations along the stiff eigenvectors must
equal zero. Next, he invoked a harmonic approximation with respect to the
full set of mass-weighted Cartesian coordinates, ultimately obtaining the
same equations as those presented here, within a change of coordinate
system.
The IRC step in a flexible subset of coordinates away from the transition
state also follows from subdividing the step vector and the Jacobian matrix.
Begin with the IRC equation in mass-weighted Cartesian coordinates,
Eq. (7). Substitute for x using Eq. (66), but note that only a subset of J 0 is
required, since qA 0 for each stiff coordinate A. The elements of J 0 that
multiply a nonzero value of qi correspond exactly to the elements of Jf 0
defined in Eq. (70). Similarly, only the gradients with respect to flexible
generalized coordinates are nonzero
T
rqf V Jf 0 rx V
94
since the gradient with respect to stiff degrees of freedom equals zero in
either the rigid or the infinitely stiff model. Premultiplying Eq. (7) by Jf 0T ,
substituting with the flexible equivalent of Eq. (66), and using Eq. (94)
results in
T
Jf 0 Jf 0 qf rqf V
95
From Eq. (72) above, the product Jf 0T Jf 0 equals the covariant metric tensor
in the reduced coordinate system, yielding Eq. (15).
To evaluate the rate, it is necessary to combine expressions for the
partition functions at the local minima and at the transition state. The latter
partition function is one dimension smaller than that of the former, since
integration is limited to the (3N 1)-dimensional dividing surface that
passes through the transition state. An exact differential equation for this
surface exists [112], but its evaluation is difficult; it amounts to following the
3N 1 intrinsic reaction coordinates defined by taking the first step along
an eigenvector associated with a positive eigenvalue. A common assumption
is to approximate the dividing surface near the transition state with a
hyperplane orthogonal to the eigenvector corresponding to the negative
eigenvalue; the dividing plane is orthogonal to the step taken away from the
transition state. The resulting equation can be written in a flexible subset of
generalized coordinates as
x xz x
qf qzf T a0 qf
0
96
477
where x
and qf
are the eigenvectors that correspond to the negative
eigenvalue in mass-weighted Cartesian coordinates and in the flexible subset
of generalized coordinates. In terms of the normal modes that correspond
to the flexible generalized coordinates, this surface restricts integration to
modes with real frequencies. From the orthogonality of eigenvectors, all
these modes are normal to the first step along the reaction coordinate.
APPENDIX C:
478
Greenfield
0
full dimensionality are followed (i.e., use of a1 or a rather than a1 or a0 in
the IRC equations), even infinitesimal steps would induce changes in the
stiff degrees of freedom, since each change depends on all elements of the
gradient,
dqA
X X @qA @q j @V
j
@xk @xk
@q j
97
Z
Vqf
2kB T 3N=2
0
det
J
exp
d
q
f
h2
kB T
98
1 T 0
1
0
q_ a q_ Pr T a1 Pr
2
2
99
479
100
Z
Vqf
2kB T 3N=2
0 1=2
exp
d qf det a
Q
h2
kB T
101
or
Hence the appearance of the metric tensor inside the integral is due to the
nondiagonal nature of the kinetic energy in the constrained space.
Go and Scheraga also showed that the Jacobian determinant in the
flexible model could be expressed as the Jacobian determinant used in the
rigid model multiplied by a number of classical harmonic oscillators, each
contributing kinetic and potential energy to the Hamiltonian. This
relationship is summarized by Eq. (83). Integration over the kinetic
energy part led to cancellation of (det a0)1/2 in the configurational integral,
while integration over the potential energy led to an arbitrary constant
multiplying Q. The treatments of the two polymer models are identical in
terms of only the potential energy. Within the IRC formalism, dissipation of
kinetic energy is instantaneous, and the kinetic energy need not be treated
until rate constants (and partition functions) are evaluated. In summary, the
method of following the IRC described above only specifies the Riemannian
geometry of the reaction path; it does not choose a physical model of the
classical statistical mechanics. Such a choice is made only when calculating
the rate constants.
480
Greenfield
xi hxi i 2h2 i
1=2
1
t0i
@ u0i A
v0i
102
Below we leave off each atom subscript i from the deviation vector in order
to simplify the notation. Differential changes in (xi hxii) and (t0 , u0 , v0 )T are
related by a Jacobian
3=2
dxi hxi idyi hyi idzi hzi i 2h2 i dt0 du0 dv0
The square of each penetrantpolymer atom separation can then be
written as
p 2
p 2
D2i rp x hxi i 2h2 it0 rp y hyi i 2h2 iu0
p 2
rp z hzi i 2h2 iv0
The contribution to the configurational integral can then be written as
(using a Lennard-Jones potential as an example)
3=2
Zji rp
1 1
"
ip6
4ip ip12
exp
kB T D2 6 D2 3
1
1
!#
02
02
02
481
f x expx2 dx
1
n
X
wk f xk
104
k1
3=2
X
kz
V ip
exp
expt02 expu02 expv02 dt0 du0 dv0
kB T
1 1
2
2
0
1
BX
6X
6 4
ip
B
6
6
wkz 6
wky B
wkx exp6
@k
4k
4 kB T
y
0
B
ip12
B
B
p 2 6
@
2h2 it0k 2 2h2 iu0k 2 rp z hzi i 2h2 iv0k
1313
C7C7
ip6
C7C7
7C7
2 3 C
p
A5A5
2
2
0
0
0
2
2
2
2h itk 2h iuk rp z hzi i 2h ivk
105
For an odd order n, one GaussHermite root equals zero, meaning the
energy is evaluated at the voxel center (t0 , u0 , v0 ) (0, 0, 0). Each individual
configurational integral is thus determined by the summation in Eq. (105).
The rotation of the interatomic
vector
to along the z axis is useful because
it allows the function Zji jrp hxij in Eq. (105) to be pretabulated in one
dimension (the separation), rather than evaluated from scratch for each
penetrantpolymer atom separation in each voxel. The pretabulated function thus depends on jrp hxij=, h2 i1=2 =, and /kBT. The same procedure
482
Greenfield
106
1
2
drr expr
1
"
!#
ip6
4ip ip12
dcosexp
kB T D2i 6 D2i 3
107
483
A numerical scheme such as the trapezoidal rule or Simpsons rule [109] can
be used for the integral over cos , with GaussHermite integration used
to integrate over r.
In comparing the two methods at a fixed separation of 5 A, we found
essentially identical results so long as enough terms were used. For the first
approach, 79 terms within each sum were sufficient (73 343 terms in all).
For the second approach, 79 GaussHermite terms were sufficient as well,
but in combination with of order 104 trapezoidal terms. The first method
(wholly in Cartesian coordinates) is thus recommended for pretabulating
the function Zji(Di).
Even within the approach based on Cartesian coordinates, more
GaussHermite terms are required for each coordinate at shorter separations. Figure 7 compares the modified potential energy kBT lnZji(r) as
calculated using 316 terms. Within the repulsive regime, using too few
terms leads to oscillations about the potential of mean force calculated with
more terms. The smaller the separation included, the more terms required.
FIG. 7 Modified potential energy between a penetrant and a single polymer atom,
within the average fluctuating polymer method. Different lines correspond to the
numerical results based on different numbers of GaussHermite terms. The same
number of terms was used for each of the x, y, and z coordinates. Increasing the
number of terms decreases oscillations at small separations.
484
Greenfield
REFERENCES
1. Vieth, W.R. Diusion in and Through Polymers: Principles and Applications;
Hansen: New York, 1991.
2. Paul, D.R.; Yampolskii, Y.P. Polymeric Gas Separation Membranes; CRC
Press: Boca Raton, FL, 1994; Chapter 1.
3. Koros, W.J.; Fleming, G.K. J. Membr. Sci. 1993, 83, 180.
4. Nagai, K.; Masuda, T.; Nakagawa, T.; Freeman, B.D.; Pinnau, I. Prog. Polym.
Sci. 2001, 26, 721798.
5. Jorgensen, W.L.; Maxwell, D.S.; Tirado-Rives, J. J. Am. Chem. Soc. 1996, 118,
1122511236.
6. Rigby, D.; Sun, H.; Eichinger, B.E. Polym. Internat. 1997, 44, 311330.
7. Sun, H. J. Phys. Chem. B 1998, 102, 73387364.
8. Martin, M.G.; Siepmann, J.I. J. Phys. Chem. B 1998, 102, 25692577.
9. Martin, M.G.; Siepmann, J.I. J. Phys. Chem. B 1999, 103, 45084517.
10. Chen, B.; Siepmann, J.I. J. Phys. Chem. B 1999, 103, 53705379.
11. Wick, C.D.; Martin, M.G.; Siepmann, J.I. J. Phys. Chem. B 2000, 104,
80088016.
12. Chen, B.; Poto, J.J.; Siepmann, J.I. J. Phys. Chem. B 2001, 105, 30933104.
13. Nath, S.K.; Escobedo, F.A.; de Pablo, J.J. J. Chem. Phys. 1998, 108,
99059911.
14. Nath, S.K.; de Pablo, J.J. Mol. Phys. 2000, 98, 231238.
15. Smith, G.D.; Jae, R.L.; Yoon, D.Y. Macromolecules 1993, 26, 298304.
16. Smith, G.D.; Jae, R.L.; Yoon, D.Y. J. Phys. Chem. 1993, 97, 1275212759.
17. Muller-Plathe, F. Acta Polymerica 1994, 45, 259293.
18. Smith, G.D.; Borodin, O.; Bedrov, D. J. Phys. Chem. A 1998, 102,
1031810323.
19. Hedenqvist, M.S.; Bharadwaj, R.; Boyd, R.H.; Macromolecules 1998, 31,
15561564.
20. Holt, D.B.; Farmer, B.L. Polymer 1999, 40, 46674772.
21. Neyertz, S.; Pizzi, A.; Merlin, A.; Maigret, B.; Brown, D.; Deglise, X. J. Appl.
Polym. Sci. 2000, 78, 19391946.
22. Allen, M.P.; Tildesley, D.J. Computer Simulation of Liquids; Oxford University
Press: New York, 1987.
23. Pant, P.V.K.; Boyd, R.H. Macromolecules 1993, 26, 679686.
24. Gusev, A.A.; Muller-Plathe, F.; van Gunsteren, W.F.; Suter, U.W. Adv.
Polym. Sci. 1994, 116, 207247.
25. Gee, R.H.; Boyd, R.H. Polymer 1995, 36, 14351440.
26. Takeuchi, H. J. Chem. Phys. 1990, 93, 44904491.
27. Takeuchi, H.; Okazaki, K. J. Chem. Phys. 1990, 92, 56435652.
28. Takeuchi, H.; Roe, R.-J.; Mark, J.E. J. Chem. Phys. 1990, 93, 90429048.
485
486
Greenfield
65. Sonnenburg, J.; Gao, J.; Weiner, J.H. Macromolecules 1990, 23, 46534657.
66. Rottach, D.R.; Tillman, P.A.; McCoy, J.D.; Plimpton, S.J.; Curro, J.G.
J. Chem. Phys. 1999, 111, 98229831.
67. Muller-Plathe, F. J. Chem. Phys. 1995, 103, 47454756.
68. Catlow, C.R.A.; Mills, G.E. Electrochim. Acta 1995, 40, 20572062.
69. Payne, V.A.; Lonergan, M.C.; Forsyth, M.; Ratner, M.A.; Shriver, D.F.;
de Leeuw, S.W.; Perram, J.W. Solid State Ionics 1995, 81, 171181.
70. Neyertz, S.; Brown, D. J. Chem. Phys. 1996, 104, 37973809.
71. Ennari, J.; Elomaa, M.; Sundholm, F. Polymer 1999, 40, 50355041.
72. Borodin, O.; Smith, G.D. Macromolecules 2000, 33, 22732283.
73. Snyder, J.F.; Ratner, M.A.; Shriver, D.F. J. Electrochem. Soc. 2001, 148,
A858A863.
74. Aabloo, A.; Thomas, J. Solid State Ionics 2001, 143, 8387.
75. Vishnyakov, A.; Neimark, A.V. J. Phys. Chem. B 2001, 105, 95869594.
76. Ferreira, B.A.; Muller-Plathe, F.; Bernardes, A.T.; De Almeida, W.B.
Solid State Ionics 2002, 147, 361366.
77. Muller-Plathe, F.; Rogers, S.C.; van Gunsteren, W.F. J. Chem. Phys. 1993, 98,
98959904.
78. Sunderrajan, S.; Hall, C.K.; Freeman, B.D. J. Chem. Phys. 1996, 105,
16211632.
79. Takeuchi, H. J. Chem. Phys. 1990, 93, 20622067.
80. Muller-Plathe, F.; Laaksonen, L.; van Gunsteren, W.F. J. Mol. Graphics 1993,
11, 118120.
81. Muller-Plathe, F.; Laaksonen, L.; van Gunsteren, W.F. J. Mol. Graphics 1993,
11, 125126.
82. Glasstone, S.; Laidler, K.J.; Eyring, H. The Theory of Rate Processes;
McGraw-Hill: New York, 1941.
83. Chandler, D. J. Chem. Phys. 1978, 68, 2959.
84. Voter, A.F.; Doll, J.D. J. Chem. Phys. 1985, 82, 8092.
85. June, R.L.; Bell, A.T.; Theodorou, D.N. J. Phys. Chem. 1991, 95, 88668878.
86. Fichthorn, K.A.; Weinberg, W.H. J. Chem. Phys. 1991, 95, 10901096.
87. Greeneld, M.L.; Theodorou, D.N. Macromolecules 1993, 26, 54615472.
88. Greeneld, M.L.; Theodorou, D.N. Macromolecules 1998, 31, 70687090.
89. Rallabandi, P.S.; Thompson, A.P.; Ford, D.M. Macromolecules 2000, 33,
31423152.
90. Greeneld, M.L.; Theodorou, D.N. Mol. Simul. 1997, 19, 329361.
91. Greeneld, M.L.; Theodorou, D.N. Macromolecules 2001, 34, 85418553.
92. Jagodic, F.; Borstnik, B.; Azman, A. Makromol. Chem. 1973, 173, 221231.
93. Arizzi, S. Diusion of Small Molecules in Polymeric Glasses: A Modelling
Approach. Ph.D. thesis, Massachusetts Institute of Technology, 1990.
94. Gusev, A.A.; Arizzi, S.; Suter, U.W.; Moll, D.J. J. Chem. Phys. 1993, 99,
22212227.
95. Bell, A.T.; Maginn, E.J.; Theodorou, D.N. In Handbook of Heterogeneous
Catalysis, Ertl, G., Knozinger, H., Weitkamp, J., Eds.; VCH: Weinheim, 1997.
96. Auerbach, S.M. Int. Rev. Phys. Chem. 2000, 19, 155198.
103.
104.
105.
106.
107.
108.
109.
110.
111.
112.
113.
114.
115.
116.
117.
118.
119.
120.
121.
122.
123.
487
488
Greenfield
161.
162.
163.
164.
165.
166.
167.
168.
169.
170.
171.
172.
173.
174.
175.
176.
177.
178.
179.
180.
181.
182.
183.
184.
489
Mills, G.; Jonsson, H.; Schenter, G.K. Surf. Sci. 1995, 324, 305337.
Zimmer, M.F. Phys. Rev. Lett. 1995, 75, 14311434.
Petropoulos, J.H. J. Membr. Sci. 1990, 53, 229255.
Theodorou, D.N. In Diusion in Polymers; Neogi, P., Ed.; Marcel Dekker:
New York, 1996; 67142.
Widom, B. J. Chem. Phys. 1963, 39, 28082812.
Widom, B. J. Phys. Chem. 1982, 86, 869872.
Muller-Plathe, F. Macromolecules 1991, 24, 64756479.
Tamai, Y.; Tanaka, H.; Nakanishi, K. Macromolecules 1995, 28, 25442554.
Boone, T.D. Prediction of Glass-Melt Behavior and Penetrant Sorption
Thermodynamics in Vinyl Polymers via Molecular Simulations. Ph.D. thesis,
University of California, Berkeley, 1995.
Spyriouni, T.; Economou, I.G.; Theodorou, D.N. Phys. Rev. Lett. 1998, 80,
44664469.
Kofke, D.A. Adv. Chem. Phys. 1999, 105, 405441.
Panagiotopoulos, A.Z. Mol. Phys. 1987, 61, 813826.
Panagiotopoulos, A.Z. Mol. Simul. 1992, 9, 124.
Shing, K.S.; Gubbins, K.E. Mol. Phys. 1982, 46, 11091128.
Knopp, B.; Suter, U.W.; Gusev, A.A. Macromolecules 1997, 30, 61076113.
Knopp, B.; Suter, U.W. Macromolecules 1997, 30, 61146119.
van der Vegt, N.F.A.; Briels, W.J. J. Chem. Phys. 1998, 109, 75787582.
Escobedo, F.A.; de Pablo, J.J. J. Chem. Phys. 1995, 103, 27032710.
Escobedo, F.A.; de Pablo, J.J. J. Chem. Phys. 1996, 105, 43914394.
Nath, S.K.; de Pablo, J.J.; DeBellis, A.D. J. Am. Chem. Soc. 1999, 121,
42524261.
de Pablo, J.J.; Laso, M.; Suter, U.W.; Cochran, H.D. Fluid Phase Equilibria
1993, 83, 323331.
Zervopoulou, E.; Mavrantzas, V.G.; Theodorou, D.N. J. Chem. Phys. 2001,
115, 28602875.
Nath, S.K.; de Pablo, J.J. J. Phys. Chem. B 1999, 103, 35393544.
Kumar, S.K.; Szleifer, I.; Panagiotopoulos, A.Z. Phys. Rev. Lett. 1991, 66,
29352938.
Kikuchi, H.; Kuwajima, S.; Fukuda, M. J. Chem. Phys. 2001, 115,
62586265.
Deitrick, G.L.; Scriven, L.E.; Davis, H.T. J. Chem. Phys. 1989, 90, 23702385.
Stapleton, M.R.; Panagiotopoulos, A.Z. J. Chem. Phys. 1990, 92, 12851293.
Fukuda, M. J. Chem. Phys. 2000, 112, 478486.
de Pablo, J.J.; Laso, M.; Suter, U.W. J. Chem. Phys. 1992, 96, 61576162.
Siepmann, J.I.; Frenkel, D. Mol. Phys. 1992, 75, 5970.
Dodd, L.R.; Boone, T.D.; Theodorou, D.N. Mol. Phys. 1993, 78, 961996.
de Pablo, J.J.; Yan, Q.; Escobedo, F.A. Annu. Rev. Phys. Chem. 1999, 50,
377411.
Ferguson, D.M.; Siepmann, J.I.; Truhlar, D.G., Eds. Monte Carlo Methods in
Chemical Physics (Adv. Chem. Phys.); John Wiley & Sons: New York, 1999;
Vol. 105.
490
Greenfield
185. Bezus, A.G.; Kiselev, A.V.; Lopatkin, A.A.; Du, P.Q. J. Chem. Soc. Faraday
Trans. II 1978, 74, 367379.
186. Gusev, A.A.; Suter, U.W. Phys. Rev. A 1991, 43, 64886494.
187. McGreevy, R.L.; Howe, M.A. Annu. Rev. Mater. Sci. 1992, 22, 217242.
188. Alfrey, T., Jr.; Gurnee, E.F.; Lloyd, W.G. J. Polym. Sci. C 1966, 12, 249261.
189. Havlin, S.; Ben-Avraham, D. Adv. Phys. 1987, 36, 695798.
190. Haus, J.W.; Kehr, K.W. Phys. Rep. 1987, 150, 263406.
191. Stauer, D.; Aharony, A. Introduction to Percolation Theory, 2nd Ed.; Taylor
& Francis: London, 1992.
192. Muller-Plathe, F.; Rogers, S.C.; van Gunsteren, W.F. Chem. Phys. Lett. 1992,
199, 237243.
193. Weber, H.; Paul, W. Phys. Rev. E 1996, 54, 39994007.
194. Brandrup, J.; Immergut, E.H.; Grulke, E.A. Polymer Handbook, 4th Ed.;
WileyInterscience: New York, 1999.
195. Fixman, M. Proc. Nat. Acad. Sci. USA 1974, 71, 30503053.
196. Go , N.; Scheraga, H.A. J. Chem. Phys. 1969, 51, 47514767.
197. Helfand, E. J. Chem. Phys. 1979, 71, 50005007.
198. Fixman, M. J. Chem. Phys. 1978, 69, 15271537.
199. Abramowitz, M.; Stegun, I.A. Handbook of Mathematical Functions; National
Bureau of Standards: Washington, DC, 1964; Reprint published by Dover
Publications: Mineola, New York, 1965.
15
Coarse-Graining Techniques
K. BINDER and WOLFGANG PAUL
Mainz, Mainz, Germany
Johannes-Gutenberg-Universitat
I.
Eidgenossische Technische
For many materials, it is difficult to span the wide range of scales of length
and time from the atomistic description to that of macroscopic phenomena.
For polymers, this multiscale problem is particularly severe, since a single
flexible macromolecule already exhibits structure on many length scales
(Fig. 1), and is a slowly relaxing object [1,2]. In addition, one often wishes to
consider systems with multiple constituents (concentrated polymer solutions
or blends, mixtures of homopolymers and block copolymers or polymers of
different architecture) which exist in several phases, and then the scales of
length and time that are of physical interest truly extend from the subatomic
scale (relevant for the electronic structure of polymers, their chemical
reactions, excited states, etc.) to the macroscopic scale (e.g., phase separated
domains of the scale of micrometers or larger, glassy relaxation on the scale
of seconds or hours, etc.) [3].
Since there is no hope of treating all these scales equally well in a single
type of computer simulation, the idea of coarse-graining has emerged
[4,5], where a mapping was attempted from an atomistically realistic
description of polymers such as polyethylene (PE) or bisphenol-Apolycarbonate (BPA-PC) towards the bond fluctuation model [6,7]; see
Fig. 2. Now coarse-grained lattice models such as the self-avoiding walk
model or the bond fluctuation model in its standard form have been used for
a long time to model polymers qualitatively, but this has the obvious
drawback that the information on the chemical structure of the considered
polymer is lost completely, and thus nothing can be said on structure
property relationships. Though these standard coarse-grained models are
492
Binder et al.
FIG. 2 A chemically realistic description of a polymer chain (bisphenol-Apolycarbonate [BPA-PC] in the present example) is mapped approximately onto
the bond fluctuation model by using suitable potentials for the length of the
effective bonds and the angles between them. In this example (3:1 mapping) one
chemical repeat unit of BPA-PC containing n 12 covalent bonds along the
backbone of the chain is translated into three effective bonds. From Paul et al. [5].
Coarse-Graining Techniques
493
nevertheless useful for some purposes (see Chapter 4), the idea to be
described here is the attempt to keep some information about the chemical
structure indirectly also on the coarse-grained level. Thus, although the
lattice spacing of the model in Fig. 2 corresponds to about 2 A, and hence
clearly the detailed packing of the various atoms or atomic groups in the
polymer melt or glassy state cannot be directly represented, some
information on such properties is kept indirectly, via the potentials U(),
V(#) for the lengths, , of the effective bonds and the angles, #, between
them. Chemically different polymers are then described by different
potentials.
Obviously, constructing the effective potentials explicitly will not be very
straightforward [8,9], and also the description of physical properties by
means of such a coarse-grained lattice model cannot be expected to be
perfect. Thus, the idea has arisen to use the coarse-grained model only as an
intermediate step for a simulation that equilibrates the long wavelength
degrees of freedom of the polymer and then to carry out an inverse
mapping where in Fig. 2 the arrow is simply read backwards! However, this
step is technically very difficult, and although it has been under discussion for
the bond fluctuation model for more than a decade now, it has not yet
actually been performed successfully (only for another type of lattice model
on the face-centered cubic lattice, obtained from chains on the diamond
lattice where every second monomer is eliminated, see [10,11], has a reverse
mapping been possible). The explicit construction of potentials for coarsegrained degrees of freedom [such as U(), V(#)] and the inverse mapping has
successfully been carried out for a coarse-grained continuum model rather
than a lattice model, e.g., a bead-spring model [12,13]. The basic idea is the
same in all these methods (Fig. 3), but there is considerable freedom in
defining the details of this procedure. For example, in the case of bisphenolA-polycarbonate a 2:1 mapping to a bead-spring chain was performed
[12,13], i.e., one chemical monomer (representing 12 consecutive covalent
bonds along the chain backbone, see Fig. 2) is mapped onto two effective
monomers. Taking the geometrical centers of the carbonate group and of the
isopropylidene group as the sites of the effective monomers, it is actually
advantageous to allow for a difference between these effective monomers:
thus, the coarse-grained chain representing BPA-PC is not a bead-spring
model of a homopolymer A, A, A, A . . . , but rather a bead-spring model of
an alternating copolymer ABAB. . . . The crucial point of all these models,
however, is such that on the coarse-grained length scale the effective potentials U(), V(#) are much smaller than the atomistic potentials for the lengths
of covalent bonds or the angles between them, and torsional potentials on the
coarse-grained level can be neglected completely. As a result, on the coarsegrained level all potentials are on the same scale of energy, unlike the
494
Binder et al.
FIG. 3 Illustration of the coarse-graining procedure, starting out from a unitedatom model for polyethylene as an atomistic description (the larger spheres show a
segment of polyethylene at a temperature T 509K taken out of a molecular
dynamics melt). One coarse-grained bond (straight line) represents the end-to-end
distance of n 5 consecutive united-atom bonds. The effective monomers (black
dots) actually are represented by the eight sites of an elementary cube of the simple
cubic lattice, if a mapping on the bond fluctuation model is used; if a mapping on a
bead-spring chain is used, the effective monomers (black dots) are represented by the
repulsive part of a suitable Lennard-Jones potential, and the effective bonds are
anharmonic springs described by the FENE potential. From Tries et al. [9].
atomistic level where the scales for the potentials of covalent bond lengths
and angles are much higher than the weak Lennard-Jones type nonbonded
forces. Hence, while on the atomistic scale a Molecular Dynamics time step
on the femtosecond scale is needed, the molecular dynamics time step for a
coarse-grained bead-spring model can be several orders of magnitude larger.
As a result, one can reach equilibrium for distinctly lower temperatures using
the coarse-grained model than with the atomistic simulation, and somewhat
larger length scales are accessible, too.
Nevertheless, one must admit that this step emphasized in Figs. 2 and 3
is only a modest first step towards the goal of bridging the gap between
the length and time scales of atomistic simulations and the scales of
macroscopic phenomena in materials science. The same caveat applies to
other interesting alternatives to this mapping procedure, which we shall not
describe in detail here, such as mapping to ellipsoidal rather than spherical
effective monomers [14,15], or generating configurations of dense polymer
melts by constructing an ensemble of curves of constant curvature, using
stream lines of a vector field that never intersect [16,2]. But how can one get
from the scales of, say, 10 nm and 100
sec, that are only accessible to the
coarse-grained models, to still much larger scales?
One line of attack to this problem is to try to use a next mapping
step from a bead-spring chain to a still much cruder model. For instance,
Coarse-Graining Techniques
495
Murat and Kremer [17] try to map whole chains of soft ellipsoidal particles,
which then must be allowed to overlap strongly in the melt. This approach
seems particularly useful for a description of heterophase structures in
polymeric materials, such as polymer blends that are unmixed on
nanoscopic scales. It seems promising to combine this model or related
approaches [18] with efficient techniques for describing the time evolution,
making use of the fact that very soft potentials between the effective
particles allow rather large time steps, such as dissipative particle dynamics
[19]. While these approaches clearly are very promising avenues for further
research, we feel it is premature to review them here simply because this field
is at its beginning and at the same time rapidly developing.
Another line of attack is the idea of bypassing this step-to-step mapping
from one scale to the next by directly making connections between the
atomistic scale and the macroscopic continuum scale by performing
calculations on both scales simultaneously and appropriately connecting
them [20,21]. We shall briefly describe the essential ideas of this approach,
emphasizing mechanical properties, in Section III of this chapter (a first
application to plastic deformation of bisphenol-A-polycarbonate can be
found in Chapter 12 of this book). In the following section, we shall explain
the mapping alluded to in Figs. 2 and 3 in some more detail, while Section
IV summarizes a few conclusions.
N
p 1
X
j1
H lj
N
p 2
X
H j
j1
N
p 2
X
H j
j2
HLJ ~rij
j6i
where H(j) / (j cc) with a fixed bond length j cc 1.53 A, while the
potential for the cosine of the bond angles j was assumed harmonic,
H j
f
cos j cos 0 2 ,
2
0 1108,
f 120 kcal=mol
496
Binder et al.
3
X
ak cosk j
k0
exp H x~
fx~g
where 1=kB T and the sum is to be interpreted as an integral for continuous degrees of freedom. Let us further assume that we have chosen a set of
mesoscopic degrees of freedom (e.g., the lengths and angles of the coarse~ . Then we can write
grained bonds in Fig. 3), denoted by m
Z
XX
exp H x~
fm~ g fx~gm~
~.
where the sum over fx~g is restricted to a fixed set of mesoscopic variables m
~ , T by writing
Introducing a generalized free energy F m
Z
~; T
exp F m
fm~ g
we see that this defines a mapping from the probabilities for the microscopic
configurations px~ 1=Z expfHx~g to those for the degrees of free~ 1=Z expfFm
~ , Tg.
dom pm
On the mesoscopic scale we are dealing with a free energy and not with a
Hamiltonian, so that such a mapping has to be done for each temperature
Coarse-Graining Techniques
497
498
Binder et al.
FIG. 4 Probability distribution of (a) the coarse-grained bond length L and (b) the
bond angle between two such coarse-grained bonds, choosing n0 10, m 2. From
Tries et al. [9].
Coarse-Graining Techniques
499
7
2
where h1=Ti denotes an average over the temperatures where input information from the atomistic calculation (such as shown in Fig. 4) is used [8].
From this fit the eight parameters u0, u1, 0 , 1 , v0, v1, c0, c1 are obtained,
as well as the time constant 0 that is necessary to translate the Monte Carlo
time units [ one attempted Monte Carlo step (MCS) per effective
monomer] into a physical time. As is well known, the random hopping
algorithm for the bond fluctuation model, where one selects an effective
monomer randomly and tries to displace it in a randomly chosen lattice
direction, can represent the Rouse-like or reptation-like dynamics of dense
polymer melts in a rough way [2,26]; see also Chapter 4. But the Monte
Carlo method never has an intrinsic physical time scale of its own, and just
as the density of polyethylene needs to be used to give the length unit of the
lattice model a physical meaning, one also needs physical input for the time
unit. It is desirable to map the mobility of the lattice model onto the average
jump rate of the torsional degrees of freedom of the chain, since these
motions dominate the relaxation of its overall configuration. Thus a
temperature-dependent time unit MC(T ), which one attempted MCS per
monomer corresponds to, is introduced via
1
4H
1 1 X
min 1; exp
Ator T
MC T
kB T
0 Ntor tor
10
500
Binder et al.
min1; expH=kB T ABFL 1 exphE i Emin =kB T
11
FIG. 5 Time rescaling factor MC(T ) for the coarse-grained Monte Carlo model for
polyethylene plotted vs. temperature. The constant 0 in Eq. (10) was fixed by fitting
the self-diffusion constant DN for C100H202 to experiment at T 450K. From Tries
et al. [9].
Coarse-Graining Techniques
501
MC(T ) has increased to about a nanosecond. Since Monte Carlo runs of the
order of 107 MCS are easily feasible, this means that relaxation processes on
the millisecond time range can be explored. Figure 6 shows that in this way a
reasonable prediction for the temperature dependence of both static and
dynamic properties of polyethylene melts over a wide temperature range
is obtained, including the melt viscosity [27], with no further adjusted parameters. Of course, this modeling accounts neither for the breaking of
covalent bonds at high temperatures, nor for crystallization of polyethylene
(that occurs for T<Tm 414K), and a further weak point at the present
stage is that attractive intermolecular interactions have been omitted as well.
The off-lattice coarse-grained models [12,13] have not included attractive
intermolecular forces either, and thus none of these approaches yields a
502
FIG. 6
Binder et al.
Continued.
reasonable equation of state. It appears that using the pressure of the system
as an additional input variable into the mapping of bead-spring type offlattice models to fix the strength of the attractive part of the Lennard-Jones
interaction between nonbonded effective monomers is a promising recipe to
eliminate this weakness of current coarse-graining procedures [28].
Coarse-Graining Techniques
503
A number of approaches to connect multiple-scale simulation in finiteelement techniques have been published [3134]. They are able to describe macroscopically inhomogeneous strain (e.g., cracks)even dynamic
simulations have been performed [35]but invariably require artificial
constraints on the atomistic scale [36]. Recently, an approach has been
introduced that considers a system comprising an inclusion of arbitrary
shape embedded in a continuous medium [20]. The inclusion behavior
is described in an atomistically detailed manner [37], whereas the continuum
is modeled by a displacement-based Finite-Element method [38,39]. The
atomistic model provides state- and configuration-dependent material
properties, inaccessible to continuum models, and the inclusion in the
atomistic-continuum model acts as a magnifying glass into the molecular
level of the material.
The basic concept to connect both scales of simulation is illustrated in
Fig. 7. The model system is a periodic box described by a continuum,
tessellated to obtain finite elements, and containing an atomistic inclusion. Any overall strain of the atomistic box is accompanied by an identical
strain at the boundary of the inclusion. In this way, the atomistic box does
not need to be inserted in the continuum, or in any way connected (e.g., with
the nodal points describing the mesh at the boundary of the inclusion). This
coupling via the strain is the sole mechanism to transmit tension between the
continuum and the atomistic system. The shape of the periodic cells is
described by a triplet of continuation (column) vectors for each phase (see
also [21]), A, B, and C for the continuous body, with associated scaling
matrix H [ABC], and analogously a, b, and c for the atomistic inclusion,
with the scaling matrix h [abc] (see Fig. 8).
The material morphology is specified by a set of nodal points in the
continuum description. The inclusion boundary is defined by a mesh of
vertices xbi (b for boundary). The exterior of the inclusion contains the
vertices xci (c for continuum). Inside the atomistic system, the (affine)
transformations obtained by altering the scaling matrix from h0 to h can be
expressed by the overall displacement gradient tensor matrix Mh hh1
0 .
The Lagrange strain tensor [40] of the atomistic system is then
1
" MT M 1
2
12
As is convenient and customary in infinite periodic systems, scaled coordinates are employed for the continuum part as well as for the atomistic part
(see Fig. 8). In the continuum part, sbi H1 xbi and sci H1 xci of the nodal
points are used as degrees of freedom (the natural state variables of the
continuum are, thus, sbi , sci , and H). Similarly, the scaled coordinates sai of
504
Binder et al.
FIG. 7 The key idea of the atomistic-continuum model is to connect the inclusion
boundary behavior and the atomistic cell behavior via the strain transformation they
undergo. The strain behavior of the atomistic box is identical to the strain behavior
of the inclusion boundary.
atom i are related to the Cartesian coordinates through the scaling matrix,
sai h1 xai .
Since, as stated above, our model [20] requires that an overall strain of
the atomistic box is accompanied by an identical strain at the boundary of
the inclusion, we constrain the vertices xbi of the inclusion boundary and the
atomistic scaling matrix h to behave in concert (the vertices xci in the matrix
are not affected) according to xbi h Mhxb0i , where xb0i are the original
coordinates of the boundary vertices i. Note that the coordinates xbi h are
no longer degrees of freedom, because they depend on the scaling matrix h.
Changes in the system cell matrix H act as homogeneous deformations of
the entire continuous system [20] and the nodal points of the continuum and
the atomic positions in the inclusion should be displaced accordingly. This
Coarse-Graining Techniques
505
FIG. 8 Sketch of the four types of degrees of freedom used in the atomisticcontinuum model: H, DH, and the scaled coordinates sc and sa of the nodes in the
continuum and the atoms in the atomistic phase. The system unit cell matrix H and
the scaled coordinates sc and sb entirely determine the finite-element configuration of
the continuum. h and sa are necessary to determine the atomistic configuration. The
nodal points on the inclusion boundary sb and the atomistic scaling matrix h are
calculated from the variables H and DH.
13
1X
V " T C "
2
14
506
Binder et al.
Coarse-Graining Techniques
507
IV. CONCLUSIONS
The recently proposed atomistic-continuum model is useful for modeling the
deformation of three-dimensional solids. In this approach, an inclusion of
arbitrary shape is surrounded by a continuum, which is modeled by a
variable-metric finite-element method. The inclusion behavior is directly
driven by that of an atomistic system modeled in detail by molecularmodeling techniques. In this manner, a solution is given to the problem of
finding a constitutive description for a material. The atomistic-continuum
model comprises four types of degrees of freedom, the appropriate combination of which permits calculation of the configuration of both the
atomistic system and the mesh elements of the continuum. The behavior of
both levels of modeling is described by their corresponding potential C- and
kinetic-energy functions which are added to give the total energy of the
system.
The major drawback of the atomistic-continuum approach lies in the
limitation to uniform strain on the level of the atomistic periodic box and,
hence, the inclusion perimeter. Crack propagation, for instance, cannot be
treated with this method.
A partial compensation for this weakness might lie in the flexibility of
the atomistic-continuum concept. The approach can be extended, in
principle, to
.
508
Binder et al.
other overall properties than stress and strain; among the properties that can be simulated are those describable by scalar fields in a
continuum [39], such as dielectric constants, thermal conductivity,
electrical conductivity, solute permeability, etc.
ACKNOWLEDGMENTS
We thank Jorg Baschnagel and Andrei A. Gusev for many illuminating
discussions and Michael Rozmann and Volker Tries for their help and
kind support. S. S. and U. W. S. also gratefully acknowledge the financial support provided by the Swiss National Science Foundation
(Schweizerischer Nationalfonds zur Forderung der wissenschaftlichen
Forschung). K. B. and W. P. acknowledge support from the Bundesministerium fur Bildung und Forschung (BMBF, Germany) under Grants
No 03N8008C and 03N6015.
REFERENCES
1. Binder, K. Macromol. Chem.: Macromol. Symp. 1991, 50, 1.
2. Baschnagel, J.; Binder, K.; Doruker, P.; Gusev, A.A.; Hahn, O.; Kremer, K.;
Mattice, W.L.; Muller-Plathe, F.; Murat, M.; Paul, W.; Santos, S.; Suter, U.W.;
Tries, V. Adv. Polymer Sci. 2000, 152, 41.
3. Kremer, K. Multiscale problems in polymer simulations. In Multiscale
Simulations in Chemistry and Physics, Proceedings NATO ARW, Eilat/Israel,
April 58, 2000, Brandt, A., Bernholc, J., Binder, K., Eds.; IOS Press:
Amsterdam, 2001.
4. Baschnagel, J.; Binder, K.; Paul, W.; Laso, L.; Suter, U.W.; Batoulis, J.; Jilge,
W.; Burger, T. J. Chem. Phys. 1991, 95, 6014.
5. Paul, W.; Binder, K.; Kremer, K.; Heermann, D.W. Macromolecules 1991, 24,
6332.
6. Carmesin, I.; Kremer, K. Macromolecules 1988, 21, 2819.
7. Deutsch, H.-P.; Binder, K. J. Chem. Phys. 1991, 94, 2294.
8. Paul, W.; Pistoor, N. Macromolecules 1994, 27, 1249.
9. Tries, V.; Paul, W.; Baschnagel, J.; Binder, K. J. Chem. Phys.1997, 106, 738.
Coarse-Graining Techniques
509
510
Binder et al.
16
CONNFFESSIT: Simulating
Polymer Flow
MANUEL LASO
I.
INTRODUCTION
ttinger
Laso and O
512
A. Some Definitions
Throughout the succeeding sections some terms are used frequently,
sometimes in a nonstandard way. The following definitions are included
in order to avoid confusion:
.
.
.
.
.
Domain: a region of d-dimensional space in which we seek the solution to the equations of conservation of mass and momentum coupled
to a description of the fluid (constitutive equation).
Cell: in d dimensions, the simplest d-dimensional geometric object in
which the domain is subdivided.
Element: synonymous with cell.
Finite element: a cell or element augmented with nodal points and
shape functions.
Particle: the smallest entity of the micro calculation that does not
directly interact with other similar entities and which contributes to
the stress; it can be as simple as a Hookean dumbbell, or as complex
as a whole Brownian Dynamics simulation of a colloidal dispersion
under periodic continuation or boundary conditions.
CONNFFESSIT
.
.
.
513
Finally, throughout this work, and unless stated otherwise, the term
microscopic refers either i) to simplified mechanical models for polymer
molecules in which no chemical detail is explicitly present or ii) to models for
particulate fluids in which the individual particles are explicitly treated and
in which the solvent is treated stochastically in an average sense and not
resolved in individual molecules. This usage of the term microscopic is
adopted here in order to be consistent with the meaning of the term micro
as used in the literature on methods that combine continuum mechanics
with a more detailed description of the flowing material. Strictly speaking,
models falling into category i) are more microscopic than those under ii).
Atomistic and ab initio (density functional) polymer modeling reside at
increasingly more fundamental levels.
A. Computational Rheology
Although it is young, the field of continuum-mechanical computational
rheology has a lively and eventful history. What was initially expected to be
little more than an extension of Newtonian CFD methods to fluids with
more complex but in principle harmless constitutive equations turned out to
be a source of frustration and of unexpectedly challenging mathematical
problems in spite of the vast store of knowledge on numerical techniques
for Newtonian fluids that had been accumulating during the past
half century. The analysis and solution of what has come to be known
as the high Weissenberg-number* problem have engaged the attention of
*Rheologists resort to dimensionless groups such as the Weissenburg and Deborah numbers
and the stress ratio SR to characterize the elasticity of the ow. High WS corresponds to high
uid elasticity.
ttinger
Laso and O
514
a large number of both applied mathematicians and engineers and spurred a great amount of analytical and numerical diagnostic work. The
book by Crochet et al. (1984) and the review articles by Crochet (1989)
and Keunings (1989) give comprehensive if slightly dated overviews of
the field; current research in computational rheology is reported in
Moldenaers and Keunings (1992), Gallegos et al. (1994), and Ait-Kadi et
al. (1996).
For the purposes of the present work the current situation in nonNewtonian CFD can be summarized as follows:
.
CONNFFESSIT
515
C. Particle Methods
Besides the Finite-Difference and Finite-Element mainstream, there has
appeared a handful of alternative computational methods in Newtonian and
non-Newtonian CFD that can be collectively characterized as particulate,
in the sense that they involve discrete computational entities in addition
to or in lieu of the standard continuum-mechanical discretization devices.
In some of these methods (the first four and the last one in the following
ttinger
Laso and O
516
As early as 1973 Chorin (1973, 1989, 1994) introduced the twodimensional random vortex method, a particle method for the solution of the NavierStokes equations. These particles can be thought
of as carriers of vorticity. Weak solutions to the conservation equations are obtained as superpositions of point vertices, the evolution of
which is described by deterministic ODEs. A random walk technique
is used to approximate diffusion, and vorticity creation at boundaries
to represent the no-slip boundary condition. The extension to three
dimensions followed in 1982 (Beale and Majda 1982). An important
improvement in stability and smoothness was achieved by Anderson
and Greengard (1985) by removing the singularities associated with
point vertices. Anderson and Greengard (1988) and Marchioro and
Pulvirenti (1984) have written comprehensive reviews of the method.
Smoothed-Particle Hydrodynamics (SPH) was introduced by Gingold
and Monaghan (1977) and Lucy (1977). SPH is a three-dimensional
free-Lagrange algorithm that in its most basic form is truly grid-free.
SPH treats fluid elements as extended clouds of material; their centers
of mass move according to the conservation laws of fluid mechanics.
It was first applied to the solution of complex astrophysical problems,
such as colliding planets and stars. In its most widely used version,
SPH is based on integral interpolants for information at nonordered
points. The fluid is represented by spherically symmetric particles,
each having a mass that has been smoothed out in space with a
density distribution given by a suitable interpolating kernel or distribution. Ordinary differential equations for the smoothed quantities at
the location of the particles are obtained by multiplying the original
governing equations by the kernel and integrating over the domain. A
critical evaluation of the quantitative performance of SPH has been
presented in Cloutman (1991). The method seems to be most useful
for exploratory studies where modest accuracy is acceptable.
Another technique, widely used in a different context (kinetic theory,
plasma simulation) is the so-called Vortex-in-Cell Method (Hockney
and Eastwood 1981, Birdsall and Langdon 1985). In the Vortex-inCell Method a two-dimensional computational domain is divided into
cells and the vorticity is counted in each cell. The Poisson equation
relating the vorticity and the Laplacian of the stream function is
subsequently solved on the grid by an FTT algorithm, the velocity
CONNFFESSIT
517
518
ttinger
Laso and O
CONNFFESSIT
519
FIG. 1
ttinger
Laso and O
520
FIG. 2
CONNFFESSIT
521
522
ttinger
Laso and O
Special care of empty elements (if any) must be taken and the array
FIRST created:
DO I 2, NUMEL 1
IF (LAST(I) .EQ. 0) LAST(I) LAST(I-1)
ENDDO
COUNT(1) LAST(1)
DO I 2, NUMEL 1
COUNT(I) LAST(I) - LAST(I-1)
ENDDO
DO I 1, NUMEL 1
FIRST(I) LAST(I) 1 - COUNT(I)
ENDDO
CONNFFESSIT
523
This compact algorithm performs the required task and is fully vectorizable. The indexing array and the indirect addressing capability offered
by the information contained in the FIRST and LAST pointer arrays (above)
act de facto as an interface that makes the particles appear as if they actually
were contiguous:
The crucial point of this procedure is that it relies on efficient sorting for
restoring contiguity. As we will see in Section IV.B.1, particle index sorting
using the binary radix sort has optimal O(Npr logNpr) complexity. Besides,
indirect addressing is fully supported (and in many cases hardwareexecuted) by all vector computers. Thus, the computational overhead
ttinger
Laso and O
524
C. Point-Inclusion Algorithm
The second key operation lying at the heart of a CONNFFESSIT calculation consists in determining if a particle, the spatial coordinates of
which are known, is located within or outside an element. This seemingly
minor task acquires immense importance in more than one dimension,
since it has to be executed an inordinate number of times: it is almost
always the single most-frequently performed basic operation invoked
in CONNFFESSIT.
Generalized point-inclusion algorithms are the object of much research
in computational geometry (Mulmuley 1994, Preparata and Shamos
1990); the goal of research in this field is the creation of optimal
algorithms with respect to the number of sides of the polygon, i.e., how to
optimally determine whether or not a point is included in a polygon of a
very large number of sides. An extensive search of available literature
showed that, somewhat surprisingly, the question of how to determine
whether a point is included in a triangle or quadrilateral in the fastest
possible way is amply ignored. The only algorithm found in the literature
turned out to be a geometrically elegant one frequently used in computer
graphics and based on the evaluation of a vector product for each of
the sides but which can certainly not compete in speed with the following
procedure (Fig. 3).
During the initialization phase, before the start of the actual timemarching computation:
.
.
For subsequent use, i.e., in order to determine if the lth particle at (xl, yl)
is located within the kth element:
.
k kj
Evaluate sign fjk akj
i , xl , yl ; if and only if sign fj ai , xl , yl
k kj
j
j
and sign fj ai , Xc , Xc are equal for all j, is the particle within the
element.
CONNFFESSIT
525
FIG. 3 Particle-inclusion algorithm in two dimensions. The same scheme is applicable to cells with higher number of edges or in higher dimensions.
*Since the particle fails the inclusion test as soon as a dierence in sign is detected.
526
ttinger
Laso and O
CONNFFESSIT
527
*A static mesh remains the same throughout the calculation, both the coordinates of the nodal
points and the numbering of the cells.
528
ttinger
Laso and O
the mesh, typically an order of magnitude larger than the hardware vector
length:
DO I 1, NUMEL
J1
DO ICORNER 1, 4
DO II 1, NUMEL
IF (II .NE. I) THEN
DO IICORNER 1, 4
IF (VERTEX(II, IICORNER) .EQ. VERTEX(I, ICORNER)) THEN
REPEAT .FALSE.
DO ICHECK 1, J
IF (II .EQ. NEI1(ICHECK, I)) THEN
REPEAT .TRUE.
ENDIF
ENDDO
IF (.NOT. REPEAT) THEN
NOFNEI1(I) NOFNEI1(I) 1
NEI1(NOFNEI1(I), I) II
ENDIF
J MAX(NOFNEI1(I), 1)
ENDIF
ENDDO
ENDIF
ENDDO
ENDDO
ENDDO
CONNFFESSIT
529
REPEAT2 .TRUE.
ENDIF
ENDDO
IF (.NOT. REPEAT2) THEN
DO IICORNER 1, 4
IF (VERTEX(II,IICORNER).EQ.
VERTEX(ITEST,ICORNER)) THEN
REPEAT .FALSE.
DO ICHECK 1, J
IF (II .EQ. NEI2(ICHECK, I)) THEN
REPEAT .TRUE.
ENDIF
ENDDO
IF (.NOT. REPEAT) THEN
NOFNEI2(I) NOFNEI2(I) 1
NEI2(NOFNEI2(I), I) II
ENDIF
J MAX(NOFNEI2(I), 1)
ENDIF
ENDDO
ENDIF
ENDIF
ENDDO
ENDDO
ENDDO
ENDDO
The slightly greater complication of this procedure stems from the need
to avoid double counting the neighbors of the first neighbors. The reward
for the more complex code is that again, in spite of the deep nesting, the DO
over the number of cells is well vectorizable and by far outperforms the
recursive construct.
The lists of first neighbors (NEI1) and second neighbors (NEI2)
generated in this way can be used as is for the task of relocating the
particles that have flowed out of a given cell. The search needs only be
performed in the cells contained in the neighbor lists.
However, if the velocity field is known, the search can be further sped
up by ordering the neighbors so that those that are downstream of the given
cell are searched first. This is the idea behind the velocity-biased neighbor
list. The natural way to implement it is to sort the list of neighbors in order
of increasing absolute value of the angle formed between the velocity vector
at some characteristic point of the central cell (e.g., the centroid) and
the vector joining the centroid of the central cell with the centroid of
a neighboring cell (angle in Fig. 4). Thus, for deterministic particle
530
ttinger
Laso and O
trajectories, all particles that have left a cell will necessarily be located in
downstream cells; during the search only a small fraction of the neighbor list
(seldom more than the first three elements) needs to be scanned. The scheme
just presented is equally valid for stochastic particle trajectories, although
due to their inherent variability it is no longer guaranteed that particles
leaving a cell will be located in cells downstream. However, given the
smallness of the typical displacement per step with respect to the cell size,
the velocity-biased neighbor list should be just as efficient as in the
deterministic case.
As a final remark it should be said that although the procedures for first
and second neighbor lists just presented fully exploit vectorizability, they are
still ONel2 algorithms. Their use is justified for fixed meshes because they
are straightforward, well vectorizable, and have to be invoked only once.
For moving meshes, however, the neighbor lists may have to be generated at
every integration step. In this case, ONel2 is unacceptable. As a matter of
fact, the experience gathered so far shows that, unless the neighbor list
problem is dealt with efficiently, neighbor list generation and the subsequent
particle location will become the limiting steps in large-scale calculations
(Section IV.B).
CONNFFESSIT
531
532
ttinger
Laso and O
calculations for integral constitutive equations. Figure 5 shows the trajectory of a particle within the recirculation region in the 4:1 contraction
geometry obtained by tracking it in a constant steady-state velocity field for
an Oldroyd-B model.
In spite of the use of a special type of macroelement (the MDF Finite
Element of Nagtigall et al. 1979, Mercier 1979, Bernstein et al. 1985) which
allows analytical tracking, the result of the cumulation of round-off errors in
an otherwise exact technique is striking. Given the low velocities and shear
rates in this particular instance, the consequences of tracking errors are very
minor.
However, for flows with strong recirculation, the erratic behavior of the
discretized trajectories leads to a rapid depletion or enrichment of particles
in certain regions. For sufficiently strong recirculation rates and sufficiently
long integration times, the spurious migration of particles with the wrong
flow history into the wrong elements and the spurious depletion/enrichment
effect render calculations meaningless. The only remedy for this ailment is
the use of smaller time steps or higher-order numerical schemes for ODEs,
or both.
The use of higher-order schemes is often the preferred solution, in
spite of requiring several function evaluations per time step for each of
CONNFFESSIT
533
the particle trajectories being integrated. This additional effort may pay
back since:
.
.
Trajectory integration does not constitute a large fraction of the overall computational load.
Relatively inexpensive velocity-field updates can be performed independently of and between two consecutive particle configuration
updates with higher frequency than the latter if so required.
The optimal order of the tracking algorithm is thus determined by the ratio
of particle tracking to particle simulation.
Figure 6 shows a comparison of a trajectory integration similar to that
depicted in Fig. 5, but using first-order Euler (Fig. 6a), second- order, fixedstep trapezoidal rule (Fig. 6b), and 12th-order predictor-corrector Adams
Bashforth with self-adaptive step, (Fig. 6c) schemes for a fixed macroscopic
velocity field. While the increase in order does result in increased stability,
even the phenomenally accurate 12th-order scheme fails to yield closed-loop
trajectories. For simple dynamic models including a few degrees of freedom
(Oldroyd-B, CurtissBird) on average-size meshes (a few thousand cells)
with simple elements (P1-C0) and for the current hardware generation,
the best ratio of tracking accuracy to computational time is obtained
for second-order, fixed-step explicit schemes. Higher-order schemes are
advantageous only for very smooth velocity fields, where a very large
integration step can be employed.
The noisy character of micro/macro methods makes it very unlikely that
higher-order schemes will significantly increase overall performance. The
same applies to richer FEs for which velocity updates are increasingly more
expensive.
The second effect of discretization and truncation is not related to the
long-time cumulation of errors, but makes itself observable at every time
step, as particles close to impenetrable boundaries (which can only be
approached asymptotically by trajectories) actually cross them due to a
finite time integration step. In general, the existence of a nonzero outwardpointing component of the velocity of a particle leads to its leaving the
integration domain if the particle is sufficiently close to the boundary.
This effect is in general of an order of magnitude commensurate with the
time step: it acts on a fraction of all particles given by the ratio of the
perimeter @ (or bounding surface in 3D) of the domain times the average
outward-pointing velocity component times the integration step to the total
area (or volume in 3D) of the domain . This is a small ratio for any
nonpathological domain shape. In most cases, the fraction of particles
that leave the integration domain at any given time step due to the
discretization error produced by a first-order explicit Euler scheme with
534
ttinger
Laso and O
FIG. 6 Effect of order of tracking algorithm on closedness of trajectories: (a) firstorder Euler, (b) trapezoidal rule, (c) 12th-order adaptive AdamsBashforth PC.
CONNFFESSIT
535
reasonable t lies between 104 and 5 103 and is therefore of no consequence for the numerical accuracy of the solution. Particles that are not in
the region close to the boundary do not exit the domain. Their trajectories
will drift, however, towards or away from the boundary at a rate dictated
both by the local outward- or inward-pointing velocity component and by
the discretization error.
Figure 7 shows this effect for three particles close to the reentrant corner
in the contraction geometry for P1-C0 quadrilaterals. The closer the initial
position of the particle to the separation line, the more difficult it is to turn
around the corner. Low-order algorithms lead to particle trajectories that
intersect impenetrable boundaries.
Unfortunately, even if only a small fraction of all particles leave the
domain at each time step, these events cannot be ignored. Even though they
have a negligible impact on the numerical values of the stresses computed as
element-wise averages, if a fraction of O(104103) of the total number of
particles leave the domain at each time step due to discretization and
particles drift toward the boundary at a comparable rate, it does not take
more than a few thousand steps for a large fraction of particles to disappear
from the computational domain.
The most straightforward solution to this question consists in reflecting
back into the domain all lost trajectories, i.e., those that have left it at a
given time step. This reflection is acceptable and does not result in a
reduction of the order of the algorithm. However, the crucial aspect of
the problem is that it utterly destroys the vectorizable character we strive
for, since such particle-exit events occur in a geometrically localized and
structured region (the boundary region), but purely at random when considered from the viewpoint of the particle numbering scheme. Hence the
need to develop a scheme to recover and replace such lost particles.
An efficient method to handle this difficulty and retain vectorizability
must: i) make use of the numbering schemes, neighbor lists, and pointinclusion techniques explained in Section III, in order to efficiently
determine which particles have left the domain and where they are located
and ii) obtain this information within and extract it from a loop for
subsequent processing without jeopardizing vectorizability. An optimal twostep algorithm works in the following way:
.
Determine the indices of all particles that have left the element in
which they were located at the previous time step. This is performed
in the double-loop construct already explained in Section III.D using
indirect addressing to point at all particles (inner loop in the following
FORTRAN excerpt) within a given element (outer loop). If a particle
has left the element it was in, add its address to the list of particles
536
ttinger
Laso and O
CONNFFESSIT
537
The second double loop finds which particles are in first neighbors
and spots those particles that have left the domain:
LOST 0
DO I 1, COUNTLEFT1
IHOLD LEFT1(I)
DO II 1, NOFNEI1(WHERE(IHOLD))
III NEI1(II, WHERE(IHOLD))
IF (particle IHOLD is in first neighbor element III)
THEN
WHERE(IHOLD) III
GOTO 10
ENDIF
ENDDO
LOST
LOST 1
NOTFOUND(LOST) IHOLD
10 CONTINUE
ENDDO
In the second step a simple pass is performed over all particles that
have left the domain in order to reflect their positions back into the
domain.
DO I 1, LOST
IHOLD NOTFOUND(I)
ttinger
Laso and O
538
ELEMENT WHERE(IHOLD)
reflect position of particle back into domain
ENDDO
This would seem to complete the procedure, since all particles must
have either been located in nearest neighbors or they must have been
reflected back into the domain. We are, however, not done yet: any
reflection procedure has a loophole which is a consequence of the
finite integration time step: if a particle leaves the domain through
an element boundary that is close to an external nodal point (see Figs.
8 and 9), there will always exist the possibility that the position after
reflection is still outside of the domain or within the domain but not
in the element the particle was originally in.
These two possibilities make it necessary to perform a third and
last pass (not explicitly presented here because of its trivial vectorizability) in order to check that all reflected particle positions are
indeed within the domain and within the intended element.
The root of both of these failures is that, within the discretization
error, it is not possible to decide whether the particle moved from
ri(tj) to ri(tj1) outside the domain directly from element n or through
element n 1. The assignment of a new position to the lost particle is
therefore arbitrary, but since the number of such failures has always a
higher error order, it is irrelevant where the position of such particles should be reinitialized, apart from obviously poor choices like
ri(tj) ri(tj 1).
FIG. 8
CONNFFESSIT
FIG. 9
539
540
ttinger
Laso and O
CONNFFESSIT
541
ttinger
Laso and O
542
Left shifting the vertex index by NBITS and bitwise ORing it with the cell
number, produces an array VC, the elements of which contain:
<- - - - - - - - - - - - - - one word - - - - - - - - - - - - ->
0
vertex number
cell number
vertex number
cell number
vertex number
cell number
...
...
...
vertex number
cell number
Starting with the rightmost bit of the word to be sorted, i.e., with each
element of VC, determine whether the corresponding bit position in
each word is a 0 or a 1. Compress all numbers that have 0 as the
rightmost key-bit in a second scratch array of the same length.
Following this, compress out all numbers that have 1 as the rightmost
key-bit into the remainder of the scratch array. Finally, copy the
scratch array back into the original array. Both the scratch array
and the original array now contain all the words or numbers in
the original array, sorted by their least significant bit in ascending
order.
Proceed in the same fashion with the second least significant bit of
each element in the array: all numbers with 0 as the second-rightmost
key-bit are gathered into the second array. Following this, compress
out all numbers with 1 as the second-rightmost key-bit into the
remainder of the scratch array and subsequently copy the scratch
array back into the original array.
The original array now contains all the numbers sorted on their
rightmost two key-bits; first will be all numbers with rightmost bits
00, followed by all numbers with rightmost bits 01, followed by all
numbers with rightmost bits 10, followed by all numbers with
rightmost bits 11. This process continues from right to left until the
numbers have been sorted on each of the key-bit positions.
The entire array is now sorted with the keys in ascending order. In the
following tables we see the way the binary radix sort operates on a sequence
CONNFFESSIT
543
Although in this example the binary radix sort operates on all bits of
a word it is, however, perfectly possible to sort words or numbers based
on only a few or just one of its bits by applying the bit sort to these bit
positions only.
The following simple implementation of the binary radix sort vectorizes
well on all vector architectures and extraordinarily well on processors where
544
ttinger
Laso and O
*btest(N,K) tests the Kth bit of argument N. The value of the function is .TRUE. if the bit is
1 and .FALSE. if the bit is 0.
y
Even so, while not the most ecient available, this routine runs at a good fraction of the
maximum possible speed for this algorithm on the Cray Y-MP and the NEC SX-3.
CONNFFESSIT
545
with each of them appearing up to four times (for quadrilateral cells) and
becomes an O(Nel log Nel) process.
Most vertices are connected to four cells and have four entries in the
array. Those on the edges of the mesh, however, will have fewer. Vertices in
the corners of the grid (if any) may have only one. It is now necessary to
determine where in the sorted array VC each vertex begins and ends. The
following code finds each place where the vertex number changes and enters
this information into two index arrays: ILASTI, giving the last index into
the array VC for each vertex, and IGATHR, which initially is set to the first
index minus one for each vertex. IGATHR is used to extract successive cells
connected to each vertex:
ENTRY 0
DO I 1, INDEX - 1
IF (IBITS(VC(I),NBITS,NBITS) .NE. IBITS(VC(I1),NBITS,
NBITS)) THEN
ENTRY ENTRY 1
ILASTI(ENTRY) I
ENDIF
ENDDO
ENTRY ENTRY 1
ILASTI(ENTRY) INDEX
IGATHR(1) 0
DO I 2, ENTRY
IGATHR(I) ILASTI(ENTRY-1)
ENDDO
The next step consists in building up a four-column table: for each separate vertex enter one of the connected cells in a separate column. If a vertex
is shared by fewer than four cells, repeat the last cell number for that vertex
as necessary. In a last pass, any such duplicated cell numbers are set to zero:
DO I 1, 4
DO J 1, ENTRY
IF (IGATHR(J) .LT. ILASTI(J)) SIGATHR(J) IGATHR(J)1
CTABLE(J) IBITS(VC(IGATHR(J), 0, NBITS))
ENDDO
ENDDO
DO I 4, 2 -1
DO J 1, ENTRY
IF (CTABLE(J, I) .EQ. CTABLE(J, I-1)) THEN CTABLE(J, I) 0
ENDDO
ENDDO
For each cell number, the only nearest neighbor cells are those that occur
in one or more rows along with it in the array CTABLE. The following code
ttinger
Laso and O
546
generates a table CELNBR that contains all cell-neighbor pairs and only cellneighbor pairs:
PAIRS 0
DO I 1, 4
DO J 1, 4
IF (I .NE. J) THEN
DO K 1, ENTRY
IF ((CTABLE(K,I) .NE. 0) .AND. (CTABLE(K,J) .NE. 0)) THEN
PAIRS PAIRS 1
CELNBR(PAIRS)
IOR (LSHIFT(CTABLE(K,I),NBITS),CTABLE(K,J))
ENDIF
ENDDO
ENDIF
ENDDO
ENDDO
cell number
neighbor
cell number
neighbor
cell number
neighbor
...
...
neighbor
cell number
neighbor
There will now be some duplicates since some cell pairs (those where two
cells share more than one vertex) will appear in CELNBR twice. Additionally,
since the cell numbers will be in no particularly useful order, sort the array
CELNBR using the cell number in each entry as the key (in the middle of the
word).
The sorted CELNBR contains at most twelve neighbor entries for each
cell.* A neighbor that shares only a corner appears once, a neighbor that
shares two corners (an edge) appears twice. Now eliminate the duplicates by
sorting on the entire cell-neighbor pair rather than on the cell number only.
But if we take into account that duplicate entries can be no more than eleven
words apart (in the event they are first and last in the neighbor list), we can
*In the case of a structured mesh, where the number of neighbors is xed.
CONNFFESSIT
547
replace this sorting by the more elegant device of comparing each entry in
the array, as sorted on cell number only, with the previous eleven entries:
DO I 1, 11
DO J I1, PAIRS
IF (CELNBR(J) .EQ. CELNBR(J-I)) CELNBR(J) 0
ENDDO
ENDDO
UNIQUE 0
DO I 1, PAIRS
IF (CELNBR(I) .NE. 0) THEN
UNIQUE UNIQUE 1
CELNBR(UNIQUE) CELNBR(I)
ENDIF
ENDDO
The last step consists in finding the first and last entries for each cell
number, which can be done in exactly the same fashion as for the array VC
sorted on vertex number and will therefore not be repeated here. After that,
reformatting the table of nearest neighbors into whatever format we wish is
a trivial matter.
We thus see that while a frontal attack results in the elegant and
straightforward but ONel2 algorithm of Section III.D, recasting the
problem as a quite convoluted sequence of sorting and indirect addressing
yields an O(Nel log Nel) process and therefore optimal.* The break-even
point between the two algorithms will obviously depend on the particular
mesh being used, in particular on the ratio of boundary cells to internal cells.
The figures in Table 1 apply to the two-dimensional eccentric cylinder
meshes.
TABLE 1 Peformance Comparison for Increasingly Refined Meshes
Mesh
M0
M1
M3
M4
M5
Number
of vertices
Number
of cells
CPU-time ratio
(Cray Y-MP):
algorithm section III.D/
algorithm section IV.B.1
55
220
840
1680
16800
50
200
800
1600
16000
0.81
2.5
8.3
25
130
CPU-time ratio
for scalar processor
0.75
2.2
8.0
23
110
*The algorithm just described is a constructive proof of the equivalence of building a neighbor
list and a search.
ttinger
Laso and O
548
We thus see that the break-even point lies at about a few tens of cells, i.e.,
except for the coarsest grids, which in practice never occur, the algorithm
just presented is significantly more efficient. For realistic 2D computations
and, above all in three dimensions, where grid sizes of O(104105) cells are
standard, the algorithm of Section IV.B.1 is at least two orders of magnitude
faster and the only viable alternative. Finally, it is worth mentioning that the
growing CPU-time ratios apply to absolute CPU-times that in turn grow
with problem size.
Interestingly, the effort invested in developing vectorizable algorithms
pays off on scalar architectures as well. The CPU ratio between the nestedloop algorithm and the algorithm just presented is slightly more favorable
for the latter on vector machines since the nested DO loop constructs of
Section III.D contain comparatively more or deeper IFs and low-count
DOs that must be unrolled.
2. Particle Relocation
As we saw in IV.B, the problem of determining in which element a given
particle is located after the mesh has been created, renumbered, or modified
is of O(Npr)O(Nel), i.e., quadratic in problem size, instead of O(Npr) as is the
case when the mesh is static and neighbor lists can be used. The reason is
that knowledge of the neighboring cells is purposeless if the location of the
particle at the previous time step is unknown, as is the case when the domain
is reenmeshed.* A partial remedy for the O(Npr)O(Nel) complexity could
be achieved by reenmeshment procedures that keep renumbering to a
minimum. Although the amount of literature on reenmeshing procedures is
overwhelming, little attention has been given to the question of minimumrenumbering algorithms.
While the quadratic complexity of particle relocation is unavoidable the
very first time a domain is initialized,y it is possible in practice to bypass this
complexity in subsequent steps. The reconstructed neighbor list can be used
in order to avoid brute-force particle location no matter how extensive the
renumbering is. The two-step idea is straightforward and does not require
pseudo-code to illustrate it:
.
*Even if the reenmeshment would only renumber cells, the new cell number would bear no
relationship whatsoever with the number before the reenmeshment.
y
Even in this case, particle relocation can be avoided by systematic cell ll-up, resulting
in O(Npr).
CONNFFESSIT
549
ttinger
Laso and O
550
APPENDIX A: SYMBOLS
akj
i
fjk
fnl,. . .
l
n
n
Nel
Np
Npr
Nsim
Nver
Nsides
Q
ri(t)
T
t
t
u
u
Ws
Xij
Yij
Xcj
Ycj
xi
yi
zi
d
s
CONNFFESSIT
551
APPENDIX B: ABBREVIATIONS
CE
CFD
FE
MC
MD
NEMD
ODE
PC
PDE
SDE
SPH
WR
Constitutive Equation
Computational Fluid Dynamics
Finite Elements
Monte Carlo
Molecular Dynamics
Non-Equilibrium Molecular Dynamics
Ordinary Differential Equation
Predictor-Corrector
Partial Differential Equation
Stochastic Differential Equation
Smoothed-Particle Hydrodynamics
Weighted Residuals
REFERENCES
Ait-Kadi, A.; Dealy, J.M.; James, D.F.; Williams, M.C., Eds., Proceedings of the
XIIth International Congress on Rheology, Laval University, Quebec City,
1996.
Allen, M.P., Tildesley, D.J. Computer Simulation of Liquids; Clarendon Press:
Oxford, 1987.
Anderson, C.; Greengard, C. SIAM J. Math. Anal. 1985, 22, 413440.
Anderson, C.; Greengard, C., Eds., Vortex Methods; Springer: Berlin 1988.
Ashurst, W.T.; Hoover, W.G. Phys. Rev. A 1975, 11, 658667.
Baaijens, F.P.T. J. Non-Newtonian Fluid Mech. 1998, 75, 119138.
Balasubramanian, K.; Hayot, F.; Saam, W.F. Phys. Rev. A 1987, 36, 22482252.
Baudet, C.; Hulin, J.P.; Lallemand, P.; dHumieres, D. Phys. Fluids A 1989, 1,
507518.
Beale, J.T.; Majda, A. Math. Comp. 1982, 39, 127.
Bell, T.W.; Nylond, G.H.; de Pablo, J.J.; Graham, M.D. Macromolecules 1997, 30,
18061812.
Bernstein, B.; Malkus, D.S.; Olsen, E.T. Int. J. Num. Meth. Fluids 1985, 5,
4370.
Bird, R.B.; Curtiss, C.F.; Armstrong, R.C.; Hassager, O. Dynamics of Polymeric
Liquids; John Wiley: New York, 1987b; Vol. II.
Birdsall, C.K.; Langdon, A.B. Plasma Physics via Computer Simulation; McGrawHill: London, 1985.
Bonvin, J.C.; Picasso, M. J. Non-Newtonian Fluid Mech. 1999, 84, 191215.
552
ttinger
Laso and O
CONNFFESSIT
553
554
ttinger
Laso and O
BIBLIOGRAPHY
Ackerson, B.J.; Pusey, P.N. Phys. Rev. Lett. 1988, 61, 10331036.
Aggarwal, R.; Keunings, R.; Roux, F.-X. J. Rheology 1994, 38 (2), 405419.
Akl, S.G. Parallel Sorting Algorithms; Academic Press: New York, 1985.
Allen, M.P. Simulations and phase behaviour of liquid crystals. In Observation,
Prediction and Simulation of Phase Transitions in Complex Fluids, Baus, M.,
Rull, L.F., Ryckaert, J.-P., Eds.; NATO ASI Series, 1995; 557591.
Ballal, B.Y.; Rivlin, R.S. Trans. Soc. Rheol. 1976, 20, 65101.
Bartlett, P.; van Megen, W. Physics of Hard-Sphere Colloidal Suspensions, Mehta, A.
Ed.; Springer: Berlin, 1994.
Beris, A.N.; Armstrong, R.C.; Brown, R.A. J. Non-Newtonian Fluid Mech. 1983,
13, 109148.
Beris, A.N.; Armstrong, R.C.; Brown, R.A. J. Non-Newtonian Fluid Mech. 1984,
16, 141171.
Beris, A.N.; Armstrong, R.C.; Brown, R.A. J. Non-Newtonian Fluid Mech. 1986,
19, 323347.
Beris, A.N.; Armstrong, R.C.; Brown, R.A. J. Non-Newtonian Fluid Mech. 1987,
22, 129167.
Bird, R.B.; Armstrong, R.C.; Hassager, O. Dynamics of Polymeric Liquids; John
Wiley: New York, 1987a; Vol I.
Bird, R.B.; Dotson, P.J.; Johnson, N.L. J. Non-Newtonian Fluid Mech. 1980, 7,
213235.
Brady, J.F.; Bossis, G. J. Fluid Mech. 1985, 155, 105129.
CONNFFESSIT
555
556
ttinger
Laso and O
Lakshmivarahan, S.; Dhall, S.K.; Miller, L.L. Parallel sorting algorithms. Advances
in Computers, 1984, 23, 295354.
Laun, H.M.; Bung, R.; Hess, S.; Loose, W.; Hess, O.; Hahn, K.;
Hadicke, E.; Hingmann, R.; Schmidt, F.; Lindner, P.; J. Rheol. 1992, 36 (4),
743787.
Lawler, J.V.; Muller, S.J.; Brown, R.A.; Armstrong, R.C. J. Non-Newtonian Fluid
Mech. 1986, 20, 5192.
Lee, S.J.; Denn, M.M.; Crochet, M.J.; Metzner, A.B.; Rigins, G.J. J. NonNewtonian Fluid Mech. 1984, 14, 301325.
Lees, A.W.; Edwards, S.F. J. Phys. C: Solid State Phys. 1972, 5, 19211929.
Lesieur, M. Turbulence in Fluids: Stochastic and Numerical Modelling; Kluwer:
Dordrecht, 1990.
Lighthill, J. An Informal Introduction to Theoretical Fluid Mechanics; Oxford Science
Publications: Oxford, 1986.
Mochimaru, Y., J. Non-Newtonian Fluid Mech. 1983, 12, 135152.
MPIF (Message Passing Interface Forum), Technical Report CS-94-230, University
of Tennessee, Knoxville, TN, USA, April 1994; also J. of Supercomputer
Applications 1994, 8, also available from:
http://www.mcs.anl.gov/mpi/mpi-report/mpi-report.html
Nagtigaal, J.; Parks, D.M.; Rice, J.R. Comp. Meths. Appl. Mech. Eng. 1974, 4,
153178.
Nickell, R.E.; Tanner, R.I.; Caswell, B. J. Fluid. Mech. 1981, 65, 219248.
Olivari, D. Introduction to the Modeling of Turbulence, Von Karman Institute for
Fluid Dynamics, 1993.
Ottinger, H.C. J. Chem. Phys. 1990, 92, 45404549.
Ottinger, H.C.; Laso, M. Bridging the gap between molecular models and viscoelastic ow calculations. In Lectures on Thermodynamics and Statistical
Mechanics; World Scientic, Singapore, 1994; Chapter VIII, 139153.
Parrinello, M.; Rahman, A. Phys. Rev. Lett. 1980, 45, 11961199.
Parrinello, M.; Rahman, A. J. Appl. Phys. 1981, 52, 71827190.
Parrinello, M.; Rahman, A. J. Chem. Phys. 1982, 76, 26622666.
Peterlin, A. Makr. Chem. 1961, 4446, 338346.
Peterlin, A. Kolloid-Zeitschrift 1962, 182, 110115.
Pironneau, O. Finite Element Methods for Fluids; John Wiley: New York, 1989.
Phan-Thien, N.; Tanner, R.I. J. Non-Newtonian Fluid Mech. 1981, 9, 107117.
Phan-Thien, N.; Sugeng, F.; Tanner, R.I. J. Non-Newtonian Fluid Mech. 1984, 24,
301325.
Powell, R.L. J. Stat. Phys. 1991, 62, 10731094.
Pusey, P.N. Colloidal suspensions. In Liquids, Freezing and Glass Transition;
Hansen, J.P., Levesque, D., Zinn-Justin, J., Eds.; North-Holland,
Amsterdam, 1991; 765942.
Rajagopalan, D.; Byars, J.A.; Armstrong, R.C.; Brown, R.S. J. Rheol. 1992, 36 (7),
13491375.
Reiner, M.; Hanin, M.; Harnoy, A. Israel J. Tech. 1969, 7, 273279.
Rigos, A.A.; Wilemski, G. J. Phys. Chem. 1992, 96, 39813986.
CONNFFESSIT
557
17
Simulation of Polymers by Dissipative
Particle Dynamics
W. K. DEN OTTER
I.
INTRODUCTION
560
and fast variables [4,5]. The strength of the DPD method lies in the softness
of the forces, and in the specific construction by which all forces act between
particle pairs following Newtons third law. Because of this latter condition,
momentum is conserved locally. Hydrodynamic interactions, therefore, arise
automatically during the simulation, so the system willwhen viewed on a
sufficiently large time and length scalebehave like any regular fluid [4].
As it turns out, surprisingly few particles are needed in practice to recover
hydrodynamic phenomena.
Hoogerbrugge and Koelman tested their method on a quiescent fluid, a
fluid undergoing Couette flow, and a fluid flow around an array of cylinders
[1]. The objects in the latter simulation were made by locally immobilizing
the liquid. But their pie`ce de resistance was calculating the viscosity of a
colloidal suspension (with rigid groups of particles acting as moving and
rotating colloids) as a function of the colloid volume fraction, in good
agreement with the celebrated Einstein expression at low fractions [2].
Research in this area has since continued in the group of Coveney and
Boek [6]. Using DPD, liquid mixtures have been simulated by the simple
expedient of assigning a color to every particle [79]. A rapid quench to
temperatures below the spinodal was then mimicked by instantaneously
increasing the repulsion between unlike particles to a value larger than that
between like particles. The scaling law for the growth of phase-separated
domains with time agreed well with the theoretical predictions for systems
with hydrodynamic interactions [79].
Polymers, the topic of this book, have been simulated by introducing
springs between the DPD particles, as in a coarse-grained bead-spring model
[10,11]. The results of these simulations are discussed in detail below. In
a related area, the rich phase behavior of aqueous surfactants has been
characterized; these molecules were modeled as a rigid unit of one
hydrophobic and one hydrophilic particle [12]. Even cell membranes [13]
and vesicles [14] have been simulated, using surfactants with flexible tails.
Thermotropic liquid crystalline phases have been characterized for a
system of rigid rods [15]. A list of all articles relating to DPD can be
found in [16].
In Section II we look more closely at the computational aspects of DPD,
before focusing attention on the specific application to polymer systems.
Section III describes the matching of simulation parameters to the
properties of real polymer systems, with an emphasis on the relation
between the conservative force field and the common FloryHuggins
parameter for mixtures. The dynamics of individual polymer chains in a
solvent and in a melt are discussed in Section IV, and the ordering dynamics
of quenched block copolymer systems is described in Section V. A summary
and conclusions are given in Section VI.
561
where rij |rij| and r^ ij rij =rij are, respectively, the length and the direction
of the difference vector rij ri rj. Here !c describes the distance dependence of the force, which is usually taken to be !c(r) (1 rij/rc), with rc the
cutoff distance of the potential. The parameter a controls the magnitude
of the force. In the case of polymers an additional conservative force f sij is
added to model the bond connecting the two particles. The force constant
and the equilibrium length of these harmonic or FENE springs are chosen
such that the average bond length coincides with the first peak in the radial
distribution function.
Between every pair of particles, whether they are bound or not, there also
acts a friction force
f dij !d rij vij r^ ij ^rij
The strength and range of these two forces are related by a conventional
fluctuation-dissipation theorem, which in the current context reads as [17]
2kB T!d rij 2 !2r rij
562
there still are no correlations between particle pairs, nor memory effects
between consecutive steps.
Note how the friction and random forces differ from their counterparts
in Brownian Dynamics (BD), where these forces act relative to a fixed
background. In BD, therefore, a bulk flow is not sustained, and the absence
of momentum conservation leads to a screening of the hydrodynamic
interactions. By making all forces explicitly Galilean invariant, as in Eqs. (1)
through (3), these deficiencies are avoided. The friction and random forces,
in addition to their physical interpretation, also serve as a thermostat to the
simulation; see Eq. (4). More elaborate models in which the particles are
equipped with an internal energy have been proposed [18,19], allowing for
a temperature gradient. This may prove a particularly valuable addition
to DPD, as it augments mass and momentum conservation with energy
conservation to obtain the complete set of five hydrodynamical equations
describing any real liquid [4,20]. Particles with an angular momentum
[21,22] and internal elasticity [23] have also been suggested.
The equations of motion, obtained from the above force expressions in
combination with Newtons second law,
dxi
vi
dt
dvi X c
6
7
563
cutoff radius of the potential as the unit of length, and the thermal energy
kB T 13 mhv2i i as the unit of energy. The unit of time is then implicitly
defined; a unit of temperature is not needed, since the temperature only
appears as the product kBT. These units, together with a field parameter a
and a friction of the order ten, suffice for qualitative results, like the role
played by hydrodynamics in the microphase separation of block copolymers. A further tuning is required for more quantitative results. We will
summarize here the method proposed by Groot and Warren [24], including
modifications of recent date [13].
One first chooses nmap, the number of atoms to be represented by a single
DPD particle, thereby fixing the mass of the particle. Next one chooses a
density of particles per r3c , which combines with the weight density of the
real fluid to give a value for rc. In simulations typical values for are 3, 4, or
higher, to ensure a sufficiently high number of nearest neighbor interactions
for a fluid-like structure.
The strength of the conservative force field can be obtained from the
requirement that the density fluctuations of the real fluid should be
reproduced. As the static structure factor for a coarse-grained system is
smaller than that of the atomic system by a factor nmap for low wave
numbers, it follows that kB TT DPD n1
map kB TT atom . The isothermal
compressibility coefficient T of the DPD fluid is readily obtained from its
surprisingly simple equation of state [24], p kBT 0.1a2, and this set of
equations is then solved for a.
Finally a friction coefficient is chosen, and the unit of time is set by
comparing the self-diffusion coefficient D of a particle with that of nmap
atoms. Typical values obtained this way for water with nmap 3, 3[l3],
and 3[mlt3/2] are a 80[ml2t2] and the units 1[l ] 3 A, 1[t] 90 ps, and
1[m] 9 1026 kg [13]. As the implications of nmap for the conservative
field have been realized only recently, most simulations have actually been
run using softer a values in the range of 10 to 25.
This procedure for choosing the simulation parameters is by no means
unique. One could, for instance, in the above scheme opt to define the unit
of time by matching the thermal velocity of a particle with that of nmap
atoms. The friction coefficient would then have to be tuned by matching the
diffusion coefficient, using the readily understood observation D1
D1 0 g6kB T=1 , where both D(0) and g are determined by the
conservative field. Note that there will be no physically meaningful solution
if the conservative force field itself already gives rise to a diffusion
coefficient below the experimental value. An altogether different approach,
using theoretical predictions [22,3032] for the ideal fluid (a 0) to match
dynamical properties such as the viscosity and the sound velocity, has also
been proposed [33].
564
565
566
of swelling due to excluded volume interactions; the overshoot of the theoretical value is typical for short chains. Scaling theory predicts that for
kRg 1 the rescaled dynamic structure factors S(k, kxt)/S(k, 0) coalesce
for x 3 (Zimm) or x 4 (Rouse) [10,11]. Although the numerical value
of x 2.7 was lower than expected, in the good company of previous polymer simulations, it evidences the impact of hydrodynamic interactions on
the motion of the polymer. A further indication hereof was found in the
distribution of relaxation times in the decay of the conformational time
correlation function,
Ct
N
1X
ri t Rcom t ri 0 Rcom 0
N i1
where Rcom denotes the center of mass position. In a nutshell, their approach
was to fit this function with
Ct /
i2h exp
10
1 i2h
5
3
p
N1
A12
theory, A 4=3
3:00:05 a=ass
simulations
11
This suggests a simple linear relationship between and a, as was also
observed by Groot and Warren [24].
567
The effect of solvent quality on the dynamics of the polymer was studied
by calculating the conformational time correlation function as a function of
N and a [34]. While the scaling exponent decreased from nearly 2 in the
good solvent to 0.9 in the bad solvent, the ratio / remained relatively
constant between 3 and 3.3 (theory: 3). A sheared simulation box with
LeesEdwards boundary conditions was used to calculate the impact of a
single dissolved polymer on the viscosity of the box [4446]. The largest
zero-shear viscosities were observed for the good solvent, the lowest for the
bad solvent. Upon increasing the shear rate _ , the solution was found
to shear-thin in the region surrounding _ 1 1, the effect being most
pronounced in the good solvent and nearly absent in the bad solvent. All of
these static and dynamic results, save for the 0.61 from the static
structure factor S(k,0), confirmed the rather unsatisfactory result that the
athermal solvent acted as a near-theta solvent.
The case of a single polymer in a quiescent athermal solvent was recently
revisited by Spenley [47], using chains of 2 to 100 beads in a weak spring
model. For the end-to-end vector he found the exponents 0.58 0.04
and 1.80 0.04, both in excellent agreement with the predictions for a
polymer with excluded volume and hydrodynamic interactions. It is not
clear what caused this marked difference with the above discussed results.
The verification of the scaling law for the self-diffusion coefficient, D / N
,
was more involved, as D is rather sensitive to the size of the simulation box.
After correcting for this complication, a satisfying agreement with theory
was observed.
In the melt [47] the polymers adopted a Gaussian distribution,
0.498 0.005, as predicted by Flory. The dynamical coefficients,
1.98 0.03 and
1.02 0.02, and the viscosity / N, all followed
the Rouse model, showing that the interpolymer interactions were
sufficiently strong to screen the hydrodynamic interactions, despite the
softness of the potential. In fact, the potential was so soft that polymers
were seen to cross one another [47,48]; a transition for large N from Rouse
dynamics to de Gennes reptation is therefore not to be expected. In cases
where the entanglement of polymers is an issue, DPD is readily combined
with recently developed algorithms which explicitly reintroduce the
uncrossability lost in coarse-grained soft potentials [49,50].
We have performed simulations at polymer concentrations ranging from
a dilute system to a melt, for polymers of 20 to 100 beads, to study the
impact of interpolymer interactions on the dynamics [51]. The polymer
solvent repulsion was tuned to a theta point condition, rendering the
exponent and the radius of gyration at fixed N virtually independent of
the concentration. For the lowest concentrations we found a reasonable
agreement with the Zimm theory of the theta state. When the polymer bead
568
V. BLOCK COPOLYMERS
In an important series of articles the Unilever group have described the
use of DPD to characterize microphase separation phenomena in linear
diblock copolymers of the general formula AnBm [24,58,59], making detailed
comparisons with both experiments and the predictions of self-consistent
field theory [60]. Block copolymers are both interesting and useful because
of their property of forming complicated aggregated structures, as the
natural tendency for segregation of the blocks composed of A and B
monomers is frustrated by the connectivity of the copolymer. The system
can therefore only reduce its free energy by coalescing the A- and B-rich
569
domains, forming e.g., sheets, rods, or spheres. The phase diagram is fully
determined by just two parameters, N and the fraction f of A beads.
The basic parameterization used was as discussed above, with / a.
Groot et al. [24,58] showed that in DPD the interfacial tension between
the component homopolymers followed the same scaling law, as a function of N and , as experimental data for PS-PMMA. This is an important result, since it is the interfacial tension which drives microphase
separation. It also provides validation for the comparison of experimental
data on long chains with DPD data on short chains (but correspondingly
much larger ).
With N 41, the simulations were well outside the weak segregation
regime, and when the melts were quenched the expected order of equilibrium structures was found as the copolymers become successively more
asymmetriclamellae ( f 0.37), perforated lamellae (0.31 f 0.37), hexagonally arranged rods (0.26 f 0.31), fluid peanut-shaped micelles (at
f =0.2), and a disordered phase (at f 0.1). These stable structures emerged
via nontrivial pathways, forming a series of metastable states before
equilibrium is reached. Discrepancies between the exact positions of phase
transition points and those predicted by theory were explained in terms of
the short length of the DPD copolymer chains, N 10, which effectively
reduced the FloryHuggins parameter to N 20 [58]. The finding of
a hexagonally perforated lamellar phase is interesting since this structure
has been identified in recent experiments, while current mean field theory
predicts a gyroid phase in this region. Finite sample size effects cannot,
however, be discounted as a reason for the nonappearance of the gyroid
phase in the simulations.
In a later article [59] the importance of hydrodynamic interactions in
the formation of certain microphases was demonstrated by close comparison
of simulations using DPD and Brownian dynamics (BD). Whilst both
simulation methods describe the same conservative force field, and hence
should return the same equilibrium structure, they differ in the evolution
algorithms. As mentioned above, DPD correctly models hydrodynamic
interactions, whereas these are effectively screened in BD. Distinctly
different equilibrium structures were obtained using the two techniques
in long simulations of a quenched 2400 A3B7 polymer system, as shown in
Fig. 1. Whilst DPD ordered efficiently into the expected state of hexagonal
tubes, BD remained trapped in a structure of interconnected tubes. By
way of contrast, both DPD and BD reproduced the expected lamellar
equilibrium structure of A5B5 on similar time scales, see Fig. 2.
There is experimental evidence [61] that the formation of microphases occurs on at least three different time and length scales. In the first
stage there is phase separation on the level of the interacting beads. This is
570
R
FIG. 2 The global order parameter Sk ln Sk d 3 k as a function of time for
quenched block copolymers, as obtained in DPD and BD simulations. For A3B7
(HEX) DPD shows an activated transition from a disordered network of tubes to a
hexagonal phase of rods, whereas BD remains stuck in the former. The A5B5 (LAM)
immediately form a lamellar structure in both DPD and BD. (From Ref. [59].)
571
VI. CONCLUSIONS
Dissipative particle dynamics has proved to be a very versatile simulation
technique for complex flows on the mesoscopic scale, where thermodynamics, hydrodynamics, and thermal noise play important roles. The
polymer simulations reviewed here cover but a part of the phenomena that
have been, and will be, studied using DPD.
The conservative force field in DPD is extremely soft compared to those
used in atomic simulations, yet after parameterization using the compressibility or the FloryHuggins parameter it is able to capture many of the
572
REFERENCES
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
573
24. Groot, R.D.; Warren, P.B. J. Chem. Phys. 1997, 107, 4423.
25. den Otter, W.K.; Clarke, J.H.R. Int. J. Mod. Phys. C 2000, 11, 1179.
26. Pagonabarraga, I.; Hagen, M.H.J.; Frenkel, D. Europhys. Lett. 1998,
42, 377.
27. Besold, G.; Vattulainen, I.; Karttunen, M.; Polson, J.M. Phys. Rev. E 2000, 62,
R7611.
28. den Otter, W.K.; Clarke, J.H.R. Europhys. Lett. 2000, 53, 426.
29. Lowe, C.P. Europhys. Lett. 1999, 47, 145.
30. Marsh, C.A.; Backx, G.; Ernst, M.H. Phys. Rev. E 1997, 56, 1676.
31. Masters, A.J.; Warren, P.B. Europhys. Lett. 1999, 48, 1.
32. Espanol, P.; Serrano, M. Phys. Rev. E 1999, 59, 6340.
33. Dzwinel, W.; Yuen, D.A. Int. J. Mod. Phys. C 2000, 11, 1.
34. Kong, Y.; Manke, C.W.; Madden, W.G.; Schlijper, A.G. J. Chem. Phys. 1997,
107, 592.
35. Flekky, E.G.; Coveney, P.V.; Fabritiis, G.D. Phys. Rev. E 2000, 62, 2140.
36. Espanol, P.; Serrano, M.; Zuniga, I. Int. J. Mod. Phys. C 1997, 8, 899.
37. Serrano, M.; Espanol, P. Phys. Rev. E 2001, 64, 046115.
38. Wijmans, C.M.; Smit, B.; Groot, R.D. J. Chem. Phys. 2001, 114, 7644.
39. van Vliet, R.E.; Hoefsloot, H.C.J.; Hamersma, P.J.; Iedema, P.D. Macromol.
Theory Simul. 2000, 9, 698.
40. Pagonabarraga, I.; Frenkel, D. J. Chem. Phys. 2001, 115, 5015.
41. Pagonabarraga, I.; Frenkel, D. Mol. Sim. 2000, 25, 167.
42. Monaghan, J.J. Annu. Rev. Astron. Astrophys. 1992, 30, 543.
43. Schlijper, A.G.; Hoogerbrugge, P.J.; Manke, C.W. J. Rheol. 1995, 39, 567.
44. Schlijper, A.G.; Manke, C.W.; Madden, W.G.; Kong, Y. Int. J. Mod. Phys. C
1997, 8, 919.
45. Kong, Y.; Manke, C.W.; Madden, W.G.; Schlijper, A.G. Tribol. Lett. 1997,
3, 133.
46. Zhang, K.; Manke, C.W. Mol. Sim. 2000, 25, 157.
47. Spenley, N. Europhys. Lett. 2000, 49, 534.
48. Gibson, J.B.; Zhang, K.; Chen, K.; Chynoweth, S.; Manke, C.W. Mol. Sim.
1999, 23, 1.
49. Terzis, A.F.; Theodorou, D.N.; Stroeks, A. Macromol. 2000, 33, 1397.
50. Padding, J.T.; Briels, W.J. J. Chem. Phys. 2001, 115, 2846.
51. den Otter, W.K.; Ali, I.; Clarke, J.H.R. In preparation.
52. Kong, Y.; Manke, C.W.; Madden, W.G.; Schlijper, A.G. Int. J. Thermophys.
1994, 15, 1093.
53. Willemsen, W.M.; Hoefsloot, H.C.J.; Iedema, P.D. J. Stat. Phys. 2002, 107,
53.
54. Malfreyt, P.; Tildesley, D.J. Langmuir 2000, 16, 4732.
55. Irfachsyad, D.T.D.; Malfreyt, P. Phys. Chem. Chem. Phys. 2002, 4, 3008.
56. Wijmans, C.M.; Smit, B. Macromolecules 2002, 35, 7138.
57. Gibson, J.B.; Chen, K.; Chynoweth, S. J. Colloid Interface Sci. 1998, 206,
464.
58. Groot, R.D.; Madden, T.J. J. Chem. Phys. 1998, 108, 8713.
574
59. Groot, R.D.; Madden, T.J.; Tildesley, D.J. J. Chem. Phys. 1999, 110, 9739.
60. Matsen, M.W.; Schick, M. Curr. Opin. Colloid Interface Sci. 1996, 1, 329.
61. Balsara, N.P.; Garetz, B.A.; Newstein, M.C.; Bauer, B.J.; Prosa, T.J.
Macromolecules 1998, 31, 7668.
62. Zhang, K.; Manke, C.W. Comput. Phys. Comm. 2000, 129, 275.
18
Dynamic Mean-Field DFT Approach
for Morphology Development
A. V. ZVELINDOVSKY, G. J. A. SEVINK, and
J. G. E. M. FRAAIJE Leiden University, Leiden, The Netherlands
I.
INTRODUCTION
576
Zvelindovsky et al.
XZ
577
Here is the partition functional for the ideal Gaussian chains in the
external field UI, and F nid[] is the contribution from the nonideal interactions. The free energy functional (1) is derived from an optimization criterion (see [11]) which introduces the external potential as a Lagrange
multiplier field. The external potentials and the concentration fields are
related via a density functional for ideal Gaussian chains:
I U r n
N
X
K
Is0 Trc r Rs0
s0 1
0
Here K
Is0 is a Kronecker delta function with value 1 if bead s is of type I and
0 otherwise. The trace Trc is limited to the integration over the coordinates
of one chain
Z
Trc N
VN
N
Y
dRs
s1
N is a normalization constant.
tion function
1 H G PN Us Rs
s1
e
H G
N
3 X
Rs Rs1 2
2
2a s2
with a the Gaussian bond length parameter. The density functional is bijective; for every set of fields {UI} there is exactly one set of fields {I}. Thus
there exists a unique inverse density functional UI[] . There is no known
closed analytical expression for the inverse density functional, but for our
purpose it is sufficient that the inverse functional can be calculated
efficiently by numerical procedures (see Appendices A and B).
We split the nonideal free energy functional formally into two parts
F nid F c F e
578
Zvelindovsky et al.
I r
I r I r
c
UI r
I r
eI r
I r
5
6
cI r
Z Z
F c X
I
J
Z
V
7
8
3
2a2
IJ jr r0 j 0IJ
3=2
h
i
2
exp 3=2a2 r r0
The excluded volume interactions can be included via the correction factor
or insertion probability for each bead, c [12]:
F e
XZ
I
I rlncr dr
10
579
11
12
III. APPLICATION
A. Pluronics in Water Mixtures
The aqueous solution of triblock copolymer Pluronic L64 EO13PO30EO13
[23] is modeled as a compressible system consisting of ideal Gaussian chain
molecules in a mean field environment. The free energy is a functional of a
set of particle concentrations {} [11,12]:
Z
n
pp ns s X
UI rI r dr
Ffg kT ln
np !ns !
V
I
Z
1X
I I r I dr
2 V I
13
where np (ns) is the number of polymer (solvent) molecules, is the intramolecular partition function, I is a component index (EO, PO, or solvent),
and V is the system volume. The external potential UI is conjugate to the
particle concentration I via the Gaussian chain density functional (2).
The average concentration is 0I and I is the particle volume. The cohesive
interactions have kernels "IJ (7) [11]. The Helfand compressibility parameter
is H [12].
The dynamics of the system is governed by the diffusion-convection
equation (11) with the periodic boundary conditions as described in [11,17].
The dynamic equations are closed by the expression for the free energy (13)
and the Gaussian chain density functional (2). Seven dimensionless
580
Zvelindovsky et al.
581
(a)
(b)
(c)
(d)
0.10
0.25
FIG. 1 Isosurface representation of PL64 in water mixture for 70% (a), 60% (b),
55% (c), and 50% (d) polymer content. The isolevel is EO EO 0.3.
FIG. 2 Detail of the simulation snapshots from Fig. 1. (a) LAM, (b) GYR, (c)
HEX, and (d) micellar phases are clearly visible.
C. Modulation by Shear
It is known that flow fields affect mesoscale structures in complex liquids
and polymer systems, giving rise to global orientation [28,29]. Because of its
582
Zvelindovsky et al.
FIG. 3 Isosurface representation of the A3B3C3A3 melt (a). The system develops
into one lamellar (b) and two cylindrical (c,d) mesostructures.
583
584
Zvelindovsky et al.
FIG. 5 Isosurface representation of the A8B8 melt at 75,000 and shear flow
vx _ y, vy vz 0, using the structure at 500 as a starting structure. The isolevel
is A A 0.3. One can clearly observe the global lamellar orientation. The x- and
y-axes are indicated.
FIG. 6 Mesophases of 55% Pluronic L64 in water under shear at time 12,500,
using the structure at 2500 from Figs.1c and 2c as a starting structure before
shearing. The isosurfaces are at P PP 0.33. The x-axis is the velocity direction,
the y-axis is the velocity gradient direction, and the z-axis is a neutral direction.
585
FIG. 7 Mesophases of 60% Pluronic L64 in water at the end of shearing (top) and
after a long relaxation after cessation of shear (bottom).
D. Modulation by Reactions
In dynamic mean field density functional theory, the dynamics of the
polymer melt under simple steady shear is described by the time evolution of
density fields I. The dynamics is governed by a diffusion-convection
equation with sheared periodic boundary conditions and can readily be
extended with (second order) reaction terms in the following manner:
N
X
@I
F
Mr I r
_ yrx I
kJK J K I
I
@t
J1, K1
14
586
Zvelindovsky et al.
F
_ yrx hA khA hB hA
hA
F
Mr hB r
_ yrx hB khA hB hB
hB
F
Mr dA r
_ yrx dA khA hB dA
dA
F
Mr dB r
_ yrx dB khA hB dB
dB
Mr hA r
15
Here, hA (hB) is the density of A (B) beads in the homopolymer, dA (dB)
is the density of A (B) beads in the diblock copolymer, and k is the reaction
rate of the coupling reaction.
Figure 8 gives an example of formation of double layer droplets in a
sheared reactive polymer system. Initially the A/B blend was subject to shear
which resulted in the formation of elongated droplets. Then reaction on the
surface of the droplets took place after switching off the shear. That leads to
relaxation of the elongated shape of the droplets towards spherical. The
excess of polymer formed at the interface goes inside the droplets which
forms a double layer structure.
I I r 0I dr
16
2 V I
except for an extra fourth term that contributes only in the direct vicinity of
the filler particles. This accounts for the interaction of a polymer melt with
587
s U r CMr
VN
17
588
Zvelindovsky et al.
0
1
r 2V 0
r 2V=V 0
The time evolution of the density field I(r) can be described by a time
dependent LandauGinzburg type equation (11). The boundary conditions
that are used on the simulation box are periodic boundary conditions. For
the diffusion flux in the vicinity of the filler particles, rigid-wall boundary
conditions are used. A simple way to implement these boundary conditions
in accordance with the conservation law is to allow no flux through the filler
particle surfaces, i.e.,
r
I n 0
18
where n is the normal pointing towards the filler particle. The same boundary conditions apply to the noise I. Figure 9 demonstrates formation of
lamellar structure in an A8B8 melt in the presence of interacting walls.
In Fig. 10 one can see the same system but confined between neutral
walls. This confinement leads the system to form perpendicular to the
walls lamellae.
Thickness of the slit/film can have a drastic effect on the polymer
morphology. In Fig. 11 the same 55% PL64 in water system was confined
in between two different slits. The bulk morphology of this system is
cylindrical, Figs. 1c, 2c, and 6. Adopting conformational freedom the system
develops into cylinders, which are either parallel or perpendicular to the
wall, depending on the slit width.
589
590
Zvelindovsky et al.
FIG. 10 Lamellar formation of an A8B8 copolymer melt in the presence of the same
filler particle as Fig. 9. Moreover, there is no interaction between the polymer beads
and boundaries of the filler particles. (e) View of filler particle in simulation box, (f)
morphology of A beads (isolevel A 0.5) in one simulation box at 500, (g) the
same for 4000, (h) final morphology at 10,000.
FIG. 11 Cylinder formation of 55% PL64 in water mixture in two different slits,
12h (left) and 15h (right), forming parallel (left) or perpendicular (right) cylinders.
591
19
20
592
Zvelindovsky et al.
with G0 r Ginv
N1 r 1. The linkage operator f r is defined as a
convolution with a Gaussian kernel
Z
3 2=3
f r
exp 3=2a2 r r0 2 f r0 dr0
21
2
2a
V
Note that the Green propagator method to calculate the density functional
(2) yields an exact answer.
Now, the crucial point of the algorithm is a numerical representation of
the linkage operator , given a cubic uniform grid with mesh width h, which
is both efficient and accurate. Since we work on large cubic grids and
use domain decomposition for the parallelization, FFTs and Gaussian
Quadrature rules are not considered. Efficiency is crucial, since the linkage
operation has to be repeated 2(N1) times for a single density functional
calculation. Furthermore, if the numerical representation of the linkage
operator is not accurate, its repeated application will result in error
accumulation, especially in the small q range, i.e., on chain length scales.
Large errors in the high q range are also undesirable, since they may disturb
the bijectivity of the mapping between density field and external potential.
A 27-point stencil is the most compact stencil that meets all constraints.
We reduce the number of parameters by invoking symmetry rules, similar to
the polyhedra rules used in multidimensional integration of symmetric
kernels [41]. The linkage operator on a uniform grid is expressed as:
f r c0 f r
X ci
i
f r ri f r ri
22
ea jqj =6
P3
Pd i
j1 cosq rij
i0 ci
continuum
discrete
593
594
Zvelindovsky et al.
The way the time integration is carried out is as follows: we start with a
homogeneous distribution of Gaussian chains I I (where is the mean
density of the fields) and UI 0. We solve the CN equations (to be defined
below) by an iterative process in which the external potential fields are
updated according to a steepest descent method. The updated external
potential fields are related, by the chain density functional, to unique density
fields. We calculate the intrinsic chemical potential from the set of external
potential fields and density fields and update the density fields. The density
fields are only accepted as an update for the next time step if the L2 norm of
the residual is below some predefined upper boundary. An overview of this
scheme is shown in Fig. 12.
The general expression of the CN equations for this problem is
kI k1
k1
k
I
!1 Mh2 DkI D
kI 1 !1 Mh2 Dk1
I D
I I
t
23
where k is the time index, h the mesh size, !i the CN parameters determining
the degree of explicitness versus implicitness of the method, and h2D(fDg)
the discrete operation on the grid representing r frg. The computational
procedure for the div-gradient operation is again important because of the
risk of introducing grid artifacts. We have developed a procedure similar to
the procedure for the numerical integration. As computational cell, again a
595
Since we do not have an explicit expression for the external potential fields
in terms of the densities, we solve this equation for the external potential
fields. The external potential fields UI are found by an iterative scheme
UI,k 0 UI,k1
final
UI,k p UI,k p1 reskI, p1 p 1
25
26
REFERENCES
1. Qiwei He, D.; Nauman, E.B. Chem. Engng. Sci. 1997, 52, 481496.
2. Gonnella, G.; Orlandini, E.; Yeomans, J.M. Phys. Rev. Lett. 1997, 78,
16951698.
3. Patzold, G.; Dawson, K. J. Chem. Phys. 1996, 104, 59325941.
4. Kodama, H.; Komura, S. J. Phys. II France 1997, 7 (1), 714.
5. Ohta, T.; Nozaki, H.; Doi, M. Phys. Lett. A 1990, 145 (6,7), 304308.
596
6.
7.
8.
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
35.
Zvelindovsky et al.
Cahn, J.W.; Hilliard, J.E. J. Chem. Phys. 1958, 28, 258267.
Oono, Y.; Puri, S. Phys. Rev. Lett. 1987, 58, 836839.
de Gennes, P.G. J. Chem. Phys. 1980, 72, 47564763.
Maurits, N.M.; Fraaije, J.G.E.M. J. Chem. Phys. 1997, 106, 67306743.
Fraaije, J.G.E.M. J. Chem. Phys. 1993, 99, 92029212.
Fraaije, J.G.E.M.; van Vlimmeren, B.A.C.; Maurits, N.M.; Postma, M.; Evers,
O.A.; Homann, C.; Altevogt, P.; Goldbeck-Wood, G. J. Chem. Phys. 1997,
106 (10), 42604269.
Maurits, N.M.; van Vlimmeren, B.A.C.; Fraaije, J.G.E.M. Phys. Rev. E 1997,
56, 816825.
van Vlimmeren, B.A.C.; Fraaije, J.G.E.M. Comp. Phys. Comm. 1996, 99, 2128.
Maurits, N.M.; Altevogt, P.; Evers, O.A.; Fraaije, J.G.E.M. Comp. & Theor.
Pol. Sci. 1996, 6, 18.
Kawakatsu, T. Phys. Rev. E 1997, 56, 32403250.
Hasegawa, R.; Doi, M. Macromolecules 1997, 30, 30863089.
van Vlimmeren, B.A.C.; Maurits, N.M.; Zvelindovsky, A.V.; Sevink, G.J.A.;
Fraaije, J.G.E.M. Macromolecules 1999, 32, 646656.
Fleer, G.J., et al. Polymers at Interfaces; Chapman & Hall: London, 1993.
Doi, M. Introduction to Polymer Physics; Oxford University Press: Oxford,
1996.
Matsen, M.W.; Schick, M. Phys. Rev. Lett. 1994, 72, 26602663.
Drolet, F.; Fredrickson, G.H. Phys. Rev. Lett. 1999, 83, 43174320.
Hamley, I.W. The Physics of Block Copolymers; Oxford University Press:
Oxford, 1998.
Alexandridis, P.; Olsson, U.; Lindmann, B. Macromolecules 1995, 28,
77007710.
Leibler, L. Macromolecules 1980, 13 (6), 16021617.
Malcolm, G.N.; Rowlinson, J.S. Trans. Faraday Soc. 1957, 53, 921931.
Hill, T.L. An Introduction to Statistical Thermodynamics; Addison-Wesley:
Reading, MA, 1962.
Krevelen, D.W. Properties of Polymers, Their Correlation with Chemical
Structure; Their Numerical Estimation and Prediction from Additive Group
Contribution; Elsevier: Amsterdam, third edition, 1990.
Kodama, H.; Doi, M. Macromolecules 1996, 29 (7), 26522658.
Fredrickson, G.H. J. Rheol. 1994, 38, 10451067.
Zvelindovsky, A.V.; Sevink, G.J.A.; van Vlimmeren, B.A.C.; Maurits, N.M.;
Fraaije, J.G.E.M. Phys. Rev. E 1998, 57, R4879R4882.
Doi, M.; Chen, D. J. Chem. Phys. 1989, 90, 52715279.
Ohta, T.; Enomoto, Y.; Harden, J.L.; Doi, M. Macromolecules 1993, 26,
49284934.
Chen, Z.R.; Korneld, J.A.; Smith, S.D.; Grothaus, J.T.; Satkowski, M.M.
Science 1997, 277, 12481253.
Zvelindovsky, A.V.; Sevink, G.J.A.; Fraaije, J.G.E.M. Phys. Rev. E 2000, 62,
R3063R3066.
Scheutjens, J.M.H.M.; Fleer, G.J. J. Phys. Chem. 1979, 83, 16191635.
597
36. Scheutjens, J.M.H.M.; Fleer, G.J. J. Phys. Chem. 1980, 84, 178190.
37. Leermakers, F.A.M.; Scheutjens, J.M.H.M. J. Chem. Phys. 1988, 89, 3264.
38. Evers, O.A.; Scheutjens, J.M.H.M.; Fleer, G.J. Macromolecules 1990, 23,
52215233.
39. de Gennes, P.-G. Scaling Concepts in Polymer Physics; Cornell University
Press: Ithaca, NY, 1979.
40. Doi, M.; Edwards, S.F. The Theory of Polymer Dynamics; Clarendon: Oxford,
1986.
41. Zwillinger, D. Handbook of Integration; Jones and Barlett Publishers: Boston,
1993.