0% found this document useful (0 votes)
58 views56 pages

Rudnick Notes

Download as pdf or txt
Download as pdf or txt
Download as pdf or txt
You are on page 1/ 56

Lectures on the random walk

Joseph Rudnick
Department of Physics and Astronomy, UCLA

Opening statment: generalities


I have been told that I can talk about whatever I want, as long as it has to
do with random walks. This is fine with me. However, it leaves me with a bit
of a dilemma. There is so much to say about the subject and three lectures
does not give me all that much time. I think that what I will do is focus on
a few topics that amuse me and that I think you will find interesting and
useful as well. My first job will be to provide a general introduction to the
subject and in particular to convince you of the importance and utility of a
few key concepts and approaches, principle among them the method of the
generating function. For those of you who are interested in pursuing the
subject further, there are a number of excellent books. However, given that I
like many others operate from a position of healthy self-interest, let me refer
you to a book co-authored by me on the subject [Rudnick and Gaspari, 2004].
We will start with the discussion of some basic quantities.

The random walker


First, we need to talk about what the random walker is, exactly. The simplest
version of a random walker is that that it is something that takes a set of
successive steps and that the direction in which it moves is, to some extent,
random. In the simplest, “ideal,” case, the direction is entirely random. This
leaves open two other questions
1. How much ground does the walker cover at each step? Is this quantity
fixed or random?
2. Does the walker take the steps in a regular time sequence (e.g. a step
each and every second) or is there randomness in that choice as well?

1
Again, in the simplest case the length of each step is fixed and the walker
takes steps at regular time intervals. We will consider variations on the above
standard set of conditions later on.

Number of walks
Now we turn to quantities of fundamental interest in discussions of the prop-
erties of random walkers and random walks. The most widely used, and
generally most widely applicable, of those quantities is the number of ran-
dom walk, assuming that certain conditions apply. For example, consider
the following simple question:

A random walker starts at the location ~x, takes N steps and


ends up at the location ~y . How many ways are there for this to
happen?

A bit more concisely and technically: how many N -step walks are there that
start at ~x and end at ~y ? Figure 1 shows three random walks starting and
ending at the same point. The question is, how many of such paths are there?
This quantity, which we’ll call C(N ; ~x, ~y ) plays the same role in random walk

Figure 1: Three random walks starting and ending at the same point.

statistics as the partition function plays in statistical mechanics. To see what

2
I mean by this, consider the following definition of the partition function
X
ZN = e−βEstate (1)
N −particle states

where the sum above is–as clearly indicated–over all of the states of a given
N -particle system. The exponential is the standard Boltzmann factor, β
being 1/kB T and Estate the energy of the given state. By comparison, the
quantity C(N ; ~x, ~y ) is given by the equation
X
C(N ; ~x, ~y ) = (2)
N −step walks starting at ~
x and ending at ~
y

In the case of the sum on the right hand side of (2), the “Boltzmann factor”
is one. We can alter this if the conditions on the walk change, and in fact we
will do so shortly.

Recursion relations
So, how do we evaluate the sum in question? One very useful method makes
use of a relationship between the number of N step walks and the number
of walks that require N + 1 steps. The argument leading to this relationship
is fairly transparent. Here is how it goes. If I have managed to find out how
many N -step walks start at a given point ~x and terminate at all possible end
points ~y 0 , then counting all N + 1 step walks starting at ~x and terminating
at a particular end point ~y just involves adding up all N step walks that
start at ~x and end up at a point from which the walker can get to y in one
step. Suppose I know the value of C(N ; ~x, ~y ) for all endpoints, ~y . Then,
C(N + 1, ~x, ~y ) equal is given by
X
C(N + 1; ~x, ~y ) = C(N ; ~x, w
~ i) (3)
w
~i

Here, the w ~ i ’s are the locations of the points from which the walker can make
it to the location ~y in one step. Figure 2 illustrates of the process summarized
in Eq. (3)1 .
1
In the figure, as noted in the caption, the process depicted is appropriate to a random
walker on a lattice. In general, I will not be precise about whether this assumption applies,
as most results are qualitatively independent of that particular restriction

3
w4

w1 y w3

w2

Figure 2: An illustration of the recursion relation summarized in Eq. (3).


Here, the random walker is confined to the vertices of a two-dimensional
lattice and all steps are between a lattice point and its nearest neighbor.

If we play some games with this recursion relation, we end up with a


useful, general and nearly ubiquitous equation. The key assumption is going
to be that we can make use of power series expansions. First, we’ll assume
that we can express the dependence of the number of walks on the number
of steps taken in terms of a Taylor expansion in the latter quantity. This
means we can write

C(N + 1; ~x, ~y ) = C(N ; ~x, ~y ) + ∂C(N ; ~x, ~y )/∂N + · · · (4)

Furthermore, we will assume that the same sort of expansion is possible with
respect to the locations of the starting and ending points. Let’s focus our
attention on the kind of walk on a lattice shown in Fig. 2. Given that we
can write

w ~ i − ~y )
~ i = ~y + (w
~i
≡ ~y + ∆ (5)

we can then replace the right hand side of the recursion relation equation (3)

4
by
X
~ i)
C(N ; ~x, ~y + ∆
∆i
!
2
X
~i·∇
~ y C(N ; ~x, ~y ) + 1 X ∂
= C(N ; ~x, ~y ) + ∆ ∆i,l ∆i,m C(N ; x~,~y ) + · · ·
2 l,m ∂yl ∂ym
~i

(6)

where ∆i.m is the mth component of the vector displacement ∆ ~ i . To continue


in our analysis of the random walk along these lines, we will consider the case
of a walker restricted to the vertices of a cubic lattice. This means that the
~ are aligned the the x, y or z axis. Because of this, we have the following
∆’s
results
X
= 6 (7)
~i

X
= 0 (8)
~i

2
∂2 ∂2 ∂2
 
XX ∂
∆i,l ∆i,m = 2a2 + +
l,m
∂yl ∂ym ∂y12 ∂y22 ∂y32
~i

~2
= a2 ∇ (9)
y

The quantity a on the right hand side of (9) is the magnitude of each ∆ ~ i,
th
and yk is the k component of the position vector ~y .
Collecting all the results together, we are left with the following equation
for the quantity C(N ; ~x, ~y ).

C(N ; ~x, ~y ) + ∂C(N ; ~x, ~y )/∂N


= 6C(N ; ~x, ~y ) + a2 ∇ ~ 2 C(N ; ~x, ~y ) (10)
y

Rearranging, we end up with the following partial differential equation


∂C(N ; ~x, ~y )
= 5C(N ; ~x, ~y ) + a2 ∇2y C(N ; ~x, ~y ) (11)
∂N
This looks a lot like the Schrödinger equation for a particle in a constant
potential. The principal difference is that the number of steps, N , plays the

5
role of an imaginary time. As it turns out, the equation is not quite right.
We have made too much of an approximation in utilizing the first two terms
in the Taylor expansion in N in our treatment of the left hand side of the
recursion relation. To see this, let’s figure out what the equation tells us
about the number of N -step walks that start at x and end up anywhere. We
do this by simply summing over end-points, y. Making use of the fact that
the last term on the right hand side of (11) can be expressed in terms of
perfect derivatives,and that the integral over all ~y of the function C(N ; ~x, ~y )
converges, because when ~y is sufficiently far from ~x no N -step walk will
connect the two points, we have on integrating both sides of (11) over ~y
Z Z
∂ 3
C(N ; ~x, ~y )d y = 5 C(N ; ~x, ~y )dy (12)
∂N
or, defining Z
C(N ; ~x) = C(N ; ~x, ~y )dy (13)

∂C(N ; ~x)
= 5C(N ; ~x) (14)
∂N
The solution to this equation is

C(N ; ~x) = C(0; ~x)e5N (15)

Equation (15) tells us that the number of N step walks increases exponen-
tially in the number of steps. Does this make sense? Consider the process.
At each step, the walker on the cubic lattice has a choice of six neighboring
sites to visit. This means that the total number of walks increases by a factor
of 6 at each step, or the total number of N step walks goes as 6N = eN ln 6 .
This is qualitatively consistent with the result displayed in (15). Is it quan-
titatively consistent? According to that result the number of walks increases
by the factor e5 = 148.4 at each step. We are right about the exponential
growth in the number of walks with steps taken but wildly off with respect to
the actual rate of exponential increase. The moral is cautionary. Be careful
when you treat a discrete quantity as if it were continuous.
There is another quantity related to the number of N -step walks for which
the kind of equation displayed in (11) is absolutely relevant and asymptoti-
cally correct. This quantity is equal to the fraction of N -step walks starting
at ~x and ending at ~y . To obtain this quantity, we divide C(N ; ~x, ~y ) by the

6
total number of N step walks, which we’ve seen is 6N . Let’s call this quantity
D(N ; ~x, ~y ). The recursion relation corresponding to (3) is
1X
D(N + 1, ~x, ~y ) = D(N ; ~x, w
~ i) (16)
6
w
~i

Carrying out the same truncated Taylor expansions as before, we end up


with the differential equation
∂D(N ; ~x, ~y ) a2
= ∇2y D(N ; ~x, ~y ) (17)
∂N 6
If we integrate both sides of this equation with respect to ~y , we end up with
the quite proper result
Z

D(N ; ~x, ~y )d3 y = 0 (18)
∂N
This tells us that if we sum the fraction of walkers that end up at a given
site over all possible end points, we end up with a constant. In fact, we know
that the constant in question is one.
Now, (17) is a well known and widely studied equation. It is, in fact the
diffusion equation. Among other phenomena, it describes the spreading of
a cloud of suspended particles in a liquid or vapor host, an example being a
drop of ink or dye that has been deposited in water. This allows for another
interpretation of the function D(N ; ~x, ~y ). Imagine a collection of walkers that
have al been simultaneously deposited at the point ~x. They then walk, in the
simplest case all in lockstep, out from that point. The quantity D(N ; ~x, ~y )
is the concentration of those walkers, normalized to one. Figure 3 shows the
case when there are only five walkers, after they have taken 100 steps.
This interpretation of the function will help lead us to some very useful
approaches to classic problems.

Solution of the diffusion equation


Given our new, and very familiar looking equation, let’s solve it. If you have
gone through the exercise of calculating the evolution of the wave packet of a
free quantum mechanical particle then the steps below will be very familiar.
We start by Fourier transforming the equation (17) in space. Explicitly, we
conjecture a solution to the equation of the form
D(N ; ~x, ~y ) = d(N, ~q)ei~q·(~y−~x) (19)

7
Figure 3: Five random walkers have stated at a common point and have
taken 100 steps away from it.

This leads to the following differential equation for the function d(N, ~q)
∂d(N, ~q) a2 q 2
=− d(N, ~q) (20)
dN 6
The solution to this equation is
2 q 2 N/6
d(N, ~q) = d(0, ~q)e−a (21)

The next step is to reconstitute the full solution by taking a superposition


of the d(N, ~q)’s. We can do this, given knowledge of the form of D(N ; ~x, ~y )
at N = 0, i.e. before the first step has been taken. A standard initial
condition, in fact one consistent with the scenarios described above, has the
D(0; ~x, ~y ) = δ(~y − ~x). Given what we know about the Fourier transform of
the Dirac delta function, we then have d(0, ~q) = (2π)−3 for all ~q, and
Z
2 2
D(N ; ~x, ~y ) = d(0, ~q)ei~q·(~y−~x)−a N q /6 d3 q
Z
1 2 2
= 3
ei~q·(~y−~x)−a q /6 d3 q
(2π)

8
 3/2
1 6π 2 /2a2 N
= e−3|~y−~x| (22)
(2π)3 a2 N

In getting to the last line of (22), I made use of the fact that the integral over
~q can be broken up into three independent integrations of the components of
that vector and then of the classic result for Gaussian integrations
Z ∞ r
Ax−Bx2 π A2 /4B
e dx = e (23)
−∞ B

Three properties of the solution are worth noting:

1. It is in the form of a Gaussian.



2. Its width scales with N .

3. Its value at the origin goes as N −3/2 .

All three properties represent key qualitative features of diffusion and the
random walk. As a test of the validity of our expression, we can integrate
the expression on the R last line of (22) over all values of ~y . Making use of
(23), we verify that D(N ; ~x, ~y )d3 y = 1.
Given our solution for D(N ; ~x, ~y ), what can we say about the original
quantity of interest, C(N ; ~x, ~y )? In light of the relationship between the two,
it is almost a triviality to obtain the second function from the first. In fact,
we have

C(N ; ~x, ~y ) = 6N D(N ; ~x, ~y )


3/2
6N

6π 2 2
= 3 2
e−3|~y−~x| /2a N (24)
(2π) a N

This tells us that the number of N step walks starting at ~x and ending at ~y has
the first two properties listed above, and additionally increases exponentially
in the number of steps. In fact a general feature of random walks will be this
exponential growth in the number of possibilities with the number of steps.
The power-law modification at the origin that is noted as the third item in
the list should also be kept in mind.

9
The current density of walkers
In the case of dye particles diffusing in water, we can express the invasion
of those impurities in terms of a current density, through the continuity
equation
∂d ~ · ~j
= −∇ (25)
∂N
In light of Eq. (17), we can easily intuit the current density
2
~j = − a ∇
~ y D(N ; ~x, ~y ) (26)
6
The relationship (26) between current density and the density of diffusers is
an example of Fick’s Law.
Suppose, now that there is an absorber, say a wall that sucks up any
walker that hits it. How does that effect the density of walkers? We know
that the overall effect has to be to decrease them, as walkers that hit the wall
are removed from the distribution. Without going into details, allow me to
simply say that the consequence on the density of walkers is to force it to
zero at the points of impact of the walkers and the absorber. To be more
precise, the density extrapolates to zero just beyond the point of impact, but
the difference between the more proper and precise effect and the effective
boundary condition that we will apply is not important for the phenomena
that will be discussed here.
Let’s start by considering the case of a walker, or a collection of walkers,
that begin some distance from an absorbing wall. We want a distribution
that starts out as a delta function, that obeys the equations above in the
region in which the walkers take their steps and that obeys the boundary
conditions that were asserted above at the wall. To be specific, let’s look
at the case of a collection of walkers that start out a distance l from a wall
that occupies the plane x1 = 0. The walkers start out in the half space
x > 0. We will locate the starting point at (l, 0, 0). Then, the solution for
the distribution D(N ; ~x, ~y ) that is consistent with the boundary conditions
is
 3/2
1 6π
D(N ; ~x, ~y ) =
(2π)3 a2 N
 2 2 2 2 2 2 2 2

× e−3((y1 −l) +y2 +y3 )/2a N − e−3((y1 +l) +y2 +y3 )/2a N (27)

10
This is reminiscent of the image charge solution for the electrostatic potential
in the presence of a conducting surface. In fact, one way to derive it is to
posit the existence of “image walkers” that meet and annihilate walkers that
impinge on the absorbing surface. See Fig. 4
x=0

image point starting point

ending point

Figure 4: Illustrating the way in which an image walker eliminates a path


by the original walker that passes over the boundary at x = 0. For every
path which crosses the boundary and returns to the ending point, there is a
path starting at the image point that can reach the same ending point in the
same number of steps, N . This is not the case for walks that do not cross the
boundary. Note that the portion of the image walker’s path that precedes
its first contact with the boundary is the mirror image of the corresponding
portion of the real walker’s path and that after that first contact the image
walker shadows the real one.

Steady state distributions and boundary conditions


We can make even closer contact to the image charge method, and in fact to
the calculation of electrostatic potentials, by altering the situation associated
with a swarm of walkers. The equation (17) expresses the fact that the
only way in which the distribution of walkers will change is through the
propagation of walkers into or out of the region in which the distribution
is being monitored. Suppose, alternatively, that walkers are steadily being
inserted. That is, suppose we amend the equation governing the distribution
of walkers by introducing a source term. This leads to a change in both (17)
and the equations following from it and to the continuity equation (25). Here
is our new version of (17)
∂D(N ; ~x) a2
= ∇2 D(N ; ~x) + s(N, ~x) (28)
∂N 6
11
The last term on the right hand side of (28) is the new source term. Now,
let’s take the source term to be independent of N , so that walkers are being
introduced at a steady rate. Given this, let’s see what the distribution of
walkers settles down to. That is, let’s see what the steady state distribution
is, corresponding to a constant infusion of walkers diffusing out from the
locations at which they are being introduced. In this case, both s and D are
independent of N , and (28) becomes

a2 2
∇ D(~x) = −s(~x) (29)
6
Here, I have removed N from the equation, as there is no dependence on that
quantity.
Of course, we have all seen Eq. (29). It is Laplace’s equation. We all
know how to solve it in special cases. If the source term is a delta function,
corresponding to the introduction of walkers at a particular point in space,
say at ~x = 0, then
3 1
D(~x) = (30)
2πa2 |~x|
The steady state distribution of walkers is governed by the same equation,
and it takes on the same form, as the electrostatic potential. Let’s make
use of this fact to consider a couple of classic problems associated with the
random walk.

Gambler’s ruin
The problem of the gambler’s ruin has been cited by Montroll and Shlesinger
as the first example of a situation that can be analyzed in terms of a random
walk [Montroll and Shlesinger, 1983]. The solution is due to the celebrated
mathematician de Moivre. The question being asked is the following one.
Given a game of chance between two players in which neither one has an
advantage, suppose the first player starts out with an amount of money
equal to A and the second player has an amount of money equal to B. Each
will play until he or she has either won all the money or has exhausted his
or her resources. What is the likelihood that the first player will walk away
the winner?
A way to think of the problem is in terms of a random walker in one
dimension that starts off in a bounded interval. The walker is a distance

12
A from one of the boundaries and B from the other one. As soon as the
walker hits one of the boundaries it is absorbed, corresponding to the “ruin”
of one of the gamblers. Now, to apply the steady state model to the analysis
of Gambler’s Ruin, we imagine a constant source of walkers at the starting
point, located at x = A as shown in Fig. 5. This replaces the single pair
of gamblers by an ensemble of them. We can think of having a very large
tournament in which games are being started at a constant rate and continued
until a player is out of money. This source leads to a steady state distribution

x
A B
Figure 5: The source, at the location of the vertical dashed line, induces the
steady state distribution d(x), as shown.

of walkers, also indicated in the figure. The distribution of walkers, d(x)


obeys the one dimensional version of the Poisson equation of electrostatics:
a2 d2 d(x)
= −s(x)
6 dx2
= s0 δ(x − A) (31)
Given that the density of walkers vanishes at the absorbing boundaries, we
can figure out the solution to this equation. As the second derivative is equal
to zero, we know that d(x) has to be a linear function. The delta function
generated by the second derivative is the consequence of a discontinuity in
the slope of the function. If the function itself were discontinuous at x = A,
then its first derivative would give rise to a delta function and the second to
the derivative of a delta function. This means that d(x) has the form
 K
A
x 0<x<A
d(x) = K (32)
B
(A + B − x) B < x < A + B
The coefficient K follows from matching the magnitude of the slope dis-
continuity, K(1/A + 1/B), to the required amplitude of the delta function.

13
Fortunately for the purposes of completing our analysis we do not need to
work out the value of K. We simply need to ratio of random walker currents,
and this is easy given (26). According to that version of Fick’s law, we have
for the walker current density
d
j(x) ∝ − d(x)
 dx
−1/A 0 < x < A
∝ (33)
1/B A<x<A+B

In other words, walkers are going off the the left and eventually being ab-
sorbed by the boundary at x = 0 at a relative rate 1/A, while walkers are
wandering off to the right at a relative rate equal to 1/B. As exit to the left
corresponds to the ruin of the gambler with resources amounting to A, while
exit to the right corresponds to success for that gambler and ruin for the
gambler that starts with an amount equal to B. Let’s call the first gambler
Gambler A and the second one Gambler B. Then the probability PA that
Gambler A wins divided by the probability that Gambler B wins is given by

PA 1/B
=
PB 1/A
A
= (34)
B
The gambler with more resources is more likely to win, and the relative
likelihood is equal to the ratio of initial resources.
And that is why you are statistically doomed to lose in Las Vegas. Even
if the odds are not against you, unless you can match the resources of the
house you will eventually end up on the short end.

Walkers near an absorbing sphere


We can also ask what happens to walkers in the vicinity of an absorbing
sphere. First, we’ll look at the case of a collection of walkers that are supplied
by a set of sources that are infinitely far away from the sphere. The sources
are such that the distribution of walkers is uniform at an infinite distance
from the sphere. To find what the concentration of walkers is at an arbitrary
distance from the sphere, which has a radius equal to r0 , we search for a
solution of Laplace’s equation (because there are no sources in the region

14
of interest) that is equal to zero at the surface of the sphere and that is a
constant infinitely far away from the sphere. Placing the origin at the center
of the sphere, we can also demand spherical symmetry about that origin.
The solution of interest is then readily intuited. It is
 r0 
d(~r) = c0 1 − (35)
r
where r is the distance from the center of the sphere, and the expression on
the right hand side of (35) holds when r > r0 .
We can use this formula to calculate the rate at which the walkers are
absorbed by the sphere. To do this, we calculate the current density of
walkers by taking the gradient of the right hand side of (35), and then by
calculating the current flux into the sphere. From (26), we have

~j(~r) = −βr̂ ∂ c0 1 − r0
 
∂r r
βc0 r0
= −r̂ 2 (36)
r
where I have replaced the constant depending on the size of the walker’s step
by the all-purpose symbol β. This tells us that the flux of walkers into the
sphere is equal to
r0
βc0 2 × 4πr02 = 4πβc0 r0 (37)
r0
We learn that the number of walkers that are absorbed by the sphere scales
linearly with the radius of the sphere. If the sphere represents a cell, and the
walkers are nutrients in the broth in which the cell sits, the rate at which
the cell takes those nutrients in is proportional to its linear size. On the
other hand, if the sphere really is a cell, it has metabolic requirements that
scale as its volume—in other words, as r03 . Ultimately, those requirements
will overwhelm the cell’s ability to absorb nutrients through diffusion, as the
size of the cell increases. In the case of an immobile cell, these considerations
place a limit on the maximum size that it can be. In general, the fact that
metabolic needs will exceed the rate at which nutrients can be gathered as
they diffuse inwards will mandate a different strategy for the acquisition of
biological fuel for any organism that is larger than a certain size [Berg, 1993].
It is also possible to calculate the distribution of walkers when there is
a point source outside of an absorbing sphere. In this case, we make use of
a modification of the image charge. If the source is a distance R away from

15
the center of a sphere with radius r0 , then the image source is a distance
ρ = r02 /R from the center of the sphere, and the “strength” of the source
is equal to −r0 /R times the strength of the original, external source. It is
possible to calculate how many of the walkers escape from the sphere by
taking the integral of the current flux through a surface that surrounds both
the source and the absorbing sphere. We can short-circuit this calculation
by noting that a version of Gauss’s law holds here, which tells us that the
net flux through a surface surrounding a set of sources is proportional to the
total strength of those sources. In this case, the total strength is equal to
the strength of the original source plus that of the image source, which is the
strength of the original source multiplied by 1 − r0 /R. The fraction of the
total number of walkers that emanate from the source that also escape from
the sphere is (R − r0 )/R.

r0 original source
ρ R

image source

Figure 6: The source and the image source in the case of an absorbing sphere.

What we have also achieved here is a derivation of the probability of


escape by a three-dimensional walker from a spherical region in the vicinity
of its point of origin.

The analogy with capacitance


There is a mathematical connection between an absorbing surface into which
matter diffuses from a distance and the element of a capacitor . Because of
the relationship between the concentration and the electrostatic potential,
we can imagine an analogy in which the surface is that of a conductor, sur-
rounded by a spherical shell some distance away. If the potential difference
between the two is ∆φ, and the charge on the inner surface is Q, then the
capacitance of the system is c = Q/∆φ. The connection between charge and
the electrostatic potential is ρ(~r) = −4π∇2 φ(~r). The total charge on the

16
inner surface is, then as given by Gauss’s law
Z
1 ~ r) · dS
~
Q=− ∇φ(~ (38)

Making use of the relationships we’ve already established between electro-
static quantities and those in steady state diffusion, we can state that the
following holds
~
− ~j(~r) · dS
R
= 4πβC (39)
c∞
where c∞ is the concentration of walkers far away from the absorbing surface
and C is the capacitance of a capacitor consisting of the absorbing surface
surrounded, at a great distance, by a spherical shell. See Figure 7. For

inner
surface

outer
sphere

Figure 7: The capacitor consisting of an inner surface surrounded by a spher-


ical shell. The radius of the surrounding shell is much greater than the size
of the inner surface.

reference, recall that the capacitance of a sphere is given by


Csphere = r0 (40)
where r0 is the sphere’s radius. This and (39) yields (37) for the total current
absorbed by the sphere, where we have replaced c0 by c∞ .

A sphere covered with receptors


This allows us to consider what happens if the nutrients are not absorbed
uniformly throughout the sphere. Suppose, instead, that the sphere has a

17
number of absorbing sites, or receptors, distributed across its surface. Imag-
ine that there are n of those receptors, and that the radius of a given site
is a. We’ll also assume that na  r0 , where r0 is the sphere’s radius. This
means that the total surface area accounted for by the receptors is a small
fraction of the surface area of the sphere. Let’s represent this collection of
receptors as a network of small conducting surfaces, arranged in the shape
of a spherical shell. This network is utilized as one of the elements in a
capacitor . What is the capacitance of the resulting capacitor? √
First, note that the distance between the receptors is the order of r0 / n.
Then, notice
√ that the potential a distance r away from the sphere, where
r  r0 / n, is going to be the same as the electrostatic potential generated
by a uniform distribution of the receptors, smeared out over the sphere. This
is because at such a distance, the difference between a set of discrete charge
and a uniformly distributed charge is negligible, as far as the electrostatic
potential goes. If a charge of Q/n is placed on each receptor, then the electro-
static potential of this array of charges goes as Q/(r + r0 ). The electrostatic
potential near the sphere is, then, well approximated by Q/r0 . This means
that the capacitance of the spherical arrangement is substantially equal to
the capacitance of a uniform sphere. Making use of the electrostatic analogy,
we find that the collection of receptors will absorb nutrients at the same rate
as if the entire surface of the sphere were a receptor.
We can be a little more explicit about the potential immediately above
the surface of the network. The electrostatic potential right next to one
of the small components of the network due to the charge carried by that
component will go as Q/na, because each of the components carries a charge
equal to Q/n, and each has an effective size equal to a. On the other hand,
the potential due to all the other components will be essentially the same
as if the charges on them were uniformly distributed over the surface of the
sphere. This potential is equal to Q(n − 1)/nr0 . If n is large enough that
na  r√ 0 , which is possible for sufficiently large n, since all we require is that
a  r/ n, the potential is dominated by the second contribution, which, in
the limit of large n, goes to Q/r0 . The potential at a point near the surface
of the sphere that is far away from one of the small components compared
to its size will be absolutely dominated by the second term. Thus, to a very
good approximation, the electrostatic potential in the immediate vicinity of
the network is the same as the electrostatic potential right next to a sphere
carrying a uniform charge equal to Q.
For a more extended discussion of the issued addressed in the last two

18
sections see the references [Berg, 1993] and [Berg and Purcell, 1977].

Finally, the generating function


At the outset I made reference to the utility of the generating function. Now
it is time to actually talk about the object. Briefly, the generating function,
G(z; ~x, ~y ) has the same relationship to the function C(N ; ~x, ~y ) that the grand
partition function in statistical mechanics has the partition function. Here,
explicitly, is the expression for the random walk generating function

X
G(z; ~x, ~y ) = C(N ; ~x, ~y )z N (41)
N =0

The quantity z plays the role of the fugacity, eβµ in statistical mechanics.
The function G(z; ~x, ~y ) is called the generating function because when it is
expanded in a power series in z the coefficients of that series are the functions
C(N ; ~x, ~y ). Of course, this relationship is pretty obvious, given the definition
of G(z; ~x, ~y ).
Thus, the generating function encapsulates information about the statis-
tical properties of random walks of all lengths. However, what makes the
generating function so absolutely valuable is the fact that its determination
is in many cases so much more straightforward than is the case for functions
describing random walks with a fixed number of steps. In this respect, it
possesses the same advantages as the grand partition function in many case,
and example being quantum statistical mechanics, the natural approach to
which is in the grand canonical ensemble.
To see how recourse to the generating function simplifies our consideration
of random walk statistics, consider the recursion relation (3). Let’s multiply
both sides of this equation by Z N +1 and sum from N = 0 to ∞. Then, we
find

X ∞
X
N +1
Z C(N + 1; ~x, ~y ) = z M C(M ; ~x, ~y ) − C(0; ~x, ~y )
N =0 M =0
= G(z; ~x.~y ) − δ~x,~y

X X
= z zN C(N ; ~x, ~y )
N =0 w
~i

19
X
= z G(z, ~x, w
~ i) (42)
w
~i

In deriving the expression on the second line of (42), we made use of the
fact that the the number of zero step walks starting at ~x and ending at ~y
is equal to one if the two locations are identical and is zero otherwise. The
relationship that can be abstracted from (42) is
X
G(z; ~x, ~y ) = z G(z; ~x, w
~ i ) + δ~x,~y (43)
w
~i

The recursion relation becomes an equation that is algebraic in the variable


z. To further reduce it, we will focus on the situation in which there are no
boundaries or any other object providing a reference point in space, which
means that translational invariance holds. The generating function will then
depend on the positions x and y only through their difference. In other words
we can write the generating function in the form G(z; ~y − ~x). If we introduce
the Fourier transformed generating function g(z; ~q), given by
X
g(z; ~q) = e−i~q·(~y−~x) G(z; ~y − ~x) (44)
~
x

then, multiplying Eq. (43) by e−i~q·(~y−~x) and summing over ~x, we obtain
X
g(z; ~q) = e−i~q·(~y−~x) G(z; ~y − ~x)
~
x
XX X
= z e−i~q·(~y−~x) G(z; w
~ i − ~x) + e−i~q·(~y−~x) δ~x,~y
~
x w
~i ~
x
X X
= z ei~q·(w~ i −~y) e−i~q·(w~ i −~x) G(z; w
~ i − ~x) + 1
w
~i x
X
= z ei~q·(w~i −~y) g(z; ~q) + 1
w
~i
= zχ(~q)g(z; ~q) + 1 (45)

The
P function χ(~q) in the last line of (45) is shorthand for the quantity
q ·(w
i~ ~ i −~
y)
~i e
w . Again, we abstract an equation for g(z; ~q) from the several
lines of (45). The equation is

g(z; ~q) = zχ(~q)g(z; ~q) + 1 (46)

20
The solution is easy. It is
1
g(z; ~q) = (47)
1 − zχ(~q)
To extract the generating function in real space, we evaluate the Fourier
transform of the function g(z; ~q). We find
Z
G(z; ~y − ~x) ∝ g(z; ~q)ei~q·(~y−~x) dd q

ei~q·(~y−~x) d
Z
= d q (48)
1 − zχ(~q)
Note that I have left the dimensionality of the integration in (48) unspecified.
This because our general result is valid in all dimensions. The proportional-
ity sign and the lack of explicit indications of the range of integration over
the wave vector variable ~q arises from the lack of specification with regard
to the exact conditions under which the random walker moves. Once that
specification is supplied (Is the walker confined to the vertices of a lattice?
If so, what kind of lattice? If not, what kind of randomness is there in the
walk—in direction only or both in direction and in the length of the step?),
both the overall multiplying constant and the range of ~q-integration follows.
Given the generating function, we are now in a position to extract, if we so
desire, the statistical properties of the N -step walk. Given that we’ve gone to
the trouble to derive the expression on the last line of (48) for the generating
function, let’s see what we get when we do the power series expansion in z
of it. Fortunately the first step is easy. We have

1 X
= z n χ(~q)n (49)
1 − zχ(~q) n=0

Performing the expansion in the integrand in the last line of (48), we end up
with the following result for C(N ; ~x, ~y ).
Z
C(N ; ~x, ~y ) ∝ ei~q·(~y−~x) χ(~q)N dd q (50)

The exact result will depend on the exact form of the function χ(~q). How-
ever, we can extract the important properties of the generating function by
expanding that function as a power series in ~q. To make contact with earlier

21
results, we will look at a walker that is, again restricted to the vertices of a
cubic lattice. The distance between neighboring points on that lattice will
be a, as previously specified. Then
X
ei~q·(w~ i −~y)
w
~i

= eiaq1 + e−iaq1 + eiaq2 + e−iaq2 + eiaq3 + e−iaq3


= 6 − a2 (q12 + q22 + q32 ) + O(ql4 ) (51)

We now recast the integral leading to the generating function and make use
of the expansion in (51).

C(N ; ~x, ~y )
Z
∝ ei~q·(~y−~x) eN ln χ(~q) d3 q
Z
ei~q·(~y−~x) exp N ln 6 − a2 (q12 + q22 + q32 ) d3 q
 
=
N a2 2 3
Z  
q ·(~
i~ y −~
x)
= e exp N ln 6 − |~q| d q
6
 3/2
6π 2 2
= 6 N
2
e−3|~y−~x| /2N a (52)
Na
Compare this to (24). We have the number of walks to within an overall
multiplicative constant, which, after all, was left as an open item.
Now, let’s turn our attention to the properties of the generating function
itself. As it turns out, the most important properties are determined by
the low order terms in the expansion of χ(~q). In other words, it suffices to
consider the quantity
1 1

1 − z(χ(0) + q 2 χ00 (0)) 1 − (z/zc ) + Azq 2
1
→ (53)
1 − z/zc + zc Aq 2
We can, for instance, make use of the last line of (53) to reconstruct the
generating function in real space:
e−~q·(~y−~x)
Z
G(z; ~x, ~y ) ∝ dd q (54)
1 − z/zc + zc Aq 2

22
We can perform the integral in (54) in three dimensions. We have
eiq|~y−~x| cos θ sin θ
Z Z
2
G(z; ~x, ~y ) ∝ 2π q dq dθ
1 − z/zc + zc Aq 2
Z ∞
4π sin(q|~y − ~x|)
= qdq
|~y − ~x| 0 1 − z/zc + zc Aq 2
√ √
2π 2 e− 1−z/zc |~y−~x|/ zc A
= (55)
Azc |~y − ~x|
The last line of (55) applies for |~y − ~x| not too small. The fact that the
expression diverges as ~y → ~x is an artifact of our approximations and of
the absence of a restriction on the integration over ~q. One thing to note is
that, aside from the divergent term—which is absent if we do the integration
properly—the
√ leading order contribution to the generating function as ~y → ~x
goes as zc − z. This allows us to work out the dependence on N of the
number of walks that return to the point from which they started. First,
though, a digression on the art of extracting the coefficient of z N in the
power series of a function.

Extraction of coefficient of z N in the power series ex-


pansion of a function with various singularities
In the cases of interest to us, the generating function displays singular struc-
ture as a function of the variable z in the vicinity of a “critical value,” zc .
Here is a summary of the ways in which the coefficients of z N behave for
N  1 in the cases of various possible singularities.

A simple pole
Let’s start with one of the simplest examples of a function with an infinite
power series expansion in z: f (z) = 1/(zc − z). If | z |<| zc |, we have
"  2 #
1 1 z z
= 1+ + + ···
zc − z zc zc zc
∞  n
1 X z
= , (56)
zc n=0 zc
−(N +1)
so the coefficient of z N is zc .

23
Two or more simple poles
Now, suppose zc and zd are both real, positive numbers and zc < zd . Fur-
thermore, let
a b
f (z) = + . (57)
zc − z zd − z
Then, if z is suficiently small(z < zc ),
∞  n ∞  n
a X z b X z
f (z) = + , (58)
zc n=0 zc zd n=0 zd

−(N +1) −(N +1)


so the coefficient of z N is cN = azc + bzd .
We can also write
"  N +1 #
b zc
cN = azc−(N +1) 1 + .
a zd

Now, let zc = zd (1 − ∆), where ∆ > 0. Then


 
−(N +1) b N +1
cN = azc 1 + (1 − ∆)
a
 
−(N +1) b (N +1) ln(1−∆)
= azc 1+ e
a
 
−(N +1) b −δ(N +1)
= azc 1+ e , (59)
a

where δ = − ln(1 − ∆) and δ > 0. As N gets larger and larger the second
term in brackets in (59) vanishes exponentially. Thus, for very large N the
coefficient of z N in a/(zc −z) + b/(zd −z) is essentially equal to the coefficient
of a/(zc − z). We will eventually generalize this result as follows:

If the functions fa (z) and fb (z) have poles or branch points at za and zb ,
respectively, and if zb > za > 0 (za and zb both real), then, when N is large,
the coefficient of z N in Afa (z) + Bfb (z) is, for all practical purposes, equal
to the coefficient of z N in Afa (z).

24
Higher Order Poles and Branch Points
What about the more general case f (z) = (zc − z)−α , where the exponent α
need not be an integer? One way of finding the coefficient of z N is to use the
binomial expansion formula. Another way is the use the following identity:
Z ∞
tA e−xt dt = x−A−1 Γ(A + 1), (60)
0

where Γ(A) is the gamma function. When A is an integer, n, Γ(n + 1) = n!.


With the use of (60) we have
Z ∞
−α 1
(zc − z) = tα−1 e−t(zc −z) dt. (61)
Γ(α) o

To find the coefficient ofz N in (zc − z)−α we expand the right hand side of
the above equation with respect to z. The coefficient of z N in that expansion
is
Z ∞
1 1 1 1 −(N +α)
tα−1+N e−tzc dt = z Γ(α + N )
Γ(α) N ! 0 Γ(α) N ! c
Γ(α + N )
= zc−(N +α) . (62)
Γ(α)Γ(N + 1)

Now, we use Stirling’s formula for the gamma function of a large argument

Γ(N ) = e(N −1) ln(N −1)−(N −1) (63)


N 1

When N is large, the coefficient of interest is

zc−N −α (α+N −1) ln(α+N −1)−(α+N −1)−N ln N + N zc−N −α (α−1) ln(N )


e = e
Γ(α) Γ(α)
z −N −α α−1
= c N (64)
Γ(α)

Logarithmic Singularities
One more complication: suppose the function is of the form −(zc −z)−α ln(zc −
z). We obtain the coefficient of z N by noting that this function is just

25
d
(z
dα c
− z)−α . Taking the derivative with respect to α of the last term in the
equation above, we have for the desired coefficient

d zc−N −α α−1 z −N −α N α−1 Γ0 (α)zc−N −α α−1 zc−N −α N α−1


N = ln(zc ) c − N + ln(N )
dα Γ(α) Γ(α) Γ(α)2 Γ(α)
−N −α
  
z 1
= ln(N ) c N α−1 1 + O (65)
Γ(α) ln(N )

By the same token, we can find the coefficient of z N in −(zc −z)−α / ln(zc −z)
by extracting the coefficient of z N in the integral
Z ∞
(zc − z)y dy, (66)
−α

Using (64) to find the coefficient of z N in the integrand, one is left with
Z ∞ −N +z ∞ Z ∞
N z−1 zc−N +z
 −N +z 
zc z−1 1 z−1 d zc
N dz = − − N dz
−α Γ(−z) ln N Γ(−z) −α ln N −α dz Γ(−z)
N α−1 zc−N −α
  
1
= 1+O . (67)
ln(N )Γ(α) ln N

The first equality in (67) results from an integration by parts. An additional


integration by parts establishes the second equality.

Back to the generating function


Now, let’s see how the generating function depends on the difference zc − z
when we let the end point approach the location from which the walker starts.
We will look this behavior in various dimensions. We start by setting ~y = ~x.
Then,

dd q
Z
G(z; ~x = ~y ) ∝
1 − z/zc + Aq 2
q d−1 dq
Z
∝ (68)
1 − z/zc + Aq 2
To simplify expressions, I have replaced the combination zc A by A in the
denominator. There is a quick and dirty way to extract the singular depen-
dence of the integral on the last line of (68) on zc − z, and that is to scale the

26
combination out of the integrand. We do this with a change of integration
variables, replacing q by (1 − z/zc )1/2 q 0 . This leaves us with the integral

(1 − z/zc )d/2 q 0 d−1 dq 0


Z Z 0 d−1 0
(d−2)/2 q dq
0
= (1 − z/zc )
(1 − z/zc ) + Aq 2 1 + Aq 0 2
= (1 − z/zc )(d−2)/2 C (69)

Our ability to successfully scale (1 − z/zc ) out of the integral requires


that we not have to worry about limits on the integration. By and large,
this does not prove to be a problem, except in the special case d − 2. In two
dimensions, we have the following integral to perform
 Q
Z Q
q dq 1  2
2
= ln 1 − z/zc + Aq
0 1 − z/zc + Aq 2A 0
1  
ln 1 − z/zc + ln Q2 − ln [1 − z/zc ] (70)

=
2A
A couple of points. First, I put an upper limit, Q, on the integral, because
otherwise it is not convergent. The second is that the value of Q is actually
irrelevant to the singularity of interest, which arises from the contribution
from the lower limit at Q = 0.
Now, making use of the methods described on pages 23–26, we find that
the coefficient of z N in the expansion of the generating function G(z; ~x = ~y )
goes as N −d/2 . I should say that we would not have had to go through the
indirect route of the generating function to obtain that result. We could
have obtained the same result by generalizing the integral over ~q in (22) to
d dimensions.

Recurrence
One of the important properties of a random walk is related to the likelihood
that a walker will visit a particular site. An issue related to this property
has to do with the question of recurrence, which is to say, the problem of
calculating the likelihood that a walker returns to its point of origin. As it
turns out, the answer to this question depends importantly on the dimension
in which the random walk is executed. Here we will work out an expression
that allows us to answer the question of recurrence, and, when properly
extended, to work out the number of different sites visited by a random
walker. First, though, an altered version of the generating function

27
A new generating function
A useful—but apparently little known2 —quantity enables one obtain some
key results with remarkably little effort. This quantity is an expanded version
of the generating function we’ve been utilizing, and it refers to a walk that
may or may not visit a special site. Suppose the quantity C(N, M ; ~x, ~y , w)
~ is
the number of N -step walks that start at the location ~x, end at the location
~y and visit the site at location w
~ exactly M times in in the process of getting
from ~x to ~y . The quantity of interest is

X
0
C (N, t; ~x, ~y , w)
~ = C(N, M ; ~x, ~y , w)(1
~ − t)M (71)
M =0

Clearly, terms in the sum on the right-hand side of the above equation for
which M > N + 1 will not count, as there is no walk that visits a site more
times than it leaves footprints. That is C(N, M ; ~x, ~y , w)
~ = 0 if M > N + 1.
We obtain the coefficients of the power series expansion in (1 − t)n that
produces this generating function in the standard way. It is easy to show
that
M

1 d 0

~ = (−1)M
C(N, M ; ~x, ~y , w) C (N, t; ~
x , ~
y , w)
~ (72)
M ! dtM
t=1
Note that in this case the value to which the expansion parameter is set is
not zero, but rather one.

Derivation of the new generating function


We derive the new generating function by introducing a weighting factor.
The weighting factor W has the following form for an N -step walk that
visits the site si at the ith step
N
Y
W = w(si ) (73)
i=1

where factor w(si ) is given by

w(si ) = 1 − tδsi ,S (74)


2
Little known among practicing solid state physicists, that is.

28
Here, S is the special site of interest. The overall weighting factor for a given
walk is then
YN
(1 − tδsi ,S ) (75)
i=1
Suppose t = 1. Then, the weighting factor will have the effect of excluding
any walk that visits the special site S; all other walks will have a weighting
factor of one. This means that if we multiply all walks by the weighting
factor above, set t = 1 and sum, we will obtain the number of walks that
never visit the site S. In the case of N -step walks that start at ~x and end
up at ~y , this is just C(N, 0; ~x, ~y , w),
~ where w
~ is the position vector of the site
S. Suppose we take the derivative of the weighted sum with respect to t,
and then set t = 1. In that case, we will end up with −1 times the number
of walks that visit the site only once. Next, take the nth derivative of the
weighted sum over walks with respect to t, multiply by 1/n!, and then set
t = 1. This yields (−1)n times the number of walks that visit the special
site exactly n times. This is because each derivative generates a factor equal
to −δsj ,S and all terms that “escape” the derivative become (1 − δsk ,S ) when
t = 1. The factor 1/n! compensates for the n! ways in which the n derivatives
with respect to t operate on the product in (75).
Now, we can evaluate the weighted walk by expanding the weighting
factor, (75), in powers of t.3 At first order we generate the quantity
N
X
−t δsi ,S (76)
i=1

When walks are multiplied by this weighting factor and summed we end up
with −t times the sum of all N -step walks that visit the site S at one step,
with no restriction on what happens either before or after that step. At
second order in the expansion we have
N X
X i−1
t2 δsi ,S δsj ,S (77)
i=1 j=1

When walks are multiplied by this second-order weighting factor and summed,
we end up with t2 times the sum of all N -step walks that visit the site S
3
We are thus treating t as if it were a small quantity and expanding in it. This technique
will be used later for a different expansion parameter in the case of self-avoiding walks, in
which case we will generate a virial expansion.

29
twice with no restriction on what happens before, after or between those
two visitations. A graphical representation for the expansion in t of the new
generating function is shown in Fig. 8. If the starting and end-point of the

-t

3
+t2 -t

Figure 8: The graphical representation of the virial expansion of the gen-


erating function defined in (71). The large dots in the diagrams lie at the
location w.
~

walk are fixed, and if the site in question is at the location v, then the sum
in Fig. 8 is given by
X
C(N ; ~x, ~y ) − t C(N1 ; ~x, w)C(N
~ 2 ; w,
~ ~y )
N1 +N2 =N
X
2
+t C(N1 ; ~x, w)C(N
~ 2 ; w,
~ w)C(N
~ ~ ~y ) + · · · (78)
3 ; w,
N1 +N2 +N3 =N,N2 ≥1

The inequality that applies to N2 in the sum above simply requires the walker
to take at least one step before revisiting the site at w.
~ Otherwise, we would
count zero step “walks” in the sum.
Now, we take the step of multiplying our new function by z N and sum-
ming. This has the effect of removing the convolution over the Ni ’s, and we
obtain

X
G(z, t; ~x, ~y , w)
~ = z N C 0 (N, t; ~x, ~y , w)
~
N =0
= G(z; ~x, ~y ) − tG(z; ~x, w)G(z;
~ w,
~ ~y )

30
+t2 G(z; ~x, w)G
~ 1 (z; w,~ w)G(z;
~ ~ ~y ) + · · ·
w,
G(z; ~x, w)G(z;
~ w,
~ ~y )
= G(z; ~x, ~y ) − t (79)
1 − tG1 (z; w,
~ w)
~

The subscript 1 in G1 (z; w,


~ w)~ indicates that it is a generating function for
walks of at least one step.
The quantity

dM

M 1

(−1) G(z, t; ~
x , ~
y , w)
~
M ! dtM
t=1
M
 
M 1 d G(z; ~x, w)G(z;
~ w,
~ ~y )
= (−1) G(z; ~x, ~y ) − t (80)
M ! dtM 1 − tG1 (z; w,
~ w)
~
t=1

is, then, the generating function for all walks that start at ~x, end up at ~y
and visit the site at w
~ exactly M times.

Dimensionality and the Probability of Recurrence


Our result allows us to determine whether or not a random walk is recurrent.
If it is, then almost all long random walks that start out at a given point
will revisit that point. If it is not, then only a finite fraction of those walks
do so, and the walk is called transient. Recurrent walks are also referred
to as “persistent.” Pólya [Pólya, 1919] was the first to demonstrate that
walks occuring in one and two dimensions return to their starting point with
absolute certainty, if they consist of an infinite number of steps and that in
higher dimensions the walker has a non-zero probability of never revisiting
its starting point, no matter how long its walk. Let’s see what (79) tells us
about the recurrence of random walks on a lattice. We want to know how
many of the walks that start out at ~x revisit their point of origin, so we set
w
~ in (79) equal to ~x. The generating function takes the form

G1 (z; 0)G(z; ~x − ~y ) G(z; ~x − ~y )


G(z; ~x − ~y ) − t = . (81)
1 + tG1 (z; 0) 1 + tG1 (z; 0)

The reason for the subscript 1 in the numerator of the second term on the left
hand side of (81) is that we want to count only those walks that take at least
one step from the starting point at x before revisiting that point. Otherwise,
we count walks that “revisit” their point of origin after zero steps.

31
To find out how many N -step walk start out at ~x and end up at ~y , never
having revisited ~x, we need to calculate the coefficient of z N in the function

G(z; ~x − ~y ) G(z; ~x − ~y )
= . (82)
1 + G1 (z; 0) G(z; 0)

The right hand side of the equality in (82) follows from the fact that the
contribution to the generating function G(z; 0) of walks consisting of no steps
is exactly one, by convention.
Let’s be even less restrictive and ask how many of the walks that start
out at ~x and end up anywhere never revisit the starting point ~x. We simply
sum the expression above over all possible end-points ~y —excluding ~x—and
see what we have. Using the relation between G(z; ~x − ~y ) and its spatial
fourier transform, g(z; ~q) we have
X G(z; ~x − ~y ) g(z; 0) G(z; 0)
= −
G(z; 0) G(z; 0) G(z; 0)
y 6=~
~ x
g(z; 0)
= − 1, (83)
G(z; 0)

where we have used the identity


X
g(z, 0) = G(z, ~x − ~y ) (84)
~
y

for the fourier coefficient g(z, ~q). From Chapter 2, we know that
1
g(z, ~q = 0) =
1 − zχ(~q = 0)
1
≡ (85)
1 − z/zc

The last line of (85) serves as a definition of the quantity zc . This tells us
that the number of N -step walks that start out at ~x and end up anywhere
is the coefficient of z N in (1 − z/zc )−1 , while the number of walks that start
at ~x and end up anywhere, not having ever revisited the point of origin, ~x,
is the coefficient of z N in
zc 1
(86)
zc − z G(z; 0)

32
As it turns out, the z-dependence of the generating function G(z; 0) in two
dimensions or less is different in a very important respect from its behavior
in three and higher dimensions. This will lead to fundamentally different
results for the “recurrence” of walks in two and one dimensions from what
we will find in the case of three dimensional walks.

The Notion and Quantification of Shape


Shape is an intuitively accessible notion. We organize visual information
in terms of shapes, and the shape of an object represents one of the first
of its qualities referred to in an informal descriptive rendering of it. While
our language presents us with a wide repertoire of verbal images for the
approximate portrayal of the shape of a physical entity (“round,” “oblong,”
“crescent,” “stellate” . . . ) the precise characterization of a shape, in terms of
a number, or set of numbers, has remained elusive. This is with good reason.
It is well-known to mathematicians that the class consisting of the set of all
curves is a higher order of infinity than the set of all real numbers. This
means that there can be no one-to-one correspondence between curves and
real numbers. As shapes, intuitively at least, bear a conceptual relationship
to curves, it is plausible that the set of all shapes dwarfs in magnitude the
set of real numbers, or of finite sets of real numbers.
On the other hand, if one is willing to content oneself with a general
paradigm for the measurement of shape, there are ways of quantifying it in
terms of numbers that have a certain descriptive and predictive utility. In
fact, the numerical specification of shapes has acquired a certain urgency of
late, in light of the widespread use of computer imaging and the concommi-
tant focus on the development of codes for the creation and manipulation of
pictorial quantities.
In this Chapter, we will look at different ways of characterizing and mea-
suring the shape of a random walk. We will focus on one particular method,
based on calculations of the width of the distribution of steps about the “cen-
ter of mass” of the walk. The particular quantity studied is the radius of
gyration tensor, and the shape of the polymer is quantified in terms of the
eigenvalues of this tensor, termed the principal radii of gyration of the walk.
We will look at a particular combination of the principal radii of gyration
that provides information with respect to the deviation from spherical sym-
metry of the shape of the walk. It will turn out that shape as a concept is,

33
as one might expect, a bit elusive. For one thing, there is no generic “shape”
for a random walk. However, statistical statements can be made, with regard
to the probability that a walk takes on a particular shape, at least as char-
acterized by the principal radii of gyration. In addition, there is one limit in
which the shape of the trail left by a walker is fixed and predictable. That
is the limit of a walker in an infinite dimensional space (d = ∞). We will
discuss the construction of an expansion about that limit, the 1/d-expansion.
This expansion yields the shape distribution of a random walker’s trail when
the walker wanders in a high spatial dimension environment. As we will see,
this expansion is—at least for some puroposes—respectably accurate in three
dimensions.

Figure 9: A “cloud” consisting of the paths of 1,000 random walkers, each of


which has taken 100 steps from a common point of origin.

Anisotropy of a random walk


When we talk about the distribution of points visited by a random walker,
we generally do so in the context of ensemble averages. That is, we ask on
average how many walks visit a given point. Looking at things this way can
obscure the detailed structure of a given random walk. For example, if we are

34
interested in how many time a given point at location ~r1 is visited by a walker
that starts out at the location ~r0 , we find, after suitable averaging, that the
answer depends only on the distance between those points in space, |~r1 − ~r0 |.
This is true because for every walker that tends to go off in one direction
there will be another walker that ends up going in the opposite direction.
The statistical distribution of places visited is rotationally symmetric about
the point of origin. In other words, the totality of walkers in the ensemble
create a “cloud” that is spherically symmetric. Figure 9 shows just such a
cloud, which consists of the paths of 1,000 random walkers each of whom has
taken 100 steps from a common point of origin. The near spherical symmetry
of the cloud is evident from the figure.
This result of averaging obscures the fact that a given random walk can
be quite anisotropic spatially. Figure 10 is a stereographic pair of images of
a single 1,000-step random walk. The elongated nature of the walk shown
in this figure is not a statistical anomaly. Figure 11 shows several examples
of 1,000-step walks. Note that not one of those walks is reminiscent of the
cloud of walkers shown in Figure 9.

Figure 10: Stereographic pair of images of a 1,000 step three-dimensional


random walk.

On the other hand, as Figure 11 makes abundantly clear, no typical, or


definitive, shape can be assigned to a random walk. How, then, to quantify
the shape of a walk?

35
Figure 11: Several examples of a 1,000-step random walk.

Measures of the shape of a walk


There are in the literature a number of algorithm for the characterization
of the shape of an object [Bookstein, 1978, Costa and Cesar, 2001]. Here,
we choose one that is particularly well-suited to our needs. We construct a
moment-of-inertia-like tensor (See Supplement 1 at the end of this chapter).
By diagonalizing this tensor we are able to extract numbers that quantify the
linear “size” of the walk in various directions, particularly in the directions
in which it has the greatest linear extent and the direction in which it is most
compact. For a visualization of this, see Figure 12, in which the extensions
in both directions of a two-dimensional random walk is illustrated. The
thick lines indicate the directions in which random walk has the greatest and
the smallest extension, as determined by the tensor that we are about to
introduce. Note that these two lines provide a quantitative representation of
both the overall orientation of the walk and of its spatial anisotropy—that is,
degree to which the shape of the walker’s path differs from that of a sphere.
As we will see, the amount of anisotropy exhibited by the walks in Figures
10 11 and 12 is not at all atypical. 4
4
For a characterization of the anistropy of a random walker using the notion of spans,
see [Weiss and Rubin, 1976]. Here, our approach will be somewhat different.

36
Figure 12: The anisotropic nature of a 1,000-step two-dimensional walker.
The red lines indicate the directions in which its linear extent is the greatest
and the smallest. The lines also run parallel to the eigenvectors of the matrix
defined in Eqs. (87)–(89). The point of intersection of those two lines is the
“center of gravity” of the walk.

The radius of gyration tensor


The tensor that we are about to define is also utilized to determine the
rotational inertia of a three-dimensional object. Supplement 1 at the end of
this chapter reviews its use in that context. What this means is that the
results to be derived here are relevant to the rotational motion of an object
that mimics the form of the path followed by a random walker, assuming that
the constituents of this object have an inertial mass, that they are uniformly
distributed along the path it imitates, and that the object is, itself, rigid.
Here is how the tensor is constructed [Solc and Stockmayer, 1971]. Given
the location, ~ri (1 ≤ i ≤ N ), of each step of a walker in d dimensions that

has left N footprints, we construct a d dimensional tensor, T with entries
N
1 X
Tkl = (rjk − hrk i) (rjl − hrl i) (87)
N j=1

Here, rjk is the k th component of the position vector of the j th step, and
hrk i is the average of the k th component of the locations of the steps of the

37
walker:
N
1 X
hrk i = rjk (88)
N j=1

For example, a walker in two dimensions has radius of gyration tensor T
with the following form
PN !
1 2 1
PN

j=1 (x j − hxi) (x
j=1P j − hxi) (y j − hyi)
T= 1
PN N
N
1 N 2 (89)
N j=1 (x j − hxi) (y j − hyi) N j=1 (yj − hyi)

The eigenvectors and eigenvalues of this tensor quantify the linear dimensions
of the walker—its girth—in various directions. The eigenvectors point in the
direction in which this span is maximized, and the direction in which it
is minimized. The eigenvalues tell us how extended the walk is in those
extremal directions. In fact, the lines in Fig. 12 lie along the directions in
which those two eigenvectors point. The lengths of those lines are directly

proportional to the eigenvalues of the matrix T appropriate to the walk in
that figure.
For a discussion of the relationship between the eigenvectors and eigen-

values of T and the maximal and minimal spans of a walk see the section
beginning on page 47 in Supplement 1 at the end of this chapter.


Eigenvalues of the matrix T : the asphericity of a random
walk

The eigenvalues of the matrix T are the squares of the principal radii of
gyration, Ri , of the object in question. They are essentially the mean square
deviations of the steps of the walker from the “center of gravity” of the walk.
In Fig 12 the walk’s center of gravity lies at the point of intersection of the
two thick lines, each of which lie in the direction of the eigenvectors of the
↔ ↔
matrix T for that walk. This means that diagonalized, the matrix T takes
the form
 

λ 1 0 0
T =  0 λ2 0 
0 0 λ3

38
 
R12 0 0
≡  0 R22 0  (90)
0 0 R32

The relative magnitudes of the eigenvalues of the radius of gyration tensor T
then tell us to what extent the object in question has a shape that differs from
that of a sphere. Clearly if all Ri2 ’s in (90) are equal, then the linear span of
the object will be the same in all directions, and it is reasonable to attribute
a kind of spherical symmetry to it. However, if, for example, R12  R22 , R32 ,
which means that R1 is significantly larger than R2 and R3 , then the object
can be thought of as greatly elongated, and not at all spherical.
The eigenvalues of an object’s radius of gyration tensor are invariant
with respect to the overall orientation of the object. That is, a rotation of
the object will not change those eigenvalues. On the other hand, the tensor
itself does change as the object is rotated. If the brackets h· · ·ir stand for
averaging with respect to overall orientation, then the average hRi2 ir is just

the same as Ri2 . On the other hand, performing the same average over T
produces a matrix altered by the averaging process. In fact, it is pretty
straightforward to argue that h(x − hxi)(y − hyi)ir will average to zero, while
h(x − hxi)2 ir = h(y − hyi)2 ir = h(z − hzi)2 ir This means that
 

T̄ 0 0
hT i =  0 T̄ 0  (91)
0 0 T̄
The eigenvalues of this matrix are clearly all equal to T̄ . In averaging the
radius of gyration tensor, we are performing the kind of ensemble average
that destroys information regarding the non-spherical shape of the object in
question. This clearly means an informative characterization of the shape of
the random walk is not contained in the averaged radius of gyration tensor.
We can, nevertheless extract useful shape information by averaging quan-
tities that are directly derivable from the radius of gyration matrix. What
we need to do is use quantities that are invariant with respect to rotations
and reflections in space (the matrix is automatically invariant with respect
to translations). All of these quantities are directly related to the eigenval-
ues of the matrix. In the case of a three dimensional matrix there are three
independent invariants. One choice of those three is

Tr T = T11 + T22 + T33

39
= R12 + R22 + R32 (92)
↔2
Tr T = (R12 )2 + (R22 )2 + (R32 )2 (93)
↔3
Tr T = (R12 )3 + (R22 )3 + (R32 )3 (94)
Another well-known invariant of the tensor, its determinant is obtained as
follows

Det T = R12 R22 R32
1h 2 3
R1 + R22 + R32 + 2 (R12 )3 + (R22 )3 + (R32 )3

=
6
−3 R12 + R22 + R32 (R12 )2 + (R22 )2 + (R32 )2
 
 
1  ↔3 ↔3 ↔ ↔2
= Tr T + 2 Tr T −3 Tr T Tr T (95)
6
Now, it is possible to average the three invariants defined in Eqs. (92)–(94).
These averages retain important information regarding the devation from
spherical symmetry of the shape of the “average” random walk. Consider,
for example, the following combination of eigenvalues
2 2 2
R12 − R22 + R12 − R32 + R22 − R32
2
= 3 (R12 )2 + (R22 )2 + (R32 )2 − R12 + R22 + R32


↔2  ↔2
= 3 Tr T − Tr T (96)

Both sides of this equation can be averaged over all orientations of an object,
and, given the fact that they are invariants with respect to translation, rota-
tion and reflection, they will remain unchanged. In the case of the random
walk, this means that if we average the last line of (96) over an ensemble of
walkers we are left with a quantity that tells us something about the differ-
ences between the various principal radii of gyration. That is, we find out
how different the shape a random walk is, on the average, from that of a
sphere.
To construct a quantity that interpolates between zero when all principal
radii of gyration are equal and one when one of the Ri ’s is much greater than
the others we will divide by
3
X ↔
2h( Ri2 )2 i = 2hTr T i2 (97)
i=1

40
We can, in fact, generalize this quantity and define the mean asphericity,
Ad , of d-dimensional random walks as follows [Aronovitz and Nelson, 1986,
Theodorou and Suter, 1985, Rudnick and Gaspari, 1986]:
Pd 2
− Rj2 )2 i
i>j h(Ri
Ad = (98)
(d − 1)h( di=1 Ri2 )2 i
P

The numerator of (98) can be rewritten as follows:


2 2 2
h R12 − R22+ R12 − R32 + · · · Rd−12
− Rd2 i
↔2  ↔2
= d TrhT i − h Tr T i
2
i − hT11 T22 i + d2 (d − 1)hT12
2

= d(d − 1) hT11 i (99)

The last line of (99) follows from the equations for the trace of a tensor and
2 2 2
of its square. It also follows from the fact that hT11 i = hT22 i = · · · hTdd i, and
2
similar equalities for hTii Tjj i and hTij i. The denominator of the last line of
(98) can be reduced in the same way, leading to the following expression for
the asphericity:
2
d(d − 1) (hT11 i − hT11 T22 i) + d2 (d − 1)hT12
2
i
Ad = 2 2
d(d − 1)hT11 i + d(d − 1) hT11 T22 i
2 2
(hT11 i − hT11 T22 i) + dhT12 i
= 2
(100)
hT11 i + (d − 1)hT11 T22 i

The calculation of the asphericity reduces to the problem of determining the


average values of powers of the entries in the radius of gyration tensor. The
details, which are a bit involved, are described later on in this lecture. The
end-result of the calculation is the following general expression for the mean
asphericity of a d-dimensional random walk:
4 + 2d
Ad = (101)
4 + 5d
The three-dimensional walk has a mean asphericity of 10/19, or a little more
than a half, so in this sense the three-dimensional walk is, on the average,
somewhere between an isotropic object and a highly elongated one.
Of course the notion of the mean asphericity of a random walk does not
necessarily imply that there is a characteristic shape for three dimensional

41
1750
1500
1250
1000
750
500
250

0.2 0.4 0.6 0.8

Figure 13: The distribution of the individual asphericities of 20,000 three-


dimensional 100 step walks.

walks. Given the examples depicted in Fig. 11, it seems much more likely
that random walks come in a wide variety of shapes and that a quantity such
as the mean asphericity provides a very broad-brush characterization of that
property of random walks. Figure 13 illustrates this point. It is a histogram
of the distribution of the individual asphericities of 20,000 three-dimensional
walks, each comprised of 100 steps. Note that the distribution spans the
range from 0 to 1, and that no narrow region dominates.
It is also important to note that what is presented in (101) is not, strictly
speaking, the average of the individual asphericities of the walks, which is
given by
2 2
*P  +
2
0 i<j Ri − Rj
Ad = (102)
(d − 1) N 2
P
k=1 Rk

This quantity can also be found exactly in the case of the ordinary d-
dimensional walk. The analytical result for this quantity is [Diehl and Eisenregler, 1989]
 
0 d 4 d
Ad = 3 + − Md/2 (103)
4 d 2

where Z ∞
Mp = xp+1 sinh−p x dx (104)
0

In three dimensions, A0d = 0.394274 . . .. The average of the individual as-


phericities is somewhat smaller than the mean asphericity.

42
Shape of a self-avoiding random walk
Work on the shape of a self-avoiding walk has been performed [Aronovitz and Nelson, 1986].
The calculation is based on an expansion in the difference between the di-
mensionality in which the walk takes place and an “ upper critical dimen-
sionality,” equal to four. The quantity  = 4 − d is the expansion parameter.
To first order in 
2d + 4
Ad = + 0.008 (105)
5d + 4
The main conclusion to be gleaned from this result is that self-avoidance
plays a non-trivial, but far from decisive, role in the shape of a random walk.
5

Principal radii of gyration and rotational mo-


tion
Recall Newton’s second law:
d~p
F~ = (106)
dt
Here, F~ is the applied force and p~ is the linear momentum of the particle
or collection of particles to which that force is applied. When the system
undergoing the change in motion is a single particle, p~ = m~v , where ~v is the
particle’s velocity. Now, assume that we have a system of particles. Newton’s
second law, as it refers to each point particle, is, then
d~vl
F~l = ml (107)
dt
where l is the subscript that identifies the particle. If we sum up these
equations we end up with (106), where F is the net external force (internal
forces cancel because of Newton’s third law), and p~ is the total momentum:
X
p~ = ml~vl (108)
l

5
Notice that the first term on the right hand side of (105) has not been expanded about
d = 4. This is a (minor) violation of the spirit of the expansion in  = 4 − d, which does
not materially affect the conclusion stated above.

43
So far, so good. This is all pretty elementary. Now, let’s focus on ro-
tational motion. We derive Newton’s second law, as it applies to rotational
motion, by taking the cross product of the position vectors ~rl with the cor-
responding equation in the set (107). Defining the total torque ~τ as the sum
of the ~rl × F~l ’s, we end up with the equation
X d~vl
~τ = ml~rl ×
l
dt
d X
= ml~rl × ~vl
dt l
dL~
≡ (109)
dt
The last two lines of (109) constitute a definition of the angular momentum,
~ of a system of point particles. Note that the precise definition of angular
L
momentum depends on the origin with respect to which the position of each
particle is defined. It is often convenient to place the origin in at the center
of mass of the set of particles. If the internal force between each pair of
particles is along the line joining them, then the internally generated torques
cancel, and the total torque, ~τ , is entirely due to external forces.
Now, suppose that the motion of the system is entirely rotational, about
some point R. ~ Then,
 
~vl = ω ~
~ × ~rl − R (110)
Here, ω~ is the angular velocity of the system of particles. See Figure 14.
Now, we can choose R ~ as the center of our system of coordinates, so that
~ is replaced by ~rl . In this case, the angular momentum becomes
~rl − R
X
ml~rl × (~ω × ~rl ) (111)
l

We can rewrite the above relationship with the use of the standard identity
for the triple product:
X
~ =
L ml (~ω (~rl · ~rl ) − ~rl (~rl · ω
~ )) (112)
l

Suppose, now, we define the matrix T as follows:
X
Tij = ml rl,i rl,j (113)
l

44
ω

r-R
v
Figure 14: The angular velocity, ω
~ , and its relation to the velocity,~v , of a
particle.

~ and ω
Then, the relationship between L ~ is
↔ ↔
~ = Tr T ω
L ~ − T ·~ω (114)

Defining
↔ ↔ ↔ ↔
C =T − I Tr T (115)

where I is the identity operator we find that

~ = − C ·~ω
L (116)

Note that the trace of the operator C has a trace equal to twice the trace of

T . This is because the trace of the identity operator is three. The matrix

C is the moment-of-inertia matrix. The fact that the angular velocity and
the angular momentum are not parallel is just one of the complications of
rotational motion. ↔
Now, because the matrix T is real and symmetric, we know that it has
real (in fact, positive) eigenvalues. Those eigenvalues have a name. They are
known as the principal radii of gyration, Ri2 . If ω
~ points in the same direction
as the eigenvector of one of them, say R12 , then the angular momentum and
the angular velocity point in the same direction, and the relationship between
the two becomes
~ = R2 + R2 ω

L 2 3 ~ (117)
↔ ↔
This is because Tr T = R12 + R22 + R32 , while T ·~ω = R12 ω
~.

45

The eigenvalues of the tensor T can also be used as a measure of the
extent to which the system of particles possesses spherical symmetry, at
least in terms of rotational motion. If all the egenvalues are equal, then the
system has approximately the same weighted extent. In fact, in this case, L ~
is always parallel to ω
~.
Now imagine that all the masses, ml , are equal to one. Then, the matrix
measures the extent to which the particles are in a spherically symmetric
distribution. While equality of the principal radii of gyration is not equivalent
to spherical symmetry, it provides a very useful quantitative measure of that
property, and of departure from it.

We can construct our tensor, Q, as follows
↔ ↔ 1 ↔ ↔
Q=T − I Tr T (118)
3
The trace of this matrix is equal to zero, and, if all the principal radii
of gyration are the same, then the diagonalized form of this matrix has all
entries equal to zero.
Position vectors transform under rotations about the center of mass as
follows X
rk0 = Rkl rl (119)
l
where Rkl are the elements of the rotation matrix. This matrix has the
property that its transpose is also its inverse. That is
↔ ↔T ↔
R · R =I (120)
A matrix whose transpose is also its inverse is known as an orthogonal matrix.

Then, the matrix T with elements going as rl rk transforms as follows
X
Tk0 1 k2 = Rk1 l1 Tl1 l2 Rk2 l2
l1 ,l2
X
= Rk1 l1 Tl1 l2 RlT2 k2
l1 ,l2
X
= Rk1 l1 Tl1 l2 Rl−1
2 k2
(121)
l1 ,l2

or, in shorthand,
↔0 ↔ ↔ ↔−1
T =R · T · R (122)

46

The same relationship clearly holds for the tensor Q. The demonstration
that traces of powers of this tensor are invariant under rotations follows from

this equation for the way in which rotations give rise to changes in Q.


Just a little bit more on the meaning of the operator T .
Suppose we are interested in finding the direction in which an object is has
the greatest spatial extent. We start by assuming a vector ~n, which points

Figure 15: Looking for the direction in which an object has the greatest
spatial extent.

along the direction of interest. We will set the length of the vector at unity.
See figure 15. Then the extent of the object in the direction established by
~n is
X XX
(~rl · ~n)2 = ni rl,i rl,j nj
l l i,j

= ~n· T ·~n (123)

Remember that the position vectors are drawn with their tails at the center
of mass of the object. If we wish to extremize the above quantity with respect
to ~n, subject to the condition that its length is held constant, we take the
derivative with respect to each component of ~n of the expression below

~n· T ·~n − λ~n · ~n (124)

47
The quantity λ in (124) is a Lagrange multiplier. The extremum equation
easily reduces to

T ·~n = λ~n (125)
That is, in order to find the direction of greatest (or least) extent of the object,

we solve the eigenvalue equation of the operator T . The largest eigenvalue
is the greatest extent, as defined by (123), with ~n chosen to extremize the
quantity.6 The smallest extent, similarly defined, is given by the smallest

eigenvalue of T .

Calculations for the mean asphericity


The average hT11 i
First, we slightly recast the general expression for the mean asphericity
 2 
hT11 i hT11 T22 i
2 i
hT12
− 2 i
hT12
+d
Ad = 2 i
hT11
(126)
hT11 T22 i
+ d − 1

This means that to calculate the mean aspericity we need to find ratios of
averages, rather than the averages themselves. This simplifies our task a bit.
Although the averages that we have to perform in order to arrive at a
numerical result for the asphericity of the random walk involve squares of
entries in the radius of gyration tensor, or products of two entries, it is useful
to look at the average of a single element of that tensor, lying along the
diagonal. Eventually, we will perform this calculation in another way when
we develop an expansion in 1/d for the principal radii of gyration. However,
we will start out by showing how the calculation can be done with the use of
the generating functions that have proven so useful in the study of random
walk statistics. As a first step, we recast the expression for the entries Tkl : 7

N
1 X
Tkl = (rjk − hrk i) (rjl − hrl i)
N j=1
6
In other words, the extremizing choice for the vector ~n is the eigenvector of the operator

T with the largest eigenvalue.
7
Here, we make no distinction between N and N + 1

48
N N
1 XX
= (rik − rjk ) (ril − rjl ) (127)
2N 2 i=1 j=1

The first line in the above equation is a recapitulation of (87). The second
line can be established by inspection. Consider, now the average
N N
1 XX
hT11 i = h(xi − xj )2 i (128)
2N 2 i=1 j=1

We have reverted here to the notation appropriate to a three-dimensional


walk, and replaced r1 by x. The average in the sum on the right hand
side of (128) is directly proportional to a function that can be graphically
represented as shown in Fig. 16.

Figure 16: Graphical representation of h(xi − xj )2 i.

The dashed curve joins the ith and j th footprints on the walk. The solid
lines stand for the walk that begins at the leftmost end of the three-line
segment and ends at the righttmost point. There will, in general, be N1
steps from the far left point to the leftmost vertex at which the dashed curve
touches the line, N2 steps in the central segment of the walk and N3 steps in
the far right segment of the walk. Subject to the overall constraint that the
Ni ’s add up to the total number of steps in the walk, we sum over all values
of N1 , N2 and N3 . The evaluation of the sum represented by this diagram
is most conveniently carried out in the grand canonical ensemble, with the
use of the generating function. We seek the coefficient of z N −1 in the direct
product
Z Z Z
1
× 2 d r0 d r2 dd r3 G(z; ~r1 − ~r0 )G(z; ~r2 − ~r1 )G(z; ~r3 − ~r2 )(x2 − x1 )2
d d
2N 2
(129)
The “missing” integration in (129), over ~r1 , would yield a factor equal to the
volume of the portion of space in which the random walk occurs. The factor

49
of two multiplying the integral represent the two possible orderings of the
indices i and j (i > j and i < j). As the next step, we rewrite the generating
functions in terms of their spatial Fourier transforms,
Z
1 ~
G(z; ~r) = d
dd kg(z; ~k)e−k·~r (130)
(2π)

Making use of this representation, we find that the expression in (129) reduces
to
2
 
1 ~k) − ∂ ~k)

g(z; g(z; k) g(z; (131)
N2 ∂kx2 ~
k=0
The second derivative follows from the identity

~ ∂ 2 i~k·(~r1 −~r2 )
(x1 − x2 )2 eik·(~r1 −~r2 ) = − e (132)
∂kx2

and an integration by parts. Once this identity and integration by parts


has been implemented, the integrations over the ~ri ’s produces Dirac delta
functions in the ki ’s, and we are immediately led to (131)
Now, in the case of a random walk in d dimensions, we can write
1
g(z; ~k) = (133)
1 − zzc−1 + k 2 l2 /2d

where l represents the mean distance covered by the walker in each step.
This leaves us with the following result for the expression (131):

1 l2 1
(134)
N 2 d (1 − zzc−1 )4

We now extract the coefficient of z N −1 in the power series expansion of


(134).8 This is straightforward given what we know about the process.9 We
find for this coefficient
1 l2 −(N −1) (N + 2)(N + 1)N 1 l2 −(N −1) N 3
z ≈ z (135)
N2 d c 6 N2 d c 6
8
The relevant power is N − 1 because there are N − 1 steps in a walk that leaves N
footprints
9
See Supplement 3 in Chapter 2

50
This result is the desired value of hT11 i, multiplied by the total number of
random walks with N − 1 steps. To obtain the average, we divide this by the
total number of N − 1-step walks, which is equal to the coefficient of z N −1 in
Z
dd rG(z; ~r) = g(z; ~k = 0)
1
= (136)
1 − zzc−1
−(N −1)
The coefficient in question is zc . We thus find
l2
hT11 i = N (137)
6d

The average of the trace of the tensor T is, by symmetry, equal to dhT11 i.
Given (137) we have

hTr T i = N l2 /6 (138)
The quantity above is also known as the mean radius of gyration .

Calculation of the asphericity


2 2
The quantities that contribute to the asphericity are hT11 i, hT11 T22 i and hT12 i.
Again, reverting to standard cartesian notation we find
2
hT11 i
N N N N
1 XXXX
= 4
h(xi − xj )2 (xk − xl )2 i (139)
4N i=1 j=1 k=1 l=1

Now, the graphical representation of the numerator in the expression lead-


ing to the desired average is a bit more complicated, in that there are three
different forms, as shown in Fig. 17. The dotted curves are a stand-in for
the differences (xm − xn )2 , and the three types of graphical representations
correspond to the three “topologically distinct” possibilities for the sequence
of the indices i, j, k, l in (139). The calculation proceeds along the same line
as the one leading to a result for hT11 i. As we can see from (98), the only in-
2
formation we need to extract from our calculation is the ratios hT11 i/hT11 T22 i
2
and hT11 i/hT12 i. We will go over the determination of the average hT11 i in
detail. The other averages are determined in a similar way.

51
a b c

Figure 17: The three different forms of the graph involved in the calculation
2
of hT11 i.

2
Determination of hT11 i.
We proceed diagram-by-diagram.

Diagram a
Here, the calculation proceeds as it did in the evaluation of hT11 i. Taking
second derivatives and performing integrations by parts, we are left with the
following expression
 2   2 
1 ~ ∂ ~ ~ ∂ ~ ~

4
× 8 g(z; k) 2
g(z; k) g(z; k) 2
g(z; k) g(z; k) (140)
4N ∂kx ∂kx ~k=0

The factor of 8 in the above expression counts the number of ways of con-
structing diagram a, exchanging end-points of the two dashed curves, and
permuting the two curves amomng themselves. The quantity of interest is,
of course, the coefficient of z N −1 in (140). We will defer the power series
expansion in the fugacity z. Making use of (133) for the Fourier transformed
generating function, we end up with the result
 2
2 l2 1
(141)
N 4 d (1 − zzc−1 )7

Diagram b
The evaluation of this diagram is a bit more involved. Utilizing the generating
function in real space, we have the average of interest proportional to the
following expression:
Z Z Z Z Z
d r0 d r1 d r2 d r3 dd r4 G (z; ~r1 − ~r0 )
d d d d

×G (z; ~r2 − ~r1 ) G (z; ~r3 − ~r2 ) G (z; ~r4 − ~r3 ) G (z; ~r5 − ~r4 )
× (x4 − x1 )2 (x3 − x2 )2 (142)

52
The next step is to express the generating functions in terms of their Fourier
transforms. We end up with a product containg the factor
~ ~ ~ ~ ~
(x4 − x1 )2 (x3 − x2 )2 eik0 ·(~r1 −~r0 ) eik1 ·(~r2 −~r1 ) eik2 ·(~r3 −~r2 ) eik3 ·(~r4 −~r3 ) eik4 ·(~r5 −~r4 )
 2  2
∂ ∂ ∂ ∂ ~ ~ ~
= + + eik0 ·(~r1 −~r0 ) eik1 ·(~r2 −~r1 ) eik2 ·(~r3 −~r2 )
∂k2x ∂k1x ∂k2x ∂k3x
~ ~
×eik3 ·(~r4 −~r3 ) eik4 ·(~r5 −~r4 ) (143)

After a series of integrations of parts in the variables ~ki the derivatives above
act on the Fourier transforms of the generating functions. The integrations
over the ~rl ’s produce delta functions, and we are left with the following result
 2  2
∂ ∂ ∂ ∂
+ +
∂k2x ∂k1x ∂k2x ∂k3x
         
g z; ~k0 g z; ~k1 g z; ~k2 g z; ~k3 g z; ~k4

~ ~ ~ ~ ~ k0 =k1 =k2 =k3 =k4 =0
(144)

The polynomial expressions in the derivatives are now expanded, discarding


in the process all terms that evaluate to zero when the ~ki ’s are equal to zero.10
This yields
 2
∂2 ∂2 ∂2 ∂4


2 2
+ 2 2
+ 4
∂k1x ∂k2x ∂k3x ∂k2x ∂k2x
         
g z; ~k0 g z; ~k1 g z; ~k2 g z; ~k3 g z; ~k4

~ ~ ~ ~ ~ k0 =k1 =k2 =k3 =k4 =0
(145)

The remainder of the calculation is fairly straightforward. Inserting combi-


natorial factors noted above and taking the appropriate derivatives of the
generating function, we end up with the contribution
 2 2
16 l 1
(146)
N4 d (1 − zzc−1 )7
10
The general operating principle is that any term that contains an odd-order derivative
with respect to a kix will evaluate to zero.

53
Diagram c
In this case, the relevant identity is
~ ~ ~ ~ ~
(x3 − x1 )2 (x4 − x2 )2 eik0 ·(~r1 −~r0 ) eik1 ·(~r2 −~r1 ) eik2 ·(~r3 −~r2 ) eik3 ·(~r4 −~r3 ) eik4 ·(~r5 −~r4 )
 2  2
∂ ∂ ∂ ∂ ~ ~ ~
= + + eik0 ·(~r1 −~r0 ) eik1 ·(~r2 −~r1 ) eik2 ·(~r3 −~r2 )
∂k1x ∂k2x ∂k2x ∂k3x
~ ~
×eik3 ·(~r4 −~r3 ) eik4 ·(~r5 −~r4 ) (147)

The same set of steps as outined immediately above leads to the following
non-vanishing contributions to the diagram, combinatorial factors having
been left out,
 2
∂2 ∂2 ∂2 ∂2 ∂2 ∂4


2 2
+ 2 2
+ 2 2
+ 4
∂k1x ∂k2x ∂k3x ∂k2x ∂k1x ∂k3x ∂k2x
         
g z; ~k0 g z; ~k1 g z; ~k2 g z; ~k3 g z; ~k4

~ ~ ~ ~ ~ k0 =k1 =k2 =k3 =k4 =0
(148)

Taking the derivatives indicated, evaluating the ~ki = 0 limits and inserting
the required combinatorial factors, we end up
 2
18 l2 1
(149)
N 4 d (1 − zzc−1 )7

Summing all diagrams


Adding (141), (146) and (148), we end up with the following total contribu-
2
tion to the generating function yielding hT11 i
 2 2
36 l 1
(150)
N4 d (1 − zzc−1 )7
Actually, the generating function yields the numerator in a fraction. The
denominator is the total number of N − 1-step walks. However, as we are
interested in ratios of averages, the common denominator is not important
for our present purposes. For the same reason, we are not required to extract
the coefficient of z N −1 in the power series expansion of (150), as that produces
a common factor that cancels out when we evaluate ratios.

54
The ratios
2
The calculations of expressions contributing to hT11 T22 i and hT12 i proceed
along the lines laid out above. For each of these averages, there are three
contributions, corresponding to the three diagrams in Fig. 17. Carrying out
the required computations, we find
2
hT11 i 9
= (151)
hT11 T22 i 5
2
hT11 i 9
2
= (152)
hT12 i 2

Inserting these results into the right of (126), we obtain (101) for the mean
asphericity of a d-dimensional random walk. .ep

References
[Aronovitz and Nelson, 1986] Aronovitz, J. A. and Nelson, D. R. (1986).
Universal features of polymer shapes. Journal de Physique, 47(9):1445–56.

[Berg, 1993] Berg, H. C. (1993). Random walks in biology. Princeton Uni-


versity Press, Princeton, N.J., expanded edition.

[Berg and Purcell, 1977] Berg, H. C. and Purcell, E. M. (1977). Physics of


chemoreception. Biophysical Journal, 20(2):193–219.

[Bookstein, 1978] Bookstein, F. L. (1978). The measurement of biological


shape and shape change. Lecture notes in biomathematics ; 24. Springer-
Verlag, Berlin ; New York.

[Costa and Cesar, 2001] Costa, L. d. F. and Cesar, R. M. (2001). Shape


analysis and classification : theory and practice. Image processing series.
CRC Press, Boca Raton, FL.

[Diehl and Eisenregler, 1989] Diehl, H. W. and Eisenregler, E. (1989). Uni-


versal shape ratios for open and closed random walks: exact results for all
d. Journal of Physics A (Mathematical and General), 22:L87–L91.

[Montroll and Shlesinger, 1983] Montroll, D. and Shlesinger, M. F. (1983).


The wonderful world of random walks. In Falk, H., editor, CCNY physics

55
symposium in celebration of Melvin Lax’s sixtieth birthday, CCNY physics
symposium in celebration of Melvin Lax’s sixtieth birthday, page 364, New
York. City College of New York Physics Dept.

[Pólya, 1919] Pólya, G. (1919). Quelques problèmes de probabilité se rappor-


tant à la ‘promenade au hasard’ (some problems of probability associated
with the ‘random walk’. L’Enseignement Mathématique, 20:444–445.

[Rudnick and Gaspari, 1986] Rudnick, J. and Gaspari, G. (1986). The as-
pherity of random walks. Journal of Physics A (Mathematical and Gen-
eral), 19(4):L191–3.

[Rudnick and Gaspari, 2004] Rudnick, J. A. and Gaspari, G. D. (2004). El-


ements of the random walk : an introduction for advanced students and
researchers. Cambridge University Press, Cambridge ; New York.

[Solc and Stockmayer, 1971] Solc, K. and Stockmayer, W. H. (1971). Shape


of a random flight chain. Journal of Chemical Physics, 54(2756-2757).

[Theodorou and Suter, 1985] Theodorou, D. N. and Suter, U. (1985). Shape


of unperturbed linear polymers: polypropylene. Macromolecules, 18:1206–
1214.

[Weiss and Rubin, 1976] Weiss, G. H. and Rubin, R. J. (1976). The theory of
ordered spans of unrestricted random walks. Journal of Statistical Physics,
14(4):333–50.

56

You might also like