TIFR
TIFR
By
H. Rademacher
By
H. Rademacher
Notes by
K. Balagangadharan
and
V. Venugopal Rao
2 Lecture 11
3 Lecture 17
4 Lecture 23
5 Lecture 30
6 Lecture 39
7 Lecture 46
8 Lecture 55
II Analysis 59
9 Lecture 60
10 Lecture 67
11 Lecture 74
12 Lecture 82
13 Lecture 89
iii
CONTENTS iv
14 Lecture 95
15 Lecture 100
17 Lecture 118
18 Lecture 124
19 Lecture 129
20 Lecture 136
21 Lecture 143
22 Lecture 150
23 Lecture 155
24 Lecture 160
25 Lecture 165
26 Lecture 169
27 Lecture 174
28 Lecture 179
29 Lecture 183
30 Lecture 188
31 Lecture 194
32 Lecture 200
CONTENTS v
34 Lecture 214
35 Lecture 219
36 Lecture 225
37 Lecture 232
38 Lecture 237
39 Lecture 242
40 Lecture 246
41 Lecture 251
42 Lecture 256
43 Lecture 261
44 Lecture 264
45 Lecture 266
46 Lecture 272
Part I
1
Lecture 1
Introduction
In additive number theory we make reference to facts about addition in 1
contradistinction to multiplicative number theory, the foundations of which
were laid by Euclid at about 300 B.C. Whereas one of the principal concerns
of the latter theory is the deconposition of numbers into prime factors, addi-
tive number theory deals with the decomposition of numbers into summands.
It asks such questions as: in how many ways can a given natural number be
ecpressed as the sum of other natural numbers? Of course the decompostion
into primary summands is trivial; it is therefore of interest to restrict in some
way the nature of the summands (such as odd numbers or even numbers or per-
fect squares) or the number of summands allowed. These are questions typical
of those which will arise in this course. We shall have occasion to study the
properties of V-functions and their numerous applications to number theory,
in particular the theory of quadratic residues.
Formal Power Series
Additive number theory starts with Euler (1742). His tool was power series.
His starting point was the simple relation xm . xn = xm+n by which multiplica-
tion of powers of x is pictured in the addition of exponents. He therefore found
it expedient to use power series. Compare the situation in multiplicative num-
ber theory; to deal with the product n.m, one uses the equation n s m s = (nm) s ,
thus paving the way for utilising Dirichlet series.
While dealing with power series in modern mathematics one asks ques- 2
tions about the domain of convergence. Euler was intelligent enough not to ask
this question. In the context of additive number theory power series are purely
formal; thus the series 0! + 1! x + 2! x2 + · · · is a perfectly good series in our
2
1. Lecture 3
In this expression the middle term is not zero while all the other terms are
zero. Therefore dk+ j , 0 and so A.B , 0, which is a contradiction.
From this property follows the cancellation law:
If A , ◦ and A.B = A.C, then B = C. For, AB − AC = A(B − C). Since
A , 0, B − C = ◦ or B = C.
If the ring of coefficients has a unit element so has the ring of power series.
As an example of multiplication of formal power series, let, 4
A=1−x and B = 1 + x + x2 + · · ·
∞
X
A= a n xn , where a0 = 1, a1 = −1, and an = 0 for n ≥ 2,
n=0
X∞
B= b n xn , where bn = 1, n = 0, 1, 2, 3, . . .
n=0
X∞
C= cn xn , where cn = a0 bn + a1 bn−1 + · · · + an b0 ;
n=0
then
c0 = a0 b0 = 1, cn = bn − bn−1 = 1 − 1 = 0, n = 1, 2, 3, . . . ;
so (1 − x)(1 + x + x2 + · · · ) = 1.
We can very well give a meaning to infinite sums and products in certain
cases. Thus
A1 + A2 + · · · = B,
C1C2 · · · = D,
both equations understood in the sense module xN , so that only a finite number
of A′ s or (C − 1)′ s can contribute as far as xN .
Let us apply our methods to prove the identity:
Let
C = (1 + x)(1 + x2 )(1 + x4 ) . . .
(1 − x)C = (1 − x)(1 + x)(1 + x2 )(1 + x4 ) . . .
= (1 − x2 )(1 + x2 )(1 + x4 ) . . .
= (1 − x4 )(1 + x4 ) . . .
1. Lecture 5
1 + x + x2 + x3 + · · · = (1 + x)(1 + x2 )(1 + x4 ) · · ·
If we replace x by x3 and repeat the whole story, modulo x3N , the coeffi-
cients of these formal power series will still be equal:
Similarly
On the right side will occur factors of the form (1 + xN ). But N can be
written uniquely as xλ .m where m is odd. That means for each N, 1 + xN will
occur once and only once on the right side. We would like to rearrange the
factors to obtain (1 + x)(1 + x2 )(1 + x3 ) · · ·
This may be done for the following reason. For any N, that part of the 6
formal power series up to xN is a polymial derived from a finite number of
1. Lecture 6
factors. Rearranging the factors will not change the polynomial. But since this
is true for any N, the entire series will be unchanged by the rearrangement of
factors. We have thus proved the identity
Y ∞
Y
(1 + xk + x2k + x3k + · · · ) = (1 + xn ) (1)
k odd n=1
∞
P
This is an equality of two formal power series and could be written a n xn
n=0
∞
P
= bn xn . Let us find what an and bn are. On the left we have
n=0
∞
Q
Now (1 − x2n ) , 0, and by the law of cancellation, we may cancel it from
n=1
both sides of the equation obtaining,
∞
Y ∞
Y
(1 + xn ) (1 − x2n−1 ) = 1.
n=1 n=1
For the same reason as before, we may rearrange the order of the factors on
the left.
∞
Y ∞
Y
(1 + xn ) 1 + x2n−1 1 + x2n−1 + x2(2n−1) + · · ·
n=1 n=1
∞
Y
= 1 + x2n−1 + x2(2n−1) + · · · .
n=1
However,
∞
Y
1 + x2n−1 1 + x2n−1 + x2(2n−1) + · · · = 1,
n=1
Since this is true for any N, we may interchange the order of factors in the
entire product and get
∞
Y 1 1
k)
= Q
n odd
(1 − x (1 − xk )
k odd
1. Lecture 9
(1 + xz)(1 + x2 z)(1 + x3 z) · · ·
= 1 + zx + zx2 + (z + z2 )x3 + (z + z2 )x4 + (z + 2z2 )x5 + · · ·
= 1 + z(x + x2 + x3 + · · · ) + z2 (x3 + x4 + 2x5 + · · · ) + · · ·
= 1 + zA1 (x) + z2 A2 (x) + z3 A3 (x) + · · · (2)
= 1 + zxA(N) 2 2 (N)
1 (x) + z x A2 (x) + · · ·
So
(1 + zx)PN (zx, x) = 1 + zxN+1 PN (z, x),
1. Lecture 10
(1 + zx) 1 + zxA(N) N (N)
1 (x) + · · · + (zx) AN (x)
= 1 + zxN+1 1 + A(N) 2 (N)
1 (x) + z A2 (x) + · · ·
We may now compare powers of z on both sides since these are polynomi- 11
als. Taking zk , k ≤ N, we have
xk
A(N)
k (x) = k
1 − xN+1−k A(N)
k−1 (x),
1−x
xk
A(N)
k (x) ≡ A(N) (x) (mod xN ).
1 − xk k−1
From this recurrence relation we immediately have
x
A(N)
1 (x) ≡ (mod xN ),
1−x
x · x2
A(N)
2 (x) ≡ (1 − x)(1 − x2 ) (mod xN )
x3
≡ (mod xN )
(1 − x)(1 − x2 )
.....................
.....................
xk(k+1)/2
A(N)
k ≡ (mod xN )
(1 − x)(1 − x2 ) · · · (1 − xk )
Hence
∞
Y zx z 2 x3 z 3 x6
(1 + zxn ) ≡ 1 + + +
n=1
1 − x (1 − x)(1 − x ) (1 − x)(1 − x2 )(1 − x3 )
2
+··· (mod xN )
Lecture 2
∞
Y ∞
X
(1 + zxk ) = zk Ak (x), (1)
n=1 k=0
where
xk(k+1)/2
Ak (x) = (2)
(1 − x)(1 − x2 ) · · · (1 − xk )
We shall look upon the right side of (1) as a power series in x and not as a
power-series in z, as otherwise the infinite product on the left side would have
no sense in our formalism. Let us inerpret (1) arithmetically. If we want to
decompose m into k summands, we have evidently to look for zk and then for
xm , and the coefficient of zk xm on the right side of (1) gives us exactly what we
want. We have
∞ ∞ ∞
1 X
n1
X
2n2
X
= x x · · · xknk
(1 − x)(1 − x2 ) · · · (1 − xk ) n =0 n =0 n =0
1 2 k
∞
X
= p(k) m
m x ,
m=0
m = n1 + 2n2 + · · · + knk , n j ≥ 0,
and p(k)
m tells us how often m can be represented by k dfferent summands (with
possible repetitions). On the other hand the coefficient of xm on the left-side
11
2. Lecture 12
and
xk
Bk (x) = (4)
(1 − x)(1 − x2 ) · · · (1 − xk )
The left-side of (3) gives m with the representation
m = n1 + 2n2 + 3n3 + · · ·
with the provide that C◦ (x) = 1. The trick is the same. One studies temporatily 14
QV
a truncated affair (1 + zxV ), replaces z by zx2 and evaluates Ck (x) as in
V=1
Lecture 1. This would be perfectly legitimate. However one could proceed as
2. Lecture 13
Euler did - this is not quite our story. Multiplying both sides by 1 + zx2 , we
have
X∞ ∞
X
zk Ck (x) = (1 + zx2 ) zk x2k Ck (x).
k=0 k=0
Now compare powers of z on both sides - and this was what required some
2
extra thought. Ck (x) begins with x1+3+···+(2k−1) = xk ; in fact they begin with
later and later powers of x and so can be added up. We have
C0 = 1,
Ck (x) = x2k Ck (x) + x2k−1Ck−1 (x), k > 0,
x2k−1
or Ck (x) = Ck−1 (x)
1 − x2k
from this recurrence relation we obtain
x
C1 (x) = ,
1 − x2
3
x x4
C2 (x) = C 1 (x) = ,
1 − x4 (1 − x2 )(1 − x4 )
x5 x9
C3 (x) = C 2 (x) = ,
1 − x6 (1 − x2 )(1 − x4 )(1 − x6 )
...............
...............
2
xk
Ck (x) = ,
(1 − x2 )(1 − x4 ) · · · (1 − x2k )
carrying on the same rigmarole.
Now note that all this can be retranslated into something.
Let us give the number theoretic interpretation. The coefficient of zk xm 15
gives the number of times m can be expressed as the sum of k different odd
1
summands. On the other hand, the coefficients in the expansion of (1−x2 )···(1−x 2k )
xk
Dk (x) = ,
(1 − x2 ) · · · (1 − x2k )
leading to the
Theorem 5. m is the sum of k odd parts as often as m − k is the sum of even
parts not exceeding 2k, or m−k m−k
2 is the sum of even parts not exceeding k. ( 2
again is integral).
Some other methods
Temporarily we give up power series and make use of graphs to study par-
titions. A partition of N may be represented as an array of dots, the number of
dots in a row being equal to the magnitude of a summand. Let us arrange the
summands according to size.
For instance, let us consider a partition of 18 into 4 different parts 16
Draw a diagonal upward starting from the last but one dot in the column on
the extreme left. All the dots to the right of this diagonal constitute a partition of
12 into 4 parts. For each partition of 18 into 4 different parts there corresponds
thus a partition of 18 − 4.32 = 12 into parts. This process works in general for a 17
partition of n with k different parts. If we throw away the dots on and to the left
of the diagonal (which is drawn from the last but one point from the bottom in
order to exsure that the number of different parts constinues to be exactly k),
we are left with a partition of n − (1 + 2 + 3 + · · · + (k − 1)) = n − k(k−1)
2 . This
partition has exactly k parts because each row is longer by at least one dot than
the row below it, so an entire row is never discarded. Conversely, starting with
a partition of n − k(k−1)
2 into k parts, we can build up a unique partition of n into
k different parts. Add 1 to the next to the smallest part, 2 to the next longer, 3 to
the next and so on. This one-to-one correspondence proves that the number of
partitions of n into k different parts equals the number of partitions of n − k(k−1)
2
into k parts.
We can prove graphically that the number of partitons of n into k odd sum-
mands is the same as the number of partitions of n − k2 into even summands
not exceeding k. The last row of the
diagram contains at least one dot, the next higher at least three, the one
above at least five, and so on. Above and on the diagonal there are 1 + 3 + 5 +
· · · + (2k − 1) = k2 dots. When these are removed, an even number of dots is
left in each row, althogether adding up to n − k2 . This proves the result.
Theorem 1 can also be proved graphically, although the proof is not quite 18
as simple. The idea of the proof is examplified by considering the partitons of
35. We have
35 = 10 + 8 + 7 + 5 + 4 + 1
=5×2+1×8+7+5+1×4+1
= 5(2 + 1) + 7 × 1 + 1(8 + 4 + 1)
2. Lecture 16
· · · + }1
7 + 5 + 5 + 5 + 1| + {z
13 times
2 4 6 3
13 times
20
Each part is represented by a row of dots with the longest row at ehe top,
second longest next to the top, etc. The oddness of the parts allows uo to
place the rows symmetrically about a central vertical axis. Now connect the 19
dots in the following way. Connect the dots on this vertical axis with those on
the left half of the top row. Then connect the column to the right of this axis
to the other half of the top row. We continue in this way as indicated by the
diagram drawing right angles first on one side of the centre and then on the
other. We now interpret this diagram as a new partition of 35 each part being
represented by one of the lines indicated. In this way we obtain the partition
20+6+4+3+2 of 35 into different parts. It can be proved that this method works
in general. That is, to prove that given a partition of n into odd parts, this
method transforms it into a unique partition of n into distinct parts; conversely,
given a partation into distinct parts, the process can be reversed to find a unique
partition into odd parts. This establishes a one-to-one correspondence between
the two sorts of partitions. This proves our theorem.
Lecture 3
∞
P
The series zk Ak (x) that we had last time is itself rather interesting; the Ak (x) 20
k=0
have a queer shape:
xk(k−1)/2
Ak (x) =
(1 − x)(1 − x2 ) · · · (1 − xk )
Such series are called Euler series. Such expressions in which the factors in
the denominator are increasing in this way have been used for wide generalisa-
tions of hypergeometric series. Euler indeed solved the problem of computing
the coefficients numerically. The coefficient of zk xm is obtained by expanding
1
(1−x)···(1−xk )
as a power series. This is rather trivial if we are in the field of com-
plex numbers, since we can then have a decomposition into partial fractiions.
Euler did find a nice sort of recursion formula. There is therefore a good deal
to be said for a rather elementary treatment.
We shall, however, proceed to more important discussions the problem of
unrestricted partitions. Consider the infinite product (this is justifiable modulo
xN )
∞ ∞ ∞
Y 1 YX
= xmn
m=1
1 − xm m=1 n=0
∞
X ∞
X ∞
X
= xn1 x2n2 · x3n3 · · ·
n1 =0 n2 =0 n j =0
2
= 1 + x + 2x + · · ·
X∞
=1+ p n xn (1)
n=1
17
3. Lecture 18
7 2 0 1 5 12 22
−5 −2 1 4 7 10
3 3 3 3 3
Write down the coefficients by picking up 0, 1 and every other alternate
term, and continue the row towards the left by putting in the remaining coeffi-
cients. Now we find that the second differences have the constant value 3. But
an arithmetical progression of the second order can be expressed as a polyno-
mial of the second degree. The typical coefficient will therefore be given by an
expression of the form
3. Lecture 19
a(2λ + 1) + b a(2λ + 3) + b
1 3 6 10 15
2 3 4 5
1 1 1
1 4 9 16 25
3 5 7 9
2 2 2
The pentagons however do not fit together like this. We start with one pen-
tagon; notice that the vertices lie perspectively along rays through the origin.
So take two sides as basic and magnify them and add successive shelves. The
second differences now are always 3:
1 5 12 22
4 7 10
3 3
In general we can have r-gonal numbers where the last difference are all
r − 2.
We go back to equation (4): 24
∞
Y ∞
X
(1 − xm ) = (−)λ xλ(3λ−1)/2
m=1 λ=−∞
1 − x −Ex2 − (1 − x2 )x 3 2 3 4
MMM − (1 − x )(1 − xRR)x − (1 − x2 )(1 − x3 )(1 4 5
JJJ − x )x
EE RRRRR
EE MMM RRR JJJ
EE MMM RRR JJJ
EE M R JJJ
+x3 +(1 − x2 )x4 +(1 − x2 )(1 − x3 )x5
So 25
1 − x − x2 + x5 + x7 (1 − x2 ) + x9 (1 − x2 )(1 − x3 ) + · · ·
= 1 − x − x2 + x5 + 27 :+(1 − x3 )x 9 3
MMM +(1 − x )(1 −CCx )x
4 11
:: MMM CC
:: MMM CC
:: M CC
CC
9 3 11
−x −(1 − x )x
1 − x − x2 + x5 + x7 − x12 (1 − x3 )x15 − (1 − x3 )(1 − x4 )x18 − · · ·
When this is continued, we get some free terms at the beginning followed
by a typical remainder
xm + (1 − xk+1 )xm+k + (1 − xk+1 )(1 − xk+2 )xm+2k − xm+k − (1 − xk+1 )xm+2k (*)
= xm − xm+2k+1 − (1 − xk+1 )xm+3k+2 − (1 − xk+1 )(1 − xk+2 )
xm+4k+3 − · · · (**)
We have two free terms with opposite signs at the beginning. In (*) the
difference between exponents in successive terms is k, while in (**) this in-
creases to k + 1; this difference is in both cases the exponent of x in the first
factor. The remainder after the free terms begine with −, so that the sequence
of signs is + − − + + − − · · · This process perpetuates itself and the question 26
remains which powers actually appear. It is sufficient to mark down a scheme
for the exponents which completely characterises the expansion. The scheme
is illustrated by what follows.
3. Lecture 22
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
2 3 4 5 6 7 8 9 10 11 12 13 14 15
5 7 9 11 13 15 17 19 21 23 25 27 29 31
3 4 5 6 7 8 9 10 11 12 13 14
12 15 18 21 24 27 30 33 36 39 42 45
4 5 6 7 8 9 10 11 12 13
22 26 30 34 38 42 46 50 54 58
5 6 7 8 9 10 11 12
35 40 45 50 55 60 65 70
6 7 8 9 10 11
51 57 63 69 75 81
We write down the sequence of natural numbers in a row; the sequence less
the first two membere is repeated in a parallel row below leaving out the first
three placess at the beginning. Adding up we get
below which is placed the original sequence less the first three members, again
translating the whole to the right by two places. We again add up and repeat
the procedure. A typical stage in the procedure is exhibited below.
m m+k m + 2k m + 3k m + 4k m + 5k
k+1 k+2 k+3 k+4 k+5
m + 2k + 1 m + 3k + 2 m + 4k + 3 m + 5k + 4 m + 6k + 5
2+3=5 3 + 4 + 5 = 12
3+4=7 4 + 5 + 6= 15,
and in general:
λ(3λ − 1)
λ + (λ + 1) + · · · + (2λ − 1) = ,
2
λ(3λ + 1)
(λ + 1) + (λ + 2) + · · · + 2λ = ,
2
which are the only exponents appearing. We thus have
∞
Y ∞
X
(1 − xn ) = (−)λ xλ(3λ−1)/2
n=1 λ=−∞
Lecture 4
∞
Y ∞
X
(1 − xm ) = (−)λ xλ(3λ−1)/2 (1)
m=1 λ=−∞
We do not need these identities for their own sake, but for their applications
to number theory. We have the same sort of power-series on both sides; let us
compare the coefficients of xn . On the left side n appears as the sum of different
exponents. But in contradiction to previous situations, the coefficients appear
with both positive and negative signs, so that when we collect the terms there
may be cancellations. There are gaps in the powers that appear, but among
those which appear with non-zero coefficients, we have a pair of positive terms
followed by a pair of negative terms and vice versa. In most cases the coef-
ficients are zero; this is because of cancellations, so that roughly terms with
positive and negative signs are in equal number. A positive sign appears if we
multiply an even number of times. otherwise a negative sign. So an even num-
ber of different summands is as frequent generally as an odd number. Hence
the following theorem:
The number of decompositions of n into an even number of different parts
is the same as the number of decompositions into an odd number, with the
exception that there is a surplus of one sort or the other if n is a pentagonal
number of the form λ(3λ − 1)/2.
Before proceeding further let us examine a number of concrete instances.
Take 6 which is not a pentagonal number. The partitions are 6, 1 + 5, 2 + 4, 29
1 + 2 + 3, so that there are two decompositions into an even number of different
parts, and two into an odd number. Next take 7, which is a pentagonal number,
7 = λ(3λ+1)
2 with λ = 2. We can actually foresee that the excess will be in the
even partitions. The partitions are 7, 1 + 6, 2 + 5, 3 + 4, 1 + 2 + 4. Take 8 which
23
4. Lecture 24
with r dots, and the last slope, with s dots i.e., points on or below a segn-
ment starting from the dot on the extreme right of the last row and inclined at
45◦ (as in the diagram). We make out two cases.
1. s < r. Transfer the last slope to a position immediately above the first
row. The diagram is now as shown below:
The uppermost row is still shorter than the others. (because in our case 30
s < r). By this procedure the number of rows is changed by 1. This
establishes the one-one correspondence between partition of the ‘odd’
type and ‘even’ type.
4. Lecture 25
Take the uppermost row away and put it parallel to the last slope. This
diminishes the number of rows by 1, so that a partition is switched over
from the ‘even’ class to the ‘odd’ class or conversely.
Therefore there exists a one-one correspondence between the two classes.
So we have proved a theorem, which is a wrong one! because we have not
taken account of the exceptional case of pentagonal numbers. The fallacy lies
in having overlooked the fact that the last slope may extend right up to the
first row; the slope and the row may very well interfere. Let us take one such
instance. Let again s < r.
If we place the last slope above the first row this works because the number 31
of points in the first row is also diminished by one, in fact by the disputed point
(notice again that no two rows are equal for s < r − 1). So the interference is of
no account. With s ≥ r we may again have an interfering case. We again place
the top row
4. Lecture 26
behind the last slope, this time with a punishment. We have now shortened the
slope by 1. For s − 1 ≥ r the method is still good. So the only cases of earnest
interference are:
Here we have something which can no longer be overcome. These are the
cases of pentagonal numbers. In (ii) the total number of dots is equal to
s + (s + 1) + (s + 2) + · · · + (2s − 1)
s(3s − 1)
=
2
In (i) this number = (s + 1) + (s − 2) + · · · + 2s
s(3s + 1)
=
2
These decompositions do not have companions. In general every partition
into one parity of different summands has a companion of the other parity of
different summands; and in the case of pentagonal numbers there is just one in 32
excess in one of the classes.
We now come to the most important application of identity (1). Since
∞
1 X
Q∞ m
= p(n)xn ,
m=1 (1 − x ) n=0
This tells us the follwing story. All the coefficients on the right side of (2)
excepting the first must be zero. The typical exponent in the second factor on
the right side is λ(3λ − 1)/2 = ωλ , say. (The first few ω′λ s are 0, 1, 2, 5, 7, 12,
15, . . .). Now look for xn . Since the coefficient in the first factor is p(n) and
that in the second always ±1, we have, since xn (n , 0) does not appear on the
left side
or
X
p(n − ωλ )(−)λ = 0 (3)
0≤ωλ ≤n
p(0) = 1
p(1) = p(1 − 1) = p(0) = 1
p(2) = p(2 − 1) + p(2 − 2) = 2
p(3) = p(3 − 1) + p(3 − 2) = 3
p(4) = p(4 − 1) + p(4 − 2) = 5
p(5) = p(5 − 1) + p(5 − 2) − p(5 − 5) = 7
(Watch! a pentagonal number - and a negative sign comes into action!). These 33
formulae get longer and longer, but not excessively so. Let us estimate how
λ(3λ − 1)
long these will be. Since ωλ ≤ n we have to look for λ satisfying ≤
2
n, which gives
12λ(3λ − 1) ≤ 24n,
36λ2 − 12λ ≤ 24n,
(5λ − 1)2 = 24n + 1,
√
|6λ − 1| = 24n + 1,
4. Lecture 28
1 1√
|λ − | ≤ 24n + 1.
6 6
1√ 2√
Hence roughly there will be 24n = 6n summands on the left side
3 3
of (3). So their number increases with the square root of n- the expressions do
not get too long after all (for n = 100, we have 17 terms).
These formulae have been used for preparing tables of p(n) which have
been quite useful. For instance Ramanujan discovered some of the divisibility
properties of p(n) by using them. In the famous paper of Hardy and Ramanu- 34
jan (1917) there is a table of p(n) for n ≤ 200. These were computed by
Macmahon, by using the above formulae and the values were checked with
those given by the Hardy-Ramanujan formula. The asymptotic values were
found to be very close to what Macmahon computed. Gupta has extended the
table for p(n) up to 600.
Before making another application of Euler’s pentagonal theorem, we pro-
ceed a bit further into the theory of formal power series. We add now one more
formal procedure, that of formal differentiation. Let
A = a ◦ + a 1 x + a 2 x2 + · · ·
A′ = a1 + 2a2 x + 3a3 x2 + · · ·
(A + B)′ = A′ + B′ ,
(c A)′ = c A′
(A · B)′ = A′ B + A B′.
Let us look into this situation. Start with the simplest case, A = xm , B = xn .
Then
A′ = mxm−1 , B′ = nxn−1
4. Lecture 29
(AB)′ A′ B′
= + ,
AB A B
which is the rule for logarithmic differentiation. (It is identical with the proce-
dure in the calculus, as soon as we speak of functions). For A, B and C,
λ(3λ − 1)
with ωλ = . Taking the logarithmic derivative - and this can be done
2
piecemeal-
∞
P
∞ (−)λ ωλ xωλ−1
X −mxm−1 λ=−∞
= ∞
1 − xm P
m=1 (−)λ xωλ
λ=−∞
Multiplying both sides by x,
∞
P
∞ (−)λ ωλ xωλ
X −mxm λ=−∞
= ∞ (2)
1 − xm P
m=1 (−)λ xωλ
λ=−∞
The left side here is an interesting object called a Lambert series, with a
structure not quite well defined; but it plays some role in number theory. Let
us transform the Lambert series into a power series; it becomes
∞
X ∞
X X∞X
− m xkm = − mxkm ,
m=1 k=1 k1 m=1
30
5. Lecture 31
and these are all permissible power series, because though there are infinitely
many of them, the inner ones begin with later and later terms.
Rearranging, this gives 37
∞
XX ∞
X X
− mxn = − xn m
n=km m=1 n=1 m/n
X∞
=− σ(n)xn ,
n=1
P
where σ(n) denotes the sum of the divisors of n, σ(n) = d.
d|n
(Let us study σ(n) for a moment.
for a prime p. And σ(n) = n + 1 implies that n is prime. σ(n) is not too big;
there can be at most n divisers of n and so roughly σ(n) = O(n2 ). In fact it is
known that σ(n) = O(n1+epsilon ), ∈> 0, that is, a little larger than the first power.
We shall however not be studying σ(n) in detail).
Equation (2) can now be rewritten as
∞
X ∞
X ∞
X
σ(n)xn (−)λ xωλ = (−)λ−1 ωλ xωλ
n=1 λ=−∞ λ=−∞
Let us look for the coefficient of xm on both sides. Remembering that the
first few ω′λ s are 0, 1, 2, 5, 7, 12, 15 · · · , the coefficient of xm on the left side is
On the right side the coefficient is 0 most frequently, because the pentago-
nal numbers are rather rare, and equal to (−)λ−1 ωλ exceptionally, when m = ωλ .
0
usually,
σ(m) − σ(m − 1) − σ(m − 2) + + − − · · · =
(−)λ−1 ωλ for m = ωλ .
This is an additive recursion formula for σ(n). We can make it even more
striking. The inhomogeneous piece on the right side is a little annoying. σ(m −
m) can occur on the right side only for m = ωλ ; σ(0) does not make sense;
however, for our purpose let us define
σ(m − m) = m.
Comparing coefficients of xn ,
X
np(n) = σ(m)p(k),
m+k=n
5. Lecture 33
or more explicitly,
∞
X
np(n) = σ(m)p(n − m) (6)
m=1
This is a bilinear relation between σ(n) and p(n). This can be proved di-
rectly also in the following way. Let us consider all the partitions of n; there
are p(n) such:
n = h1 + h2 + · · ·
n = k1 + k2 + · · ·
n = ℓ1 + ℓ2 + · · ·
.........
Adding up, the left side gives np(n). Let us now evaluate the sum of the
right sides. Consider a particular summand h and let us look for those partitions
in which h figures. These are p(n −h) partitions in which h occurs at least once,
p(n − 2h) in which h occurs at least twices; in general, p(n − rh) in which h
occurs at least r times. Hence the number of those partitions which contain h
exactly r times is p(n − nh) − p(n − n + 1h). Thus the number of times h occurs 40
in all partitions put together is
X n o
n p(n − nh) − (n − n + 1h)
nh≤n
on applying partial summation. Now summing over all summands h, the right
side becomces
X X XmX
h p(n − nh) = p(n − m),
h nh≤n n/m
n m≤n
This last formula enables us to find out the sum of the divisors provided
that we know the partitions. This is not just a curiosity; it provides a useful
check on tables of pertitions computed by other means.
We go back to power series leading up to some of Ramanujan’s theorems.
Jacobi introduced the products
∞
Y
(1 − x2n )(1 + zx2n−1 )(1 + z−1 x2n−1 ).
n=1
This is a power series in x; though these are infinitely many factors they
start with progressively higher powers. The coefficients this time are not poly-
nomials in z but from the field R(z), the field of rational functions of z, which
is a perfectly good field. Let us multiply out and we shall have a very nice
surprise. The successive coeffieicnts are:
1
x : z + z−1 (note that this is unchanged when z → z−1 )
x2 : (1 + 1) = 0
x3 : (z + z−1 − z − z−1 ) = 0
x4 : (−1 − 1 + z2 + 1 + 1 + z−2 ) = z2 + z−2 (again unchanged when z → z−1 )
............
5. Lecture 35
We observe that non-zero coeffieicnts are associated only with square ex- 42
ponents. We may threfore provisionally write
∞
Y ∞
X 2
−1 2n−1
2n
(1 − x )(1 + zx 2n−1
)(1 + z x )=1+ (zk + z−k )xk
n=1 k=1
∞
X 2
= z k xk (7)
k=−∞
ΦN (x, z) = C◦ (x) + (z + z−1 )C1 (x) + (z2 + z−2 )C2 (x) + · · · + (zN + z−N )C N (x).
N
Y
ΦN (x, zx2 ) = (1 − x2n )(1 + zx2n+1 )(1 + z−1 x2n−3 ).
n=1
Compare ΦN (x, zx2 ) and ΦN (x, z); these are related by the equation
ΦN (x, zx2 )(1 + zx)(1 + z−1 x2N−1 ) = ΦN (x, z)(1 + zx2n+1 )(1 + z−1 x−1 )
5. Lecture 36
The negative power in the last factor on the right is particularly disgusting;
to get rid of it we multiply both sides by xz, leading to
N
2
−2N+1 1 − x4N Y
C N−1 (x) = xN (1 − x2n )
1 − x2 n=1
N
Y
2
= x(N−1) (1 − x2n ) · (1 − x4N );
n=2
N
Y
2
C N−2 (x) = x(N−2) (1 − x2n ) · (1 − x4N )(1 − x4N−2 )
n=3
............
In general,
N
Y j−1
Y
2
C N− j (x) = x(N− j) (1 − x2n ) (1 − x4N−2m )
n= j+1 m=0
or, with j = N − n,
N
Y N−n−1
Y
2
Cn (x) = xn (1 − x2n ) (1 − x4N−2m ) (8)
n=N−n+1 m=0
5. Lecture 37
Equation (8) leads to some congruence relations. The lowest terms of Cn (x)
have exponent
n2 + 2(N − n + 1) = 2N + (n2 − 2n + 1) + 1 ≥ 2N + 1
Hence
2
Cn (x) ≡ xk (mod x2N+1 ) (9)
From the original formula,
N
Y
ΦN (x, z) = (1 − x2n )(1 + zx2n+1 )(1 + z−1 x2n−1 )
n=1
≡ 1 + (z + z−1 )x + (z2 + z−2 )x4 + · · · (mod x2N+1 )
X∞
2
≡ zk xk (mod x2N+1 ),
k=−∞
since the infinite series does not matter, the higher powers being absorbed in 45
the congruence. Hence
∞
Y
ΦN (x, z) ≡ (1 − x2n )(1 + zx2n−1 )(1 + z−1 x2n−1 ) (mod x2N+1 )
n=1
Thus both expansions agree as far as we wish, and this is what we mean
by equality of formal power series. Hence we can replace the congruence by
equality, and Jacobi’s identity (7) is proved.
As an application of this identity, we shall now give a new proof of the
pentagonal numbers theorem. We replace x by y3 , as we could consistently in
the whole story; only read modulo y6N+3 . Then we have
∞
Y ∞
X 2
(1 − y6n )(1 + zy6n−3 )(1 + z−1 y6n−3 ) = zk y3k
n=1 k=−∞
by y3 nobody could forbid that. Since z occurs in negative powers, the powers
of y might be lowered too by as much as N. We obtain polynomials in y alone
on both sides, but true modulo y5N+3 , because we may have lowered powers of 46
y. With this proviso it is justified to replace z by −y; so that ultimately we have
∞
Y ∞
X 2
(1 − y6 )(1 − y6n−2 )(1 − y6n−4 ) = (−)k y3k +k
(mod y5N+3 )
n=1 k=−∞
We can carry over the old proof step by step. Since we now have only even
powers of y, this leads to
∞
Y ∞
X
(1 − y2m ) = (−)k yk(3k+1)
m=1 k=−∞
∞
Y ∞
X 2
(1 − x2n )(1 + zx2n−1 )(1 + z−1 x2n−1 ) = z k xk (1)
n=1 k=−∞
and this was a polynomial of the proper size and everything went through.
When we replace z by −zx and multiply out, the negative powers might accu-
mulate and we might be destroying xN possibly; nevertheless the congruence
relations would be true this time modulo xN+1 instead of x2N+1 as it was previ-
ously; but this is all we went. So the old proof can be reproduced step by step
and every thing matches modulo xN+1 . (Let us add a side remark. In the proof
of (1) we had to replace z by zx2 - and this was the essential step in the proof.
We cannot do the same here as this would lead to congruences mod x only.
Before we had the congruences we had identities and there we could carry out
any substitution. Then we adopted a new point of view and introduced congru- 48
ences; and that step bars later the substitution z → zx2 .
So let us make the substitution z → −zx without further compuncton. This
39
6. Lecture 40
gives us
∞
Y ∞
X 2
(1 − x2n )(1 − zx2n )(1 − z−1 x2n−2 ) = (−)k zk xk +k
n=1 k=−∞
In the field R(z) we can replace z by 1. We can do what we like in the field
and that is the essence of the power series method. So putting z = 1,
∞
Y ∞
X
(1 − x2n )3 = (−)k xk(k+1) (2k + 1).
n=1 k=0
∞
Y ∞
X
(1 − yn )3 = (−)k (2k + 1)yk(k+1)/2 (2)
n=1 k=0
6. Lecture 41
The second part on the right side here is of interest, because it is the gener- 50
ating function of the partition. We had earlier the formula
∞
Y ∞
X
(1 + zx2n−1 ) = zmCm (n),
n=1 m=0
2
(4)
xm
Cm (x) =
(1 − x2 ) · · · (1 − x2m )
and these are permissible power series, beginning with later and later powers
of x, and so the right side of (4) makes sense, as a formal power series in x.
Substituting (4) in (3), we have
∞ ∞ ∞
X X X 2 1
znCr (x) z−sC s (x) = z k xk ∞ (5)
Q
r=0 s=0 k=−∞ (1 − x2n )
n=1
We can compare zO on both sides for, for very high xN the left side will
contain only finitely many terms and all otheres will disappear below the hori-
zon; we can also add as many terms as we wish. So equating coefficients of zO ,
we have
∞
X 1
Cr (x)Cr (x) = ∞ ,
Q
r=0 (1 − x2n )
n=1
∞ 2
X x2r 1
or = ∞
(1 − x ) · · · (1 − x2n )2
2 2 Q
r=0 (1 − x2n )
n=1
6. Lecture 42
y y4 y9
1+ 2
+ 2 2 2
+ +···
(1 − y) (1 − y) (1 − y ) (1 − y) (1 − y2 )2 (1 − y3 )2
2
1
= ∞ (6)
Q
(1 − yn )
n=1
This formula is found in the famous paper of Hardy and Ramanujan (1917) 51
and ascribed by them to Euler. It is very useful for rough appraisal of asymp-
totic formulas. Hardy and Ramanujan make the cryptic remark that it is “a
formula which lends itself to wide generalisations”. This remark was at first
not very obvious to me; but it can now be interpreted in the following way. Let
us look for zk in (5). Then
2
X xk
Cr (x)C s (x) = ∞
Q
r,s
r−s=k
(1 − x2n )
n=1
or, replacing r by s + k, and writting C s for C s (x), the left side becomes
∞ 2 2
X xk x1+(k+1)
C sC s+k =1· + +
s=0
(1 − x2 ) · · · (1 − x2k ) (1 − x2 )2 (1 − x4 ) · · · (1 − x2k+2 )
2
x4+(k+2)
+ +···
(1 − x2 )2 (1 − x4 )2 (1 − x6 ) · · · (1 − x2k+4 )
2
Let us divide by xk . The general exponent on the right side is ℓ2 + (k + ℓ)2 , 52
so on division it becomes 2ℓ2 + 2kℓ. Every exponent is even, which is a very
nice situation. Replace x2 by y, and we get the ‘wide generalisation’ of which
Hardy and Ramanujan spoke:
1 yk+1
+
(1 − y)(1 − y2 ) · · · (1 − yk ) (1 − y)2 (1 − y2 ) · · · (1 − yk+1 )
y2(k+2)
+ +···
(1 − y)2 (1 − y2 )2 (1 − y3 ) · · · (1 − yk+2 )
yl(k+l) 1
2 l 2 l+1 k+l
+··· = ∞ (7)
(1 − u) · · · (1 − y ) (1 − y ) · · · (1 − y ) Q
(1 − yn )
n=1
6. Lecture 43
We are now slowly coming to the close of our preoccupation with power
series; we shall give one more application due to Ramanujan (1917). In their
paper Hardy and Ramanujan gave a surprising asymptotic formula for p(n).
It contained an error term which was something unheard of before, O(n−1/4 ),
error term decreasing as n increases. Since p(n) is an integer it is enough to take
a few terms to get a suitable value. The values calculated on the basis of the 53
asymptotic formula were checked up with those given by Macmahon’s tables
and were found to be astonishingly close. Ramanujan looked at the tables and
with his peculiar insight discovered something which nobody else could have
noticed. He found that the numbers p(4), p(9), p(14), in general p(5k + 4) are
all divisible by 5; p(5), p(12), · · · p(7k + 5) are all divisible by 7; p(11k + 6) by
11. So he thought thiswas a general property. A divisibility property of p(n) is
itself surprising, because p(n) is a function defined with reference to addition.
The first and second of these results are simpler than the third. Ramanujan in
fact suggested more. If we chose a special progression modulo 5λ , then all the
terms are divisible by 5λ . There are also special progressions modulo 72λ−1 ; so
for 11. Ramanujan made the general conjecture that if δ = 5a 7b 11c and 24n ≡ 1
(mod δ), then p(n) ≡ 0 (mod δ). In this form the conjecture is wrong. These
things are deeply connected with the theory of modular forms; the cases 5 and
7 relate to modular forms with subgroups of genus 1, the case 11 with genus 2.
Let us take the case of 5. Take p(5k + 4). Consider Σp(n)xn ; it is nicer
to multiply by x and look for x5k . We have to show that the coefficients of
x5k in xΣp(n)xn are congruent to zerp modulo 5. We wish to juggle around
with series a bit. Take Σan xn ; we want to study x5k . Multiply by the series
1 + b1 x5 + b2 x10 + · · · where the b′ s are integers. We get a new power series
X X
an xn · (1 + b1 x5 + b2 x10 + · · · ) = cn xn ,
which is just as good. It is enough if we prove that for this series every fifth 54
coefficient ≡ 0 (mod 5).
For,
P
X cn xn
a n xn =
1 + b1 x + b2 x10 + · · ·
5
6. Lecture 44
X
= cn xn , (1 + d1 x5 + d2 x10 + · · · ), say.
Then if every fifth coefficient of Σcn xn is divisible by 5, multiplication by
Σdn x5n will not disturb this. For a prime p look at
! ! ! !
p p 2 p 3 p p
(1 + x) p = 1 + x+ x + x +···+ x .
1 2 3 p
All except the first and
lastp!coefficients on the right side are divisible by p,
for in a typical term qp = (p−q)!q! , the p inm the numerator can be cancelled
only by a p in the denominator. So
(1 + x) p ≡ 1 + x p (mod p).
This means that the difference of the two sides contains only coefficients
divisible by p. This
(1 − x)5 ≡ 1 + x5 (mod 5)
We now go to Ramanujan’s proof that p(5k + 4) ≡ 0 (mod 5) We have 55
X x
x p(n)xn = Q
(1 − xn )
It is irrelevant here if we multiply both sides by a series containing only
x5 , x10 , x15 , · · · . This will not ruin our plans as we have declared in advance.
So
∞ ∞
X Y x Y
x p(n)xn (1 − x5m ) = Q n)
(1 − x5m )
m=1
(1 − x m=1
∞
x Y
≡Q n
(1 − xm )5 modulo 5
(1 − x ) m=1
Y Y
(1 − x5m ) − (1 − xm )5 has only coefficients divisible by 5
∞
Y
≡x (1 − xm )4 modulo 5
m=1
∞
Y ∞
Y
=x (1 − xm ) (1 − xm )3 .
m=1 m=1
For both products on the right side we have available wonderful expres-
sions. By (2) and (2a),
Y Y ∞
X ∞
X
x (1 − xm ) (1 − xm )3 = x (−)λ(3λ−1)/2 (−)k (2k + 1)xk(k+1)/2
λ=−∞ k=0
6. Lecture 45
The exponent = 1 + λ(3λ − 1)/2 + k(k + 1)/2, and we want this to be of the
form 5m. Each such combination contributes to x5m . We want 56
λ(3λ − 1) k(k + 1)
1+ + ≡0 (mod 5)
2 2
Multiply by 8; that will not disturb it. So we want
A2 ≡ 0, 1, 4(5),
2B2 ≡ 0, 2, 3(5);
and so A2 + 2B2 ≡ 0(5) means only the combination A2 ≡ 0(5) and 2B2 ≡ 0(5);
each square must therefore separately be divisible by 5, or
2k + 1 ≡ 0(5)
x x4 x9
1+ + +
1 − x (1 − x)(1 − x ) (1 − x)(1 − x2 )(1 − x3 )
2
1
+··· = Q ; (1)
n>0 (1 − xn )
n≡±1 (mod 5)
x2 x2·3 x3·4
1+ + +
1 − x (1 − x)(1 − x2 ) (1 − x)(1 − x2 )(1 − x3 )
1
+··· = Q (2)
n>0 (1 − xn )
n≡±2 (mod 5)
The right hand sides of (1) and (2), written down explicitly, are respectively
1
(1 − x)(1 − x4 )(1 − x6 )(1 − x9 ) . . .
1
(1 − x )(1 − x )(1 − x7 )(1 − x8 ) . . .
2 3
46
7. Lecture 47
k2 = 1 + 3 + 5 + · · · + (2k − 1),
each partition of n into summands not exveeding k can be enlarged into a par-
tition of n + k2 into summands which differ by at least two, for we can adjoin
k2 dots on the left side, putting one in the lowest row, three in the next, five
7. Lecture 48
in the one above and so on finally 2k − 1 in the top most row. Conversely any
partition of n into
P xk(3k−1)/2
no surprise. In fact would give us the number of
(1 − x)(1 − x2 ) · · · (1 − xk )
partitions into parts differing by at least 3. And for 4 the story is similar.
The unexpected element in all these cases is the association of partitions
of a definite type with divisibility properties. The left-side in the identities
is trivial. The deeper part is the right side. It can be shown that there can
be no corresponding identities for moduli higher than 5. All these appear as
wide generalisations of the old Euler theorem in which the minimal difference
between the summands is, of course, 1. Euler’s theorem is therefore the nucleus
of all such results.
We give here a proof of the Roger-Ramanujan identities which is in line
with the treatment we have been following, the motiod of formal power series.
It is a transcription of Roger’s proof in Hardy’s ‘Ramanujan’, pp.95-98. We
use the so-called Gaussian polynomials.
Let us introduce the Gaussian polynomials in a much neater notation than
usual. Consider for first the binomial coefficients:
!
n n(n − 1)(n − 2) · · · (n − k + 1)
=
m 1 · 2 · 3 · ···k
(Observe that both in the numerator and in the denominator there are k fac- 62
tors, which are consecutive integers, and that the factors of equal rank in both
numerator and denominator always add up to n + 1). The nk are all integers, as
is obvious from the recursion formula
! ! !
n+1 n n
= +
k k k−1
n n
n = 1, of course, and by definition, 0 = 1 We also define nk = 0 for k > n
n
or for k < 0. Observe also the eymmetry: nk = n−k
The Gaussian polynomials are something of a similar nature. We define the
Gaussian polynomial
" # " #
n n
=
k k x
" #
n (1 − x )(1 − xn−1 ) · (1 − xn−k+1 )
n
by =
k (1 − x)(1 − x2 ) · · · (1 − xk )
where 2
xk
Ck (x) =
(1 − x2 ) · · · (1 − x2k )
Now write 64
x2 = X, 1 − X = X1 − X 2 = X2 , . . . , 1 − X k = Xk ;
(1 − X)(1 − X 2 ) · · · (1 − X k ) = X, X2 . . . Xk = Xk !
Notice that the powers of z occur with the same parity as n. Now (2) can 65
be re-written as
∞
P 2
∞ z l xl
X Bn (z, x) l=−∞
= ∞
Xn ! Q
n=0 (1 − x2n )
n=1
7. Lecture 52
Both sides are formal power series in x of the appropriate sort. The Bn (z, x)
are linear combinations of power series in x with powers of z for coeffieicnts.
We can now compare powers of z. We first take only even exponents z2m ; we
then have infinitely many equations of formal power series. We multiply the
equation arising from z2m by (−)m xm(m−1) and add all these equations together;
(amd that is the trick, due to Rogers) we can do this because of linearity. Then
∞
P 2
∞ (−)m xm(m−1) x(2m)
X β2l (x) m=0
= ∞ , (3)
X2l ! Q
l=0 (1 − x2n )
n=1
2l "
X #
2l 2 2
where β2l (x) = x(2l−r) +r (−)l−r x(l−r)(l−r−1)
r X
r=0
Writting l − r = s,
X l " #
2l 2 2
β2l (x) = x2l +2s (−) s x s(s−1)
l−s
s=−l
Xl " #
2l2 2l 2
=x (−) s x3s − s
l+s
s=−l
(because of the symmetry between l − s and l + s). Separating out the term
corresponding to s = 0 and folding together the terms corresponding to s and 66
−s,
" # l " #
2l2 2l
X 2l
s s(3s−1) 2s
β2l (x) = x + (−) x (1 + x )
l
l+s
s=1
l " # l " #
X s 2l X 2l
2l2 s(3s−1) s s(3s+1)
=x (−) x + (−) x
l+s l+s
s=1 s=0
l " # l " #
X
2l X 2l
2l2 s+1 (s+1)(3s+2) s s(3s+1)
=x (−) x + (−) x (4)
l+s+1 l+s
s=0 s=0
Then
l # " !
2l2
X 2ls s(3s+1) 1 − X l−s 4s+2
β2l (x) = x (−) x 1− x
l+s 1 − X l+s+1
s=0
l " #
2
X 2l 1 − X 2s+1
= x2l (−) s x s(3s+1)
l+s 1 − X l+s+1
s=0
7. Lecture 53
2 l " #
x2l X
s 2l + 1
= (−) x s(3s+1) (1 − x4s+2 ) (5)
1 − X 2l+1 s=0 l+s+1
Let us now compute β2l+1 (x). For this we compare the coefficients of z2m+1 ,
multiply the resulting equations by (−)m xm(m−1) and add up. Then
∞
P 2
∞ (−)m xm(m−1) x(2m+1)
X β2l+1 (x) m=0
= ∞ , (6)
X2l+1 ! Q
l=0 (1 − x2n )
n=1
where 67
2l+1
X" #
2l + 1 (2l+1−r)2
β2l+1 (x) = x + r2 (−)l−r x(l−r)(l−r−1)
r
r=0
Writting l − r = s, this gives
l " #
X 2l + 1 (l+1−s)2 +(l−s)2 s s(s−1)
β2l+1 (x) = x (−) x
l−s
s=−l−1
l " #
X 2l + 1 2 2 2
= (−) s x3s +s+l +(l+1)
l−s
s=−l−1
l " #
X
2l + 1
2l2 +2l+1 s
=x (−) x s(3s+1)
l+s+1
s=0
l " #
X 2l + 1
s+1 (−s−1)(−3s−2)
+ (−) x
l+s+1
s=0
l " #
2l2 +2l+1
X
s 2l + 1
=x (−) x s(3s+1) (1 − x4s+2 ) (7)
l+ s+1
s=0
This expression for β2l+1 (x) is very neat; it is almost the same as β2l (x) but
for trivial factors. Let us go back to β2l+1 (x) in its best shape.
(" #
2l2 +2l+1 2l + 1
β2l+1 (x) = x
l
X l " # " # !
2l + 1 s s(3s+1) 2l + 1
s s(3s−1)
+ (−) x + (−) x
l+s+1 l+s
s=1
" # X l " # !
2l2 +2l+1 2l + 1
2l + 1 s s(3s−1) 1 − X l−s+1 2s
=x + (−) x 1+ x
l
l+s 1−X l+s+1
s=1
7. Lecture 54
Since 68
This fits with β2l+2 . Now we can read off the recursion formulae. The
consequences are too very nice facts. The whole thing hinges upon the courage
to tackle these sums. We did not do these things ad hoc.
Let us compare β2l+1 with β2l
Last time we obtained the two fundamental formulae for β2l , β2l+1 , from which 69
we deduced the recurrence relations:
β2m+1 = x2m+1 (1 − x2(2m+1) )β2m ,
1 − x2(2m+2) (1)
β2m+2 = x2m+1 β2m+1
1 − x2(m+1)
β2m came from B2m by a substitution which was not yet plausible. Let us cal-
culate the first few β′ s explicitly. By definition
B0 = 1 = β0
β1 = x(1 − x2 ) β0 = x(1 − x2 )
1 − x4
β2 = x β1 = x2 (1 − x4 )
1 − x2
β3 = x (1 − x6 )
3
β2 = x5 (1 − x4 )(1 − x6 )
1 − x8
β4 = x3 β3 = x8 (1 − x6 )(1 − x8 );
1 − x4
and in general,
2
β2m = x2m (1 − x2m+2 )(1 − x2m+4 ) · · · (1 − x4m )
2 X2m !
= Xm (with X = x2 ); (2)
Xm !
and similarly,
2
β2m+1 = x2m +2m+1
(1 − x2m+2 )(1 − x2m+4 ) · · · (1 − x4m+2 )
2 X2m+1 !
= X m +m x · (3)
Xm !
55
8. Lecture 56
so that by (2)
∞
P
(−)l X l(5l−l)/2 ∞ 2
l=0
X Xm
∞ = (4)
Q X !
m=0 m
(1 − xm )
m=1
Similarly we had
∞
P 2
(−)l x5l +3l+1
∞
l=0
X β2m+1
∞
= ,
Q X
m=0 2m+1
!
(1 − x2m )
m=1
so that by (3)
∞
P
(−)l X l(5l+3)/2 ∞
l=0
X X m(m+1)
∞ = (5)
Q Xm !
(1 − xm ) m=0
m=1
Now the right side in the Rogers-Ramanujan formula is
∞
Q
(1 − x5m )(1 − x5m−2 )(1 − x5m−3 )
1 m=1
∞
= ∞
Q Q
(1 − x5m−1 )(1 − x5m−4 ) (1 − xm )
m=1 m=1
The numerator is the same as the left side of Jacobi’s triple product formula:
∞
Y ∞
X 2
(1 − x2m )(1 − zx2m−1 )(1 − z−1 x2m−1 ) = (−)l zl xl ,
m=1 l=−∞
8. Lecture 57
now
∞
P 2
∞
P 2
∞
P 2
+l (2l)2
(−)l x5l +l (−)k zk xk ∞ (−)l xl x
l=−∞ k=−∞
X Bn (z, x) l=−∞
∞ = ∞ = = ∞ ,
Q Q Xn ! Q
(1 − x2m ) (1 − x2m ) n=0 (1 − x2m )
m=1 m=1 m=1
on replacing z2l by (−)l xl(l+1) , and this we can do because of linearity. Hence 72
∞
P
(−)l X l(5l−1)/2
l=−∞ 1
∞
= ∞
Q Q
(1 − X m ) (1 − x5m−1 )(1 − x5m−4 )
m=1 m=1
Similarly,
∞
Q
(1 − x10m )(1 − x10m−2 )(1 − x10m−8 )
1 m=1
∞ = ∞
Q Q
(1 − X 5m−2 )(1 − X 5m−3 ) (1 − X m )
m=1 m=1
∞
P 2
(−)l x5l +3l
l=−∞
= ∞ .
Q
(1 − X m )
m=1
These formulae are of extreme beauty. The present proof has at least to do
with things that we had already handled. The pleasant surprise is that these
things do come out. The other proofs by Watson, Ramanujan and other use 73
8. Lecture 58
completely unplausible combinations from the very start. Our proof is sub-
stantilly that by Rogers given in Hardy’s Ramanujan, pp.96-98, though one
may not recognize it as such. The proof there contains completely foreign
elements, trigonometric functions which are altogether irrelevant here.
We now give up formal power series and enter into an entirely different
chapetr - Analysis.
Part II
Analysis
59
Lecture 9
Theta-functions
A power series hereafter shall for us mean something entirely different from 74
∞
P
what it did hitherto. x is a complex variable and an xn will have a value, its
n=0
sum, which is ascertained only only after we introduce convergence. Then
∞
X
f (x) = a n xn ;
n=0
x and the series are coordinated and we have a function on the complex domain.
We take for granted the theory of analytic functions of a complex variable; we
shall be using Caushy’s theorem frequently, and in a moment we shall have
occasion to use Weierstrass’s double series theorem.
Let us go back to the Jacobi identity:
∞
Y ∞
X 2
(1 − x2n )(1 + zx2n−1 )(1 + z−1 x2n−1 ) = z k xk
n=1 k=−∞
∞
X 2
=1+ (zk + z−k )xk , (z , 0),
k=1
which is a power series in x. Two questions arise. First, what are the domains
of convergence of both sides? Second, what does equality between the two
sides mean? Formerly, equality meant agreement of the coefficients up to any 75
stage; what it means now we have got to explore. The left side is absolutely
convergent - and absolute convergence is enough for us - for |x| < 1; (for the
Q P
infinite product (1+an ) is absolutely convergent if |an | < ∞; z is a complex
60
9. Lecture 61
variable which we treat as a parameter). For the right side we use the Cauchy-
Hadamard criterion for the radius of convergence:
1
ρ= √
lim n |an |
1
= p
k2
lim |z + z−k |
Suppose |z| > 1; then *****************, and
with increasing N, when |x| < 1, and indeed uniformly so in |x| ≤ 1 − δ < 1.
On the left side we have a sequence of polynomials:
N
Y ∞
X
fN (x) = (1 − x2n )(1 + zx2n−1 )(1 + z−1 x2n−1 ) = a(N) m
m x , say.
n=1 m=0
(of course the coefficients are all zero beyond a certain finite stage). Now we
know that the left side is a partial product of a convergent infinite product; in
fact fN (x) tends uniformly to a series, f (x), say. Now what do we know about
9. Lecture 62
am = lim a(N)
m .
N→∞
The coefficients of the limit function have got something to do with the 77
original coefficients. Now
Z
(N) 1 fN (x)
am = dx
2πi xm+1
|x|=1−δ
(Weirestrass’ own proof of this theorem was what we have given here, in
some disguise; he takes the values at the roots of unity and takes a sort of mean
value).
P 2
Now what are the coefficients in 1 + (zk + z−k )xk ? Observe that the con-
(N) (N)
vergence of am to am is a peculiar and simple one. am indeed converges to a
(N)
known am ; as a matter of fact am = am for N sufficiently large. They reach a
limit and stay put. And this is exactly the meaning of our formal identity. So
the identity has been proved in the function-theoretic sense:
∞
Y ∞
X ∞
X
2 2
(1 − x2n )(1 + zx2n−1 )(1 + z−1 x2n−1 ) = 1 + (zk + z−k )xk = z k xk .
n=1 k=1 k=−∞
These things were done in full extension by Jacobi. Let us employ the ususl 78
symbols; in place of x write q, |q| < 1, and put z = e2πiv . Notice that the right
side is a Laurent expansion in z in 0 < |z| < ∞ (v is unrestricted because we
hace used the exponential). We write in the traditional notation
∞
Y
(1 − q2n )(1 + q2n−1 e2πiv )(1 + q2n−1 e−2πiv )
n=1
∞
X 2
= qn e2πinv
n=−∞
9. Lecture 63
= v3 (v, q)
v3 (and in fact all the theta functions) are entire functions of v. We have taken
|q| < 1; it is customary to write q = eπiτ , so that |q| < 1 implies
V3 (V + 1, q) = V3 (V , q)
∞
Y
V3 (V + τ, q) = (1 − q2n )(1 − q2n−1 e2πiV e2πiτ ) × (1 + q2n−1 e−2πiV e−2πiτ )
n=1
X∞
2
= qn e2πinV e2πinV ,
n=−∞
or
∞
1 + q−1 e−2πiV Y
(1 − q2n )(1 + q2n−1 e2πiV )(1 + q2n−1 e−2πiV )
1 + qe2πiV n=1
∞
X 2
= q−1 e−2πiV q(n+1) e2πi(n+1)V
n=−∞
−1 −2πiV
=q e V3 (V , q)
2πiV −1
= (qe ) V3 (V , q)
9. Lecture 64
V3 (V + τ, q) = q−1 e−2πiV V3 (V , q)
which is denoted V4 (V , q)
Again 81
∞
τ Y
V3 (V + , q) = (1 − q2n )(1 + q2n−1 e2πiV eπiτ )(1 + q2n−1 e−2πiV e−πiτ )
2 n=1
∞
X 2
= qn e2πinV eπinτ
n=−∞
∞
Y
i.e., (1 + e−2πiV ) (1 − q2n )(1 + q2n e2πiV )(1 + q2n e−2πiV )
n=1
∞
X 2
= qn +n e2πinV
n=−∞
9. Lecture 65
∞
X 2
= q−1/4 e−πiV q(n+1/2) e(2n+1)πiV
n=−∞
= q−1/4 e−πiV V2 (V , q)
∞
P 2
where V2 (V , q) = q(n+1/2) e(2n+1)πiV , by definition. (Here q−1/4 does not
n=−∞
contain an unknown 4th root of unity as factor, but is an abbreviation for e−πiτ/4 ,
so that it is well defined). So
∞
Y
V2 (V , q) = 2q1/4 cos πV (1 − q2n )(1 + q2n e2πτV )(1 + q2n e−2πiV )
n=1
Finally 82
!
1+τ 1 1
V3 (V + , q) = q−1/4 e−πi(V + 2 ) V2 V + , q
2 2
!
1 1
= q−1/4 e−πiV V2 V + , q
i 2
X∞
2n+1 2
= q1/4 e−πiV (−)n q( 2 ) e(2n+1)πiV
n=−∞
!
2 1 −πiV
= cos π V + e
i 2
Y∞
(1 − q2n ) 1 − q2n e2πiV × 1 − q2n e−2πiV
n=1
Now define !
1
V1 (V , q) = V2 V + , q ,
2
or
∞
Y
V1 (V , q) = 2q1/4 sin πV (1 − q2n )(1 + q2n e2πiV )(1 − q2n e−2πiV )
n=1
∞
X 2n+1 2
= iq−1/4 (−)n q( 2 ) e(2n+1)πiV
m=−∞
∞
X 2n+1 2
= (−)n q( 2 ) sin(2n + 1)πV
n=0
∞
X 2n+1 2
V2 (V , q) = 2 q( 2 ) cos(2n + 1)πV
n=0
∞
X 2
V3 (V , q) = 1 + 2 qn cos 2nπV
n=1
∞
X 2
V4 (V , q) = 1 + 2 (−)n qn cos 2πnV
n=1
Observe that the sine function occurs only in V1 . Also if q, V are rel these
reduce to trigonometric expansions.
Lecture 10
∞
1 X 2n+1 2
V1 (V , q) = (−)n q( 2 ) e(2n+1)πiV
i n=−∞
∞
X 2n+1 2
=2 (−)n q( 2 ) sin(2n + 1)πV
n=∞
∞
Y
= 2q1/4 sin πV 1 − q2m 1 − q2m e2πiV 1 − q2m e−2πiV (1)
m=1
∞
X 2n+1 2
V2 (V , q) = q( 2 ) e(2n+1)πiV
n=−∞
X∞
2n+1 2
=2 q( 2 ) cos(2n + 1)πV
n=0
∞
Y
= 2q1/4 cos πV 1 − q2m 1 + q2m e2πiV 1 + q2m e−2πiV (2)
m=1
∞
X 2
V3 (V , q) = qn e2πiV
n=−∞
∞
X 2
=1+2 qn cos 2nπV
n=1
∞
Y
= 1 − q2m 1 + q2m−1 e2πiV 1 + q2m−1 e−2πiV (3)
m=1
67
10. Lecture 68
∞
X 2
V4 (V , q) = (−)n qn e2nπiV
n=−∞
∞
X 2
=1+2 (−)n qn cos 2nπV
n=1
∞
Y
= (1 − q2m )(1 − q2m−1 e2πiV )(1 − q2m−1 e−2πiV ) (4)
m=1
Let us study how these functions alter when the argument V is changed by
1, τ, 1/2, τ/2, (1 + τ)/2. V → V + 1 is trivial; V → V + 1/2 is also easy to see
by inspection. Let us take V + τ. (We suppress the argument q for convenience
of writing).
!
1 1/4 2πiV 1+τ
V1 (V ) = q e V3 V +
i 2
!
1 1/4 πi(V +τ) 1+τ
∴ V1 (V + τ) = q e V3 V + τ +
i 2
!
1 1/4 πiV −1 −2πi(V +1+τ/2) 1+τ
= q e qq e V3 V +
i 2
= e−2πiV e−πi(1+τ) V1 (V , q)
= −AV1 (V , q),
where A = q−1 e−2πiV ; the other conspicuous factor which occurs in similar 86
contexts is denoted B = q−1/4 e−2πiV .
10. Lecture 69
The other transformations can be worked out in a similar way by first going
over to V3 . We collect the results below in tabular form.
1 3 1+τ
V +1 V +τ V + 2 V + 2 V + 2
It may be noticed that each column in the table contains all the four func-
tions; so does each now.
The systematique of the notation for the V -functions is rather questionable.
Whittaker and Watson write V instead of πV , which has the unpleasant con-
sequence that the ‘periods’ are then π and πτ. Our notation is the same as in 87
Tannery and Molk. An attempt was made by Kronecker to systematise a little
the unsystematic notation. Charles Hermite introduced the following notation:
∞
X 2r+µ 2
Vµν (V , q) = (−)νn q 2 e(2n+µ)πiV
n=−∞
X∞ 2n+µ 2
πiτ (2n+µ)πiV
= (−)νn e 2 e
n=−∞
V00 (V , q) = V3 (V , q)
V01 (V , q) = V4 (V , q)
V10 (V , q) = V2 (V , q)
V11 (V , q) = iV1 (V , q).
∂
Vα′ (V , q) = Vα (V , q) (α = 1, 2, 3, 4)
∂ν
10. Lecture 70
∂ ∂2
4πi Vµν (V /τ) = 2 Vµν (V /τ) (7)
∂τ ∂ν
This is a partial differential equation of the second order, a parabolic equa-
tion with constant coefficients. It is fundamental to write iτ = −t; (7) then be-
comes the differential equation for heat conduction. V -functions are thus very
useful tools in applied mathematics; they were used by Poisson and Fourier in
this connection.
Again,
∞ !2 2
∂ X 2n + µ 2n+µ
Vµν (V , q) = (−)ν q 2 −1 e(2n+µ)πiV ,
∂q n=−∞
2
∂ ∂2
− 4π2 q Vµν (V , q) = 2 Vµν (V , q), (8)
∂q ∂ν
which is another form of (7). Here the uniformity of notation was helpful; it
was not necessary to discuss the different functions separately.
We now pass on to another important topic. The zeros of the theta - func-
tions.
The V -functions are more or less periodic. The exponential factor that is
picked up on passing from one parallelogram to another is non-zero and can
accumulate. It is evident from the definition that
V , (0, q) = 0.
0 1
V1 (m1 + m2 τ) = 0
!
1
V 2 m1 + m2 τ + =0
2
!
1+τ
V 3 m1 + m2 τ + =0
2
τ
V 4 m1 + m2 τ + =0
2
It is of interest to study Vα (0, q) (usually written Vα ).
V1 (0) = 0
X∞
1 2
V2 (0) = q(2n+ 2 )
n=−∞
∞
Y
= 2q1/4 (1 − q2m )(1 + q2m )2
m=1
10. Lecture 72
= V2
X∞
2
V3 (0) = qn
n=−∞
∞
Y
= (1 − q2m )(1 + q2m−1 )
m=1
X∞
2
V4 (0) = (−)n qn
n=−∞
Y∞
= (1 − q2m )(1 − q2m−1 )2
m=1
or, replacing q2 by x,
∞
Y ∞
X
(1 − xn )3 = (−)n (2n + 1)xn(n+1)/2
n=1 n=0
We had proved this earlier by the method of formal power series. Here we
can differentiate with good conscience.
Now
∞ 2
′
Y 2m 2m−1 2m−1
πV2 V3 V4 = V1 (1 + q )(1 + q
)(1 − q )
m=1
∞ 2
′
Y 2m 4m−2
= V1 (1 + q )(1 − q ) ,
m=1
10. Lecture 73
∞
Q
However, (1 + xn )(1 − x2n−1 ) = 1. We therefore have the very useful and
m=1
pleasant formula
V1′ = πV2 V3 V4
Lecture 11
74
11. Lecture 75
0 1
i.e., the sum of the residues at the poles =0. This means that either the
pole is a double is a double pole with zero residue, or there are two simple
poles with residues equal in magnitude but opposite in sign. However neither
of these is the case. So there is no necessity for any further experimentation.
Let us therefore consider the squares
f22 (V ), f32 (V ), f42 (V )
these are indeed doubly periodic functions. And they are even functions. So
the expansion in the neighbourhood of the pole will not contain the term of
power −1. Hence the pole must be a double pole with residue zero. So they
are closely related to the Weierstrassian function P(V ), and must indeed be
of the form CP(V ) + C1 .
So we have constructed doubly periodic functions. They are essentially 95
P(V ). ω1 and ω2 of P(V ) are our 1 and τ. In order to get a better insight
we need the exact values of the functions. Let us consider their pole terms.
Expanding in the neighbourhood of the origin,
V ′′
Vα (V , q) Vα + 2!α V 2 + · · ·
= V′ V1′′′
V1 (V , q) 3
1! V + 3! V + · · ·
1
′′
1 Vα 2
Vα 1 + 2 Vα ν + · · ·
= ′′′
νV1′ 1 + V1 ν2 + · · ·
6V1′
! !2
Vα 1 Vα′′ 2 V ′′′
= 1 + ν +··· 1 − 1 ν2 + · · · + (· · · )3 − · · ·
νV1′ 2 Vα 6V1 ′
′′ ′′′ ! !
Vα V V
= 1 + ν2 − 1 ′ +···
νV1′ 2Vα 6V1
!2
Vα (V , q)
∴ fα2 =
V1 (V , q)
11. Lecture 76
! !
Vα2 1 2 Vα
′′ V1′′′
= ′ 1+V − +···
V 21 V 2 V1 3V1′
The left side is a doubly periodic function without a pole and so a constant
V ′′ V ′′
C; the right side is therefore just 2 − 4 . The vanishing of the other terms
V2 V4
on the other terms on the right side, of course, implies lots of identities.
So we have already computed C in one way:
V2′′ V4′′
C= −
V2 V4
1 τ
To evaluate C in other ways we may take in (*) V = , V = or V =
2 2
(1 + τ)/2. From the table,
! !
1 1+τ
V1 , q = V2 V1 , q = q−1/4 V3
2 2
! !
1 1+τ
V2 , q = −V1 = 0 V2 , q = −iq−1/4 V4
2 2
! !
1 1+τ
V4 , q = V3 V4 , q = q−1/4 V2
2 2
Also
′ ′
V1 2 V42 V1 2 V22
C = 2 − 2 − 2 2
V2 V3 V4 V3
′ ′
π2 V1 2 V44 π2 V1 2 V24
=− −
π2 V22 V32 V42 π2 V22 V32 V42
= −π2 V44 − π2 V24
We have also
V4′′ V2′′
π2 V34 = − (2)
V4 V2
Now let us look at (1) and do a little computing. Explicitly (1) states:
∞ 4 ∞ 4
X n2 4 1/4 X
∞
n(n+1)
X n n2
q = q q + (−) q (3)
n=−∞ n=−∞ n=−∞
on the right. If N is even, it is trivial that both sides are in agreement because
the first term on the right gives only odd powers of q, and the coefficient of qn
in the second term on the right is
X
(−)n1 +n2 +n3 +n4
n21 n22 +n23 +n24 =N
Since N is even either all ni ’s are odd, or two of them odd, or none. It is not
transperent. What happens when N is odd.
Take the more interesting formula (2):
V4′′ V2′′
πV34 = −
V4 V2
By the differential equation,
" #
∂2
Vα′′ = Vα (V , q)
∂V 2 V =0
11. Lecture 78
" #
∂
= −4π2 q Vα (V , q)
∂q V =0
!
4 1 ∂V 2 1 ∂V 4
∴ V3 = 4q −
V2 ∂q V4 ∂q
∂ V2
= 4q log
∂q V4
Now 99
∞
Q
(1 − q2n )(1 + q2n )2
V2 n=1
= 2q1/4 ∞
V4 Q
(1 − q2n )(1 − q2n−1 )
n=1
Q∞
(1 − q2n )2 (1 + q2n )2
n=1
= 2q1/4 ∞
Q
(1 − q2n )2 (1 − q2n−1 )2
n=1
Q∞
(1 − q4n )2
n=1
= 2q1/4 ∞
Q
(1 − qn )2
n=1
1
2q /4
= Q
(1 − qn )2
4∤n
We have not partitions this time, but representation as the sum of squares.
We agree to consider as distinct these representations in which the order of the
components has been changed. In partitions we abstracted from the order of 101
the summands; here we pay attention to order, and also to the sign (i.e., one
representation n21 + n22 + n23 + n24 is actually counted, order apart, as 16 different
representations (±n1 )2 +(±n2 )2 +(±n3 )2 +(±n4 )2 , if n1 , n2 , n3 , n4 are all different
from 0).
As an example, take m = 10. The different representations as the sum of
four squares are
r4 (10) = 6 × 16 + 6 × 8 = 144
11. Lecture 80
0 1 2
This may be deduced by checking that the poles on both sides are the same,
Further they are odd functions and so the constant term in the difference must
1
vanish. Put V = on both sides.
2
11. Lecture 81
1
V1′ 41 τ2 V1′ 43 τ2
πV32 =
. − .
2 V1
1 τ 3
V1 4 2 τ
4 2
f (V ) = V1 (V τ/τ)
Then
f (V + 1) = V1 ((V + 1)τ/τ)
= V1 (V τ + τ/τ)
= −e−πiτ e−2πiV τ V1 (V τ/τ)
= e−πiτ e−2πiV τ f (V )
82
12. Lecture 83
= − f (V )
So f is a sort of V -function which picks up simple factors for the ‘periods’ 1 106
and − 1τ . f (V ) has clearly zeros at 0 and τ′ = − τ1 , or generally at V = m1 +m2 τ′ ;
m1 , m2 integers, which is a point-lattice similar to the old one turned around.
Similarly let us define
!
′ 1
g(V ) = V1 (V /τ ) = V1 V / −
τ
g(V + 1) = V1 (V + 1/τ′ )
= −V1 (V /τ′ )
= −g(V )
!
1
g V − = g(V + τ′ )
τ
= V1 (V + τ′ /τ′ )
′
= −e−πiτ e−2πiV V1 (V /τ′ )
= −eπi/τ e−2πiV g(V )
f (V )
Φ(V ) =
g(V )
f (V + 1)
φ(V + 1) =
g(V + 1)
= −e−πiτ e−2πiV τ φ(V )
!
1 f (V + τ′ )
Φ V − =
τ g(V + τ′ )
f (V )
= πi/τ −2πiV
e e g(V )
= e−πi/τ e2πiV Φ(V )
Φ takes on simple factors in both cases of this peculiar sort that we can
eliminate them both at one stroke. We write
Let us try the following trick. Let us supplement Φ(V ) by an outside func-
tion h(V ) so that the combined function Φ(V ) is totally doubly periodic. Write
or
We can solve both at one stroke. Since on the right side we have a linear
function of V in both cases, a quadratic polynomial will do what we want.
2 V1 (V τ/τ)
Ψ(V ) = eπiτV
V1 V / − 1τ
I. !
2 1
eπiτV V1 (V τ/τ) = C(τ)V1 V / −
τ
What we need now are the corresponding formulas for the other functions.
Replacing V by V + 12 ,
! ! !
πiτ(V + 21 )2 1 1. 1
e V1 V + τ/τ = C(τ)V1 V + − ,
2 2 τ
!
πiτ(V 2 +V +1/4) −πiτ/4 −πiV τ 1
or e ie e V4 (V τ/τ) = C(τ)V2 V / −
τ
We notices here that two different V -functions are related. This gives
II. !
πiτV 2 1
ie V4 (V τ/τ) = C(τ)V2 V / − .
τ
Replacing in I V by V + τ′ /2 = V − 1/(2τ), we get
III. !
πiτV 2 1
ie V2 (V τ/τ) = C(τ)V4 V / −
τ
1
Finally putting V + 2 for V In, III,
IV. !
2 1
ieπiτV V3 (V τ/τ) = C(τ)V3 V / −
τ
The way the functions change over in I-IV is quite plausible. For consider 110
the location of the zeros.
When we take the
parallelogram and
turn it around what
was originally a
zero for V4 becomes
one for V2 and vice 3
4
versa; and what
used to be in the
middle, the zero of 1 2
V3 , Remains in the
middle. So the for-
mulae are plausible
in structure.
The most important thing now is, what is C(τ)? To evaluate C(τ) let us
differentiate I and put V = 0. We have
12. Lecture 86
V. !
1
τV11 (0/τ) = C(τ)V11 0/ −
τ
From II, III, and IV, putting V = 0,
!
1
iV4 (0/τ) = C(τ)V2 0/ −
τ
!
1
iV2 (0/τ) = C(τ)V4 0/ −
τ
!
1
iV3 (0/τ) = C(τ)V3 0/ −
τ
1
= C 2 (τ),
iτ
r
1
or C(τ) = ±
iτ
C(τ)
In II, III, IV, it is i that appears; so let us write this is
r r
C(τ) 1 1 i
=± =±
i i iτ τ
C(τ)
Now k(i/τ) > 0 · i is completely determined, analytically, in particular
by IV:
P πiτn2
C(τ) V3 (0/τ) e
= 1
= P πiτ′ n2
i V3 (0/ − τ ) e
Both the numerator and denominator are analytic functions if Im τ > 0. So
C(τ)
i is analytic and therefore continuous. i/τ must lie in the right half-plane,
q
and thus τi in either of the sectors with central angle π/2, but because of
continuity it cannot lie on the border lines. So it is in the interior of entirely
one sector. To decide which one it is enough if we make one choice.
12. Lecture 87
∞ r ∞
X
−πin2 /τ τ X πin2 τ
e = e
n=−∞
i n=−∞
∞ r ∞
−πiτ/12
Y
−2πim/τ τ πiτ/12 Y
e (1 − e )=ǫ e (1 − e2πimτ )
m=1
i m=1
Then ! r
1 τ
η − =ǫ η(τ)
τ i
3
This . pis challenging; we have to decide which ǫ to take: ǫ = 1. The quotient
η(− 1τ ) τi η(τ) is an analytic (hence contains) function in the upper half-plane
and so must be situated in on of the three open sectors. Now make a special
choice; put τ = i. Then η(i) = ǫ(+1)η(i), or ǫ = 1.
! r
1 τ
∴ η − = η(τ)
τ i
What we have done by considering the lattice of periods can be done in 114
more sophisticated ways. One can have a whole general theory of the transfor-
aτ + b
mations from 1, τ, to 1, . The quotients appear first and can be carried
cτ + d
over. We start with V1 and come back to it; there may be difficulty, however in
deciding the sign.
Lecture 13
89
13. Lecture 90
Be sure that we want to go from 1 to τ′ through an angle less than π in the 116
positive sense. For this we want τ′ to have a positive imaginary parts Imτ̄′ > 0,
or
τ′ − τ̄′
>0
i !
1 aτ + b aτ̄ + b
i.e., − >0
i cτ̄ + d cτ̄ + d
1 adτ + bcτ̄ − ad τ̄ − bcτ
i.e., >0
i |cτ + d|2
1 (ad − bc)(τ − τ̄)
i.e., >0
i |cτ + d|2
or since τ − τ̄ is purely imaginary,
ad − bc = ±1.
We could do the same thing in all our different steps. The most important
step, however, cannot be carried through, because we get lost at an important
point; and rightly so, it becomes cumber some because a number-theoretic
problem is involved there. Let us see what we have done. Compare
. aτ + b !
V1 ((cτ + d)V /τ) and V1 V
cτ + d
We want periods 1, τ′ ; indeed all things obtainable from ω1 = cτ + d and
ω2 = aτ + b; or m1 ω1 + m2 ω2 must in their totality comprise all periods. For
the first cτ + d is indeed a period, and for the second aτ + b.
Now define 117
f (V ) = V1 ((cτ + d)V /τ)
f (V + 1) is essentially f (V ):
f (V + 1) = V1 ((cτ + d)V + cτ + d/τ)
2
= (−)c+d e−c πiτ −∗∗∗∗∗(cτ+d)V
e V1 ((cτ + d)V /τ), from the table,
= (· · · · · · ) f (V )
′
f (V + τ ) = V1 ((cτ + d)ν + aτ + b/τ)
2
= (−)a+b e−a πiτ −2πia(cτ+d)ν
e V1 ((cτ + d)V /τ)
= (· · · · · · ) f (V )
′
f (V ) has, leaving
trivial
. factors aside, periods 1, τ *****. So too for the
aτ+b
second function V1 V cτ+d .
13. Lecture 91
We take one liberty from now on. Take µ, ν to be arbitrary integers, no 118
longer 0, 1. That will not do very much harm either. In fact,
For
X 2
Vµ+2,ν (v/τ) = (−)ν eπiτ(n+µ/2) e2πiν(n+1+µ/2)
n
X 2
= (−)ν (−)v(n+1) eπiτ(n+1+µ/2) e2πiv(n+1+µ/2)
n
= (−)v Vµν (v/τ),
119
This one formula has the whole table in it.
We now turn to our purpose, viz. To consider the quotient
V1 ((cτ + d)ν/τ)
.
V1 ν aτ+b
cτ+d
Φ takes up exponential factors which contain ν linearly. As before ewe can 121
submerge this under a general form. Define
Ψ(ν) = Φ(ν)eh(ν) ,
we therefore want
2
eh(ν+1)−h(ν) (−)c+d+1 e−c πiτ−2πic(cτ+d)ν = 1,
′ 2
πiτ+πiτ′ +2πiν −2πia(cτ+d)ν
eh(ν+τ )−h(ν) (−)a+b+1 e−a e = 1.
h(ν) = Aν2 + B
for each separately and see whether it works for both. 122
A + B = πic(cτ + d),
!2 !
aτ + b aτ + b πic(aτ + b)2
A +B =
cτ + d cτ + d cτ + d
13. Lecture 94
C = C(τ; a, b, c, d)
More generally we can have a parallel formula for any µ, ν. As before we 123
get an equation for C 2 . And there the thing stops. Formerly we were in a very
good position with the special matrix
a b 0 −1
= .
c d 1 0
We were considering the behaviour of V11 (ν/τ) under the general modular 124
transformation:
. aτ + b ! 2
V11 ν = C(τ)eπiC(cτ+d)ν V11 ((cτ + d)ν/τ), (1)
cτ + d
a b
a, b, c, d integers with = +1.
c d
We want to determine C(τ) as far as possible. We shall do this up to a ±
sign. ν is unimportant at the moment; even if we put ν = 0, C(τ) survives. Put
′ ′
ν = 21 , τ2 , 1+τ
2 in succession, ans use out auxiliary formula which contracted
the whole table into one thing:
!
k lτ . 2
Vµν v + + τ = ikµ e−πiτl /4 e−πiν Vµ+l,v+k (v/τ) (*)
2 2
1 aτ + b
Putting ν = 2 in (1), and writing τ′ = ,
cτ + d
! !
1. ′ πic(cτ+d)/4 cτ + d .
V11 τ = C(τ)e V11 τ (2)
2 2
This is the right moment to call for formula (*). From (*) with ν = 0,
µ = ν = 1, k = 1, l = 0, we get
!
1. ′
V11 τ = iV12 (0/τ′ )
2
Also from (*) with ν = 0, µ = ν = 1, k = d, l = C, we get
!
cτ + d . 2
V11 τ = id C −πic /4 V1+c,1+d (o/τ).
2
95
14. Lecture 96
Substituting these two formulas in the left and right sides of (2) respec- 125
tively, we get
2
iV12 (0/τ′ ) = C(τ)eπic(cτ+d)/4 id e−πiτc /4
V1+c,1+d (0/τ)
Making use of (*) in succession on the left and right sides (with proper
choice of indices) as we did before, this gives
′ ′ 2
e−πiτ /4 V12 (0/τ′ ) = C(τ)eπicτ (aτ+b)/4i be −πia τ/4
V1+a,1+b (0/τ),
and this, after slight simplification of the exponents on the right sides, gives in
view of (**),
−V01 (0/τ′ ) = C(τ)ib eπiab/4 V1+a,1+b (0/τ) (4)
Putting ν = (1 + τ′ )/2 in (1), 126
! !
1 + τ′ . ′ πic/4(1+τ′ )(la+c)τ+b+d) (a + c)τ + l + d .
V11 τ = C(τ)e V11 τ
2 2
Observe that the sum of the first subscripts on the right side = 3+2a+2c ≡ 1 127
(mod 2). So either all three numbers 1 + a, 1 + c, 1 + a + c are odd, or one of
them is odd and two even. Then first case is impossible since we should then
have both a and c even and so ac db , 1. So two of them are even and one
odd. The even suffixes can be reduced to zero and the odd one to 1 by repeated
application of (**). Similarly for the second suffixes. So the V -factors on the
right will be V00 , V01 , V10 . What we hate is the combination 1, 1 and this does
not occur. (If it did occur we should have V11 which vanishes at the origin).
Although we can not identify the V -factors on the right, we are sure that we
get exactly the combinations that are desirable: 01, 10, 00. The dangerous
combination is just out.
Let us reduce the subscripts by stages to 0 or 1 as the case may be. When
we reduce the second subscript nothing happens, whereas when we reduce
index by steps of 2, each time a factor ±1 is introduced, by virtue of (**). By
1+c
the time the subscript 1 + c is reduced to 0 or 1, a factor (−)[ 2 ] (1 + d) will
have accumulated in the case of V1+c,1+d . Similarly in the case of V1+a,1+b and
V1+a+c,1+b+d . Altogether therefore we have a factor 128
1+c
(−)[ 2 ](1+d)+[ 1+a 1+a+c
2 ](1+b)+[ 2 ](1+b+d) ,
and when this compensating factor is introduced we can write V00 , V11′ and V10 .
Hence our formula becomes
where
# " " # " #
1+c 1+a 1+a+c
α=b+d+ (1 + d) + (1 + b) + (1 + b + d)
2 2 2
real fixed points -the hyperbolic case. Here the fixed are not accessible to us
because they are quadratic algebraic numbers on the real axis.
In the elliptic case with |a + d| < 2 we could finish the thing without much
trouble. In the parabolic case we are already in a fix. Much more difficult is
the hyperbolic case.
If ξ1 and ξ2 are the fixed
points, τ and τ′ will lie on the
same circle through ξ1 and ξ2 ,
and repetitions of the transfor-
mation would give a sequence
of points on the same circle
which may converge to either ξ1
or ξ2 . So the ambiguity in the ±
sign will remain.
It will be much more difficult when we pass from V to η, because then we
shall have to determine a cube root.
Lecture 15
We were discussing the possibility !of getting a root of unity determined for 131
. aτ + b
the transformation of V11′ ν . There do exist methods for determining
cτ + d
this explicitly. Only we tried to carry out the analogue with the special case
as far as possible, not with complete success. other methods exist. The first of
these is due to Hermite, done nearly 100 years ago. He used what are called
Gaussian sums. There are difficulties there too and we want to avoid them.
Another method is that of Dedekind using Dedekind sums.
In the special case of the transformation from τ to τ′ = − 1τ we were faced
with an elliptic substitution ac bd . These are of two sorts:
1. a + d = 0
2. a + d = ±1
In both cases we can completely forget about the root of unity if we remember
the following fact. Our formula had the following shape:
r
. aτ + b ! cτ + d cτ + d
′
V11 0 = · ρ(a, b, c, d)V11′ (0/τ) (*)
cτ + d i i
where ρ is a root of unity which is completely free of τ. we can then get things
straightened out. We have only to consider the fixed points of the transforma-
tion given by p
a − d ± (a + d)2 − 4
ξ=
2c
aξ + b
Put ξ on both sides of the formula; since ξ = , both sides look alike 132
cξ + d
′
and V11 does not vanish for appropriate τ in the upper half-plane (we may take
c > 0), so that ρ is given directly by the formula.
100
15. Lecture 101
Case 1. a + d = 0 √
−2d ± −d −d + i
ξ= =
2c c
(reject the negative sign since we want a point in the upper half-plane).
cξ + d i
= =i
i 1
∴ V11′ (0/ξ) = ρ(a, b, c, d)V11′ (0/ξ)
1 = e∓πi/4 ρ(a, b, c, d)
∴ ρ(a, b, c, d) = e±πi/4
when a + d = ±1, (we may take c > 0; the case c = 0 is uninteresting and
if c < 0 we can make it c > 0).
There are unfortunately no more cases like these.
Parabolic case. The analysis here is a little longer but it is worth while working
it out. Now a + d = ±2, and there is only one fixed point
a − d −2d ± 2 −d ± 1 δ
ξ= − = =−
2c 2c c γ
where (γ, δ) = 1 and we may choose γ > 0. The fixed point is now a rational
point on the real axis. We try to approach it. This is a little difficult because
15. Lecture 102
we do not know what the function will do thee. But by an auxiliary transfor-
mation we can throw this point into the point at infinity. Consider the auxiliary
transformation
Aτ + B A B
T= , =1
γτ + δ γ δ
Dividing, we get
q ′ 3
γτ +δ
V11′ (0/T ′ ) i
′ ′
V11 (0/τ )
= q (1)
V11′ (0/T ) γτ+δ V11′ (0/τ)
i
The left side gives the behaviour at infinity. We cannot of course put τ = ξ.
Put τ = ξ + it, t > 0, and later make t → 0. τ is a point in the upper half-plane.
τ − ξ = it,
aτ + b aξ + b
τ′ − ξ = −
cτ + d cξ + d
τ−ξ
=
(cτ + d)(cξ + d)
it
=
1 ± ict
This is also in the upper half-plane. τ′ → ξ as t → 0
Let us calculate T and T ′ . For this consider
Aτ′ + B Aτ + B
T′ − T = −
γτ′ + δ γτ + δ
15. Lecture 103
τ′ − τ
=
(γτ′ + δ)(γτ + δ)
c
=∓
γ2
135
This is quite nice; the difference is a real number.
Aτ + B A(ξ + it) + B
T= =
γτ + δ γ(ξ + it) + δ
Ait + Aξ + B
= , since γξ + δ = 0,
γit
δ
A −A γ + B
= +
γ γit
A Bγ − Aδ
= +
γ γτt
A 1
= + , since Bγ − Aδ = −1
γ γ2 t
A c i
T′ = ± +
γ γ2 r 2 t
→ i∞ as t → 0
∞
Y
V11′ (0/T ) = 2πieπiT/4 (1 − e2πinT )3
n=1
A i
Let T= γ + γ2 t
. Then
2
e2πinT = e2πinA/γ e−2πnt/γ → 0
∴ V11′ (0/T ) = 2πieπiT/4 (a factor tending to 1)
We do not know what happens to eπiT/4 . But we need only the quotient. So
V11′ (0/τ) ′
∼ eπi(T −T )/4
V11′ (0/τ)
2
= e∓πic/4γ (2)
15. Lecture 104
Consider similarly the quotient V11′ (0/τ′ )/V11′ (0/τ). We have, since γξ+δ =
0,
γτ + δ γ(ξ + it) + δ
= = γt
i i
r
γτ + δ √
or = νt, where we take the positive square root
i
it
γτ′ + δ γ ξ + 1±ict + δ γt
= =
i i 1 ± ict
r r
γτ′ + δ √ 1
∴ = rt (both branches principal)
i 1 ± ict
√
∼ γt as t → 0.
r
γτ′ + δ . γτ+δ
q
Hence the quotient i behaves like 1. 137
i
And so we have what we were after:
V11′ (0/τ′ ) 2
→ e∓πic/4γ as t → 0 (3)
V11′ (0/τ)
a−d
cτ + d c(ξ + it) + d + cit + d
= = 2
i i i
a + d2 + cit 1 + cit
= =±
i i
= ∓i + ct
→ ∓i as t → 0.
What will the square root of this do?
r
cτ + d √
= ct ∓ i, and this
i
does lie in the proper half plane
because ct > 0. For small t
it will be very near the imagi-
nary axis near ∓i. So the square
root lies in the sector, in the
lower half plane if we choose
√
−i = e−πi/4 , and in the√upper
half-plane if r
we choose +i =
cτ + d
eπi/4 Hence → e∓πi/4
i
as t → 0.
15. Lecture 105
Using this fact as well as (2) and (3) in (*) we get 138
2
e∓πic/4γ = e∓3πi/4 ρ(a, b, c, d)
πi
∴ ρ(a, b, c, d) = e∓ (c/γ2 − 3)
4
We observe that the common denominator (γ, δ) = 1 plays a role, however
a and b do not enter.
Hyperbolic case. The thing could also be partly considered in the hyperbolic
case. It will take us into deeper things like real quadratic fields and we do not
propose to do it.
Let us return to what we had achieved in the specific case. We had a formula
for η(τ): r
′ cτ + d
η(τ ) = ǫ(a, b, c, d)η(τ),
i
where ǫ(a, b, c, d) is a 24th root of unity. This shape we have in all circum-
stances. The difficulty is only to compute ǫ. We shall not determine it in
general, and we can do away with it even for the purpose of partitions by using
a method developed recently by Selberg.
However in each specific case we can compute ǫ.
! r
1 τ
η − = η(τ)
τ i
Now
∞
Y
η(τ) = eπiτ/12 (1 − e2πinτ ),
n=1
η(τ + b) = eπib/12 η(τ)
Out of these two facts we can get every other one, because the two substi- 139
tutions
1 1 0 −1
S = ,
T =
0 1 1 0
form generators of the full modular group. This can be shown as follows. Take
c > 0.
3τ + 4
Let us break into simpler substitutions,
2τ + 3
3τ + 4 1
τ3 = =1− ,
2τ + 3 τ2
τ2 = −2 + τ1 ;
1
τ1 = −
τ+1
! r
1 τ+1
η(τ1 ) = η − = η(τ + 1)
τ+1 i
r
τ + 1 πi/12
= e η(τ).
i
η(τ2 ) = η(τ1 − 2) = e−πi/6 η(τ1 )
r
τ + 1 −πi/12
= ·e η(τ)
i
! r
1 τ2
η − = η(τ2 )
τ2 i
r r
τ + 1 τ2 −πi/12
= e η(τ)
i i
The two square roots taken separately are each a principal branch, but taken 140
15. Lecture 107
Here we are faced with a question: which square root are we to take? 141
√ πi/4
q
2τ+3
We write 3 + 2τ = e i
Let us look into each root singly. For τ = it where do they go?
r r
τ+1 it + 1
=
i i
→ ∞ with argument 0 as t → ∞.
s s
−2τ − 3 −2it − 3 √
= → 2i as t → ∞,
i(τ + 1) i(τ + 1)
or its argument = π/4
r r
τ + 1 −2τ − 3
The product has here argument π/4, so that it continues
i i(τ + 1)
to be the principal branch. Of course in a less favourable case, if we had two
other arguments, together they would have run into something which was no
longer a principal branch. Finally,
! r
3τ + 4 πi/4 2τ + 3
η =e η(τ)
2τ + 3 i
and here there is no ambiguity. Actually in every specific case that occurs one
can compute step and make sure what happens.
There does exist a complete formula which determines ǫ(a, b, c, d) explic-
itly by means of Dedekind sums S (h, k).
Part III
108
Lecture 16
Our aim will be now to get an explicit formula for p(n) and things connected 142
with it. Later we shall return to the function η(τ) and the discussion of the sign
of the square root. That will again lead us into some aspects of the theory of
V -functions connected with quadratic residues.
Let us come to our topic. Euler had, as we know, the identity:
∞
X 1
p(n)xn = ∞
.
Q
n=0 (1 − xm )
m=1
109
16. Lecture 110
h + h′
, provided that the denominator is of the proper size. Following Hardy
k + k′
h + h′ h h′ h
and Littlewood we call ′
the ‘mediant’ between and ′
. We have <
k + k k k k
h + h′ h′
< ′ , so that the order is automatically the natural order. We call
k + k′ k
that row which has denominator k ≤ N, the Farey series of order N. We get
this by forming mediants from the preceding row. Farey made the following
observation. Take two adjacent fractions in a row; then the determinant formed
by their numerators and denominators is equal to −1. For instance, in the fifth
1 2
1 2
row and are adjacent and = −1***********. If we now prove that
3 5
3 5
new fractions are always obtained by using mediants, then we can be sure, by 145
induction, that this determinant is always −1. For, let
h h′
= −1; then
k k′
h h + h′ h + h′ h′
= −1 =
k k + k′ k + k′ k′
−Hk2 + Kh2 = µ,
so that λ and µ are integers > 0. Solving for H and K by Cramer’s rule, 146
λ −h1 λ −k
1
µ h2 µ k2
H = , K =
h k2
h2 k2
2
h1 k1
h1 k1
H = λh2 + µh1
K = λk2 + µk1
H λh2 + µh1
or = .
K λk2 + µk1
What do we know about K? K did not appear in a series of order K − 1;
k1 and k2 are clearly less than K. What we have found out so far is that any
h1 h2 λh2 + µh1
fraction lying between and can be put in the form . Of these
k1 k2 λk2 + µk1
only one interests us - that one with lowest denominator. This comes after the
ones used so far. Look for the one with lowest denominator; this corresponds
to the smallest possible λ, µ, i.e., λ = µ = 1. Hence first among the many later
H h1 + h2
appearing ones is = , i.e., if in the next Farey series a new fraction
K k1 + k2
is called for, that is produced by a mediant. So what was true for K − 1 is true
for K; and the thing runs on.
One remark is interesting, which was used in the Hardy - Littlewood- Ra-
h1 h2
manujan discussion. In the Farey series of order N, let and be adjacent
k1 k2
h1 h2 h1 + h2
fractions. < · does not being these. It is of higher order. This
k1 k2 k1 + k2
says that k1 + k2 > N. For two adjacent fractions in the Farey series of order N, 147
the sum of the denominators exceeds N. Both k1 and k2 ≤ N, so
2N ≥ k1 + k2 > N.
k1 and k2 are equal only in the first row, otherwise it would ruin the deter-
minant rule. So
2N > k1 + k2 > N, N > 1.
16. Lecture 113
This was very often used in the Hardy - Ramanujan discussion. (The Farey
series is an interesting way to start number theory with. We can derive from it
Euclid’s lemma of decomposition of an integer into primes. This is a concrete
way of doing elementary number theory).
We now come to the special path of integration. For this we use Ford
Circles (L.R. Ford, American Mathematical Monthly, 45 (1938), 568-601).
We describe a series of circles in the upper half-plane. To each proper fraction
h h i 1
we associate a circle Chk with centre τhk = + 2 and radius 2 , so the
k k 2k 2k
circles all touch the real axis.
Take another Ford circle Ch′ k′ , with centre at τh′ k′ . Calculate the distance
between the centres.
!2 !2
2 h h′ 1 1
|τhk − τh′ k′ | = − + − .
k k′ 2k2 2k′2
1 1
The sum of the radii = 2k2
+ 2k′2
148
! !2
2 1 1 h h′ 1
|τhk − τ | −
h′ k′ + = − − 2 ′2
2k2 2k′2 k k′ k h
(hk′ − h′ k)2 − 1
= ≥ 0,
k2 k′2
since h, k are coprime and so h k is an integer , 0. So two Ford circles never
h′ k′
intersect. And they touch if and only if
h k
= ±1,
′ ′
h k
h h′
i.e., if in a Farey series , ′ have appeared as adjacent fractions.
k k
16. Lecture 114
h1 h
In a certain Farey series of order N we have adjacent fractions < <
k1 k
h2
. (We know exactly which are adjacent ones in a specific series). Draw also 149
k2
the Ford circles Ch1 k1 and Ch2 k2 . These touch Chk . Take the arc γhk of Chk from
one point of contact to the other in the clockwise sense (the arc chosen is the
one not touching the real axis). This we do for all Farey fractions of a given
order. We call the path belonging to Farey series of order N PN . Let us describe
this in a few cases.
We fix α = i and pass to α + 1 = i + 1. Take N = 1; we have two circles of
i i
radii 2 each with centres at and 1 +
2 2
0 1
arc from − 12 + 2i to i. Next consider Farey series of order 2; and are no
1 1
longer adjacent. The path now comprises the arc of C01 from i to the point of
contact with C12 , the arc of C12 from this point to the point of contact with C11 150
and the arc of C11 from this point to i + 1. Similarly at the next stage we pass
from i on C01 to i+1 on C11 through the appropriate arcs on the circles C13 , C12 ,
C23 in order. So the old arcs are always retained but get extended and new arcs
spring into being and the path gets longer and longer. At no stage does the path
intersect itself, but these are points of contact. The path is complicated and was
not invented in one sitting. The Farey dissection of Hardy and Ramanujan can
be pictured as composed of segments parallel to the imaginary axis. Here it is
more complicated.
We need a few things for our consideration. We want the point of contact
of Chk and Ch′ k′ . This is easily seen to be the point
1 1 ! !
2k′2 2k2 h i k2 h′ i k′2
τhk + τh′ k′ = + 2 2 ′2
+ ′
+
1
2k2
+ 1
2k′2
1
2k2
+ 1
2k′ 2
k 2k k + k k 2k k + k′2
′2 2
!
h h′ h k′2 i
= + ′ − +
k k k k2 + k′2 k2 + k′2
′ k′ i
where ζhk = + 2 . We notice that the imaginary part 1/(k2 +k′2 )
k(k2 + k ) k + k′2
′2
gets smaller and smaller as k + h′ lies between N and 2N. Each arc runs from
h ′ h ′′
k + ζhk to k + ζhk . Such an arc is the arc νhk . No arc touches the real axis.
We continue our study of the integral. Choose a number N, the order of the
Farey series, and cut the path of integration PN into pieces γhk .
Z
p(n) = f (e2πiτ )e−2πinτ dτ
PN
X Z
= f (e2πiτ )e−2πinτ dτ
(h,k)=1 γ
0≤h<k≤N hk
h ′ h
′′
(γhk goes from k + ζhk + ζhk
to k ; these are all arcs of radii 1/2k2 ). We make 152
iz
a further substitution: put ζ = 2 , so that we turn round and have everything
k
in the right half-plane, instead of the upper half-plane. (All these are only
preparatory changes; there is no actual mathematical progress as yet). Then
′′
Zzhk
X e−2πinh/k h iz 2
p(n) = i f (e2πi( k + k2 ) )e2πnz/k dz
(h,k)=1
k2
z′hk
0≤h<k≤N
k2 + ikk′
z′hk =
k2 + k′2
k2 − ikk′′
z′′hk = 2
k + k′′2
So what we have achieved so for is to cut the integral into pieces. 153
16. Lecture 117
′′
Zzhk
X i −2πinh/k h z 2
p(n) = e f e2πi( k + k2 ) e2πinz/k dz
(h,k)=1
k2
z′hk
0≤h<k≤N
n is a fixed integer here, n R 0 and p(n) = 0 for n < 0; and this will be of some
use later, trivial as it sounds. N is the order of the Farey series. We have to deal
with a complicated integrand and we can foresee that there will be difficulties
as we approach the real axis. However, f is closely related to η:
h iz
For us τ = + 2 .
k k
We can now use the modular transformation. We want to make Im τ large so
aτ + b
that we obtain a big negative exponent. So we put τ′ = , a, b, c, d being
′
cτ + d
chosen in such a way for small τ, τ becomes large. This is accomplished by 155
taking kτ − h in the denominator; kτ − h = 0 when z = 0 and close to zero
aτ + b
when z is close to the real axis. We can therefore put τ′ = where a, b
kτ − h
a b
should be integers such that k −h = 1. This gives −ah − bk = 1 or ah ≡ −1
118
17. Lecture 119
(mod k). Take a solution of this congruence, say h′ i.e., choose h′ in such a
way that h′ h ≡ −1 (mod k), which is feasible since (h, k) = 1. As soon as h′ is
found, we can find b. Thus the matrix of the transformation would be
a b h′ − hh′ +1
= k
c d k −h
and there is no doubt about the square root - it is the principal branch. We write
πi h h′
−
= ǫe 12 k k
ωhk
P′
where denotes summation over h and k, (h, k) = 1. Now what is the 157
advantage we have got? Realise that
∞
X
f (x) = p(n)xn = 1 + x + · · ·
n=0
Then
Zhk z′′
X′ iωhk 2
−2πinh/k
p(n) = 5/2
e e2πnz/k Ψk (z)dz+
0≤h<k≤N
k
z′hk
′′
Zzhk
X iωhk −2πinh/k ′
h i 2
e Ψk (z) f (e2πi k + z ) − 1 e2πnz/k dz
o≤h<k≤N
k5/2
z′hk
where the second term compensates for the mistake committed on taking 158
f (x) = 1. The trick will be now to use the first term as the main term and
17. Lecture 121
to use an estimate for the small contribution from the second term. We have
∗
now to appraise this. Writing Ihk and Ihk for the two integrals, we have
X′ iωhk X′ iωhk
p(n) = 5/2
e−2πinh/k Ihk + 5/2
e−2πinh/k Ihk
∗
o≤h<k≤N
k o≤h<k≤N
k
∗
Here we stop for a moment and consider only Ihk and see what great ad-
vantage we got from our special path.
This is the arc of the circle |z − 12 | = 21 from z′hk to z′′hk described clockwise.
∞
P ∗
Since f (x)−1 = p(ν)xν , the integrand in Ihk is regular, and so for integration
m=1
we can just as well run across, along the chord from z′hk to z′′hk . Let us see what
happens on the chord. We have
∞
√ R π Rz π2 (− 12 +2n) X (2πi h′ 2π
− )ν
= ze 12z e k × p(ν)e k z
ν=1
∞
√ X 1 π π 1
≤ | z| p(ν)e−R z (2πν− 12 ) e k2 (− 12 +2n)Rz
ν=1
1 1 x
R =R = 2 ,
z x + iy x + y2
17. Lecture 122
k2 kk1
z′hk = 2 2
+i 2 ,
k1 + k k1 + k 2
k2 kk2
z′′hk = +i
k22 + k2 k22 + k2
k4 + k2 k12 k2
|z′hk |2 = =
(k12 + k2 )2 k2 + k12
Now we wish to appraise this in a suitable way.
2k2
|z′hk |2 ≤
N2
√
′ 2·k
or |zhk | ≤ ;
N
√
2·k
also |z′′hk | ≤
N
So the inequality becomes 161
√ 1/2 X∞
f (e2πih′ /k−2π/z ) − 1 Ψk (z)e2πnz/k2 ≤ 4 2 · k
1 2
p(ν)e(2πν−π/12) eπ(− 2 +2/n1)/k
N 1/2 ν=1
k1/2
≤ Ce2π|n|
N 1/2
∞
P
where C is independent of ν, since the series p(ν)e−(2πν−π/12) is convergent.
ν=1√
Since the length of the chord of integration < 2 2 · k/(N + 1)m, we have
∗ k3/2
Ihk < C1 e2π|n| 3/2
N
17. Lecture 123
Then
X′ iωhk −2πnh/k ∗ X′ 1
5/2
e Ihk ≤ C1 e2π|n| 3/2
0≤h<k≤N k 0≤h<k≤N
kN
X 1
≤ C1 e2π|n| ,
0<k≤N
N 3/2
Since the summation is over h < k with (h, k) = 1, so that there are only 162
ϕ(k) terms and this is ≤ k. So the last expression is
C1 e2π|n| N −1/2
The formula that we had last time looked like this: 163
X′
p(n) = iωhk e−2πinh/k k−5/2 Ihk + RN ,
o≤h<k≤N
This is an improper integral with both ends going to zero. That it exists is
′
Rzhk
clear, for what do we have to compensate for that? we have to subtract · · ·
0
124
18. Lecture 125
R0
and · · · , and we prove that these indeed contribute very little. What is after 164
z′′hk
all Ψk (z)?
√ 12π ( 13 − z2 )
Ψk (z) = ze k
so that the integrand is bounded. Hence the limit exists. This is indeed very
astonishing, for Ψ has an essential singularity at the origin; but on the circle it
does not do any harm. Near the origin there are value which are as big as we
want, but we can approach the origin in a suitable way. This is the advantage
of this contour. The earlier treatment was very complicated.
R0
The same estimate holds good for · · · . Hence introducing.
z′′hk
Now everything is under our control. N appeared previously tacitly in z′hk , 166
because z′hk depends on the Farey arc. Now N appears in only two places. So
p(n) is the limit of the sum which we write symbolically as
∞
X Z
2
p(n) = i Ak (n)k−5/2 Ψk (z)e2πnz/k dz
k=1
K (−)
18. Lecture 126
(n R 0, integral, and p(n) = 0 for n < 0). So we have an exact infinite series
for p(n).
∞ 1+i∞
Z
X
−5/2 π 2 2πn dω
p(n) = − Ak (n)k ω−1/2 e 12 (ω−1/k ω)
e k2 ω ·
k=1
ω2
1−i∞
∞ 1+i∞
Z
1X π π
= Ak (n)k−5/12 ω−5/2 e 2 + 12k2 ω (24n−1) dω
i k=1
1−i∞
One more step is advisable to get a little closer to the customary notation. 167
πω
We then get traditional integrals known in literature. Put = s,
12
π
12 +i∞
∞ Z
1 π 3/2 X π2
p(n) = Ak (n)k−5/2 s−5/2 e s+ 12k2s (24n − 1)ds
i 12 k=1 π
12 −i∞
One could look up Watson’s ‘Bessel Functions’ and write down this inte-
gral as a Bessel function. But since we need the series anyway we prefer to
compute it directly. So we have to investigate an integral of the type
c+i∞
Z
1 ν
L(ν) = s−ν−1 e s+ s ds
2πi
c−i∞
It does not matter what c > 0 is because it means only a shift to a parallel
line, and the integrand goes to zero for large imaginary part of s. For absolute
convergence it is enough to have a little more than s−1 . So take Rν > 0; in our
case ν = 3/2.
18. Lecture 127
For better understanding we take a specific loop. On the lower bank of the 169
negative real axis we proceed only up to −ǫ,
then go round a circle of radius ǫ in the positive sense and proceed thence
along the upper bank, the integrand now having acquired a new value-unless ν
is an integer. This we take as a standardised loop. We now prove that Lν (V ) is
actually differentiable and that the derivative can be obtained by differentiating
18. Lecture 128
under the integral sign. For this we take {Lν (V + h) − Lν (V )} /h and compare
it with what we could foresee to be L′ν (V ) and show that the difference goes to
zero as h → 0.
(0+)
Z
Lν (V + h) − Lν (V ) 1 ν
− s−ν−2 e s+ s ds
h 2πi
−∞
Z(0+) v+h ν ν
1 s −e s
−ν−1 s e
es
= s e − ds
2πi h
s
−∞
(0+)
Z h
1 ν es − 1 1
s−ν−1 e s+ s
= − ds
2πi h
s
−∞
Now 170
2
h h h
es − 1 1 s + s2 ,s! + · · · 1
− = −
h s ( h s )
1 h
=h 2 + + ···
s · 2! s3 · 3!
Lv (ν+h)−Lv (ν)
So the limit lim h exists and Lν (ν) is differentiated uniformly in a
h→0
circle of any size. Since the differential integral is of the same shape we can
differentiate under the integral as often as we please.
Lecture 19
∞ Z(0+)
1 π 3/2 X 1 π 2
p(n) = Ak (n)k−5/2 s−5/2 e s+ s ( 12k ) (24n−1) ds
i 12 k=1 −∞
This integral is again of the same sort as before; so we can repeat differenti-
ation under the integral sign. Clearly then Lv (ν) is an entire function of ν · Lv (ν)
has the expansion in a Taylor series:
∞
X L(r)
v (0) r
Lv (ν) = ν
r=0
r!
L(r)
v (ν) can be foreseen and is clearly
(0+)
Z
1 v
s−v−1−r e s+ s ds
2πi
−∞
129
19. Lecture 130
So 172
(0+)
∞ Z
X νr 1
Lv (ν) = s−v−1−r e s ds
r=0
r! 2πi
−∞
and what we have proved therefore is that we can interchange the integration
and summation. We have
L′v (ν) = Lv+1 (ν).
Having this under control we can put it back into our formula and get a
final statement about p(n).
π 3/2 X∞ π 2 !
p(n) = 2π Ak (n)k−5/2 L3/2 (24n − 1)
12 k=1
12k
This is not yet the classical formula of Hardy and Ramanujan. One trick 173
one adopts is to replace the index. Remembering that L′v (ν) = Lv+1 (ν), we have
π 2 ! π 2 !
′
L3/2 (24n − 1) = L1/2 (24n − 1)
12k 12k
19. Lecture 131
!
6k2 d π 2
= 2 L1/2 (24n − 1)
π dn 12k
Let us write the formula for further preparation closer to the Hardy Ra-
manujan notation:
π 1/2 X∞ π 2 !
−1/2 d
p(n) = Ak (n)k L1/2 (24n − 1)
12 k=1
dn 12k
1
Now it turns out that the L-functions for the subscript 2 are elementary
functions. We introduce the classical Bessel function
∞
X (−)r (z/2)2r+v
Jv (z) =
r=0
r!Γ(v + r + 1)
and the hyperbolic Bessel function (or the ‘Bessel function with imaginary
argument’)
∞
X (z/2)2r+v
Iv (z) =
r=0
r!Γ(v + r + 1)
How do they belong together? We have 174
!
z2 z −v
Lv = Iv (z) ,
4 2
!
z2 z −v
Lv − = Jv (z) ,
4 2
connecting our function with the classical functions. In our case therefore we
could write in particular
π 2 ! π √ π √ −1/2
L1/2 (24n − 1) = I1/2 24n − 1 24n − 1
12k 6k 12k
This is always good, but we would come into trouble if we have 24n−1 ≤ 0.
It is better to make a case distinction; the above holds for n ≥ 1, and for n ≤ 0,
n = −m, we have
π 2 ! π 2 !
L1/2 (24n − 1) = L1/2 − (24m + 1)
12k 12k
π √ π √ −1/2
= J1/2 24m + 1 24m + 1
6k 12k
19. Lecture 132
So we have: n ≥ 1.
√
π 1/2 X∞ I1/2 π 24n − 1
d 6k
Ak (n)k−1/2 √
p(n) = 1/2
12 k=1
dn π
24n − 1
12k
n = −m ≤ 0 175
√
π 1/2 X∞ J1/2 π 24m + 1
d
p(n) = p(−m) = − Ak (−m)k−1/2 · √6k
1/2
12 dm π
k=1 12k 24m + 1
We are not yet quite satisfied. It is interesting to note that the last expression
is 1 for n = 0 and 0 for n < 0. We shall pursue this later.
We have now more or less standardised functions. We can even look up
tables and compute the Bessel function. However I1/2 and J1/2 are more ele-
mentary functions.
∞
X (−)r (z/2)2r+1/2
J1/2 (z) =
r=0 r!Γ(r + 23 )
∞ 1
X (−)r (z/2)2r+ 2
=
1 1 1 1
r=0 r! r + 2 r − 2 · · · 2 Γ 2
!1/2 X∞
2 (−)r z2r+1
=
πz r=0
(2r + 1)!
!1/2
2
= sin z.
πz
∞ π
√
1 X −1/2 d
sinh 6k 24n − 1
p(n) = √ Ak (n)k √
3 k=1 dn π 24n − 1
6k
r r r
π √ π 2 1 c 1 2
or with 24n − 1 = (n − ) = n − ,C = π ,
6k k 3 24 k 24 3
q
c 1
d sinh k n − 24
∞
1 X
p(n) = √ Ak (n)k1/2 q
π 2 k=1 dn
n− 124
n = −m ≤ 0
q
∞ sin c m + 1
1 X 1 d k 24
p(n) = p(−m) = − √ Ak (−m)k 2 q
π 2 k=1 dm
m+ 1 24
This is the final shape of our formula -a convergent series for p(n). 177
The formula can be used for independent computation of p(n). The terms
become small. It is of interest to find what one gets if one breaks the series off,
say at k = N
N
π5/2 X
p(n) = √ · · · + RN
12 3 k=1
Let us appraise RN · |Ak (n)| ≤ k, because there are only ϕ(k) roots of unity.
We want an estimate for L3/2 . For n ≥ 1,
2 r
! X ∞ π
π 2
6k2
n
L3/2 (24n − 1) ≤
12k r=0 r!Γ r + 2
5
2 r
∞ π
X 6(N+1) 2 n
≤
1 1 3
r=0 r!Γ 2 · 2 · r+ 2
2 r
∞ π 2r+1
1 X 6(N+1)2 n · 2
(sincek > N in RN ) = √
π r=0 (2r + 1)!(r + 32 )
r
∞ 2π2
2 X 3(N+1)2 n
≤ √
π r=0 (2r + 1)!
19. Lecture 134
r
∞ 2π2
2 2 X 3(N+1)2 n
≤ · √
3 π r=0 (2r)!
4 π
√ 3n
< √ e N+1 3
3 π
π2 π
√ 2n X ∞
1
∴ |RN | ≤ √ e N+1 3 3/2
9 3 k=N+1
k
π2 π
√ 2n Z∞ dk
≤ √ e N+1 3
9 3 k3/2
N
2π2 π √ 1
∴ |RN | < √ e N+1 2n3 1/2
9 3 N
And this is what Hardy and Ramanujan did. Their work still looks different.
They did not have infinite series. They had replaced the hyperbolic sine by the
most important part of it, the exponential.
q The series
converges in our case
c 1
since sinh x ∼ x as x → 0, so that sinh k n− 24 behaves roughly like kc . On
differentiation we have k12 so that along with k in the numerator we get k−3/2
1/2
So
!!
1 3/4 1
|R∗ | = O n + √ n5/4
n n
1
−4
=O n
The constants in the O-term were not known at that time so that numerical 180
computation was difficult. If the series was broken off at some other place the
terms might have increased. Hardy and Ramanujan with good instinct broke
off at the right place.
We shall next resume our function-theoretic discussion and cast a glance at
the generating function for p(n) about which we know a good deal more now.
Lecture 20
We found a closed expression for p(n); we shall now look back at the generat- 181
ing function and get some interesting results.
∞
1 X
f (x) = ∞ = p(n)xn ,
Q
(1 − xn ) n=0
n=1
and we know p(n). p(n) in its simplest form before reduction to the traditional
Bessel functions is given by
π 3/2 X∞ π 2 !
p(n) = 2π Ak (n)k−5/2 L3/2 (24n − 1) ,
12 k=1
12k
!! ∞
X π2 (n − 1 )r
π2 1
where L3/2 n − = 6k
24
6k2 24 5
r=0 r!Γ 2 + r
We wish first to give an appraisal of L and show that the series for p(n)
converges absolutely. The series is
∞ ∞ !
π 3/2 X X π2
f (x) = 2π xn Ak (n)k−5/2 L3/2 (n − α) ,
12 n=0 k=1
6k2
1
where we write 24 = α for abbreviation - it will be useful for some other
purposes also to have a symbol there instead of a number.
We make only a crude estimate.
! X 2 r
∞ π
π2 6n
L3/2 2
(n − α) ≤
6k
r=0 r!Γ 5 + r
2
136
20. Lecture 137
∞ π2 r
6 n
X
=
1 1
r=0 r!Γ 2 · · 23 · · ·
2
3
2 +r
r
∞ 2π
22 X 3 πn
√
π r=0 (2r + 1)!(3 + 2r)
∞ √ r
X (C n)2r 2
≤4 ,C = π ,
r=0
(2r)! 3
∞ √
X (C n)ρ
≤4
ρ=0
ρ!
√
= 4eC n
∞ √ ∞
X X 1
constant x |x|n eC n
3/2
n=1 k=1
k
where the middle sum is a finite sum. This is all good for |x| < 1. Now call 183
∞ !
X π2
Φk (z) = L3/2 (n − α) zn
n=0
6k2
We have now a completely new form for our function. It is of great in-
terest to consider Φk (z) for its own sake; it is a power series (|z| < 1) and the
20. Lecture 138
∞
P
where ϕr (z) is the power series (n−α)r zn . Actually it turns out to be a rational
n=0
function. Φk (z) can be extended over the whole plane.
∞
X 1
ϕr (z) = zn = .
n=0
1−z
20. Lecture 139
∞
X
ϕ′r (z) = n(n − α)r zn−1 ,
n=0
X∞
zϕ′r (z) = n(n − α)r zn ,
n=0
∞
X
αϕr (z) = α(n − α)r zn ;
n=0
so,
∞
X
zϕ′r (z) − αϕr (z) = (n − α)r+1 zn = ϕr+1 (z)
n=0
This says that we con derive ϕr+1 (z) from ϕr (z) by rational processes and
differentiation. This will introduce no new pole; the old pole z = 1 (pole
for ϕ◦ (z)) will be enhanced. So ϕr (z) is rational. Let us express the function
1
a little more explicitly in terms of the new variable u = z−1 or 1u + 1 = z.
Introduce (−)r+1 ϕr (z) = (−)r+1 ϕr (1 + u) = ψr (u), say he last equation which
was a recursion formula now becomes
!
1
(−)r+2 ψr+1 (u) = + 1 (−)r u2 ψ′r (u) − α(−)r+1 ψr (u)
u
! ! !
1 1 1 1
because ψ′r (u) = (−)r+1 ϕ′r 1 + − 2 = (−)r ϕ′r 1 +
u u u u2
′
∴ ψr+1 (u) = u(u + 1)ψr (u) + αψr (u)
ψ0 (u) = u
ψ1 (u) = u(u + 1) + αu = (1 + α)u + u2
ψ2 (u) = u(u + 1)(2u + 1 + α) + α(1 + α)u + αu2
= (1 + α)2 u + (2α + 3)u3 + 2u3
ψr (u) is a polynomial of degree r+1 without the constant term. The coefficients 186
are a little complicated. If we make a few more trials we get by induction the
following:
20. Lecture 140
Theorem.
r
X
ψr (u) = ∆ j (α + 1)r u j+1 ,
j=0
∆ f (x) = f (x + 1) − f (x),
∆2 f (x) = ∆∆ f (x) = ∆ f (x + 1) − ∆ f (x)
= f (x + 2) − 2 f (x + 1) + f (x)
How does the formula for ψr fit? For induction one has to make sure that 187
the start is good.
ψ0 (u) = (α + 1)◦ u = u
ψ1 (u) = (α + 1)′ u′ + ∆(α + 1)′ u2 = (α + 1)u + u2
ψ2 (u) = (α + 1)2 u′ + ∆(α + 1)2 u2 + ∆2 (α + 1)2 u3
= (α + 1)2 u + (α + 2)2 − (α + 1)2 u2 + 2u3
= (α + 1)2 u + (2α + 3)u2 + 2u3
To show that the last factor is ∆ j (α + 1)r+1 , we need a side remark. Intro- 188
duce a theorem corresponding to Leibnitz’s theorem on the differentiation of a
product. We have
and observe that f being linear permits only 0th and 1 st differences;
! !
k k
∆k (α + 1)r = ∆k−1 (α + 1)r + (α + k + 1)∆k (α + 1)r
k−1 k
= k∆k−1 (α + 1)r + (α + k + 1)∆k (α + 1)r
Xr
∴ ψr (u) = ∆ j (α + 1)r u j+1
j=0
20. Lecture 142
We have rewritten the generating function f (x) as a sum consisting of certain 190
functions which we called Φk (x):
∞
π 23 X X′ h
f (x) = 2π k−5/2 ωhk Φk xe−2πi k
12 k=1 h mod k
∞ !
X π2
where Φk (z) = L3/2 2
(n − α) zn ,
n=0
6k
1
with α = 24 . Φk (z) could also be written as
2r
∞ π
√
X k 6
Φk (z) = ϕr (z′′ )
5
r=0 r!Γ 2 +r
143
21. Lecture 144
and in general
∆k f (x) = f k (ξ), x < ξ < x + k.
This was to be expected. Take |1 − z| ≥ δ, 0 < δ < 1 so that z is not too
1
close to 1. > 1 and 0 < α < 1
δ
r
X 1
|ϕr (z| ≤ r(r − 1) · · · (r − j + 1)(1 + α + j)r− j ·
j=0
δ j+1
r
X (α + 1 + r)r
<
j=0
δ j+1
(α + 1 + r)r
< (r + 1)
δr+1
(α + 1 + r)r+1
<
δr+1
Originally we know that the formula for f (x) was good for |x| < 1. From
this point on we give a new meaning to ϕr (z) for all z , 1.
This is a new step. We prove that the series for Φk (z) is convergent not 192
merely for |z| < 1 but also elsewhere. The sum in Φk (z) is majorised by
2 r
∞ π
1X 6 (α + 1 + r)r+1
·
δ r=0 r!Γ 5 + r δr
2
π 3/2 X∞ X′
5
f (x) = −2π k− 2 ωhk
12 k=1 h mod k
2 r
X∞ − π6k ∞
X
j 1
∆α (α + 1)r · −2πih/k − 1) j+1
5
r=0 r!Γ 2 + r j=0
(xe
π 3/2 X∞ X′
5
f (x) = −2π k− 2 ωhk
12 k=1 h mod k
2 r
∞ h ∞ π
X e−2πi k ( j+1) X j − 6k 2
r
2πih/k j+1
∆α (α + 1)
j=0
(x − e ) r= j r!Γ 5 + r
2
P P∞
However, if we replaced ∞ r= j by r=0 it would not to any harm because the
summation is applied to a polynomial of degree r and the order of the difference
is one more than the power. We can therefore write, taking ∆ outside,
21. Lecture 146
π 3/2 X∞ X′
5
f (x) = −2π k− 2 ωhk
12 k=1 h mod k
∞ ∞ π2 r
X e2πi hk ( j + 1) X − 6k 2
∆j
2πih/k ) j+1 α
(α + 1)r
j=0
(x − e r=0 r!Γ 52 + r
∞ ∞ h !
π −5/2 X 5
X′ X e2πi k ℓ ℓ−1 π2
= −2π k2 ωhk ∆ L 3/2 − (α + 1)
12 k=1 h mod k ℓ=1
(x − e2πih/k )ℓ α 6k2
∞ ∞ !
π 3/2 X X π2
f (x) = −2π x−m Ak (−m)k−5/2 L3/2 (−m − α)
12 m=1 k=1
6k2
And here is a surprise which could not be foreseen! By its very meaning
p(−m) = 0. So
f (x) ≡ 0
outside the unit circle. This was first conjectured by myself and proved by
H.Petersson by a completely different method. Such expressions occur in the
theory of modular forms. Petersson got the outside function first and then the
inner one, contrary to what we did.
21. Lecture 148
The function is represented by a series inside the circle, and it is zero out-
side, with the circle being a natural boundary. There exist simpler examples of
this type of behaviour. Consider the partial sums:
x 1
1+ =
1−x 1− x
x x2 1 x2 1
1+ + = + =
1 − x (1 − x)(1 − x2 ) 1 − x (1 − x)(1 − x2 ) (1 − x)(1 − x2 )
x x2 x3
1+ + + + · · · to n + 1 terms
1 − x (1 − x)(1 − x2 ) (1 − x)(1 − x2 )(1 − x3 )
1
=
(1 − x)(1 − x2 ) · · · (1 − xn )
1
For |x| < 1, the partial sum converges to ∞ . For |x| > 1 also 198
Q
(1 − xm )
m=1
it has a limit; the powers of x far outpace 1 and so the denominator tends to
infinity and the limit is zero. The Euler series here is something just like our
complicated function. Actually the two are the same. For suppose we take the
1
partial sum and break it into partial fractions. We
(1 − x)(1 − x2 ) · · · (1 − xn )
get the roots of unity in the denominator, so that we have a decomposition
X Bh,k,l,n
h ℓ
x − e2πi k
k ≤ n and ℓ not too high. For a higher n we get a finer expression into partial
fractions. Let us face one of these, keeping h, k, ℓ fixed:
Bh,k,l,n
h
ℓ
x − e2πi k
B0,1,1,n
is itself very complicated. Let us evaluate the principal formula for 199
x−1
f (x) and pick out the terms corresponding to h = 0, k = l, ℓ = √1.
6 12 3 1
L3/2 is just the sine function and terns out to be − − . Since 1−x =
25 75π
1
− x−1 , −1 is the first approximation to B0,1,1,n. If we take the partial fraction
decomposition for
1 1 ·· ·· ··
, = + + ,
(1 − x)(1 − x2 ) (1 − x)(1 − x2 ) (x − 1)2 (x − 1) (1 + x)
the numerator of the second term would give the second approximation. If in-
deed these successive approximations converge to B0,1,1,n we could get a whole
new approach to the theory of partitions. We could start with the Euler series
and go to the partition function.
We are now more prepared to go into the structure of ωhk . We shall study
next time the arithmetical sum Ak (n) and the discovery of A.Selberg. We shall
then go back again to the η-function.
Lecture 22
We shall speak about the important sum Ak (n) which appeared in the formula 200
for p(n), defined as X′
Ak (n) = ωhk e−2πinh/k .
h mod k
Here, as we know,
1
f (x) = Q∞
m=1 (1 − xm )
Y∞
and as η(τ) = eπiτ/12 (1 − e2πimτ ),
m=1
f (e2πiτ ) = eπiτ/12 (η(τ))−1
150
22. Lecture 151
shall not use the product formula for η(τ), but the infinite series from the pen-
tagonal numbers theorem. This was carried out at my suggestion by W.Fischer
(Pacific Journal of Mathematics; vol. 1). However we shall not copy him. We
shall make it shorter and dismiss for our purpose all the long and complicated
discussions of Gaussian sums
k
X 2
G(h, k) = e2πiν h/k
v=1
We can therefore get what we are after if we specify the formula by these
particular values. ! !
h′ + iz √ h + iz
η = ǫ zη
k k
√
with the principal value for z. We wish to determine ǫ defined by this. We
shall expand both sides and compare the results. For expansion we do not use
the infinite product but the pentagonal numbers formula.
∞
X
η(τ) = eπiτ/12 (−)λ e2πiτλ(3λ−1)/2
λ=−∞
∞
X πiτ 2
−12λ)
= (−)λ e 12 (1+36λ
λ=−∞
22. Lecture 152
∞
X 2
= (−)λ e3πiτ(λ−1/6)
λ=−∞
Most determinations of η(τ) make use of the infinite product formula; the
infinite series is simpler here
! X ∞
h + iz h+iz 2
η = (−)λ e3πi k (λ−1/6)
k λ=−∞
In order to get the root of unity a little more clearly exhibited, we replace λ 203
mod 2k.
λ = 2kq + j, j = 0, 1, . . . , 2k − 1 and q runs from −∞ to ∞. So
! X ∞ 2k−1
h + iz X h 1 2 z 1 2
η = (−) j e3πi k (2kq+ j− 6 ) e−3π k (2kq+ j− 6 )
k q=−∞ j=0
∞
X 2
V3 (ν/τ) = eπiτn e2πinν ,
n=−∞
22. Lecture 153
which is the formula quoted. We now apply this deep theorem and get some-
j − 1/6
thing completely new. Putting t = 12zk and α = ,
k
! 2k−1 ∞
h + iz X 1 2 1 X πm2
(−) j e2πi k ( j− 6 ) √ e− 12zk e k ( j− 6 )
h πim 1
η =
k j=0 12kz m=−∞
second teaches us that by no means do all sequences appear in the exponent. 206
Only m2 = (6λ − 1)2 can occur. There is no constant term in the second expres-
sion. So m has the form |6λ − 1| = 6λ ± 1, λ > 0. Make the comparison; the
coefficients are identical. They are almost always zero. In particular T (0) = 0.
T (m) for m other than ±1 (mod 6) also vanish. So we have the following iden-
tification.
1 πih′
√ (T (6λ − 1) + T (−6λ + 1)) = ǫ −1 (−)λ e 12k (6λ − 1)2
3k
Realise that we have acknowledged here that a transformation formula ex-
ists. The root of unity ǫ is independent of λ. This we can assume but W.
Ruscher does not. Take in particular λ = 0. Then we have for m = ±1,
1 πih′
√ (T (−) + T (1)) = ǫ −1 e 12k
3k
This is proved by Fischer by using Gaussian sums. Therefore
′
− πih 2k−1 2k−1
X 3h 1 2 6 j−1
e 12k
X 3h i 2 6 j−1
πi j+ k ( j− 6 ) + 6k
ǫ −1 = √ πi( j+ k ( j− 6 ) )− 6k
e + e
3k j=0
j=0
The sum appears complicated but will collapse nicely; however compli-
cated it should be a root of unity. In Ak (n) the sums are summed over h and for
that purpose we shall not need to compute the sums explicitly.
Lecture 23
1 πi(h−h′ ) πi
X πi
2
ǫ −1 = √ e 12k e− 6k e k 3h j j (k−h+1)
2 3k j mod 2k
1 πi(h−h′ ) πi
X πi 2
+ √ e 12k e 6k e k (3h j + j(k−h−1))
2 3k j mod 2k
155
23. Lecture 156
πi ′
∴ ǫ = ωhk e 12k (h −h)
π ′
or ωhk = ǫe− 12k (h −h)
208
In the first formula we have obtained an expression for ǫ −1 . However, we
could make a detour and act ǫ directly instead of ǫ −1 . Even otherwise this could
be fixed up, for after all it is a root of unity. We have ǫ ǭ = 1 or ǫ = ǭ −1 . So
consistently changing the sign in the exponents, we have
πi ′ 1 πi
X πi 2
ωhk = ǭ −1 e 12k (h−h ) = √ e 6k e− k (3h j + j(k−h+1))
2 3k j mod 2k
1 πi
X πi 2
+ √ e− 6k e− k (3h j + j(k−h−1))
2 3k j mod 2k
We now have the ωhk that we need. But the ωhk are only of passing interest;
we put them back into Ak (n);
X′
Ak (n) = ωhk e−2πinh/k
h mod k
This formula has one unpleasant feature, viz. (h, k) = 1. But this would not
do any harm. We can use a lemma from an unpublished paper by Whiteman
which status that if (h, k) = d > 1, then
X πi 2
e− k (3h j + j(k−h±1)) = 0
j mod 2k
For proving Whiteman status put h = dh∗ , k = dk∗ and j = 2k∗ l + r, 209
0 ≤ 1 ≤ d − 1, 0 ≤ r ≤ 2k∗ − 1. Then
∗
X d−1 2k
X X −1
− πik (3h j2 + j(k−h±1)) πi ∗ ∗
1+r)2 +(2k∗ ℓ+r)(dk∗ −dh∗ ±1)))
e = e− dk∗ (3dh (2k
j mod 2k ℓ=0 r=0
∗
2k
X −1 d−1
X
πi 2
= e− k (3hr +r(k−h±1))
e∓2πiℓ/d ,
r=0 ℓ=0
and the inner sum = 0 because it is a full sum of roots of unity and d , 1.
This simplifies the matter considerably. We can now write
1 πi
X X πi 2 h
Ak (n) = √ e 6k e− k (3h j + j(k−h+1)) e−2πin k
2 3k h mod k j mod 2k
23. Lecture 157
1 πi
X X πi 2 h
+ √ e− 6k e− k (3h j + j(k−h−1)) e−2πin k
2 3k h mod k j mod 2k
1 πi
X πi
X 2πi j(3 j−1)
Ak (n) = √ e 6k e− k (k+1) j e− k (n+ k )h
2 3k j mod 2k h mod k
1 πi
X πi
X 2πi j(3 j−1)
+ √ e− 6k e− k (k−1) j e− k (n+ 2 )h
2 3k j mod 2k h mod k
The inner sum is equal to the sum of the kth roots of unity, which is 0 or k,
k if all the summands are separately one, i.e., if
j(3 j − 1)
n+ ≡0 (mod k)
2
Hence
r r
1 k πi X
j − πik j 1 k − πi X πi j
Ak (n) = e 6k (−) e + e 6k (−) j e k
2 3 j mod 2k
2 3 j mod 2k
j(3 j−1) j(3 j−1)
2 ≡−n
(mod k) 2 ≡−n
(mod k)
In the summation here we first take all j′ s modulo 2k (this is the first sieving
out), and then retain only those j which satisfy the second condition modulo k
(this is the second sieving out). Combining the terms,
r
1 k X n πi πi o
Ak (n) = (−) j e− 6k (6 j−1) + e 6k (6 j−1)
2 3 j mod 2k
j(3 j−1)
2 ≡−n (mod k)
r
k X π(6 j − 1)
= (−) j cos
3 j mod 2k
6k
j(3 j−1)
2 ≡−n(mod k)
211
This formula is due to A.Selberg. It is remarkable how simple it is. We shall
change it a little, so that it could be easily computed. We shall show that the
Ak (n) have a certain multiplicative property, so that they can be broken up into
prime parts which can be computed separately. Let us rewrite the summation
condition in the following way.
In both terms the range of summation is j mod 2k and there is the further
condition which restricts j. So
r
1 k X πi
Ak (n) = (−) j e− 6k (6 j±1)
2 3 j mod 2k
(6 j±1)2 ≡ν (mod 24k)
And one final touch. The ranges for ℓ in the two conditions are modulo 12k 213
and modulo 24k. Make these ranges the same. Then
r
1 k X
(−){ 6 } e 6k
ℓ πiℓ
Ak (n) =
4 3 ℓ mod 24k
ℓ2 ≡ν (mod 24k)
23. Lecture 159
We prefer the formula in this form which is much handler. We shall utilise
this to get the multiplicative property of Ak (n).
Lecture 24
We derived Selberg’s formula, and it looked in our transformation like this: 214
r
1 k X
(−){ 6 } e 6k ,
ℓ πiℓ
Ak (n) =
4 3 2
l ≡γ (mod 24k)
where ν = 1 − 24n, or ν ≡ 1 (mod 24). We write this Bk (ν); this is defined for
ν ≡ 1 (mod 24), and we had tacitly (ℓ, 6) = 1. We make an important remark
about the symbol (−){ 6 } . This repeats itself for ℓmodulo12. The values are
ℓ
ℓ= 1 3 7 11
(−){ } = 1
ℓ
6 −1 −1 1
{ ℓ
}
But (−) 6 can be expressed in terms of the Legendre symbol:
! !
{ ℓ
} ℓ −1
(−) = 6
3 ℓ
ℓ−1
when (ℓ, 6) = 1. We can test this, noticing that −1 ℓ = (−1) 2 . Since 1, 7
are quadratic residues and 5, 11 quadratic non-residues modulo 3, we have for
ℓ
ℓ = 1, 5, 7, 11, (−){ 6 } = 1, −1, −1, 1 respectively; this agrees with the previous
ℓ
list. It is sometimes simpler to write (−){ 6 } in this way, though it is an after-
thought. It shows the periodicity.
Let us repeat the formula: 215
r ! !
1 k X ℓ −1 πiℓ
Bk (ν) = e 6k
4 3 2 3 ℓ
ℓ ≡ν (mod 24k)
This depends upon how k behaves with respect to 24. It has to be done
separately for 2, 3, 4, 6. For this introduce d = (24, k3). We have
160
24. Lecture 161
d = 1 if (k, 24) = 1,
3 if 3 | 4, k odd,
8 if k is even and 3 ∤ k
24 if 6 | k.
Let us introduce the complementary divisor e, de = 24. So e = 24, 8, 3 or
1. (d, e) = 1. Also (c, k) = 1.
All this is a preparation for our purpose. The congruence ℓ2 ≡ ν (mod 24k)
can be re-written separately as two congruences: ℓ2 ≡ ν (mod dk), ℓ2 ≡ ν
(mod e).
The latter is always fulfilled if (ℓ, 6) = 1. Now break the condition into two
subcases. Let r be a solution of the congruence
(er)2 ≡ ν (mod dk);
then we can write ℓ = er + dk j, where j runs modulo e and moreover ( j, e) = 1.
To different pairs modulo dk and e respectively belong different ℓ modulo 24k.
Bk (ν) can then be written as
r
1 k X X er + dk j ! −1
!
πi
Bk (ν) = e 6k (er+dk j)
4 3 2 j mod e
3 er + dk j
(er) ≡ν (mod dk)
( j,e)=1
where ! !
X′ er + dh j −1 πid j
S k (r) = e 6k
j mod e
3 er + dk j
We compute this now in the four different cases implied in the possibilities
d = 1, 3, 8, 24.
Case 1. d = 1, e = 24
! !
X′ k j −1 πi j
S k (r) = e6
j mod 24
3 kj
! !
k −1 X′ j j−1 πi j
= (−) 2 e 6
3 k j mod 24 3
24. Lecture 162
There are eight summands, but effectively only four, because they can be
folded together.
! !
k −1 X′ j π−1
S k (r) = 2 (−) 2 eπi j
3 k j mod 12 3
! !
h −1 n πi 5πi 7πi 11πi o
=2 e6 −e 6 −e 6 +e 6
3 k
ℓ
(We replaced the nice symbol (−){ 6 } by the Legendre symbol because we
did not know a factorisation law for the former. So we make use of one special
character that we know).
! ! !
k −1 π 5π
S k (r) = 4 cos − cos
3 k 6 6
! !
k −1 √
=4 3
3 k
k−1
and since 3k 3k = (−) 2 ·1 = −1 k , this gives gives 217
!
√ 3
S k (r) = 4 3
k
Case 2. d = 3, e = 8.
! !
X′ 8r −1 πi2 j
S k (r) = e
j mod 8
3 3k j
−r −1 ! X′ −1 ! πi j
= e2
3 3k j mod 8 j
r −1 ! X′ −1 ! πi j
=2 e2
3 k j mod 4 j
r −1 !
=2 (i + i)
3 k
r −1 !
= 4i .
3 k
Case 3. d = 8, e = 3.
! !
X′ 8k j −1 4πi j
S k (r) = e 3
j mod 3
3 3r
24. Lecture 163
! ! X
k −1 ′ j 4πi
= e3
3 r
j mod 3
3
! !
k −1 4πi 8πi
= e 3 −e 3
3 r
! !
k −1 2π
= −2i sin
3 r 3
! !
1 √ k −1
= 3
i 3 r
218
Case 4. d = 24, e = 1.
! ! !
k −1 3
S k (r) = =
3 r r
Case 2. r !
k −1 X r 4πir
Bk (ν) = i e 3k
3 k 3
(8k)2 ≡ν (mod 3k)
The i should not bother us because r and −r are solutions together, so they
combine to give a real number.
r ! r
k −1 X 4πr
Bk (ν) = − sin
3 k 2
3 3k
(8r) ≡v (mod 3k)
Case 3.
! !
1 √ k X −1 πir
Bk (ν) = k e 3k
4i 3 r
(3k)2 ≡ν (mod 8k)
! !
1√ k X −1 πr
= k sin
4 3 2
r 2k
(3r) ≡ν (mod 8k)
219
24. Lecture 164
Case 4. r !
1 k X 3 πir
Bk (ν) = e 6k
4 3 r
r2 ≡ν (mod 24k)
d=3 !r
−1 k X r
Bk (ν) = 2i e4πir/3k
k 3 3
(8r)2 ≡ν (mod 3k)
d=8 ! !
1 k √ X −1 πir/2k
Bk (ν) = k e
4i 3 r
(8r)2 ≡ν (mod 8k)
d = 24
There is nothing new; we get the old formula back.
We wish first to anticipate what we shall use later and get An (n) for prime
powers which will be the ultimate elements. Again we have to discuss several
cases.
First take k = pλ , p a prime exceeding 3. Then, by case 1 above (since 221
(24, k3 ) = 1),
!λ
3 X λ
Bk (ν) = pλ/2 e4πir/p
p 2 λ
(24r) ≡ν (mod p )
165
25. Lecture 166
Look into the condition of summation. It is quite clear that this implies
(24r)2 ≡ ν (mod p) i.e., ν is a quadratic residue modulo p. Hence
!
v
B p λ(ν) = 0 if = −1. (1)
p
This again involves two cases, λ even and λ odd. If λ is even, λ = 2µ and
the sum becomes
X pµ
µ
e4πi j/p
j=1
and this is 0, being a full sum of roots of unity. Hence in this case
Bk (ν) = 0 (4)
25. Lecture 167
r = pµ+1 · j, j = 0, 1, . . . , pµ − 1.
Again
Bk (ν) = 0 (8)
We now take up the case p = 3. This corresponds to p = 3. If k = p = 3λ , λ
224
λ−1
X r
λ+1
B3 λ(ν) = i(−)λ 3 2 e4πir/3
2 λ+1
3
(8r) ≡ (mod 3 )
25. Lecture 168
ν
ν ≡ 1 (mod 24) or ν ≡ 1 (mod 3). So 3 = 1. There are two solutions, r and
2
−r for the congruence (8r) ≡ ν (mod 3 λ+1
). Since −r
3 =− 3 ,
r
rλ−1
4πir 4πir
B3λ (ν) = i(−)λ 3 2 e 3λ+1 − e− 3λ+1
3
λ−1
λ+1 r 4πr
= 2(−) 3 2 sin λ+1 (9)
3 3
Finally, we take p = 2; then d is 8. Let k = 2λ . Then
!
1 λ λ/2
X −1 4πir/2λ+1
B2 λ(ν) = (−) 2 e
4i 2 λ+3
r
(3r) ≡ν (mod 2 )
ν ≡ 1 (mod 8) implies that (3r2 ) ≡ ν (mod 8) has four solutions, and these
solutions are inherited by the higher powers of the modulus. The solutions are
r ≡ 1, 3, 5, 7 (mod 8). In general the congruence x2 ≡ ν (mod 2µ ), µ ≥ 3 has
four solutions
±r + h2µ−1 , h = 0, 1
Then 225
o −1 !
1 n λ+1 λ+1 λ+1 λ+1
B2λ (ν) = (−)λ 2λ/2 e4πir/2 − e−4πir/2 + e4πir/2 − e−4πir/2
4i r
r−1
and since −1
r = (−)
2 ,
!
−1 4πr
B2λ (ν) = (−)λ eλ/2 sin λ+! (10)
r 2
1) d = 1 !
3 √ X
Bk (ν) = k e4πir/k
k
(24π)2 ≡ν (mod k)
2) d = 3
!r !
−1 k X −1 πir/2k
Bk (ν) = i e
k 3 r
(24r)2 ≡ν (mod 3k)
3) d = 8 ! !
1 k √ X −1 πir/2k
Bk (ν) = k e
4i 3 r
(3r)2 ≡ν (mod 8k)
(ii) None is prime to 6. But since (k1 , k2 ) = 1, 2/k1, 3/k1 . Under the circum-
stances prevailing these are the two mutually exclusive cases.
169
26. Lecture 170
Case 1. Utilise d = 1.
! r
3 p 1 k2 X
Bk1 (ν1 ) · Bk2 (ν2 ) = k1 · e4πir/k1
k1 4 3
(24r)2 ≡ν 1 (mod k1 )
! !
X ℓ−1 πiℓ/6k2
· e
3 ℓ
ℓ2 ≡ν2 (mod 24k2 )
!r ! !
1 3 k1 k2 XX πi
6k1 k2 (24k2 r+k1 l)
l −1
= e
4 k1 3 2
3 l
(24r) ≡ν1 (mod k1 )
ℓ2 ≡ν2 (mod 24k2 )
k1 and 24k2 are coprime moduli. If r runs modulo k1 and ℓ runs modulo 24k2 ,
24k2 r + k1 ℓ would then run modulo 24k1 k2 .
Write
24k2 r + k, ℓ ≡ t (mod 24k1 k2 )
Then
t2 = (24k2 + k1 ℓ)2 ≡ (24k2 r)2 (mod k1 )
≡ k22 ν1 (mod k1 ), since (24r)2 ≡ ν1 (mod k1 )
Similarly
t2 ≡ (k1 ℓ)2 (mod 24k2 )
≡ k12 ν2 (mod 24k2 ), since ℓ2 ≡ ν2 (mod 24k2 ).
So in order to get both conditions of summation, we need only choose 228
t2 ≡ ν (mod 24k1 k2 ); and this can be done by the Chinese remainder theorem.
So !r ! !
1 3 k X ℓ −1 πit/6k
Bk1 (ν1 )Bk2 (ν2 ) = , e
4 k 3 2 3 ℓ
t ≡ (mod 24k,k2 )
This already looks very much like the first formula though not quite. What
we have in mind is to compare it with
r t −1 !
1 k X
Bk (ν) = eπit/6k
4 3 2 3 t
t ≡ν (mod 24k)
So find out
t −1 ! 24k2 r + k1 ℓ
!
−1
!
=
3 t 3 24k2 r + k1 ℓ
26. Lecture 171
! !
k1 ℓ −1
=
3 k1 ℓ
! ! ! !
k1 −1 ℓ −1
=
3 k1 3 ℓ
! ! !
3 ℓ −1
= ,
k1 3 ℓ
by the reciprocity law. So the formulas agree: Bk1 (ν1 )Bk2 (ν2 ) = Bk (ν); and we
have settled the affair in this case by
Theorem 1. If k22 ν1 ≡ ν (mod k1 ) and k12 ν2 ≡ ν (mod 24k2 ), (k, 6) = 1, then 229
Since (k1 , k2 ) = 1, (8k1 , 3k2 ) = 1 and so 3k2 r + 8k1 s = t runs through a full
system of residues modulo 24k1 k2 . So
! !r !
1 k1 −1 k X −1 s πit/(6k1 k2 )
Bk1 (ν1 )Bk2 (ν2 ) = e
4 3 k2 3 2 r 3
t ≡ν (mod 24k1 k2 )
As before 230
t2 = (3k2 r + 8k1 s)2 ≡ (3k2 r)2 ≡ (3k2 r)2 ≡ k22 ν1 (mod 8k1 )
2 2
t = (8k1 s) ≡ k12 ν2 (mod 3k2 )
Now determine ν such that ν ≡ k22 ν1 (mod 8k1 ) and ν ≡ k12 ν2 (mod 3k2 ),
again by the Chinese remainder theorem. So t2 ≡ (mod 24k1 k2 ). Now
t −1 ! !
8k1 s −1
!
=
3 t 3 3k1 r
26. Lecture 172
! ! !
k1 −1 s −1
=
3 k2 3 r
(since 8 and −1 are quadratic non-residues modulo 3). So
r t −1 !
1 k X
Bk1 (ν1 )Bk2 (ν2 ) = eπit/6k
4 3 2 3 t
t ≡ν (mod 24k)
= Bk (ν)
Let us give an example of what this is good for. Calculate A10 (26). Since
we can reduce modulo 10, A10 (26) = A10 (6).
ν = 1 − 24n = −143.
A10 (26) = A10 (6) = B10 (−143) = B10 (−23)
= B5 (ν1 )B2(ν2 )
λ √
|B3λ (ν)| ≤ 3 2 2 3,
λ
|B pλ (ν)| ≤ 2p 2 , p > 3.
We see that although An (n) has ϕ(k) summands and in general all that one
knows is that ϕ(k) ≤ k − 1, because of strong mutual cancellations among the
1
roots of unity, the order is brought down to that of k 2 +ǫ . This reminds us of
other arithmetical sums like the Gaussian sums and the Kloosterman sums.
Lecture 27
We now give a proof of the transformation formula for η(τ). η(τ) we first in- 233
troduced by Dedekind in his commentary on a fragment on modular functions
by Riemann; it is natural in the theory of elliptic functions.
∞
Y
πiτ
η(τ) = e 12 (1 − e2πimτ )
m=1
aτ + b
We want to replace τ by τ′ = . Actually in the whole literature there
cτ + d
is no full account except in a paper by W.Fischer (Pacific Journal of Mathe-
matics, Vol. 1). We know what happens in the special cases − τ1 and τ + 1. We
get the explicit form in which the root of unity appears in the transformation
formula if we put together some things from the theory of modular functions.
There some discussion in Tannery-Molk; they write h(τ) instead of η(τ). (η(τ))3
is up to a factor V11 (o/τ). It turns out for quite other reasons that (η(τ))8 can
be discussed too; it has to do with the modular invariant J(τ). Dedekind did
something more than what is needed here. He studied log η(τ). For Im τ > 0,
η(τ) is a function in the interior of the unit circle (if we set x = e2πiτ ) free from
zeros and poles. So the logarithm has no branch points and is fully defined
without ambiguity.
∞
πiτ X
log η(τ) = + log(1 − e2πimτ )
12 m−1
(For purely imaginary τ, the logarithms on the right side are real).
The multiplicative root of unity now appear as something additive. This 234
is what Dedékind investigated. Recently (Mathematika, vol.1, 1954) Siegel
1
published a proof for the particular case − , using logarithms. Actually Siegel
τ
174
27. Lecture 175
proves much more than the functional equation for η(τ). He proves that
1 τ
log η(−τ−1 ) = log η(τ) + log
2 i
We shall extend his proof to the more general case. The interesting case
where a root of unity appears explicitly has not been dealt with by Siegel.
We write the general modular transformation in the form
h + iz ′ h′ + i/z
τ= ,τ = , hh′ ≡ −1 (mod k)
k k
We wish to prove that
! !
h′ + i/z h + iz 1
log η = log η + log z + πiC(h, k) (*)
k k 2
m = qk + µ; µ = 1, . . . , k; q = 0, 1, 2, . . . .
Then
! ∞ k ∞
h + iz πih πz X X X 1 2πiµ rh −2π(1k+µ) rz
log η = − − e k e k ,
k 12k 12k q=0 ν=1 r=1 r
∞ ∞ ∞
πih πz X X 1 2πiµ rh −2πµ rz X −2πqrz
− − e k e k e
12k 12k µ=1 r=1 r q=0
k ∞
πih πz X X 1 2πiµ rh e−2πµrz/k
= − − e k
12k 12k µ=1 r=1 r 1 − e−2πrz
27. Lecture 176
h′ +i/z
Substituting in (*), with similar expansion for η k , we have
k ∞ 2πνr
πih π XX 1 h′ e− kz
− − e2πiνr k
12k 12kz ν=1 r=1 r 1 − e−2πr/z
k ∞ rz
1 πih πz X X 1 2πiµ rhk e−2πµ k
= log z + πiC(h, k) + − − e
2 12k 12k µ=1 r=1 r 1 − e−2πrz
∞
k X k ∞
X 1 h′ e−2πνr/kz XX 1
2πiµrh/k e
−2πµrz/k
e2πiνr k . − e
ν=1 r=1
r 1 − e−2πr/z µ=1 r=1 r 1 − e−2πr/z
!
π 1 πi 1
+ −z + (h − h′ ) + πiC(h, k) = − log z.
12k z 12k 2
We now follow Siegel’s idea to get the whole thing as a sum of residues of
a certain function. Clearly there is r in it. Being integers r can be produced
1 1
by something like which has poles with residue − 2πi at every integral
1 − e2πix
valued x. So let us study a function like
−2πµxz/k
1 1 2πiµxh/k e
e
x 1 − e2πix 1 − e−2πxz
We may have to sum this from µ = 1 to µ = k. This should somehow be
the form of the function that we wish to integrate. We do not want it in the
whole plane. In fact, we can either take a wider and wider path of integration,
or multiply the function by a factor and magnify it; we prefer to do the latter. 237
We shall put xN for x, keep the path fixed and take N = n + 12 , n integer,
to avoid integral points, and then make n → ∞. The term corresponding to
µ = k should be treated separately, as otherwise the factor e−2πxz would stop
convergence. Also µh and µ should appear symmetrically for reasons which
we shall see. So introduce µ∗ ≡ µh (mod k), µ = 1, 2, . . . , k − 1, and choose
1 ≤ µ∗ ≤ k − 1. It turns out, taking all this together, that the following thing
will do. Write
k−1
1 πN x X 1 e2πµN x/k e−2πiµN x/kz
Fn (x) = − cot hπN x cot + · ·
4ix z µ=1
x 1−e 2πN x 1 − e−2πiN x/z
We then let n → ∞.
The poles of Fn (x) are indicated by the denominators and the cotangent
factors. These are
rz ir
x = 0, x=− , x= , r integer.
N N
x = 0 is a triple pole for the first summand.
1 πN x 1 1 z
− cot hπN x cot =− ·
4ix z 4ix πN x πN x
( ) ( )
(πN x)2 (πN x/z)2
1+ +· × 1− +···
3 3
27. Lecture 178
We had 240
k−1 ∗
1 πN x X 1 e2πµN x/k e−2πiµ N x/kz
Fn (x) = − cot hπN x cot + · × ,
4ix z µ=1
x 1−e 2πN x 1 − e−2πiN x /z
179
28. Lecture 180
( )
2πiNz (2πN 2 /z)2
× 1+ + +···
z 2
241
1
Fishing out the term in , the residue at x = 0 from this summand becomes
x
!2 !2 !2
iz 1 2πµN
1 2 1 2πµ∗ N 1 2πN N
+ (2πN) − − − 2πµ πN
4π2 N 2
2 k 12 2 kz 12 z k
2 2 2 2 ∗ 2 2 2 2 ∗ 2
)
N i 2π iµN 2π iµ N π iN 2π µ N
−4π2 µµ∗ 2 + + − +
k z kz kz z kz2
( ) 2
iz 2 2 1 2µ −2µ∗
i
1 2µ∗
= k + − + − +
4 µ2 3 k k2
4z 3 k
( )
i 4iµµ∗ 2iµ 2iµ∗
+ − 2 + + −i
4 k k k
( 2 ) 2 ! !
∗ ∗
µ µ 1 1 µ
µ 1
µ 1 µ∗ 1
= iz − + + − + + − − (*)
2k2 2k 12 2k2 2k 12
iz k 2 k 2
which appears here very simply as a sum of residues. The last expression
becomes !
k−1 1
− iz + + s(h, k)
12k iz
So the total residue at x = 0 is 243
! ! !
1 1 k−1 1 1 1
iz + − iz + + s(h, k) = iz + + s(h, k)
12 iz 12k iz 12k iz
Next, we consider the simple poles of Fn (x) at the points x = Nir (r , 0).
The coth factor is periodic and so the residue at any of these poles is the same
as that at the origin, which is π1 . Hence the residue of Fn (x) at x = Nir (r , 0)
becomes
k−1 ∗
N 1 πir X N −1 2πiµ r e2πµ r/kz
· cot + e k
4r πN z µ=1
ir 2πN 1 − e2πr/z
ey + e−y 2e−y
coth y = −y
=1+ y
y
e −e e − e−y
−2y
2e
=1+
1 − e−2y
coth y is an odd function so that 1y cot h y is even. Hence summing up over all 244
the poles corresponding to r = ±1, . . . , ±n, we get the sum of the residues
n ( ) k−1 n ∗
1 X1 2e−2πr/z 1 X X 1 2πih′ µ∗ r/k e−2πµ r/kz
= 1+ + e
2πi r=1 r 1 − e−2πr/z 2πi µ∗ =1 r=1 r 1 − e−2πr/z
28. Lecture 182
k−1 n ∗
1 X X 1 2πih′ (k−µ∗ )r/k e2πµ r/kz
− e ,
2πi µ∗ =1 r=1 r 1 − e2πr/z
where we have made use of the fact that hh′ ≡ −1 (mod k), so h′ µ∗ ≡ hh′ µ ≡
−µ (mod k), or µ ≡ −h′ µ∗ (mod k). In the last sum replace µ∗ by k − µ∗ ; then
the previous sum is duplicated and we get
n k−1 n ∗
1 X1
2e−2πr/z 1 X X 1 2πih′ µ∗ r/k e−2πµ r/kz
1 + −2πr/z
+ e −2πr/z
2πi r=1 r
1−e πi µ∗ =1 r=1 r 1−e
n k n
1 X 1 1 X X 1 2πih′ νr/k e−2πνn/kz
= + e
2πi r=1 r πi ν=1 r=1 r 1 − e−2πr/z
This accounts for all the poles on the imaginary axis (except the origin 245
which has been considered separately before).
Finally we have poles x = rz
N (e , 0) on the other diagonal of the parallelo-
gram. The same calculation goes through verbatim and we get the sum of the
residues at these poles to be
n k n
i X 1 i X X 1 2πihνr/k e−2πνrz/k
+ e
2π r=1 r π ν=1 r=1 r 1 − e−2πrz
Lecture 29
We had 246
k−1 ∗
1 πN x X 1 e2πµN x/k e−2πiµ N x/kz
Fn (x) = − cot hπN x cot +
4ix z µ=1
x 1 − e2πN x 1 − e−2πN x/z
The residue at x = 0 is
!
1 1
iz + + s(h, k),
12k iz
s(h, k), which will interest us for some time, being
k−1 ! " # !
X µ 1 hµ hµ 1
− − − .
µ=1
k 2 k k 2
ir
The residues at the points x = N (r , 0) amount to
n k n
1 X 1 1 X X 1 2πih′ ν r e−2πνr/kz
+ e k ;
2πi r=1 r πi ν=1 r=1 r 1 − e−2πr/z
zr
and the residues at the points x = N (r , 0)
n k n
i X 1 i X X 1 2πihµ r e−2πµrz/k
+ e k
2π r=1 r π µ=1 r=1 r 1 − e−2πrz
P
When we add up, the sums nr=1 1r , the disagreeable ones which would have 247
gone to infinity, fortunately destroy each other; so the sum of the residues of
Fn (x) at all its poles becomes
183
29. Lecture 184
! k n
1 1 1 X X 1 2πih′ νr/k e−2πνr/kz
− z + s(h, k) + e
12ki z πi ν=1 r=1 r 1 − e−2πr/z
k n
1 X X 1 2πihµr/k e−2πµrz/h
− e
πi µ=1 r=1 r 1 − e−2πrz
x = ρi + σz; ρ, σ ≤ 0, ρ + σ = 1.
Actually we take only ρ, σ > 0; we shall exclude the points i and z them-
selves. Then this becomes
1 eπN(ρi+σz) + e−πN(ρi+σz) eπiN(ρi+σz)/z + e−πiN(ρi+σz)/z
− × i ×
4i eπN(ρi+σz) − e−πN(ρi+σz) eπiN(ρi+σz)/z − e−πiN(ρi+σz)/z
The size of the first factor is determined by the terms eπNσz and e−Nπσz
in the numerator; the first term becomes big and the other goes to zero as
N → ∞(σ > 0 and Re z > 0). So we divide by the first term. Similarly for the
29. Lecture 185
1 πN x 1 eπN(−ρi+σz) + −e−πN(−ρi+σz)
− cot hπN x cot = − πN(−ρi+−σz)
4i z 4e − e−πN(−ρi+σ−z)
−ρi −ρi
eπiN z +σ
+ e−πiN z +−σ
× −ρ ρi
eπiN z
i+σ
− e −πiN − z +σ
ρi
1 1 + e−2πN(−ρi+σz) 1 + e−2πiN − z +σ
=− −2πN(−ρi+σz)
−× ρ−i
41−e 1 − e−2πiN − z +σ
Let N → ∞. Assuming that the denominator is going to behave decently,
this goes to − 41 . The other pieces go to zero for the same reason as before. And
all this is good for the opposite side too.
29. Lecture 186
We now have got to show that the convergence it nice and the denominators
do not make any fuse. This we can clarify in the following way. Consider the
denominator 1 − e−2πN(ρi+σz) .
Difficulties will arise if the exponent comes close to an even multiple of πi.
So we should see that it stays safely away from these points.
251
And actually it stays away from the danger spots by the same distance, for 251
the exponent is −2N(πiρ + πzσ) i.e., a point on the segment joining (2r + 1)πi
and (2r + 1)πz. Since ez is periodic there is a minimal amount by which it
stays away from 1. The second denominator looks a little different. We have
π
z instead of πz. But we have only to turn the whole thing around. We see how
essential it was to take N = n + 21 = (2n + 1) 21 = on odd multiple of 21 .
So the convergence is nice, but not uniform. We can nevertheless say that
xFn (x) → ± 14 boundedly on the sides of ρ except for the vertices where it does
not converge but oscillates finitely. But bounded convergence is enough for
1
interchanging integration and summation. Fn (x) → ± 4x and the x does not
ruin anything because it stays away from zero everywhere on ρ. Hence
Z
1
lim Fn (x)dx
n→∞ 2πi p
z is in the positive half-plane; we can take the principal branch of the logarithm, 252
so that we get on integration, since log i is completely determined,
1 π πi 1
− log z − log z + =− log z
4πi 2 2 2πi
which is the complete formula in all its details. The mysterious s(h, k) enjoys 253
certain properties. It has the group properties of the modular group behind it
and so must participate in them.
Lecture 30
Last time we had the formula of transformation of log η in the following shape: 254
! !
h′ + i/z h + iz 1 πi ′
log η = log η + log z + (h − h) + πis(h, k),
k k 2 12k
where s(h, k) is the Dedekind sum, which, by direct computation of residues,
was seen to be
k−1 ! " # !
X µ 1 hµ hµ 1
− − − .
µ=0
k 2 k k 2
Then
k !!
X µ hµ
s(h, k) = .
µ=1
k k
Now ((x)) is an odd function; for x integer, trivially ((−x)) = −((x)), and
for x not an integer,
1
((−x)) = −x − [−x] −
2
1
= −x + [x] + 1 − , since [−x] = −[x] − 1,
2
= −((x)).
∗∗∗∗∗∗∗∗∗∗∗∗∗∗∗∗∗∗∗∗∗∗∗
188
30. Lecture 189
and since hµ also runs through a full system of residues mod k when µ does 256
so, as (h, k) = 1, the second sum is zero, and we can therefore write
k !!
X µ hµ
s(h, k) =
µ=1
k k
30. Lecture 190
Let us now rewrite this in a form in which the modular substitution comes
into play
h′ + i/z h + iz
τ′ = , τ= ;
k k
so kτ − h = iz, and
h′ − 1/(kτ − h) h′ kτ − hh′ − 1
τ′ = =
k k(kτ − h)
h′ τ − (hh′ + 1)/k
=
kτ − h
′
( hhk+1 is necessarily integral for hh′ ≡ −1 mod k). So the modular substitution
is
′ −hh′ +1
h k = a b , c > 0.
k −h c d
The transformation formula for log η now reads
!
aτ + b 1 cτ + d πi
log η = log η(τ) + log + (a + d) − πis(d, c),
cτ + d 2 i 12c
since s(−d, c) = − − s(d, c).
Let us take in particular 257
a b 0 −1
= ;
c d 1 0
then we obtain !
1 1 τ
log η = log η(τ) + log ,
τ 2 i
the special case discussed by Siegel.
Let us now make two substitutions in succession:
aτ′ + b 1
τ′′ = , τ′ = − .
cτ′ + d τ
Then
−a/τ + b bτ − a
τ′′ = =
−c/τ + d dτ − c
We suppose c > 0, d > 0; (c, d) = 1. Then
1 cτ′ + d πi
log η(τ′′ ) = log η(τ′ ) + log + (a + d) − πis(d, c);
2 i 12c
30. Lecture 191
1 dτ − c πi
log η(τ′′ ) = log η(τ) + log + (b − c) − πis(−c, d).
2 i 12d
Sub tracting, and observing that
1 τ
log η(τ′ ) − log ηn(τ) = log ,
2 i
we have 258
1 τ 1 cτ′ + d 1 dτ − c
0= log + log − log
2 i 2 i 2 i !
πi a + d b − c
+ − − πi(s(d, c) − s(c, d))
12 c d
The sum of the logarithms on the right side is determinate only up to a
multiple of 2πi:
τ cτ′ + d dτ − c τ (−c/τ + d)/i
log + log − log = log + 2πik
1 i i i (dτ − c)/i
!
1
= log + 2πik
i
πi
= − + 2πik
2
Now each logarithm above has an imaginary part which is strictly less than
π
2 in absolute value; so
( )
τ cτ′ + d dτ − c 3π
Im log + log − log <
i i i 2
So the only admissible value of k is zero.
Hence we have 259
!
πi πi a + d b − c
0=− + − − πi(s(d, c) + s(c, d)) ,
4 12 c d
or since ad − bc = 1,
!
1 1 d c 1
s(d, c) + s(c, d) = − + + + .
4 12 c d cd
This is the reciprocity law for Dedekind sums. It is a purely arithmetical
formula for which I have given several proofs; here I reproduce the proof that
I gave originally, by lattice-point enumeration.
30. Lecture 192
This reduces to something that looks familiar; indeed the square brackets
appear in lattice-point enumeration. Here (h, k) = 1, but in a paper with White-
man I have also discussed the case where h, k are not coprime.
Enumerating by rows and columns parallel to the µ− and ν− axes, the num- 261
ber of lattice-points in the integer a the rectangle
30. Lecture 193
194
31. Lecture 195
Consider the rectangular parallelopiped with three concurrent edges along 263
the axes of µ, ν and ρ, the lengths of these edges being h, k, hk respectively. Dis-
sect the parallelopiped into three pyramids having a common apex at the origin
and having for bases the three rectangular faces which do not pass through the
origin, viz. ABCD, BCFE and CDGF. We now compute the number of lattice
points in each pyramid. Take for example the pyramid O(BEFC). Consider
a section parallel to the (ρ, ν)-plane at a distance µ along the µ-axis. The lat-
tice points lie in such sheets. The edges of this section are hµ and µ hk . The
number of lattice points on this sheet (including possibly those on the edges) is
h i k−1 h i
hµ µh hµ µh
P
k . So for the whole pyramid the number = k . For the pyramid
µ=1
h−1
P h i
νk
O(ABCD), the one facing us, the number is kν h
ν=1
Of course are some points on the common edge. Finally there is a pyramid 264
of exceptional sort which lies upside down. Consider a section at a height
h parallel to the (µ, ν) plane the numberr of lattice points on and inside this
pyramid is seen to be
hk−1
X ρ ρ
.
ρ=1
h k
So altogether we have
k−1 " # Xh−1 " # hk−1
X ρ ρ
X µh νk
hµ + kν +
µ=1
k ν=1
k ρ=1
h k
31. Lecture 196
points, including some points which have been counted twice over. But the
number of lattice points inside: the parallelopiped is equal to (h − 1)(k − 1)(hk −
1). Hence making a correction for the lattice points on the cleaving surfaces
through the edges CF and CD which have been counted twice (the surface
along BC has no points on it because (h, k) = 1), we have
k−1 " # Xh−1 " # hk−1
X ρ ρ
X µh νk
hµ + kν +
µ=1
k ν=1
h ρ=1
k h
= (h − 1)(k − 1)(hk − 1) + (h − 1)(k − 1)
= hk(h − 1)(k − 1)
Now write
hk−1
Xρ ρ
S =
ρ=1
h k
ρ ρ 1 ρ ρ ρ
= − − , if h ∤ ρ; − , if h ∤ ρ.
h h 2 h h h
So 265
hk−1
X( )( )
ρ 1 ρ ρ 1 ρ
S = − − − −
ρ=1
h 2 h 2 2 k
So
hk−1
X( )( ) X hk ! X hk !
ρ 1 ρ 1 ρ ρ 1 ρ ρ 1
S = − − − − − −
ρ=1
h 2 k 2 ρ=1
h k 2 ρ=1
k h 2
hk k−1 ( ) h−1 ( )
X ρ ρ 1 X hσ 1 1 X 1 kτ 1
+ + − + −
ρ=1
h k 2 σ=1 k 2 2 τ=1 2 h 2
31. Lecture 197
P µ
Since k = 0, this becomes
µ mod k
hk−1
X( 2 ) hk hk
ρ 1 ρ ρ 1 1 X ρ 1 X ρ
S = − + + − ρ − ρ
ρ=1
hk 2 h k 4 4 ρ=1 k h ρ=1 k
hk ! !
X ρ ρ 1 h(k − 1) k − 1 1 k(h − 1) h − 1
+ + − + −
ρ=1
h k 2 2 2 2 2 2
ρ = hr + s; r = 0, 1, . . . , k − 1; s = 1, . . . , h.
hk ρ X k−1 Xh !!
X hr + s
ρ = (hr + s)
ρ=1
h r=0 s=1
h
k−1 X
X h s k−1 X
X h s
= hr + s
r=0 s=1
h r=0 s=1
h
Xh s
=k s
s=1
h
next, consider
ik
X ρ ρ
ρ=1
h k
Look at the denominator of s(h, k). At worst it can have for factors 2 and 269
31. Lecture 199
(k − 1)(k − 5)
s(2, k) =
24k
Let us calculate s(5, 27).
1 12 + 52 + 272
s(5, 27) + s(27, 5) = − +
4 12 × 5 × 27
1 12 + 22 + 52
s(2, 5) + s(5, 2) = − +
4 12 × 2 × 5
s(5, 2) = 0 = s(1, 2), and on sub traction,
s(5, 27) = 35/(6 × 27); and we know that
the denominator could be at most 2.27(27, 6) = 6 × 27.
Lecture 32
We shall study a few more properties of Dedekind sums. We had the reciprocity 271
law !
1 1 h k 1
s(h, k) + s(k, h) = − + + + .
4 12 k h hk
From this we deduced as a consequence
Now when do the Dedekind sums vanish? Let us write s(h, k) in the more
flexible form:
X µ µh !!
s(h, k) =
µ mod k
k k
Let hh∗ ≡ 1 (mod k). Since (h∗ , k) = 1, h∗ µ runs through a full residue
system modulo k, and so
!! !!
X µh∗ µhh∗
s(h, k) =
µ mod k
k k
X µ µh∗ !!
=
µ mod k
k k
= s(h∗ , k)
This is of some significance. We came to s from the substitution a b and
c d
since ad ≡ 1 (mod c), s(d, c) = s(a, c). hh′ ≡ −1 (mod k), and
X µ µh !!
s(h, k) =
µ mod k
k k
200
32. Lecture 201
!! !!
X h′ µ µhh′
=
µ mod k
k k
X µ h′ µ !!
= −
µ mod k
k k
= −s(h′ , k)
s(h, k) = −s(h, k)
or s(h, k) = 0 if h2 ≡ −1 (mod k)
to do with the existence if certain quadratic forms with minimum values close
to zero for integers. 1, 1, 2 is a Markoff triple. If we keep two of them fixed, for
the third we get a quadratic equation of which one root we know to be rational.
So the other root is rational too. For instance if a, b = 1 are fixed, we have
c2 − 3c + 2 = 0 or (c − 1)(c − 2) = 0; – the triples pre 1, 1, 1 and 1,1,2. If we
take the triple a, 1, 2, then a2 + 5 = 6a or a = 1, 5; we have the triples b, 1, 2;
1, 1, 2. T = 0 only if a, b, c being to a Markoff triple. For such a triple,
µ=1
k j=1 µ=1
= kcl ,
k
1 X µ −2πiµ ℓ
i.e., cl = e k
k µ=1 k
So if k | l, then
1 k(k − 1) k − 1
cℓ = − = 0.
k2 2 2k
In particular
ck = 0.
If k ∤ l, then writing
k−1
X ℓ
S = µe−2πiµ k ,
µ=1
k−1
X
ℓ µ+1
S e−2πi k = µe−2πi k ℓ
µ=1
k
X ℓ
= (ν − 1)e−2πiν k
ν=2
k
X
l l l
= S − e−2πi k + k − e−2πiν k + e−2πi k
ν=1
32. Lecture 204
So 276
k
S =
e−2πiℓ/k − 1
Hence, if k ∤ l, then
−1 1
cℓ = +
k(1 − e2πiℓ/k ) 2k
−2 + 1 − e−2πiℓ/k
=
2k(1 − e−2πiℓ/k )
ℓ
1 1 + e−2πi k
=− ·
2k 1 − e−2πi kℓ
i πℓ
= cot
2k k
So we have what is essentially Eisenstein’s formula:
µ k−1
i X πℓ µ
= cot e2πi j k
k 2k j=1 k
This is an explicit formula for µk as a finite Fourier series. We utilise it
for Dedekind sums.
X µ hµ !!
s(h, k) =
µ mod k
k k
k−1 k−1
1 X X π j 2πi j µ X πℓ µ
=− cot e k × cot e2πiℓh k
4k2 µ mod k j=1
k ℓ=1
k
k−1 k−1
1 XX πj πl X 2πi µ ( j+hℓ)
=− 2
cot cot e k
4k j=1 ℓ=1 k k µ mod k
k−1
1 X πℓ −πhℓ
=− cot cot ,
4k ℓ=1 k k
since in the summation with respect to µ only those terms remain for which 277
j + hℓ ≡ (mod k). Then
k−1
1 X πℓ πhℓ
s(h, k) = cot cot .
4k ℓ=1 k k
32. Lecture 205
So the residue at z = 0 is
! !
k2 π2 π2 π2 h2 k k 1 h
− − 2− =− + +
π2 h 3 3k 3k2 3π h hk k
So
! k−1
X k k h 1 1X πℓ πhℓ
(Res) = − + + + cot cot
3π h k hk π ℓ=1 k k
32. Lecture 206
h−1
k X πrh πh
+ cot cot
πh k=1 h h
! !
k k h 1
= − + + + 12s(h, k) + 12s(k, h)
3π h k hk
And this is equal to Z 279
1
f (z)dz
2πi R
where R is the rectangle. On the vertical lines the function the same value (by
periodicity) end so the integrals cancel out. Hence
Z
−iΩ+k
Z iΩ+k
Z
1 1
f (z)dz = −
2πi R 2πi
−iΩ iΩ
Now
eiω + e−iω
cot ω = i , ω = x + iy,
eiω − e−iω
eix−y + e−ix+y
= i ix−y ;
e − e−ix+y
x varies from o to k and y = ±Ω, for this
−i, as y = Ω → ∞
→ uniformly
i, as y = −Ω → −∞
Therefore
Z
1 1 n3 o
lim f (z)dz = i k − (−i)3 k
Ω→∞ 2πi R 2πi
2ki3 k
= =−
2πi π
280
! !
k h k 1
∴ − + + + 12s(h, k) + 12s(k, h)
3π k h hk
k
=−
π !
k h 1
or 12s(h, k) + 12s(k, h) = −3 + + + ,
h k hk
which is the reciprocity formula.
Part IV
Representation by squares
207
Lecture 33
We wish to begin the study of the representation of a number as the sum of 281
squares:
n = n21 + n22 + · · · + n2r
We shall develop in this connection the Hardy-Littlewood circle method.
Historically it is an off shoot of the Hardly-Ramanujan method in partition-
theory, though we did not develop the latter in its original form in our treat-
ment. The circle method has been applied to very many cases, and the problem
of squares is a very instructive one for finding out the general thread. We
shall later replace the problem by that of the representation of n by a posi-
tive quadratic form. This would involve only the general Poisson summation
formula. In the case of representation as the sum of squares there is some
simplification, because the generating ing function is the rth power of a simple
V function. We shall deal with the asymptotic theory. Later we may go into
Siegel’s theory of quadratic forms.
Let us write
X∞ X∞
2 2
Θ(x) = xn = 1 + 2 xn ,
n=−∞ n=1
on collecting the terms with exponent n, where Ar (n) is the number of times n 282
208
33. Lecture 209
where ξhk are the arcs over which one integration piecemeal the prime denoting
that (h, k) = 1. Consider on each piece ξhk the neighbourhood of a root of unity: 283
h
x = e2πi k −2πξ
Re z < 0, and set z = δN − iϕ, so chat we have a little freedom along both real
and imaginary axes.
h
x = e2πi k −2πδN +2πiϕ .
The choice of the little arc ϕ is also classical. hk is a certain Farey fraction,
with adjacents hk11 and hk22 , say. hk11 < hk < hk22 . We limit ϕ on the seperate arcs.
Introduce the mediants:
h1 h1 + h h h2 + h h2
< < < < ,
k1 k1 + k k k2 + k k2
h +h h +h
h
So that the interval k11 +k , k22 +2 gives the movement of k + ϕ. So ϕ runs
between
h1 + h h h2 + h h
−Vhk′ = − ≤ϕ≤ − = Vhk′′
k1 + k k h2 + k k
33. Lecture 210
1 1
−Vhk′ = − ; Vhk′′ = ;
(k1 + k)k (k2 + k)k
k1 +k
and since 2N > k2 +k > N, we have necessarily
1 1
≤ |Vhk | ≤
2Nk Nk
Now changing the variable of integration to ϕ, we can write
′′
X′ ZVhk h
−2πi hk n
Ar (n) = e Θr e2πi k −2πz e2πnz dϕ
0≤h<k≤N
−Vhk′
284
The trick is to overcome the difficulty in the integral by replacing on each
arc the highly transcendental function by a simpler function. Here we stop for
a moment to see what we can do with the integrand.
h X∞
e(2πi k −2πz)n
h 2
Θ · e2πi k −2πz =
n=−∞
k−1
X X
h 2 2
= e2πi k j e−2πzn
j=0 n≡ j (mod k)
k−1
X ∞
X
h 2 2
(q+ kj )2
= e2πi k j e2πzk ,
j=0 q=−∞
where we have written n = kq + j. We can now handle this from our V -series
formula. We proved (Lecture 12) that
C(τ) 1 2
V3 (V / − ) = eπiτV V3 (V τ/τ)
i τ
r
C(τ) i
and = , Imτ > 0.
i τ
Since 285
∞
X
πin2 τ 2πinV
V3 (V /τ) = e e ,
n=−∞
∞
X 2
= e−πt(n+V )
n=−∞
j
Replacing n by q, V by k and t by 2zk2 , we have
k−1 ∞ 2
h X h 2 i X − πq j
Θ e2πi k −2πz = e2πi k j p e 2zk2 e2πiq k
2
2zk q=−∞
j=0
∞ πq2
1 X − 2zk
= √ e 2 T q (h, k)
k 2z q=−∞
k−1
X h j2 +q j
where T q (h, k) = e2πi k
j=0
k−1
X h 2
T 0 (h, k) = e2πi k j = G(h, k),
j=0
where G(h, k) are the so-called Gaussian sums which we shall study in detail.
They are sums of roots of unity raised to a square power, Θ is actually a V3 ,
and when we evaluate T q we get some other V .
We now write
h 1
Θ e2πi k −2πz = √ {G(h, k) + H(h, k; z)}
k 2z
∞
X 2
− πq
where H(h, k; z) = T q (h, k)e 2k2 z
q=−∞
q,0
We shall throw H into the error term. Let us appraise T q (h, k), not explic-
itly: that will take us into Gaussian sums.
k−1
X h j2 +q j
T q (h, k) = e2πi k
j=0
k−1 X
X k−1
2πi 2 i 2
|T q (h, k)|2 = e k (h j +q j) e−2π k (hℓ +qℓ)
j=0 ℓ=0
33. Lecture 212
X X i 2
−ℓ2 )+q( j−ℓ))
= e2π k (h(ℓ j
j mod k ℓ mod k
X X i
= e2π k ( j−ℓ)(h( j+ℓ)+q) ,
j mod k ℓ mod k
The inner sum is a sum of the roots of unity. Two cases arises, according
as k | 2a or k ∤ 2a. k odd implies that a = 0 and k even implies that a = 0 or
k | 2. In case k | 2a, the sum is zero. We then have
!
2
2 2π ki h k4 + 2k q
|T q (h, k)| = k, if k is odd; k 1 + e , if k is even
= k 1 + eπi( 2 +q) , if k is even
hk
= o or 2k if k is even
hk
It is of interest to notice that T q = 0 only if k is even and 2 + q is an odd
integer. In any case, √
|T q (h, k)| ≤ 2k,
and this cannot be improved. We then have
∞ √
X πq2 1
|H(h, k; k; z)| ≤ 2 2ke− 2k2 R z (q = 0 is not involved here).
q=1
√ − π R1 X ∞
(q2 −1) 1
= 2 2ke 2k 2 z e−π 2k2 R z
q=1
√ ∞
X
π 1 3πm 1
= 2 2ke− 2k2 R z e− 2k2 R z
m=0
√ π 1 1
= 2 2ke− 2k2 R z 3π
− R 1z
1−e 2k2
1 1 1 1 δN
2
R =R 2 =R 2 = 2 2
k z k z k (δN − iϕ) k (δN + ϕ2 )
1 1 δN 1
∴ 2
R ≥ 2 1
, since |Vhk | ≤ ,
k z 2
k δN + N 2 kN
δN 1
≥ =
N 2 δ2N + N12 N 2 δN + N 21δN
We want to make this keep away from 0 as far as possible. This gives a
desirable choice of δN . Make the denominator as small as possible. Since x + 1x
is minimised when x = 1, we have
1 1 1
R ≥ ,
k2 z 2
1
√ {G(h, k) + H(h, k; z)}
k 2z
√ π 1
where |H(h, k; z)| < C ke− 2k2 R z
Now ′′
290
ZVhk h
P′ h
Ar (n) = e−2πi k n 2πznΘ e2πi k n−2πz
0≤h<k≤N
−Vhk′
214
34. Lecture 215
1 1
Now 2kN ≤ Vhk ≤ kN and Vhk′ ≤ ϕ ≤ Vhk′′ , while δN = N12 . Putting 291
δN r π
X = k2 (δ2 +ϕ2 ) , the integrand becomes X 4 e− 2 X which remains bounded. (It was
N
for this purpose that in our estimate of H(h, k; z) earlier we retained the factor
2 1
e−π/(2k )·R z ). Hence the last expression is less than or equal to
′′
X ZVhk
2π Nn2 r n r
C e N 2 dϕ = Ce2π N2 N 2 ,
o≤h<k≤N −Vhk′
since the whole Farey dissection exactly fills the interval (0, 1).
In the next stage of our argument we take the integral
′′
ZVhk
e2πnz
k
z4
Vhk′
and write it as
Z∞ Z∞ Z −Vhk′
2πnz
− − e
r/2 dϕ
z
−∞ Vhk′′ −∞
′
−V
R hk
(Here and in the estimate of the other integral , we make use of the
−∞
fact that the interval from Vhk′ to Vhk′′ is neither too long nor too short. This
argument arises also in Goldbach’s problem and Waring’s problem). The right
side is equal to
Z∞ Z∞
r−2 2π Nn2 N 2 dϕ 2π Nn2 r−2 dψ
N e r = e N
4 2
(1 + N ϕ ) 4 (1 + ψ2 )r/4
1 N
− 2kN 2k
Z∞
2π Nn2 r−2 dψ
<e N
ψr/2
N
2k
This appears crude but is nevertheless good since ϕ never comes near 0;
N/2k > 12 , and the ratio of ψ2 to 1 + ψ2 is at least 13 and so we lose no essential
order of magnitude. The last integral is equal to
n
N − 2r +1
Ce2π N2 N r−2 , r ≥ 3,
2k
n r r
= Ce2π N2 N 2 −1 k 2 −1
−Vhk′
R
A similar estimate holds for also. So, 293
−∞
!r Z∞ 2πnz
P ′ h G(h, k) e
Ar (n) − e−2πi k n √ r/2
dϕ
0≤h<k≤N k 2 z
−∞
2π n
r/2 P′ 1 2π n
< Ce N2 N +C e N2 N r/2−1 kr/2−1
0≤h<k≤N kr/2
34. Lecture 217
n h
< Ce2π N2 N r/2 + Ce2π N2 N r/2−1
P′
0≤k≤N
n
= Ce2π N2 N r/2 .
This, however, does not go to zero as N → ∞; we have no good luck here as
we had in partitions. So we make the best of it, and obtain an asymptotic result.
Let n also tend to infinity. We shall keep n/N 2 bounded, without lotting; it go
to zero, as in the latter case the exponential √ factor would become 1. We have
to see to it that n ≤ CN 2 i.e., N is at least n. Otherwise the error term would
increase fast. Making N bigger would not help in the first factor and √ would 294
make the second worse. So the optical choice for N would be N = [ N]. The
error would now be r
O n4
We next evaluate the integral
Z∞
e2πnz
dϕ
zr/2
−∞
which exists for r > 2, and is actually the Hankel loop integral, and hence equal
to
2π(2πn)r/2 − 1
Γ(r/2)
Hence, for f ≥ 3. We hence the number of representations of n as the sum 295
of r squares:
r
(2π)r/2 n 2 −1 P′ G(h, k)r −2ri h
Ar (n) = · r/2 · e k + O(nr/4 ).
Γ(r/2) 2 0≤h<k≤N kr
34. Lecture 218
πr/2 r −1
Ar (n) = n 2 S r (n) + O(nr/4 ),
Γ(r/2)
X∞
where S r (n) = Vk (n)
k=1
S r (n) is the singular series. We shall show that S r (n) remains bounded at
least for r ≥ 5.
Lecture 35
After we reduced our problem to the singular series in which the Gaussian 297
sums appear conspicuously, we have to do something about them before we
proceed further. The Gaussian sums are defined as
X h 2
G(h, k) = e2πi k ℓ , (h, k) = 1
ℓ mod k
For, put ℓ = rk1 + sk2 ; when r runs modulo k2 and s modulo k1 , ℓ runs
through a full residue system modulo k1 k2 . Hence
X X
2πi h (k r+k s)2
G(h, k1 k2 ) = e k1 k2 1 2
k mod k2 s mod k1
X X
2πi k hk (k12 r2 +k22 s2 )
= e 1 2
r mod k2 s mod k1
X hk1 2 X hk2 2
2πi k2 r 2πi k1 s
= e e
r mod k2 s mod k1
= G(hk1 , k2 )G(hk2 , k1 ).
219
35. Lecture 220
write ℓ = mpα−1 + r;
r=0,p|r s=0
α−2
pX−1
h
2πi s
=p e pα−2
s=0
= pG(h, pα−2 )
ℓ = m2λ−1 + r; m = 0, 1; r = 0, 1, . . . , 2λ−1 − 1
35. Lecture 221
2λ−1
X−1 2λ−1
X−1
h 2 h λ−1
+r)2
G(h, 2λ ) = e2πi 2λ r + e2πi 2λ (2
r=0 r=0
This, however is not a Gaussian sum. The substitution for ℓ does not work;
to be effective, then we take
ℓ = m2λ−2 + r; m = 0, 1, 2, 3; r = 0, 1, . . . , 2λ−2 − 1.
This is not Gaussian sum either. But is is of the form (*). We therefore 301
have, for λ ≥ 4, G(h, 2λ) = 2G(h, 2λ−2). If λ = 4, we need go down to only
22 = 4 and if λ = 5 to 23 = 8. So we need separately G(h, 8) and G(h, 4); and
of course G(h, 2). These cases escape us, while formerly only G(h, p) did. For
λ ≥ 4, we may write
λ λ
G(h, 2λ) = 2[ 2 ]−1G h, 2λ−2[ 2 +2] (2)
35. Lecture 222
Before we return to G(h, p), p > 2, we shall a digression an connect to the 302
whole thing with the Legendre-Jacobi symbols
p−1
X h 2
G(h, p) = e2πi p ℓ
ℓ=0
X h
=1+2 e2πi p a ,
a
the summation over all quadratic residues a modulo p, since along with ℓ, p − ℓ
is also a quadratic residue. We can write this in a compact form, so arranging
it that the non-residues get cancelled and the residues appear twice:
X ( r
!)
h
G(h, p) = 1+ e2πi p r
r mod p
p
!
X r 2πi hp r
= e
r mod p
p
This would appear in a completely new aspect if we utilised the fact that hr
runs through a full system of residues modulo p. Then
X h ! hr ! h
G(h, p) = e2πi p r
k mod p
p p
! X !
h r 2πi rp
= e
p r mod p p
35. Lecture 223
!
h
= G(h, p).
p
This is very useful if we new go to the Jacobi symbol. For prime p, the 303
Legendre symbol has the multiplicative property:
! ! !
r1 r2 r1 r2
=
p p p
Jacobi has the
following generalisation.
r
Define pq by
! ! !
r r r
= .
pq p q
Si it is ±1; if it is +1 it does not necessarily mean that r is a quadratic residue
modulo pq. The Jacobi symbol no longer discriminates between residues and
non residues. From the definition then
! !α !β
a a a
= ··· .
pα qβ · · · p q
The Jacobi symbol has the properties of a character, as can be verified by
using the Chinese remainder theorem.
We can now write
!α
h
G(h, pα ) = G(1, pα )
p
under all circumstances. How does this come about? Separate the cases: α
even, α odd.
taking the second and third factors together, and also the last two. And this is
!
h
G(1, pα qβ )
pα qβ
We shall do a little more than that; we shall study them in a more flexible
form. Define
k−1
X hg 2
S (h, k) = eπi k ℓ ,
ℓ=0
h, k > 0 but not necessarily coprime. We cannot now take the summation over
ℓ modulo k. For if k is odd, (ℓ + k)2 = ℓ2 + 2ℓk + k2 and k2 may give rise to
an odd multiple of πi in the exponent and hence introduce a change of sign,
We should therefore insist on this particular range of summation. S (h, k) are
connected with the Gaussian sums; indeed
G(h, k) = S (2h, k)
225
36. Lecture 226
0 1 2
Where C is the parallelogram with vertices at ±(1 + i)Ω, ±(1 + i)Ω + k, 307
with the slant sides inclined at 45◦ (infact this may be anything less than 90◦ )
to the real axis, and making a detour round 0 and k. When we push Ω to ∞, the
integrals along the horizontal sides will tend to zero. For instance on the upper
side, z = (1 + i)Ω + x, 0 ≤ x ≤ k, and the integrand is therefore
h 2 h 2 2
eπi k ((1+i)Ω+x) eπi k (2iΩ +2(1+i)Ωx+x )
2πi((1+i)Ω+x)
= 2πi(Ω+x)−2πΩ
e −1 e −1
h 2 h 2
e−π k (2Ω +2Ωx)+πi k (2Ωx+x )
=
e−2πΩ+2πi(Ω+x) − 1
h
→ 0 uniformly as Ω → ∞ since k > 0. Hence the integral can be written as 308
(1+i)∞+k
Z (1+i)∞
Z h 2
eπi k (z)
− dz
e2πz − 1
−(1+i)∞+k −(1+i)∞
= dz
e2πiz − 1
−(1+i)∞
36. Lecture 227
Let us assume from now on that hk is even. Then we can actually divide 309
out and the integral becomes
(1+i)∞
Z
h 2 X h−1
eπi k z e 2πiλz dz
−(1+i)∞ λ=0
Write z + λk/h = ω; and shift the integral back to the line from −(1 + i)∞
to (1 + i)∞ - this we can do since the integrand tends to zero along a horizontal
segment. This gives
(1+i)∞
h−1
X Z
−πi hk λ2 h 2
e eπi k ω dω,
λ=0 −(1+i)∞
q q
h h
or writing t = ω k, k > 0, 310
r (1+i)∞ r
h−1 Z h−1
k X −πi h λ2 πit2 k X −πi k λ2
e k e dt = A e h
h λ=0 h λ=0
−(1+i)∞
Hence r
k
S (h, k) = AS (−k, h).
h
In order to evaluate A, take a simple case: h = 1, k = 2
√
S (1, 2) = A 2S (−2, 1)
πi √
i.e., 1 + e 2 = A 2,
36. Lecture 228
√
So A = (1 + i)/ 2, an eighth root or unity.
So our reciprocity formula becomes complete:
r
1+i k
S (h, k) = √ S (−k, h).
2 h
Let us develop some corollaries.
1) h = 2, k arbitrary: 311
(1 + i)(1 + (−i)k ) √
G(1, k) = k.
2
We mention the four cases separately:
√
k if k ≡ 1 (mod 4)
0
if k ≡ 2 (mod 4)
G(1, k) =
√
i k if k ≡ 3 (mod 4)
√
(1 + i) k if k ≡ 0 (mod 4)
√
Hence the absolute value of G(1, k) can be 0, k or 2k.
So far k was only positive. The case k odd deserves some special mention.
k − 1 is even and
√
k if k−1
2 is even
G(1, k) =
√
i k if k−1 is odd.
2
2
k−1
2 ≡ 0, 1 (mod 4) according as k−12 is even or odd; so we can write this 312
as
k−1 2 √
G(1, k) = i( 2 ) k.
36. Lecture 229
But the left side is only G(l, k)G(1, k), and this is always > 0. So
!
−1 k−1
(−) 2 k > 0,
k
and since k > 0 by nature, !
−1 k−1
= (−) 2
k
which is Euler’s criterion for the Jacobi symbol.
2) h = 2, k odd.
r
1+1 k
G(2, k) = S (4, k) = √ b S (−k, 4)
2 4
1+i√ n πik πik o
= √ k 1 + e− 4 + e−πik + e− 4
2 2
1 + i 1 √ √ −πi k
= √ 2 ke 4
2 i
πi √
= e− 4 (k−1) k
πi k−1
√
= e− 2 2 2
k−1 √
= i− 2 k
36. Lecture 230
k
= i− 2 (1+ 2 )
k−1 k−1
k2 −1
= i− 4
k2 −1
= i−2 8
k2 −1
= (−) 8
So
(h−1)(k−1)
ib = i2 4 an odd number
(h−1)(k−1) (h−1)(k−1)
= (−) 4 (odd number) = (−) 4
! !
h k (h−1)(k−1)
∴ = (−) 4 .
k h
We have finished to some extent Gaussian sums; we treated then only in view 317
of their occurrence in the singular series defined as
∞
X
(r)
S (n) = Vk(r) (n)
k=1
!r
X G(h, k) −2πi h n
with Vk(r) (n) = Vk (n) = e k ,
h mod k
k
(h,k)=1
which appeared as the principal term in the expression for the number of rep-
resentation of n as the sum of r squares:
πr/2 r −1 (r)
Ar (n) = n 2 S (n) + O nr/4 ,
r
Γ 2
p prime.
We first prove the multiplicative property of Vk (n): for (k1 , k2 ) = 1,
232
37. Lecture 233
We had a similar situation in connection with Ak (n) for the partition func- 318
tion; but there the multiplication was more complicated. Here we have
1 X
−2πi hn
Vk1 k2 (n) = r
G(h, k1 k2 )r e k1 k2 .
(k1 , k2 ) h mod k k
1 2
(h,k1 k2 )=1
1 XX −2πi h n
G(k2 h1 + k1 h2 , k1 k2 )e k1 k2
(k1 k2 )r h h
1 2
1 XX
= G ((k2 h1 + k1 h2 )k1 , k2 )r
(k1 k2 )r h h
1 2
−2πi(k2 h1 +k1 h2 ) k nk
G((k2 h1 + k1 h2 )k2 k1 )r e 1 2
Now X 319
h 2 2
ℓ
G(ha2 , h) = e2πi k a
ℓ mod k
If (a, k) = 1, al also runs modulo k when ℓ does, so that the right side is
X h 2
e2πi k m = G(h, k)
n mod k
say; this is an absolutely convergent product. This simplifies matters consid- 320
erably. We have to investigate V only for those G′ s is which prime powers
appear.
We first take p = 2. then
γ2 (n) = 1 + V2 (n) + V22 (n) + · · ·
1 X n
V2λ (n) = λr G(h, 2λ)r e−2πih 2λ
2 λ
h mod 2
2∤h
λ λ
G(h, 2λ) = 2 2 −1 2(1 + ih ) = 2 2 (1 + ih )
1 X h
V2λ (n) = λr 2λr/2 (1 + ih )r e−2πi 2λ n
2 h mod 2λ
2∤h
1
X
r −2πi 2hλ n
X
r −2πi 2hλ n
= λr/2 (1 + i) e + (1 − i) e
2
h≡1 (mod λ4)
h≡− (mod 4)
h mod 2 h mod 2λ
2r/2
X
πi 4r −2πi 2hλ n
X
−πi 4r −2πi 2hλ n
= λr/2 e e + e e
2
h≡1 (mod 4) h≡−1 (mod 4)
1
πi 4r −2πi 2rλ X s
= λ−1 e e−2πi 2λ−2 n +
2 2 r
s mod 2λ−2
X
−πi 4r +2πi 2rλ s
−2πi λ−2 n
+e e 2
λ−2
s mod 2
= 0, if 2λ−2 + n;
2λ−2 r ν
λ−1
cos π − 2π , if 2λ−2 /n, n = 2λ−2 .ν
2 2 r 4 4
1 π
i.e., (λ−1)( r −1) cos (2ν − r)
2 4 4
37. Lecture 235
321
Hence, for λ even, λ ≥ 4,
if 2λ−2 + n;
0,
V2λ (n) =
π
(*)
cos 4 (2ν−r)
r , if 2λ−2 .ν = n.
2(λ−1)( 2 −1)
(iii) λ odd, λ ≥ 3.
λ−3
G(h, 2λ) = 2G(h, 2λ−2) = 2 2 G(h, 23)
λ−3 λ+1
= 2 2 4eπih/4 = 2 2 eπih/4
1 λ+1 X r h
V2λ (n) = λr 2 2 eπih 4 e−2πi 2λ n ,
2 λ
h mod 2
2∤h
2λ−3
1 XX n
= λ−1
eπitr/4 e−2πi(8s+t) 2λ
2 2 r
t s=1
2λ−3
1 X X λ−3
= λ−1
e πitr/4−2πitn/2λ
e−2πisn/2
2 2 r
t s=1
= 0, if 2λ−3 ∤ n.
2λ−3 X
λ−1
eπit∤4(r−ν) = o, if 4/(r − ν);
2 2 r
t
λ−1
2
λ−1
eπi(r−ν)|4 , if 4/(r − ν)
2 2 r
1 ν−r
i.e., r · (−) 4 .
2(λ−1)( 2 −1)
Hence for λ odd, λ ≥ 3, 323
0, if 2λ−3 ∤ n;
V2λ (n) =
0, if 2λ−3 . | n, n = 2λ−3 ν, 4 ∤ (ν − r); (**)
ν−r
(−) 4
(λ−1)( r −1) , if 2λ−3 | n, 4 | (ν − r)
2 2
37. Lecture 236
Now, given n, only a finite number of powers of 2 can divide it. So the
situation 2λ−3 /n will occur sometime or the other, so that γ2 (n) is always a
finite sum.
∞
X 1
|γ2 (n) − 1| ≤ r
λ=2
2(λ−1)( 2 −1)
1 1
= ·
2 2 −1 1 − 1/2 2r − 1
r
1
= r/2−1 ;
2 −1
and this is valid for r ≥ 3. so the singular series behaves much better than we
expected.
Lecture 38
V2 (n) = 0.
For V2λ (n), λ > 1, we have to make a distinction between λ even and λ odd.
λ even.
if 2λ−2 ∤ n;
0,
V2λ (n) =
π
cos 4 (2ν−r)
r , if 2λ−2 ∤ n, n = 2λ−2 .ν.
2(λ−1)( 2 −1)
λ odd.
0, if 2λ−3 ∤ n;
V2λ (n) =
0, if 2λ−3 | n, n = 2λ−3 ν, ν − r . 0 (mod 4)
ν−r
(−) 4
(λ−1)( r −1) , if 2λ−3 | n, n = 2λ−3 ν, ν − r ≡ 0 (mod 4)
2 2
So for r = 3,
=1+ √ + ,
2 2
237
38. Lecture 238
where the last summand has to be replaced by 0 if (n − 3)/4 is not an integer. 325
Since 2n − 3 is odd, we have
π 1
| cos (2n − 3)| = √ ,
4 2
and thus clearly,
1 1
|γ2 (n)| ≤ 1 ++ =2
2 2
Moreover, γ2 (n) can vanish. This would require
n−3
(−) 4 = 1
π 1
and cos (2n − 3) = − √
4 2
simultaneously. But this is the case for
n ≡ 7 (mod 8),
1. 2 ∤ n. Then from relations (*) and (**) proved in lecture 37, we have
2. 2 | n Let n = 2α n′ , 2 | n′ . Then (*) and (**) show that V2λ (n) = 0 for 326
λ > α + 3. But actually V2λ (n) = 0 also for λ = α + 3. Indeed, for α odd,
λ = α + 1 is the last even, λ = α + 2 the last odd index for non-vanishing
V2λ (n). For α even, λ = α + 2 is the last even index: λ = α + 3 is odd and
since 4 ∤ (n′ − 4), we have also V2λ (n) = 0 for λ = α + 3.
α+2
X
∴ γ2 (2α n′ ) = 1 + V2λ (n)
λ=2
and V2λ (n) = 0 for λ = α + 2 since then 4 ∤ 2α−λ+1 . The numerators of the 327
non-vanishing V2λ (n) are −1 upto the last one, which is 1. And thus
1 1 1 1
γ2 (2α n′ ) = 1 − − 2 − · · · − α−1 + α
2 2 2 2
1 1 3
= α−1 + α = α
2 2 2
Although here γ2 (2α n′ ) > 0, we see that for α sufficiently large γ2 (n) can
come arbitrarily close to 0.
We now consider γ p (n) for p ≥ 3.
Now
!
h
G(h, pλ ) = G(1, pλ)
pλ
!λ pλ −1 2
h λ
= i 2 p2
p
2
pλ −1
r 2
!
i X h −2πi phλ n
∴ V pλ (n) = e
pλr/2 p
h mod pλ
p∤h
38. Lecture 240
So 2
pλ −1
r 2
i X −2πi phλ n
V pλ (n) = e
pλr/2
h mod pλ
p∤h
The inner sum here is a special case of the so-called Ramanujan sums:
X h
Ck (n) = e2πi k n
h mod k
(h,k)=1
Classify the λ′ s in S k (n) according to their common divisor with k. Then 329
X X λ
S k (n) = e2πi k n
d|k λ mod k
(λ,k)=d
X X λ n
= e2πi d · k/d
d|k λ mod k
( λk , dk )=1
X X µn
= e2πi k/d
d|k µ mod k
d
( µ, dk )=1
38. Lecture 241
X
= C dk (n)
d|k
X
= Cd (n).
d|k
The relation
X µ(n) X 1
s
=1
nX ns X
then implies that 0= µ( j) · 1 = µ(d), n > 1.
jk=n d|n
Lecture 39
242
39. Lecture 243
if α < λ − 1, n = pα n′ , p ∤ n′ ;
0,
−1 × pλ−1 = − − pα .
if α = λ − 1;
=
−1 × pλ−1 + pλ
= pλ (1 − 1 ),
if α ≥ λ.
p
For obtaining these values we observe that in the summation on the right
λ
side we have to take into account only such divisors d that pα is at most p. This
leads in the first case α < λ − 1 to a vacuous sum. In the second case the only
admissible divisor is pλ−1 ; in the last we have two divisors pλ−1 and pλ . Thus
V pλ (n) = 0
for λ > α + 1; we get again a finite sum for γ p (n)
We now take λr odd. We want
!
X h −2πi phλ n
e
λ
p
h mod p
p∤h
This is zero when pλ−1 ∤ n (because the inner sum vanishes). Otherwise,
let n = pλ−1 ν and p ∤ ν; then it is again zero because we have only a sum of
quadratic residue symbols (since the character is not the principal character).
If p | ν, the sum becomes
!
ν p−1 2
√
pλ−1G(ν, p) = pλ−1 i 2 p
p
So if n = pα · n′ where p ∤ n′ , then
0, if λ − 1 > α;
n′ p−1 2 √
V pλ (n) =
pα p i 2 p, if λ − 1 = α;
0, if 0 ≤ λ − 1 < α.
39. Lecture 244
So the only non vanishing term in the case α + 1 odd is V pα+1 (n).
Let us put things together now. Let r be even. If p ∤ n, then
p−1 2
r
i 2
γp = 1 + Vp = 1 −
pr/2
r p−1
(−) 2 2
=1−
pr/2
334
If p | n, n = pα · n′ , then
γ p = 1 + V p + V p2 + · · · + V pα + V pα +1
ǫp ǫ p2
=1+ (p − 1) + p(p − 1) + · · ·
pr/2 p2 r/2
ǫ pα ǫ pα+1
+ α pα−1 (p − 1) − (α+1)r/2 pα ,
p r/2 p
where ǫ p = (−)r(p−1)/4 for r , 4
! !
ǫp ǫp ǫp
= 1 − r/2 + r/2 1 − r/2
p p −1 p
2
ǫ pα
! !
ǫp ǫp ǫp
+ 1 − r/2 + · · · + 1 −
pα( 2 −1)
r
p2 2r − 1 p pr/2
!
ǫ pα+1 ǫ p −1
!
ǫ p
= 1 − r/2 1 − 1 − r/2−1
p(α+1)( 2 −1)
r
p p
For r = 4, the thing becomes critical: Let us look at it more specifically. 335
r(p − 1)
is even now and so ǫ p = 1. Hence
4
! 1
1 1 − pα+1
γp = 1 − 2
p 1 − 1p
P 1
The product is convergent since p2
< ∞. So
! 1
Y 1 Y 1 − p2
|S 4 (n)| ≥ γ2 1− 2
p
p p|n 1 − 1p
!2
Y 1 Y 1
≤ γ2 1− 2 1
p
p p|n 1 − p
Q
1 − 1p diverges to zero in the infinite product senses. So S 4 (n) is not
bounded. S 4 (n) could become very small if we keep the odd factors fixed
and introduce more even factors.
S 4 (n) is unbounded in both senses; it can be as large as we please or as
small as zero.
For r ≥ 5 we are again on the safe side. In this case the first term comes
from V pλ . We have
!
Vp
S 5 (n) ∼ 1 ± 5/2
p
! !
Y 1 Y 1
or C2 1 + 2 < S 5 (n) < C1 1− 2
p p
336
For r = 7 the situation is similar. For r = 6 the series again converges. So
for r ≥ 5.
0 < C1 < S r (n) < C2
This is of importance in the application to our problem.
We had
πr/2 r −1
Ar (n) = n 2 S r (n) + O(nr/4 )
Γ(r/2)
If r ≥ 5, 2r − 1 > 4r , and since S r(n) being bounded does not raise the order
in the term,
πr/2 r −1
Ar (n) ∼ n 2 S r (n)
Γ(r/2)
If, however, if r = 4, the sharpness of the analysis is lost. Both the first
factor and the error term are O(r) and S r (n) may contribute to a decrease in the
first term. If there are many odd factors for n, the main term is still good. But
if there are many powers of 2, it would be completely submerged.
For r = 4 the exact formula was given by Jacobi.
We shall consider also representation of n in the form an21 + bn22 + cn23 + dn24
in which connection the Kloosherman sums appear. We shall also cast a glance
at the meaning of the singular series in the sense of Siegel’e p-edic density.
Lecture 40
r ≡ 0 (mod 4).
In this case we need not bother about the sign of the Gaussian sums; the
fourth power of the coefficient becomes 1.
V2 (n) = 0
(−)r/4
(λ−1)( 2r −1)
. if λ < α + 1;
2
V2λ (n) = (−)r/4
− (λ−1) , if λ = α + 1;
2 ( 2r −1)
0, if λ > α + 1
So
1
r/4 1 1 1
γ2 (n) = 1 + (−) r + + · · · + −
22( 2 −1) 2(α−1)( 2 −1) 2α 2r − 1
−1 r r
22
α n
r
X (−) 2µ
= 1 + (−) 4 ,
µ=1 2 2r − 1
µ
246
40. Lecture 247
(2π)2k B2k
ζ(2k) = (−)k−1 , k ≥ 1,
2(2k)!
where B2k are the Bernoulli numbers. 339
!
1
B1 = − , B3 = B5 = B7 = · · · = 0; B2k , 0; sgnB2k = (−)k−1
2
2r/2 2 2r !
P1 = r/2 ×
2 − 1 (2π)r/2|Br/2 |
πr/2 r
Ar (n) ∼ n 2−1 S n (n)
Γ 2r
= Cr (n),
40. Lecture 248
say, where
πr/2 2r/2 2 2r ! r
X 1
Cr (n) = r/2 r/2
n 2 −1 γ2 (n) r
−1
Γ(r/2) 2 − 1 (2π) |Br/2| d|n,dodd
d 2
(a divisor sum! which is interesting, but not surprising, because the Jacobi 340
formula contains it).
r r
X 1
Cr (n) = r n 2 −1 r
2 2 −1|Br/2| d −1
2
d|n,d odd
r ≡ 0 (mod 8)
r
X 1
n 2 −1 γ2 (n)·
d|n,d odd
dr/2−1
1 + r1 + · · · + 1 1 1
r
X
=n 2 −1 − .
(α−1)( 2r −1) r
2 2 −1)
(
−1 r
2 2 2 α
d|n,d odd
d 2 −1
r
X (−) nδ
= n 2 −1 r , if n is even;
δ|n
δ 2 −1
r
X 1
n 2 −1 r , if n is odd.
δ|n
δ 2 −1
So for any n,
r
X 1 r
X (−) nδ
n 2 −1 γ2 (n) =n 2 −1 (−) n
r r
d|n,dodd
d 2 −1 δ|n
δ 2 −1
X n 2r −1
= (−)n (−)n/δ
δ|n
δ
X r
= (−)n (−)t t 2 −1
t|n
So 341
X r
Cr (n) = Qr (−)n (−)t t 2 −1 ,
t|n
40. Lecture 249
r
here Qr = r
2 2 −1 |B 2r |
This is exactly what appears for r = 4 in the Jacobi formula.
r ≡ 4 (mod 8)
r
X 1
n 2 −1 γ2 (n) r
d 2 −1
d|n,dodd
!X
r
−1 1 1 1 1
=n 2 1− r −···− r + , d odd r −1
2 2 −1 2(α−1)( 2 −1) α(
2 2
r
−1) d2
d|n
r
X (−)δ+ nδ +1
= n 2 −1 r
δ|n
δ 2 −1
X n
n 2r −1
= (−)δ+ δ +1
δ|n
δ
X n r
= (−) t +t+1 2 −1
t , if n is even;
t|n
X r
t 2 −1 , if n is odd;
t|n
22k − 1 ≡ 0 (mod q)
It might be of interest to take Cr (n), the main term in the formula for Ar (n) and 345
make some remarks about it.
X r −1
(−)n+d+ 4 ( d +1)d 2
r n
Cr (n) = Qr
d|n
this will give a sort of partial fraction decomposition. In the case where r ≡ 0
(mod 8), it is simpler:
∞
X X r
Hr (x) = 1 + Qr xn (−)n+d d 2 −1
n=1 d|n
∞
X X r
= 1 + Qr (−x)n (−)d d 2 −1
n=1 d|n
∞
X ∞
X
r
= 1 + Qr (−)d d 2 −1 (−x)qd
d=1 q=1
∞
X r (−x)d
= 1 + Qn (−)d d 2 −1
d=1
1 − (−x)d
∞
X r xd
= 1 + Qr d 2 −1
d=1
1 − (−x)d
( )
1.x r x2 r x3
= 1 + Qr + 2 2 −1 + 3 2 −1 +···
1+x 1 − x2 1 + x3
251
41. Lecture 252
πr/2 r −1
Ar (n) ∼ n 2 S r (n) (*)
Γ 2r
This comes out as a nice formula. Now could we not make some sense out 347
of
this formula?
r What is its inner meaning? We shall show that the first factor
πr/2 /Γ(r/2) n 2 −1 gives the average value of the number of representations of
n as the sum of r squares; the second factor also is an average, in the p-adic
measurement. We shall show that
X πr/2 X r −1
Ar (n) ∼ n2
n≤x
Γ(r/2) n≤x
41. Lecture 253
So for each
r individual n, S r (n) gives the deviation of Ar (n) from
πr/2 /Γ(r/2) n 2 −1 ; but on the average there is no deviation.
P
Let us first look at n≤x Ar (n).
X X X
ar (n) = 1
n≤x n≤x m21 +···m2r =n
X
= 1 ,
m21 +···+m2r ≤x
πr/2 r/2
= x
Γ(r/2)
348
The difference will not bezero but of the order of magnitude of the surface
of the sphere, i.e., O xr/2 − 1
Now consider the other side.
Z x
πr/2 X r −1 πr/2 r
n2 ∼ V 2 −1 dV
Γ(r/2) n≤x Γ(r/2) 0
πr/2 xr/2
=
Γ 2r + 1
So the first factor on the right of (*) gives the average. S r (n) has to be
adjusted. S r (n) is also, surprisingly, an average. It was defined as
S r (n) = γ2 (n)γ3 (n)γ5 (n) · · · γ p (n) · · · ,
and γ p (n) in turn was given by 349
∞
X
γ p (n) = 1 + V pλ (n)
λ=1
∞
X 1 X −2πi phλ n 1 X −2πi phλ n
=1+ G(h, pλ)r e G(h, pλ )r e
pλr pλr
λ=1 h mod pλ h mod pλ
p∤h p∤h
41. Lecture 254
1 X
X
2πi h ℓ
X 2πi h ℓ
= λr e pλ 1 e pλ 2
p
h mod pλ ℓ mod pλ
1 ℓ mod pλ 2
p∤h
X h
2πi pλ
ℓr 2 −2πi phλ n
e
e
ℓr mod pλ
1 X X 2πi h
= λr e pλ (ℓ12 + · · · + ℓr2 − n)
p
ℓ1 ,...,ℓ4 mod pλ h mod pλ
1 X
= λr
p
ℓ1 ,...,ℓr mod pλ
X
2πi psλ (ℓ12 +···+ℓr2 −n)
X t
2πi λ−1 (ℓ12 +···+ℓr2 −n)
e − e p
s mod pλ
t mod pλ−1
1 X X 2πi s (ℓ2 +···+ℓr2 −n)
= λr e pλ 1
p
ℓ1 ,...ℓn mod pλ s mod pλ
X X 2πi t (ℓ2 +···+ℓ22 −n)
− e pλ−1 1
ℓ1 ,...,ℓr mod pλ−1 t mod pλ−1
The sum on the right gives the number of times the congruence
ℓ12 + · · · + ℓr2 ≡ n (mod pλ ) can be solved, N pλ (n), say. Then
1
W pλ (n) = N pλ (n)
pλ(r−1)
We have therefore divided the number of solutions of the congruence by
pλ(r−1) . Now how many ℓ1 , . . . , ℓr mod pλ do we have? There are pλr possi-
bilities discarding n. n is one of the numbers modulo pλ . So dividing by pr , the
41. Lecture 255
N pλ (n)
average number of possibilities is pλ(r−1) . Hence is the average density
pλ(r−1)
modulo pλ of the solution of the congruence. And since the W pλ (n) eventu-
ally becomes γ p (n), each factor γ p (n) acquires a density interpretation, viz. the
p-adic density of the lattice points modulo pλ .
Lecture 42
The error term in the formula for the number of representations of n as the 351
sum of r squares, r ≥ 5, was O(nr/4 ). For r = 4 this did not suffice. We
shall therefore study the problem by Kloosterman’s method and find out what
happens when we want to decompose n in the form n = n21 + n22 + n23 + n24 .
1
We shall see that we can diminish the order in the error term by nearly 18 .
When Kloosterman did this for the first time (Act a Mathematic as 1927) he
took a slightly more general problem, that of representing n in the form n =
an21 + b22 + cn23 + dn24 , a, b, c, d integers. This works nicely; we get the singular
series and an error term which is smaller than before. The difficult not will be
about the arithmetical interpretation. The singular series will now be a difficult
phenomenon; we shall have multiplicativity, but the interpretation of the factors
γ p becomes complicated. We shall content ourselves with the analytical power
of the discussion. The generating function which will have to be discussed is
quite clear:
256
42. Lecture 257
(We could very well unsedes also the ‘semi-integral’ case). Let S be a pos-
itive definite integral symmetric matrix and x a column vector with elements
x1 , x2 , . . . xr in r-space. x′ is the transposed row-vector. x′ S x is a quadratic
form in r variables. The question is how often can we express an integer n by
integer vectors with respect to this quadratic form in r variables.
The generating function to be studied this time is
X ′
Fr (x) = xn S n , |x| < 1,
n
∞
X
=1+ Ar (n)xn .
n=1
uncertainty at all about Vhk′ , Vhk′′ ; and we could single them out if we insisted
on that.
For example, let hk = 95 , N = 12; what are the neighbours? hk11 < 5/9 < hk22 .
First determine h′ . 5h′ ≡ −1 (mod 9) or h′ = 7. Then 12 − 9 < k1 ≤ 12 and
k1 ≡ −7 (mod 9), so k1 = 11. Similarly 3 < k2 ≤ 12, k2 ≡ 7 (mod 9) so
k2 = 7. We need only k1 and k2 ; but for our own enjoyment let us calculate h1
and h2 .
h, 5 5 h
= −1, 2 = −1,
11 9 9 7
6
or h1 = 6, h2 = 4, so that we have 11 < 59 < 47 as adjacent fractions in the Farey
series of order 12. We do not need to display the whole Farey series.
Now utilise this in the following way.
1
k(k+k2 )
X′ Z h
2πi hk n
Ar (n) = e Fr e2πi k −2πz e2πnz dϕ
o≤h<k≤N
− k(k 1+k)
1
Kloosterman does the following investigation. In any case we are sure that k1 ,
k2 can at most become N. If we take k1 and k2 big we have a small interval of
integration. Since
k1 + k < k1 + 1 + k < · · · < N + k,
k2 + k < k2 + 1 + k < · · · < N + k,
1 1 1 1
> , > ,
k1 + k N + k k2 + k N + k
so that the interval of integration should be at least as big as the interval 356
−1/k(k + N) to 1/k(k + N). This interval is always present whatever be k1
and k2 . So Ar (n) is equal to the always present kernel
1
k(k+N)
Z
X′ h
e2πi k n (· · · )dϕ,
0≤h<k≤N −1
k(k+N)
There is no doubt about the integrals. The limits are all well-defined. This
will help us to appraise certain roots of unity closely-by the Kloosterman sums. 357
We shall now return to the integrand; that is a V -function and requires the
usual V treatment. Consider the r-fold V -series:
X ′
Θ(t) = e−πtn S n , Re t > 0.
n
which is an integral over the unit cube W, and so on translation with respect to 358
the vector n, becomes
XZ Z
′
· · · e−πt(V S V ) e−2πim V dV 1 · · · dVr
n
W+n
where
Z ∞ Z
′
S V −2πim′ V
c(m) = ··· e−πtV e dV1 . . . dVr
−∞
To get this into shape, consider the quadratic complement
π π
− (tV ′ + im′ S −1 )S (tV + iS −1 m) = −πtV ′ S V − πiV ′ m − πim′ V + m′ S −1 m
t t
Since m′ V = V ′ m,
Z ∞ Z
π ′ −1
m − πt (tV ′ +im′ S −1 )S (tV +iS −1 m)dV1 ...dVr
c(m) = ··· e− t m S e
−∞
Z ∞ Z √ √
− πt m′ S −1 m −π( tV ′ + √i t m′ S −1 )S ( tV + √i t S −1 m)
=e ··· e dV1 · · · dVr
−∞
√ 1
Put tV = w and µ = √
t
m′ S −1 . Then 360
π ′ −1 Z ∞ Z
e− t m S m ′ ′
c(m) = √ ··· e−π(w +iµ )S (w+iµ) dw1 · · · dwr
( t)r
−∞
261
43. Lecture 262
Since every positive definite quadratic form may be turned into a sum of
squares, we can put S = A′ A, so that the exponent in the integrand become
−π(w′ A′ + iµA′ )(Aw + iAµ); and writing Aw = z, we have
π ′ −1 Z ∞ Z
e− t m S m
′ ′
)(z+iV ) dz1 . . . d zr
c(m) = √ ··· e−π(z +iV
( t)r |A|
−∞
Let us now we back to our study of Ar (n). We had integrals with now limits 361
which were the special feature of the Kloosterman method.
1
k(k+N)
Z h N−1
X N−1
X
P′ −2πi hk n
Ar (n) = e Fr e2πi k −2πz e2πnz dz + ···+ ···
0≤h<k≤N ℓ=0 ℓ=0
1
− k(k+N)
Now
X ∞
X
′
Fr (x) = xn S n = 1 + Ar (n)xn
n n=1
X
2πi hk −2πz (2πi hk −2πz) ′
Fr e = e n Sn
n
X h ′ ′
= e2πi k n S n e−2πzn S n
n
43. Lecture 263
n = kq + ℓ, ℓ = (ℓ1 , . . . , ℓn ), 0 ≤ ℓ j < k.
So dismissing multiples of k,
X X ℓ′
ℓ
h ′
h ′ 2
(q′ + k )S q+ k
Fr (e2πi k −2πz ) = e2πi k ℓ S ℓ e−2πzk ,
ℓ mod k q
and applying the transformation formula we derived earlier, with t = 2zk2 and
α = 1k ℓ, this becomes
1 X h ′
X − π
m′ S −1 m 2πim′ ℓ
′
√ e2πi k ℓ S ℓ e 2zk2 e k
1 X − 2zkπ 2 m S −1 m
′
= √ e T k (h, m),
Dkr (2z)r/2 m
T k (h, 0) will be the most important; the others we only estimate. We require
a little more number theory for this. We cannot tolerate the presence of a both
a quadratic form and a linear form in the exponent. There will be a common
denominator in m′ S −1 m and that will have to be discussed.
Lecture 44
We had 363
2πi hk −2πz 1 X − π m′ S −1 m
Fr (e )= r r/2 1/2
e 2zk2 T k (h, m),
k (2z) D m
X 2πi ′ ′
and T k (h, m) = e k (hℓ S ℓ+m ℓ)
ℓ mod k
′ −1
The common denominator in m S m will be at most D, the determinant;
define k∗ and Dk by
kD = k · (k, D) · Dk = k∗ Dk , (Dk , k) = 1,
so that Dk is D stripped of all its common divisors with k. Suppose first that k
is odd. Let ρ be a solution of the congruence
4hDk ρ ≡ 1 (mod k∗ )
X
2πi h (ℓ′ S ℓ+4Dk ρm′ ℓ)
T k (h, m) = e k
ℓ mod k
X h ′ ′ −1
)S (ℓ+2Dk ρS −1 m) −(4D2k ρ2 m′ S −1 m)2πi hk
= e2πi k (ℓ +2Dk ρm S e .
ℓ mod k
′ −1
X ′ ′ −1
h 2 2 h
)S (ℓ+2Dk ρS −1 m)
= e−2πi k ·4Dk ρ m S m
e2πi k (ℓ +2Dk ρm S
ℓ mod k
D ρ
−2π kk m′ S −1 m
=e Uk ,
say, (using the definition of ρ), where Uk = Uk (h, m) is periodic in m with 364
period (k, D); it is enough if we take this period to be D itself. So
2πi hk −2πz 1 X X D ρ
− π 2 +2πi kk m′ S −1 m
Fr (e )= r Uk (h, s) e 2zk
k (2z)r/2 D1/2 s mod D m≡s (mod D)
264
44. Lecture 265
x 1
The power series goes in powers of D because D remains silent inside. This
is for k odd.
For k even, define. σ by
X h ′ ′
T k (h, m) = e2πi 4k (4ℓ S ℓ+4Dk σm ℓ)
ℓ mod k
h 2 2 ′ −1
X h ′ ′ −1
−2πi 4k Dk σ m S m )S (2ℓ+Dk σS −1 m)
=e e2πi 4k (2ℓ +Dk σm S
ℓ mod k
D ′ −1
−2πi 4kk σm S m
=e Uk (h, m),
where Uk again has a certain periodicity; we can take the period to be 2D and 365
forget about the refinement. So
h 1 X X − π
+ 2πi ′ −1
4k Dk σ m S m
Fr e2πi k −2πz = Uk (h, s) e 2zk2
kr (2z)r/2 D1/2 s mod 2D m≡s (mod 2D)
We have to get a clear picture of that we are aiming at. We are discussing the 366
function under the integral sign. We get it as
h 1 X
Fr e2πi k −2πz = Uk (h, s)
kr (2z)r/2 D1/2 s mod 2D
X π Dk σ
− +2πi 4k
e 2k2 z m′ S −1 m
m≡s (mod 2D)
Lemma 2.
|T k (h, m)| ≤ Ckr/2
We have
X h ′
X h ′
S λ +σm′ ℓ)
|T k (h, m)|2 = e2πi k (ℓS ℓ+σm ℓ) e−2πi k (λ
ℓ mod k λ mod k
266
45. Lecture 267
X ′
h
S ℓ − λ′ S λ +σm′ (ℓ − λ))
= e2πi k (ℓ ,
ℓ,λ
ℓ′ S ℓ − λ′ S λ = (ℓ′ − λ′ )S (ℓ + λ) + λ′ S ℓ − ℓ′ S λ
= (ℓ′ − λ′ )S (ℓ + λ) + ℓ′ S ℓ − ℓ′ S λ
= (ℓ′ − λ′ )S (ℓ + λ),
this is equal to
X ′
h
− λ′ )(S (ℓ + λ)+σm)
e2πi k (ℓ
ℓ,λ
X X h ′
= e2πi k α (S (ℓ + λ)+σm)
α mod k ℓ − λ≡α mod k
X X h ′
= e2πi k α (S (2 λ +α)+σm)
α mod k ℓ −λ≡ mod k
X h ′
X h ′
= e2πi k α (S α+σm) e2π k 2α S λ
α mod k λ mod k
0 otherwise
|T k (h, m)|2 ≤ 2r |S |r kr ,
45. Lecture 268
We shall now outline the main argument a little more skillfully, putting the 369
thing back on its track. Ar (n) is the sum of integrals over the finer-prepared
Farey arcs of Kloosterman:
1
k(k+N)
Z
1 P′ −2πi hk n e2πnz
Ar (n) = r/2 1/2 e
2 D 0≤h<k≤N zr/2
1
− k(k+N)
X π 2πi 1
− + D σ
Uk (h, s)Θ s e 2k2 z 4k k dϕ +
s mod 2D
2r/2 D1/2
1 1
k(k+ℓ) − k(k+ℓ+1)
N−1 Z N−1 Z
X h
X 1 X h
X
e−2πi k n + e−2πi k n ··· ,
h,k ℓ=k2
2r/2 D1/2 h,k ℓ=k1
1 1
k(k+ℓ+1) − k(k+ℓ)
where
X ′ −1
Θ s (x) = xm S m
;
m≡s (mod 2D)
= S 0 + S 2 + S 1 , say,
X X X
= S 00 + S 0m + S 20 + S 2m + S 10 + S 1m
m,0 m,0 m,0
2D
Corollary of Lemma 1. Uk (h, m) for k even depends on h only modulo Dk =
∧, say.
1
l(k+N)
Z
1 X h e2πnz
S 00 = r/2 1/2 e−2πi k n T k (h, o) dϕ
2 D 0≤h<k≤N zr/2
1
− k(k+N)
This goes into the principal term. We shall make it a little more explicit
later.
1
k(k+N)
Z − 2kπ2 z
1 P′ −2πi hk n
D
2πi 4kk σm′ S −1 m e m′ S −1 m
S om = r/2 1/2 e Uk (h, m)e dϕ
2 D 0≤h<k≤N zr/2
1
− k(k+N)
1
k(k+N)
N Z
1 X
= r/2 1/2 Kk (n, m) ··· ,
2 D k=1
1
− k(k+N)
where 371
P′ −2πi hk 2π 4ki Dk σm′ S −1 m
Kk (n, m) = e Uk (h, m)e
h mod k
1 X X 2πiahn.4+2πiV σ
= Uk (λ, m) e− 4k∗ ,
a λ mod ∧ h≡λ (mod ∧)
h mod 4k∗
∗
where 4k∗ = ak, a ≤ 4D, and V = kk Dk m−1 S −1 m
We defined σ by
hDk σ ≡ 1 (mod 4k∗ )
Let
DD̄k ≡ 1 (mod 4k∗ ), hH̄ ≡ 1 (mod 4k∗ )
Then
1 X X −2πiahn+2πiV σ
Kk (n, m) = Uk (λ, m) e 4k∗ Dk D̄k
a λ mod ∧ h≡λ (mod ∧)
h mod k∗ ,(h,k∗ )=1
1 X
e 4k (−4anh+V D̄k h̄)
P′ 2πi
= Uk (λ, m)
a λ mod ∧ h≡λ (mod ∧)
h mod k∗
The inner sum here is a Kloosterman sum. It has essentially 4k∗ terms. A 372
trivial estimate of this would be O(k), and this is what we had tacitly assumed
in the older method. The advantage here is, however, that they can be appraised
letter. We shall not estimate them here but only quote the result as
45. Lecture 270
Lemma 3.
X i
Kk (u, ν) = e2π k (uh+νh̄) , ∧ | k, hh̄ ≡ 1 (mod k)
h≡λ (mod ∧)
h mod k
= O k1−α+ǫ (u, k)α
There has been a lot of discussion about the size of the α in this formula.
Kloosterman and Esternann proved that α = 14 (Hamb, Ab. 1930), Salie’ that
α = 1/3 and A.Weil that α = 12 (P.N.A.S’, 48) Weil’s was a very complicated
and deep method going into the zeta-functions of Artin type and the Riemann
hypothesis for these functions.
We thus save a good deal in the order of magnitude. The further S’s will be
nearly similar; the complete Kloosterman sums will be replaced by sums with
certain conditions.
1
k(k+N) π
N Z −R ·1
(n, k)α π ′ −1 1
X e 2k2 z D
|S om | ≤ C k1−α+ǫ r/2 e− 4 (m S m− D ) dϕ
k=1
k |z|r/2
1
− k(k+N)
Since R k12 z ≥ 1
k on the Farey arc, the integrand is majorised by
π δN
r/4 π δN
− 2D 1 δN − 2D
δ−r/4 e k2 (δN +ϕ2 )
k2 |δ2 +ϕ2 2 2 −r/4 2
e N |k (δ2N + ϕ )| = N
2 2 2
k (δN + ϕ )
= O(nr/4 )
√
n
X π ′ −1 1
|S om | ≤ Cn r/4
k1−α+ǫ (n, k)α e− 4 m S m
√ ,
k=1
k n
√
since the path of integration has a length of the order 1/k n. Now summing 373
over all m , 0,
√
X X n
r 1
S om ≤ Cn 4 − 2 k−α+ǫ (n, k)α
m,0 k=1
r 1
X X
−
< Cn 4 2 dα (dt)−α+ǫ
√
d|n dt≤ n
r 1
X X
= Cn 4 − 2 dǫ t−α+ǫ
√
d|n t≤ n
d
45. Lecture 271
√ !1−α+ǫ
r 1
X n
< Cn 4 − 2 dǫ
d|n
d
r α ǫ
X
= Cn 4−2+2 dα−1 ,
d|n
All the other sums that we have to estimate behave some what similarly. We 375
take as specimen S 20 .
1
k(k+ℓ)
N−1 Z
1 P′ h 1 X e2πnz
S 20 = e−2πi k n T k (h, o) × r × dϕ
D1/2 2r/2 o≤h<k≤N k ℓ=k
zr/2
2 1
k(k+ℓ+1)
1
k(k+ℓ)
N N−1 Z
1 X 1 X e2πnz X h
= 1/2 r/2 e−2πi k n T k (h, o)dϕ
D 2 k=1 kr ℓ=N−k+1
zr/2 h mod k
1
k(k+ℓ+1) N−k<k2 ≤ℓ
The original interval for k2 was bigger: N − k < k2 ≤ N. Now the full
interval is not permissible, i.e., we have admitted not all residues modulo k, but
only a part of these, and the N may lie in two adjacent classes of residues.
Here we have a new type of sum of interest. We know how to discuss T k ; h
plays a role there. The sums we have now get are
X P′ h
T k (λ, o) e−2πi k n
λ mod ∧ h≡λ (mod ∧)
N−k<k2 ≤ℓ
272
46. Lecture 273
1), and the interesting fact is that they also permit the same appraisal, viz.
O kr/2 k1−α+ǫ (k, n)α
From there on things go just as smoothly as before.
1
√n k(N+1)
Z
X dϕ
S 2o = O k−r/2 k1−α+ǫ (k, n)α
k=1
(δ2N + ϕ2 )r/4
1
k(k+N)
and here for convergence of the integral we want r ≥ 3. This would give again
the old order. Similar estimates hold for the other pieces:
X α
S 2m = O nr/4 − + ǫ
m,o
2
Now
Z∞ Z∞
dϕ dϕ 1−r/2
2
(δN + ϕ2 )r/4
= ϕ 2 r/4 δN
1 1 1 + δN
kN kN
Z∞
r −1 dψ
= O n 2
2 r/4
(1 + ϕ )
N
k
and the integral being the Hankel integral for the gamma-function, 379
√
n
(2π)r/2nr/2 − 1 X Hk (n) 1
Ar (n) = 1/2 r/2 r r/2
+ O nr/4−α/2+ǫ
D 2 k=1
k Γ( )
∞
πr/2 X Hk (n)
= nr/2−1 r
+ O nr/4−α/2+ǫ
r
Γ 2 D1/2 k=1
k
46. Lecture 275
∞ r
2 +1−α+ǫ (k, n)α
r X k
−1
+ O n 2 r
√ k
k= n+1
and then we use the old appraisal (k1−α+ǫ (k, n)α with α = 1 + ǫ). So we have 380
√ !−r/2+1
r X n r 1 X
O n 2 −1 d1+ǫ−r/2 = O n 4 − 2 dǫ = O nr/4−1/2+2ǫ
d|n
d d|n
This is of smaller order than the old error term. So we have our final result:
∞
πr/2 r/2−1
X Hk (n)
r/4−α/2+ǫ
Ar (n) = n + O n ;
Γ(r/2)D1/2 k=1
kr
Hk (n) H p2 (n)
where γp = 1 + + +···
pr p2r
The arithmetical interpretation now becomes difficult, because all the prop- 381
erties that the quadratic form may have will have to show up. One or other of
the factors γ p may be zero in which case we have no representation.
46. Lecture 276
We should like to throw some light on the Kloosterman sums. We take for
granted the estimate
P′ i
e2π k (uh+V h̄) = O k1−α+ǫ · (k, u)α
h mod k
hh̄≡1 (mod k)
where
2πi j
1 e− k − e−2πi j(a+1)/k
cj = , j , k,
k 1 − e−2πi j/k
a
ck =
k
2
|c j | ≤
k sin π j/k
The sum becomes 384
k−1
X P′ 2πi P′ 2πi
cj e k (uh+(V + j)h̄)
+ ck e k (uh+V h̄)
Since sin α ≥ π2 ,
k−1
j
X 1 πX 1
2 ≤2
j=1 sin πkj 2 πj
k
46. Lecture 278
X 1
=k = O(k log k)
k−1
j
j≤ 2
Kloosterman first discussed his method for a diagonal quadratic form. Later
on he applied it to modular forms and for this he could derive on the investiga-
tions by Hecke comparing modular forms with Eisenstein series. In this case
the theory becomes simpler: we can subtract suitable Eisenstein series and the
principle term then becomes zero. The r-fold theta-series that we had are in
fact modular forms.