LN Dist
LN Dist
LN Dist
CHRISTIAN REMLING
1. Basic properties
The infamous Dirac δ-function is an object much beloved by theoret-
ical physicists. It cannot be given a mathematically sound definition
(that is, as a function), but rather is usually imagined as having the
values (
∞ x=0
δ(x) = ,
0 x 6= 0
and thenRone assumes that the singularity at x = 0 can be fine tuned
∞
to make −∞ δ(x) dx = 1.
Despite these inauspicious beginnings, the δ-function often seems a
convenient tool at least in formal calculations, and so it is tempting to
try to build a rigorous theory around the idea.
To do this, we take seriously the frequently heard excuse that the
“definition” is not to be taken literally, but that everything will make
sense when put under an integral sign. We can then in fact hope,
slightly more ambitiously, that
Z ∞
f (x)δ(x) dx = f (0)
−∞
that
∂α
(1.1) |u(ϕ)| ≤ C sup ϕ(x)
x∈U,|α|≤N ∂xα
for all ϕ ∈ D with supp ϕ ⊆ K.
We will be mostly interested in the case U = Rd , in fact most of the
time with d = 1, but sometimes the extra flexibility in the definition
will be useful.
I have used multi-index notation for partial derivatives, so α =
(α1 , . . . , αd ), and
∂ αf ∂ α1 ∂ αd
= · · · αd f (x1 , . . . , xd ).
∂xα ∂xα1 1 ∂xd
Recall that for well behaved functions such as f ∈ D, it does not matter
in which order these derivatives are taken, so this is well defined. We
also set |α| = α1 + . . . + αd .
A functional on a vector space X over C is, by definition, a map u :
X → C. The notation X 0 or X ∗ for the space of linear functionals (often
with extra continuity conditions imposed) is common. The action of
u ∈ D0 on a test function ϕ ∈ D is often written as (u, ϕ) instead of
u(ϕ). This suggests the viewpoint of a pairing between D and its dual
space D0 .
The supremum (really: maximum) on the right-hand side of (1.1)
is also denoted by kϕkN . Condition (1.1) can reformulated as follows,
and this indeed has the feel of a continuity condition. (In fact, it is
possible to put a topology T on D in such a way that (1.1) becomes
equivalent to continuity with respect to T and the standard topology
on C, but this T is rather inconvenient to work with, and it is not
useful at all for our purposes.)
Exercise 1.1. Let u : D → C be a linear functional. Show that the
following are equivalent: (a) u satisfies the condition from Definition
1.1; (b) If ϕn ∈ D, supp ϕn ⊆ K for some fixed compact set K ⊆ Rd ,
and kϕn kN → 0 for all N ≥ 0, then u(ϕn ) → 0.
Example 1.1. The motivating example, δ(ϕ) = ϕ(0), is a distribution
in this sense. Clearly, the map ϕ 7→ ϕ(0) is linear, and |ϕ(0)| ≤
sup |ϕ(x)| = kϕk0 , as required.
Example 1.2. The functional u(ϕ) = ϕ0 (0) (and here d = 1) also defines
a distribution, since u is obviously linear and |u(ϕ)| ≤ sup |ϕ0 (x)| ≤
kϕk1 .
DISTRIBUTIONS 3
P
Example 1.3. u(ϕ) = n∈Z ϕ(n) defines a distribution: Note, first of
all, that there are no convergence issues because ϕ(n) 6= 0 for only
finitely many n for any given ϕ ∈ D. Moreover, if supp ϕ ⊆ [−N, N ],
then |u(ϕ)| ≤ (2N + 1)kϕk0 .
Note that in the last example, the constant in estimate (1.1) does de-
pend on the support of the test function, as anticipated as a possibility
in the definition.
Exercise 1.2. Find an example of a distribution u (perhaps similar to
the one from Example 1.3) for which N from (1.1) also necessarily
depends on the support of test function.
If f ∈ L1loc (U ), then
Z
u(ϕ) = f (x)ϕ(x) dx
U
R
defines a distribution since if supp ϕ ⊆ K, then |u(ϕ)| ≤ kϕk0 K |f |.
Moreover, f can be recovered from the distribution it generates, up
to a change of its values on a null set. So we can identify f with
the distribution u = uf and think of a locally integrable function as a
distribution when this is convenient.
Theorem 1.2. Let f, g ∈ L1loc (U ) and suppose that uf = ug . Then
f = g a.e.
Exercise 1.3. Prove Theorem 1.2. If desired, you can focus on the
case U = Rd exclusively, though the general case isn’t much differ-
ent. Suggestion: It suffices to show that if uf = 0, then f = 0 a.e.
Take convolutions of f with suitable functions and interpret these as
applications of uf to test functions.
Exercise 1.4. Show that Lp ⊆ L1loc for all p ≥ 1.
Given a general u ∈ D0 , it is sometimes interesting to ask if u = uf
for some f ∈ L1loc or, as this is often put, if u ∈ D0 is a function.
Exercise 1.5. Show that δ ∈ D0 is not a (locally integrable) function in
this sense.
More generally, if µ is a Borel measure on Rd (and this, as usual,
includes the requirement
R that µ(K) < ∞ for all compact sets K ⊆ Rd ),
then u(ϕ) = f dµ is another distribution. Again, µ can be recovered
from u, and thus we can think of measures as distributions if desired.
Example 1.4. Define
Z
1 ϕ(x)
(1.2) PV , ϕ = lim dx.
x h→0+ |x|>h x
4 CHRISTIAN REMLING
Here, PV stands for principal value, and what is meant by this is the
regularization of the integral by omitting the neighborhood (−h, h) of
the singularity at x = 0 and then sending h → 0. Some such device is
needed since ϕ(x)/x is not integrable if ϕ(0) 6= 0.
Before we can use (1.2) as a definition, we must in fact establish that
the limit exists. By the mean value theorem, ϕ(x) = ϕ(0) + ϕ0 (ξ)x,
for some ξ = ξ(x) between 0 and x. If we also fix an L > 0 with
supp ϕ ⊆ [−L, L], then we can write the integral as
Z Z
dx
ϕ(0) + ϕ0 (ξ) dx.
h<|x|<L x h<|x|<L
TheR first integral equals zero for all h > 0, and the second one converges
L
to −L ϕ0 (ξ) dx, by DC. Thus PV(1/x) is well defined, and we also
conclude that |(PV(1/x), ϕ)| ≤ 2Lkϕk1 . So we have indeed defined a
distribution.
Exercise 1.6. Show that PV(1/x) is not a function. Suggestion: If it
were, what would this function have to be equal to away from x = 0?
2. Operations on distributions
If f ∈ C 1 (R), then its derivative f 0 , being continuous, is a locally
integrableR function, so may be viewed as a distribution, which acts as
∞
(f 0 , ϕ) = −∞ f 0 (x)ϕ(x) dx. By an integration by parts, we can rewrite
this as Z ∞
0
(f , ϕ) = − f (x)ϕ0 (x) dx.
−∞
Exercise 2.1. Prove this in more detail. Why are there no boundary
terms?
This motivates:
Definition 2.1. Let u ∈ D0 (R). The distributional derivative u0 ∈ D0
is defined as the distribution (u0 , ϕ) = −(u, ϕ0 ).
We must in fact show that this indeed defines a new distribution,
but this is easy: if K ⊆ R is compact and |(u, ϕ)| ≤ CkϕkN for ϕ ∈ D
with supp ϕ ⊆ K, then |(u0 , ϕ)| ≤ CkϕkN +1 .
In the same way, we can more generally define partial derivatives of
distributions u ∈ D0 (U ), U ⊆ Rd as (∂u/∂xj , ϕ) = −(u, ∂ϕ/∂xj ).
Note that any distribution has (distributional) derivatives, in fact of
any order, since the derivatives are themselves distributions to which
the definition can be applied. For example, (δ 0 , ϕ) = −(δ, ϕ0 ) = −ϕ0 (0)
and, more generally, δ (n) (ϕ) = (−1)n ϕ(n) (0).
DISTRIBUTIONS 5
2 2
Exercise 2.2. Show that u ∈ D0 (U ) satisfies ∂x∂j ∂x
u
k
= ∂x∂k ∂x
u
j
(that is,
higher order partial distributional derivatives can be taken in any or-
der).
Example 2.1. The standard calculus example for a function that is not
differentiable at a point is f (x) = |x|. Clearly, f ∈ L1loc , so f does
have a distributional derivative, which we’ll denote simply by f 0 (but
perhaps more circumspect notation would have been u0f ). We compute
Z ∞
0 0
(f , ϕ) = −(f, ϕ ) = − |x|ϕ0 (x) dx
−∞
Z 0 Z L
0
= xϕ (x) dx − xϕ0 (x) dx
−L 0
Z 0 Z L
0 L
= xϕ(x) − −L
ϕ(x) dx − xϕ(x) 0
+ ϕ(x) dx
−L 0
Z ∞
= sgn(x)ϕ(x) dx,
−∞
with sgn(x) = 1 for x > 0 and sgn(x) = −1 for x < 0. We also took
L > 0 large enough so that supp ϕ ⊆ [−L, L] in this calculation. We
have found, unsurprisingly, that f 0 ∈ D0 is in fact a function, and it is
equal almost everywhere to the pointwise derivative f 0 (x) = sgn(x) of
f.
6 CHRISTIAN REMLING
We now again split this last integral into the two parts x < −h and
x > h and then integrate by parts in those. This produces
Z
0 ϕ(x)
(f , ϕ) = lim dx + log h(ϕ(h) − ϕ(−h)) ,
h→0+ |x|>h x
and here the second term goes to zero because ϕ(±h) = ϕ(0) + O(h)
and h log h → 0.
Let’s summarize:
Theorem 2.2. In D0 , |x|0 = sgn(x), |x|00 = 2δ, (log |x|)0 = PV(1/x).
As our next example, let’s take an increasing function F : R → R.
Such an F is bounded on any compact set, so is clearly locally integrable
and thus generates a distribution. What is its distributional derivative?
Before we answer this, recall that a monotone F is differentiable, in
the classical sense, at almost all x ∈ R. Moreover, F generates a Borel
measure µ = mF , by setting µ((a, b]) = F (b) − F (a), if we also make
F right-continuous here. This changes F at at most countably many
points, so will not affect the distribution F ∈ D0 , and we may thus
indeed insist on this normalization at no cost. It will also be convenient
to make F (0) = 0, by adding a constant to the original function. It
seems intuitively clear that this should not affect F 0 , but let’s establish
this carefully before we proceed.
Theorem 2.3. Let u, v ∈ D0 , c ∈ C. Then (u+v)0 = u0 +v 0 , (cu)0 = cu0 ,
and c0 = 0.
DISTRIBUTIONS 7
Let
R me look at the last integral in more detail: we can write µ((0, x]) =
(0,x]
dµ(t). We obtain an iterated (double) integral, to which we apply
Fubini-Tonelli. Note that the integration is over the set 0 < t ≤ x, so
this produces
Z Z ∞ Z
0
− dµ(t) dx ϕ (x) = ϕ(t) dµ(t).
(0,∞) t (0,∞)
Of course, the Rfirst integral from (2.1) can be given a similar treatment:
it is equal to (−∞,0] ϕ(t) dµ(t). Putting things together, we thus see
that (F 0 , ϕ) = R ϕ(x) dµ(x). In other words, F 0 = µ, the distribution
R
generated by the measure that F induces. Let’s state this formally.
Theorem 2.4. If F : R → R is increasing, then F 0 = µ, with
µ((a, b]) = F (b+) − F (a+).
In particular, note that the distributional derivative need not equal
the pointwise derivative. This holds only if µ is absolutely continuous.
If µ has a singular part, then F 0 is not a function. (You may wonder
how a measure can ever be a function, but in fact it happens, with our
use of terminology, since the measure f dx and the function f are the
same distribution, so in that sense the measure f dx is a function also.)
We in fact already saw this principle in action in the simple example
F (x) = sgn(x), which gave F 0 = 2δ (and this of course is a measure,
though we didn’t emphasize this originally.)
This condition, of having a positive measure as its distributional de-
rivative, characterizes the increasing functions. This criterion is not
particularly useful (usually there will be easier ways to check that a
given function is increasing than computing its distributional deriva-
tive), but it is elegant and satisfying.
8 CHRISTIAN REMLING
and ψ(x) = 0 for all small x. Our assumption on ϕ makes sure that
ψ(x) = 0 also for all large x. So ψ has compact support.
Proof of Theorem 2.6. Fix once and for all a ϕ0 ∈ C0∞ withR ϕ0 = 1.
R
∞
R arbitrary ϕ ∈ C0 , write ϕ = aϕ
For 0 + (ϕ − aϕ0 ), with a = ϕ. Then
(ϕ − aϕ0 ) = 0, so ϕ − aϕ0 = ψ 0 for some ψ ∈ C0∞ , by the lemma.
However, (u, ψ 0 ) = −(u0 , ψ) = 0, and thus
Z ∞
(u, ϕ) = a(u, ϕ0 ) = (u, ϕ0 ) ϕ(x) dx.
−∞
as claimed.
Proof. We have
1 ϕ(x + h − t) − ϕ(x − t)
((u ∗ ϕ)(x + h) − (u ∗ ϕ)(x)) = u, ,
h h
and here we apply u to the difference quotient as a function of t, for
fixed x, h. If we now send h → 0, then this expression will converge
to ϕ0 (x − t). This is clear in pointwise sense, but this isn’t quite good
enough here. Rather, what we need is that the functions
ϕ(x + h − t) − ϕ(x − t)
gh (t) = − ϕ0 (x − t) ∈ C0∞
h
together with all their derivatives converge to zero uniformly in t and
their supports are contained in one fixed compact set. This is condition
(b) from Exercise 1.1, and by the result of that Exercise, we can then
conclude that (u, gh ) → 0 as h → 0.
Exercise 2.6. Prove this claim about gh in more detail.
It follows that u ∗ ϕ is indeed differentiable, and (u ∗ ϕ)0 = u ∗ ϕ0 .
Since this is still of the same general form (convolution of a distribution
with a test function), we can repeat this argument, and we have in fact
shown that u ∗ ϕ ∈ C ∞ .
Finally,
u0 ∗ ϕ = (u0 , τx Rϕ) = −(u, (τx Rϕ)0 ) = (u, τx Rϕ0 ) = u ∗ ϕ0 .
Definition 2.18. Let g ∈ C ∞ (U ), u ∈ D0 (U ). Then gu ∈ D0 is defined
as the distribution (gu, ϕ) = (u, gϕ).
R
R This definition is of course motivated by the trivial formula (gu)ϕ =
u(gϕ), which is valid for nice functions u. Note that we have defined
the product of a distribution and a (smooth) function; there is no natu-
ral notion of a product of two distributions. This is not surprising. For
example, it is indeed hard to come up with a reasonable interpretation
of what δ · δ might be.
Exercise 2.7. Show that gu indeed is a distribution.
Exercise 2.8. Show that the product rule is valid in the (limited) con-
text of Definition 2.18: if g ∈ C ∞ , u ∈ D0 , then (gu)0 = g 0 u + gu0 .
Example 2.4. What is gδ? We compute (gδ, ϕ) = (δ, gϕ) = g(0)ϕ(0).
This is what the (constant) multiple g(0)δ would have done on ϕ, so
gδ = g(0)δ.
Exercise 2.9. Compute similarly gδ 0 , for g ∈ C ∞ (R).
14 CHRISTIAN REMLING
3. Convergence of distributions
Definition 3.1. Let un , u ∈ D0 . We say that un → u (in D0 , or in the
sense of distributions) if (un , ϕ) → (u, ϕ) for all ϕ ∈ D.
This is a rather weak requirement. For example, einx → 0 in D0 , if
we view these functions as distributions.
Exercise 3.1. Prove this. Suggestion: Riemann-Lebesgue lemma
Can you in fact also show that nN einx → 0 in D0 for any N ≥ 1?
Similarly, for example nδn → 0, where (δn , ϕ) = ϕ(n). To confirm
this, simply observe that (nδn , ϕ) = nϕ(n) will be zero once n is outside
the support of ϕ.
Example 3.1. Let fn (x) = nχ(−1/2n,1/2n) (x). Then fn → δ. To prove
this, we use the mean value theorem to write ϕ(x) = ϕ(0) + ϕ0 (ξ)x =
ϕ(0) + O(x) and compute
Z 1/2n Z 1/2n
(fn , ϕ) = n ϕ(x) dx = ϕ(0)+n O(1/n) dx = ϕ(0)+O(1/n),
−1/2n −1/2n
After these easy introductory examples, we are now ready for a more
substantial convergence statement.
Theorem 3.3 (Sokhotski-Plemelj formula). In D0 (R),
1 1
lim = PV + iπδ.
h→0+ x − ih x
Proof. We write
1 x h
= 2 2
+i 2 ,
x − ih x +h x + h2
and we’ll treat the real and imaginary parts separately. The imaginary
part is actually covered by (the continuous version of) Exercise 3.2
because, in the notation used there,
R h/(x2 + h2 ) = f1/h (x), with f (x) =
1/(1 + x2 ), and (by calculus) f = π. But a direct argument is also
easy:
Z ∞ Z ∞ Z ∞
h 1 dt
2 2
ϕ(x) dx = 2
ϕ(ht) dt → ϕ(0) 2
= πϕ(0)
−∞ x + h −∞ 1 + t −∞ 1 + t
The convergence follows from DC, and in this step we use that ϕ is
continuous and bounded.
Next, we look at the real part. Since we anticipate the limit being
the principal value distribution, we first establish that we can remove
a small interval about x = 0. More precisely,
Z h Z h Z h
x x x
2 2
ϕ(x) dx = ϕ(0) 2 2
dx + 2 2
O(x) dx,
−h x + h −h x + h −h x + h
and this goes to zero because the first integral on the right-hand side
has this value, and the second one is O(h).
So it now suffices to show that
Z
x 1
lim − ϕ(x) dx = 0.
h→0+ |x|>h x2 + h2 x
To do this, we write
x 1 −h2
− =
x2 + h2 x (x2 + h2 )x
and again use the Taylor expansion ϕ(x) = ϕ(0) + O(x) of the test
function. The term with ϕ(0) doesn’t contribute
R since the integrand is
odd, and what is then left is of order O(h2 ) |x|>h dx/(x2 + h2 ). Since
R
R
h/(x2 + h2 ) dx = π, this goes to zero.
Exercise 3.3. Show that log(x2 + h2 ) → 2 log |x| in D0 .
16 CHRISTIAN REMLING
A slightly slicker proof of the second part can be based on this Ex-
ercise and Theorem 3.2: Take derivatives to obtain
(log(x2 + h2 ))0 → 2(log |x|)0 .
Now recall that this latter distribution equals 2(PV(1/x)); see Theorem
2.2. Moreover, log(x2 + h2 ) is a smooth function for fixed h > 0, so its
distributional derivative is its classical derivative, and thus
2x
(log(x2 + h2 ))0 = 2 .
x + h2
We earlier defined convolutions, and one of their main uses (in the
classical setting) is the approximation of general functions by nice ones.
The convolution from Definition 2.16 can be put to similar use in the
realm of distributions.
Theorem 3.4. Let u ∈ D0 (R), and fix a ϕ ∈ C0∞ (R) with ϕ = 1.
R
Put ϕn (x) = nϕ(nx). Then u ∗ ϕn → u in D0 .
We originally defined u ∗ ϕ as a function (not a distribution), but
we then saw that this is a smooth function, so is in particular locally
integrable and can be viewed as a distribution after all. This is of
course the interpretation of u ∗ ϕn that we need to adopt here, for the
statement to make sense.
So we see that any distribution can be approximated by smooth
functions. We can do even better:
Exercise 3.4. (a) Fix a ψ ∈ C0∞ (R) with ψ(x) = 1 in a neighborhood
of x = 0, and let ψn (x) = ψ(x/n). Show that ψn u → u for any u ∈ D0 .
(b) Combine this with Theorem 3.4 to conclude that for any u ∈
D (R), there are ϕn ∈ C0∞ (R), ϕn → u in D0 .
0
4. Tempered distributions
The Fourier transform of an f ∈ L1 (Rd ) is defined as
Z
(4.1) fb(t) = f (x)e−2πit·x dx.
Rd
We would now like to define the Fourier transform u b also for distribu-
tions u ∈ D0 . We follow the usual strategy of moving the operation
over to the test function, taking the case of a nice function u as our
18 CHRISTIAN REMLING
So S is the space of all smooth functions that together with all their
derivatives decay faster than any power. Clearly, C0∞ ⊆ S, but S also
contains (many) not compactly supported functions such as f (x) =
2
e−x .
The norms k · kj,n can be combined into a metric d, for example as
follows:
X kf − gkj,n
d(f, g) = 2−j−n
j,n≥0
1 + kf − gkj,n
Exercise 4.1. Show that d(fk , f ) → 0 if and only if kfk − f kj,n → 0 for
all j, n ≥ 0.
That we have made the right choice is confirmed by
Theorem 4.1. The Fourier transform is a continuous bijection on S.
I don’t want to prove this in detail here, so will just make a few
general remarks. The proof is straightforward, but somewhat tedious
to write down. Essentially, we have to show that the Fourier transform
maps S continuously back into itself; the rest will then follow quickly
from Fourier inversion. The two defining properties, rapid decay and
smoothness, are dual to each other in the sense that one will become
the other after taking Fourier transforms, so the combination of both
is preserved. If this is done carefully, we will then also obtain control
on kfbkj,n in terms of the kf kj 0 ,n0 , and this gives the asserted continuity
of the Fourier transform.
DISTRIBUTIONS 19
and this gives the correct answer, but is not a formally correct ma-
nipulation because we tried to apply δ to e−2πitx , which is not a test
function.
Now let’s take a closer look at the relation between S 0 and D0 . We
already observed that D ( S. Because of this, any tempered distri-
bution u : S → C can be restricted to D (and don’t confuse this type
of restriction with the one from Definition 2.12). Then u|D will lie in
D0 : By assumption, we have |u(ϕ)| ≤ Ckϕkj,n for some j, n. If now
also ϕ ∈ D, supp ϕ ⊆ B(0, R), then kϕkj,n ≤ (1 + Rj )kϕkn , and this
latter norm is the one from Definition 1.1. This verifies that u ∈ D0 ,
as claimed; more precisely, it is the restriction u|D that is in D0 , but
we will often not make this distinction explicit in the notation.
Moreover, the original tempered distribution u ∈ S 0 can be recovered
from its restriction u|D ∈ D0 . This follows from the fact that D ⊆ S is
dense in S with respect to d.
Exercise 4.3. Prove this. More explicitly, suppose that ϕ ∈ S, and
then show that there are ϕk ∈ C0∞ such that kϕk − ϕkj,n → 0 for each
j, n ≥ 0. Suggestion: Fix a ψ ∈ C0∞ with ψ = 1 near x = 0 and then
try ϕk (x) = ψ(x/k)ϕ(x).
However, if we conversely start out with a u ∈ D0 , then this may
or may not have an extension to a tempered distribution u0 ∈ S 0 ; if
it does, then u0 is unique, as we just discussed. As an example for
such a u ∈ D0 with no continuousRlinear extension, we can consider
the function u(x) = ex . So u(ϕ) = ex ϕ(x) dx, and it is already clear
intuitively what will go wrong here: there seems to be only one natural
way to attempt an extension to S, namely, try to use this formula for
general ϕ ∈ S also, but then such a function is not guaranteed to have
enough decay to make the integral convergent.
To make a proof out ofR this intuition, we need to proceed differently.
Fix a ψ ∈ C0∞ (R) with ψ 6= 0, and let ϕn (x) = ψ(x − n)e−x ∈ C0∞ ⊆
S. Then kϕn kj,N → 0.
Exercise 4.4. Prove this in detail.
R R
However, u(ϕn ) = ψ(x − n) dx = ψ(x) dx does not converge to
zero. So u has no continuous extension to S.
What we actually showed is that u is already discontinuous on D,
at ϕ = 0, when we use the metric d on D. So the general message is
that the continuity requirement on a tempered distribution u ∈ S 0 is
stronger than the one on a u ∈ D0 (when only test functions ϕ ∈ D are
considered). Let’s summarize:
DISTRIBUTIONS 21
While distributions that do not grow too fast near infinity will be
tempered, the converse is not true. Consider the function f (x) =
ex cos ex . This does become large (some of the time, at least) for large
x. However, we have f ∈ S 0 anyway, and this follows from Theorem
4.8 because f = g 0 , with g = sin ex ∈ S 0 .
Definition 4.9. Let un , u ∈ S 0 . We say that un → u in S 0 if (un , ϕ) →
(u, ϕ) for all ϕ ∈ S.
DISTRIBUTIONS 23
(d) as an illustration:
((τa u)b, ϕ) = (τa u, ϕ)
b = (u, τ−a ϕ)
b
This test function equals (e−2πiax ϕ)b, by Proposition 4.1(e), and un-
packing this, we then obtain
(u, (e−2πiax ϕ)b) = (b
u, e−2πiax ϕ) = (e−2πiax u
b, ϕ).
Since this is true for any ϕ ∈ S, we have (τa u)b= e−2πiax u
b, as claimed.
Only part (f) presents some technical challenges, similar actually to
the ones from the proof of Theorem 3.4. Let ψ ∈ S be an arbitrary
test function. Then
Z
((u ∗ ϕ)b, ψ) = (u ∗ ϕ, ψ) = ψ(x)(u,
b b ϕ(x − t)) dx.
R
I will again rewrite this as (u, ψ(x)ϕ(x−t)
b dx) without fully justifying
R
this step. We can then interpret ψ(x)ϕ(x − t) dx = (ψb ∗ Rϕ)(t), and
b
Fourier inversion says that the inverse operation of b is R b. Thus
ψb ∗ Rϕ = fb, with
Theorem 4.12 right away and get around this difficulty by observing
that
1 1
lim χ(−L,L) (x) =
L→∞ x − ih x − ih
0
in S ; this is a routine application of DC, but do it in more detail
please if it’s not immediately clear to you. Now this cut off function is
integrable, so its Fourier transform is given by
Z L −2πitx
e
(4.4) dx.
−L x − ih
A similar calculation is possible for t > 0, and in this case we find that
Z L −2πitx
e
lim dx = 0
L→∞ −L x − ih
(because we now close the contour in the lower half plane, to make the
exponential small, and there will be no singularities inside our region
this time).
So, summing up, we have shown that
Z L −2πitx (
e 2πie2πht t < 0
lim dx = .
L→∞ −L x − ih 0 t>0
On reflection, we have actually shown this as a pointwise limit, but
this is not quite what we need here because what we know is that the
cut off integrals will converge to the Fourier transform of 1/(x − ih)
in S 0 . However, it is easy to go over the argument one more time and
confirm that this follows, too. We have shown that
b
1
(t) = 2πiχ(−∞,0) (t)e2πht .
x − ih
26 CHRISTIAN REMLING
This or also more general versions of the statement can be and usu-
ally are proved directly.
δn ∈ S 0 .
P
Exercise 4.11. Prove that
P
Proof of Theorem 4.15. Clearly, u = δn has the following two invari-
ance properties: τn u = u and e2πinx u = u, for all n ∈ Z. By taking
Fourier transforms, this implies that v = u b has the same two properties.
In particular, (e2πix − 1)v = 0, and this implies that supp v ⊆ Z, by
arguing as follows: Let V ⊆ R be an open set with V ∩ Z = ∅, and let
ϕ ∈ D be arbitrary with supp ϕ ⊆ V . Since supp ϕ is a compact subset
of V , it has positive distance to V c , and thus ψ = ϕ/(e2πix − 1) ∈ D
also. Thus
(v, ϕ) = (v, (e2πix − 1)ψ) = ((e2πix − 1)v, ψ) = 0.
We have shown that V ∩ supp v = ∅, and this holds for any open V
with V ∩ Z = ∅, thus supp v ⊆ Z, as claimed.
Consider now the restriction v0 of v to (−3/4, 3/4), say. Actually,
we want an element of D0 (R), not of D0 (−3/4, 3/4), so we really take
v0 = ψv with a v ∈ C0∞ with supp ψ ⊆ (−1, 1), ψ = 1 on (−3/4, 3/4).
(The purpose is to make Theorem 4.14 applicable, but this is a technical
point and not essential; the argument would also work with the actual
restriction of v, if Theorem 4.14 is slightly adapted instead.)
We have supp v0 ⊆ {0}, so
N
X
(4.8) v0 = cj δ (j) ,
j=0
by Theorem 4.14. We can say more here: in fact, v0 = cδ. To see this,
assume that N ≥ 1 and consider a trigonometric polynomial p(x) =
2πinx
P
|n|≤L an e , and here we want to choose L ≥ 1 and then the an ∈ C
such that p(0) = 1, p(j) (0) = 0 for 1 ≤ j ≤ N − 1, and p(N ) (0) 6= 0.
Exercise 4.12. Show that this is possible.
Finally, we build a test function by multiplying p by a ϕ ∈ C0∞ .
Then
(ϕp)(j) (0) = ϕ(j) (0) (j < N ), (ϕp)(N ) (0) = ϕ(N ) (0) + p(N ) (0)ϕ(0),
and thus, if v0 is as in (4.8) and invariant under multiplication by e2πinx ,
as we know it is, then
(v0 , ϕ) = (pv0 , ϕ) = (v0 , pϕ)
N
X
= cj (−1)j ϕ(j) (0) + cN (−1)N p(N ) (0)ϕ(0).
j=0
DISTRIBUTIONS 29
This leaves only the case p = 1/2, which we can handle by a limiting
argument, sending p → 1/2, p 6= 1/2.
Finally, this argument also works in higher dimensions if we establish
one additional symmetry before entering the final part of the argument:
u
b is spherically symmetric.
Exercise 4.16. Give a precise definition of what is meant by this final
claim, and then prove it.
With additional trickery, the constants cp,d can actually be identified.
One finds
Γ((d − p)/2)
(4.9) cp,d = π −d/2+p .
Γ(p/2)
The differential operator
∂2 ∂2
∆= + . . . +
∂x21 ∂x2d
is called the Laplacian. The corresponding partial differential equa-
tion ∆u = 0 is called the Laplace equation, and its solutions are also
called harmonic functions or distributions, if we admit distributional
solutions.
Theorem 4.18. Let u ∈ S 0 , ∆u = 0. Then u is a (harmonic) polyno-
mial.
The assumption that the distribution is tempered is crucial here.
The theorem does not hold for general distributions, and in fact there
are many counterexamples that are smooth functions such as u(x, y) =
ex cos y.
Exercise 4.17. Find all u ∈ D0 (R) with ∆u = 0.
Corollary 4.19 (Liouville). A bounded harmonic function is constant.
Proof. Such a function u can be viewed as a tempered distribution, so
is a polynomial by Theorem 4.18, but the only bounded polynomials
are the constants.
In the same way we can establish the more general statement that a
harmonic function u of at most polynomial growth, |u(x)| . (1 + |x|)N ,
actually is a polynomial of degree at most N .
Proof of Theorem 4.18 (sketch). Take Fourier transforms in ∆u = 0
and use Theorem 4.11 to conclude that
(4.10) −4π 2 |t|2 u
b = 0.
32 CHRISTIAN REMLING
Note that |t|2 = t21 + . . . + t2d is a smooth function satisfying the as-
sumptions of Theorem 4.8, so |t|2 u b ∈ S 0 is well defined.
Since 1/|t|2 is smooth away from t = 0, (4.10) implies that supp u ⊆
{0}. We now need the multi-dimensional analog of Theorem 4.14.
Assuming this, we obtain
X ∂α
u
b= cα δ,
∂tα
|α|≤N
and
P then we|α| take (inverse) Fourier transforms to deduce that u =
cα (−2πi) xα .
Proof. To motivate what the theorem says is true, we can again take
Fourier transforms in ∆G = δ as in the proof of Theorem 4.18, to see
that (in S 0 ) −4π 2 |t|2 G b = −1/(4π 2 |t|2 ).
b = 1. This suggests to try G(t)
Theorem 4.17 then shows that indeed G(x) = −cd |x|2−d , with cd =
1/(4π 2 cd−2,d ).
Exercise 4.18. The way I wrote it up, the argument doesn’t quite follow
the required logic. We should really define G(x) as in the theorem and
then check that ∆G = δ. Clean up the argument by reorganizing it
along these lines, and also derive the formula for cd from (4.9).
DISTRIBUTIONS 33
Since
ϕ(−h) + ϕ(h) − 2ϕ(0)
→0
R h
and |x|>h dx/x2 = 2/h, we can rewrite this as
0
ϕ(0) − ϕ(x)
Z
1
PV , ϕ = lim dx.
x h→0+ |x|>h x2
We summarize:
34 CHRISTIAN REMLING
Theorem 4.21.
ϕ(x) − ϕ(0)
Z
(u, ϕ) = lim dx
h→0+ |x|>h x2
defines a (tempered) distribution. We have u = −2π 2 |t|
b and also u =
0
−(PV(1/x)) .