2 Linearoperators
2 Linearoperators
c1 v1 + c2 v2 + . . . + ck vk
where the c’s are just (real or complex) scalars. One property of vector spaces is that any linear
combination of elements in S is also in S. This is easily verified in most cases - for example, Rn
(the set of n-dimensional vectors) and C 0 (the set of continuous functions) are vector spaces.
More generally, if f1 , f2 are elements of a vector space S, then a linear operator L is a mapping
from S to some other vector space (frequently also S) so that
Av = λv.
Except in rare cases (for so called “non-normal” matrices), there are exactly n linearly indepen-
dent eigenvectors v1 , v2 , . . . , vn corresponding to eigenvalues λ1 , λ2 , . . . , λn (the eigenvalues can
sometimes be the same for different eigenvectors, however). Why is this a useful fact? Note that if
x is any n-dimensional vector, it can be written in terms of the basis given by the eigenvectors, i.e.
x = c1 v1 + c2 v2 + . . . + cn vn . (3)
1
The ci solve a linear system Vc = x where V has columns of eigenvectors and c is the vector
(c1 c2 . . .)T . When A is self-adjoint (see below), there is a much easier way to find the coefficients.
Note that action of a linear transformation A on the vector x can be written simply as
Ax =A(c1 v1 + c2 v2 + . . . + cn vn )
=c1 Av1 + c2 Av2 + . . . + cn Avn
=c1 λ1 v1 + c2 λ2 v2 + . . . + cn λvn .
In other words, eigenvectors decompose a linear operator into a linear combination, which is a fact we
often exploit.
2
1.5 Self-adjointness and expansion in eigenvectors
Sometimes an operator is its own adjoint, in which case its called self-adjoint. Self-adjoint operators
have some very nice properties which we will exploit. The most important are
1. The eigenvalues are real.
2. The eigenvectors corresponding to different eigenvalues are orthogonal.
Suppose a matrix A is symmetric and therefore self-adjoint, and we know its eigenvectors. As
in (3), we can try to write any vector x as a linear combination of the eigenvectors. Taking the
inner product of (3) with any particular eigenvector vk and using (4), we have
hx, vk i = hc1 v1 + c2 v2 + . . . + cn vn , vn i
= c1 hv1 , vk i + c2 hv2 , vk i + . . . + c1 hvn , vk i (7)
= ck hvk , vk i
since vk is orthogonal to all eigenvectors except itself. Therefore we have a simple formula for any
coefficient ck :
hx, vk i
ck = . (8)
hvk , vk i
In many cases the eigenvectors are rescaled or normalized so that hvk , vk i = 1, which means that
(8) simplifies to ck = hx, vk i.
Lf = 0, (homogeneous equations),
Lf = g, (inhomogeneous equations),
3
where f is the “solution” (the function to be found), L is some differential linear operator, and
g is another given function. As a rule of thumb, identifying a linear equation is just a matter of
making sure that each term in the equation is a linear operator acting on the unknown function,
or something which is already known.
This basic idea can be amended for inhomogeneous equations. In this case, one needs to find
any single particular solution fp which solves Lfp = g, so that the difference h = f − fp solves a
homogeneous equation Lh = 0. The result is
d d
L= p(x) + q(x). (9)
dx dx
(the notation means that to apply L to a function f (x), we put f on the right hand side and
distribute terms so Lf = (p(x)f 0 )0 + q(x)f (x)) The functions p and q are prescribed, and in the
4
most basic cases are simply constant. The space L acts on will be C0∞ [a, b], which means that
f (a) = f (b) = 0 (zero boundary conditions) and f is infinitely differentiable on the interval [a, b]
where it lives.
We can now show that the Sturm-Liouville operator (9) acting on C0∞ [a, b] is self-adjoint with
respect to this inner product. How does one compute the adjoint of a differential operator? The
answer lies in using integration by parts (or in higher dimensions Green’s formula). For any two
functions f, g in C0∞ [a, b] we have
Z b
d df
hLf, gi = p(x) g(x) + q(x)f (x)g(x)dx
a dx dx
Z b
df dg df b
= −p(x) + q(x)f (x)g(x)dx + p(x)g(x)
a dx dx dx a
Z b
d dg dg b
= p(x) f (x) + q(x)f (x)g(x)dx − p(x)f (x)
a dx dx dx a
= hf, Lgi.
Integration by parts was used twice to move derivatives off of f and onto g. Therefore (compar-
ing to (6) ) L is its own adjoint. Note that the boundary conditions were essential to make the
boundary terms in the integration by parts vanish.
As another example, consider the weighted inner product on C0∞ [0, R] defined by
Z R
hf, gi = rf (r)g(r)dr. (11)
0
Then for the linear operator Lf = r−1 (rf 0 )0 , integration by parts twice gives
Z R df
−1 d
hLf, gi = r r r g(r)dr
0 dr dr
Z R
df dg
=− r dr, (int. by parts)
0 dr dr
Z R
d dg
= r f (r) dr, (int. by parts)
0 dr dr
Z R dg
−1 d
= r r r f (r) dr
0 dr dr
= hf, Lgi.
Note in the final step, a weight factor of r was needed inside the integral to obtain the inner
product (11). Thus L is self adjoint with respect to the weighted inner product. It turns out not to
be self adjoint, however, with respect to the unweighted one (10).
5
What about non-self-adjoint operators? The simplest example is L = d/dx acting on C0∞ [a, b].
We again compute
Z b Z b
df dg
hLf, gi = g(x) = − f (x)dx = hf, −Lgi.
a dx a dx
It follows that L† = −L. Operators which are their negative adjoints are called skew symmetric.
It should be noted that sometimes the eigenvalue problem is written instead like Lv(x)+λv(x) = 0
(which reverses the sign of λ), but the theory all goes through just the same. The functions v(x)
which satisfy this are called eigenfunctions and each corresponds to an eigenvalue λ. As with
eigenvectors, we can rescale eigenfunctions: if v(x), λ is an eigenvector-value pair, so is (cv(x), λ)
for any scalar c 6= 0.
What can we expect to happen in these problems? Since we are working in infinite dimensions,
there are often an infinite number of eigenvalues and eigenfunctions. The set of eigenvalues is
also called the discrete spectrum (there is another part called the essential or continuous spectrum
which we will not get into). The eigenfunctions constitute a linearly independent set, so one
might think that any function can be written as a (infinite) linear combination of eigenfunctions,
analogous to (3). This is not always the case, but when it does happen, the set of eigenfunctions is
called complete. This property is extremely valuable by allowing functions to be written as linear
combinations of the eigenfunctions. The most famous example is the Fourier series, which we
discuss later.
Self-adjoint operators have some properties equivalent to self-adjoint matrices. In particular,
their eigenvalues are real, and their eigenfunctions are orthogonal. Orthogonality is taken with
respect to the same inner product which gives self-adjointness.
2. The eigenfunctions vn (x) corresponding to each eigenvalue λn form a complete set, i.e. for
any f ∈ C0∞ [a, b], we can write f as a (infinite) linear combination
∞
X
f= cn vn (x).
n=1
6
A technical point: in what sense does the series converge? For functions, it should be noted
that there are different ways convergence can be defined. One way is just to fix each point x and
consider the limit of function values, which is called “pointwise” convergence. We actually have
something stronger here: the rate at which the sequences at all x converge is the same. This is
called uniform convergence.
Here is the simplest example of a Sturm-Liouville eigenvalue problem. Consider the operator
L = d2 /dx2 on the vector space C0∞ [0, π]. The eigenvalue problem reads
d2 v
+ λv = 0, v(0) = 0, v(π) = 0. (13)
dx2
This is just a second order differential equation, and writing down the general solution is easy.
√
Recall that we “guess” solutions of the form v(x) = exp(rx). Provided λ > 0, we get r = ±i λ,
which means that the general solution has the form
√ √
v(x) = A cos( λx) + B sin( λx).
Not all of these solutions are valid; we require that v(0) = 0 = v(π).√ Therefore A cos(0)+B sin(0)√=
0, so that A = 0. The other boundary condition implies B sin(π λ) = 0 which is only true if π λ
is a multiple of π. Therefore
λn = n2 , n = 1, 2, 3, . . . .
Corresponding to each of these eigenvalues is the eigenfunction vn (x) = sin(nx), n = 1, 2, 3, . . ..
Recall that we don’t care about the prefactor B since eigenfunctions can always be rescaled.
Properly speaking, we also need to consider the cases λ = 0, λ < 0. For the first case, v(x) =
Ax + B,pand it’s obvious p
one needs A = B = 0 to satisfy the boundary conditions. If λ < p0, v(x) =
A exp( p |λ|x) + B exp(− |λ|x). The boundary conditions imply A + B = 0 and A exp( |λ|π) +
B exp(− |λ|π) = 0, which written in matrix form is
1 1p A 0
p = .
exp( |λ|π) exp(− |λ|π) B 0
This is the famous Fourier sine series - just one of several types of Fourier series.
7
Series Space of functions Orthogonal expansion Coefficients
type for f (x)
RL
A0
+ ∞ nπx 1
f (x) cos( nπx
P
Fourier f (x) : [−L, L] → R n=1 An cos( L ) An = L )dx
2P
∞
L
R−L
f (−L) = f (L) + n=1 Bn sin( nπx
L ) Bn = 1 L nπx
L −L f (x) sin( L )dx
f 0 (−L) = f 0 (L)
P∞ nπx 2
RL
Sine f (x) : [0, L] → R n=1 Bn sin( L ) Bn = L 0 f (x) sin( nπx
L )dx
f (0) = 0 = f (L)
A0 P∞ nπx 2
RL
Cosine f (x) : [0, L] → R 2 + n=1 An cos( L ) An = L 0 f (x) cos( nπx
L )dx
f 0 (0) = 0 = f 0 (L)
P∞ −inπx 1
RL inπx
Complex f (x) : [−L, L] → C n=−∞ cn exp( L ) cn = 2L −L f (x) exp( L )dx
f (−L) = f (L)
f 0 (−L) = f 0 (L)
Table 1: Various Fourier series
The question that arises is, how do we actually compute the coefficients Bn ? We have already
answered this question in a more general setting. Because the eigenfunctions sin(nx) are orthog-
onal, we can use (7) and (8). For the present case, this is equivalent to taking an inner product of
(14) with each eigenfunction sin(nx). This gives the equivalent of (7), namely
Rπ
hf (x), sin(nx)i f (x) sin(nx)dx
Bn = = 0R π 2 .
hsin(nx), sin(nx)i 0 sin (nx)dx
It should be emphasized that Fourier coefficient formulas like this one don’t need to be memo-
rized. They arise quite simply from the basic idea of finding coefficients of a linear combination
of orthogonal vectors.
There are other types of Fourier series involving cosines or complex exponentials. These func-
tions are all eigenfunctions of the second derivative operator L = d2 /dx2 , but with different vector
spaces of functions on which it acts (remember the technical point earlier: properties of differential
operators crucially depend on the vector space in question). Table 2.8 summarizes the standard
types of Fourier series. The coefficient formulas all derive from taking inner products with each
RL
eigenfunction (note that for the complex series, the inner product is hu, vi = −L uvdx). We shall
see later that other eigenvalue problems give rise to new orthogonal sets of functions: the Bessel
functions, and the spherical harmonics.
8
We will not encounter many equations which have integral operators, but some of our solu-
tions will involve integral operators. This is because integral operators are often inverses of dif-
ferential operators. This means that the inhomogeneous equation Lu = g has a solution u = L−1 g
where L−1 is an integral operator.
Integral operators have adjoints. For example, using the L2 inner product, the operator given
in (15) satisfies
Z Z
hLu, vi = v(y) k(x, y)u(x) dxdy
ZΩ ZΩ
= u(x) k(x, y)v(y) dydx = hu, L† vi
Ω Ω
This is not the same as L; the inputs to the kernel k(y, x) have been reversed.