notes_spectral_theorems
notes_spectral_theorems
From section 6.3, we know that every linear operator T on a finite dimensional inner prod-
uct space V has an adjoint. (T ∗ is defined as the unique linear operator on V such that
hT (x), yi = hx, T ∗ (y)i for every x, y ∈ V – see Theroems 6.8 and 6.9.) When V is infinite
dimensional, the adjoint T ∗ may or may not exist.
One useful fact (Theorem 6.10) is that if β is an orthonormal basis for a finite dimen-
sional inner product space V , then [T ∗ ]β = [T ]∗β . That is, the matrix representation of the
operator T ∗ is equal to the complex conjugate of the matrix representation for T .
For a general vector space V , and a linear operator T , we have already asked the ques-
tion “when is there a basis of V consisting only of eigenvectors of T ?” – this is exactly when
T is diagonalizable. Now, for an inner product space V , we know how to check whether vec-
tors are orthogonal, and we know how to define the norms of vectors, so we can ask “when
is there an orthonormal basis of V consisting only of eigenvectors of T ?” Clearly, if there is
such a basis, T is diagonalizable – and moreover, eigenvectors with distinct eigenvalues must
be orthogonal.
(a) T is normal if T ∗ T = T T ∗
(b) T is self-adjoint if T ∗ = T
Notes
1. Self-adjoint ⇒ Normal (Clearly, if T ∗ = T , then T ∗ and T commute!)
Spectral Theorem 2 Let T be a linear operator on a finite dimensional real inner prod-
uct space V . Then, T is self-adjoint if and only if there exists an orthonormal basis (for V )
consisting of eigenvectors of T .
We’ll prove the simple direction first: Assume that there is an orthonormal basis β con-
sisting of eigenvectors of T . Then, we know that [T ]β is diagonal, and also that [T ∗ ]β = [T ]∗β
is diagonal. Since diagonal matrices commute, we have that
[T ∗ T ]β = [T ∗ ]β [T ]β = [T ]β [T ∗ ]β = [T T ∗ ]β .
We’ll see in the lemmas below that T normal implies that every eigenvector of T with
eigenvalue λ is also an eigenvector of T with eigenvalue λ̄. Since β is a basis of eigenvectors
of either T or T ∗ , we have that both [T ]β and [T ∗ ]β are diagonal matrices with their eigen-
values on the diagonal. In the case F = R, every eigenvalue of T is real, so λ = λ̄, and we
must have that [T ∗ ]β = [T ]β . Therefore, in the case of a real inner product space, we know
that T is not only normal, but also self-adjoint!
Lemma (pg. 269) Let V be a finite-dimensional inner product space and T ∈ L(V ). If
T has an eigenvector with eigenvalue λ, then T ∗ has an eigenvector with eigenvalue λ̄
See the book for the proof. Note that the above lemma is true for any linear operator T
– but we do not know that the eigenvectors for T and T ∗ are related! The next theorem
states that when T is normal, we know the eigenvectors for T and T ∗ are the same – this is
the fact we used in the proof of the spectral theorem above.
Theorem 6.15 (pg. 371) Let V be an inner product space and T ∈ L(V ) be a normal
operator. Then,
(c) If T (x) = λx (for some x ∈ V and some λ ∈ F ), then T ∗ (x) = λ̄x.
(d) If λ1 and λ2 are distinct eigenvalues of T with eigenvectors x1 and x2 , then hx1 , x2 i = 0.
Sketch of proof: First prove that, since T is normal, kT (x)k = kT ∗ (x)k for every x ∈ V . Then,
since U = T − λI is normal, and U ∗ = T ∗ − λ̄I, we know that U (x) = 0 if and only if
U ∗ (x) = 0. In other words, T (x) = λx if and only if T ∗ (x) = λ̄I.
For (d), Simply compute λ1 hx1 , x2 i = λ2 hx1 , x2 i. Since λ1 6= λ2 , we must have hx1 , x2 i = 0.
We’re almost ready to finish the proof of the first spectral theorem, but first, we’ll need
the following theorem:
Proof: Induct on n = dim V . Of course, if n = 1, take any basis β = {x} with kxk = 1. Then,
[T ]β = (a) (where a = T (x)) is upper-triangular. Now, assume the result for n − 1: that is,
suppose that if V is any finite-dimensional inner product space with dimension n − 1 and if
T is any linear operator such that the characteristic polynomial splits, then there exists an
orthonormal basis for V such that [T ]β is upper triangular. We wish to show this result for n.
Fix an n−dimensional inner product space V and a T ∈ L(V ) such that the character-
istic polynomial of T splits. This implies that T has an eigenvalue λ. Therefore, T ∗ has
an eigenvalue λ̄ (by Lemma, pg. 369). Let z be a unit eigenvector of T ∗ , so T ∗ z = λ̄z.
Consider the space W = span({z}). Clearly, W ⊥ is n − 1 dimensional, and TW ⊥ is a linear
operator on W ⊥ , so by the induction hypothesis, there exists a basis γ such that [TW ⊥ ]γ is
upper-triangular. (Notice that we know that TW ⊥ splits since W ⊥ is T −invariant! Check
this fact below:)
Assume v1 , v2 , ..., vk−1 are all eigenvectors Let λ1 , λ2 , ..., λk−1 be the corresponding eigen-
values. Then, since A is upper-triangular, look at the k th column of A and note Amk = 0 if
m > k: This lets us compute
From Theorem 6.15, the fact that T is normal implies that T ∗ (vj ) = λ̄j vj . Since the basis
is orthonormal, we have the formula Ajk = hT (vk ), vj i for the coefficients in the equation
above. Computing,
Before we can prove the second version of the spectral theorem, for F = R, we need the
following lemma:
Proof of (b): When F = C, we already know the characteristic polynomial splits. Let β
be any orthonormal basis for V ; we know that A = [T ]β satisfies A∗ = A. Define, for every
x ∈ Cn ,
S(x) = Ax.
S has the same characteristic polynomial as T , but since S is a linear operator on a com-
plex inner product space, we know the characteristic polynomial of S splits. Moreover,
since S is self-adjoint, each eigenvalue λ1 , ..., λn is real, and the polynomial pT (t) = pS (t) =
(λ1 − t)(λ2 − t)...(λn − t) splits over the field F = R.
Proof of Spectral Theorem 2: Assume F = R and T is a self-adjoint linear operator on
V . Then, the characteristic polynomial of T splits, and Schur’s theorem implies that there
exists an orthonormal basis β for V such that A = [T ]β is upper-triangular. The same proof
as for the first spectral theorem now works since T is normal, but it is easier to note that
since T ∗ = T , we know that both A and AT = A∗ = A are upper triangular. Therefore, A
is diagonal.
Finally, we will show the next theorem, which includes the fact that unitary (and orthog-
onal) operators must be normal!
Theorem 6.18 Let V be a finite dimensional inner product space and T ∈ L(V ). The
following are equivalent:
(a) T ∗ T = T T ∗ = I
Proof:
Let β = {v1 , v2 , ..., vn } be an orthnormal basis for V . Then T (β) = {T (v1 ), T (v2 ), ..., T (vn )}
is an orthonormal basis for V (since, from (b), hT (vi ), T (vj )i = hvi , vj i = δij ).
Pn Pn
Given any x ∈ V , write x = i=1 ai vi , we can easily compute kxk2 = i=1 |ai |2 .
Pn Pn
Since T is linear, T (x) = i=1 ai T (vi ); the same computation shows kT (x)k2 = i=1 |ai |2
(e) ⇒ (a) Assume (e) kT (x)k2 = kxk2 for every x ∈ V .
For all x ∈ V ,
hx, xi = kxk2 = kT (x)k2 = hT (x), T (x)i = hx, T ∗ T (x)i
This implies, for all x ∈ V , that
For all x ∈ β, U (x) = λx. By (?), 0 = hx, U (x)i = λ̄ hx, xi. Since x 6= 0, we must
have λ = 0. This implies that [U ]β is the zero matrix! Hence, U = I − T ∗ T = To (the zero
transformation on V ), so T ∗ T = I. Since V is finite dimensional, this proves that T ∗ must
be the inverse of T and therefore T T ∗ = I.