0% found this document useful (0 votes)
11 views

notes_spectral_theorems

Uploaded by

y100000agrawal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
11 views

notes_spectral_theorems

Uploaded by

y100000agrawal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

Math 108b: Notes on the Spectral Theorem

From section 6.3, we know that every linear operator T on a finite dimensional inner prod-
uct space V has an adjoint. (T ∗ is defined as the unique linear operator on V such that
hT (x), yi = hx, T ∗ (y)i for every x, y ∈ V – see Theroems 6.8 and 6.9.) When V is infinite
dimensional, the adjoint T ∗ may or may not exist.

One useful fact (Theorem 6.10) is that if β is an orthonormal basis for a finite dimen-
sional inner product space V , then [T ∗ ]β = [T ]∗β . That is, the matrix representation of the
operator T ∗ is equal to the complex conjugate of the matrix representation for T .

For a general vector space V , and a linear operator T , we have already asked the ques-
tion “when is there a basis of V consisting only of eigenvectors of T ?” – this is exactly when
T is diagonalizable. Now, for an inner product space V , we know how to check whether vec-
tors are orthogonal, and we know how to define the norms of vectors, so we can ask “when
is there an orthonormal basis of V consisting only of eigenvectors of T ?” Clearly, if there is
such a basis, T is diagonalizable – and moreover, eigenvectors with distinct eigenvalues must
be orthogonal.

Definitions Let V be an inner product space. Let T ∈ L(V ).

(a) T is normal if T ∗ T = T T ∗

(b) T is self-adjoint if T ∗ = T

For the next two definitions, assume V is finite-dimensional: Then,


(c) T is unitary if F = C and kT (x)k = kxk for every x ∈ V

(d) T is orthogonal if F = R and kT (x)k = kxk for every x ∈ V

Notes
1. Self-adjoint ⇒ Normal (Clearly, if T ∗ = T , then T ∗ and T commute!)

2. Unitary or Orthogonal ⇒ Invertible (Show that N (T ) = {0}.)

3. Unitary or Orthogonal ⇒ Normal (This is because, when T is unitary (or orthogonal),


we must have T ∗ T = T T ∗ = I, the identity operator on V – see Theorem 6.18.)
Example A simple example of an orthogonal operator is the rotation operator Tθ on R2 . In
the standard basis,  
cos θ − sin θ
Aθ = [Tθ ]β = .
sin θ cos θ
Notice that A is orthogonal. Also, A∗ = AT 6= A, so A is not self-adjoint. However, it is
easy to check that AA∗ = A∗ A = I, so A is normal. If the matrix A were over the complex
field, it would have a orthonormal basis of eigenvectors by the spectral theorem (see below)!
However, over the reals, we see that the characteristic polynomial of A does not split (unless
sin θ = 0) – therefore, there are no real eigenvalues and therefore no eigenvectors at all for
the operator Tθ on R2 .

Spectral Theorem 1 Let T be a linear operator on a finite dimensional complex in-


ner product space V . Then, T is normal if and only if there exists an orthonormal basis (for
V ) consisting of eigenvectors of T .

Spectral Theorem 2 Let T be a linear operator on a finite dimensional real inner prod-
uct space V . Then, T is self-adjoint if and only if there exists an orthonormal basis (for V )
consisting of eigenvectors of T .

We’ll prove the simple direction first: Assume that there is an orthonormal basis β con-
sisting of eigenvectors of T . Then, we know that [T ]β is diagonal, and also that [T ∗ ]β = [T ]∗β
is diagonal. Since diagonal matrices commute, we have that

[T ∗ T ]β = [T ∗ ]β [T ]β = [T ]β [T ∗ ]β = [T T ∗ ]β .

Hence, T ∗ T = T T ∗ , so T is normal. This holds whether F = C or F = R.

We’ll see in the lemmas below that T normal implies that every eigenvector of T with
eigenvalue λ is also an eigenvector of T with eigenvalue λ̄. Since β is a basis of eigenvectors
of either T or T ∗ , we have that both [T ]β and [T ∗ ]β are diagonal matrices with their eigen-
values on the diagonal. In the case F = R, every eigenvalue of T is real, so λ = λ̄, and we
must have that [T ∗ ]β = [T ]β . Therefore, in the case of a real inner product space, we know
that T is not only normal, but also self-adjoint!

Lemma (pg. 269) Let V be a finite-dimensional inner product space and T ∈ L(V ). If
T has an eigenvector with eigenvalue λ, then T ∗ has an eigenvector with eigenvalue λ̄

See the book for the proof. Note that the above lemma is true for any linear operator T
– but we do not know that the eigenvectors for T and T ∗ are related! The next theorem
states that when T is normal, we know the eigenvectors for T and T ∗ are the same – this is
the fact we used in the proof of the spectral theorem above.
Theorem 6.15 (pg. 371) Let V be an inner product space and T ∈ L(V ) be a normal
operator. Then,
(c) If T (x) = λx (for some x ∈ V and some λ ∈ F ), then T ∗ (x) = λ̄x.
(d) If λ1 and λ2 are distinct eigenvalues of T with eigenvectors x1 and x2 , then hx1 , x2 i = 0.

Sketch of proof: First prove that, since T is normal, kT (x)k = kT ∗ (x)k for every x ∈ V . Then,
since U = T − λI is normal, and U ∗ = T ∗ − λ̄I, we know that U (x) = 0 if and only if
U ∗ (x) = 0. In other words, T (x) = λx if and only if T ∗ (x) = λ̄I.

For (d), Simply compute λ1 hx1 , x2 i = λ2 hx1 , x2 i. Since λ1 6= λ2 , we must have hx1 , x2 i = 0.

We’re almost ready to finish the proof of the first spectral theorem, but first, we’ll need
the following theorem:

Schur’s Theorem Let T be a linear operator on a finite-dimensional inner product space


V . Suppose the characteristic polynomial of T splits. Then, there exists an orthonormal
basis β for V such that the matrix [T ]β is upper triangular.

Proof: Induct on n = dim V . Of course, if n = 1, take any basis β = {x} with kxk = 1. Then,
[T ]β = (a) (where a = T (x)) is upper-triangular. Now, assume the result for n − 1: that is,
suppose that if V is any finite-dimensional inner product space with dimension n − 1 and if
T is any linear operator such that the characteristic polynomial splits, then there exists an
orthonormal basis for V such that [T ]β is upper triangular. We wish to show this result for n.

Fix an n−dimensional inner product space V and a T ∈ L(V ) such that the character-
istic polynomial of T splits. This implies that T has an eigenvalue λ. Therefore, T ∗ has
an eigenvalue λ̄ (by Lemma, pg. 369). Let z be a unit eigenvector of T ∗ , so T ∗ z = λ̄z.
Consider the space W = span({z}). Clearly, W ⊥ is n − 1 dimensional, and TW ⊥ is a linear
operator on W ⊥ , so by the induction hypothesis, there exists a basis γ such that [TW ⊥ ]γ is
upper-triangular. (Notice that we know that TW ⊥ splits since W ⊥ is T −invariant! Check
this fact below:)

Let y ∈ W ⊥ . We want to show that T (y) is in W ⊥ . For every , compute

Finally β = γ ∪ {z} is orthonormal, and [T ]β is upper-triangular: Letting [T z]β = a ∈ F n ,


 
[TW ⊥ ]γ |
[T ]β =  a . 
0 |
Proof of Spectral Theorem 1 Assume T is normal. Since F = C, we know (from the fun-
damental theorem of algebra) that the characteristic polynomial of T splits. By Schur’s
theorem, we can find an orthonormal basis β = {v1 , v2 , ..., vn } such that A = [T ]β is upper-
triangular. Clearly v1 is an eigenvector of T (since T (v1 ) = A11 v1 .) The fact that T is normal
will imply that v2 , ..., vn are also eigenvectors!

Assume v1 , v2 , ..., vk−1 are all eigenvectors Let λ1 , λ2 , ..., λk−1 be the corresponding eigen-
values. Then, since A is upper-triangular, look at the k th column of A and note Amk = 0 if
m > k: This lets us compute

T (vk ) = A1k v1 + A2k v2 + ...Akk vk

From Theorem 6.15, the fact that T is normal implies that T ∗ (vj ) = λ̄j vj . Since the basis
is orthonormal, we have the formula Ajk = hT (vk ), vj i for the coefficients in the equation
above. Computing,

Ajk = hT (vk ), vj i = hvk , T ∗ (vj )i = vk , λ̄j vj = λj hvk , vj i = 0.

Therefore, we have simply that T (vk ) = Akk vk , so vk is an eigenvector for T .

By induction, we have shown that v1 , v2 , ..., vn are all eigenvectors of T , so β is an or-


thonormal basis consisting of only eigenvectors of T , and the spectral theorem is proven. 

Before we can prove the second version of the spectral theorem, for F = R, we need the
following lemma:

Lemma (pg. 373) Let T be a self-adjoint operator on a finite-dimensional inner product


space V . Then the following two facts hold (whether we have F = R or F = C)

(a) Every eigenvalue of T is real.

(b) The characteristic polynomial of T splits.

Proof of (a): From Theorem 6.15, if x is an eigenvalue of T , we have both T (x) = λx


for some λ ∈ F and T ∗ (x) = λ̄x. However, since T is self-adjoint, T (x) = T ∗ (x). Therefore,
λx = λ̄x. Since x 6= 0, we have that λ = λ̄; i.e., λ is real.

Proof of (b): When F = C, we already know the characteristic polynomial splits. Let β
be any orthonormal basis for V ; we know that A = [T ]β satisfies A∗ = A. Define, for every
x ∈ Cn ,
S(x) = Ax.
S has the same characteristic polynomial as T , but since S is a linear operator on a com-
plex inner product space, we know the characteristic polynomial of S splits. Moreover,
since S is self-adjoint, each eigenvalue λ1 , ..., λn is real, and the polynomial pT (t) = pS (t) =
(λ1 − t)(λ2 − t)...(λn − t) splits over the field F = R. 
Proof of Spectral Theorem 2: Assume F = R and T is a self-adjoint linear operator on
V . Then, the characteristic polynomial of T splits, and Schur’s theorem implies that there
exists an orthonormal basis β for V such that A = [T ]β is upper-triangular. The same proof
as for the first spectral theorem now works since T is normal, but it is easier to note that
since T ∗ = T , we know that both A and AT = A∗ = A are upper triangular. Therefore, A
is diagonal.

Finally, we will show the next theorem, which includes the fact that unitary (and orthog-
onal) operators must be normal!

Theorem 6.18 Let V be a finite dimensional inner product space and T ∈ L(V ). The
following are equivalent:

(a) T ∗ T = T T ∗ = I

(b) hT (x), T (y)i = hx, yi for every x, y ∈ V

(e) kT (x)k = kxk for every x ∈ V

Proof:

(a) ⇒ (b) Assume (a) T ∗ T = T T ∗ = I.

Given x, y ∈ V , since T ∗ T (x) = I(x) = x, we have hx, yi = hT ∗ T (x), yi = hT (x), T (y)i.

(b) ⇒ (e) Assume (b) hx, yi = hT (x), T (y)i for every x, y ∈ V

Let β = {v1 , v2 , ..., vn } be an orthnormal basis for V . Then T (β) = {T (v1 ), T (v2 ), ..., T (vn )}
is an orthonormal basis for V (since, from (b), hT (vi ), T (vj )i = hvi , vj i = δij ).
Pn Pn
Given any x ∈ V , write x = i=1 ai vi , we can easily compute kxk2 = i=1 |ai |2 .

Pn Pn
Since T is linear, T (x) = i=1 ai T (vi ); the same computation shows kT (x)k2 = i=1 |ai |2
(e) ⇒ (a) Assume (e) kT (x)k2 = kxk2 for every x ∈ V .

For all x ∈ V ,
hx, xi = kxk2 = kT (x)k2 = hT (x), T (x)i = hx, T ∗ T (x)i
This implies, for all x ∈ V , that

(?) hx, (I − T ∗ T )(x)i = 0.

Let U = I − T ∗ T . Check that U is self-adjoint: U ∗ = I ∗ − (T ∗ T )∗ = I − T ∗ T . Then, by the


spectral theorem there exists an orthonormal basis β consisting of eigenvectors of U .

For all x ∈ β, U (x) = λx. By (?), 0 = hx, U (x)i = λ̄ hx, xi. Since x 6= 0, we must
have λ = 0. This implies that [U ]β is the zero matrix! Hence, U = I − T ∗ T = To (the zero
transformation on V ), so T ∗ T = I. Since V is finite dimensional, this proves that T ∗ must
be the inverse of T and therefore T T ∗ = I. 

You might also like