0% found this document useful (0 votes)
88 views16 pages

Section 6.1 - Inner Products and Norms: M×N I, J I, J

This document discusses inner product spaces and related concepts. It defines inner products, norms, orthogonality, orthogonal complements and bases. It proves several theorems regarding these topics, including the Cauchy-Schwarz inequality, Gram-Schmidt process, and existence of orthonormal bases. It also discusses linear functionals on inner product spaces.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
88 views16 pages

Section 6.1 - Inner Products and Norms: M×N I, J I, J

This document discusses inner product spaces and related concepts. It defines inner products, norms, orthogonality, orthogonal complements and bases. It proves several theorems regarding these topics, including the Cauchy-Schwarz inequality, Gram-Schmidt process, and existence of orthonormal bases. It also discusses linear functionals on inner product spaces.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 16

Section 6.

1 - Inner Products and Norms


Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to
every ordered pair of vectors x and y in V , a scalar in F , denoted hx, yi, such that for all x, y, and z in V
and all c in F , the following hold:
1. hx + z, yi = hx, yi + hz, yi
2. hcx, yi = chx, yi
3. hx, yi = hy, xi, where the bar denotes complex conjugation.
4. hx, xi > 0 if x 6= 0
Note that if z is a complex number, then the statement z 0 means that z is real and non-negative.
Notice that if F = R, (3) is just hx, yi = hy, xi.
Definition. Let A Mmn (F ). We define the conjugate transpose or adjoint of A to be the n m matrix
A such that (A )i,j = Ai,j for all i, j.
Theorem 6.1. Let V be an inner product space. Then for x, y, z V and c F ,
1. hx, y + zi = hx, yi + hx, zi
2. hx, cyi = chx, yi
3. hx, 0i = h0, xi = 0
4. hx, xi = 0 x = 0
5. If hx, yi = hx, zi, for all x V , then y = z.
Proof. (Also see handout by Dan Hadley.)
1.
hx, y + zi = hy + z, xi = hy, xi + hz, xi
= hy, xi + hz, xi = hx, yi + hx, zi
2. hx, cyi = hcy, xi = chy, xi = c hy, xi = chx, yi
3. hx, 0i = hx, 0xi = 0hx, xi = h0x, xi = h0, xi
4. If x = 0 then by (3) of this theorem, hx, xi = 0. If hx, xi = 0 then by (3) of the definition, it must be
that x = 0.
5. hx, yi hx, zi = 0, for all x V . hx, y zi = 0, for all x V . Thus, hy z, y zi = 0 and we have
y z = 0 by (4). So, y = z.

Definition. Let V be an inner product space. For x V , we define the norm or length of x by ||x|| =

hx, xi.

Theorem 6.2. Let V be an inner product space over F . Then for all x, y V and c F , the following are
true.
1. ||cx|| = |c| ||x||.
2. ||x|| = 0 if and only if x = 0. In any case ||x|| 0.
3. (Cauchy-Schwarz Inequality) | < x, y > | ||x|| ||y||.
4. (Triangle Inequality) ||x + y|| ||x|| + ||y||.
1

Proof. (Also see notes by Chris Lynd.)


1. ||cx||2 = hcx, cxi = c c hx, xi = |c|2 ||x||2 .
p
p
2. ||x|| = 0 iff (hx, xi) = 0 iff hx, xi = 0 x = 0. If x 6= 0 then (hx, xi) > 0 and so (hx, xi) > 0.
3. If y = 0 the result is true. So assume y 6= 0. Finished in class.
4. Done in class.

Definition. Let V be an inner product space, x, y V are orthogonal or ( perpendicular ) if hx, yi = 0. A


subset S V is called orthogonal if x, y S, hx, yi = 0. x V is a unit vector if ||x|| = 1 and a subset
S V is orthonormal if S is orthogonal and x S, ||x|| = 1.

Section 6.2
Definition. Let V be an inner product space. Then S V is an orthonormal basis of V if it is an ordered
basis and orthonormal.
Theorem 6.3. Let V be an inner product space and S = {v1 , v2 , . . . , vk } be an orthogonal subset of V such
that i, vi 6= 0. If y Span(S), then
k
X
hy, vi i
y=
vi
||vi ||2
i=1
Proof. Let
y=

k
X

ai vi

i=1

Then
k
X
h
ai vi , vj i

hy, vj i =

i=1
k
X

ai hvi , vj i

i=1

=
=
So aj =

aj hvj , vj i
aj ||vj ||2

hy,vi i
||vi || .

Corollary 1. If S also is orthonormal then


y=

k
X

hy, vi ivi

i=1

Corollary 2. If S also is orthogonal and all vectors in S are non-zero then S is linearly independent.
Pk
ii
Proof. Suppose i=1 ai vi = 0. Then for all j, aj = h0,v
||vi || = 0. So S is linearly independent.
Theorem 6.4. (Gram-Schmidt) Let V be an inner product space and S = {w1 , w2 , . . . , wn } V be a
linearly independent set. Define S 0 = {v1 , v2 , . . . , vn } where v1 = w1 and for 2 k n,
vk = wk

k1
X
i=1

hwk , vi i
vi .
||vi ||2

Then, S is an orthogonal set of non-zero vectors and Span(S 0 ) =Span(S).


2

Proof. Base case, n = 1. S = {w1 }, S 0 = {v1 }.


0
Let n > 1 and Sn = {w1 , w2 , . . . , wn }. For Sn1 = {w1 , w2 , . . . , wn1 } and Sn1
= {v1 , v2 , . . . , vn1 },
0
0
we have by induction that span(Sn1 ) = span(Sn1 ) and Sn1 is orthogonal. To show Sn0 is orthogonal, we
just have to show j [n 1], < vn , vj >= 0.

vn

= wn

n1
X
k=1

< vn , vj >

< wk , vk >
vk
||vk ||2

= < wn , vj > <

n1
X
k=1

= < wn , vj >

n1
X
k=1

< wk , vk >
vk , vj >
||vk ||2

< wk , vk >
< vk , vj >
||vk ||2

< wn , vj >
= < wn , vj >
< vj , vj >
||vj ||2
= =0
We now show span(Sn ) = span(Sn0 ). We know dim(span(Sn )) = n, since Sn is linearly independent.
We know dim(span(Sn0 )) = n, by Corollary 2 to Theorem 6.3.
0
). So, we just need to show vn span(()Sn ).
We know span(Sn1 ) = span(Sn1
vn = wn

n1
X

aj vj

j=1
0
) and
for some constants a1 , a2 , . . . , an1 . For all j [n 1], vj span(Sn1 ) since span(Sn1 ) = span(Sn1
0
vj Sn1 . Therefore, vn span(Sn ).

Theorem 6.5. Let V be a non-zero finite dimensional inner product space. Then V has an orthonormal
basis . Furthermore, if = {v1 , v2 , . . . , vn } and x V , then
x=

n
X

hx, vi ivi .

i=1

Proof. Start with a basis of V Apply Gram Schmidt, Theorem 6.4 to get an orthogonal set 0 . Produce
from 0 by normalizing 0 . That is, multiply each x 0 by 1/||x||. By Corollary 2 to Theorem 6.3, is
linearly independent and since it has n vectors, it must be a basis of V . By Corollary 1 to Theorem 6.3, if
x V , then
n
X
x=
hx, vi ivi .
i=1

Corollary 1. Let V be a finite dimensional inner product space with an orthonormal basis = {v1 , v2 , . . . , vn }.
Let T be a linear operator on V , and A = [T ] . Then for any i, j, (A)i,j = hT (vj ), vi i.
Proof. By Theorem 6.5, x V , then
x=

n
X

hx, vi ivi .

i=1

So
T (vj ) =

n
X

hT (vj ), vi ivi .

i=1

Hence

[T (vj )] = (hT (vj ), v1 i, hT (vj ), v2 i, . . . , hT (vj ), vn i)t

So, (A)i,j = hT (vj ), vi i.


3

Definition. Let S be a non-empty subset of an inner product space V . We define S (read S perp) to be
the set of all vectors in V that are orthogonal to every vector in S; that is, S = {x : hx, yi = 0, y S}.
S is called the orthogonal complement of S.
Theorem 6.6. Let W be a finite dimensional subspace of an inner product space V , and let y V . Then
there exist unique vectors u W , z W such that y = u + z. Furthermore, if {v1 , v2 . . . , vk } is an
orthonormal basis for W , then
k
X
hy, vi ivi .
u=
i=1

Proof. Let y V . Let {v1 , v2 , . . . , vk } be an orthonormal basis for W . This exists by Theorem 6.5. Let
u=

k
X

hy, vi ivi .

i=1

We will show y u W .
It suffices to show that for all i [k], hy u, vi i = 0.
hy u, vi i

= hy

k
X

hy, vj ivj , vi i

j=1

= hy, vi i

k
X

hy, vj i hvj , vi i

j=1

= hy, vi i hy, vi i
= 0
Thus y u W . Suppose x W W . Then x W and x W . So, hx, xi = 0 and we have that
x = 0.
Suppose r W and s W such that y = r + s. Then
r+s
ru

= u+z
= zs

This shows that both r u and z s are in both W and W . But W W = so it must be that
r u = 0 and z s = 0 which implies that r = u and z = s and we see that the representation of y is
unique.
Corollary 1. In the notation of Theorem 6.6, the vector u is the unique vector in W that is closest to
y. That is, for any x W , ||y x|| ||y u|| and we get equality in the previous inequality if and only if
x = u.
Pk
Proof. Let y V , u = i=1 < y, vi > vi . z = y u W . Let x W . Then < y u, x >= 0, since
z W .
By Exercise 6.1, number 10, if a is orthogonal to b then ||a + b||2 = ||a||2 + ||b||2 . So we have
||y x||2

= ||(u + z) x||2
= ||(u x) + z||2
= ||u x||2 + ||z||2
||z||2
= ||y u||2

Now suppose ||yx|| = ||yu||. By the squeeze theorem, ||ux||2 +||z||2 = ||z||2 and thus ||ux||2 = 0,
||u x|| = 0, u x = 0, and so u = x.
4

Theorem 6.7. Suppose that S = {v1 , v2 , . . . , vk } is an orthonormal set in an n-dimensional inner product
space V . Then
(a) S can be extended to an orthonormal basis {v1 , . . . , vk , vk+1 , . . . , vn }
(b) If W =Span(S) then S1 = {vk+1 , vk+2 , . . . , vb } is an orthonormal basis for W .
(c) If W is any subspace of V , then dim(V ) = dim(W )+ dim(W ).
Proof. (a) Extend S to an ordered basis. S 0 = {v1 , v2 , . . . , vk , wk+1 , wk+2 , . . . , wn }. Apply Gram-Schmidt
to S 0 . First k do not change. S 0 spans V . Then normalize S 0 to obtain = {v1 , . . . , vk , vk+1 , . . . , vn }. (b)

S1 = {vk+1 , vk+2 , . . . , vn } is an orthonormal


set and linearly
Pn
Pn independent Also, it is a subset of W . It

must span W since if x W , x = i=1 < x, vi > vi = i=k+1 < x, vi > vi .


(c) dim(V ) = n = k + n k = dim(W ) + dim(W ).

Section 6.3:
Theorem 6.8. Let V be a finite dimensional inner product space over F , and let g : V F be a linear
functional. Then, there exists a unique vector y V such that g(x) = hx, yi, x V .
Proof. Let = {v1 , v2 , . . . , vn } be an orthonormal basis for V and let
y=

n
X

g(vi )vi .

i=1

Then for 1 j n,
hvj , yi =

hvj ,

n
X

g(vi )vi i

i=1

n
X

g(vi )hvj , vi i

i=1

=
=

g(vj )hvj , vj i
g(vj )

and we have, x V, g(x) = hx, yi.


To show y is unique, suppose g(x) = hx, y0 i, x. Then, hx, yi = hx, y0 i, x. By Theorem 6.1(e), we have
y = y0 .
Example. (2b) Let V = C2 , g(z1 , z2 ) = z1 2z2 . Then V is an inner-product space with the standard inner
product h(x1 , x2 ), (y1 , y2 )i = x1 y1 + x2 y2 and g is a linear operator on V . Find a vector y V such that
g(x) = hx, yi, x V .
Sol: We need to find (y1 , y2 ) C2 such that
g(z1 , z2 ) = h(z1 , z2 ), (y1 , y2 )i, (z1 , z2 ) C2 .
That is:
z1 2z2 = z1 y 1 + z2 y 2 .
Using the standard ordered basis {(1, 0), (0, 1)} for C2 , the proof of Theorem 6.8 gives that y =
So,
(y1 , y2 ) =

g(1, 0)(1, 0) + g(0, 1)(0, 1)


= z1 (1, 0) + 2z2 (0, 1)
= (1, 0) 2(0, 1)
= (1, 2)

Check (1): LHS= z1 2z2 and for y1 = 1 and y 2 = 2, we have RHS= z1 2z2 .
5

Pn
i=1

(1)
g(vi )vi .

Theorem 6.9. Let V be a finite dimensional inner product space and let T be a linear operator on V . There
exists a unique operator T : V V such that hT (x), yi = hx, T (y)i, x, y V . Furthermore, T is linear.
Proof. Let y V
Define g : V F as g(x) = hT (x), yi, x V .
Claim: g is linear.
g(ax + z) =
=
=
=

hT (ax + z), yi
haT (x) + T (z), yi
haT (x), yi + hT (z), yi
ag(x) + g(z).

By Theorem 6.8, there is a unique y0 V such that g(x) = hx, y0 i.


So we have hT (x), yi = hx, y0 i, x V .
We define T : V V by T (y) = y0 . So hT (x), yi = hx, T (y)i, x V
Claim: T is linear. We have for cy + z V , T (cy + z) equals the unique y0 such that hT (x), cy + zi =
hx, T (cy + z)i. But,
hT (x), cy + zi =

chT (x), yi + hT (x), zi


chx, T (y)i + hx, T (z)i
hx, cT (y) + T (z)i

=
=
Since y0 is unique, T (cy + z) = cT (y) + T (z).
Claim: T is unique.
Let U : V V be linear such that
hT (x), yi =
hx, T (y)i =
=

hx, U (y)i, x, y V
hx, U (y)i, x, y V
T (y) = U (y), y V

So, T = U .
Definition. T is called the adjoint of the linear operator T and is defined to be the unique operator on V
satisfying hT (x), yi = hx, T (y)i, x, y V . For A Mn (F ), we have the definition of A , the adjoint of A
from before, the conjugate transpose.
Fact. hx, T (y)i = hT (x), yi, x, y V .
Proof.
hx, T (y)i =
=
=

hT (y), xi
hy, T (x)i
hT (x), yi

Theorem 6.10. Let V be a finite dimensional inner product space and let be an orthonormal basis for V .
If T is a linear operator on V , then
[T ] = [T ]
Proof. Let A = [T ] and B = [T ] with = {v1 , v2 , . . . , vn }.
By the corollary to Theorem 6.5,
(B)i,j

hT (vj ), vi i

hvi , T (vj )i

hT (vi ), vj i

(A)j,i

= (A )i,j
6

Corollary 2. Let A be an n n matrix, then LA = (LA ) .


Proof. Use , the standard ordered basis. By Theorem 2.16,
[LA ]
and [LA ]

= A
= A

By Theorem 6.10, [LA ] = [LA ] , which equals A by (2) and [LA ] by (3).
Therefore, [LA ] = [LA ] LA = LA .
Theorem 6.11. Let V be an inner product space, T, U linear operators on V . Then
1. (T + U ) = T + U .
2. (cT ) = cT , c F .
3. (T U ) = U T (composition)
4. (T ) = T
5. I = I.
Proof.

1.
h(T + U )(x), yi =

hx, (T + U ) (y)i

and
h(T + U )(x), yi =
=
=

hT (x), yi + hU (x), yi
hx, T (y)i + hx, U (y)i
hx, (T + U )(y)i

And (T + U ) is unique, so it must equal T + u .


2.
hcT (x), yi =

hx, (cT ) (y)i

and
hcT (x), yi =
=
=

chT (x), yi
chx, T (y)i
hx, cT (y)i

3.
hT U (x), yi =

hx, (T U ) (y)i

hT U (x), yi =

hT (U (x)), yi

and

=
=

hU (x), T (y)i
hx, U (T (y))i

hx, (U T )(y)i

(2)
(3)

4.
hT (x), yi =

hx, (T ) (y)i

by definition and
hT (x), yi =

hx, T (y)i

by Fact .
5.
hI (x), yi = hx, I(y)i = hx, yi, x, y V
Therefore, I (x) = x, x V and we have I = I.

Corollary 1. Let A and B be n n matrices. Then


1. (A + B) = A + B .
2. (cA) = cA , c F .
3. (AB) = B A
4. (A ) = A
5. I = I.
Proof. Use Theorem 6.11 and Corollary to Theorem 6.10. Or, use below.
Example. (Exercise 5b) Let A and B be m n matrices and C an n p matrix. Then
1. (A + B) = A + B .
2. (cA) = cA , c F .
3. (AC) = C A
4. (A ) = A
5. I = I.
Proof.

1.
(A + B)i,j = (A + B)j,i = (A)j,i + (B)j,i = (A)j,i + (B)j,i

and
(A + B )i,j = (A )i,j + (B )i,j = (A)j,i + (B)j,i
2. Let c F .
(cA)i,j = (cA)j,i = c(A)j,i = c(A)j,i
and
(cA )i,j = c(A )i,j = c(A)j,i
8

3.
((AC) )i,j

(AC)j,i
n
X

k=1
n
X

k=1
n
X

k=1
n
X

(A)j,k (C)k,i
(A)j,k (C)k,i
(A )k,j (C )i,k
(C )i,k (A )k,j

k=1

= (C A )i,j

Fall 2007 - The following was not covered


For x, y F n , let hx, yin denote the standard inner product of x and y in F n . Recall that if x and y are
regarded as column vectors, then hx, yin = y x.
Lemma. Let A Mmn (F ), x F n , and y F m . Then
hAx, yim = hx, A yin .
Proof.

hAx, yim = y (Ax) = (y A)x = (A y) x = hx, A yin .

Lemma. Let A Mmn (F ), x F n . Then rank(A A) =rank(A).


Proof. A A is an n n matrix. By the Dimension Theorem, rank(A A)+nullity(A A) = n. We also have,
rank(A)+nullity(A) = n.
We will show that the nullspace of A equals the nullspace of A A. We will show A Ax = 0 if and only
if Ax = 0.
0 =
0 =
0 =
0 =

A Ax
hA Ax, xin
hAx, A xim
hAx, Axim

Ax

Corollary 1. If A is an m n matrix such that rank(A) = n, then A A is invertible.


Theorem 6.12. Let A Mmn (F ) and y F m . Then there exists x0 F n such that (A A)x0 = A y and
||Ax0 y|| ||Ax y|| for all x F n . Furthermore, if rank(A) = n, then x0 = (A A)1 A y.
Proof. Define W = {Ax : x F n } = R(LA ). By the corollary to Theorem 6.6, there is a unique vector
u = Ax0 in W that is closest to y. Then, ||Ax0 y|| ||Ax y|| for all x F n . Also by Theorem 6.6,
z = y u is in W . So, z = Ax0 y is in W . So,
hAx, Ax0 yim = 0, x F n .
9

By Lemma 1,
hx, A (Ax0 y)in = 0, x F n .
So, A (Ax0 y) = 0. We see that x0 is the solution for x in
A Ax = A y.
If, in addition, we know that rank(A) = n, then by Lemma 2, we have that rank(A A) = n and is therefore
invertible. So,
x0 = (A A)1 A y.

Fall 2007 - not covered until here.


Section 6.4:
Lemma. Let T be a linear operator on a finite-dimensional inner product space V . If T has an eigenvector,
then so does T .
Proof. Let v be an eigenvector of T corresponding to the eigenvalue . For all x V , we have
0 = h0, xi = h(T I)(x), xi = hv, (T I) (x)i = hv, (T I)(x)i
So v is orthogonal to (T I)(x) for all x. Thus, v 6 R(T I) and so the nullity of (T I) is not 0.
There exists x 6= 0 such that (T I)(x) = 0. Thus x is a eigenvector corresponding to the eigenvalue
of T .
Theorem 6.14. (Schur). Let T be a linear operator on a finite-dimensional inner product space V . Suppose
that the characteristic polynomial of T splits. Then there exists an orthonormal basis for V such that the
matrix [T ] is upper triangular.
Proof. The proof is by mathematical induction on the dimension n of V . The result is immediate if n = 1.
So suppose that the result is true for linear operators on (n 1)-dimensional inner product spaces whose
characteristic polynomials split. By the lemma, we can assume that T has a unit eigenvector z. Suppose
that T (z) = z and that W = span({z}). We show that W is T -invariant. If y W and x = cz W ,
then
hT (y), xi =
=

hT (y), czi = hy, T (cz)i = hy, cT (z)i = hy, czi


c hy, zi = c(0) = 0.

So T (y) W .
By Theorem 5.21, the characteristic polynomial of TW divides the characteristic polynomial of T and
hence splits. By Theorem 6.7(c), dim(W ) = n 1, so we may apply the induction hypothesis to TW
and obtain an orthonormal basis of W such that [TW ] is upper triangular. Clearly, = {z} is an
orthonormal basis for V such that [T ] is upper triangular.
Definition. Let V be an inner product space, and let T be a linear operator on V . We say that T is normal
if T T = T T . An n n real or complex matrix A is normal if AA = A A.
Theorem 6.15. Let V be an inner product space, and let T be a normal operator on V . Then the following
statements are true.
(a) ||T (x)|| = ||T (x)|| for all x V .
(b) T cI is normal for every c F
(c) If x is an eigenvector of T , then x is also an eigenvector of T . In fact, if T (x) = x, then T (x) = x.
10

(d) If 1 and 2 are distinct eigenvalues of T with corresponding eigenvectors x1 and x2 , then x1 and x2
are orthogonal.
Proof. (a) For any x V , we have
2

kT (x)k

= hT (x), T (x)i = hT T (x), xi = hT T (x), xi


2

= hT (x), T (x)i = kT (x)k .


(b)
(T cI)(T cI)

= (T cI)(T cI)
= T (T cI) cI(T cI)
= T T cT cT c
cI
cI
= T T cT cT c
= T T cT cT c
cI

= T (T cI) cI(T cI)


= (T cI)(T cI)
= (T cI) (T cI)

(c) Suppose that T (x) = x for some x V . Let U = T I. Then U (x) = 0, and U is normal by (b).
Thus (a) implies that

0 = kU (x)k = kU (x)k = (T I)(x) = T (x) x .


Hence T (x) = x. So x is an eigenvector of T .
(d) Let 1 and 2 be distinct eigenvalues of T with corresponding eigenvectors x1 and x2 . Then using
(c), we have
1 hx1 , x2 i =
=

h1 x1 , x2 i = hT (x1 ), x2 i = hx1 , T (x2 )i

x1 , 2 x2 = 2 hx1 , x2 i .

Since 1 6= 2 , we conclude that hx1 , x2 i = 0.


Definition. Let T be a linear operator on an inner product space V . We say that T is self-adjoint (
Hermitian) if T = T . An n n real or complex matrix A is self-adjoint (Hermitian) if A = A .
Theorem 6.16. Let T be a linear operator on a finite-dimensional complex inner product space V . Then T
is normal if and only if there exists an orthonormal basis for V consisting of eigenvectors of T .
Proof. The characteristic polynomial of T splits over C. By Shurs Theorem there is an orthonormal basis
= {v1 , v2 , . . . , vn } such that A = [T ] is upper triangular.
() Assume T is normal. T (v1 ) = A1,1 v1 . Therefore, v1 is an eigenvector with associated eigenvalue
A1,1 .
Assume v1 , . . . , vk1 are all eigenvectors of T .
Claim: vk is also an eigenvector.
Suppose 1 , . . . , k1 are the corresponding eigenvalues. By Theorem 6.15,
T (vj ) = j vj T (vj ) = j vj .
Also, T (vk ) = A1,k v1 + A2,k v2 + + Ak,k vk .
We also know by the Corollary to Theorem 6.5 that Ai,j =< T (vj , vi >.
So,
Aj,k

= < T (vk , vj >


= < vk , T (vj ) >
= < vk , j vj >
= j < v k , v j >

0
j 6= k
=
Ak,k j = k
11

So T (vk ) = Ak,k vk and we have that vk is an eigenvector of T . By induction, is a set of eigenvectors


of T .
() If is orthonormal basis of eigenvectors then T is diagonalizable by Theorem 5.1.
D = [T ] is diagonal. Hence [T ] is diagonal and equals [T ] . Diagonal matrices commute. Hence,
[T ] [T ] = [T ] [T ] .
Let x V , x = a1 v1 + a2 v2 + + an vn .
We have that T T (x) = T T (x), and hence, T T = T T .
Lemma. Let T be a self-adjoint operator on a finite-dimensional inner product space V . Then
(a) Every eigenvalue of T is real.
(b) Suppose that V is a real inner product space. Then the characteristic polynomial of T splits.
Proof.

(a) Let be an eigenvalue of T . So T (x) = x for some x 6= 0. Then

x = T (x) = T (x) = (x)

=0
( )x
= 0, and we have that =
so is real.
But x 6= 0 so
(b) Let dim(V ) = n and be an orthonormal basis for V . Set A = [T ] . Then
A = [T ] = [T ] = [T ] = A
So A is self-adjoint.
Define TA : Cn Cn by TA (x) = Ax. Notice that [TA ] = A where is the standard ordered basis,
which is orthonormal.
So TA is self-adjoint and by (a) its eigenvalues are real.
WE know that over C the characteristic polynomial of TA factors into linear factors, t , and since
each is R, we know it also factors over R. But TA , A, and T all have the same characteristic
polynomial.
Theorem 6.17. Let T be a linear operator on a finite-dimensional real inner product space V . Then T is
self-adjoint if and only if there exists an orthonormal basis for V consisting of eigenvectors of T .
Proof. ( ) Assume T is self-adjoint. By the lemma, the characteristic polynomial of T splits. Now by
Shurs Theorem, there exists an orthonormal basis for V such that A = [T ] is upper triangular. But
A = [T ] = [T ] = [T ] = A
So A and A are both upper triangular, thus diagonal and we see that is a set of eigenvectors of T .
( ) Assume there is an orthonormal basis of V of eigenvectors of T . We know D = [T ] is diagonal with
eigenvalues on the diagonal. D is diagonal and is equal to D since it is real. But then [T ] = D = D =
[T ] = [T ] . So T = T .

Fall 2007 - The following was covered but will not be included on the final.
Section 6.5:
Definition. Let T be a linear operator on a finite-dimensional inner product space V (over F ). If ||T (x)|| =
||x|| for all x V , we call T a unitary operator if F = C and an orthogonal operator if F = R.
Theorem 6.18. Let T be a linear operator on a finite-dimensional inner product space V . Then the following
statements are equivalent.
1. T T = T T = I.
12

2. hT (x), T (y)i = hx, yi, x, y V .


3. If is an orthonormal basis for V , then T () is an orthonormal basis for V .
4. There exists an orthonormal basis for V such that T () is an orthonormal basis for V .
5. ||T (x)|| = ||x||, x V .
Proof. (1) (2)
hT (x), T (y)i

= hx, T T (y)i
= hx, yi

(2) (3) Let vi , vj . Then 0 = hvi , vj i = hT (vi ), T (vj )i, so T () is orthogonal.


By Corollary 2 to Theorem 6.3, any orthogonal subset is linearly independent and since T () has n
vectors, it must be a basis of V .
Also,
1 = ||vi ||2 = hvi , vi i = hT (vi ), T (vi )i.
So, T () is an orthonormal basis of V .
(3) (4) By Graham-Schmit, there is an orthonormal basis for V . By (3), T () is orthonormal.
(4) (5) Let = {v1 , v2 , . . . , vn } be an orthonormal basis for V . Let
x = a1 v1 + a2 v2 + an vn .
Then
||x||2 = a21 hv1 , v1 i + a22 hv2 , v2 i + + a2n hvn , vn i = a21 + a22 + + a2n
And
||T (x)||2 = a21 hT (v1 ), T (v1 )i + a22 hT (v2 ), T (v2 )i + + a2n hT (vn ), T (vn )i = a21 + a22 + + a2n
Therefore, ||T (x)|| = ||x||.
(5) (1) We are given ||T (x)|| = ||x||, for all x.
We know hx, xi = 0 if and only if x = 0 and hT (x), T (x)i = 0 if and only if T (x) = 0.
Therefore, T (x) = 0 if and only if x = 0.
So N (T ) = {0} and therefore, T is invertible.
We have hx, xi = hT (x), T (x)i = hx, T T (x)i. Therefore, T T (x) = x for all x, which implies that

T T = I. But since T is invertible, it must be that T = T 1 and we have that T T = T T = I.


Lemma. Let U be a self-adjoint operator on a finite-dimensional inner product space V . If hx, U (x)i = 0
for all x V , then U = T0 . (Where T0 (x) = 0, x.)
Proof. u = U U is normal.
If F = C, by Theorem 6.16, there is an orthonormal basis of eigenvectors.
If F = R, by Theorem 6.17, there is an orthonormal basis of eigenvectors.
Let x . Then U (x) = x for some . x,
< x, x >
0 =< x, U (x) >=< x, x >=
= 0, so = 0 and U (x) = 0, x.
Therefore,
Corollary 1. Let T be a linear operator on a finite-dimensional real inner product space V . Then V has
an orthonormal basis of eigenvectors of T with corresponding eigenvalues of absolute value 1 if and only if
T is both self-adjoint and orthogonal.
13

Proof. () Suppose V has an orthonormal basis {v1 , v2 , . . . , vn } such that i, T (vi ) = i vi and |i | = 1.
Then by Theorem 6.17, T is self-adjoint. Well show T T = T T = I. Then by Theorem 6.18, ||T (x)|| =
||x||, x V T is orthogonal.
T T (vi ) = T (T (vi )) = T (i vi ) = i T (vi ) = i i vi = i2 vi = vi
So T T = I. Similarly, T T = I.
() Assume T is self-adjoint. Then by Theorem 6.17, V has an orthonormal basis {v1 , v2 , . . . , vn } such
that T (vi ) = i vi , i.
If T is also orthogonal, we have
|i | ||vi || = ||i vi || = ||T (vi )|| = ||vi || |i | = 1

Corollary 2. Let T be a linear operator on a finite-dimensional complex inner product space V . Then V
has an orthonormal basis of eigenvectors of T with corresponding eigenvalues of absolute value 1 if and only
if T is unitary.
Definition. A square matrix A is called an orthogonal matrix if At A = AAt = I and unitary if A A =
AA = I. We say B is unitarily equivalent to D if there exists a unitary matrix Q such that D = Q BQ.
Theorem 6.19. Let A be a complex n n matrix. Then A is normal if and only if A is unitarily equivalent
to a diagonal matrix.
Proof. () Assume A is normal. There is an orthonormal basis for F n consisting of eigenvectors of A, by
Theorem 6.16. Let = {v1 , v2 , . . . , vn }. So A is similar to a diagonal matrix D by Theorem 5.1, where the
matrix S with column i equal to vi is the invertible matrix of similarity. S 1 AS = D. S is unitary.
(Lef tarrow) Suppose A = P DP where P is unitary and D is diagonal.
AA = (P DP )(P DP ) = P DP P D P = P DD P
Also A A = P D DP , but D D = DD .
Theorem 6.20. Let A be a real nn matrix. Then A is symmetric if and only if A is orthogonally equivalent
to a real diagonal matrix.
Theorem 6.21. Let A Mn (F ) be a matrix whose characteristic polynomial splits over F .
1. If F = C, then A is unitarily equivalent to a complex upper triangular matrix.
2. If F = R, then A is orthogonally equivalent to a real upper triangular matrix.
Proof. (1) By Shurs Theorem there is an orthonormal basis = {v1 , v2 , . . . , vn } such that [LA ] = N
where N is a complex upper triangular matrix.
0
Let 0 be the standard ordered basis. Then [LA ] 0 = A. Let Q = [I] . Then N = Q1 AQ. We know
that Q is unitary since its columns are an orthonormal set of vectors and so Q Q = I.

Section 6.6
Definition. If V = W1 W2 , then a linear operator T on V is the projection on W1 along W2 if whenever
x = x1 + x2 , with x1 W1 and x2 W2 , we have T (x) = x1 . In this case,
R(T ) = W1 = {x V : T (x) = x} and N (T ) = W2 .
We refer to T as the projection. Let V be an inner product space, and let T : V V be a projection. We
say that T is an orthogonal projection if R(T ) = N (T ) and N (T ) = R(T ).
14

Theorem 6.24. Let V be an inner product space, and let T be a linear operator on V . Then T is an
orthogonal projection if and only if T has an adjoint T and T 2 = T = T .
Compare Theorem 6.24 to Theorem 6.9 where V is finite-dimensional. This is the non-finite dimensional
version.
Theorem 6.25. (The Spectral Theorem). Suppose that T is a linear operator on a finite-dimensional inner
product space V over F with the distinct eigenvalues 1 , 2 , . . . , k . Assume that T is normal if F = C and
that T is self-adjoint if F = R. For each i(1 i k), let Wi be the eigenspace of T corresponding to the
eigenvalue i , and let Ti be the orthogonal projection of V on Wi . Then the following statements are true.
(a) V = W1 W2 Wk .
(b) If Wi0 denotes the direct sum of the subspaces Wj for j 6= i, then Wi = Wi0 .
(c) Ti Tj = i,j Ti for 1 i, j k.
(d) I = T1 + T2 + + Tk .
(e) T = 1 T1 + 2 T2 + + k Tk .
Proof. Assume F = C.
(a) T is normal. By Theorem 6.16 there exists an orthonormal basis of eigenvectors of T . By Theorem 5.10,
V = W1 W2 Wk .
(b) Let x Wi and y Wj , i 6= j.
Then hx, yi = 0 and so Wi0 = Wi .
But from (1),
X
dim(Wi0 ) =
dim(Wj ) = dim(V ) dim(Wi ).
j6=i

By Theorem 6.7(c), we know also that


dim(Wi ) = dim(V ) dim(Wi ).
Hence Wi0 = Wi .
(c) Ti is the orthogonal projection of T on Wi . For i 6= j, x V , x = w1 + w2 + + wk , wi Wi .
Ti (x) =
Tj Ti (x) =
Ti (wi ) = wi
Ti (Ti (x)) = Ti (x)

wi
Tj (wi ) = 0

(d)
(T1 + T2 + + Tk )(x)

= T1 (x) + T2 (x) + + Tk (x)


= w1 + w2 + + wk = x

(e) Let x = T1 (x) + T2 (x) + + Tk (x).


So, T (x) = T (T1 (x)) + T (T2 (x)) + + T (Tk (x)).
For all i, Ti (x) Wi . So
T (Ti (x)) =
=
=

i Ti (x)
1 T1 (x) + 2 T2 (x) + + k Tk (x)
(1 T1 + 2 T2 + + k Tk )(x).

15

Definition. The set {1 , 2 , . . . , k } of eigenvalues of T is called the spectrum of T , the sum I = T1 + T2 +


+ Tk is called the resolution of the identity operator induced by T , and the sum T = 1 T1 + 2 T2 +
+ k Tk is called the spectral decomposision of T .
Corollary 1. If F = C, then T is normal if and only if T = g(T ) for some polynomial g.
Corollary 2. If F = C, then T is unitary if and only if T is normal and || = 1 for every eigenvalue of
T.
Corollary 3. If F = C and T is normal, then T is self-adjoint if and only if every eigenvalue of T is real.
Corollary 4. Let T be as in the spectral theorem with spectral decomposition T = 1 T1 + 2 T2 + + k Tk .
Then each Tj is a polynomial in T .

16

You might also like