MA212 LT Notes
MA212 LT Notes
Dr James Ward
Lecture 1, page 1
Vector spaces
Lecture 1, page 3
Vector spaces
Lecture 1, page 4
Vector spaces
Also, every vector in the vector space has an inverse under the
operation of vector addition
Inverse vectors: For all v 2 V , there is a vector v 2 V such
that v + ( v ) = 0.
It should come as no surprise that the vector v is the vector v
multiplied by the scalar 1, i.e. v = ( 1)v .
Lecture 1, page 5
Vector spaces
⌅ u + v = v + u,
⌅ u + (v + w ) = (u + v ) + w ,
⌅ ↵(u + v ) = ↵u + ↵v ,
⌅ (↵ + )u = ↵u + u,
⌅ ↵( u) = (↵ u).
And these should all make sense given what you know about how
vectors and scalars behave.
Lecture 1, page 6
Example
The set 80 1 9
>
< a 1 >
=
3 B C
R = @a 2 A a 1 , a 2 , a 3 2 R ,
>
: >
;
a3
where vector addition and scalar multiplication are defined by
0 1 0 1 0 1 0 1 0 1
a1 b1 a1 + b1 a1 ↵a1
B C B C B C B C B C
@a2 A + @b2 A = @a2 + b2 A and ↵ @a2 A = @↵a2 A ,
a3 b3 a3 + b3 a3 ↵a3
Lecture 1, page 7
Example
The set ( ! )
a1 a2
a1 , a2 , a3 , a4 2 R ,
a3 a4
where vector addition and scalar multiplication are defined by
! ! !
a1 a2 b1 b2 a1 + b1 a2 + b2
+ =
a3 a4 b3 b4 a3 + b3 a4 + b4
and ! !
a1 a2 ↵a1 ↵a2
↵ = ,
a3 a4 ↵a3 ↵a4
is a real vector space.
!
0 0
Observe that the zero vector in this space is 0 = .
0 0
Lecture 1, page 8
Example
Lecture 1, page 9
Subspaces
Lecture 1, page 10
Example
is a subspace of R3 .
② CUVA : take
any fyz ) (few )
,
e u ,
so 2x y -12=0 and
-
zu -
U -1W -
-
t ( 2K
=
U is
-
o CUVA
-
=
③×µz )
Cush
EU and AER . XX -
dy tot =
aux y -12-1=0
-
u is CUSH
Lecture 1, page 11
Example
ftge U
So f- E EO , D →R .
Cts .
Lf C- U
so U is a
subspace of Flo D .
Lecture 1, page 12
Linear independence
↵ 1 v1 + ↵ 2 v2 + · · · + ↵ k vk = 0
is ↵1 = ↵2 = · · · = ↵k = 0.
Lecture 1, page 13
Example
is linearly dependent.
-2
(}) t 3
( §) t
tf ! ) =/ ! )
so not just the trivial solution LD
Lecture 1, page 14
Example
is linearly independent.
let a
( }) t da
f! ) t 23
/! ) -
-
o
21--0
122--0
{ 341-1292=0
Lecture 1, page 15
A test for linear independence
linearly independent?
-
I
d- Rz
)
Rz -7 l
(
l O
no 200 rows in REE
-
- o , z
Rz -7 -
Rs LI
O O I
Lecture 1, page 17
Example
linearly independent?
( ! ! ! )r3→m-R3- ( g ! ! ) ( g ! :o)
'
HI, '
( y
Rz -74k
I 3 I
→
o i Yz the set is LD
O O O
Lecture 1, page 18
More tests for linear independence
Lecture 1, page 19
Other thoughts on linear independence
When we are just dealing with two vectors, we also have the scalar
multiple test for linear independence.
Also remember that any set that contains the zero vector is linearly
dependent.
Lecture 1, page 20
Linear combinations
S = {u1 , u2 , . . . , un }.
↵ 1 u1 + ↵ 2 u2 + · · · + ↵ n un for scalars ↵1 , ↵2 , . . . , ↵n
Lecture 1, page 21
Linear span
S = {u1 , u2 , . . . , un }.
Lecture 1, page 22
Linear spans give subspaces
Lecture 1, page 23
Example
Lecture 1, page 24
Example
Lecture 1, page 25
Linear independence and linear span
Lastly, we can use linear spans to say some useful things about
linear independence.
Lecture 1, page 26
MA212 Further Mathematical Methods
Lecture 2: More revision and something new
Dr James Ward
Lecture 2, page 1
Information
⌅ Exercises 11 is on Moodle.
I Attempt questions 2, 3, 5, 6 and 9.
I Revision of MA100 linear algebra you should know.
I Submit your written solutions on Moodle by Friday
or follow any instructions laid down by your class teacher.
Lecture 2, page 2
Bases
Lecture 2, page 3
Example
Lecture 2, page 4
Finite-dimensional vector spaces
Lecture 2, page 5
Dimension
Lecture 2, page 6
Example (Just for fun!)
Lecture 2, page 7
Functions: Developing a test for linear independence
Lecture 2, page 8
Looking...
A fo )
suppose { fi is LD then aft Kfz + asf, o must have a
. ,
-
.
non -
trivial solution ( ie .
at least one of a .
As as to )
aifi'M tdzfz
" "
CX) tazfz Cx) -
O
l::* :* :*:L H :L
fix) fix) fs IX) O
must be 0 .
Lecture 2, page 9
Concluding...
!!! , *
Is:
if ff f. to }
, ,
is LD , then WH)=0 for all XEEO .
D
contrapositive
:
if WCXHO ,
then {fi fz.to }
.
is LI
Lecture 2, page 10
The Wronskian
for x 2 D.
That is, the Wronskian has a column for each function, the first
row is just the functions and each successive row is the derivative
of the previous row until we get enough rows for a square matrix.
Lecture 2, page 11
The Wronskian test
Note: To apply this test, we just need to find at least one value
of x in the domain that gives us a non-zero Wronskian.
Lecture 2, page 12
Example
2×2+1 /
x4xtI
(
" x -11
WH) -
- O
l
= 2 to for at least one AER
O O
{ fi fz fz }
, ,
is LI
Lecture 2, page 13
Example
( I
I Sihx sihzx
WCX)
Stl -45nA)
=
= -
C- sin x ) czcoszx )
O cost zcoszx
o -
Sim -
48in21
,
= -
say x= -9 ,
W (E) = . .
-
=
-2=10
next to -
Lecture 2, page 14
A warning!
BUT
If W (x) 6= 0 for no values of x in the domain, i.e.
Lecture 2, page 15
Example
Show that W (x) = 0 for all x 2 R but the set {f1 , f2 } is linearly
independent.
any
II! I:/
xxo.mx' o
o.mn/I;g/--o
-
any "
-
- -
-
.
af, t
tzfz -
O , for all XEIR , x. fix ) -1 Afzal) -
O
At X= -
I, Alf , L I) t
-
so
only solution is Aida = o
{fi fz }
, is LI
Lecture 2, page 16
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Exercises 11: Assumed background
For these and all other exercises on this course, you must show all your working.
1. Determine if the following matrices have inverses and, if they do, find them.
0 1
0 1 0 1 1 1 0 0 0 1
1 1 1 1 1 0 B0 0 1 2 1
1 1C
(i) @1 2 3 A , (ii) @0 1 1A , (iii) B @1 0
C , (iv) @1 4 1 A.
0 1A
1 3 5 1 0 1 1 8 1
0 1 1 0
For any set that is not linearly independent, find a basis for the subspace spanned by the set.
3. For each n
✓ 2, find a basis B for the subspace
V = {(v1 , v2 , . . . , vn )t 2 Rn : v1 + · · · + vn = 0},
of Rn . You should show that your set B is linearly independent and spans V .
4. Calculate the rank of the following matrix and find a basis for its nullspace.
0 1
1 2 3 1
B2 3 1 2C
B C
B 1 0 7 1C
B C
@3 4 1 3A
1 3 8 1
:
0 1
1 2 1
@2 4 3 A x = b
1 2 2
when (i) b = (1, 2, 1)t and (ii) b = (1, 2, 2)t . In case (i), write down a basis for the subspace of which
the solution set is a translate.
6. Let 1 p, q n be given with p 6= q. Let the n ⇥ n matrix A be defined by aii = 1 for all
1 i n, apq = a, and otherwise aij = 0. The matrix A represents the row operation that adds a
times row q to row p. What is the inverse of A?
✓
9. Suppose we have a matrix A, and we obtain A0 from A by the elementary row operation of
adding the first row of A to the second row. (That is, the second row of A0 is the sum of the first
two rows of A.)
(a) Show that the row space of A is equal to the row space of A0 .
(You need to show that every vector in the row space of A0 is in the row space of A, and vice
versa).
(b) Suppose the columns c10 , . . . , cn0 of A0 satisfy ↵1 c10 + · · · + ↵n cn0 = 0 for some scalars ↵1 , . . . , ↵n .
Show that the columns c1 , . . . , cn of A also satisfy ↵1 c1 + · · · + ↵n cn = 0.
(c) Deduce that, if the column vectors of A are linearly independent, then the column vectors of A0
are also linearly independent.
You should be able to see that (a)-(c) also hold for any other elementary row operation.
Ci ) det l l lo 9) (5-3)-1 113-2 )
2-11=4
-
i. - -
I = I -
so no inverse
Cii ) det = I Cl -
O) t ICI -
O) =
2
I l l
(
-
) )
l
(
l -
I
l I l
f )
adj CA )
- -
Am =
A cot =
y l
- I I
-
I
y l
, ,
l l
-
l l l
l l l -
l
± I d-
( )
then A -1=4
-
adj CA) =
E
iz f -
V
-
± , ±
⑤
deify ly ! ) fly g)
ciii dei i. de' t.int *
9
-
- t in =
" i. i 0
Ii:÷ :
""
÷: ÷:
"" " ""
SO A
-
I = -
I d- Is - I
± - I ± i
inverse
-
no
Is ,
×
Civ ) det =
I .-
L
-
12 ) -
I .
6 -1 In b = -
12
f! ! ! ) off! I ÷ ÷)
AM " admit"
-
-
÷)
* * an" -
' = I I -
. :
-
T
take negative
iii. in :÷÷÷n:÷÷÷n : : : :L
2- (a) 2 l l l l O O O
O O 0
"
l: : : →
e: : : →
l : : : :L →
Iii : :S .si : : :
doesn't
since the REE have a row
of zeros , the vector is
linearly independent
✓
Is:÷÷H : : :÷t
Cc ) The vector can be written as 2 l l l l o O
43
µ;) g) I
-
-
3 .
Suppose B -
fei , ez , - -
-
, en
) is a basis for V .
1+021 1=0
then an the only
=a
a t - - -
t
,
so solution is area = - - -
mm
Hence B is
linearly independent more general proof
Az -_
U2
, , .
, .
Mr
Iv
dim span CB) -
-
dim V
ti : : ¥ :÷÷÷t
2 3
' '
I I
rank CA) =
dim ( RS CA ) ) = 2
✓
g) (g) )
in parametric term '
= ×, + xp
u
(g) )
so a basis of NH ) =
{ )
-
,
✓
f ! § ! / g ! ! ) ( to ! ! / ÷ ! ! )
s .
! §)
Therefore
"
A =
! HI %,
" " "'
Therefore
µ}) (g) (! ) Sw
the solution set is
= t
I :÷÷¥÷÷÷÷÷ ÷:
" "" ...
solution
so no
: : :O .
6 The inverse of derived times
.
A can be by subtracting a how f to row p ,
so A-
"
will be defined by aii - I
for all Isis n
{ da !" "
otherwise
-
key understand
:
how it works
tbh ! ) µ )
7 .
suppose =
a + c
Then we have a c
-
I and thus a = £
{ I ::
- -
'
zatbxzc ,
at .me .
This means u, =
I uz Eds -
- I U4
l :÷ : ÷:
""
-
b 544
-
43=44-2
712=0
714--94
42
Iii: : :÷¥l÷÷ :
Cb ) I 2 3 4 2
9
4
13
T2
-
9
Therefore
I
the equation has
unique solution X 73=-4
-
a
z
-
,
12=-1-4 74=-433
faa :?) fan :÷:a9ai)
s.ca , we can write matrix a- -
:&: a
.
so a-
:b hi .
i. :
Ail Aiz - - -
Aig
Aaj ) t . -
.
t dj ( Ail Air - - -
Aij )
'
row (A ) = a, ( Aa Ar - -
-
Aijtazj ) t
- - -
t dj ( Ail air . . .
Aij )
=
di ( All AK - - -
Aig ) t 92 ( All Ak - - .
Aij ) t Az ( Az , Azz . . .
Azj ) t .
. .
t Nj ( Ail Aiz
- - -
Aij )
=
(911-02) ( Al , Ar - - -
Aij ) t Az ( AH Azz . - -
92J ) t . -
-
t dj ( Ail Air - - -
Aij )
if
'
Cb ) a c, t
'
. . .
+ dncn =
0 then a cat Aa ) t ' .
-
QQ -10 , Az , t - -
-
t tncnt an Azn =D
1- An Azn O
-
, 9,9 t -
- i
t An Ch =D
cc )
if column vectors of A ate
linearly independent ,
the
then
only solution for diaz, f- Az Azz t . - -
+ an Azn O -
is a , -_9z= -
- -
= an =D
,
.
. . .
. a
A
a an =D
linearly independent
'
so A is also
- - - - -
=
.
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Solutions to Exercises 11: Assumed background
This document contains answers to the exercises. There may well be some errors; please do let me
know if you spot any.
1. The first and the third matrices are singular, meaning that there are no inverses. One way to
see this is to note that, in the first matrix, adding the first and the third rows gives twice the second
row. That means that the rows are not linearly independent. Similarly for the third matrix, where
you can spot that the sum of the top two rows equals the sum of the bottom two rows.
0 1
1/2 1/2 1/2
The inverse of the second matrix is @ 1/2 1/2 1/2A.
1/2 1/2 1/2
0 1
1 1/2 1/2
The inverse of the fourth matrix is @ 1/6 0 1/6 A.
1/3 1/2 1/6
Techniques for finding inverses are covered in MA100.
2. For (a), we place the vectors as the rows of a matrix, and carry out row reductions.
0 1 0 1 0 1
2 1 1 1 1 2 1 1 R2 2R1 ! R2 1 2 1 1
B1 C B
1C R1 $ R2 B2 1C B 1C
B 2 1 1 1 C R3 3R1 ! R3 B0 3 3 C
@3 1 0 2A ! @3 1 0 2 A ! @0 5 3 1A
1 2 3 4 1 2 3 4 R4 R1 ! R4 0 0 4 3
0 1 0 1
1 21 1 1 2 1 1
1 B C B 1 1/3C
3 R2 ! R2 B0 11 1/3C R3 + 5R2 ! R3 B0 1 C
! @0 35 1A ! @0 0 2 2/3A
0 0
4 3 0 0 4 3
0 1
1 2 1 1
B
R4 + 2R3 ! R4 B0 1 1 1/3 C
C
! @0 0 2 2/3 A
0 0 0 13/3
The final matrix arising is in “upper triangular form” or “reduced form”, and no zero row has arisen,
so the set of vectors is linearly independent.
For (b), the only way a pair of vectors can be linearly dependent is if one is a multiple of the other;
that’s clearly not the case here, so the pair is linearly independent.
For (c), the set is linearly dependent; here are four ways to demonstrate that.
(i) 2(2, 1, 1, 1)t + 2(1, 2, 1, 1)t 3(3, 1, 0, 2)t + (3, 3, 0, 2)t = (0, 0, 0, 0)t .
(ii) All four of them lie in the 3-dimensional hyperplane normal to ( 2, 0, 1, 3)t .
(iii) The matrix we get by putting these as the rows of a matrix has determinant zero.
(iv) Carrying out row reductions as in (a) above yields a zero row.
The easiest way to find a basis for the space spanned by this set of vectors, let’s call it C, is to notice
that (by part (a)) the first three vectors form a linear independent set and they span Lin(C), so they
form a basis for Lin(C). Another answer can be found by taking the non-zero rows arising at the
end of the row-reduction process.
3. For j = 2, . . . , n, let uj be the vector with a 1 in position j and a 1 in position 1, with all
other entries zero. Let B = {u2 , . . . , un }. Clearly each element of B is in V . We claim that B is a
basis of V .
First we check that B is linearly independent. Suppose we have ↵2 u2 + · · · + ↵n un = 0. Then,
considering the jth co-ordinate (j > 1), we see that all the ui have a 0 in position j except uj , which
has a 1; so ↵j = 0. This means that all of ↵2 , . . . , ↵n are zero, and the set B is indeed linearly
independent.
Now we check that B spans V . We could do this by a dimension argument: Lin(B) is a subspace
of dimension at least n 1, contained in V . As V is not the whole of Rn , the dimension of V is no
greater than n 1, so Lin(B) = V . More directly, if x = (x1 , . . . , xn )t is any vector in V , we have
x1 = x2 x3 · · · xn , and we can write x = x2 u2 + · · · + xn un .
4. Subtracting appropriate multiples of the first row from all the other rows reduces the given
matrix A to 0 1
1 2 3 1
B0 1 5 0C
B C
B0 2 10 0C
B C.
@0 2 10 0A
0 1 5 0
Now we see that all the rows except the first are multiples of each other, and one further round of
row reductions gets the matrix into reduced row-echelon form:
0 1
1 0 7 1
B0 1 5 0C
B C
B=B B 0 0 0 0C.
C
@0 0 0 0A
0 0 0 0
As there are only two non-zero rows in the reduced matrix, the rank of the matrix is 2.
The nullspace of A is the same as the nullspace of B. To write down the solution of the matrix
equation Bx = 0, we treat the variables corresponding to rows without a leading 1 as “free”. So, if
x = (x, y, z, w), the solution is given by x = 7z w and y = 5z for z, w 2 R. This can then be
written as 0 1 0 1 0 1
x 7 1
By C B 5C B0C
B C = zB C + wB C,
@z A @1A @0A
w 0 1
and these two vectors, i.e. (7, 5, 1, 0)t and ( 1, 0, 0, 1)t , form a basis for the nullspace of A.
This represents the equivalent equations x + 2y = 1 and z = 0 which means that the solution set can
be written as 80 1 0 1 0 1 9
< x 1 2 =
@y A = @0A + y @ 1 A : y 2 R .
: ;
z 0 0
Geometrically, this is the line through (1, 0, 0)t in the direction of the vector ( 2, 1, 0)t and so it’s a
translate of the subspace with basis {( 2, 1, 0)t }.
For (ii), repeating the process gives us an inconsistency and so there are no solutions.
6. The inverse of A is the matrix B where bii = 1 for every i, bpq = a, and otherwise bij = 0.
You can either use row operations to find this, or you can note that the operation that “undoes” the
e↵ect of A is the addition of a times row q to row p, i.e. left-multiplication by the matrix B above.
Thus multiplying on the left by BA leaves the original matrix unchanged, so BA = I or B = A 1 .
9. This is an important result, as it underlies our use of row operations to solve all kinds of
practical matrix problems.
For (a), the new second row of A0 is the sum of two rows of A, so it is also in the row space of A.
Therefore all rows of A0 are in the row space of A, and so the row space of A0 is contained in the row
space of A. This holds in reverse too: the second row of A can be obtained from the rows of A0 by
taking the second row of A0 and subtracting the first row. So the row space of A is contained in the
row space of A0 , and we may conclude that the two row spaces are equal.
For (b), we can think of the ↵1 c01 + · · · + ↵n c0n = 0 as a linear equation satisfied by each row of A0 .
In particular, the entries of the first two rows of A0 satisfy the equation, and hence so do the entries
in the second row of A. So the equation is also satisfied by the columns of A.
For (c), we need to show that, if the columns c1 , . . . , cn of A are linearly independent, then the
corresponding columns c01 , . . . , c0n of A0 are linearly independent. This can be done by showing that
the contrapositive of this statement (which is logically equivalent to the original statement) is true.
That is, we show instead that, if the columns c01 , . . . , c0n of A0 are linearly dependent, then the
corresponding columns c1 , . . . , cn of A are linearly dependent.
To see why this is true, consider that if the columns c01 , . . . , c0n of A0 are linearly dependent, then
there are scalars ↵1 , . . . , ↵n , not all zero, such that
↵1 c01 + · · · + ↵n c0n = 0.
But, by (b), we know that this means that these very same scalars ↵1 , . . . , ↵n , not all zero, are also
such that
↵1 c1 + · · · + ↵n cn = 0.
That is, the corresponding columns c1 , . . . , cn of A are linearly dependent, as required.
MA212 Further Mathematical Methods
Lecture 3: Linear transformations and matrices
Dr James Ward
⌅ Linear transformations
⌅ Using matrices to represent linear transformations
⌅ Coordinate vectors and change of basis
Lecture 3, page 1
Linear transformations
Lecture 3, page 3
Example
is a linear transformation.
suppose
( Yu;) µ;)
'
e IR
,
,
lie :II: I
"
167%7 I µ; ) if;)
transformation
= a T is a linear
e at e
-
t = t
. u, , a,
Lecture 3, page 4
Example
T ( ax t Py ) =
A ldxt BY ) =
AA X t BAY = a Ttx) t BT ly )
T is a linear transformation
Lecture 3, page 5
Example
is a linear transformation.
Te aft BS I =
I:@ft PS ) Cx ) dx =
If Kfar t fax ) ) dx
=
dftoflxdxt toffs G) DX = a TH t ft CS)
T is a linear transformation
Lecture 3, page 6
Representing linear transformations by matrices
T (x ) = AT x for all x 2 Rn ,
Lecture 3, page 7
Proof: Any x 2 Rn can be written as
x = ↵1 e1 + ↵2 e2 + · · · + ↵n en = (↵1 , ↵2 , . . . , ↵n )t .
T (x ) = T (↵1 e1 + ↵2 e2 + · · · + ↵n en )
= ↵1 T (e1 ) + ↵2 T (e2 ) + · · · + ↵n T (en )
0 1
0 1 ↵1
B C
B CB ↵2 C
= @T (e1 ) T (e2 ) · · · T (en )A B C
B .. C = AT x
@ . A
↵n
as required.
Lecture 3, page 8
Example
Y) l?)
tf! ) I 491--171
-
-
That ? ? 9) x
a-
AT
Lecture 3, page 9
Generalising what we have seen
Lecture 3, page 10
Coordinate vectors
u = ↵ 1 u1 + ↵ 2 u2 + · · · + ↵ n un ,
Lecture 3, page 11
A quick note on ordered bases...
Lecture 3, page 12
...And coordinate vectors
u = 2u1 + 3u2 .
Lecture 3, page 13
Coordinate vectors: A special case
Lecture 3, page 14
Change of basis
Lecture 3, page 15
Proof: Any u 2 Rn can be written as
u = ↵ 1 u1 + ↵ 2 u2 + · · · + ↵ n un ,
0 1
0 1 ↵1
B C
B CB ↵2 C
u = @u1 u2 · · · un A B C
B .. C = MB [u]B ,
@ . A
↵n
as required.
Lecture 3, page 16
Change of basis continued
Lecture 3, page 17
Representing linear transformations by matrices again
MBK ) B
ACT ,B [x ]B ,
-
[T (x )]C =
T T
input
TH) -
Mc [Tix)) c
output
where ACT ,B 1
= MC A T MB .
MC [T (x )]C = AT MB [x ]B =) [T (x )]C = MC 1 AT MB [x ]B ,
as required.
Lecture 3, page 18
Example
Find ACT ,B .
Lecture 3, page 19
fi y )
4 I 91
f! ! ! )
es . At -
- .
MB -
-
Me
-
-
mi
'
-
-
Eff it
( III )
'
ai
f: it C i 71
'
Mia, mis I
fo : :)
-
= -
- .
-
.
Lecture 3, page 20
Example continued
0 1
4
B C
Verify that this works by finding T @4A using both AT and ACT ,B .
6
Tfg ) Atf ! ) ( T T 7)
4) ( it )
① AT
using
-
-
-
=
-
's
ai
ffg ) ) ai
② using "
.
④ Pre I
-
B
-
. '
④ iii. Is
.
-
-
.
or -
s
'll to + "'
"
ai
it '
ftp.t : : it I Ha
-
-
c -
-
t exact Ic or i -
elitist =
Lecture 3, page 21
Representing linear transformations by matrices yet again
Lecture 3, page 22
Proof: Any u 2 U can be written as
u = ↵ 1 u1 + ↵ 2 u2 + · · · + ↵ n un ,
T (u) = T (↵1 u1 + ↵2 u2 + · · · + ↵n un )
Lecture 3, page 23
Now, suppose that C = (v1 , v2 , . . . , vm ) is the ordered basis for V .
As T (u1 ), T (u2 ), . . . , T (un ) 2 V , we can write them as
Lecture 3, page 24
So, substituting these into our expression for T (u), we get
T (u) = ↵1 (a11 v1 + a12 v2 + · · · + a1m vm ) +
↵2 (a21 v1 + a22 v2 + · · · + a2m vm ) + · · · +
↵n (an1 v1 + an2 v2 + · · · + anm vm )
Lecture 3, page 25
Thus, noting that matrix multiplication gives us
0 1 0 10 1
↵1 a11 + ↵2 a21 + · · · + ↵n an1 a11 a21 · · · an1 ↵1
B C B CB C
B ↵1 a12 + ↵2 a22 + · · · + ↵n an2 C Ba12 a22 · · · an2 C B↵2 C
B C=B . .. C B C
B .. C B . .. ..
. C B .. C ,
@ . A @ . . . A@ . A
↵1 a1m + ↵2 a2m + · · · + ↵n anm a1n a2n · · · anm ↵n
we can write
0 1
B C
[T (u)]C = @[T (u1 )]C [T (u2 )]C ··· [T (un )]C A [u]B ,
as required.
€13
Lecture 3, page 26
Example
Find ACT ,B .
Lecture 3, page 27
Tf ! ) K ) ti )
-
- -
-
to IT Ic -
-
Tok
'
II ) I out if I
-
-
i le c
'
II ) H tilt ¥ pit k
-
-
-
-
i
. ,
,
Lecture 3, page 28
Example
{ Ifans 1Pa
(a) Show that B = (f0 , f1 , f2 ) is anh
ordered
mm
basis of P2 .
(b) Show that the function T : P2 ! P2 given by
is a linear transformation.
(c) Find the matrix AB,B
T .
Lecture 3, page 29
(9) Cfo fi fz)
, ,
is LI
using Wroskian
Cb ) select f.Sept a. BE IR
for f- C- Ah , f -
-
do fo t 9. fit Kfz with No hi AER, , . we can be sure THIER as well since
"
. .
".
-
-
o
*on. .
f! )
TH) Cx) fi cxtt)
-
-
=
Htt =
fit fo Clt)) pi
,
yyfz, ex) =
fzcxtt) -
-
Cath =
442kt I =
fit # t fo
#gyps ,
f !) ,
Lecture 3, page 30
MA212 Further Mathematical Methods
Lecture 4: Similar matrices
Dr James Ward
Lecture 4, page 3
Proof: T : u →v is a linear transformation . That Bd ) = a TCU) t BTN)
claim Kent) -
: -
: -
im the IV GV
is a
subspace of u eu
d
im 4) to as 7109 - o ,
so o c- im 4)
kept) EU ,
Kent) -40 as Ot KENT)
Cura : Oi , Ouimet)
because Tco )
-
- O
⇐u
P A
Ev
Ii -
Tah ) and 02=17142) for some Ui Uze
,
U
Uit Uz -
Tuli) -17142 ) TCU, that
CUVA Ui U2 C-KENT)
-
TL Ui ) O TCU2) =D
: "
-
-
, -
Lecture 4, page 4
Example
Kerch -
-
f ER
'
l T (5) = o
} { Cy ) EIR
-
-
'
l ( 1×7,1=0 } =
Iff ) HR Kay ,
'
}
=
{ (g) TRY x
-
-
y } Ling ( f ) }
=
im th -
-
f ER I T -
-
flu as
} Ling l !) }
-
- - =
=
Lecture 4, page 5
Kernels and images, ranges and null spaces
T (x ) = AT x
{ At 71=0 }
"
Proof: KENT) =
HER l T CX) - o ) = =
N LAT)
TCU) ERM } AM I
"
in CT) = { I IE IR =
{ AT U E IER }
" =
R CAT )
Lecture 4, page 6
Nullity and rank
Note: This agrees with what we know about the nullity and rank
of a matrix.
Lecture 4, page 7
The rank-nullity (or dimension) theorem
Lecture 4, page 8
Similar matrices
Lecture 4, page 10
Proof:
Lecture 4, page 11
Diagonalisability
1
P AP = D,
Ax = x for some x 6= 0
Lecture 4, page 13
Finding eigenvalues
Ax = x =) (A In )x = 0,
|A In | = 0.
Lecture 4, page 14
Characteristic polynomials
pA ( ) = | In A|.
Lecture 4, page 15
Characteristic polynomials, traces and determinants
n n 1
pA ( ) = + an 1 + · · · + a1 + a0 ,
Lecture 4, page 16
Example
1 2
A= .
3 4 IA 1=4 6=-2 -
informal
1×12 Al
=/ ?-31 /
a-ha -47
PALM b N sa z
-
- -
- -
- -
- - -
,
a
Truth -
Ai -_
5 and IAI -
- C- 1590=-2
Lecture 4, page 17
Similarity and characteristic polynomials
Proof:
Lecture 4, page 18
A warning!
have the same characteristic polynomial but they are not similar.
PAIN -
-
HI2 -
A I
=/ to Oa f
=P and PB CN =
HIz B
-
I =/ do I - I =
Paul)
p
-
-
= =
,
B cannot be similar to A
Lecture 4, page 19
Similarity and eigenvalues, traces and determinants
i. e .
Paix) =
KI n -
A I = in t an i
-
am ' t . . .
tant Ao =
ItIn -
B I =
PB Id)
same
-
- -
Tr LA) =
-
an -
I
=
Tr KB) and so A. B have the same trace
"
IAI -
- C 1)
-
do =
IBI and so A. B have the same equivalent
Lecture 4, page 20
Eigenvalues, traces and determinants
of Algebra
pA ( ) = ( 1 )( 2) · · · ( n ).
Lecture 4, page 21
But, expanding out these brackets, we get
n n 1
pA ( ) = ( 1 + 2 + ··· + n) + · · · + ( 1)n 1 2··· n,
Tr(A) = an 1 = 1 + 2 + ··· + n
and
|A| = ( 1)n a0 = ( 1)2n 1 2··· n = 1 2··· n,
as required.
Lecture 4, page 22
example session ( Wh
fix)
{ No
recall : -
ni and fix) -
-
Xzo are functions f. fz :
IR → R
X- O
Afi Afzal =D
Take X=l ,
Cl ) 1- Ax -192=0
F) 42=0
take X - -
I, asf H ) tlzfzti )
,
=
o A, -192103=0 di -
0
Hence ,
only trivial solution di - Az -
-
o .
So {fhfz } is
linearly independent
?
1 . Are these functions LI or LD
'
Cl ) I . X . X LI
CZ ) I . Sim , cos X LI
-
③ 1. sin 2x , Cds X LD cosh = I -
Sinn
(4) I , X. XZ , ( HX )
'
LD ( HXJZ =
It 2X -192
(5) I , X , CHAZ LI
2- lecture 3 page 29
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Exercises 12: Wronskians and Linear Transformations
For these and all other exercises on this course, you must show all your working.
1.
✓ Use the Wronskian to show that the set {ex , xex , x2 ex } is linearly independent.
e↵x , e x , e x .
Hence show that this set of functions is linearly independent if ↵, and are all di↵erent.
[Hint: Use column operations to simplify the determinant before you expand it.]
:
( (
x2 if x 0 0 if x 0
f1 (x) = and f2 (x) = .
0 if x 0 x2 if x 0
Show that the Wronskian, W (x), for these functions is zero for all values of x, but f1 and f2 are
linearly independent.
f (x) = ak xk + · · · + a2 x2 + a1 x + a0 .
(a) Show that P3 is a subspace of the vector space of all functions from R to R.
(b) For j = 0, 1, 2, 3, let gj (x) = xj . Show that B = (g0 , . . . , g3 ) is an ordered basis for P3 .
(a) If B = (1, 0, 1)t , (1, 1, 0)t , (0, 1, 1)t , what is the matrix AB,B
T representing the linear transforma-
tion T with respect to the ordered basis B? [Hint: Question 1 of Exercises 1 may help.]
(b) Given that C = (2, 0, 2)t , (3, 3, 0)t , (1, 4, 3)t is also an ordered basis, show that AC,B
T is the
identity matrix I3 . Why does that happen?
7. In R3 , the basis is changed from the standard ordered basis (e1 , e2 , e3 ) to the ordered basis
B = (1, 2, 3)t , (3, 1, 2)t , (2, 3, 1)t .
Find the matrix Q such that, if x = (x1 , x2 , x3 ) is an element of R3 written in terms of the standard
basis, then Qx is the same element written in terms of the ordered basis B.
det(xI A) = xn + an 1x
n 1
+ · · · + a1 x + a0 ,
Show that ( 1)n a0 is the determinant of A and that an 1 is the trace of A (i.e. the sum of its
diagonal entries).
Subspace :
Bz EIR & 1134/0 i C USAA ; CUVA
-
NH ) tf -1191
I .
denote f. Cx ) =
, fzlx) = Xe
"
. fzcx) -
-
He
fi
'
then f ,
ex ) = ex ,
Ix) -
f ! CX ) ex f " "
Mex
"
= ,
Ix) - e text x ex ,
f- six) = Ze t 2x ex t 2x ex t
e Ze
"
t Xe
"
= Next e't ze
"
Hii:: Ii ÷:&:* I
'
hence WIX) -
-
::*:*: .
Xlzxtx 2)
'
= ( HX ) CR -144 -12) KTX ) LZXTXZ ) XIX -14×-12) -1×42-17) 041T X )
-
t
-
-
nx/
l
/
l l
He Be GEREMIA
x x
a B
-
r) ( p p ) ( f -
o
)
filmed? fix)=eB× fsu,
=
= r =
. . -
semi
-
2 denote
42 BZ p2
. ,
em em
:: : :c:c::::.
l ::: ::: :: l : :÷÷÷÷÷÷.ie
"
"
"
then who = e "
So if at Btr ,
WH) to
" B
I @+ a) em
"
=
e c B a)e (
-
r -
a) em . -
cesta) em ]
T
not so obvious
i
. - - -
-
. .
for BR f-
Of incomplete
Aso ax O t
-
O
Wronskian doesn't work the other around
-
-
,
× way
Hence by definition f .
,
and £ are linearly independent ie .
next -
-
o # LD
at X I 9=0
f
-
-
-
f and fz
.
, LI
at X =
I , Az =
O
4 As X't Ask t ai
.
Ca )
by definition . Bz = Xt Ao
taint ago C- Ps
\ /
Cb) WCHB ) = I X XZ 93 =
" " × " =
" *°
O l 2X 3×2
O O 2 tox
O O O 6
so vectors in 13=180 .
. - i
.
Ss ) are linearly independent
Lin { So Si 92,93 }
, ,
= Ao fo t Alf , t Ask -1 9383 =
433 , so B spans 03
CC ) select f. SEP
3
and a. BE IR
( aft BS ) H) aftx)
'
f! ) ( Ig )
ed ) TC fo ) CX) =
To Cx) =
0 Then ETCSODB -
(TCSDTB =
'
1- ( Sl ) CX ) = 8 , CX) = I -
fo B ,
( Og )
Ft 93) ) B
(µ
T 1941×1--9247--24=291 ITCH B) = -
AFB ( (Tl )
( Og § ! ! )
Hence -
1B (THD B ITCH ) B [TOMB =
* dim ( Ker) t dim ( im) -
- dim 37=4
=D ( dim 1 )
im (D) =
{ felt 3 I beg ) =
f for some SEIP
3
} =
{ ftp./Ait2XAzt372a3--AotAiXtazxI-A3X3 }
=
{ f Elp '
l (Ai -
Ao ) t Raz -
ai ) x + Gas -
Az) Xd -
As 93=0 } = Bz ( dim 3)
②
"
ask.fi/I--ATfy/=fI/=fg/ ,
me
! !) . Mei -
tf! ! ! )
µ ) Atf ! ) (g) ( %) ( § ! !)
's
AE ME'A¥BmB=
( ) '
- -
- - =
- .
-
r
method
( I ( Y)
standard
µ) =/ ! )
this
( 4)
T = AT = is the
Hence AFB
( 2g § ! )
.
f! ) ( g ) Tf ! ) ( ! )
②
(g) ( Yo )
'
Cb) T T AFB '
=
=
; i =
want : = Mi AT MB Is -
-
, , ,
instead of finding Mi!
Therefore AFB =
Is check AT MB -
-
Mo
melt it .
mis -
ft ! ! ) me
÷÷÷÷÷ )
ill : :L -4 ! ! ;)
'
Ii i Hi II
'
* -
-
H ol f :)
'
mi -
-
'
mi -
-
(I ÷ ! )
'
MB -
¥¥¥¥s :
-
'
,
' X
( Ig
:B
at (I f f ) Oo } )
'
'
Hence = ME At mis -
-
ATB .
*
MB is the matrix converts standard coordinates into B coordinates
÷!)
motto
l 's ! ! ) ! !) ! ;)
" " .
. " m.
f } ! ! ) Coo ! ! ) f } ! ;)
'
men
-
a- mi- -
-
-
-
-
8 . (* proof )
since Ao -
-
fu ) ,
then def C-A) = Ao
"
and by both sides dell A) det CA )
multiplying by C- IS ao
=
i =
-
-
so dei LA ) C- lmao
-
for the
polynomial . rewrite it into ex -
an ) . . -
ex -
Ann )
the TNA )
-
This document contains answers to the exercises. There may well be some errors; please do let me
know if you spot any.
We don’t need to evaluate this determinant in general. All we need to know is that the Wronskian
is non-zero for at least one value of x. The easiest value to consider is x = 0.
1 0 0
W (0) = e0 1 1 0 = 2.
1 2 2
The fact that W (x) is not identically zero means that the three functions are linearly independent.
e↵x e x e x 1 1 1
W (x) = ↵e↵x e x e x = e↵x e x e x
↵ .
↵2 e↵x 2e x 2e x ↵2 2 2
This last determinant is independent of x, so we need to show that it’s non-zero whenever ↵, , are
all di↵erent. (Note: showing that the Wronskian is zero whenever some pair are equal – why is this
obvious from the start? – is not the same thing at all.)
As in the hint, perhaps the easiest way to do this is via columns operations.
1 1 1 0 0 1
↵ = ↵
↵2 2 2 ↵2 2 2 2 2
↵ 1 1
= = (↵ )( )
↵2 2 2 2 ↵+ +
= (↵ )( )( ↵),
If you don’t think it’s obvious that this expression is non-zero, then it’s fundamentally dishonest to
write “therefore W (0) 6= 0 as long as ↵, and are all di↵erent”. But, one way forward is to note
that, since W (0) is zero if (say) = , then is a factor of W (0). That is, we can now write:
⇥ ⇤
W (0) = ( ) + ↵2 ↵ ↵ =( )( ↵)( ↵).
0 x2 x2 0
W (x) = and W (x) = ,
0 2x 2x 0
4. Before we get going, it’s important to understand what the set P3 consists of. It’s a set, whose
elements are functions: in particular, it’s a subset of the vector space of all functions from R to R.
A function from R to R is in the space P3 if it can be written as a polynomial of degree at most 3.
So, for instance, the function f given by f (x) = 3x3 x + 11 is in P3 , as is the constant function f
given by f (x) = 9. The general form of a function f in P3 is f (x) = a3 x3 + a2 x2 + a1 x + a0 , where
the ai are real numbers.
I want to contrast the notations f and f (x): f denotes a function (an element of P3 , in this question),
while f (x) denotes the value of that function at the real number x — so f (x) is a real number,
depending on f and x. When we talk about adding two functions, we mean adding them point-wise:
the function f + g is defined by saying that its value (f + g)(x) at the real number x is given by
(f + g)(x) = f (x) + g(x).
(a) We need to show that, if f and g are two elements of P3 , then f + g is in P3 , as is any scalar
multiple of f .
Suppose that f (x) = c3 x3 + c2 x2 + c1 x + c0 and g(x) = d3 x3 + d2 x2 + d1 x + d0 ; then we have
(f + g)(x) = (c3 + d3 )x3 + · · · + (c0 + d0 ), so f + g is indeed in P3 . Also, for any real number a,
(af )(x) = (ac3 )x3 + · · · + ac0 , so af is in P3 .
(b) By its definition, P3 is the linear span of the set B, so B spans P3 . To check that B is linearly
independent, one good way is to evaluate the Wronskian. (Details omitted, but it’s easy, and
very similar to an example in lectures.)
(c) We need to observe that, for f and g in P3 , D(f + g) = (f + g)0 = f 0 + g 0 = D(f ) + D(g).
Also, for a 2 R, D(af ) = (af )0 = af 0 = aD(f ). We also need to notice that, whenever f is a
polynomial of degree at most 3, so is f 0 : indeed f 0 is a polynomial of degree at most 2.
(d) The vector space P3 has dimension 4, so we know that the function D we’ve just shown to be a
linear transformation from this space to itself will be represented by a 4 ⇥ 4 matrix AB,B
D , with
respect to the ordered basis B.
We have to think of functions in P3 in terms of their “B-coordinates”: we know that every
function in P3 can be expressed uniquely as a linear combination ↵0 g0 + · · · + ↵3 g3 of the basis
elements: that means its B-coordinates are (↵0 , . . . , ↵3 )t . The matrix takes the set of coordinates
corresponding to a function f , and maps them to the coordinates of the function D(f ).
For instance, in terms of the ordered basis B, the function g3 (x) = x3 has co-ordinates (0, 0, 0, 1)t ,
meaning that g3 = 0 ⇥ g0 + · · · + 1 ⇥ g3 . Now D(g3 )(x) is the derivative of the function x3 , which
is 3x2 = 3g2 (x). In terms of the basis B, D(g3 ) has co-ordinates (0, 0, 3, 0)t . This means that
the matrix AB,B
D representing D with respect to B must map (0, 0, 0, 1)t to (0, 0, 3, 0)t . Thinking
about how matrix multiplication works, this means that the fourth column of AB,B D is (0, 0, 3, 0)t .
In general, the ith column is the image of the ith basis vector, written in terms of the basis. We
have D(g0 ) = 0, D(g1 ) = g0 , D(g2 ) = 2g1 , D(g3 ) = 3g2 , so the full matrix is
0 1
0 1 0 0
B0 0 2 0C
AB,B
D =B@0 0 0 3A .
C
0 0 0 0
(e) The kernel of D is the set of functions f in P3 such that D(f ) = 0, and of course this is the set
of constant functions, which is Lin({g0 }).
The image of D is the set of polynomial functions of degree at most 2. We’ve already remarked
that D(f ) is a polynomial of degree at most 2, for all f 2 P3 . There are a number of ways of
showing that every polynomial g of degree at most 2 is in the image of D: the most direct is to
let f be an indefinite integral of g — which is a polynomial of degree at most 3, and note that
D(f ) = f 0 = g, and so g is in the image of D.
There’s a lot of simple arithmetic to be done in the next few exercises. Probably the majority of
students will make some kind of arithmetic error along the way. That’s understandable. You may be
reassured to know that, in an exam, most of the marks are for applying the right method, not for your
accuracy at arithmetic. However, you shouldn’t be too careless, and you are supposed, whenever it’s
reasonable, to check your answers. One particular case is when you are working out the inverse
of a matrix: once you’ve got your answer, it’s easy to check whether it’s right. So you should never
get an inverse wrong in your homework. My rule of thumb on marking is to delete one mark for a
small number of arithmetic errors (within reason), and to delete another mark if the final answer is
wrong and could have been checked easily. You have been warned!
5. For (a), we first form MB , the matrix whose columns are the give vectors, and work out the
inverse. (If you’ve done Question 1 from Exercises 1, then referring back and quoting the answer
from there is legitimate, indeed positively recommended!)
0 1 0 1
1 1 0 1 1 1
1
MB = @0 1 1A and MB 1 = @ 1 1 1A .
2
1 0 1 1 1 1
Now
0 1 0 10 1
1 1 1 4 2 0 1 1 0
1@ 1
AB,B
T
1
= M B AT M B = 1 1 1A @ 1 7 1 A @ 0 1 1A
2 2
1 1 1 1 1 5 1 0 1
0 10 1
1 1 1 4 6 2
1@
= 1 1 1A @0 6 8A
4
1 1 1 4 0 6
0 10 1 0 1
1 1 1 2 3 1 2 0 0
1@ A @ A @
= 1 1 1 0 3 4 = 0 3 1A .
2
1 1 1 2 0 3 0 0 3
Tip: when manipulating matrices, you’ll be much less prone to arithmetic errors if you use integers
as far as possible, as in the calculation above.
For (b), we could just calculate MC 1 , and then verify that AC,BT = MC 1 AT MB is equal to the
identity matrix. A tiny amount of thought will save you a lot of calculation: all we really need to do
is check that AT MB = MC , as indeed it is.
General tip: if you are told what the outcome to some calculation is supposed to be, it is often much
easier to verify the answer than to perform the calculation.
There’s even slightly more to be gained, in the question taken as a whole, by thinking ahead and
planning what we’ll have to calculate before actually doing so. We see that we are asked to work out
AB,B
T = MB 1 AT MB and that in this part we’ll need to know AT MB ; so it pays to do the multiplication
AT MB first when working out AB,B
T — which indeed is what we did in the answer to (a) above. Note:
AT MB is the matrix that takes a vector in B-coordinates and gives its image in standard coordinates
– the question asks us about expressing this image in both B- and C-coordinates.
You were also asked why AC,B
T is the identity matrix. That’s a slightly vague question; in principle
there could be several plausible alternative answers, and it might not be clear to you what the
questioner is looking for. Tip: the one approach guaranteed to attract no credit (which for some
reason seems to be the most popular approach) is to act as though the question isn’t there.
What is the matrix AC,BT supposed to represent? It takes the B-coordinates of a vector v in R3 ,
and returns the C-coordinates of the image T (v ) = AT v . The fact that this matrix is the identity
matrix is equivalent to the observation that AT maps each vector in the basis B to the corresponding
element of the basis C (which you might care to check directly).
6. The easiest way to tackle this is first to find the matrix AT representing T with respect to the
standard bases. We know that AC,B T = MC 1 AT MB , so AT = MC AC,B 1
T MB . Then we can find the
C 0 ,B 0 C,B
matrix AT = MC 01 AT MB 0 = MC 01 MC AT MB 1 MB 0 . Here are the matrices we need. We have
1
MB in the previous answer.
0 1
1 1 0 ✓ ◆ ✓ ◆ ✓ ◆
@ 1 1 2 1 0 1
MB 0 = 0 1 1A , MC = , MC 0 = and 1
MC 0 = .
1 1 1 0 1 2
1 0 1
so that ✓ ◆
0 ,B 0 1 0 1
AC
T
1
= M C 0 AT M B 0 =
5 0 3
As an aside, we can see from the form of AT that v = (1, 1, 0)t is in the kernel of the transformation
T . Let’s see how this is reflected with respect to the other bases.
The vector v is the second element of the basis B 0 , so its B 0 -coordinates are (0, 1, 0)t . This means
0 ,B 0
that we should have AC T (0, 1, 0)t = 0, which is equivalent to saying that the second column of the
matrix is all-zero, as indeed it is.
Similarly, with respect to the basis B, we have v = (1, 0, 1)t (0, 1, 1)t ; i.e. the B-coordinates of v
are (1, 0, 1)t . Therefore we should have AC,B t t
T (1, 0, 1) = (0, 0, 0) , and indeed it is so.
7. Remember (and understand) that MB is the matrix that converts B-coordinates to standard
coordinates.
Here, if the basis in the question is B, you’re asked for the matrix that converts standard coordinates
to B-coordinates, so the matrix Q you are asked for is MB 1 .
0 1 0 1
1 3 2 5 1 7
1 @
MB = @2 1 3A and MB 1 = 7 5 1 A.
18
3 2 1 1 7 5
So, if a vector is written in standard co-ordinates as x = (x1 , x2 , x3 )t , then it is written in the new
co-ordinates as MB 1 x .
8. Since a0 = f (0), setting x to zero gives that a0 = det( A). Remember that multiplying each row
of a matrix by 1 has the e↵ect of multiplying the determinant by 1, so det( A) = ( 1)n det(A).
Imagine expanding out the expression for the determinant of xI A completely: each term will be
a product of entries, either of the form aij or of the form x aii . To get a coefficient for xn 1 in any
of these terms, one must either multiply by diagonal entries exactly n 1 times or exactly n times.
But it is impossible to multiply by diagonal entries exactly n 1 times, since one entry aij o↵ the
diagonal in the product implies that neither x aii nor x ajj will appear in the product.
Therefore the coefficient of xn 1 in the characteristic polynomial comes solely from multiplying all
the diagonal entries, and is equal to the coefficient of xn 1 in the product
(x a11 ) · · · (x ann ). This will be the sum a11 · · · ann , i.e. minus the trace of A.
MA212 Further Mathematical Methods
Lecture 5: Real inner products
Dr James Ward
Note: We will refer to (i), (ii) and (iii) as linearity in the first
argument, symmetry and positivity respectively.
Lecture 5, page 3
Real inner products are bilinear
=
x ( Ui W , t Us Uk ) t B ( Vi Wit lk wz)
'
ang a. WEIR and a BEN
(Mus)
u.
( mud ( Ww: )
>
a B
HI!:p: ) Tn)
=
.
+
is couteau
-
- -
= ace we ,.
+ ecv.ws
dis u I
. =
%) ( T! ) .
=
Ui Vitus U2 = Icu, talk =
( Y! ) Cute)
.
= V - U
( ) (a)
U, UI 2
Ciii ) U U
-
= .
= Uit Us 70
uz
(hug ) ( )
2
Ul
U U- o
'
⇐ -
o Uft Us -
o ⇐ Ui = Use o ⇐ 4=0
"
.
u · v = v t u,
Lecture 5, page 6
Example
=
asf h > , t p < 9 h>
.
Sf Hx) )
-
ciii, off > = DX so
f -
-
o
Lf f s ,
-
-
ft O -
DX -
-
o
<
f.f s = o f! If Pdx -
-
o fix) so for all a Elo I .
] f- o
Lecture 5, page 7
Remark
When we have real vector spaces like Rn or C[0, 1], we usually use
the inner products
Z 1
hu, v i = u · v or hf , g i = f (x)g (x) dx,
0
Lecture 5, page 8
Example A is symmetric .
so At = A
t
!
1 1
Is hu, v i = u t v an inner product on R2 ?
1 2
take any a. U , we R2 and a. BE IR
- A
1×1 matrix CAB 5h - BtAt
( L) ( )
Ul
Cii) cu , a > =
(Ui uz)
l =
UP -1241kt 2h22 =
Cultus ) 't uz > o
, U2
4=0 L U Us - (O -101402--0
.
so this is a IP on B
Lecture 5, page 9
Example
!
1 2
Is hu, v i = u t v an inner product on R2 ?
2 1
I
similarly . A is symmetric,
so At A
-
-
in
particular . A -
⇐ I -40 ,
but < U, Us -
22-212-31-112=-3 so
so this is not an IP on IR
'
Lecture 5, page 10
Norms and angles
for all v 2 V .
Lecture 5, page 11
Unit vectors
and so !
1 1
v̂ = p
10 3
is a unit vector in the same direction as the vector v .
Lecture 5, page 12
The Cauchy-Schwarz inequality
Lecture 5, page 13
Angles
hx , y i
1 1
kx kky k
hx , y i
cos ✓ =
kx kky k
where 0 ✓ ⇡.
Note: This ‘angle’ will only have its usual geometric significance if
we are using the dot product in Rn .
Lecture 5, page 14
Orthogonal vectors
Lecture 5, page 15
Zero vectors and orthogonality
Lecture 5, page 16
The generalised theorem of Pythagoras
kx + y k2 = kx k2 + ky k2 .
kx + y k kx k + ky k.
where 1 i, j n.
For now, the key thing about orthonormal sets is that they are
linearly independent.
Lecture 5, page 20
which, using linearity in the first argument, gives us
LI # t
Lecture 5, page 21
Orthonormal bases
Lecture 5, page 22
Gram-Schmidt orthogonalisation
Lecture 5, page 23
Step 1: Take ŵ1 to be the vector
u1
ŵ1 = .
ku1 k
{ŵ1 } is an orthonormal set which spans Lin{u1 }.
Step 2: Take ŵ2 to be the vector
w2
ŵ2 = where w2 = u2 hu2 , ŵ1 i ŵ1 .
kw2 k
{ŵ1 , ŵ2 } is an orthonormal set which spans Lin{u1 , u2 }.
Step 3: Take ŵ3 to be the vector
w3
ŵ3 = where w 3 = u3 hu3 , ŵ1 i ŵ1 hu3 , ŵ2 i ŵ2 .
kw3 k
{ŵ1 , ŵ2 , ŵ3 } is an orthonormal set which spans Lin{u1 , u2 , u3 }.
Step . . . And so on, following the obvious pattern until...
Lecture 5, page 24
Step n: Take ŵn to be the vector
n 1
X
wn
ŵn = where w n = un hun , ŵi i ŵi
kwn k
i=1
Lecture 5, page 25
Example
Consider the vector space C[0, 1] with for all f , g 2 C[0, 1] the
inner product Z 1
hf , g i = f (x)g (x) dx,
0
.
Let U = Lin{f1 , f2 , f3 } be the subspace of C[0, 1] where
Lecture 5, page 26
Gram -
Schmidt :
1st
step :
" fill
'
=L fi fi
,
> = do I dx EX ) to -
-
=
I Hell = ,
hi =
1¥11 =
fi .
So hi :[0.13 → IR ,
hi CX) -
-
I
Lfz ,
hi > Ihl =
X -
d- * I = X -
E ,
so gz -
-
fa I hi
-
1192112=42 .
Sz > = To Ix -
4)
-
dx= ⇐ Cx 4) 3) -
o
'
=
zi 119211=452
SO ha -
-
I =
TE cfz Zhi )-
so hai EO .
I ] → IR and hzlx) = TE ex -
I)
,
119211
3rd
step : 83=-13 cfs -
,
h, > hi -
of 3 hz > hz ,
hz= ¥1 = to f, where
fs :[oil) → R , fzcx) =P -
X -
I
Dr James Ward
Lecture 6, page 1
Information
⌅ Exercises 13 is on Moodle.
I Attempt questions 2, 4, 5, 9, 10 and 11.
I Follow your class teacher’s submission instructions.
I Do it! Actually submit your homework!
⌅ Classes
I Go to them.
Lecture 6, page 2
Orthogonal matrices
Thus, |A| is 1 or 1.
Of course, as |A| =
6 0, this means that A is invertible.
Lecture 6, page 3
A warning!
IAI = 2x I -
O =/
but at a- =
(f ¥ ) fo : ) ( I %)
= FIN
so A is not orthogonal
Lecture 6, page 4
Other definitions of orthogonal matrices
(3) AAt = In .
(4) The columns of A form an orthonormal set of vectors.
Lecture 6, page 5
Proof of (1) () (2), i.e. At A = In () At = A 1
At A = In =) (At A)A 1
=A 1
=) At (AA 1
)=A 1
,
RTL: If At = A 1, then
At A = A 1
A = In ,
Lecture 6, page 6
Proof of (2) () (3), i.e. At = A 1
() AAt = In
AAt = AA 1
= In ,
as required.
RTL: If AAt = In , then A is invertible as
Thus, we have
1
A (AAt ) = A 1
=) (A 1
A)At = A 1
,
Lecture 6, page 7
Proof of (1) () (4), i.e. At A = In () columns of A are ON
Lecture 6, page 8
Properties of orthogonal matrices
Lecture 6, page 9
Proof of (1) () (2)
(Au) · (Av ) = v t In u = v t u = u · v ,
as required.
RTL: Also, if for all u, v 2 Rn , we have
(Au) · (Av ) = u · v =) v t At Au = v t u = v t In u,
Lecture 6, page 10
Proof of (2) () (3)
Lecture 6, page 12
Remarks about linear transformations
In particular, notice that due to (2) and (3), we can now see that
Let’s see what this means for linear transformations that are
represented by orthogonal matrices.
Lecture 6, page 13
Orthogonal matrices in R2
Lecture 6, page 14
Case 1: First choice of v2
We can take
! !
sin ✓ cos(✓ + ⇡2 )
v2 = =
cos ✓ sin(✓ + ⇡2 )
cos ✓ sin ✓
|A| = = cos2 ✓ + sin2 ✓ = 1.
sin ✓ cos ✓
Lecture 6, page 15
Case 2: Second choice of v2
We can take
! !
⇡
sin ✓ cos(✓ 2)
v2 = = ⇡
cos ✓ sin(✓ 2)
cos ✓ sin ✓
|A| = = cos2 ✓ sin2 ✓ = 1.
sin ✓ cos ✓
Lecture 6, page 16
Orientation
Lecture 6, page 17
Remarks about rotations in R2
Lecture 6, page 18
Remarks about reflections in R2
Lecture 6, page 19
Orthogonal matrices in R3
Lecture 6, page 20
Case 1: A is a rotation by an angle ✓ about an axis
Av = v .
Thus, v 6= 0 is an eigenvector of
A with eigenvalue 1.
Lecture 6, page 21
In particular, this means that A is similar to the matrix
0 1
cos ✓ sin ✓ 0
B C
B = @ sin ✓ cos ✓ 0A .
0 0 1
Observe that |B| = 1 and so |A| = 1 too. That is, this linear
transformation is orientation preserving.
Also, if (u1 , u2 , v̂ ) is an ordered orthonormal basis of R3 where u1
and u2 are vectors in the plane of rotation, A and B are related by
0 1
| | |
B C
B = M t AM with M = @u1 u2 v̂ A .
| | |
Lecture 6, page 22
Case 2: A is a reflection about a plane
Av = v.
Thus, v 6= 0 is an eigenvector of
A with eigenvalue 1.
If u is in the plane of reflection, i.e. u ? v , then Au is just u as it
is not reflected, i.e.
Au = u.
Lecture 6, page 23
In particular, this means that A is similar to the matrix
0 1
1 0 0
B C
B = @0 1 0 A .
0 0 1
Observe that |B| = 1 and so |A| = 1 too. That is, this linear
transformation is orientation reversing.
Also, if (u1 , u2 , v̂ ) is an ordered orthonormal basis of R3 where u1
and u2 are vectors in the plane of reflection, A and B are related by
0 1
| | |
B C
B = M t AM with M = @u1 u2 v̂ A .
| | |
Lecture 6, page 24
Case 3: A is a combination of cases 1 and 2.
Av = v.
Thus, v 6= 0 is an eigenvector of
A with eigenvalue 1.
Observe that |B| = 1 and so |A| = 1 too. That is, this linear
transformation is orientation reversing.
Also, if (u1 , u2 , v̂ ) is an ordered orthonormal basis of R3 where u1
and u2 are vectors in the common plane of rotation and reflection,
A and B are related by
0 1
| | |
B C
B = M t AM with M = @u1 u2 v̂ A .
| | |
Lecture 6, page 26
extra -
example session ( W 37
I .
from lecture b- ,
slides 26
A P AP
'
i.e
-
-
.
,
z .
show :
if A is similar to In , then A In -
-
Inp =P P In
' '
there's invertible matrix P A P
- -
-1
-
an
-
s .
.
=
Ca )
'
3 .
suppose that A and B are similar , show AZ and B are similar too
'
there's A P BP
-
"
then AZ BZP
'
AA ' '
(P BP ) ( P Bp) 1) Bp
-
P
-
P
-
=
B Cpp
-
= =
=
A and
'
Cb )
assuming that A. B invertible then At and B
I
are similar
-
are , .
'
At ( p IBP 5 p
'
B- Ip
-
'
-
A- =p BP then
-
- =
I
.
so B similar recall
'
are
-
:
.
IAI to
show that invertible matrix then B invertible
CC )
if A is an .
T O
is
' K t t
I
I PAP ' I
-
PAP
-
B fo
-
-
-
- -
-
-
= -
recall :
Ix 'll =
txt l 'll
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Exercises 13: Similar matrices and real inner products
For these and all other exercises on this course, you must show all your working.
1. Suppose that A and B are similar matrices. Show that A2 and B 2 are similar. If A is invertible,
show that B is also invertible and that A 1 and B 1 are similar.
2. (a) Suppose that A and B are n ⇥ n matrices, and that A is invertible. Show that the matrices
✓
AB and BA are similar.
(b) Suppose that the n ⇥ n matrices S and T are similar. Show that there are matrices A and B,
with A invertible, such that S = AB and T = BA.
✓ ◆ ✓ ◆
1 1 1 1
(c) Let A = and B = . Show that AB is not similar to BA.
0 0 1 1
AB = BA. hemn
Wp cop
ro
,
✓
4. Suppose that A is an n ⇥ m matrix, and B is an m ⇥ k matrix.
(a) Show that R(AB) ✓ R(A) and N (B) ✓ N (AB).
(b) If also R(B) = Rm , show that R(AB) = R(A).
(c) Deduce that, if A is an n ⇥ m matrix, and B is an m ⇥ n matrix, with m < n, then AB is singular.
✓
5. Verify that
h(x1 , x2 )t , (y1 , y2 )t i = x1 y1 x 1 y2 x2 y1 + 3x2 y2 .
is an inner product in R2 .
Does h(x1 , x2 )t , (y1 , y2 )t i = x1 y1 x 1 y2 x2 y1 + x2 y2 also give an inner product on R2 ?
7. Use the Gram-Schmidt process to find an orthonormal basis for the subspace
080 1 0 191
< 2 2 =
Lin @ @1 , 0A A
A @
: ;
2 1
of R3 .
9. A function f : R ! R is odd if f ( x) = f (x) for all x, and even if f ( x) = f (x) for all x.
✓
Show that, if f is odd and g is even, then
Z 1
f (x)g(x) dx = 0.
1
10. Let P2 [ 1, 1] be the space of all real polynomial functions on [ 1, 1] with degree two or less.
:
The 3-tuple of functions (1, x, x2 ) is an ordered basis of this space, as in Question 4 of Exercises 2.
An inner product is defined on P2 [ 1, 1] by
Z 1
hf, gi = f (x)g(x) dx.
1
11. (a) Suppose A is an n ⇥ n matrix with the property that x t Ay = 0 for all x , y in Rn . Show
that A = 0, i.e. every entry aij of A is equal to 0.
(b) Deduce that, if A and B are n ⇥ n matrices with the property that x t Ay = x t By for all x , y in
Rn , then A = B.
(c) Give an example of a non-zero 2 ⇥ 2 matrix A such that x t Ax = 0 for all x in R2 .
i . Since A and B are similar matrices ,
A =
PBP
"
then "
PBP PBP
' " '
AE PB p -
I
-
AP BP PB p
-
AA
- -
i.e
-
-
= = .
'
sa A and B are similar
'
and ( S ' Ats ) B
' ' '
I
-
S A' SS AS
-
13=5 As B is invertible
-
given similarity so
-
- =
. ,
'
also similar to A -1
"
BY S A which is
-
since A is invertible ,
=
s . .
-
- = =
A C BA ) A
A- SA and
'
cb ) since s and T are similar ,
then there exists an invertible matrix St .
T = S -
-
"
ATA
A s -
-
-
. .
,
'
and T A- SA At ABA In BA =
BAU
- =
=
Co ill : : ) l: : ) fi : ) ( II 'll It
-
an and BA
'
-
-
-
-
-
✓
Pt
( f f) p =
(Oooo) tf ! I ) -
BA
so AB is not similar to BA
3 .
denote Di -
SAS
"
and Dz -
SBS -1 , so A- 5'D, s and B - 5 Dzs
'
'
-
and
-
S
-
S S Dilks BA S Da Dis
-
=
-
-
=
-
-
SYD Das
AB=BA✓
'
then S DzDis
-
i.e
-
, ,
.
th
F) ) TF }
B A
))
→
( n →
A- =
n
.
B
-
-
m
,
AB =
Rk 1pm Rn
"
4 ,
Ca) A B becomes
-
a nxk matrix B R :
"
→ Rm ; A :
RM → Rn ; AB :
IR → Rn
}
" "
{
"
RCAB ) =
AB -
{ A.IE/RnlIElR }
M
RCA) -
-
let Z BX -
-
,
then y ABX - -
AZ . so YERLA )
}
M
N (B) =
{ XEIRK I Boxee Cb) let UENCB ) .
by definition , BI -
-
I then ABI =
AI -
-
}
n
I SO UENCAB )
A-
let IE NCB ) , then BA Om and
-
- thus ABN = AIM =
In B- →) RCA
RCB)
so N ( B) E N CAB )
Cb )
by definition , reps -
{ Bae
m
, uepkg.am
let I C- RCA ) , then P =
AI for some I c- 1pm =
Reps)
I I c- Rk Srt U
Bud and thus P All ABI E RCAB)
-
so =
-
.
-
hence RCASERCAB)
def :
AB is non singular iff p CAB) = n CP is rank)
know :
PLAB ) e PLA) smh n
hence PLAB ) s n . so AB is
singular
5 .
take any x. g. ZEN and a. BER
ch ca
-18µL ) (I:) .
> =
I!! !!! ) (Ea) .
>
-
-
laxity IZ , ,
-
Kitty, ) Zz -
ftp.Z/t3dXzZzt3BYzZ2--dLXiZI-XiZz-XzZI-3X2Zz
= A XIZ, t By , Zi -
XXI Zz -
BY , Zz -
AXZZ , -
) t B ( Yi Zi -
Y , Zz -
YZZ, -13922-2)
= "
It # .
s te
' .
,
dammit iy.y.us ,
.
* g. .
Xi Yz -
Xzy , t 3 Xsyz
=
Yi Xi -
Y, Xs -
Yz Xt t 392×2
ysyt
=
a Cy , . Hi ,
Xz
t
Xslt XP 2×1×21-3 XI }
'
(3) L (Xi X 2)
, ,
( Xi ,
s = Xi Xi -
Xi Ka -
X 2X , t 3XzXz =
-
=
( Xi -
Xz) t 2X 30
- -
> =
,
, . , ,
Overall ,
L ( Xi ,
Xzlt , Cy , , yet > = X. y, -
Xzy , -
Xi Yz t 3×292
However ,
L ( Xi , y , )t , Cy , .
Yzjt > = Xi Yi -
Xiyz -
for property 3 ,
let
4=4 ) ,
, then <
I . I > = 0 . which violates the rule that
f! ¥7417 ¥
effs Rdx
Iggy
O
fix) floro then
=
Take
-
= =
X
-
-
=
,
-
, ,
o xx
-
: -
-
, , .
f )
z
7 .
let 4 -
and be o
,
2 I
11411 -
-
evi =
=3
soni -
-
¥n= 's
2
( y ) (1)
n
< Us ,
W, > = .
Is = It -5=2
sow. o. a. wi > wi -
fo ) tf ! ) 4,1-(4%3)=41: )
-
- - -
- -
HWzlt-LW2.ws =
I
therefore wi -
iffy, = -5 I
an orthonormal basis for the subspace is
{ uh uh }
,
I I
l: )
o
-
t
8 '
: 'ere
x
.
x
ime
!f ) }
'
" "Mt
Ling
-
-
take Ui =
(l l -
z 05h the l l l -2 0 It UE Bo l -2 4 I -351 by GS
( check Vita)
Us =
# ( 2,0 , I ,
-
15h 64=1370 ( 1.3.24 )
by GS
therefore . { Ui ,
Oz ,
Is , Ug} form an
orthogonal basis of 1124 s -
t
. it and Oz
form
a basis of the kernel
same thing
I
{ (g) (E) }
run) -
-
as cm) -
Lin
,
(g ) try
({ ) Bto L2 -1.5T
'
take then Ui Uz GS ( check Uit Uz)
u,
by
- -
= - -
.
,
take
any Et f ch Uz }
µ) U, ) U , Uz ) U2
43 say Uz ( Us ( Us
Lin wz Uz
-
'
- - - -
. ,
=
Us =
WSH
Wsh
therefore { us
,
,
U2 , Us } form an orthonormal basis of R ' sit .
Ui . Uz form a basis of the image
9 .
I! " fix gcx) dx = ft th) Stx) dx t f! fans dx
by the
property of symmetry
+ x) Stx) dx
-
since fi x) -
=
-
10 .
so wi -
-
1¥ E ,
-
- -
-
F-
Secondly ,
we Uz -
cuz ,
wi > Tv ,
= Uz -
I xFd⇐
, x -
⇐ x' THE =
x -
o =
x
[3×3] !, = I -1 I =
-32 ,
so 11 Wall = ¥
so uh = b-
11W 211
=
÷ =
F- X = ¥ Uz
thirdly ,
Wz =
Us -
< Us ,
uh > Wi -
< Us, WI S WI
E
=
X
'
-
Stix '
dx -
H ,
x
'
dx )x
R E
Lf Xs ] !, [ ¥x4 ] !,
'
-
= X -
-
z -
x
=
x2 -
( Ex 2) E -
x EI ¥ ] -
=P -
I
f't
l l '
Ii
'
Z
Ht
J ) DX FF f¥
' -
II wash =
sws.ws > =
LK - =
TX
-
g-
× =
,
so 11 Wz he
8¥
Xd
SO if, WI J
t) ferox ¥0
-
-
'
=
(x
= -
= =
-
H Wsh
ii. Ca) A- ( )) n t x.
y C- Rn Xt Ay O w A-I
-
-
, ,
( )( )
suppose Ato so there Iaij to so I ai ) ai to contradiction
aig
.
, .
Cb) Xt Ay -
-
Xt By
x' Ay Xt
By o
-
- -
Xt LA B) y
- -
-
O €7 A B - -
o , so A - B
(I
'
A-
)
x )
, o
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Solutions to Exercises 13: Similar matrices and real inner products
This document contains answers to the exercises. There may well be some errors; please do let me
know if you spot any.
1. Suppose that A and B are similar, and that A = SBS 1 for some invertible matrix S. Then
we have A2 = (SBS 1 )(SBS 1 ) = SB(S 1 S)BS 1 = SB 2 S 1, so A2 and B 2 are similar.
Suppose now that A is invertible. We can write B = S 1 AS, and note that (S 1 A 1 S)B =
S 1 A 1 SS 1 AS = I, so B is invertible, with inverse B 1 = S 1 A 1 S, which is similar to A 1 .
2. (a) Suppose that A has an inverse A 1. We can write AB = AB(AA 1) = A(BA)A 1, showing
that AB and BA are similar.
(b) Suppose S and T are similar, with S = P 1T P . Then we can set A = P and B = P 1T , so that
A is invertible, AB = T and BA = S.
In other words, two n ⇥ n matrices are similar if and only if they can be written as AB and BA for
some pair of n ⇥ n matrices A and B, with A invertible.
(c) Here AB is the zero matrix, while BA isn’t. Any matrix S(AB)S 1 similar to AB = 0 will also
be the zero matrix, so in particular AB and BA are not similar.
4. (a) Take any vector y in the range of AB; so y = ABx for some x 2 Rk . Now y = Az where
z = Bx 2 Rm , so y 2 R(A). This proves that R(AB) ✓ R(A).
For any vector u 2 N (B), we have Bu = 0, and so ABu = 0, so u 2 N (AB). This shows that
N (B) ✓ N (AB).
(b) Now suppose that R(B) = Rm : we need to show that the range of A is contained in the range of
AB. So, take a vector z in the range of A: that means that Ax = z for some x 2 Rm . As the range
of B is Rm , there is some y 2 Rk with By = x . Putting this together, we have ABy = Ax = z , so
z is in the range of AB, which is what we needed.
(c) Finally, we note that, if A is an n ⇥ m matrix, then the rank ⇢(AB) of AB, which is the dimension
of R(AB), is at most the dimension ⇢(A) of R(A), and this is at most m, since R(A) is spanned by
the columns of A. If AB is an n ⇥ n matrix and m < n, this implies that AB is singular.
t
5. ✓ is linear◆ in the first variables, since it can be written in the form hx , y i = x Ay ,
The function
1 1
where here A = .
1 3
The function can be seen to be symmetric, i.e, hx , y i = hy , x i, directly from its definition; you’ll see
that this is e↵ectively the same thing as checking that the matrix A above is symmetric.
You do need to mention these points in your answer, but the main thing is to check positivity. For
that, it is not enough to choose a particular x and check that hx , x i 0: you need to demonstrate
this for all vectors x .
We have h(x1 , x2 )t , (x1 , x2 )t i = x21 2x1 x2 + 3x22 and as this is just (x1 x2 )2 + 2x22 it is always
non-negative. Also, if this is zero, we must have x1 = x2 and x2 = 0, which gives us x1 = x2 = 0.
And, conversely, if x1 = x2 = 0, this just gives us zero.
So the first function we are given does indeed give us an inner product in R2 .
For the second function we are given, we don’t have an inner product in R2 as positivity fails. That is,
even though we do always have h(x1 , x2 )t , (x1 , x2 )t i = (x1 x2 )2 0, the point is that this expression
can be equal to zero for non-zero x : for instance, h(1, 1)t , (1, 1)t i = 0.
Note: giving an example, like the vector (1, 1)t above, is the most convincing way to answer the
question.
6. It’s perhaps illuminating to do the second part first: explaining why each of (i)-(iii) fails to
define an inner product. Then we’ll know which points we have to take some care on when proving
that the first example does give an inner product.
(i) This is not an inner product because it fails positivity; if f (x) is large for negative values of
x and small for positive values of x, then hf, f i will be negative. For an explicit example, set
f (x) = xR for x 0 and f (x) = 0 for x 0. This is a continuous function, so it is in C[ 1, 1], yet
0
hf, f i = 1 x3 dx = 1/4.
(ii) This is not an inner product because hf, f i takes the value 0 on some continuous functions that
are not identically zero. The function f constructed in the answer to (i) is an explicit example.
(iii) This will also fail positivity, but it is even easier to see that it is not symmetric. The expression
hf, gi hg, f i = f (0)g(1) g(0)f (1) is typically non-zero (e.g., take f (x) = 1, g(x) = x) and this
means that we can have hf, gi = 6 hg, f i.
To show that the first function is an inner product, we note that hf, gi is certainly linear in f for
each fixed g: for f1 , f2 , g 2 C[ 1, 1] and real numbers ↵ and , we have
Z 1
h↵f1 + f2 , gi = x2 (↵f1 (x) + f2 (x))g(x) dx + (↵f1 (0) + f2 (0))g(0)
1
✓Z 1 ◆ ✓Z 1 ◆
2 2
= ↵ x f1 (x)g(x) dx + f1 (0)g(0) + x f2 (x)g(x) dx + f2 (0)g(0)
1 1
which is always non-negative. This can only be zero if xf (x) = 0 for all x and f (0) = 0, which in
turn is only possible if f = 0. And, conversely, if f = 0, this just gives us zero.
7. This is rather more straightforward than the various parts of the next question, and we omit
the working.
One answer is: v1 = 13 (2, 1, 2)t and v2 = 13 (2, 2, 1)t .
0 1
1 0 1/2 1/2
8. @
We start with the kernel. The given matrix row-reduces to 0 1 1/2 3/2 A.
0 0 0 0
We can now read o↵ that the two vectors (1, 1, 2, 0)t and ( 1, 3, 0, 2)t span the nullspace of the
matrix, which is the same thing as the kernel of the linear transformation. ( This is another time
when it’s convenient to have integer vectors wherever possible. ) Take u1 = (1, 1, 2, 0)t (because it
has the simpler norm) and let v1 = p16 (1, 1, 2, 0)t . ( And here it’s best to keep it in this form; don’t
try to do any “cancellation”. )
Now we set
✓ ◆
t 1 1
w2 = ( 1, 3, 0, 2) p h( 1, 3, 0, 2)t , (1, 1, 2, 0)t i p (1, 1, 2, 0)t
6 6
1 1
= ( 1, 3, 0, 2)t (1, 1, 2, 0)t = ( 4, 8, 2, 6)t .
3 3
Finally, we normalise this to get v2 = p1 ( 2, 4, 1, 3)t .
30
So v1 and v2 form an orthonormal basis of the kernel. It’s easy to check that the two vectors are
orthogonal, so I expect you to do so.
To extend the set {v1 , v2 } to an orthonormal basis of R4 , it helps to observe that the rows of the
original matrix, or indeed the rows of the reduced matrix, are orthogonal to the kernel. So what we
need to find is an orthonormal basis {v3 , v4 } of the space spanned by either of those two pairs of
rows.
For instance, let u3 = (2, 0, 1, 1)t and u4 = (0, 2, 1, 3)t . For v3 , we take p1 (2, 0, 1, 1)t . Now take
6
1 1
w4 = (0, 2, 1, 3)t h(0, 2, 1, 3)t , (2, 0, 1, 1)t i(2, 0, 1, 1)t = (2, 6, 4, 8)t .
6 3
Finally, normalise this to get v4 = p1 (1, 3, 2, 4)t .
30
The vectors v1 , v2 , v3 and v4 form an orthonormal basis of R4 such that v1 and v2 form a basis of
the kernel.
Now we turn to the image. We know that the dimension of the image is 2, so for instance the first
two columns of the matrix span the image.
Set v1 = p1 (1, 2, 0)t and let
5
1 1
w1 = (1, 0, 2)t h(1, 2, 0)t , (1, 0, 2)t i(1, 2, 0)t = (4, 2, 10).
5 5
Now we normalise this to get v2 = p1 (2, 1, 5)t .
30
So v1 and v2 form an orthonormal basis of the image. It’s easy to check that the two vectors are
orthogonal, so I expect you to do so.
Perhaps the quickest way to find v3 is to look back to see why we got a zero row in the row-reduction.
This was because there was a non-trivial linear dependence among the rows: 2R1 R2 R3 = 0.
This means that w3 = (2, 1, 1)t is orthogonal to all the columns of the matrix. Therefore the
normalised multiple p16 (2, 1, 1)t will do for v3 .
Another way is just to write down what w3 has to be: any vector of the form (2, 1, x)t will be
orthogonal to v1 , and for such a vector to be orthogonal to v2 we require 5 + 5x = 0, so x = 1. (
This is more-or-less equivalent to choosing u3 = (0, 0, 1)t , which we can see is already orthogonal to
v1 , and adding an appropriate multiple of v2 , or of (2, 1, 5)t , to u3 . )
Either way, the vectors v1 , v2 and v3 form an orthonormal basis of R3 such that v1 and v2 form a
basis of the image.
9. If f is odd and g is even, then the product f g is odd: f ( x)g( x) = f (x)g(x). Now,
substituting y = x, we have
Z 0 Z 1 Z 1
f (x)g(x) dx = f ( y)g( y) dy = f (y)g(y) dy,
1 0 0
and therefore Z Z Z
1 0 1
f (x)g(x) dx = f (x)g(x) dx + f (x)g(x) dx = 0.
1 1 0
R1
10. Set g0 (x) = 1, g1 (x) = x, and g2 (x) = x2 . We have hg0 , g0 i = p1 g0
1 1 dx = 2, so f0 = 2
—
p
the constant function 1/ 2 — has norm 1.
We have hg0 , g1 i = 0, by the result in the previous question, as g0 is even and g1 is odd.
q q
Normalising g1 gives us the unit vector f1 = 32 g1 , i.e. f1 (x) = 32 x.
The final p function g2 is again even, so orthogonal to the odd function g1 , and hence to f1 , but
hg2 , f0 i = 2/3, so we put p
2 1
h2 (x) = g2 (x) f0 (x) = x2 ,
3 3
p
3p 5 2 1
and normalise to get f2 (x) = 2 2
(x 3 ).
11. (a) Suppose x t Ay = 0 for all x and y . Consider any entry aij of A; let x = ei and y = ej ,
the standard basis vectors. Then 0 = x t Ay = aij . This holds for every entry of A, so A is the zero
matrix.
(b) If A and B have the stated property, then x t (A B)y = 0 for all x and y . By the previous part,
we have that A B = 0, or A = B.
(c) One example is ✓ ◆
0 1
A= ,
1 0
(in fact, all examples are multiples of this). It is easy to check that x t Ax = x1 x2 x2 x1 = 0, for all
vectors x .
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Solutions to Exercises 13: Similar matrices and real inner products
This document contains answers to the exercises. There may well be some errors; please do let me
know if you spot any.
1. Suppose that A and B are similar, and that A = SBS 1 for some invertible matrix S. Then
we have A2 = (SBS 1 )(SBS 1 ) = SB(S 1 S)BS 1 = SB 2 S 1, so A2 and B 2 are similar.
Suppose now that A is invertible. We can write B = S 1 AS, and note that (S 1 A 1 S)B =
S 1 A 1 SS 1 AS = I, so B is invertible, with inverse B 1 = S 1 A 1 S, which is similar to A 1 .
2. (a) Suppose that A has an inverse A 1. We can write AB = AB(AA 1) = A(BA)A 1, showing
that AB and BA are similar.
(b) Suppose S and T are similar, with S = P 1T P . Then we can set A = P and B = P 1T , so that
A is invertible, AB = T and BA = S.
In other words, two n ⇥ n matrices are similar if and only if they can be written as AB and BA for
some pair of n ⇥ n matrices A and B, with A invertible.
(c) Here AB is the zero matrix, while BA isn’t. Any matrix S(AB)S 1 similar to AB = 0 will also
be the zero matrix, so in particular AB and BA are not similar.
4. (a) Take any vector y in the range of AB; so y = ABx for some x 2 Rk . Now y = Az where
z = Bx 2 Rm , so y 2 R(A). This proves that R(AB) ✓ R(A).
For any vector u 2 N (B), we have Bu = 0, and so ABu = 0, so u 2 N (AB). This shows that
N (B) ✓ N (AB).
(b) Now suppose that R(B) = Rm : we need to show that the range of A is contained in the range of
AB. So, take a vector z in the range of A: that means that Ax = z for some x 2 Rm . As the range
of B is Rm , there is some y 2 Rk with By = x . Putting this together, we have ABy = Ax = z , so
z is in the range of AB, which is what we needed.
(c) Finally, we note that, if A is an n ⇥ m matrix, then the rank ⇢(AB) of AB, which is the dimension
of R(AB), is at most the dimension ⇢(A) of R(A), and this is at most m, since R(A) is spanned by
the columns of A. If AB is an n ⇥ n matrix and m < n, this implies that AB is singular.
t
5. ✓ is linear◆ in the first variables, since it can be written in the form hx , y i = x Ay ,
The function
1 1
where here A = .
1 3
The function can be seen to be symmetric, i.e, hx , y i = hy , x i, directly from its definition; you’ll see
that this is e↵ectively the same thing as checking that the matrix A above is symmetric.
You do need to mention these points in your answer, but the main thing is to check positivity. For
that, it is not enough to choose a particular x and check that hx , x i 0: you need to demonstrate
this for all vectors x .
We have h(x1 , x2 )t , (x1 , x2 )t i = x21 2x1 x2 + 3x22 and as this is just (x1 x2 )2 + 2x22 it is always
non-negative. Also, if this is zero, we must have x1 = x2 and x2 = 0, which gives us x1 = x2 = 0.
And, conversely, if x1 = x2 = 0, this just gives us zero.
So the first function we are given does indeed give us an inner product in R2 .
For the second function we are given, we don’t have an inner product in R2 as positivity fails. That is,
even though we do always have h(x1 , x2 )t , (x1 , x2 )t i = (x1 x2 )2 0, the point is that this expression
can be equal to zero for non-zero x : for instance, h(1, 1)t , (1, 1)t i = 0.
Note: giving an example, like the vector (1, 1)t above, is the most convincing way to answer the
question.
6. It’s perhaps illuminating to do the second part first: explaining why each of (i)-(iii) fails to
define an inner product. Then we’ll know which points we have to take some care on when proving
that the first example does give an inner product.
(i) This is not an inner product because it fails positivity; if f (x) is large for negative values of
x and small for positive values of x, then hf, f i will be negative. For an explicit example, set
f (x) = xR for x 0 and f (x) = 0 for x 0. This is a continuous function, so it is in C[ 1, 1], yet
0
hf, f i = 1 x3 dx = 1/4.
(ii) This is not an inner product because hf, f i takes the value 0 on some continuous functions that
are not identically zero. The function f constructed in the answer to (i) is an explicit example.
(iii) This will also fail positivity, but it is even easier to see that it is not symmetric. The expression
hf, gi hg, f i = f (0)g(1) g(0)f (1) is typically non-zero (e.g., take f (x) = 1, g(x) = x) and this
means that we can have hf, gi = 6 hg, f i.
To show that the first function is an inner product, we note that hf, gi is certainly linear in f for
each fixed g: for f1 , f2 , g 2 C[ 1, 1] and real numbers ↵ and , we have
Z 1
h↵f1 + f2 , gi = x2 (↵f1 (x) + f2 (x))g(x) dx + (↵f1 (0) + f2 (0))g(0)
1
✓Z 1 ◆ ✓Z 1 ◆
2 2
= ↵ x f1 (x)g(x) dx + f1 (0)g(0) + x f2 (x)g(x) dx + f2 (0)g(0)
1 1
which is always non-negative. This can only be zero if xf (x) = 0 for all x and f (0) = 0, which in
turn is only possible if f = 0. And, conversely, if f = 0, this just gives us zero.
7. This is rather more straightforward than the various parts of the next question, and we omit
the working.
One answer is: v1 = 13 (2, 1, 2)t and v2 = 13 (2, 2, 1)t .
0 1
1 0 1/2 1/2
8. @
We start with the kernel. The given matrix row-reduces to 0 1 1/2 3/2 A.
0 0 0 0
We can now read o↵ that the two vectors (1, 1, 2, 0)t and ( 1, 3, 0, 2)t span the nullspace of the
matrix, which is the same thing as the kernel of the linear transformation. ( This is another time
when it’s convenient to have integer vectors wherever possible. ) Take u1 = (1, 1, 2, 0)t (because it
has the simpler norm) and let v1 = p16 (1, 1, 2, 0)t . ( And here it’s best to keep it in this form; don’t
try to do any “cancellation”. )
Now we set
✓ ◆
t 1 1
w2 = ( 1, 3, 0, 2) p h( 1, 3, 0, 2)t , (1, 1, 2, 0)t i p (1, 1, 2, 0)t
6 6
1 1
= ( 1, 3, 0, 2)t (1, 1, 2, 0)t = ( 4, 8, 2, 6)t .
3 3
Finally, we normalise this to get v2 = p1 ( 2, 4, 1, 3)t .
30
So v1 and v2 form an orthonormal basis of the kernel. It’s easy to check that the two vectors are
orthogonal, so I expect you to do so.
To extend the set {v1 , v2 } to an orthonormal basis of R4 , it helps to observe that the rows of the
original matrix, or indeed the rows of the reduced matrix, are orthogonal to the kernel. So what we
need to find is an orthonormal basis {v3 , v4 } of the space spanned by either of those two pairs of
rows.
For instance, let u3 = (2, 0, 1, 1)t and u4 = (0, 2, 1, 3)t . For v3 , we take p1 (2, 0, 1, 1)t . Now take
6
1 1
w4 = (0, 2, 1, 3)t h(0, 2, 1, 3)t , (2, 0, 1, 1)t i(2, 0, 1, 1)t = (2, 6, 4, 8)t .
6 3
Finally, normalise this to get v4 = p1 (1, 3, 2, 4)t .
30
The vectors v1 , v2 , v3 and v4 form an orthonormal basis of R4 such that v1 and v2 form a basis of
the kernel.
Now we turn to the image. We know that the dimension of the image is 2, so for instance the first
two columns of the matrix span the image.
Set v1 = p1 (1, 2, 0)t and let
5
1 1
w1 = (1, 0, 2)t h(1, 2, 0)t , (1, 0, 2)t i(1, 2, 0)t = (4, 2, 10).
5 5
Now we normalise this to get v2 = p1 (2, 1, 5)t .
30
So v1 and v2 form an orthonormal basis of the image. It’s easy to check that the two vectors are
orthogonal, so I expect you to do so.
Perhaps the quickest way to find v3 is to look back to see why we got a zero row in the row-reduction.
This was because there was a non-trivial linear dependence among the rows: 2R1 R2 R3 = 0.
This means that w3 = (2, 1, 1)t is orthogonal to all the columns of the matrix. Therefore the
normalised multiple p16 (2, 1, 1)t will do for v3 .
Another way is just to write down what w3 has to be: any vector of the form (2, 1, x)t will be
orthogonal to v1 , and for such a vector to be orthogonal to v2 we require 5 + 5x = 0, so x = 1. (
This is more-or-less equivalent to choosing u3 = (0, 0, 1)t , which we can see is already orthogonal to
v1 , and adding an appropriate multiple of v2 , or of (2, 1, 5)t , to u3 . )
Either way, the vectors v1 , v2 and v3 form an orthonormal basis of R3 such that v1 and v2 form a
basis of the image.
9. If f is odd and g is even, then the product f g is odd: f ( x)g( x) = f (x)g(x). Now,
substituting y = x, we have
Z 0 Z 1 Z 1
f (x)g(x) dx = f ( y)g( y) dy = f (y)g(y) dy,
1 0 0
and therefore Z Z Z
1 0 1
f (x)g(x) dx = f (x)g(x) dx + f (x)g(x) dx = 0.
1 1 0
R1
10. Set g0 (x) = 1, g1 (x) = x, and g2 (x) = x2 . We have hg0 , g0 i = p1 g0
1 1 dx = 2, so f0 = 2
—
p
the constant function 1/ 2 — has norm 1.
We have hg0 , g1 i = 0, by the result in the previous question, as g0 is even and g1 is odd.
q q
Normalising g1 gives us the unit vector f1 = 32 g1 , i.e. f1 (x) = 32 x.
The final p function g2 is again even, so orthogonal to the odd function g1 , and hence to f1 , but
hg2 , f0 i = 2/3, so we put p
2 1
h2 (x) = g2 (x) f0 (x) = x2 ,
3 3
p
3p 5 2 1
and normalise to get f2 (x) = 2 2
(x 3 ).
11. (a) Suppose x t Ay = 0 for all x and y . Consider any entry aij of A; let x = ei and y = ej ,
the standard basis vectors. Then 0 = x t Ay = aij . This holds for every entry of A, so A is the zero
matrix.
(b) If A and B have the stated property, then x t (A B)y = 0 for all x and y . By the previous part,
we have that A B = 0, or A = B.
(c) One example is ✓ ◆
0 1
A= ,
1 0
(in fact, all examples are multiples of this). It is easy to check that x t Ax = x1 x2 x2 x1 = 0, for all
vectors x .
MA212 Further Mathematical Methods
Lecture 7: Orthogonal matrices continued
Dr James Ward
basis MB -
-
I UI UI II =
-
I →
swap the order to have left handed ordered
-
basis MB -
-
C UI UI I
.
,
)
AF
''
( since
Sino
'
)
= so
with if orientation preserving and , if orientation reversing
-
t '
- -
ago
o
It
'
Ay -
-
Mps AFB mis
is
type l rotation
:
g
IAI = + I AE =
y A In
-
(! )
IA I I AE I At In \
{ !})
= -
- -
K ,
reversing
*
angle
preserving
{
: Tr CA) -
-
dos Ott
reversing
:
Trial -
zaeso -
I
* direction
I § ayy if I { Yg
find any ate .
=p
anticlockwise
clockwise
a)
type 2 :
reflection
③
types both :
rotation and reflection
Orthogonal matrices in R3
Lecture 7, page 3
Case 1: A is a rotation by an angle ✓ about an axis
Av = v .
Thus, v 6= 0 is an eigenvector of
A with eigenvalue 1.
Lecture 7, page 4
In particular, this means that A is similar to the matrix
0 1
cos ✓ sin ✓ 0
A-F
' B B C
=
B = @ sin ✓ cos ✓ 0A .
0 0 1
Observe that |B| = 1 and so |A| = 1 too. That is, this linear
transformation is orientation preserving.
Also, if (u1 , u2 , v̂ ) is an ordered orthonormal basis of R3 where u1
and u2 are vectors in the plane of rotation, A and B are related by
0 1
| | |
B C
B = M t AM with M = @u1 u2 v̂ A .
| | |
Lecture 7, page 5
Case 2: A is a reflection about a plane
Av = v.
Thus, v 6= 0 is an eigenvector of
A with eigenvalue 1.
If u is in the plane of reflection, i.e. u ? v , then Au is just u as it
is not reflected, i.e.
Au = u.
Lecture 7, page 6
In particular, this means that A is similar to the matrix
0 1
1 0 0
B B C
A-F B = @0 1 0 A .
'
0 0 1
Observe that |B| = 1 and so |A| = 1 too. That is, this linear
transformation is orientation reversing.
Also, if (u1 , u2 , v̂ ) is an ordered orthonormal basis of R3 where u1
and u2 are vectors in the plane of reflection, A and B are related by
0 1
| | |
B C
B = M t AM with M = @u1 u2 v̂ A .
| | |
Lecture 7, page 7
Case 3: A is a combination of cases 1 and 2.
Av = v.
Thus, v 6= 0 is an eigenvector of
A with eigenvalue 1.
@ sin ✓ cos ✓ 0 A.
0 0 1
Observe that |B| = 1 and so |A| = 1 too. That is, this linear
transformation is orientation reversing.
Also, if (u1 , u2 , v̂ ) is an ordered orthonormal basis of R3 where u1
and u2 are vectors in the common plane of rotation and reflection,
A and B are related by
0 1
| | |
B C
B = M t AM with M = @u1 u2 v̂ A .
| | |
Lecture 7, page 9
Recap!
Lecture 7, page 10
What about anticlockwise and clockwise?
e1 ⇥ e2 = e3
Lecture 7, page 11
Then, with this in place, taking e3 to be the direction of the axis
of rotation and Lin{e1 , e2 } to be the plane of rotation, an
anticlockwise rotation by ✓ is then defined in the R2 sense, i.e.
Lecture 7, page 12
In particular, with the cases above in mind, we want our ordered
orthonormal basis (u1 , u2 , v̂ ) to map
e1 ! u 1 , e2 ! u 2 and e3 ! v̂ ,
| | |
u1 u2 v̂ = 1 so that orientations are preserved.
| | |
And, if we do this, any rotation discussed in the cases from earlier
will be anticlockwise.
Of course, if our determinant is 1, we can make it 1 by
interchanging the vectors u1 and u2 !
And, if we want a clockwise rotation, we can use the same ordered
basis but just use a negative angle instead!
Lecture 7, page 13
What we need to do now...
Lecture 7, page 14
Finding an orthogonal matrix for a given transformation
we then have a, =
(§ ) and uh
Eff )
-
-
iii. is is
/
(
f ! -49g I
,
left handed
-
⇐ lui uh O l
, .
-
- = ICI ta ) -
- I
raaiiiiiiasiiiuu
! ) fi! i! ! )
#"
:9
f :{ s.in !
!
-
"
transformation a
-
i.
, our is
-
. .
.
" "
:
to check : AT is orthogonal
Lecture 7, page 16
Finding the transformation from a given orthogonal matrix
Lecture 7, page 18
Example
/ I
3 8 -
I 4
I Al =
(f) , g *
= . .
.
= I orientation preserving c pure rotation )
4 4 -7
Il:÷¥H ! !: Hi ÷÷l
' "ene .
"" '
Lecture 7, page 19
( Yo ) f-
( ! Is Ig ) ( ! ) f- ( Ig )
utu and find Ae
pick say E- - -
-
,
also o -
f- to = I .
I .
cos o also = -
I
therefore .
'
0=65 ( Eg ) a 15273
°
Check -
or anti clock wise ?
-
Lecture 7, page 20
Is this transformation anticlockwise or clockwise?
û ŵ v̂ = 1.
Lecture 7, page 21
That is, what do we have?
Lecture 7, page 22
û Aû v̂ = û ↵û + ŵ v̂
= û ŵ v̂
g) Cz -2oz -
ac,
= û ŵ v̂
=
Of course, using non-unit vectors u, Au and v (in that order) will
change the magnitude of but not its sign giving us the following
simple test.
Lecture 7, page 23
An anticlockwise or clockwise test
u Au v .
If it is
Lecture 7, page 24
Example continued
where v is the direction of the axis and the angle is cos 1( 8/9).
Is this rotation anticlockwise or clockwise?
| ! Ht Y / ( f
"
-119
of
0 4
171g
= =
sa so
-
- .
.
=
419 I
( §)
anticlockwise cost c- 819 )
so A represents rotation by about the axis
Lecture 7, page 25
An alternative method for finding angles of rotation
⌅ 1 if A is orientation-reversing. Tr CA ) =
Tr CAT )
Lecture 7, page 26
Example continued
g-
= -
-
-
Lecture 7, page 27
MA212 Further Mathematical Methods
Lecture 8: Complex vector spaces
Dr James Ward
⌅ Complex numbers
⌅ Complex vector spaces
⌅ Complex inner product spaces
Lecture 8, page 1
Complex numbers
C = {a + ib | a, b 2 R}
p
where i = 1 so that i 2 = 1.
Complex numbers can be added and multiplied in the usual way,
i.e. for a, b, c, d 2 R we have
and
Lecture 8, page 3
Complex conjugate and modulus
Lecture 8, page 4
Properties of complex conjugates and moduli
w |w |
|wz| = |w | |z| and = if z 6= 0.
z |z|
This should all be familiar to you from MA100 and so we will not
dwell on it here.
Lecture 8, page 5
Complex vector spaces
Lecture 8, page 6
What’s the same in complex vector spaces?
The only change is that our scalars (i.e. what we can multiply by
and what we can have as the entries of a vector or matrix) are now
complex numbers.
Lecture 8, page 7
Example
0 10 1 0 1
1 0 i z1 2+i
B CB C B C
Solve the equation @ i 1 1 + i A @z2 A = @3 2i A.
0 i 1 z3 2i
REF v
augmented matrix
RREF v
f: :÷il¥n*t: ! I :÷l"n& : :L : )
"
number
complex
+
SO Zi ti Z 3=2 Ii let 23=-1 Cl
-
E , so that 2-1=2-1 i -
it
Zz ti Zg - 2 Zz =
2- if
¥z;) (Ii ) tf )
i . the solution is
=
t .
tee
Lecture 8, page 8
Complex conjugates of complex vectors and matrices
Lecture 8, page 9
Complex conjugates of complex determinants
a11 a12
|A| = = a11 a22 a12 a21
a21 a22
a11 a12
= a11 a22 a12 a21 = = |A|.
a21 a22
Lecture 8, page 10
What’s not the same in complex vector spaces?
Lecture 8, page 11
Complex inner product spaces
Lecture 8, page 12
Complex conjugate linearity in the second argument
Lecture 8, page 14
Example
is an inner product on C2 .
Bri ) wit Haz ie uz ) WI
Ci )
WT U , it , t Va WI )
.
, = d l U, t Uz WJ ) t fC
= a
twirlers fait
di )
.f =
GUTAI =
That is Uz = u ,
it , t Usui =
( Y;) .
Citi)
(Yz ) -
= u, a + uzi .
=
Iu, 14 lurk so and Cut!) .
-
-
o ⑦ lull 't last ?o
-
⇐ Ui O U2 0
-
- =
,
a real vector
⇐
fun;) -
- O
Lecture 8, page 15
Two di↵erences between real and complex dot products
2 1 i
Is hu, v i = u t v an inner product on C2 ?
1+i 2
'
Ci ) ft I ht
. , E Cl and a, B. red coat BI ,
we > = the a. he > + focus Ws
at
( I Ii ) ( Ii Ii )
' '
note -
-
'
a
=
= -
-
✓
(I ti ) ( Eu)
< u.us -
-
ah U2 )
Lecture 8, page 17
U U> for u I C- ¢2
IV. US = < .
,
t
B_ze= ai "
( E)
I
)
note
ri z
=
( Ii ¥ ) =
=
Bz
Reo L: 4=0 we Use O LU, U > so
[ to R : cu U >
. =
O Illit 42124 I UH iU2P=O uit U2 o
-
-
and Ultimo Ui - -
U2
(Un!) -
-
Lecture 7, page 20
What do we keep from real inner product spaces?
Lecture 8, page 18
Example
q)
'
" Ali . Ui -
it -
-
=
It =
± , =
,
.
Nhl f- to be =
.
=
E t t"" t I =
It Its = ,
it ik
. =
) Yg )
.
=
to tip .
to =
f Ho- -
-
o
Lecture 8, page 19
Gram-Schmidt orthogonalisation still works...
Step . . . And so on, until we have used all of the original vectors .
Lecture 8, page 20
...BUT we must take care with the inner products!
hw2 , ŵ1 i = hu2 hu2 , ŵ1 i ŵ1 , ŵ1 i = hu2 , ŵ1 i hu2 , ŵ1 i hŵ1 , ŵ1 i = 0.
w 2 = u2 hŵ1 , u2 i ŵ1 ,
hw2 , ŵ1 i = hu2 hŵ1 , u2 i ŵ1 , ŵ1 i = hu2 , ŵ1 i hŵ1 , u2 i hŵ1 , ŵ1 i =
6 0
Lecture 8, page 21
Example continued
4=1 ! ) (t ) -
-
O
Gram -
Schmidt :
pick any us # Linge .gg say
, , us ,
fogy
(§ )
I"
Us
I ¥
.
① Ws
.
=
= 43 -
( Us A ) Ui
. -
cuz -
vz ) uz
② E- Yuta
-
-
(9) -
o -
÷ #
.
il
.
'
HH t.es .
all 02 Us >
, , is on
=
=
C-DCI )
I -
ti
i 2-14--6
Ci ) -1215 )
basis of E3
Lecture 8, page 22
extra example session cw4)
(y )
O -
I 0
what does the orthogonal matrix A- ?
i
represent
g g
.
/ I I -6 ? I
O '
' °
IAI D= orientation
' = c i ) It I
reversing
-
- - =
-1 O O
O O l
U to sit AI I LATIN d
-
-
. = -
µ ) (t)
l -
I O
ATI - Ci ta to ↳ =D U -
-
,
,
O O 2
find ate ie e-
(§) Aa
⇐ Io ! ) ( ! ) ( ! )
-
-
:
-
. . -
O E Osa
-
Ot Otl -
Lt All
.
=
HUH HA UH Cosa =
Cosa cos 0=1 0=0
c o clockwise
if 0=0 . Ay
-
-
a 14 I I 1=0 neither clockwise nor anticlockwise
a 11-161=0 nor
find the matrix that represents reflection in the plane through the origin with normal
(! ) and
2 .
a a
E- El :o) .
take E- (9) and e -
-
Eff )
/ / =/ ! For %E/=
" "
, + + i
-
Ati
'
fusion sago ! )
'
cui.ui.is mist
-
relative to .
-
- A- MBA
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Exercises 14: Orthogonal matrices
For these and all other exercises on this course, you must show all your working.
•
1. Let u = (a, b, c)t be a vector in R3 with a, c 6= 0. Show that v = (b, a, 0)t is orthogonal to u,
and that there is a value of x such that w = (a, b, x)t is orthogonal to both u and v . Find a formula
for x. can be a
useful conclusion TO
help find orthonormal vectors
2.
✓ (a) Find the orthogonal matrix A representing a rotation through an angle ⇡/3 anticlockwise
about an axis in the direction ( 1, 1, 1)t .
(b) Find the orthogonal matrix C representing a reflection in the plane through the origin with
normal ( 1, 1, 1)t .
:
3. Consider the orthogonal matrices
0 1 0 1
2/3 2/3 1/3 3/5 4/5 0
A = @ 1/3 2/3 2/3 A and B=@ 0 0 1A .
2/3 1/3 2/3 4/5 3/5 0
(ii) Why are the angles of rotation of AB and BA the same? (An earlier exercise should help.)
(iii) Why would it be incorrect to conclude that the axes of rotation of AB and BA are the same?
hx , y i = 14 kx + y k2 1
4 kx y k2 .
(You should not assume that we are dealing with any specific inner product!)
i .
given :
a-
( Fg ) ,
U
=/ ! ) ,
he -
-
< U. U> c
(§ ) Goa ) ab ba to orthogonal
=
, > = - = o . so u is to a .
therefore
(§ ) ( ! ) and
( § ) (g)
ab ab -10=0
'
or t b txc O < o
-
c o s -
-
>
= -
, , ,
,
,
-
alba
since Cfo ,
X=
c- can make us to be
orthogonal to a and I
orthonormal vectors :
a .
ca) a unit vector in the direction of the axis is it #
(! ) ✓② and it at t ut
(y) =/ ! )
'
then two orthonormal vectors are at and UI
x
14,43 1%32,1
lui uh it '
E) its t 's ) cui.ua I )
-
since Ko I left-handed
-
=
* is
-
a
-
- -
, . -
. .
, g orthogonal basis of Rs
µ.sn?j-csing?goy)=(Io-Ig ! )
"
relative to cut iv. is
.
,
our transformation is ¥ -
-
B Bpg)
combining the orthogonal vectors . we have an orthogonal matrix MB
' ' '
.
one. .
adit .
moai 'm.
"
soar
÷ : ÷:
-
"
.
(a. b C)
X
.
↳
2 (a) I = C -
l l l It
( Yf )
43 43
from Guestion of of 2/3
-
l C
u, = Cb a
Ay
,
,
-
,
= l , I , =
Uz -
- la b . .
-92¥ ) -
- c -
hi ,
-2 )
1,3 42,33
Fifi )
,
Ef :o)
,
)
'
so it -
-
E ,
-
-
,
uh -
and uh -
-
ri
,
AYE
for reflection
( to ! ! ) j
do , a .
moai 'm.'
"
t : ! ! .tl?i ! !.tf÷÷:
men
-
e- '
.
. "
. . . .
/ I, / / /
I
/ Eg GI f
3. Ca) det IAI -
5- t t t
-
=
-5×-5 t 's 's
x t -5×-5=1
so A is orientation -
preserving ✓
137g g¥ I
det 1131 I
reversing
= - =L l l -
x I - -
I . so B is orientation -
Cb)
from Ca ) , we know that transformation A only includes rotation
let it be the axis of rotation , given by AI =
I
ft
2 2 I
al f: : :D ÷ !)
"
-
'
a- '' '
z -
Z
-
I 2
(!)
Since at at 03=0 , I =
choose a
(÷ ) such that at u then Aa I
? ! ) ( Io ) ( Io )
- =
=
.
.
taking
/ ! ! I / =/ If I ! /
A = -
i
/ to :/ = -
I so ,
so A is a clockwise rotation
Since I Al -
I , Tr CA) =
zoos Otl .
i.e .
It 25 t 25 =
2 SO -11 . So cos 0=-12 and D= F-V
-25
f!)
Overall transformation A clockwise rotation of about
,
represents a axis
cc ) from Cal , we know that transformation includes both rotation and reflection
t
tf ! I +4.7:p fo 's E)
* 'e
-
-
-
-4 3 5
I ✓
Since Cz -
Cztzc , so , he
=
(1) µo¥¥q ! ) ÷)
choose e- such that at ht .
then Ba
-
-
=
hence
fi if II ,
ai
,
-
-
¥ tent! ! first ! If
,
= -
Eta .
-
Eco
Tr CB) =
zoos o -
I ,
ie . -35--2Cosa -
I , so Cosa -
- ¥ and O -
-
arc cos ¥
Overall ,
transformation
reflection about
B
represents
a
plane
a
whose
clockwise rotation
normal is ⇐
of
.
-
I
arc
,
costs
15h
about axis
) and a
I I 5-
÷÷÷÷÷)
(d) Ci ) AB -
-
and B"
¥,
"
orthogonal
product of
"
srmifianceopsd.net#ABl--detlBAl= I so
they are orientation
reversing ✓ orthogonal ?
- -
AB
.
. .
BA , no reason for I =
BE
*
4 .
¥ 11 X type -
I 11 x y 112
-
=
¥ I
llxty 112 -
Hx -9114
=
¥ ccxty xty > .
-
<x y
-
,
x y
-
s )
by property Ci) =
Iq ( ex .
xty > + ay xtys ,
-
ex , x -
y ) t ay ,
x -
y> I
=
¥ ( s Xty ,
x > t city y , > - ex -
y
. x > t ex
-
y y>)
,
=
&C c
-
X. x > t <
y . x St Lx ,
ys t
-
ay y , >
-
-
ax . x > + ay , x) tax , y> -
-
ay y > )
,
=
Th ( 4 ex y s ) ,
L X
y>
=
,
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Solutions to Exercises 14: Orthogonal matrices
This document contains answers to the exercises. There may well be some errors; please do let me
know if you spot any.
which is correct (if it hadn’t been, we should swap v1 and v2 ). Thus, we take
0 1 1 1
1 0 1 1
1
p
2
p
6
p
3
p
2
p
2
0
B p1 C B p2 C
MB = @ p12 p1
6 3A
so that MB 1 = MBt = @ p16 p1
6 6A
.
0 p2 p1 p1 p1 p1
6 3 3 3 3
since the matrix representing rotation through ⇡/3 in the anticlockwise (positive) direction in R2 is
✓ ◆
cos(⇡/3) sin(⇡/3)
.
sin(⇡/3) cos(⇡/3)
0 1
2/3 2/3 1/3
We can now see that A = AT = MB AB,B
T M t = @ 1/3
B 2/3 2/3A.
2/3 1/3 2/3
This is a product of orthogonal matrices, and you can see that it is indeed orthogonal. You can also
see that it maps ( 1, 1, 1)t to itself, as it should.
For the second matrix C, representing the reflection S in the plane normal to ( 1, 1, 1)t , the easiest
approach is to note that 0 1
1 0 0
AB,B
S = @0 1 0 A
0 0 1
as the vector v3 is mapped to v3 whereas v1 and v2 , as they lie in the plane of reflection, are fixed.
Thus 0 1
1/3 2/3 2/3
C = AS = MB AB,B
S MB =
t @2/3 1/3 2/3A .
2/3 2/3 1/3
Note: do you have to make the axis the third vector in the ordered basis? No, but the transformation
matrix AB,B
S depends on which of the three vectors in the basis B is the axis. If you are dealing with
a rotation, and you want to represent the rotation with respect to a basis using the matrix
0 1
cos ✓ sin ✓ 0
@ sin ✓ cos ✓ 0A
0 0 1
then the axis vector has to be the third vector of the basis. For this question as a whole, there are
some clear advantages to using the same change-of-basis matrix (i.e. represent the transformation
with respect to the same basis) for both parts.
3. (a) The determinant of the orthogonal matrix A is +1, so this matrix is orientation-preserving.
The determinant of B is 1, so it is orientation-reversing.
(b) To find the axis of rotation of A, we look for the eigenvector corresponding to eigenvalue 1. We
have 0 1
1/3 2/3 1/3
A I = @ 1/3 1/3 2/3 A ,
2/3 1/3 1/3
and it’s perfectly legitimate to say that x = (1, 1, 1)t is in the null space “by inspection”: perhaps
you spotted that adding the three columns together gives the zero vector.
One way to find the angle of rotation is to find a vector y perpendicular to the axis x , and find the
angle between y and Ay . For instance, the vector y = (1, 0, 1)t is perpendicular to (1, 1, 1), and
Ay = (1, 1, 0)t . If ✓ denotes the angle of rotation, then we have
y · (Ay ) 1 1
cos ✓ = =p p = .
ky k kAy k 2· 2 2
( By the way, we have kAy k = ky k as A is orthogonal. ) Thus the angle of rotation is ⇡/3.
Another approach is to use the fact that the matrix representing a rotation by ✓ with respect to
standard co-ordinates is 0 1
cos ✓ sin ✓ 0
@ sin ✓ cos ✓ 0A ,
0 0 1
which has trace (sum of the diagonal entries) 1 + 2 cos ✓, and trace is preserved under similarity. So
the trace of A, which is equal to 2, is also equal to 1 + 2 cos ✓; this again gives ✓ = ⇡/3.
To determine whether the rotation is clockwise or anticlockwise, we look at the determinant
0 1 0 1
1 1 1
@ A
det y Ay x = det 0 @ 1 1A .
1 0 1
Then, as this gives us a determinant of 3, which is negative, we can see that we have a clockwise
rotation through an angle of ⇡/3 about the axis (1, 1, 1)t .
It is also correct to say that the matrix represents an anticlockwise rotation through an angle of ⇡/3
radians about the axis ( 1, 1, 1)t .
(c) To find the axis of rotation of B, we look for the eigenvector corresponding to eigenvalue 1. We
have 0 1
8/5 4/5 0
B+I =@ 0 1 1A ,
4/5 3/5 1
and it’s perfectly legitimate to say that x = (1, 2, 2)t is in the null space “by inspection”: perhaps
you spotted that c1 2c2 + 2c3 = 0.
One way to find the angle of rotation is to find a vector y perpendicular to the axis x , and find the
angle between y and By . For instance, the vector y = (0, 1, 1)t is perpendicular to (1, 2, 2), and
By = (4/5, 1, 3/5)t . If ✓ denotes the angle of rotation, then we have
y · (By ) 8/5 4
cos ✓ = =p p = .
ky k kBy k 2· 2 5
( By the way, we have kBy k = ky k as B is orthogonal. ) Thus the angle of rotation is cos 1 (4/5)
Then, as this gives us a determinant of 18/5, which is negative, we can see that we have a clockwise
rotation through an angle of cos ✓ = 4/5 about the axis (1, 2, 2)t combined with a reflection in the
plane through the origin with this vector as its normal.
It is also correct to say that the matrix represents an anticlockwise rotation through an angle of
cos ✓ = 4/5 radians about the axis ( 1, 2, 2)t combined with a reflection in the plane through the
origin with this vector as its normal.
(d) For (i), both AB and BA are products of orthogonal matrices, so orthogonal. Their determinants
are both equal to det(A) det(B) = 1, so they’re both orientation-reversing.
For (ii), the answer to Question 2 from Exercises 13 tells us that AB and BA are similar. This
means, in particular, that they have the same trace. We know that the trace of an orientation-
reversing matrix with an angle of rotation ✓ is 1 + 2 cos ✓, so AB and BA have the same angle of
rotation ✓.
For (iii), if the axis of rotation of AB is v 6= 0, we have
ABv = v.
B(ABv ) = Bv =) (BA)Bv = Bv ,
and so, as B is orthogonal, Bv 6= 0 is the axis for BA. There is, of course, no reason to suppose that
v will give us the same direction as Bv as B will rotate and/or reflect the vector v .
4. All we have to do is expand out the right-hand side to see that
kx + y k2 = hx + y , x + y i = kx k2 + 2hx , y i + ky k2 ,
and similarly
kx y k2 = hx y, x y i = kx k2 2hx , y i + ky k2 ,
so their di↵erence is 4hx , y i, which is what we wanted.
It is worth noting that this proof works for any inner product, not just the dot product.
MA212 Further Mathematical Methods
Lecture 9: Special complex matrices
-
orthogonal ATA In
⌅ Hermitian matrices
-
-
Lecture 9, page 1
The Hermitian transpose
A⇤ = At .
Lecture 9, page 3
Hermitian transposes when A is real
Note: All real matrices are complex matrices where the imaginary
part of each entry just happens to be zero. That is,
if A is real, then A = A.
Lecture 9, page 4
Some useful facts about Hermitian transposes
Lecture 9, page 5
A useful fact about complex dot products
u · v = v ⇤ u,
Lecture 9, page 6
A further fact about complex dot products
x · (A⇤ y ) = (Ax ) · y
for all x , y 2 Cn .
Lecture 9, page 7
Hermitian matrices
Lecture 9, page 8
Properties of Hermitian matrices
AI A1 AL M1
-
- -
-
.
① x. CAL) =
I. ( UI ) tick 9)
)→ ,
Chi d)
-
LI 17=0
-
cis take a -
- U and 1=1 CI -
d) I 1=0 LI a) 11×42=0
-
I -
-
a * is real
② take Atm, ki d) CX y ) =D Ch A) ex y) O x. y o
orthogonal toy
-
x is
- - - -
- - -
ti u as reef
- to
htt
eigenvalues Lecture 9, page 9
Unitary matrices
Lecture 9, page 11
Other definitions of unitary matrices
(3) AA⇤ = In .
(4) The columns of A form an orthonormal set of vectors.
Lecture 9, page 12
Proof of (1) () (2), i.e. A⇤ A = In () A⇤ = A 1
A⇤ A = In =) (A⇤ A)A 1
=A 1
=) A⇤ (AA 1
)=A 1
,
RTL: If A⇤ = A 1, then
A⇤ A = A 1
A = In ,
Lecture 9, page 13
Proof of (2) () (3) i.e. A⇤ = A 1
() AA⇤ = In
AA⇤ = AA 1
= In ,
as required.
RTL: If AA⇤ = In , then A is invertible as
Thus, we have
A 1
(AA⇤ ) = A 1
=) (A 1
A)A⇤ = A 1
,
Lecture 9, page 14
Proof of (1) () (4), i.e. A⇤ A = In () columns of A are ON
Lecture 9, page 15
Remarks about unitary matrices
Lecture 9, page 16
Example
!
1 1
Show that the matrix A = p1 is unitary.
2 i i
Verify that its determinant, |A|, has a modulus of one.
A is unitary
I I:/
'
# Kim tines E ai
' Ah
.
-
.
- -
. i
-
as I -
it 2=1 i) ( T)
- = C i -
) i= -
iz = I modulus of IAI is I
Lecture 9, page 17
A property of unitary matrices
i. e .
AI # I , XFO , AHA In =
* * * *
( AI) CAH
-
=
CA Ay =
I A AI =
x In A = x* x
I
*
→ *x -
lap x x
(A1) # D=
-
Cays and
.
=
mix e ,
-
=
µp×* ,
't
(I 142 ) x X O
-
- -
=
X X=
.
11×11270 as XFO
IN 2--1 Idk I
Lecture 9, page 18
Normal matrices
Lecture 9, page 19
Example
!
0 2
Show that the matrix A = is normal.
2 i
Also show that A is neither Hermitian nor unitary.
an*
-
⇐ ill : =
⇐ ÷.tl ::
*
AA -
-
(I ? ) ( Iz ! ) =
(I fi )
-
) → AH AA't
-
A is normal
but At =
( I Ii ) ⇐ ? )
t -
-
A A is not Hermitian
AAA
(I ti ) (to 9)
and =
t Iz A is not
unitary
=
Lecture 9, page 20
A quick summary
⌅ A complex n ⇥ n matrix is
I Hermitian if A⇤ = A,
I unitary if A⇤ A = In ,
I normal if A⇤ A = AA⇤ .
⌅ A real Hermitian matrix is symmetric.
A real unitary matrix is orthogonal.
⌅ If we have an Hermitian matrix
I all the eigenvalues are real,
I distinct eigenvalues have orthogonal eigenvectors.
⌅ All the eigenvalues of a unitary matrix have a modulus of one.
Lecture 9, page 21
MA212 Further Mathematical Methods
Lecture 10: Diagonalisation
Dr James Ward
⌅ Exercises 15 is on Moodle.
I Attempt questions 1, 4, 6, 7 and 8.
I Follow your class teacher’s submission instructions.
I Do it! Actually submit your homework!
⌅ Classes
I Go to them.
Ax = x
|A In | = 0,
m1 m2 mk
( 1) ( 2) ···( k) = 0,
(A In )x = 0,
E (A, ) = N(A In ) = {x | (A In )x = 0} = {x | Ax = x }.
1g a .
f-zi.iq/=o...ftYaIIa-n--o
IA tI3ko
-
4 X
-
o 6
-3 O
'
di -
I, 91=2 land D= -2,9-2=1
-
see ,
y
d l f -2=1
-
I
1=1 ,
CA -
I3 ) I
+ -
I l
l
-
l; :
3
raki
i O 6 I
,
'
I
l
2tN=3 - - -
-
-
" '
l
t
l
l
only
'
(
I LI erector → 91=1 I
Lecture 10, page 6
Diagonalisation
Proof: We’ll show this for the 2 ⇥ 2 case. The proof generalises.
The key here is that, if v1 and v2 are the columns of a matrix P,
0 1 0 1
B C B C
AP = A @v1 v2 A = @Av1 Av2 A
IA dIz t o
II 1×1=0 Ntl
-
-
-
-
- o a- Ii
÷) I fair H)
ai ca - '' *o -
-
A is real . a- - i
. so eigenvector (t );
so P -
-
ft t ) and D= ( f Ii )
In MA100, you saw that a real matrix (with real eigenvalues and
eigenvectors) is orthogonally diagonalisable if and only if it is
symmetric.
We now see which matrices are unitarily diagonalisable.
Proof: We will only prove this in the LTR direction. The RTL
direction requires theory that we aren’t going to cover in this
course.
This gives us
A⇤ A = (PD ⇤ P ⇤ )(PDP ⇤ ) = PD ⇤ In DP ⇤ = PD ⇤ DP ⇤
and
ok ;) fo : ) f :X : off : : )
note :
AA
't ' -
- -
- Ata -
-
'
, ,
earlier we
found d, = i . Ui =
(f ) i Ase i , Use -
I, if
.
=
(f ) .
= I Ci ) t i C -F) =
It i 2=0 At Us orthogonal
1101112 =
(f ) (! ) = l Li ) tic F) = I -
i 2=2 ,
I =
Ft ( f )
1102112 '
f!;) =
I Ci )t C -
i ) L T)
-
= I -
i 2=2 ,
I #
=
( ) ( I %)
YR 452
" and D=
ios .iq
Proof: We’ll show this for the n = 2 case. The proof generalises.
Let {v1 , v2 } be an orthonormal basis of C2 and let ⌃ be the sum
⌃ = v1 v1⇤ + v2 v2⇤ ,
we want to show that, for all 1 i, j 2, we have
vi⇤ (⌃ I2 )vj = 0 so that we can conclude ⌃ = I2
using what we saw in Question 11 of Exercises 13.
⇤ ⇤ ⇤
A= 1 v1 v1 + 2 v2 v2 + ··· + n vn vn
Avi = i vi ,
as required.
Note: The n ⇥ n matrices given by the vi vi⇤ are sometimes
denoted by Ei so that the spectral decomposition is
A= 1 E1 + 2 E2 + ··· + n En .
evaluesdr-i.dz = - i A is normal
on erectors E- Eff ) ,
A- Eff ) I
't
unitarily diagonis able
A- = did U , t ask UI
= i .
(! ) # ci -
is this # fi ) # u ti ) spectral decomposition
l :: ::L it :
"
"
" -
"
.
For these and all other exercises on this course, you must show all your working.
:
1. Show that, for x and y in Cn ,
1 1 i i
hx , y i = kx + y k2 kx y k2 + kx + iy k2 kx iy k2 .
4 4 4 4
2. Given that the Cauchy-Schwarz inequality holds in any complex inner product space, prove
that the generalised theorem of Pythagoras and the triangle inequality also hold. That is,
(i) if x and y are orthogonal, then kx + y k2 = kx k2 + ky k2 .
(ii) for any pair x , y of vectors, kx + y k kx k + ky k;
Hint: Follow the proofs given for the real case.
4. Find an orthonormal basis for the subspace of C3 spanned by the vectors (i, 0, 1)t and (1, 1, 0)t .
✓
6. (a) Which, if any, of the following matrices are Hermitian? Which are unitary? What does
that tell you about their eigenvalues?
✓ ◆ ✓ ◆ ✓ ◆
2 2i 0 i 0 1
A= ; B= ; C= .
2i 5 i 0 i 0
✓
7. Show that all diagonal matrices are normal.
Which complex numbers can arise as an eigenvalue of a normal matrix?
✓ p p ◆
1/ 2 i/ 2
8. Find all the triples (a, b, c) of real numbers for which the matrix
✓ is
a + ib c
=
CX . x> -
icx y > + icy
,
, y>
-
Ray y )
,
/ . ¥11 type x =
I cxty , xty >
=
¥ ( ex , xty > t Cy , Atys )
=
¥ ( ex , X s t c X. y s t <
g. x > t Ly y ,
> )
-1411 x -
y 112 =
¥ ex -
y , X y> -
= ¥ ( CX X y ,
-
> -
Ly , x -
y s)
=
¥ ( ex X . > -
ex y > ,
-
ay , x> + ay y > ),
¥11 x yn -
'
= IT ( CX . X s t c X. y s t <
g. x > t < y y ,
> - < x. X > t ex y > +
,
ay, x> -
ay y s)
,
=
& c < X. y > t Ly ,
x > )
=
pi (c x , xtiy > t i < y xtiy > )
.
= TE le x. x s -
is X.
y s t icy , X s
-
icy , y > )
= I' ( CX .
x > t Ly ys) , t Td Lex y , s -
Ly ,
X s )
Ig Hx type TE ex iy X
- = -
,
-
iy >
if Kx X iy =
,
-
> -
icy ,
X iy s)
-
=
¥ ( < x. x > + i say > -
ray X s ,
-
is < y y , > )
=
Ep ( CX X , s +
ay y ,
s ) -
I ( ex y ,
s -
ay × >7
,
ight TIC =
CX .
xstcy.us) t Td Lex y , s -
Ly .
x s )
-
Ep ( CX X , s + ay y ,
s) t
-
I ( ex y
,
s -
ay × >)
,
=
I ( LX ,
y> -
Cy , X > )
,
-
-
,
= -
,
note : Z t E =
2 Re CZ )
2. Li )
from Q1 , 11×-19112 = 11×1142Relay > thy 112
since X. y orthogonal ,
ix. y> = o, so 11×-19112=11×114 Ily 112
Hxty HE 11×111-11911
! :) l! ! :) l : : :o)
' '
s
:: :
.
. -
in
. .
.
And the first two columns containing leading ones form the basis of RIMI = Lin
{ (I )
, . ( !) }
( g : Hit fool
man since '
-
4 . let it -
f! ) and
(!) E-
" UH even
.T i.it#Ti=Fti=E
-
- = =
so ni -
-
¥, ri -
-
(g)
us
!
w.
fi ) f ! ) it HI ihr
< > o
-
#
=
.
-
.
=
i .
+ i. -
son o. -
ca .
wi > wi =
+
in -
Elio ) (f)
,
-
-
+ i
=L t
" wzil-cwa.ws = I =
Mitt = B12
(÷ ) ( If
therefore wi -
Ha, -
-
E- -
Fi
.
*a .
-
⇐ ⇐ =
Ii Iii : =L :* .
it.
✓ -
(I j )
Bt
( to )t
't
B -
-
-
-
o = ' -
B
.
⇐ ill : of %) C if
*
'
' *
-
-
'
.
✓
.
u -
* =
E -
-
⇐ of ( 9
' =
to
fill : of 71--491=1
'
* a- .
v u
10 - 7h -142 -
4 = 0
( d 1) Cd b ) and
42=4
- - -
-
O ,
so diet
when a- i. save
⇐ If f) (j jiff ) → ,
where x. =
zixz
-
(%) ( ) ✓
x,
take Xz -
- i . Xi
-
-
Z
.
=
when d- b solve
€4 I I g) →
( fi f / : ) where E- ix.
-
. .
, ,
(Yj ) ¥1
take na -
- i .
x. =
-
Zi .
so =
Hence matrix A has eigenvectors
( iz ) and
(I;) with eigenvalues d - I and b
I 1=0
x i
let det CB
'
d ti 2--0 and
-
-
Izd ) -
-
O
,
i.e .
then d -
( I, I I f) fo il g )
when it ' ink
solve so na
-
.
-
→
,
take Kei Xi i
(7h ) (4) V
-
-
so
-
, .
( ! I I g) ( bio I g )
when D= -
i . solve so x. = -
ixz
→
,
( Yaz ) ( ji )
take xz I ,
-
-
Ai -
- -
i . so =
I respectively
Et Ei E Fi
let det
I ? If and
-
CC Iz d)- -
-
o ,
i.e .
- o
, then of i - =
o D= and - -
f?
when a-Et Ei solve Ei
⇐¥1 Oo )
then
( I:) : ⇐ Er )
-
+
.
E Ei
( Ya)
'
! I f)
then
when a. solve
-
-
. -
and Et Ei E Fi
⇐s ) ⇐ )
Hence matrix c has eigenvectors . with
eigenvalues a- and - -
, ,
respectively
im ×
# t Thi
i
, •
;
'
t I
,
/ TE
y
• '
O
& re
( :b o.o ! ! ! ban )
aitbii o O -
> .
write a diagonal matrix as A -
o a "
, Honi
!
by definition .
't
A = # =
Ai -
b, i . . .
O . -
-
??
dans
° - '
'
.
.
(
af tbf
)
't'
then A# A = O O all
AA so diagonal matrices are normal
- - -
=
.
o aim .
. .
o
"
an'tbn
,
O O - -
.
/ /
a, -
X O - - .
0
since IA -
Intl =
o am . .
.
o
=
(ai -
t ) ( Az N -
-
-
-
can -
d) , then all entries of the diagonal
'
be
matrix possible eigenvalues
-
-
can .
O O an i, -
( a'?b ?)
'
it
s . denote A-
b) ( af? Y )
"
( tar:b if ) (Yj
' a '
a* at
-
is at A =
require so
-
- =
.
- =
( §) ffg)
hence have restrictions ib tri and E and thus
ga
we a- o b
for TER
-
so
-
- - -
-
.
=
atib tri
.
✓
-
=
.
A* A - -
-
sofa:÷÷÷÷÷ ÷÷÷ or
need more justification & process
d
*
city require A- A -
-
't
AA
( )
AA
't
= I 952 Til 4G -
blitz ) =
ATA derived from cii )
oyrzticblf CLE) -
a4b2 to
my :÷÷÷÷ . ..
:H÷ . or
rewire
aa*=⇐:* . .
.
⇐ "
ai:*:?)
-
-
at "
I:c ! .
it :: ) "
Ytb
"
AHA & AA't Hermitian if top right then bottom left equal
fine
are is is
equal
-
, .
C b
± Ft a4bZ=I
-
ba if C- take
F- any
-
it a. o and
!!!
men 'r
-
lit .
lot.÷÷kt÷H÷!
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Solutions to Exercises 15: Complex inner products and matrices
This document contains answers to the exercises. There may well be some errors; please do let me
know if you spot any.
1. As in Question 4 from Exercises 14, we just expand out the various norms and add. The catch
here is that we need to be aware that the inner product (or dot product) in Cn is not symmetric as,
in general, hx , y i = hy , x i. Starting with the first term, we have
kx + y k2 = kx k2 + hx , y i + hy , x i + ky k2 .
and, at this point, we could observe that, for a complex number z, z + z = 2Re(z) so that we have
kx + y k2 = kx k2 + 2Rehx , y i + ky k2 .
However, it’s probably easiest to leave the expression for kx + y k in the original form. Similarly, we
have
kx y k2 = kx k2 hx , y i hy , x i + ky k2 ,
so
kx + y k2 kx y k2 = 2hx , y i + 2hy , x i.
Now we can apply this result for the pair x , iy to get
so that, multiplying this by i yields 2hx , y i 2hy , x i. Thus, the whole right-hand side is equal to
4hx , y i, as required.
It is worth noting that these proofs work for any inner product, not just the dot product.
kx + y k2 = kx k2 + 2Rehx , y i + ky k2
kx k2 + ky k2 + 2Rehx , y i kx k2 + ky k2 + 2kx k ky k.
The rank of the matrix M is therefore equal to 2. The first two columns in the RRE form have
leading 1s, so the first two columns of M form a basis of the range: {(i, 0, 1)t , (0, 1, i)t }.
We can read o↵ from the row-reduced matrix that the nullspace is spanned by the single vector
(i, 0, 1)t . ( A word of warning: in the real case, the nullspace is the set of vectors orthogonal to the
row-space. In the complex case, that isn’t right: in fact the nullspace is the set of complex conjugates
of vectors orthogonal to the row space. )
You should check that this vector, (1, 2, i)t , really is orthogonal to (i, 0, 1)t .
To finish, we normalise this second vector w2 , to get v2 = p1 (1, 2, i)t .
6
Thus, the requested orthonormal basis is { p12 (i, 0, 1)t , p16 (1, 2, i)t }. This is not the only right answer.
Note: a surprisingly large number of students go on to find a third unit vector v3 orthogonal to both
v1 and v2 , thus obtaining an orthonormal basis of C3 extending the one required. If you want to do
this, you should make it clear in your answer that you’re just doing so for sheer enjoyment (and/or
more practice), and you should display clearly the answer to the question that was actually asked.
5. Subtracting two times the first row from the second leaves
0 1
1 1 1
@0 2 1 + iA .
0 1+i 1
One way or another, we recognise the second and the third row as multiples of one another: to me, it
seems natural to multiply the third row by 1 i at this point. So we eliminate the third row, divide
the second row by two, and add it to the first row, giving us the row-reduced matrix
✓ ◆
1 0 (1 + i)/2
.
0 1 ( 1 + i)/2
The vector v1 = (1 + i, 1 + i, 2)t is in the nullspace: we’ll leave it as it is and normalise it later.
(It’s convenient to multiply up by 2; in fact, v1 is a multiple of the slightly simpler vector (1, i, 1 i)t ,
and it would save a little work to use that vector instead.)
Rather than applying Gram-Schmidt, which is going to involve us in some awkward arithmetic,
let’s use the complex analogue of the method introduced in Question 1 of Exercises 14, to get
v2 = ( 1 i, 1 + i, 0)t and v3 = (1 + i, 1 + i, 2)t .
Finally, we normalise all our vectors to get
0 1 0 1 0 1
1+i 1 i 1+i
1 1 1
u1 = p @ 1 + iA , u2 = @ 1 + iA and u3 = p @ 1 + iA .
2 2 2 2 2
2 0 2
These vectors form an orthonormal basis of C3 which extends the orthonormal basis {u1 } of N (C).
There are plenty of other correct answers.
6. (a) By inspection, A and B are Hermitian: each diagonal entry has to be real, and each
o↵-diagonal entry satisfies aij = aji . (C is not Hermitian since c12 6= c21 .) That tells us that the
eigenvalues of A and B are real.
The matrices B and C are unitary: their columns are vectors of norm 1 that are orthogonal to each
other (while the columns of A clearly don’t have norm 1). That means that the eigenvalues of B and
C have modulus 1.
Combining the two tells us that the only possible eigenvalues of B are ±1.
(b) The eigenvalues of A are 1 and 6, with corresponding eigenvectors (2i, 1)t and (1, 2i)t respec-
tively.
The eigenvalues of B are 1 and 1, with corresponding eigenvectors (i, 1)t and (i, 1)t respectively.
The characteristic polynomial of C is x2 i, so the eigenvalues of C are the square roots of i. What
are these complex numbers? The best approach is to think about complex numbers z in the form
rei✓ , where r = |z| and ✓ is the argument of z. The square roots of z then have the formula ±r1/2 ei✓/2 .
In this case z = i has modulus 1 and argument ⇡/2, so its “positive” square root has modulus 1 and
argument ⇡/4: in other words p (think about the argand diagram) it is the multiple of (1 + i) that has
modulus 1, namely (1 + i)/ 2.
The two eigenvalues, the square roots of i, are thus ± p12 (1 + i). The corresponding eigenvectors can
be written as (1, ± p12 (1 + i))t .
7. For a diagonal matrix D, DD⇤ is a diagonal matrix with entries given by di di = |di |2 . The
matrix D⇤ D is the same diagonal matrix, so D is normal.
A diagonal matrix has its diagonal entries as its eigenvalues, and every complex number can appear
as an entry on the diagonal of a diagonal matrix. So any complex number whatsoever can be an
eigenvalue of a normal matrix.
Knowing that A is Hermitian or unitary tells us a lot about its eigenvalues: knowing only that A is
normal tells us nothing.
8. (i) For a 2 ⇥ 2 matrix to be Hermitian, the entries on the diagonal have to be real, and
the o↵-diagonal entries have to be complex
p conjugates of each other. So c can be any real number,
whereas we must have a = 0 and b = 1/ 2.
(ii) For the matrix to be unitary, the two rows have to have norm 1, and be orthogonal to each other.
Here the first row has already been fixed to have norm 1. To be orthogonal to the first row, the
second row must be a multiple of (i, 1) – if the second entry is 1, then the first entry x has to satisfy
0 = (x, 1)t · (1, i)t = x + i = x i, so x = i. We want the real multiples of this vector of norm 1, and
there are two of these, namely p12 (i, 1) and p12 (i, 1).
p p
So the answer is that a = 0, and either b = c = 1/ 2 or b = c = 1/ 2.
(iii) A matrix A is normal if AA⇤ = A⇤ A. Let’s write down what that means for this matrix A:
! ✓ ◆
1 p1 (a ib + ic) 1 2 2 i
AA =⇤ 2 and ⇤
A A= 2 +a +b 2 + ac ibc
.
1 i 1
p (a + ib
2
ic) a + b + c2
2 2
2 + ac + ibc 2 +c
2
Notice, by the way, that both AA⇤ and A⇤ A are Hermitian – you can easily check that this is always
the case, and we’ll be exploring this in lectures. This means that, if the top-right entries of the two
matrices are equal, then their bottom-left entries are also equal.
For the two matrices to be equal, we need, equating real and imaginary parts of all the entries,
1 a c b 1
a 2 + b2 = , p = ac and p = bc.
2 2 2 2
p
The second of these equations says that we must have either c = 1/ 2 or a = 0.
p
• If c = 1/ 2, then the third equation is automatically satisfied, so we may take any values of
a and b such that a2 + b2 = 1/2.
p p
• If a = 0, then b = p±1/ 2. If now b = 1/ 2, then the third equationp is satisfied for any
value of c. If b = +1/ 2, then the third equation reduces to c = 1/ 2, as in the other case, so
there are no new solutions.
p
In summary, there are two types of solutions. One is where a = 0, b = 1/ 2 and c is any real p
number – this is exactly the condition for the matrix to be Hermitian. The other is where c = 1/ 2,
and a and b are any real numbers such that a2 + b2 = 1/2, so the matrix is of the form
✓ ◆
1 1 i
p ,
2 z 1
where z is a complex number of modulus 1. This includes the one example where the matrix is
unitary but not Hermitian (z = i).
We know that every matrix that is either Hermitian or unitary is also normal, so the matrices in (iii)
have to include all those in (i) or (ii).
MA212 Further Mathematical Methods
Lecture 11: Singular values
Dr James Ward
A⇤ Au = u =) u ⇤ A⇤ Au = u ⇤ u.
kAuk2
kAuk2 = kuk2 =) = 0,
kuk2
as required.
Note: Given the previous theorem, we can also use AA⇤ to find the
singular values of A since A⇤ A and AA⇤ have the same positive
eigenvalues.
11=4
②
anti : : :L
(÷ :X : : :)
**
.
(
Evokes IAA't xtzleo
'
-
D= 4 or 6 4 21 O
=
'
"'
€
E values
/A*A 4131 -
-
o 4=0.46
same
1
vi = p Aui ,
i
(I I )
ai 't
4 o AA -_
⇐I
"" ,
thinks
"" " El :
Heil )
"o Fi
) Eli )
'
-
-
or.
aui
Fil : HE
"" " "
#
xx .
④ .
"
=
's
Is
f-
a- b.
use
it
1
u i = p A ⇤ vi ,
i
II
⇐÷
ant .
A'*
*
one-Hit
a- a.
ui-aif.fi/-rli)=E(I )
I
* me.
then we have
p p p
⇤ ⇤ ⇤
A= 1 v1 u 1 + 2 v2 u 2 + ··· + k vk u k .
Consequently, we have
p p p
⇤ ⇤ ⇤
A= 1 v1 u 1 + 2 v2 u 2 + ··· + k vk u k ,
as required.
Note: The singular values decomposition has some interesting
properties and, if we have time, we will talk about them again later
in the course.
*
of A- A
tf ! )
di -
-
4
Irf ! )
A -
Ui -
-
a- b -
u.
a- Eli't
uittru.us#--f4--
the SVD is A- Tx, I,
(f) ¥1101 ) tr -
fi (7) Fil -
I -
il )
add up
check :
the decomposition and get A
Lecture 11, page 20
A quick summary
Suppose A is an m ⇥ n matrix.
This means that AA⇤ is m ⇥ m and A⇤ A is n ⇥ n.
Let 1, . . . , k be the positive eigenvalues of AA⇤ and A⇤ A.
⌅ A⇤ A has an orthonormal set of eigenvectors {u1 , . . . uk } in Cn .
? x
? 1 ? 1
? ?
The ‘link’: ?vi = p Aui or ?ui = p A⇤ vi .
y i ? i
case I :X ,
X. U 1A XII 62=63
-
t tr
U, Vz ? 03
Dr James Ward
case 2 :
X. d. d IA XII Uz
-
- Us
k d
U, 62 ? Is ? IA -
NII Us -
- lk
⌅ Exercises 16 is on Moodle.
I Attempt questions 3, 4, 5, 6, 7 and 8.
I Follow your class teacher’s submission instructions.
I Do it! Actually submit your homework!
⌅ Classes
I Go to them.
1
P AP = J,
where J is a JNF.
as A is not diagonalisable.
We need an invertible matrix P that satisfies AP = PJ, i.e.
0 1 0 1
| | | | !
B C B C 1
A @v 1 v 2 A = @v 1 v 2 A
0
| | | |
(A I2 )v2 = v1 .
I ? I! I
Erawes :
I A- Neko a -
- -
l -
I
= o .
IA think
I ) (f ) ( J) (f) yl! ) forger
(
treaters : -
o l -
-
o a -
y -
-
o .
-
-
-
, ,
go for JNF !
a
p
C f) ( )
so ' '
and J
-
-
-
-
,
o y
as A is not diagonalisable.
We need an invertible matrix P that satisfies AP = PJ, i.e.
0 1 0 10 1
| | | | | | 1 0
B C B CB C
A @v1 v2 v3 A = @v1 v2 v3 A @ 0 0A
| | | | | | 0 0 µ
(A I3 )v2 = v1
↵v1 + v2 + v3 = 0.
v1 + (µ )v3 = 0.
0 1
0 4 4
B C
Put the matrix A = @ 1 0 3A in Jordan normal form.
2 4 7
E values IA -
XI31=0 a- 2. 2.3
Erectors ca ith
Iq (f) fi )
-
'- o . M :
'
Ui -
find
(II)
need to a vast CA ' Isla U
9,343 ) =L! )
-
-
take Vi
. -
g.
,
until
*'
. .
l: :"
Lecture 12, page 14
Case 2: Thrice repeated eigenvalue with two eigenvectors
as A is not diagonalisable.
We need an invertible matrix P that satisfies AP = PJ, i.e.
0 1 0 10 1
| | | | | | 1 0
B C B CB C
A @v1 v2 v3 A = @v1 v2 v3 A @ 0 0A
| | | | | | 0 0
(A I3 )v2 = v1
↵v1 + v2 + v3 = 0.
v1 = 0.
↵v1 + v3 = 0.
pal's ! ! ) .at ! ! ! )
0 1
2 1 1
B C
Put the matrix A = @ 1 2 1A in Jordan normal form.
2 2 1
f ÷ /
Evaluesia Misko-
"
-
""t
.
.
.
erectors CA dis )
⇐ y y)
-
x
-
-
o -
l l l
-
we want U2 sat CA -
NI 3) 02=01
" t"
erector
.
not erector
!)
! !
;
-
a- a. take or so as
!f! ) ,
take us
as A is not diagonalisable.
We need an invertible matrix P that satisfies AP = PJ, i.e.
0 1 0 10 1
| | | | | | 1 0
B C B CB C
A @v1 v2 v3 A = @v1 v2 v3 A @ 0 1A
| | | | | | 0 0
(A I3 )v2 = v1
(A I3 )v3 = v2 .
↵v1 + v2 + v3 = 0.
v1 + v2 = 0.
af! i. g) f! ! !)
'
0 1
-
# and
0 0 1
B C
Put the matrix A = @1 0 3A in Jordan normal form.
0 1 3
( -31=0
evacuees IA dI3l=o
- X O l
d- i. i. ,
I a
O l 3 X -
f! ! ! ) ¥1
erectors canst
TIE: .
eareoi-41.ca#soz-- "
l
÷
a
or
l
LA KI ) hack
-
Usa
( q) Lecture 12, page 22
extra
example session (Wb )
be tricky sometimes .
( ! ! ;)
i.
find the JNF of A-
I
y y
evalues :
IA XII
-
- o 3 X -
o -
I 2 -
X -
I D- 2) 3--0
o H, X 2 ( algebraic multiplicity of 3 )
-
/ -
erectors :
CA AI ) 7=0
-
l! ! t.in;÷ 4.)
"" case .
' .
. *
.
heed to find CA XI )
-
62=4 , take the
f !) ,
then be
§) and then
63=1 !)
f! ! ! ) f! ! ! )
so e-
-
and J -
-
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Exercises 16: Complex diagonalisation and singular values
For these and all other exercises on this course, you must show all your working.
✓ ◆
0 1
1. Show that the matrix A = is normal.
1 0
Find a unitary matrix U such that U ⇤ AU is diagonal.
0 1
0 2 1
3. Consider the matrix A = @ 2 0 2A.
1 2 0
Find the eigenvalues and eigenvectors of A. Write down an invertible matrix P and a diagonal
matrix D such that D = P 1 AP . (There is no need to calculate P 1 .)
0 1
7 0 9
4. Consider the matrix A = @0 2 0A.
9 0 7
Find an orthogonal matrix P such that P t AP is diagonal.
Express A in the form 1 E1 + 2 E2 + 3 E3 where the i are the eigenvalues of A and the Ei are
3 ⇥ 3 matrices given by Ei = vi vit where the vi are the corresponding eigenvectors.
Verify that, for i 6= j, Ei Ej is equal to the zero matrix and that Ei Ei = Ei .
✓ ◆
1 1 0
5. Find the singular values of the matrix A = .
1 0 1
Find the singular values decomposition of A.
7. Let A be a non-singular square matrix, with eigenvalues 1, . . . , k. What are the eigenvalues
of its inverse, A 1 ?
8. Let A be an invertible matrix. What are the singular values of A 1 in terms of the singular
values of A?
( f)
't O
I .
since all entries of A are real . A -
-
A =
C Ilf ) f II
"
(Y )
'
(: I ) fo f )
-
then a'ta -
-
-
-
, , o o
o
't
By definition . Atta -
-
AA implies that A is normal ,
and thus can be
unitarily diagonal ised .
① find eigenvalues HH
I
I -11=0 Ii
g 7,1
let IA o a-
-
i.e
-
=
we
-
.
-
,
.
. so
② find eigenvectors
when a -
- i .
fi Iif ; ) fo )
-
-
.
then
I :} . so e .
-
-
fi )
fi if;) fo )
men x
fix :} (t)
i then so E-
-
- -
-
. .
.
-
since I I -
=
( Y ) (f) . = i. it l -
T -
-
o , it is
orthogonal to I
=/ ( f ) ( f )
-
HEH .
-
- i-T-i.TT IT -
-
=P
so E #
(; ) and I E (t)
-
- -
-
therefore , u -
- CE Es ) -
-
for
( ji ) f /
I A- XIZI and Nti
O '
O then d o d It
-
A o
-
2. ca)
- -
= -
-
- : . ,
, ,
f:)
ie
El ! )
Iff ) foot
r
f!
when " " a man
- -
-
so
- -
- -
-
. .
when d ' -
"
f ! 4) ( f ) fo )
-
-
. so a-
(;) Han -
-
r is -
-
pili )
hence P -
-
( f for
"""" "
)
for I
Cb) O l
D= : '
# Kim 's
'
o, son out
-
-
= -
:p
o
II
d -
- O .
I 0 '
'
, ,
,
.
og g) (1) (g) %)
when 't-4 '
{ II:
'
E-
.
= us -
-
Hoshi
II ! )
'
sonia ee "
. .
3 .
To find eigenvalues .
let IA -
aI3l=o . ie .
I 1,1=0
'
-
NI -19 ) -
-
o and thus D= o or ±3i
ti : :L
when D=
I :&:÷÷÷ 't :L
O 2 I
O,
'
a. "" .
" 2
H ÷t¥t f.
when D= 3i , -3 i 2 I
and "" so be -4 Si
{
-
. -
X -
zy -
zit -
-
o
U
Es 44k¥ 'THE
f ÷ y :o) I:)
men " -
" -
men 's .
÷ !ii )
: II
fgs! ! )
"
therefore
'
. P . and "
/
(4) ( ( 7- d) 2-811=0
/
4
. let IA -
o z -
a o
← o
(2 -
d) ( Ib d) f
-
-
a -27--0
f o
7-a
D= 2 , -2 ,
16
f; ; :/ 't ! )
when '
is :*
so
.
.
f! )
'" "⇐
' '
f! I g ) ( I go ! )
'
therefore ,
P -
- with D=
:L yo: : : ) y :O: Y : : :)
⇐ am .
a- - ' " '
but
(I ) ( I ooo ! ) 2/1/1010
E- " o -
it =
A- ) 2
lyre
-
-
yrs )
-
o -
,
,
x
Esiesust
(g) =L ! ! ! )
-
-
" o ' '
+ a
eye o yr )
x
verification : Ei Ez =
E, E3 =
Ez E3 =
Q
E, E , =
Ei , Ez EE Ez ,
E3 E3 =
E3
f ! ly ) ( too )
( y g)
' '
definition
(f f )
't *
A l
is
by AA
=
so =
-
-
.
. .
, ,
1--0
/ yl / ( 2 Xi
let IAA't -
d Iz .
i.e .
2 -
= 0 so -
-
I = 0
,
,
(3 -
X ) Cl d) - - O
x =
I or 3
R and B
so the singular values of the matrix A is l
-
-
when * i .
(: i) /; ) -
-
(8) .
so u. -
f ,
I, If :/ -
fo ) .
so a-
f; ) and E- Eli )
Tre A
( g g) ( ; ) Is
(I )
'
using the link . we have a
-
-
't
I =
'
=
,
⇐
,
¥ # E-
oil ! ! ) rill ) tf ! )
a- -
-
-
# f )# ( O l tr
Irl ! ) Trio
=
l
-
l -
R l l )
,
il : : it El : ill
-
+
=
6 .
let It I ,
.
. .
.
. He be the
corresponding eigenvectors of eigenvalues a .dz .
- - -
.
XK
't
since A is Hermitian .
A -
-
A
then At Ui =
Alli =
dilli where I Eisk
and CAUI )
't '
Ui Ii
't *
At Ati =
A -
-
A di =
di CA ki ) =
Xi
.
Idk I
7 .
let Ui be the
corresponding eigenvectors where i sie k since A is non -
singular ,
di -1-0
since di .
. . .
.dk are
eigenvalues . AE -
-
dilli
A- AE
' '
A die
-
-
-
"
Iti =
A villi
'
A- II =
¥ II
by definition .
the eigenvalues of A
"
ane Ii where I sick
*
't
8 .
need
T.fi ¥
'
then
'
A- s
singular values are =
proof
B*A*
*
(AB) =
't
I -
LAA " ) =
,
so CA ) =
' '
CA 5
'
(AAA )
-
'
A' ( A Y 't 't
At CA t
- - -
= = =
'
this means of CA*A5
I
singular values of A will the positive square roots
-
be
MA212 Further Mathematical Methods
Lecture 13: Using the Jordan normal form
Dr James Ward
evabues ltzi I si
(Ii Ii ) ( Itoi
:
eio-coso-iis.no
-
,
erectors
¥ ) fi )
.
'
'O
.
' '
set '
e'Ote 0=200
' '
-
'
E- PE I I' =p API
' -
.
-
-
Ay y -
-
pz -_
APE -
DE
zi-ci-ziizzeio-e-io-zis.no
(Zz ) ( Hfi Iz ) (IL)
Zi lH2ilZi
-
and
=
; ( Itzik "-2Mt
ZI= Ae zz= Be
) (
'"
(I ÷ ) (
Ae 't Be
1- PE - =
"-2Mt
, Ae ziaeutziltfzi Batiste
-
{ tell-2Mt
2- "" 't 't
and
gas
, y, # se yarn . -
zie , zieti
Yolo) o D= 2iAt2iB
etcezitye Zit )
-
"
B=I
-
"t
zietce e- )
-
=
= - -
= zetcoszt =
get sink)
Lecture 13, page 5
Example 2
(y ! ! ) ( to ! ! )
' '
'
and F- that P API
-
F- so
I. espde
€15 Q'I '
'
set y Pz APE JE
' -
-
-
pz
'
-
APE 2- =P -
I-eetdt.ee t
-
.
zie-t-zie-t-AHBZze-t-zze-t.tt
11
⇐ et ) Izzet)
' '
-
At -113 -
-
A
Zi -
-
E- ' ±
( ÷ o://AEIIIY.to/et=...Z2=tAttBlet
' -
Ak = (PDP 1
)(PDP 1 ) · · · (PDP 1
) = PD k P 1
| {z }
%)
k times
D"
where, as D is diagonal, D k is easy to find.
When we have to use a JNF instead, the same reasoning works, i.e.
Ak = (PJP 1
)(PJP 1 ) · · · (PJP 1
) = PJ k P 1
| {z }
k times
0 1 0 1
4 4 3 6 2 5 5 4 10 3
B C B C
J34 = @ 0 4 4 3 A , J35 = @ 0 5 5 4 A, etc.
0 0 4 0 0 5
(Ii Ii ) ( fi Iz )
"
F- and D=
;
y µ Akyo -
-
-
-
=
( ) (
¥q
e- iojk
)
"
Zill ( B- eiojk +
'
( Itzik t Ci -
= =
-
ziutzisk -12in silk -
ti eiojktzicrseiojk
( )
, Z ( KO )
K
=
fs
Hzierseio asinine)
⇐ real
r -
-
Htt =f5
Tano
-
-
211
"
O -
- tan 2=1 -
167rad
Lecture 13, page 15
Example 2
0 1
0 0 1
B C
Solve yk = Ayk 1 for k 2 N when A = @1 0 3A as before.
0 1 3
af! ! ! ) tf ! ! ! ) and
) ( g Ky ! )
"
÷ ' ±
!!
' ""
II
'' "
yr A' yo PJ "
'
with J
"
- -
-
- =
①
A
y: I I :*::÷÷i÷÷÷÷÷÷i: :L
"" " "
:
=
:
-
. .
' '
til "t÷÷÷H¥÷¥ Lecture 13, page 16
MA212 Further Mathematical Methods
Lecture 14: Di↵erence equations continued
'
for large K , if Cfo ,
Lk ' Ck
dominant
Dr James Ward
if f- 0 Bto
Lk E BK
.
,
subdominant
⌅ Dominant eigenvalues and long-term behaviour
⌅ Age-specific population growth
if Go Ato
,Ik=Ak(÷ )
, 13=0 .
subdominant
if o_0 , 13--0 A- 0
,
,
Lk I
-
-
⌅ Exercises 17 is on Moodle.
I Attempt questions 3, 4(a), 5 and 6.
I Follow your class teacher’s submission instructions.
I Do it! Actually submit your homework!
⌅ Classes
I Go to them.
k
yk ' c v for large k,
(3×3)×(1×3)=(1×3)
→
k
yk ' c v for large k,
happens to make this entry zero, the theorem gives us yk ' 0 for
-
-
Pelo I g) Is
"
fi )
'
I
Yo - and P
"
- "so
, .
⇐o
.
f- II)
'
::
"
÷ !! =/ ! ! ! )
"
I'
'
' Kerner's =P ''
.
en
T
subdominant
an example of
behaviour
so -
.
-
92=41--4}?¥g )
after 30yrs ,
4433£53)
-
( )
O 4 3
f- yz o o d =-3 is the unique real e.values
° 44 0
since two successive fertile classes , d is dominant
=/ ! ) ( §)
' "
"
erector for a- E . - -
a ; in the
long term, yxec ⇐ I
class 18 :b I
proportion
:
the in each age
KEY
( !) ( f ))
' " '
also ya = -
- I Ects a -3%,
4+61611=0×3
- 0
IL AIK +"
-
= -
-1=0
¥3
"3 ,
'
°
a i > cantata)
-
so a
-
-
I,
moduli : Idf -
-
I ,
lad =
¥-1435 =L , 1h31 =
-41-1735=1 Ail is not dominant
y:÷÷K:÷÷M÷ : :L
"
4 1%4.4=4%1
" ii.
.
's
10 , It =L XI ,
12=441 -13=13# Io X4=L4Xo=LXo= XI
,
= ,
- - -
K 0.3 6
'
'
-
µ.gg
-
,
,
population evolves in a
cyclical way ,
so Xk= K -
i. 4.7 ;
-
-
. K =
2.518 . .
-
Lecture 7, page 20
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Exercises 17: Jordan normal forms
For these and all other exercises on this course, you must show all your working.
1. Prove that, if 1 , 2 are distinct eigenvalues of a square matrix A, then the corresponding
eigenvectors v1 , v2 are linearly independent.
Now show that, if A has three distinct eigenvalues 1 , 2 , 3 , then the corresponding eigenvectors are
linearly independent.
✓ ◆
1 a
2. (a) Let be a 2 ⇥ 2 matrix in Jordan normal form.
0 2
Explain briefly why the entry a is either 0 or 1. If a = 1, what can you say about 1 and 2?
✓ ◆
1 a
(b) Let S be the set of all 2 ⇥ 2 matrices in Jordan normal form with 1 2.
0 2
✓
3. Put the following matrices in Jordan normal form.
0 1 0 1
0 0 1 1 1 0
(a) A = @1 0 3A and (b) B = @0 0 1A .
0 1 3 1 1 1
(a) Use the information given to determine the eigenvalues and corresponding eigenvectors of A.
✓
6. Find the general solution to the system of di↵erence equations
For which initial values (x0 , y0 , z0 ) will zk be positive for all sufficiently large k?
I .
① Since ×, and d2 are eigenvalues , we have All , di U,
-
-
and Alk -
-
Az Az
a -1
- -
-
,
Adi I t Pdz Lk =
O ①
② Similarly ,
hi .
dz .
As are eigenvalues ,
we have Alf dik -
-
.
A E- AZIZ and Aks =dz Use
, a -1 t =D
HADI t @ dy Lk t Axs) Us =D ①
2. Ca ) if a-4 , di -
-
Xz and they generate the same eigenvector .
( If I ) (Tf k )
I only take the form or
,
Cb )
suppose these exists two matrices in s that are similar ,
"
then
(
dip, =
dip, t bPs b
-
-
O or 13=0
di Ps =
dzB 4=12
app
II:p:
"" " " ' -
bP4
aps t da Pg a .
.
. . ..
:3
/ 1=0
oh let IA XI31=0 C-A) [ C-A) 13 N -13 ) t l O
-
.
-
. ice .
, so -
, * -
O l 3 X -
-
Mld -
l) -13kW -
t) -
Catt ) ca 1) =p -
(X 1) C at 3A X
-
- -
-
I 7=0
2
-
D 1) -
= 0 d, =
42=43--1
is :÷÷÷÷a
. ""
÷
men ." .
so a
(I ! ÷ )
= -
.
. .
=
¥ Hi:L
"" ⇐
÷÷÷÷i
"" " '
Es
=
we
- - -
=
;
- .
-
, , .
fi Ii g) ( g ! ! ) (÷ ! ! )
' ' "
therefore
"
A- PIP -
)
X I
/g
I
-
-
Cb) let l B -
AI 1=0 , ie .
so ( th ) lack t ) - -
I ]t l -
Lott ) -
o
y y
=
o ,
, N( 2- x) O
-
-
( if Y ) (¥ ) =/ ! ) (f) ✓
'
when arts - o.
,
so a -
A -
o and take a
-
-
i . then a-
o
Z1 O
-
=
, , ,
Xi Y , -12-1=0
-
HEE
µ ! } ) =/ ! ) f! )
then need CA ' ie E
!
we
-
-
. -
.
'
÷ I :L ÷÷÷÷÷ Hr
men 's . . .
"" "
f! ! !/ ) fo ooo ) f ! ! ! )
therefore B - " JP ' -
-
' "
need to make p invertible
O O 2
×
your P is not invertible
'
A PJ P
by remembering
-
* AP PJ
-
- -
-
from
fog 0g } ) f! I g ) (g ! !)
'
and J
!
"
4. ca) seas . A '
-
PSP where P
- -
- -
- -
- -
"
=P APE JI
' '
set y = P2 Ay APE
'
then y =P? =
so z
- -
- -
, ,
from ③ ,
Z3= Aet , then write ⑦ as Zz
'
-
Zz
-
-
2-5- Aet
et
'd ' T t I
(Att B) et
'
Eze
-
I - -
-
-
A then Zze =
AT TB Zz -
✓
- -
. ,
Zi (Att B) et
finally rewrite ① as Zi
-
- -
'd'
taking I -
-
et -
- et , G- ie
t
-
)
'
=
Att B , then z, e
t
-
4¥ ) III ) :
"
hence we nave ⇐
-
-
et
men
.ae#y.i.ii:lf:::i.tyee=f:::.:.i::::..::i :D :: .
-
..
teeny :÷ feet!:L
=
Cb )
5. Ca) from J , we get the eigenvalues from the diagonal entries :
Ai -
-
I ,
Az -
I and he -1
corresponding eigenvectors
f? ) (f ) f!) ✓
their are
e
-
-
, E- and E-
×
T
JE derived Qa ca)
'
Cb ) set y Pt so Z as at
eigenvector
-
-
- -
,
not an
my:÷i÷x÷ii÷÷÷÷÷ :
t
from ② and ③ Zz Aet and
-
we
get -
Zz Be
-
-
,
Aet
'
d'
since I=
et '
-
-
e
t
-
. He
-
t
)
'
=
A , then z, e
t
-
-
-
:÷ )
nen. ⇐ C)
n
.
yo: If :÷
+ c) Be '
so
.*
-
-
.
B to
go
# EE
( AaT¥¥÷) ( FA )
t
when t so -
, y =
→
④ since Kil -
-
taek last -
-
I . there's no dominant a .
mm
fo: Ii I ) ( g II )
"
"
" ' '
'll )
''
-
-
-
so this g. ate
can't use . .
.
.
Y: : ÷ ) ¥)
'" " " '
. ..
if
§)
as k -70 Bto yk = Bk
, ,
X
if 13=0 , Ato ,
C -
-
o , 9k =
A
§)
(§ )
"
if 13=0 , A
-
-
O , Cfo , 9k =L -
1) C
if A Bec -
-
- o
, yk -
-
e
'
Cd ) y , Ct ) → o Aet -1 BE = 0 as this A
-
- o
92107=0 Ct 13=0 Be C-
iii. Hii :S
' '
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Solutions to Exercises 17: Jordan normal forms
This document contains answers to the exercises. There may well be some errors; please do let me
know if you spot any.
1. Suppose that Av1 = 1 v1 and Av2 = 2 v2 with 1 6= 2 and the eigenvectors v1 and v2 are
non-zero.
Suppose that ↵v1 + v2 = 0 (†). Applying A to each side gives
So S consists of diagonal matrices of type (i) with 1 2 , and all matrices of type (ii).
✓ ◆
1 1
Note: A matrix like , with 1 6= 2 , is not in Jordan Normal Form!1
0 2
We know that similar matrices have the same characteristic polynomials. A matrix
✓ ◆
1 ⇤
,
0 2
has characteristic polynomial (x 1 )(x 2 ). So 1 and 2 are uniquely determined by the char-
acteristic polynomial, subject to the requirement that 1 2 .
So the only pairs of matrices in the set S with the same characteristic polynomial are those pairs
✓ ◆ ✓ ◆
0 1
= I and .
0 0
These two matrices are not similar, as the first has two linearly independent eigenvectors — indeed
the eigenspace corresponding to eigenvalue is the whole of C2 — and the second does not.
1
Such a matrix has characteristic polynomial (x 1 )(x 2 ), and thus has two linearly independent eigenvectors
(see Question 7 from Exercises 5) and can be diagonalised, so it is similar to (i) — if you think you proved that these
two matrices are not similar then you may wish to go back and find your mistake. A matrix in JNF consists of Jordan
blocks arranged on the diagonal, and each block has all the diagonal entries equal — any diagonal matrix is of this
form, with all its blocks being of size 1.
(Alternatively, if P is any invertible matrix, then P 1( I)P = I: the only matrix similar to I is
I itself.)
So no two matrices in S are similar. We establish (more or less) in lectures that every 2 ⇥ 2 matrix is
similar to a matrix in S: this exercise shows that each 2 ⇥ 2 matrix is similar to exactly one matrix
in S.
3. For (a), the characteristic polynomial of the matrix A is x3 3x2 + 3x 1 = (x 1)3 and so we
have an eigenvalue of 1 with an algebraic multiplicity of three. Following the method given in the
lectures, we then solve
0 1 0 1
1 0 1 1
(A I)x = 0 =) @ 1 1 3A x = 0 =) x = @ 2A
0 1 2 1
is the sole eigenvector. So, we take v1 = (1, 2, 1)t and seek a vector v2 such that
0 1 0 1 0 1
1 0 1 1 1
(A I)v2 = v1 =) @ 1 1 3A v2 = @ 2A =) v2 = @ 1 A ,
0 1 2 1 0
will do the job. We also need a vector v3 such that
0 1 0 1 0 1
1 0 1 1 1
(A I)v3 = v2 =) @ 1 1 3A v 3 = @ 1 A =) v3 = @0A ,
0 1 2 0 0
will do the job. Thus, taking these vectors in the required order we see that
0 1 0 1
1 1 1 1 1 0
P = @ 2 1 0A gives P 1 AP = @0 1 1A = J,
1 0 0 0 0 1
which is the required Jordan Normal Form.
For (b), the characteristic polynomial of this matrix B is pB (x) = x3 2x2 = x2 (x 2) so have
eigenvalues of 2 and 0 where the latter has an algebraic multiplicity of two. Following the method
in the lectures, we then solve
0 1 0 1
1 1 0 1
(B 0I)x = 0 =) @ 0 0 1 x = 0 =) x = 1A ,
A @
1 1 1 0
is the sole eigenvector corresponding to the eigenvalue = 0. So, we take v1 = (1, 1, 0)t and seek a
vector v2 such that
0 1 0 1
1 1 0 1
(B 0I)v2 = v1 =) @0 0 1A v2 = v1 =) v2 = @ 0 A,
1 1 1 1
will do the job. We then solve
0 1 0 1
1 1 0 1
(B 2I)x = 0 =) @0 2 1A x = 0 =) x = @ 1A ,
1 1 1 2
is an eigenvector corresponding to the eigenvalue = 2 and so we can simply take v3 = (1, 1, 2)t .
Thus, taking these vectors in the required order we see that
0 1 0 1
1 1 1 0 1 0
P = @1 0 1A gives P 1 BP = @0 0 0A = J,
0 1 2 0 0 2
which is the required Jordan Normal Form.
4. I do hope you spotted that the matrices here are exactly those in the previous question.
(a) The method is to put y = P z , where
0 1
1 1 1
P =@ 2 1 0A
1 0 0
is the change-of-basis matrix from the previous question. This transforms the system of equations to
0 01 0 10 1
z1 1 1 0 z1
@z20 A = @0 1 1A @z2 A .
z30 0 0 1 z3
The matrix here is of course the matrix P 1 AP in Jordan Normal Form that we found in the previous
question.
Now we solve the equations one after another, working from the bottom row upwards. To start, the
general solution of the equation z30 = z3 is z3 = Xet , for some constant X. Then, we find that
Z
0 t t
z2 z2 = Xe =) z2 e = X dt = Xt + Y,
z2 = Xtet + Y et ,
t2 t
z1 = X e + Y tet + Zet ,
2
for some constant Z as the general solution for z1 .
Consequently, we have
0 10 2 1 0 2 1
1 1 1 Xt /2 + Y t + Z Xt /2 + (Y X)t + (Z Y + X)
y = P z = e t @ 2 1 0A @ Xt + Y A = et @ Xt2 (2Y X)t (2Z Y ) A .
1 0 0 X Xt2 /2 + Y t + Z
If your answer doesn’t exactly match this in form, it may still be right — one can rewrite the constants
above in terms of (for example) a = X/2, b = Y X, c = Z Y + Z and get another correct solution.
Another thing to note here is that we never needed to use the inverse P 1 .
(b) However, for this question, we do need the inverse of the change-of-basis matrix P we found
before, so that P 1 BP = J. We write y 0 = By + x , set y = P z , and obtain P z 0 = P Jz + x so
that z 0 = Jz + P 1 x . Here
0 1 0 1 0 1
1 1 1 1 3 1 0 1 0
1
P = @1 0 1A , P 1 = @2 2 2A and J = @0 0 0A .
4
0 1 2 1 1 1 0 0 2
Also P 1x =P 1 (0, 0, 1)t = 14 (1, 2, 1)t . Thus the equations have been transformed to
0 01 0 10 1 0 1
z1 0 1 0 z1 1
@z20 A = @0 0 0A @z2 A + @ 2A .1
4
z30 0 0 2 z3 1
Again we solve the equations from the bottom up so that we have
Z
0 1 2t 1 1
z3 2z3 = =) z3 e = e 2t dt = e 2t
+ X,
4 4 8
as the integrating factor is e 2t and so we have
1
z3 = + Xe2t .
8
for some constant X as the general solution for z3 . Then we have
1 1
z20 = =) z2 = t + Y,
2 2
for some constant Y as the general solution for z2 . Finally, we have
✓ ◆
0 1 1 1 2 1
z1 = t+Y + =) z1 = t + Y + t + Z,
2 4 4 4
for some constant Z as the general solution for z1 .
Consequently, we have
0 1
t2 /4 + (Y 1/4)t + Xe2t + (Y + Z 1/8)
y = Pz = @ t2 /4 + (Y + 1/4)t Xe2t + (Z + 1/8) A .
t/2 + 2Xe2t (Y + 1/4)
5. (a) From the diagonal entries of J, we can read o↵ that the eigenvalues of A are +1 (with an
algebraic multiplicity of two) and 1. The corresponding eigenvectors are (0, 1, 1)t and (1, 1, 0)t , the
first and third columns of P .
Notice that J has a Jordan block of size 2: if v1 and v2 are the first two columns of P (the two
columns corresponding to this block), then (A I)v2 = v1 6= 0, and so v2 is not an eigenvector of
A corresponding to = 1, but v1 is.
In general, if J breaks up into Jordan blocks, the leftmost column of P corresponding to each block
is an eigenvector of A. (For a diagonal matrix, of course, this means that every column of P is an
eigenvector.)
(b) As usual, we first solve z 0 = Jz . The solution to this is, working up: z3 (t) = ae t , z2 (t) = bet ,
z1 (t) = btet + cet , where a, b and c are arbitrary constants.
0 1
bet + ae t
The solution to the original equation is then y = P z = @btet + cet + ae t A.
btet + (b + c)et
(c) As t ! 1, the “dominant” term is btet . If we divide the solution by that dominant term, we see
that 0 b 1 0 1
a
1 t + te2t 0
y = @b + c + a2t A ! @ b A
tet t te
b + b+c
t
b
as t ! 1. (This assumes, of course, that b 6= 0. What happens if this is not the case?)
0 1
0
We might then write y ⇡ btet @1A. Here b is an arbitrary constant, which might be negative.
1
The above expression tells us about the rate of growth of the largest of the yi (t), and the relative
sizes of the yi (t) for large t. Whether this is a satisfactory answer depends on the context. If we
want to know how y2 (t) behaves for large t, this is fine: y2 (t) behaves as btet , for some constant b. If
we want to know the approximate relationship between y2 (t) and y3 (t) for large t, this is also fine:
y2 (t)/y3 (t) ! 1 as t ! 1. We also learn that y1 (t) is “much smaller than” y2 (t) and y3 (t) for large
t, but we might be interested in knowing “how much smaller”. Of course we can read o↵ the answer
to that question as well: y1 (t)/et ! b as t ! 1, so y1 (t) is approximately y2 (t)/t for large t. (More
precisely, ty1 (t)/y2 (t) ! 1 as t ! 1.)
(d) The given conditions tell us that b = 0, and then that c = a. So the solution is: y1 (t) = ae t ,
y2 (t) = a(e t et ), y3 (t) = aet .
0 1 0 10 1
xk 3 1 0 xk 1
6. @ A
We write yk = 1@ 1 0 A @ yk 1
A and call the matrix A.
zk 1 1 2 zk 1
for s, t 2 R. That is, we find that (1, 1, 0)t and (0, 0, 1)t are two linearly independent eigenvectors
of A. Of course, we can’t diagonalise A as we are one eigenvector short, and so we go for the JNF
instead. Thus, taking note of the matrix A 2I, we choose v1 = (1, 1, 1)t to be our eigenvector so
that 0 1 0 1 0 1
1 1 0 1 1
(A 2I)v2 = v1 =) @1 1 0A v2 = @1A =) v2 = @0A ,
1 1 0 1 0
will do the job. We then take v3 = (0, 0, 1)t as the other eigenvector as it is not a scalar multiple of
the eigenvector v1 . Thus, taking these vectors in the required order we see that
0 1 0 1
1 1 0 2 1 0
P = @1 0 0A gives P 1 AP = @0 2 0A = J,
1 0 1 0 0 2
So we have
0 10 k 10 1 0 1
1 1 0 2 k2k 1 0 0 1 0 k+2 k 0
Ak = P J k P 1
= @1 0 0A @ 0 2k 0 A @1 1 0A = 2k 1@
k 2 k 0A
1 0 1 0 0 2k 0 1 1 k k 2
and the general solution of the recurrence relations is
0 1 0 1 0 1
xk x0 k(x0 y0 ) + 2x0
@ yk A = Ak @ y0 A = 2k 1 @ k(x0 y0 ) + 2y0 A .
zk z0 k(x0 y0 ) + 2z0
Lastly, zk = 2k 1 [k(x
0 y0 ) + 2z0 ] is positive for large k if either x0 > y0 , or x0 = y0 and z0 > 0.
MA212 Further Mathematical Methods
Lecture 15: Sums and complements
Dr James Ward
U + W = {u + w | u 2 U and w 2 W }
U \ W = {x | x 2 U and x 2 W }.
Proof:
Proof:
u u
W
-
- - -
- - -
•
- - -
.
.
.
few aww
j U
unw fo }
-
-
Uaw
lytbl
-
=
+ -1
{ (g)
'
'
}
utw
-
Lin
R
-
, ,
Univ XEUTW ,
so XEU , X -
(§ ) and HEW , *
(f)
f! ) (g) .
-
e-
(g) '
a
unw=linµ ) )
Lecture 15, page 6
Direct sums
U \ W = {0}.
with OE V and W
v. utw WE in
exactly one way
OE Un w so that I C- U and I C- W E e
-
- (
by restriction of one
suppose
I Eto E It I
doing
-
it )
-
-
way
C- W EU C- W
C- U
p the only member of unwise
since W is a
i. e .
Un w -
-
fo )
subspace
U \ W = {0}.
Proof: RHS -
given Un w -
-
fo }
is
with UE v and we w
say I =
Uit id and I =
Ust we with UI , KEV and Wi W ,
-
EW
C- U E W
- n
ki t wi -
-
Ust we ki ez wa we ,
- - -
as Uiw are subspaces and have CUVA
Ui -
Uzeunw
-
and wz Eze
- un w so there's actually only one
way of
-
UI E
- - o and we -
y, so as un w =
Ut w is direct
y us and
-
we we
-
-
{u1 , u2 , . . . , uk } is a basis of U,
is a complement of U.
in IRZ in
U ,
a complement
u a complement
of W
,
E ; e
of w
w , a complement :
of u
pi -
-
U⑦ W Rs = U ⑦W
(2) V = U U ?,
(3) dim(V ) = dim(U) + dim(U ? ),
(4) (U ? )? = U.
Proof:
. Ek ) is ON basis of U
'
. . s .
,
- -
C- U f Ut e u e Ut
- -
VEV be written as 1- An
now any vector can I a Uit tdkuk take, Uet, t UI
- - -
- -
=
-
so It Ut Ut ,
thus VE Ut Ut
Ut E V then Ut Ut V
clearly
-
Ut
-
,
,
② Ut
if U Eun , then UEU and UE Ut
SO LI I > .
= 0 So 1--0
hence u
-
-
not at
(3) dim IV ) -
n Kt ch K) dim ( UJ t dim cut )
- -
- =
T t
basis of Ut
a basis
of
et ) UE Utt ( proof in Qs of ex 18 )
by Ut is also
a) ,
a
subspace
by . V = U Ut and V = Uh ⑦ Utt
so dim ( UI -
- dimer ) -
dim cut ) =
dim ( Utt )
at
49,9 { =
filarian '
meats .
- tin
I} ! verify
:[ fool -141/1447 ) )
I
C- RCA) C- MCAT)
(g ! ) →
looks trader
!
=p ) .gg/=o.er-ier--o
! , for all a B. VER
(g ) ) )
,
'
Rathlin { i
to
vector RCA) is every vector
,
in t
i
,
every
I in MCAT ) ,
i. e.
MCAT ) -
-
RCA)
'
Dr James Ward
⌅ Projections
⌅ Orthogonal projections
⌅ Exercises 18 is on Moodle.
I Attempt questions 1, 2, 3, 4, 5 and 6.
I Follow your class teacher’s submission instructions.
I Do it! Actually submit your homework!
⌅ Classes
I Go to them.
/
in
Geometrically... 1=1-11
V=U④w
¥
R2 EW
in
I fue EU
e.
E- ate
e
the
✓ = U④ W i
C- U EW i
.
a
TLE) I -
TLE ) -
-
I
Lecture 16, page 3
Questions, so many questions...
Rn = R(P) N(P).
y = Px = P 2 x = P(Px ) = Py = 0,
v = u + w,
Pv = Pu + Pw = Pu + 0 = Pu 2 R(P),
Let’s now see how we can find a matrix which represents a given
projection.
We’ll look at two methods. ‘Method 1’ follows what we have seen
before whereas ‘Method 2’ will be more useful in the next lecture
and beyond.
our projection is
0 1
| | | | | |
B C
AB,B
T = @e1 e2 · · · ek 0 0 ··· 0A ,
| | | | | |
B '
{ (! ) .
(I ) ( f ) )
,
,
is an
C-U
ordered
EW
basis of IRS =
a⑦ w
IT Ati
f ! ! !)
"
mis . and -
,
onears :
Ai .
at
t
' R LAT) -
U
find Mps
-
§ !)
M LAT) = W
'
then AT =
MBABIB Mf = . . .
Suppose that
⌅ A is an n ⇥ m matrix of rank m and
⌅ B is an m ⇥ n matrix of rank m
so that the m ⇥ m matrix BA is invertible.
The projection of Rn onto R(A) parallel to N(B) is then
1
P = A(BA) B.
PA = A and BP = B,
{ a II AE for I ERM }
'
RCA) -
-
-
-
A ( BAMBI )
C- RCA SO Rep) ERA
-
ERM
"
take U' ERLA) , then I
'
-
-
AE
'
ER
=
PAE C- Rcp ) RCA) E Rcp) RCAKRIP)
③ Ncp) - N CB)
proof : NCP) -
-
ful PIE } MB) -
-
{ I 1131=0 }
take any EE NCP) ,
take any It NCB)
have O
we have Py=o we BI -
- -
It
Nlp) MLB)
-
-
Example of Method 2 (projection)
projection of Pe RCA) to
'
P ALBA) onto "
NCB)
-
-
-
B is the
RLA) - U
A -
(! ) 3×1 rank I
checks : P'=P
NCB) W -
1×3 rank I
) U
-
I
Ba ( z 3 RIP)
-
-1=4
( f) Mcp) w
BA ( 3 -3 l) 4 ( BA )
-
- = -
(1) Ics ( § Ig ! )
p ACBAHB
-
-
-
-
-3 1) =L
Geometrically... in IRS ut
intr ut w_µfE
I , I
:') I
' "
TT
→ u
e al k dad
-
Tcu ) -
-
u
U -
Utw
- -
Tulku EU C- Ut
-
C- U Gut
Lecture 16, page 16
Symmetric idempotent matrices are orthogonal projections
B C
Find the orthogonal projection of R onto Lin @1A .
3
>
: >
;
1
we need to find a basis of Ut .
Lin { (tf ) .
(I ) )
→ LI and t to
/ !)
( (!) foz) µ ))
'
ordered basis B
-
-
of R
,
,
,
w -
check :
PIP
onto H
pt =p
( ! Io I ) (g ! ! )
'
MB -
-
and AFB =
pep ) = u
,
u w
j Nlp) - Ut
"to "
find Most ,
then penn BAYBayt
'
=
. .
.
80 1 t0 19
>
< 1 0 >=
B C B C
Find the orthogonal projection of R onto Lin @ 2A , @ 1 A .
3
>
: >
;
0 1
a
II. at .
-
fo ? ;)
**
to ? I
,
)f÷ ! ) .
it *tart .tl : : )
projecting MAIN to NCB)
(
' onto
If
)
2
then D= ALATA )
-
At -2 -2
=
→ ,
y onto RLA) NAIT)
-
z -
i y
t
projecting [4 to
11
NAT )
so '
p ALBA) B
-
-
-
A CATA)
'
At
-
Co ! )
ffi ) ( do 1) ¥ )
have E- Je with J and
suppose we
-
=
' '
idk
Z, =
XZI t Zz Zz =
AZZ
zi -
AZ , = Aet Zz -
-
Ae
't
since IT = XZZ
eftdt.e.ae/dEI--fxdtzie-H-xe-Hzi--AeHe*td-dyCz,e-dt)--
If .
At -113
In 12-4 -
atte
" 't
2- it) -
-
Catt B) e
'
Zz =
Ae
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Exercises 18: Dominant eigenvalues, sums and complements
For these and all other exercises on this course, you must show all your working.
1. An n ⇥ n matrix A is called nilpotent if there exists a positive integer k such that Ak is the
zero matrix.
(a) Show that, if all the entries of A on and below the diagonal are zero, then A is nilpotent.
(b) Show that, if A is nilpotent and B is similar to A, then B is nilpotent.
(c) Show that every n ⇥ n matrix is the sum of a diagonalisable matrix and a nilpotent matrix.
(d) Show that, if is an eigenvalue of a nilpotent matrix, then = 0.
(e) Show that 0 is an eigenvalue of every nilpotent matrix.
µ :)
, .ca, as
required . write A -
- O AR AB . . -
Ain
a "
:3 %
, .
÷:÷÷÷:÷t÷÷÷÷÷
"
÷:÷÷÷ : :
""
:p ÷ :÷÷÷:÷ )
'
1
AK 923 912 Azn
:÷: : :: ::
= O O - - - -
-
÷
. . . . .
.
- - -
- L -
O O O
÷ :)
"
a a
-
- -
- -
so there exists k
-
-
h . sit . Ak Ah -
-
-
-
e
'
P AP
-
Cb ) since B is similar to A , B -
-
where P is an invertible matrix
if
"
A is nilpotent , then there exists k St -
A -
-
e
'
Bk P Akp I for invertible P
-
so
-
- = all
can as normal ,
then
( ! ! !! =/ ! ÷ :o) to !! ! )
J
* * *
" '
nilpotent it in diagonal ,
- - -
= -
-
' '
p Ipp
-
P MP
-
A= P lait D) p
-
so = t
T F
nilpotent diagonalsable
from Cb ) matrix
Cd) since d is an eigenvalue of matrix A then AI XI -
-
,
"
AK I -
=
A
""
.
AI = A
"
.
AI = x (AK
-
'
g) = d A -
"-2
CAI) =
d ( Ak
'
-
2) I = . . .
=D I
since Xto , to
Le) FA , Ak - O
then AKI -
-
MI -
-
o , since I fo ,
a
-
-
O for every nilpotent matrix A ?
2 .
if A is diagonalis able ,
then there exists an invertible p sit .
A PDP
" -
-
÷:÷÷n÷÷÷÷÷n÷÷÷÷÷i
win .
µ!
IT, O - . .
O
denote
'
D
:
- "
. . .
(P p )
- -
Cpp 1) D p D
'
hence A =P D
'
P
' ' -
D D P =
D =
'
define 'P be
A written A B
B PD diagonalis able can as
-
hen matrix
-
"
since P P also P (eigenvectors ) not
, , D are
unique , B is
unique . is
unique
×
B is not unique
"
A diagonalis able PJP
if is not ,
then A-
'
J may
( dy j ) PA
since take the form ,
cannot be written as A -
B
,
( :b: )
✓
(g : ) ( g !)
a "'
from J form ,
a c -
and blate ) =
Zab -4 b -
Ia , then J
'
-
µ r¥ )
,
does not hold when d -
O
g§ )
3. ca, at eigenvalue a -
- E .
denote a-
then LI -
-
de
l:÷÷÷÷n¥ ÷⇐ ÷÷¥÷÷÷:* :
"
¥.
it
10
(b) 12,000 x
lot ft 3 -11
' =
b. 000 ✓
"
cc ) 120001 ⇐ ) =
208
X d is already for 10 yr period ,
12,0001 D= 8,000
4
given diets C- 3 tri ) then its complex dugate a Ttt 3 ri )
-
con
- -
-
.
,
I ¥3 EEE x ¥3x¥ ,
/ 1=0
let l L
-
RI Ko , i.e .
-
x
'
then ⇐ 747 -
-
⇐
o
-5 o
AE t 2
-
° E -
,
r ✓
since last -4×4=45-17257 =
IF 15 < 1-2 ,
a- 1.2 is the dominant eigenvalue
. . .
.
(z ) fi;)
us -
-
-
-
(Ig )
"
cu 23 and
in the long term , yX e .
the
proportion in each age
chassis 18:10 :S
(! ) (§ )
""
since ya Chek [ Cll 2) 1.29kt
= = 1.2 -
) I
,
so
age groups grows at rate of 1.2
✓
*
subspace : non -
empty ; CWA ; cu SM
CU VA :
if I , I C- St , then I d. £7
-
- o t E ES and she .
I> = O VEE S
Salsa I > St }
o
FIE
-
= . -
,
test SECS It
'
SES
-
( to )
° 3
6 .
RRECAI -
-
, ,
f? ) )}
3
( too )
since then MAI
Ling
-
-
= o .
, ,
of
{ ( f ) (L ) }
since rank A 2 -
-
RCAF Lin
,
,
(! !) § !) { ( f) }
At -
. ARE CA )
' =
. so NCA's =
Lin
( %) § ) }
since rank of At then RCA)
{
-
- 2 ,
-
-
Verify
[ (¥ ) ) In µ ) )
that a t B . = -
San -
3 RB -
an -
z Br t loan -15 fr -
-
O
C- RLAT ) E N CA)
[ (3) tell ) )
a .
In D= o
C- RLA) EM CAT )
therefore , RCAT) and NCA) are orthogonal complements , as are RA) and MCAT)
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Solutions to Exercises 18: Dominant eigenvalues, sums and complements
This document contains answers to the exercises. There may well be some errors; please do let me
know if you spot any.
1. (a) For each t, if we take the tth power of such a matrix A, all entries on the first t diagonals
on and above the leading diagonal, as well as all entries below the leading diagonal, are zero. In
particular, An = 0.
(b) If B = P AP 1, where P is invertible, and Ak = 0, then B k = P Ak P 1 = 0.
(c) Write B = P JP 1 , where J is in Jordan Normal Form. Write J = D + N , where D is diagonal
and N is nilpotent (by (a)). Now B = P DP 1 + P N P 1 . The first such matrix is diagonalisable by
construction, and the second is nilpotent by (b).
(d) If is an eigenvalue with corresponding eigenvector v , then Ak v = k v , for all k. As A is
nilpotent, Ak = 0 for some k, meaning that k v = 0. Since v 6= 0, it follows that k = 0 for some
k, and so that = 0.
(e) We know that every matrix has at least one (real or complex) eigenvalue. It follows from (d) that
a nilpotent matrix has zero as an eigenvalue.
(Alternatively, if an n ⇥ n matrix A does not have zero as an eigenvalue, then it represents a trans-
formation from Rn to Rn whose image is the whole of Rn . Thus Ak also represents a transformation
from Rn to Rn whose image is the whole of Rn , for each k. So Ak is never equal to the zero matrix.)
2. Suppose that A is diagonalisable, and let P be the invertible matrix such that P 1 AP is a
diagonal matrix D with 1 , 2 , . . . , n on the diagonal of D. We have assumed that i is non-negative
for all i.
p p p
We can define D0 to be the diagonal matrix with 1, 2, . . . ,
0
n on the diagonal. Letting A be
PD P 0 1 0 0 0 1 0 1 0 0
we have that A A = P D P P D P = P D D P = P DP = A. 1 1
p p
The solution
p is not unique as we could have taken D0 to be the diagonal matrix with 1, 2,
..., n on the diagonal.
What if A is a 2⇥2 matrix that is not diagonalisable? We have to consider the case where A = P JP 1
and ✓ ◆
1
J= ,
0
for 0. In this case, we can confirm that
p 1
!
p
2
K= p
0
has K 2 = J, and we can again set A0 = P KP 1 . There is a catch in that we’d better not have = 0:
for instance the matrix ✓ ◆
0 1
0 0
does not have a square root.
3. (a) Finding the eigenvector in the usual way, we get (10, 6, 3, 1)t .
(b) The interpretation of the eigenvector is that it gives the proportions of the population in each
age-range. So here the proportions should be (roughly) 1/2, 3/10, 3/20, 1/20 in age ranges 0-10,
10-20, 20-30 and 30-40 respectively. As there are 12,000 individuals in all in 2010, roughly 6,000 of
them are aged 0-10.
(c) The interpretation of the eigenvalue is that it gives the increase in the population over (here) a
10-year period. So there will be (approximately) 12, 000 = 18, 000 individuals in 2020. Going the
other way, there were (approximately) 12, 000/ = 8, 000 individuals in 2000.
p
4. Suppose we are willing to trust the question-setter that 15 ( 3 + 2i) is an eigenvalue of matrix
L. Then p we know that, because L is a real matrix, another eigenvalue is the complex conjugate
1
5 ( 3 2i). How can we get the third eigenvalue without evaluating and factorising |L xI|? The
easiest way is to focus on the trace of L: we know that this is the sum of the eigenvalues. The trace
of L is 0, and the two eigenvalues we’ve found so far sum to 6/5, so the third one must be 6/5.
The dominant eigenvalue is the one of largest modulus, and this is 6/5.
If the eigenvector corresponding to eigenvalue 6/5 is (x, y, z)t , then we have to solve the equations
3 33 6 2 6 3 6
y + z = x, x= y and y = z.
2 25 5 3 5 5 5
We can write down a solution to the second and third of these equations instantly: x = 1, y =
10/18 = 5/9, z = y/2 = 5/18. We can take the integer eigenvector (x, y, z)t = (18, 10, 5)t for
convenience. Then it would be good practice to check that it does solve the first equation.
For the final part: the population grows by a factor of approximately 6/5 every ten years, and the
proportions of the various age-groups in the population should approach 18/33, 10/33, 5/33.
• Let v and v 0 be any vectors in S ? , and let u be any vector in S, so that hu, v i = 0 and
hu, v 0 i = 0. This means that
• Let v be any vector in S ? and let u be any vector in S, so that hu, v i = 0. If ↵ is any scalar,
we then have
hu, ↵v i = ↵hu, v i = ↵0 = 0,
so ↵v 2 S ? , i.e. S ? is closed under scalar multiplication.
To show that S ✓ (S ? )? , note that every vector u 2 S satisfies hv , ui = hu, v i = 0 for all v 2 S ? .
This is exactly what it means to say that u is in (S ? )? .
Note: if S is not a subspace, then S cannot be equal to (S ? )? , which is. In the finite-dimensional
case, if S is a subspace, then S = (S ? )? : to prove this we can use a dimension argument.
To say that v 2 D? means that v is orthogonal to every vector in D. As C ✓ D, this certainly
implies that v is orthogonal to every vector in C, i.e. v 2 C ? .
6. The range (column space) of the matrix A is a subset of R2 that does not consist of the
multiples of any single column, so it has dimension at least 2, and therefore is the whole of R2 . The
nullity of A is equal to 1 by the rank-nullity theorem. You can check that (3, 1, 1)t is in the null
space, and so the null space is the set of multiples of this vector.
The range R(At ) is equal to U = Lin({(3, 1, 10)t , (1, 2, 5)t }). The work to check that this is the
orthogonal complement of V = Lin({(3, 1, 1)t }) is the same as that needed to check that V is the
null space of A.
The null space of At is {0} (for instance, by the rank-nullity theorem). This is indeed the orthogonal
complement of R(A) = R2 : the only vector orthogonal to every vector in R2 is the zero vector.
Note that it’s not enough just to say that U = R(A) is spanned by (3, 1)t and (1, 2)t , and to note that
these vectors are both orthogonal to 0. The point is that every subspace U satisfies that condition,
but not every subspace is the orthogonal complement of {0}! Indeed, {0}? is the set of all vectors
orthogonal to 0, which is the entire space (here R2 ), so we need to observe that the columns of A do
indeed span R2 .
MA212 Further Mathematical Methods
Lecture 17: Orthogonal projections continued
Dr James Ward
If
" E PEHE HE all fora EV
Pdf Hu pull
E-
-
-
-
•→
E
HYPE
U
kv Pv k kv uk,
for any u 2 U.
kv uk2 = kv Pv k2 + kx k2
kv uk2 kv Pv k2
ky Py k ky Ax k,
for all Ax 2 R(A). That is, Ax ⇤ 2 R(A) minimises the sum of the
squares of the errors if
Ax ⇤ = Py .
Ax ⇤ = A(At A) 1 t
Ay =) x ⇤ = (At A) 1 t
A y,
y " = Ax is x ⇤ = (At A) 1 t
A y.
t 0 3 5 8 10
y 2 5 6 9 11 error in here
d
find the least squares fit for a curve of the form y = mt + c.
l÷H÷÷l¥ :p:
Z
:* :*
with
1- AI : : :
- ,
( 5)
i
- - -
198 2b
i
so
* Lata
'
-
Tia giggb )
soy .
-
* i.
I
¥435,5711
A'
=p;D
I
Lecture 17, page 10
Where are the errors?
In the last example, we assumed that the errors were in the y -data
so that we could minimise the sum of the squares of the errors
d
Find the least squares fit for a curve of the form t = ay + b.
¥ ¥)
*'
. x .
-
-
-
A 1
Tt (Iffy ) (ab! )
'
CATA 5 A'y
't
x
-
= - -
-
t=¥y 5244T -
y
-
-
244ft 54¥
t
= o.
888T -11982
T
earlier fit !
We can also find fits for curves with more complicated forms...
Example: Find the least squares fit for a curve of the form
errors in here
yt/035
d
y = at 2 + bt + c.
2 5 6 9 11
I AI
-
-
z= Aotbotcl
µ } s÷:p ! )
2
5=99-163-19
c -
P1 -
-
AI 6=9251-65-101
X
aca.im#y..axs=abatbs-ia
, . -
w -
11=9100 1- blotch
1 A
'
z*= CATA ) Atb
-
- - - -
Dr James Ward
⌅ A 1A = In ,
⌅ AA 1 = In and
⌅ Ax = b has a unique solution x = A 1b for every b 2 Rn .
depending on B
Note: If A has a left inverse, then there are usually many di↵erent
matrices B that will give us many di↵erent left inverses (BA) 1 B.
0 t 1 80 19
1 0 >
< 0 > =
B C ` B C
Find a left inverse of A = @1 1A with N(A ) = Lin @ 1 A .
>
: >
;
1 1 1
"
Al -
-
LBAJ B
NCA )
' -
-
f! ! ) ( to Y )
° '
A- want BA take B- St NLA )
-
2×3 NLB) -
. -
. .
Co ? 7)
(! ! ) (L 9) E
(39 )
then BA -
- -
-
CBA 5
'
-
I
⇐ f)
' 0
Ab 43A) B left
-
=
-
-
is the inverse
,
Ling (I ) )
'
A is not
.
'
NCA ) -
-
unique
,
Lecture 18, page 5
Right inverses
Note: If A has a right inverse, then there are usually many di↵erent
matrices B that will give us many di↵erent right inverses B(AB) 1 .
A` Ax = A` b =) In x = A` b =) x = A` b.
A(Ar b) = Im b = b.
For these and all other exercises on this course, you must show all your working.
1. An n ⇥ n matrix is called idempotent if A2 = A. Show that 0 and 1 are the only possible
eigenvalues of an idempotent matrix.
2.
✓ Let U = Lin({(1, 2, 1)t , (1, 0, 0)t }) and V = Lin({(0, 1, 1)t }) be subspaces of R3 .
Find the matrix A representing the projection onto U parallel to V .
Also find the matrix B representing the projection onto V parallel to U .
Evaluate A + B, and explain the answer you get.
:
3. Let v = (v1 , . . . , vn )t be any vector in Rn with ||v || = 1. Define the n ⇥ n matrix A = vv t
where v t is represented as a horizontal 1 ⇥ n matrix and v is represented as a vertical n ⇥ 1 matrix.
Show that A is symmetric, idempotent and of rank one. Hence show that A represents the orthogonal
projection onto Lin({v }).
With A as above, set B = I 2A. Show that B is symmetric, orthogonal and that B 2 = I. Is B
orientation-preserving or orientation-reversing? What transformation does it represent?
Hint: what are the eigenvalues of A, counted according to multiplicity? How do the eigenvalues and
eigenvectors of B correspond to those of A?
4. Let S be the subspace of R3 spanned by (1, 2, 3)t and (1, 1, 1)t . Find the matrix A representing
the orthogonal projection onto S.
5. Find an orthogonal and a non-orthogonal projection in R3 , each of rank two, that sends the
vectors (1, 1, 0)t and (0, 2, 1)t to themselves.
:
t 0 ⇡/3 2⇡/3 ⇡ 4⇡/3 5⇡/3
T 2 10 18 24 22 14
Find the best approximation to this data of the form T = f (t) = a + b cos t + c sin t.
Evaluate your function f (t) at the values t = 0, ⇡/3, . . . , 5⇡/3.
10. Consider the data points ( 3, 3), (0, 1) and (1, 4) where the first coordinate is the value for x
and second the value for y.
Find the least squares approximation to this data using only functions of the form y = ax + b and
check how closely it approximates the data.
What is the least squares approximation to this data using functions of the form y = ax2 + bx + c,
and how closely does it approximate the data?
i . A's eigenvalue satisfies : AE -
-
de and also AE =
die since AZ A -
-
so AE =
ME
{ ( { ) (fo ) f ! ) }
2 . A- is an ordered basis
of Rs = U ⑦ V
,
,
'
µ to ! I g ! ! ) (g ! ! / I ÷ ! )
we can find Mai using the 3×6 matrix '
'
, ,
) f ! ! ! ) Ko ! ! ) ( III)
Armani
'
'
f
'
: : Y
'
so ma and thus ma .
-
""
i :* : it
( Oo )
I O O
=
I
2
, need to check !
, ,
( ÷ g)9
(g! !)
' ' '
then Mps and BE =
'
µ { to / g ! ! ) ( g ! ! / ! ! ! )
we can find ME using the 3×6 matrix ' '
,
f! I y ) f ? ÷ g ) ( g ! ! ) (! I ÷ )
"
so most -
-
and thus BT -
-
MB Bi mis -
-
' '
: : n: i. il
'
( )
O O O
=
o -
I 2
O -
I 2
fg ? I I f ) fo )
From
: ::
I3
above A it Bt
? 9
-
=
t
- - -
-
.
O I 2
, l
-
O O
rank l :
as I =
(Vi .
. . .
.
An It .
It =
(Vi ,
- - -
,
Un ) ,
i÷÷ ÷:÷ :)
"
y
"
:: : : :
. on
* me .
. -
Un :
'
Uno, Unit In
.
. .
.
: : :÷÷ )
( C
I:÷÷÷ ) .
. * nasrani .
RCA) = { at a -
-
AE for all OE Rn }
= Lin, f aij ,
is i. sent this does not make sense
=
Linda) )
2612 6162 I, In
÷:
B -
-
I -
2A = I - - - - -
-
÷:
up
" .
. . .
②
symmetric t idempotent orthogonal projection
so RCA) = Lin 463 )
symmetric :
÷)
Uno ,
-265
f-
I
:÷
BI
-
I
B
-
= -
me
.
. .
.
-
hen -
Usan
.
orthogonal :
B Bt =
B- = CI 2A 5
-
= IZ -14A' -
4A =
I -
41A -
A) -
-
I since A is idempotent
-
orientation :
since A is idempotent .
o and I are the only eigenvalues proved in Qi
A with eigenvalue e
for any eigenvector of
AI =D
LI 2A ) I
BE = =
I so B has eigenvalue 1
-
for eigenvalue I of A . AE I -
-
1131=-1 B is orientation
so
reversing
-
what transformation B
represents ?
if He line { u } ) ,
then BI = -
f! ! ) (; I, )
' 2
let P =
,
then P =
# Pi Fi
Ptp
C ; ;)
( g ! ) fi ; g) fo ;)
-
-
" -
-
-
-
, ,
,
(Ig Ig Ia )
*
MPTP " M #
⇐ %) ( i I 7)
f} !)
so a
-
-
- - =
-
,
f! I ) (f I I, )
5 . orthogonal projection :
A -
-
,
At -
f! ! ) ⇐ 3.) ( 3 :)
"
Ata
fo I 7) *As
-
- - -
-
E
( Iz Iz ÷ )
to
÷) 3) ( I 9 )
then '
At
'
choose B
( f Oo ? ) St NCB) W
(f )
-
-
- .
-
-
-
(÷ ! !)
"
since BA -
-
I ,
then P =
A CBA ) B -
-
AB -
-
§ ÷ ) (Ffg )
*I
ftp./=(bI )
" "
then a* = Att* = A' * 5 # T
'
-
-
HA ' = =
✓
and hence T -
-
ft ) =
15
-
to cost -
IB sing ✓
f- I F) =
15 t 5 -
4353 x-P =
18 V
f la) =
15 -110 -
Fps x 0=25 V
f LIK) =
15 -15 t 4353×-13 = 22 ✓
f- ( 5574=15 -
5 1- 3453×-7 =
12 V
1- I
GMT
suppose A transfer x into y so I AI with 5 = to
. -
-
, ,
3 =
tmz t C
2 =
tmz tC
bnf-tc.us l =
=L! ! )
(5,9 f ) ¥6 E)
4
µ !) (
'
( I ;)
6 3 '
from
-
iz
, ,
) CI )
3
and ate
C i :) -
-
-
, ,
) ( 4,4 ) ( 4) ( II )
""
so * -
CA hi
- '
ATI
-
-
to = =
33 11/28
therefore y
-
-
Tx
t
#✓
It is wrong to
suppose this was equivalent to Xy
-
-
cat 6M is because y -
-
tf t c contains
T I
CATA s I
-
take A -
(l l - -
-
l ) , so =
Cata's
't
A ca . .
ai . -
i. ant =
's .IE ai
,
s .
ca ) at X -
o , y
-
- b then the sum of squares of vertical distance
b't
- ' 2
Cl A b) 17 4A b)
X I y Atb t 14 za b) t
- = - -
- - -
-
=
-
x
x -
-
2 ,
y = Zatb
71=4.9--4 Atb
4g ) µ ) !
y -
-
mate , we set A- and b -
¥ ( 14 ) ¥
' ( 64×-7
*
CATA ) Atb y )
-
2-
-
-
= =
Cb )
interchange the role
of x and y ,
A- nytd
µ !)
see a- and "
f)
µ !)
to ca )
.
as given ,
I =
I
-
A
a'
: i) ( 3 ni )
fi f! ! )
cattail
f: ;)
*A -
-
-
- -
Atg -
-
f
-3
,
o
, t)
(I )
=
( 5g )
Tiko ) ( I! )
'
( 3 Yo ) (YY
'
Eas a' y
't -
so n
-
- - -
=
%
-
hence I =
If It 535
g, =
six c- 3) + If = IT
Yz = # xot If = ¥5
93
=
€ x It
=
¥
÷÷Ht÷i÷d
error . . ax
.
µ § !)
Cb) given 3 fa 3b then I
-
as C
t
I
-
=
-
-
, .
I =
C
-
4- atbtc
-
ATA =
-
-
-
y
-
-
t ( 11×425 Xt 12)
in similar
a manner , E I AI
-
-
-
= - -
i
MA212: Further Mathematical Methods (Linear Algebra) 2020–21
Solutions to Exercises 19: Projections and least squares
This document contains answers to the exercises. There may well be some errors; please do let me
know if you spot any.
2. There are two ways to do this. One way involves using a change of basis where the basis
consists of the three given vectors. We’ll show how to do this the other way.
We need a matrix C whose range is U (which is easy to write down) and a matrix D whose null
space is V : for this we need two linearly independent vectors orthogonal to (0, 1, 1)t such as (1, 0, 0)t
and (0, 1, 1)t . We write
0 1
1 1 ✓ ◆
@ A 1 0 0
C= 2 0 and D = .
0 1 1
1 0
Now the matrix we want is A = C(DC) 1 D, which we calculate to be
0 1
1 0 0
A = @0 2 2A .
0 1 1
This matrix A is supposed to be idempotent, have U as its range and (0, 1, 1)t is supposed to be in
its null space: all are easy to check.
0 1
0 0 0
Repeating the process, we find that B = @0 1 2A.
0 1 2
We see that A + B = I. Why is this? Given a vector w 2 R3 , it can be written uniquely as u + v
where u is in U and v is in V . Now, for all w , Aw = u and Bw = v , so
(A + B)w = Aw + Bw = u + v = w ,
3. The (i, j) entry of the matrix A will be vi vj ; therefore the matrix A is symmetric. Alternatively,
observe that (vv t )t = v tt v t = vv t .
P P
Calculating the (i, j) entry of A2 we get k ai,k ak,j = k vk2 vi vj = vi vj where the last equality arises
since kv k = 1). Therefore the matrix A is idempotent. Alternatively, observe that v t v = kv k2 = 1
and so vv t vv t = vv t .
As A is symmetric and idempotent, it is an orthogonal projection matrix with a range equal to
the column space of A, which is Lin({v }). The columns of A are all multiples of v , and they are
not all zero, so A has rank one, and the eigenvector with eigenvalue 1 is v . Let’s note here that
the eigenvalues of A, counted according to multiplicity, are 0, 0, . . . , 0, 1. The null space of A is the
orthogonal complement of R(A) (of dimension n 1): any non-zero vector orthogonal to v is an
eigenvector with eigenvalue 0.
As A is symmetric, B = I 2A must also be symmetric. We also have
This gives us
0 1
✓ ◆ 1 1 ✓ ◆ ✓ ◆
t 1 2 3 @ A 14 0 t 1 1/14 0
AA= 2 1 = =) (A A) = ,
1 1 1 0 3 0 1/3
3 1
and so
0 1 0 1
1 1 ✓ ◆✓ ◆ 17 20 5
1 @2 3 0 1 2 3 1 @
P = A(At A) 1 At = 1A = 20 26 4 A.
42 0 14 1 1 1 42
3 1 5 4 41
5. For the orthogonal projection, we have a formula: let A be the matrix whose columns are the
two vectors, and evaluate P = A(At A) 1 At .
Let’s just pause a moment to contemplate this formula. We see that P A = A, so the two vectors in
the question are mapped to themselves. Moreover, any vector in N (At ) is also in N (P ), but N (At )
= R(A)? . So this transformation does the right thing to all the vectors in R(A) and R(A)? , so it
must be the transformation we are looking for. If this helps you to remember, or reconstruct, the
formula, that would be good.
Anyway, with this matrix A, we have
0 1
✓ ◆ 1 0 ✓ ◆ ✓ ◆
1 1 0 @ 2 2 1 5 2
At A = 1 2 A= and (At A) 1
= ;
0 2 1 2 5 6 2 2
0 1
You should expect P to be symmetric (and at least notice if you get an answer that’s not symmetric),
and you should check that P does indeed take the two given vectors to themselves.
To find a non-orthogonal projection, there are a variety of approaches. One can take pretty much
any 2 ⇥ 3 matrix B such that BA is invertible, and work out P = A(BA) 1 B. By much the same
argument as before, this is the projection onto the right range, parallel to the null space of B.
Almost all matrices will work, but there’s still scope for choosing one that cuts down on the amount
of work you’ll have to do. For instance, a really good matrix to choose for B is
✓ ◆
1 0 0
B= ,
0 0 1
Again, you should check that this matrix P does the job it’s supposed to.
6. We first need to form the matrix A, whose columns are the values of the three functions 1, cos t
and sin t at the various values of t. So
0 1
1 1 p0
B1 1/2 C
B p3/2 C
B1 1/2 3/2 C
A=BB1
C.
C
B 1 p 0 C
@1 1/2 3/2A
p
1 1/2 3/2
Now we find the vector of coefficients (a, b, c)t by plugging in the vector y of recorded data to the
equation (a, b, c)t = (At A) 1 At y . We find:
0 1 0 1
6 0 0 1/6 0 0
At A = @0 3 0A and (AT A) 1 = @ 0 1/3 0 A ;
0 0 3 0 0 1/3
f (0) = 5, f (⇡/3) = 8, f (2⇡/3) = 18, f (⇡) = 25, f (4⇡/3) = 22, f (5⇡/3) = 12.
These numbers are respectably close to the experimental data given: if you don’t get numbers that
are close to the data, then you should realise that you have made a mistake, and — at least — say
so.
1 1 1
and evaluate (m, c)t = (At A) 1 At yto get m = 22/28 and c = 11/28, so that the best fit of the given
form is ✓ ◆
6 22 11 33 11
y= + = + .
x 28 28 7x 28
What you’re doing here is minimising the sum of the values of (y 6m/x c)2 , for the four given data
points, over pairs (m, c). The alternative suggested amounts to minimising the sum of the values of
(xy 6m cx)2 . That’s not the same problem — the second formulation amounts to giving more
weight to the data with higher values of x.
8. Here we are trying to find the best fit function to the dataset (x, a1 ), (x, a2 ), . . . , (x, an ). We
can only be expected to fit a value at x, i.e. we are looking for the number a minimising the sum of
the squares of the terms (a ai )2 .
Our method applies. We take A = (1, 1, . . . , 1)t , so that (At A) 1 = 1/n, and
n
1X
(At A) 1
At (a1 , . . . , an )t = ai ,
n
i=1
9. For (a), we have the standard least-squares problem. If the points are values of (x, y), we need
to fit a line of the form y = mx + c. We set
0 1 0 1
0 1 0
B1 1C B1C
A=B C
@2 1A and b = @4A ,
B C
4 1 7
✓ ◆
m
so we want to find the value z ⇤ = such that Az ⇤ is closest to b.
c
As usual, we have z ⇤ = (At A) 1 At b so we find that
✓ ◆ ✓ ◆ ✓ ◆ ✓ ◆
t 21 7 t 1 1 4 7 t 37 ⇤ 1 64
AA= ; (A A) = ; Ab= ; z = .
7 4 35 7 21 12 35 7
1
That is, the line we are asked for is y = 35 (64x 7).
For (b), we need to interchange the roles of x and y so that we are looking for a line of the form
x = ny + d. We set 0 1 0 1
0 1 0
B1 1C B1C
A=B @4
C and b=B C
@2A ,
1A
7 1 4
and we want (n, d)t = z ⇤ = (At A) 1 At b. We find that
✓ ◆ ✓ ◆ ✓ ◆ ✓ ◆
t 66 12 t 1 1 2 6 t 37 1
⇤ 32
AA= ; (A A) = ; Ab= ; z = .
12 4 60 6 33 7 60 9
1 1
That is, the required line is now x = 60 (32y + 9), or y = 32 (60x 9).
10. This is the standard least-squares problem. The points are values of (x, y), and we need to fit
a line of the form y = ax + b. So we set
0 1 0 1
3 1 3
A= @ 0 1 A and c = 1A ,
@
1 1 4
and we want to find the vector z ⇤ = (a, b)t such that Az ⇤ is closest to c.
As usual, we set z ⇤ = (At A) 1 At c and find that
✓ ◆ ✓ ◆ ✓ ◆ ✓ ◆
t 10 2 t 1 1 3 2 t 5 ⇤ 1 1
AA= ; (A A) = ; Ac= ; z = .
2 3 26 2 10 8 26 70
1
That is, the line we are asked for is y = 26 (x + 70).
This is not a wonderfully accurate approximation: fundamentally, the three points we are given are
not all close to any straight line. This line is nevertheless the “best”, in the sense that it minimises
the sum of the squares of the vertical distances from the points to the line.
We now allow ourselves to use quadratic functions. Let’s see how the algebra works out. We set
0 1 0 1
9 3 1 3
@
A= 0 0 1 A and d = 1A ,
@
1 1 1 4
so we want to find the vector z ⇤ = (a, b, c)t such that Az ⇤ is closest to d . We know to take
z ⇤ = (At A) 1 At d , but here At and A are invertible, so this is just z ⇤ = A 1 (At ) 1 At d = A 1 d .
Thus we have 0 1 0 1
1 4 3 11
1 @ 1 1
A 1= 8 9A =) z ⇤ = A 1 d = @25A .
12 12
0 12 0 12
The “best fit” is then
1
(11x2 + 25x + 12).
y=
12
You can check that this does indeed go through all three points.
What did we do, e↵ectively? Instead of solving the problem of finding a vector z ⇤ such that Az ⇤ is
closest to d , we solved the (easier?) problem of finding a vector z ⇤ such that Az ⇤ is equal to d . In
other words, we solved the matrix equation Az ⇤ = d . Our method amounted to setting z ⇤ = A 1 d .
In general, we can fit a quadratic function through any three points with di↵erent values of x. In
even more generality, if we are given any k points in the plane with di↵erent x-values, then we can
find a polynomial of degree k 1 passing through all the points.
To see this, let the points be (x1 , y1 ), (x2 , y2 ), . . . , (xk , yk ), and suppose we are looking for a polynomial
y = a0 + a1 x + a2 x2 + · · · + ak 1 xk 1 ; we are then looking to solve:
0 1
0 2 k 11 a0 0 1
1 x1 x1 · · · x1 B C y1
B1 x2 x2 · · · xk 1 C B 1 C By2 C a
B 2 2 CB a C B C
2 C = B . C.
B .. .. .. . . .. C B
@. . . . . A B . C @ .. A
k 1
@ .. A
1 xk x2k · · · xk yk
ak 1
What we need to verify is that the matrix above is non-singular, for any set of distinct values
x1 , . . . , xk . There are various ways to approach that: here’s a rather abstract treatment. If you think
of expanding out the determinant, each term will contain one entry from each column, so each term
will be a “monomial” of the form xi1 x2i2 · · · xkik 1 , of degree 1 + · · · + (k 1) = k2 . So the whole
expression for the determinant is a polynomial of degree k2 . Now, if we have any pair of terms
equal, say xi = xj , then two of the rows of the matrix will be zero, so the determinant will be zero.
Therefore each of the terms (xi xj ) is a factor of the determinant. There are k2 such terms, so
the determinant is a constant times the product of all of them. Hence the determinant is non-zero if
all the terms (xi xj ) are non-zero. (Don’t worry if you didn’t get that.)
MA212 Further Mathematical Methods
Lecture 19: Generalised inverses
Dr James Ward
B C
Find a WGI of the matrix A = @0 1 1 0 6 A. A'BC
1 1 1 1 10
)
B 3×2
II. ⇐
ran"
I. E can
.
* rank 2
. .
on
.
' '
B P "
-
Co : : )
9:14! ) Coil
a-
-
-
(÷)
" o
.
""
Is iii.
cnn.yio.i.io.r.co ; ,
(! ! ! )
l
l O O
l
l
=
O l O ,
i I check AASA -
A
Lecture 19, page 5
Example
0 1
1 1 2 4
B 1 2 1 1C
B C
Find a WGI of the matrix A = B C.
@ 3 0 5 9A
( : :/
/ O 513 3 1 4 3 7
RRECA)=
:
"'
O O
o
O O
soecai-r.a-B.ca
4×2
-
2×4
0=49%33 ) ,
ranks "
cha QLCQ)
2g )
"
" Bl CPB)
-
'
'
-
p
a 4×2 rankle
ranks
g) % Yo)
p #
area . "
taker -
Co : : :) -
-
-
-
"
;①d②
-
-
- -
-
BEC ! !: :/
'
①
. .
as .ua .
-
Yg ! ! ! )
Lecture 19, page 6
Projections and WGIs
1
B ` = (PB) P and C r = Q(CQ) 1
Ag = C r B ` = Q(CQ) 1
(PB) 1
P.
Amongst all the possible WGIs for a matrix, one is very special...
In fact, even though a matrix can have many WGIs, it turns out
that only one of these will be an SGI.
B ` = (B t B) 1
Bt and C r = C t (CC t ) 1
,
As = C r B ` = C t (CC t ) 1
(B t B) 1
Bt.
and
As A = (C r B ` )(BC ) = C r C = C t (CC t ) 1
C.
It’s relatively easy to show that these two matrices are symmetric.
To show that the SGI is unique, see Exercises 20.
1 0 2 A 3×3
B C
Find the SGI of the matrix A = @ 1 1 0A. '
-
BC rank 2
0 1 2
B
tank ,
B'
Y) CI
T A
CI 2142£
T
"
E
fi } )
E
B -
C 2×3
rank
C -
-
to :3 ) or
*
,
a
Cz CI
duct) ( Bt BI ' Bt
( j f f)
As '
-
'
-
c
.
-
-
( y 3oz)
I 3
z
-
I
-
=
9-
- -
=
-
Ase .
.
.
-
-
same
y " = Ax is x = (At A) 1 t
A y,
y " = Ax is x = As y .
Note that, if we have to use this, this will be one of many possible
least squares solutions...
x = As y + (In As A)w ,
t.me#as---if:. 's .
'
'÷i⇐! ! ) III
a '
'
"
'
±¥µ ,
=
if, } ! ) ! !) "
¥ !)
Asa -
a. wer
Y
I
Is -
Asa -
-
E-
(÷ ) +
htgb (I ) ,
,
a. b. CER
"" "
"
H vine fit
"" " " " "
"
-
t
(4) =
ASI
Lecture 19, page 19
MA212 Further Mathematical Methods
Lecture 20: Orthogonal projections in function spaces
Dr James Ward
P(v ) = hv , u1 i u1 + hv , u2 i u2 + · · · + hv , un i un ,
v = P(v ) + [v P(v )]
= hv , ui i hv , ui i = 0
using the orthonormality of the ui .
Thus, every v 2 V can be written as
v = P(v ) + [v P(v )]
with P(v ) 2 U and v P(v ) 2 U ? so that we have V = U + U ? .
Also, if u 2 U \ U ? , then hu, ui = 0 which implies that u = 0.
That is, we also have U \ U ? = {0}.
Consequently, V = U U ? and so P(v ) is indeed the orthogonal
projection of v 2 V onto U. Lecture 20, page 5
Example
Consider the vector space C[0, 1] with the inner product
Z 1
hf , g i = f (x)g (x) dx.
0
To rx d x
5- [x-P] -5
'
<
f. his =
=
o
=
B ( 2¥
±
< f. hz > =
Sf Tx B C 2x -
1) dx =
f! -
X ) DX =
TIB
< f. has > =
If Fx 55 lb X
'
-
Gx ti ) dx -
-
S! if lb XE 69¥ txt ) dx
-
= -
Tosof
Lecture 20, page 6
PIE ) f.he > kit at hz he CI 13713
= < .
> t .
=
I .
I t TI B .
B (2x D t C
-
-
Fas ft ) if 16×2-6×-11 )
= -
4x4 Efx + Is exr
-
t
over all a, b, c 2 R.
In this sense, it is a least squares approximation to f .
eikt
hk (t) = p ,
2⇡
can be used to form the orthonormal set
{h n , . . . , h 2 , h 1 , h0 , h 1 , h 2 , . . . hn }
I
< I I>,
-
II fth Studt ,
her Ct) =
22T
if ntm .
cnn.hn > =
ce÷ tis ,
=
eI e÷
II de
¥ (eimn-mmt-g.ca
m"
¥17, ein
-
so isi no
zig ¥ 2isinln-mlk.co
-
= =
n m -
M n-
eio.e-io-zisinoe.int int
Ia Ia
e
Ka
,
if mm cnn.hn> de Idt I
⇐ >
.
.
= =
, =
÷ one:* an :i÷
.
II. ( ¥ IIfit e-
""
=
) de ) eike
n -
Ak Cf )
n
Kt
'
'
=
Ak Cf ) e
n
if Kao, Ao -
-
Ia Ia Itt at = ¥ (Ia -
t dt tf} edt )
Eal I Ia ⇐I ? I E
'
-
-
E
-
t
-
-
i"
¥17, It I e ¥ ( Ia
i" "
de )
""
if Jo t
-
Ho
- -
, ak -
-
dy = -
te de + e
""
fate
""
¥I fo't e
-
de dt )
-
-
=
t
te t.e.IT?q+c=te:iY*e-
-
ikt
Y÷ I
"'
Ste
-
mm de -
-
- de -
-
to
need k¥0
"'' ik"
fi ¥ ( Eike , e
ike
)
}
-
¥1
-
(e e )
-
=
- - +
-
-
k
Zisihlk 'd) for odd 20054587=21 1)
-
-
O k
=
Ia I + 2t I = t -
ith )
"
)
o k even
E
{÷
and
so
for k¥0 ao -
A"
-
-
-
.
,
K odd
, , , ,
""
Et En
N
⇐ eikt
2
e-
-
I
¥2
-
let
"
K -7 -
K
"
= t
k
-
-
I KH
Lk odd ) ( K odd)
[ e' ,
-
=
Is + -
+ e
K I
-
20056kt)
-
CK odd)
)
so Pnc the - I cosckt
t day kik which is real
Ck odd)
K2
k
Ckodd)
For ⇡ t ⇡, the first, third and fifth-order Fourier series (solid
curve) of f (t) = |t| (dashed curve) look like this...
P1 (t) #
=
-
cost P3 (t) =p th 4¥
,
- -
005431 P5 (t) -
-
Bce )
-
IE 25
Pnc f) it) -
-
E -
¥ I ⇐ TE En % -
,
) -
-
fit)
K= I
Ck odd )
is cos Ckt)
E ka -2
-
1<=1
To =
HI
(K odd)
Sef t -
- o
,
I -
TF I ¥ = o
K= I
Ck odd )
¥ ft
A
so I =
It j t ¥ Tt t t . . -
=
f- ¥
.
1<=1
(K odd)
For these and all other exercises on this course, you must show all your working.
✓
1. Write down the equations which u, v, w, x, y and z must satisfy in order for the matrix
0 1
u v
@w x A
y z
✓
2. Express the matrix 0 1
1 1 0 1
@
A= 0 1 1 1A
1 1 0 1
in the form A = BC where B and C are both of rank 2 and of sizes 3 ⇥ 2 and 2 ⇥ 4 respectively.
Hence find the strong generalised inverse of A.
Deduce the matrix which represents the orthogonal projection of R3 onto the column space of A.
0 1
1 0 0 1
B 1 1 0 0C
3. Find the strong generalised inverse of A = B
@0
C.
1 1 0A
0 0 1 1
✓ ◆
1 x
4. Let A be the singular 2 ⇥ 2 matrix where x and y are real numbers, not both zero.
y xy
Find a formula for the strong generalised inverse of A.
5. Let A be a real matrix, with strong generalised inverse As . Show that (As )t = (At )s , i.e. that
the transpose of As is the strong generalised inverse of At .
Deduce that, if A is symmetric, then so is As .
6. Assume that there are only two values for x, namely 0 and 1, but there are four pairs of data,
namely (0, 3), (0, 2), (1, 5) and (1, 3) where the first coordinate is the independent variable x and the
second the dependent variable y.
Using the strong generalised inverse, find the least squares approximation function to the data in
Lin({1, x, x2 }).
✓
7. Let P be the space of all polynomial functions. Define the linear transformation D : P ! P by
D(f ) = f 0 , i.e. each function f in P is mapped to its derivative.
(a) Does D have an inverse?
(b) Find a right inverse of D, i.e. a linear transformation D⇤ : P ! P such that DD⇤ is the
identity.
:
inner product given by Z 1
hf, gi = f (x)g(x) dx.
1
9. Find the best approximation of the function t in the interval ( ⇡, ⇡) by a linear combination
of the functions eikt for k in some range n k n.
Hence write the function t as a sum of functions of the form ck sin(kt) and dk cos(kt), for k > 0.
Evaluate both sides of your final expression at t = ⇡/2 to get an expression for ⇡.
10. Find the best approximation of the function t2 in the interval ( ⇡, ⇡) by a linear combination
of the functions eikt for k in some range n k n and rewrite this approximation as a real-valued
function.
X1
( 1)k+1 ⇡2
By letting n tend to 1 and evaluating at t = 0, show that = .
k2 12
k=1
I
①
(Wy Az )
u a
"
1- For to be a
right inverse of A , AA -
-
Iz
would wty
l÷÷÷÷÷i÷÷:
a I
So require
- -
we
-
( 3×1
3W
② any matrix in the form of is right inverse of A where x. WEIR
W X
-2W -
2X X
(f (f )
0
③ since
Yz )
KAI
( f I f)
RRE LAD -
-
=
-
I =
.
-
-
2+3 -
n
, ,
①
2 . Observe that Cz
-
-
E -
Ci . Et =
E -
2cL
( b f i T)
'
µ !) wand
then we can take B-
- c-
✓
(¥ !)
② E-
Bt
f t)
o
and thus -
-
, ,
its :L
ki it
f) 3)
at -
i' = eat .
-
±
* a-
( : : : ) (y !) ⇐ 3) * Bi
⇐ E)
' -
-
v. ¥¥¥¥
÷÷÷÷H÷÷,
Cicco ) ' CBTB ) Bt
-
'
-
hence As =
l O l
* income
b
l
⇐
{ (§ !))
'
③ RCA) -
- Lin
, according to theorem ,
AAS
orthogonally projects R onto RCA)
÷÷÷f÷÷tx4
I
:L
I 0
Ii : :
*"
: :
IT
¥¥iE¥
3 .
① Ca) the of D ff DX To fttldt
't
inverse #
'
is D =
)
"
Cb , let fi X. ni X be the basis of Pn
- - -
. . .
Then DC D= # Ci ) - O so D
*
Cl ) =
fl da
DIX) - ddxcx) = I 't
D Cx) =fxdx= ¥
Dhi )=ddxlX4= 2x D*Cx4=fPdx=¥
i .
t
I ,
Dam ) =
# ( xn ) -
- NX
""
D ( xn)
't
=
Janda =
# x
""
f! ! ! ÷ : ) f ! ! ;÷÷
" " and ""
asterism inverse
Ca)
for an inverse to exist :
for f Elp , D
't
Dcf) f -
-
*
DD (f) f=
8 .
from Q lo in ex 13 , the basis of < f. s > is hi Cx) - ¥
hzlx)= E- X
hslx) -
-
¥042 -
f)
< f. his -
-
I Fda E [ Thx's ; -Ec¥ E ) o
,
-
-
=
- -
-
,
= -91¥ -1¥ ) ¥ -
<
f. has > =L ,
7131K£ ) To da = Info f'yCx5 3×3) dx - =
Info ⇐ X' -
¥44 ) ! ,
=
# fo tf E It # 7=0 -
-
hence PCI ) =
af ,
his hit af ,
hz > hat < f. h3 > h3
=
IF Xo t F- F- x x t O x -34%(712--4)
= X z x3
f. if k=o ,
Ao -
- ¥172 Edt ¥ CITY % = = ¥ ( Iai Eas ) - =
o ?
eine
cK e
reika
# Tu )
-
= + - -
Tik
"i"
"
"
-
Iac t s
÷ ¥
I
=
l Tk-
zoos ( ka)
-
Zi Sin Clas )
=
Iq IT .
2.cos ( ka ) =
Iq C- 1)
k
hence Pnc f) A) =
Fen c 1)-
Keitel
iz ,¥ Keith
't
=
C 1)
-
Keil + c-1)
, m ,
=
II ,
Tj c -
Dk ceikt -
e
- ikt
)
FE i ,
C -
1) kzisinlkt)
141
n
C-1) k n ←, y
-2-2 is
=
I f-
sin ckt ) =
2
IT
k= I
sin ( KL )
7L
Set ties . Pncfjct) = I -
-
I
•
I
sin CE k)
-2¥ I
-
-
-2 [ EE ,
since Ht II ,
Ti -
H ) -
O ) =
E note when
:
k is even ,
since 14=0
odd
even
so -2 f -
I -
I ly- -
- . .
) = Is I= 14C It I t I t - - -
) x
its It 'T -9 I + tf it -9 TF t
-
.
4
- -
-2C
-
so - -
t . .
.
.
) = ⇐ -
7. (a) d) Ut w =
{ Utne / a c- V. New }
It
'
i. e. U c- U -1W
cii )
by def , V =
Ut Ut
Ut ( Vtwjt
from ci ) VE V -1W then V E V
- -
. ,
SO ⑨ + WH E ut
take Ue Ut n wt and UE V. WE W
then < U ,
Utne > = < U . U> + < Usw > = 0
SO It CU -1W )
1-
hence Utnwt c- @ + wjt
( vtw It =
Uts wt
a) cis vein
{ (g) (f) } w tin
{ (f) }
-
-
µ ;) ( f ? :)
°
take A- ,
then A"
4,0;)
* A" =-3
* A-
Yo ? ;) =L ; ;) ⇐ it
men
P=A*AÉñ=µ ;) +
⇐ 2)
4911=+4 ? / to ? :)
-
-13
t: : :)
= 2 -
l l
µ ;) 1)
8. d) '
(
Az =
Az -
A, so A- I 0 -
I -2
o l l
94 = Az -
29 ,
,
B- c-
di ) want BL be orthogonal projection , how to take B ?
( ? f)
1
take BT =
*'
?
a)
(b) 4) < 9 , It ) ,
821-1 ) > =) ! , 8114921-1) de = It , t d-1 =
1--12/1 ,
= d- -2-1=0
118,112 = < S, ,
S, > = f- I 1. Idt = It] !, = 1- c- 1) = 2 119,11 - E
hi =
÷f =
-1oz , so hit) =
¥
11 fall
-
=
f! ,
ttdt =
3--131 ! , = d- -
(-1-3)=-5 111-211=-38
h2= 1¥11 ¥-3 = = t , so hz It ) =
¥-1
<h , h, > =
# f! ,
et de -
Filet ] ! # ( e- e-1)
,
=
le
'
-
e- 1) =2e -1
Plh ) =
< h, f , > fit < h, fz > fz =
. - -
= d- le - e-1) +3 e-It
icaci )
lying ,
"r%n CHR
= ¥%¥¥=¥%¥ñ=¥
* um
✗ →A
"-
2×2-1
-
¥%¥,=¥% =¥%÷¥÷¥÷=iim¥¥÷×+
=iim¥¥¥ =&
cii ) take log ?
eb
fbacsx -
✗ 2) dllnx ) = I?(5X-ñ)- dx =/ Fcs -
WEN
•→
step function :
↳ Tix] 1h1 ✗ c- [1. 2)
{
=
1h5 •
1h2 [ 2,3 )
1h4
In } ↳ 4)
1h3
1h4 [4. 5)
1h2 -0
•
1h5 71=5
Be-10 1
>
I I
I 2 3 4 5
=
f- (2) 1h2 -1
'
f- ↳ ) -
11h3 -
=
GIN -1 61h3 -61-92-1444 -
41h3
=
21h3 -1444 g- -
-
-
→É
-
- i. -
"
÷.
o
"
¢1
"
=/
*
{
* " ↳ ✗ En -
-
""
#
.
.
.
2 [↳ 3) 2X [2. 3)
?
.
3 Esat) }x [3. 4)
i.
¥÷→
:
i
2) (4-2) 1- fl 3) Cf 6) -
1- f-(4) 116-12) + .
-
.
-1 fin ) b) +
. f? 44dm + JÉ Hd Csx) + - -
.
+ In? dcn -
DX
=
It .
2-1 34 .
3 -1 44 .
4-1 .
- .
+ n4 .
n ] + ⇐ ✗5) ? + [2 -5-+5 ] {
.
+ [ 3. ✗5) § + 1-(4-1)=5×511
. .
.
= 125+35-145-1 . . -
this ] + -5425 -1545 -
+ -5235 -5425
-
+ E- 45+-3535 + . . .
-1 ¥ n5 -15 -
-
= [ 25 -1354 .
.
.
+ n5 ] -
[ 25 + 35 -1 . . .
+ Cn -135 ] + n5 -
¥
= ¥225 -135 -1 . . .
n 5) 1- Jt [ nb -
i ]
=
,
I
cosy = 1- -5
so
Jteyasydy f ! g- dy -
go.yt-jasy-otji.FI?y-=-ey-HRey+-y-----i--c
take
finite .
I'd ¥775T DX also converges
for ② .
converge + converge =
2
-
-
- -
- -
-
± :
g- IT
o I I >
✗
.
3) DX
✗
=/ ¥ I É ,
-
+ E)dat C- É +2×5 sdx
=
(I ✗
3
+ zlnx ]i + f- ¥ + 3- ✗ b) ?
= -
TI 8 -1 21h2 ✗
-
l -
I 1- 21h 1) + c- TF + I ✗ 2° + I -
=
-
5- 1- 21h2 -1 TI -
f- p ¥ + TI d- -
= + 2hr2 ✗ f-
3 .
(a)
4. evalues C- 1-d) [(-2411-34)+1]
1
: let / A -
dI1=0 . then -1 -
X O -1 =
I -2 -
X -
I -
I (1-0)
. I -3 = ⇐ (6+5×+5+1) -1=0
D= -2
:/ 1:11 :/ at :p
IA its / 1=0 0
1:
electors :
-
I 0 -
I ✗
Ui=( f) µ !-1*41=>0--1 :|
°
take ,
then 1A -51-314--4
' °
,
,
f; ; ;) ( if ;)
Home * and "
0 -2 I
41=1 : : :/ 1¥:
rewrite XK I 0 -
I
ZK
1<(-4×-1) ( ? ? ;)
""
"=PJ"P"=µ gig ) (
' "
A 21 K ' -4 °
k O l -
l
0 C- 2)
, 0 k
0 C-2)
(f)
(b) unit vector in the direction of the axis .
Ñ=
we men have E-
(f) ii.
=ɵ )
-
ui.it/:-.Y.YE/=-ii--i--ii=i
iii.
0
Bib
( sing ;) ¢2 )
cos -4¢
* Yrs Yr 0
-
= = -
sin-44
assay o YR °
0 O O -
A7= xnBA%BMÑ =
.
.
.
Check AT is orthogonal
( & f) ( Ff ) ( ) ( )
5. cakis B*B= = 2-41 Zi = 2-41 Zi = I
-
2- i - i -
Zi I
2- = 0
cii ) B*B=BB*
(; ; ;) f. I ;)
• '
" "
' " "
' = - '
'
"
p BDB
-
-
rd with # c- 0£ ¥ #
'
1. cb]
by polar coordinates , ✗ =
rcoso ,
y -
rsino and ✗ 1-y
'
-_
and are
"> "- 42
( 2- pzj
-
y J [2- CX tyyj
-
( 2- ✗
'
= =
-
I
g- ✗
y=É
YYp.is#=Yf-.Yf--If-Yf-=rcos-o+rsirio--r
use Jacobian :
(/
Kosa
1%1
0=46 no ¥5K IN dr ) do
44
=/ o-xa.CH?FI-rFdr)do
let a- 2- M ,
then dull.dr= -
zr
JE f ÷ %-
a- %
.
do
=s÷%I÷÷ .
do
744 Kosa
=
✓
0=>46
[ no
do
=s÷
=
I _u÷du -
at # a
{
)
{
i. - = 0 ✗c- ←
-
X -
O
✗C- ( 0 , I ]
+ ✗e -4,2 ) -
I
-2 ✗= 2 → ✗ ell .2]
1
-
-
I 71=1 I -
•
2 ✗ C- 4,2 )
←É
4 A- 2
decreasing function
partition P={ Xo , ×, ,
-
"
.
An } where 710=1 and Xn=2
mi-inff.fm/7:XEEti-i.ti ) )
LHP )= ¥2 ,
mil titi ) -
✗ Iti 1)
- him Lacp )=2fH -12142)
=
Mk ( 2- d) -1mn (4-2)
=
ZMK -12mn
¥ ; FIX)= X ,
f' 1×1=1
42
(11-1-2×2)
51×1=3×-5,4 , +
dt
=
lH-iI 4t¥× ¥ +1¥ _icHt¥¥ -
I/ ✗
i. , ,
"+¥+¥+SÉ¥÷÷
=
so fire ) = 2 +4T / T2 =
To
b- X
¥
=1im¥¥tÑ
=
di ) Taylor theorem
JI =
+
JÉ + ji + fi
3. (a) d) ↳ fits } =
¥+4T ,
=
¥-5 = L
{ sinizt) ] ✗
tis] =
¥4 fit)= sin at
STS ) =
# get)= sin 2T
by convolution theorem .
Acts -
-
J ! flu) Slt -
u )du= To sinew sin Rt -24dg = ¥ sina.tl Itasca
-
cii ) L {y
"
} -14L { y } =
L{ sina.tl }
say }
'
-
y
'
lo ) -14 LEY } =
¥+2
sty -
sylo) b -
-1 44 =
¥
>
s Y -
as -
b -14T =
2/(5+4)
⇐ + 4) Y = as tbt
¥4
y=%¥4 É +
acosztt-bsinsf-izcoszt-d-sinzt-1-gls.in
y =
2-1-14b)
-
4-10052-111 -4A)