Exam2016 17s1
Exam2016 17s1
0 1
1 0 1 1
B C
B0 k 1 k
Let A = B
C
1. [14 marks] C where k is a constant.
@k k 2 0 A
k 0 k 0
(a) Use Gauss-Jordan Elimination to reduce A to the reduced row-echelon form.
(Write down the elementary row operations clearly.)
(b) Find a basis for the nullspace space of A.
0 1
4 0 2 2
B
B0 4 2 2C
Let B = B
C
3. [18 marks] C.
@0 0 2 2A
0 0 2 2
4. [22 marks] (All vectors in this question are written as column vectors.)
0 1 0 1 0 1 0 1 0 1 0 1
1 0 0 1 0 0
Let u1 = 1 , u2 = 1 , u3 = 1 , and e1 = 0 , e2 = 1 , e3 = 0A.
@ A @ A @ A @ A @ A @
1 1 1 0 0 1
(a) (i) Find the reduced row-echelon form of the following 3 6 matrix
u1 u2 u3 e1 e2 e3 :
(ii) Let S = fu1 ; u2 ; u3 g. You can assume that S is a basis for R3 .
Write down the coordinate vectors (e1 )S , (e2 )S , (e3 )S .
1
(b) De ne a linear transformation T : R3 ! R3 such that
T (x) = c1 u1 + c2 u2 for x = c1 u1 + c2 u2 + c3 u3 2 R 3 :
(i) Find the standard matrix for T .
(ii) Determine the rank and nullity of T .
(iii) Explain why T (x) is the orthogonal projection of x onto V = spanfu1 ; u2 g.
5. [17 marks] (All vectors in this question are written as column vectors.)
Let A be a square matrix of order n.
(a) Let E = fe1 ; e2 ; : : : ; en g be the standard basis for Rn .
Show that Aej = the j th column of A.
(b) Suppose Am = 0 and Am 1 6= 0 for some integer m 2.
(i) Show that there exists at least one vector u 2 Rn such that Am 1 u 6= 0.
(ii) Show that fu; Au; : : : ; Am 1 ug is linearly independent where u is the
vector obtained in Part (i).
(c) Prove that if An+1 = 0, then An = 0.
6. [17 marks] (All vectors in this question are written as column vectors.)
(a) Let B be a 2 2 symmetric matrix and let u, v be two eigenvectors of B
associated with the eigenvalues and respectively.
(i) Show that if 6= , then u v = 0.
! !
p1 p1
(ii) Suppose u = 2 ,v= 2 , = 1 and = 3.
p1 p1
2 2
Find B .
(b) Let C be a symmetric matrix of order n with a characteristic polynomial
( 1 )( 2 ) ( n )
where 1 2 n .
x 2 Rn, 1 xxTCx
T
Prove that for any nonzero vector
x n.
2
Solution:
0 1 0 1 0 1
1 0 1 1
B C R3 kR1 B10 0 1 1
C R3 R2 B01 0 1 1
C
B0 k 1 k k 1 k k 1 k
1. (a) B
@k k 2 0
C
C ! B B
k 2 k k
C
C ! B B
0 1 k
C
C
k 0 k 0
A
R4 kR1 @00 0 0 k
A @0
0 0 0
0
k
A
Case 1: k = 0.
0 1 0 1
1 0 1 1
B
0C
R3 R2 B10 0 0 1
0C
B0 0 1 0 1
B
@0 0 1 0A
C
C ! B B
C
C
0 0 0 0
R1 R2 @00 0
0
0
0
0A
0
Case 2: k = 1.
0 1 0 1 0 1
1 0 1 1
B
1C
R3 $ R4 B10 0 1 1
1C
R1 + R3 B10 0 1 0
0C
B0 1 1 1 1 1 1
B
@0 0 0 0A
C
C ! B B
@0 0 0 1A
C
C ! B B
C
C
0 0 0 1 0 0 0 0
R2 R3 @00 0
0
0
0
1A
0
Case 3: k 6= 0 and k 6= 1.
0 1 1 R2 01 0 1 0 1
1 0 1 1 1 1
B
kC
k
B 1 1C
R1 + R4 B10 0 1 0
1 C
B0 k 1 C 1 1 k R3 B0 1 1 k 0C
B
@0 0 1 k
C
0 A ! @0 0
B k
1 0A
C
C ! B B C
1 R2 R4 @00 0 1 0A
k R4
0 0 0 k 0 0 0 1 0 0 1
0 1
R1 R3 B10 0
1
0
0
0
0C
! B B
C
C
R2 k R3 @00
1 0
0
1
0
0A
1
Alternatively , for this case, since A is invertible, the reduced row-
echelon form is an identity matrix.
(b) Case 1: k = 0.
The general solution for Ax = 0 is
0 1 0 1 0 1
t 0 1
B C B C B C
sC B1C B0C
x=B
B C = sB C + tB C
@0A @0A @0A
for s; t 2 R:
t 0 1
3
Case 2: k = 1.
The general solution for Ax = 0 is
0 1 0 1
t 1
B C B
B tC 1C
x=B B
C = tB
@ t A @ 1A
C
C for s; t 2 R:
0 0
2. (a) Since
(a + b 2c; 2b c; 3c + d; a + 3b + d)
= a(1; 0; 0; 1) + b(1; 2; 0; 3) + c( 2; 1; 3; 0) + d(0; 0; 1; 1);
(b) Method 1:
0 1 0 1
1 0 0 1 1 0 0 1
B Guassian
1 2 0 3C B
B0 2 0 2C
B
B
@ 2 1 3
C
C
0A
! B
@0 0 3 3A
C
C
Elimination
0 0 1 1 0 0 0 0
Method 2:
0 1 0 1
1 1 2 0 1 1 2 0
B Guassian
B0 2 1 0C B
B0 2 1 0C
B
@0 0 3 1A
C
C ! B
@0 0 3 1A
C
C
Elimination
1 3 0 1 0 0 0 0
4
8
>
>
>
a + b 2c = 1
<
2b c = 0
, > 3c + d = 0
>
>
:
a + 3b +d = 1:
0 1 0 1
1 1 2 0 1 1 1 2 0 1
B Guassian
B0 2 1 0 0C B
B0 2 1 0 0C
B
@0 0 3 1 0A
C
C ! B
@0 0 3 1 0A
C
C
Elimination
1 3 0 1 1 0 0 0 0 0
The linear system is consistent, i.e. there exists real numbers a, b, c, d
such that (1 + a + b 2c; 2b c; 3c + d; 1 + a + 3b + d) = (0; 0; 0; 0).
So the zero vector is contained in V .
(ii) Yes.
Remark: Actually, W = V .
4 0 2 2
0 4 2 2
3. (a) det(I B) = 0 0 2 2
0 0 2 2
2 2
= ( 4)2
2 2
= ( 4)3 .
The eigenvalues of B are 0 and 4.
0 1 0 1
4 0 2 2 0
B C B C
0 4 2 2C B0C
For = 0, (I B)x = , C x = B C.
B
0 B
@ 0 0 2 2A @0A
0 0 2 2 0
0 1 0 1
4 0 2 2 0 1 0 0 1 0
B C Guass-Jordan B C
0 4 2 2 0 0 1 0 1 0
B
B
@ 0 0 2 2 0
C
C
A
B
B
@
! 0 0 1 1 0
C
C
A
Elimination
0 0 2 2 0 0 0 0 0 0
The general solution is
0 1 0 1
t 1
B
tC B
1C
x B
=B
@
C B
C = tB
tA @
C for t 2 R:
C
1A
t 1
So f (1; 1; 1; 1)T g is a basis for E0 .
5
0 1 0 1
0 0 2 2 0
B C B C
B0 0 2 2C B0C
For = 4, (I B)x = 0 , B C x = B C.
@0 0 2 2A @0A
0 0 2 2 0
0 1 0 1
0 0 2 2 0 0 0 1 1 0
B C Guass-Jordan B C
0 0 2 2 0 0 0 0 0 0
B
B
@ 0 0 2 2 0
C
C
A
!B
B
@ 0 0 0 0 0
C
C
A
Elimination
0 0 2 2 0 0 0 0 0 0
The general solution is
0 1 0 1 0 1 0 1
r 1 0 0
B C B C B C B C
sC B0C B1C B0C
x=B
B C = rB C + sB C + tB C for r; s; t 2 R:
@tA @0A @0A @1A
t 0 0 1
So f (1; 0; 0; 0)T ; (0; 1; 0; 0)T ; (0; 0; 1; 1)T g is a basis for E4 .
0 1 0 1
1 1 0 0 0 0 0 0
B C B
1 0 1 0C B0 4 0 0C
Let P C and D = B P 1BP = D.
B C
=B C. Then
@ 1 0 0 1A @0 0 4 0A
1 0 0 1 0 0 0 4
0 1
0 0 0 0
B
B0 2 0 0C
(b) For example, let C = PB CP . Then
1 C 2 = B.
C
@0 0 2 0A
0 0 0 2
0 1 0 1
1 0 0 1 0 0 Guass-Jordan 1 0 0 1 0 0
4. (a) (i) @1 1 1 0 1 0A ! B
@0 1 0 1 1
2
1 C
2 A
1 1 1 0 0 1 Elimination 0 0 1 0 1 1
2 2
(ii) (e1 )S = (1; 1; 0), (e2 )S = (0; 12 ; 12 ), (e3 )S = (0; 21 ; 1 ).
2
0 1 0 1 0 1
1 0 1
(b) (i) T (e1 ) = u1 u2 = @1A @1A = @0A.
1 1 0
0 1 0 1
0 0
T (e2 ) = 0u1 + 2 u2 = 2 @1A = @ 21 C
1 1 B
A.
1 1
2
0 1 0 1
0 0
T (e3 ) = 0u1 + 2 u2 = 2 1 = @ 2 A.
1 1 @ A B1C
1 1
2
6
Thus the standard matrix for T is 0 1
1 0 0
B
T (e1 ) T (e2 ) T (e3 ) = @0 1
2
1C :
2A
0 1 1
2 2
(ii) From the de nition of T , R(T ) = spanfu1 ; u2 g. It is obvious that
fu1; u2g is linearly independent and hence is a basis for R(T ).
So rank(T ) = dim(R(T )) = 2
and nullity(T ) = dim(R3 ) rank(T ) = 3 2 = 1.
(iii) Since u3 u1 = 0 and u3 u2 = 0, u3 is orthogonal to V .
For any x = c1 u1 + c2 u2 + c3 u3 2 R3 ,
5. (a) Let A = a1 a2 a n where a j is the j th column of A.
0 1
0
B .. C
B.C
B C
B0C
B C
Ae j = a1 a2 a B C
n B1C the j th coordinate
B0C
B C
B.C
@ .. A
0
= 0a1 + 0aj 1 + 1aj + 0aj +1 + + 0an = aj .
(b) (i) Since Am 1 6= 0, Am 1 has at least one nonzero column, say, the j th
column is nonzero.
Let u = e . Then by (a), Am 1 e = the j th column of Am 1 6= 0.
j j
c1 u + c2 Au + + cm Am 1 u = 0: ()
) c1 Am 1 u = : 0
7
Since Am 1u is a nonzero vector, c1 = 0.
Suppose we have already shown that c1 = 0, c2 = 0, : : : , ci = 0 for some
i with 1 i < m.
Pre-multiplying Am i 1 to both sides of (), we have
Am i 1 (c Ai u + c Ai+1 u + + c Am 1 u) = Am
i+1 i+2 n
i 10
) ci+1 A u + ci+2 A u + + cn A i 2 u = 0
m 1 m 2m
) ci+1 Am 1 u = 0:
Applying (b) with m = n+1, there exists u 2 Rn such that fu; Au; : : : ; Anug
is linearly independent.
But this contradicts the fact that a set of n + 1 vectors in Rn cannot be
linearly independent (by Theorem 3.2.7).
So we conclude that our assumption is wrong and hence An = 0.
(b) Method 1:
8
Since fu1 ; u2 ; : : : ; u g is orthonormal, u u
n i i = 1 and ui uj = 0 for i 6= j .
So
xTx = x x
= (c1 u1 + c2 u2 + + cn u ) (c1 u1 + c2 u2 + + cn u
n n )
= c21 + c22 + + c2n :
On the other hand,
xTCx = x (Cx)
= (c1 u1 + c2 u2 + + cn u ) (c1 Cu1 + c2 Cu2 + + cn Cu )
n n
= (c1 u1 + c2 u2 + + cn u ) (1 c1 u1 + 2 c2 u2 + + n cn u )
n n
) 1 1c1c2++c22c2++ ++c2ncn n
2 2 2
1 2 n
) 1
xTCx n:
xT x
Method 2:
) 1
xTCx n:
xT x