Matrices and Linear Algebra
Matrices and Linear Algebra
net/publication/320170178
CITATIONS READS
0 3,983
1 author:
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Akhilesh Chandra Yadav on 02 October 2017.
i
ii
My Parents
and
1 Matrix Algebra 1
1.1 Matrices . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Matrix operations . . . . . . . . . . . . . . . . . . . 5
1.3 Transpose of a matrix . . . . . . . . . . . . . . . . . 13
1.4 Symmetric and Skew-symmetric matrices . . . . . . 13
1.5 Hermitian and skew-Hermitian matrices . . . . . . . 17
1.6 Orthogonal and Unitary matrices . . . . . . . . . . . 23
1.7 Adjoint matrix . . . . . . . . . . . . . . . . . . . . . 25
2 Rank of a matrix 31
2.1 Elementary operations . . . . . . . . . . . . . . . . . 31
2.2 Elementary matrices . . . . . . . . . . . . . . . . . . 37
2.3 Echelon form . . . . . . . . . . . . . . . . . . . . . . 38
2.4 Linear dependence and independence . . . . . . . . . 43
2.5 Normal form . . . . . . . . . . . . . . . . . . . . . . 54
2.6 Inverse of a matrix . . . . . . . . . . . . . . . . . . . 61
v
vi CONTENTS
Matrix Algebra
1.1 Matrices
Definition 1.1.1 An m×n matrix A is a rectangular arrangement
of m.n numbers (real or complex) in m-rows and n - columns. Let
aij be the number that appears in ith row and j th column of A for all
i = 1, 2, . . . , m and j = 1, 2, . . . , n. Then, it is denoted by (aij )m×n
and aij is called the (i, j)th entry of A. Thus, A = (aij )m×n and
we read A is an m by n matrix. Sometimes, we also read A is a
matrix of order m cross n or A is a matrix of size m × n .
In extended form it can be written as
a11 a12 ... a1j ... a1n
a21 a22 ... a2j ... a2n
.. .. .. .. ..
. . . . .
A=
a
i1 a i2 ... aij . . . ain
.. .. .. .. ..
. . . . .
am1 am2 . . . amj . . . amn
1
2 CHAPTER 1. MATRIX ALGEBRA
Some authors also use square brackets to denote matrices, that is;
a11 a12 . . . a1j . . . a1n
a21 a22 . . . a2j . . . a2n
.. .. .. .. ..
. . . . .
A=
ai1 ai2 . . . aij . . . ain
.. .. .. .. ..
. . . . .
am1 am2 . . . amj . . . amn
Example 1.1.2
1 2
A=
0 5
is a 2 by 2 matrix and
1 2 9
B=
3 0 2
is a 2 by 3 matrix.
Example 1.1.4 The 3×3 matrix whose (i, j)−th entry is (−1)i−j
for all i, j = 1, 2, 3, is given by
1 −1 1
−1 1 −1
1 −1 1
1. A = (ij )3×3 .
1.1. MATRICES 3
(
1 if i = j
2. A = (aij )2×2 , where aij = .
0 6 j
if i =
(
1 if i + j is even
3. A = (aij )3×3 , where aij = .
0 otherwise
(
ei+j if i + j is even
4. A = (aij )2×2 , where aij = .
0 otherwise
For example,
2 0 0 0
1 0 0 0
0 2 0 and 2 0 0
0 0 2 0
0 0 1
0 0 0 2
are diagonal matrices. The second matrix is a scalar matrix
Example 1.2.2
1 0 5 4 11 7 1+4 0 + 11 5 + 7
+ =
2 3 9 6 10 13 2+6 3 + 10 9 + 13
5 11 12
= .
8 13 22
Similarly,
1 0 5 4 11 7 1−4 0 − 11 5 − 7
− =
2 3 9 6 10 13 2−6 3 − 10 9 − 13
−3 −11 −2
=
−4 −7 −4
Proposition 1.2.3 Let A, B, C be any three m×n matrices. Then
A + B = B + A and A + (B + C) = (A + B) + C.
Proof: Since matrices A, B, C are of same sizes therefore A +
B, B + A, A + (B + C) and (A + B) + C are defined. Let A =
(aij )m×n , B = (bij )m×n and C = (cij )m×n . Then
A + B = (aij + bij )
= (bij + aij ) as aij + bij = bij + aij f or all i, j
= B+A
Similarly, A + (B + C) = (A + B) + C follows from the fact that
(aij + bij ) + cij = aij + (bij + cij ) for all i, j. 2
1 0 0 0 0 1
Exercise 1.2.7 If A = 2 6 7, B = 3 5 2 and
3 1 4 1 4 4
1 1 1
C = 2 2 5 . Then, evaluate 3A, (−7)B, A + 5B and 12A −
3 1 44
17B + 9C.
b1k ai1b1k
i-th row of A b2k ai2b2k
ai1 ai2 ai3 . . . ain b3k ai3b3k
... ...
bnk ainbnk
Pn
j=1 aij bjk = ai1 b1k + ai2 b2k + . . . + ain bnk
k-th column of B
1.2. MATRIX OPERATIONS 9
Example 1.2.11
6
1 2 0 1×6+2×7+0×8
7 =
3 4 9 3×6+4×7+9×8
8
20
=
118
−1 2
Example 1.2.12 If A = then
1 3
2 −1 2 −1 2
A = AA =
1 3 1 3
2
(−1) + 1.2 (−1).2 + 2.3
=
1.(−1) + 3.1 1.2 + 3.3
3 4
=
2 11
3 4 −1
Exercise 1.2.13 Let A = and
−2 11 7
1 1 0
B = −1 0 1 . Evaluate AB. What about BA? (Hint:
0 −1 −1
BA is not defined, why?).
0 1 1 1
Example 1.2.14 If A = and then
0 0 0 0
0 0 0 1
AB = = 02×2 but BA = 6= 02×2 .
0 0 0 0
Thus, AB 6= BA (i.e; in general matrix product is non-
commutative). Here A and B both are non-zero matrices but
BA = 02×2 . Thus, the product of two non-zero matrices
may be zero.
and BA are defined and have same sizes (orders). However they
may or may not be equal( see Example 1.2.14. Thus, in general
AB 6= BA. From example 1.2.14, it also follows that product of
two non-zero matrices may be zero.
1 0 1
Example 1.2.16 Consider the matrices A = and B =
0 1 1
1 1
1 0. Then AB and BA both are defined and of orders(sizes)
0 1
1 2
2 × 2 and 3 × 3 respectively. Verify that AB = and BA =
1 1
1 1 2
1 0 1
0 1 1
3
3
1. 1 −1 3 1 and 1 1 −1 3 .
0 0
1 2 4 2 9 0
2. 0 3 1 0 3 1.
0 0 2 0 0 1
1 0 0 1 2 4 1 2 4 1 0 0
3. 0 3 0 6 3 5 and 6 3 5 0 3 0 .
0 0 2 7 8 9 7 8 9 0 0 2
Also verify that both products are same. Indeed, if A is any
square matrix of size n × n and D = cIn is any scalar matrix
of size n × n then AD = DA.
Proof: Let A = (aij )m×n , B = (bjk )n×p and C = (ckl )p×r . Then
Xn
(AB)C = aij bjk (ckl )
j=1
p
X Xn
= aij bjk ckl
k=1 j=1
n
X p
X
= aij [bjk ckl ]
j=1 k=1
p
!
X
= (aij ) bjk ckl
k=1
= A(BC)
2x + 3y = 1
x−y = 2
x
in to matrix form AX = B, where X =
y
1 2 2
Exercise 1.2.21 If A = 2 1 2 then show that A2 −4A−5I =
2 2 1
03×3 .
12 CHAPTER 1. MATRIX ALGEBRA
3 −4
Exercise 1.2.22 If A = then prove that
1 −1
k 1 + 2k −4k
A = , where k is any natural number.
k 1 − 2k
cos θ sin θ
Exercise 1.2.23 If Aθ = and
− sin θ cos θ
cos φ sin φ
Aφ = , then prove that
− sin φ cos φ
cos(θ + φ) sin(θ + φ)
Aθ Aφ = .
− sin(θ + φ) cos(θ + φ)
In particular (Aθ )n = Anθ , where n is an integer. Also prove that
Aθ Aφ = I2 if and only if θ + φ = 2πn.
Exercise 1.4.5 Let A and B be any two square matrices such that
AB = BA. Then, we have the following:
(i). AB and BA both are symmetric if either both are symmetric
or both are skew-symmetric.
(ii). AB and BA both are skew-symmetric if one of them is skew-
symmetric.
Exercise 1.4.8 Let A, B be any two row vector of size 1×n. Prove
that
(i). AB 0 = BA0 ,
(ii). A0 B − B 0 A is skew-symmetric matrix of size n × n.
A + A0 A − A0
A= + (1.4.1)
2 2
A+A0 A−A0
Let P = 2 and Q = 2 . Then
A=P +Q (1.4.2)
A=P +Q (1.4.3)
A=R+S (1.4.4)
A0 = P 0 + Q0 = P − Q (1.4.5)
0 0 0
A =R +S =R−S (1.4.6)
16 CHAPTER 1. MATRIX ALGEBRA
1 −1 3
A0 = −1 2 −1 .
0 3 0
A+A0
Consider P = 2 . Then
1 −1 3/2
P = −1 2 1
3/2 1 0
Since (i, j)-th entry of Q is the negative of (j, i)-th entry for all i
and j, i.e; Q0 = −Q. Therefore Q is skew-symmetric.
Now
1 −1 3/2 0 0 −3/2
P + Q = −1 2 1 + 0 0 2
3/2 1 0 3/2 −2 0
1 −1 0
= −1 2 3 =A
3 −1 0
is given by
1 1 − i −5 − i 0
A= .
−7 1 1 3+i
Proof: Let A = (aij )n×n be Hermitian. Then aji = aij for all
i, j = 1, 2, . . . , n. In particular, aii = aii for all i = 1, 2, . . . , n, i.e;
aii is real for all i = 1, 2, . . . , n. 2
is skew-hermitian.
For,
0 1+i 2−i
Given that B = −1 + i i 3 + 7i then
−2 − i −3 + 7i −3i
0 1−i 2+i
B = −1 − i −i 3 − 7i .
−2 + i −3 − 7i 3i
Thus,
0
0 1−i 2+i
B θ = (B)0 = −1 − i −i 3 − 7i
−2 + i −3 − 7i 3i
0 −1 − i −2 + i
= 1−i −i −3 − 7i
2 + i 3 − 7i 3i
0 1+i 2−i
= − −1 + i i 3 + 7i = −A.
−2 − i −3 + 7i −3i
A+Aθ A−Aθ
Let P = 2 and Q = 2 . Then
1 1
P θ = [ (A + Aθ )]θ = [Aθ + (Aθ )θ ] = P
2 2
1 1
Qθ = [ (A − Aθ )]θ = [Aθ − (Aθ )θ ] = −Q
2 2
as (Aθ )θ = A.
A+Aθ A−Aθ
Thus, P = 2 is Hermitian and Q = 2 is skew-Hermitian.
A + Aθ A − Aθ
A= + (1.5.1)
2 2
Then
A=P +Q (1.5.2)
A+Aθ A−Aθ
where P = 2 and Q = 2 . Now,
θ
θ 1 θ 1h θ i
P = (A + A ) = A + (Aθ )θ = P
2 2
θ
θ 1 θ 1h θ i
Q = (A − A ) = A − (Aθ )θ = −Q
2 2
as (Aθ )θ = A.
θ θ
Thus, P = A+A
2 is Hermitian and Q = A−A
2 is skew-Hermitian
and so by equation 1.5.2 A can be expressed as the sum of a Her-
mitian matrix and a skew-Hermitian matrix. For uniqueness, let
A=P +Q (1.5.3)
A=R+S (1.5.4)
1.5. HERMITIAN AND SKEW-HERMITIAN MATRICES 21
Aθ = P θ + Q θ = P − Q (1.5.5)
θ θ θ
A =R +S =R−S (1.5.6)
A = P + iQ (1.5.9)
A = R + iS (1.5.10)
Aθ = P θ + iQθ = P − iQ (1.5.11)
θ θ θ
A = R + iS = R − iS (1.5.12)
Then
θ 1−i −3
A =
2 1 + 5i
Hence
A + Aθ
1 1+i 2 1−i −3
= +
2 2 −3 1 − 5i 2 1 + 5i
1 −1/2
=
−1/2 1
Next,
A − Aθ
1 1+i 2 1−i −3
= −
2 2 −3 1 − 5i 2 1 + 5i
i 5/2
=
−5/2 −5i
1 −1/2 i 5/2
Take P = and Q = . Clearly
−1/2 1 −5/2 −5i
P is Hermitian
and Q is skew-Hermitian.
1 −1/2
Thus, P = is a Hermitian matrix and Q =
−1/2 1
i 5/2
is a skew-Hermitian matrix such that A = P + Q.
−5/2 −5i
θ θ
(ii). Consider P = A+A
2 and Q = A−A
2i . Then, we have
1 −1/2 1 −5i/2
P = and Q=
−1/2 1 5i/2 −5
1.6. ORTHOGONAL AND UNITARY MATRICES 23
1 −1/2 1 −5i/2
P +Q = +i
−1/2 1 5i/2 −5
1+i 2
= =A
−3 1 − 5i
1 −1 i 3 −1 i
(i). −1 i 1 (ii). −1 3 − i 1 + 2i
1 −1 0 1 + i −1 + i 3 − 2i
1 −1 2 3 9 0
(i). 8 i 1 (ii). −1 3 − i 0
1 −1 0 1+i 0 3 − 2i
cos φ sin φ 0
Example 1.6.2 Let A = − sin φ cos φ 0 .
0 0 1
24 CHAPTER 1. MATRIX ALGEBRA
cos φ − sin φ 0
Then A0 = sin φ cos φ 0 . Now,
0 0 1
cos φ − sin φ 0 cos φ sin φ 0
A0 A = sin φ cos φ 0 − sin φ cos φ 0
0 0 1 0 0 1
2
cos2 φ + sin φ
cos φ sin φ − sin φ cos φ 0
= cos φ sin φ − sin φ cos φ sin2 φ + cos2 φ 0
0 0 1
1 0 0
= 0 1 0 = I3
0 0 1
for all i, j = 1, 2, . . . , n.
26 CHAPTER 1. MATRIX ALGEBRA
1 2 0
For example, the cofactor A21 of a21 in A = −1 3 4 is
5 6 7
2+1 2
21
0
A = (−1) 6 7 = −14.
2
Adj A
Thus, if |A| =
6 0, then A. |A| = In , i.e; A is invertible and
Adj A
A−1 =
|A|
Exercises
Exercise 1.7.4 Write down the extended form of matrix A =
(aij )2×2 , where aij = ei−j for all i and j.
2x − 3y = 0, y = 5
1 2 3 1 1 −1
Exercise 1.7.9 Let A = , B = , C =
0 1 1 0 8 1
d 0
and D = . Verify the following
0 d
1. A + B = B + A,
2. A − B 6= B − A,
28 CHAPTER 1. MATRIX ALGEBRA
3. (A + B) + C = A + (B + C),
4. (AB)C = A(BC),
5. (A + B)C = AC + BC,
6. A(B + C) = AB + AC,
7. AB 6= BA,
8. DA = AD,
Exercise 1.7.11 Prove that the zero square matrix is the only ma-
trix which is symmetric (Hermitian) as well as skew-symmetric
(skew-hermitian).
0 0 0 0
1.7. ADJOINT MATRIX 29
Rank of a matrix
31
32 CHAPTER 2. RANK OF A MATRIX
n
X n
X
pil alk = pil alk
l=1 l=1
Xn
= δjl alk
l=1
= δj1 a2k + δj2 ajk + . . . + δj(j−1) a(j−1)k + δjj ajk
+δj(j+1) a(j+1)k + . . . + δjn ank
= 0 + 0 + . . . + 0 + 1.ajk + 0 + . . . + 0
(by definition of δij )
= ajk
Then
1 0 0 a11 a12 a11 a12
P A = 0 0 1 a21 a22 = a31 a32
0 1 0 a31 a32 a21 a22
which is obtained by interchanging second and third row.
Then
1 0 0 0
a11 a12 a13 a14 0 0 1 0
AP =
a21 a22 a23 a24 0 1 0 0
0 0 0 1
a11 a13 a12 a14
=
a21 a23 a22 a24
Proposition 2.1.8 Let A be any m×n matrix and D = (ci δij )m×m
be any diagonal matrix. Then DA (AD) is obtained from A by
multiplying the i-th row (column) of A by ci .
2.1. ELEMENTARY OPERATIONS 35
are zero except the (r, s)-th entry which is λ, i.e; (i, j)-th entry
λ ) is given by
(Ers ij
λ λ if i = r, j = s
(Ers )ij =
0 otherwise
Let P = (pij )m × m and A = (ajk )m×n . By definition of P , we
have P = Im + Ersλ . Thus,
λ
pij = δij + (Ers )ij
λ
= aik + (Ers )is ask
aik if i 6= r
=
ark + λask if i=r
is in row-echelon form because the zero row lies below to each non-
zero row and the number of zeros, before first non-zero entry, from
first row to fourth row are in ascending order.
is not a row-echelon matrix because its first row contains one 0(zero)
while the second row contains no 00 s before their first non-zero en-
try.
Example 2.3.5 Prove that every diagonal matrices whose all di-
agonal entries are non-zero, are in row-echelon form.
Applying Ri → Ri − abrs
is
R1 on matrix B, we have
0 0 . . . 0 ars ar s+1 ... arn
0 0 . . . 0 0 c1 s+1 . . . c1n
0 0 . . . 0 0 c . . . c2n
C= 2 s+1
.. .. .. .. .. .. .. ..
. . . . . . . .
0 0 . . . 0 0 cm−1 s+1 . . . cm−1 n
is given by
0 1 0 0 9 1 0 2 1 7
1 0 2 1 7 ∼ 0 1 0 0 9 by applying R1 ↔ R2
0 0 0 0 0 0 0 0 0 0
is given by
1 2 3
A = 3 1 2
5 5 8
1 2 3
R2 → R2 − 3R1
∼ 0 −5 −7
R3 → R3 − 5R1
0 −5 −7
1 2 3
∼ 0 −5 −7 R3 → R3 − R2
0 0 0
1 2 3
Example 2.3.11 The echelon form of the matrix A = 3 1 2
2 3 1
2.3. ECHELON FORM 41
is given by
1 2 3
A = 3 1 2
2 3 1
1 2 3
R → R2 − 3R1
∼ 0 −5 −7 2
R3 → R3 − 2R1
0 −1 −5
1 2 3
∼ 0 −5 −7 R3 → R3 − 51 R2
18
0 0 −5
1 2 3
R2 → −15 R2
∼ 0 1 75 −5
R3 → 18 R3
0 0 1
1 2 0
R1 → R1 − 3R3
∼ 0 1 0
R2 → R2 − 57 R3
0 0 1
1 0 0
∼ 0 1 0 R1 → R1 − 2R2
0 0 1
Observe that this echelon form of matrix is special in the sense that
first non-zero entry of each row is one together with all the entries
above and below the first non-zero in its corresponding column are
zero. Such echelon form of a matrix is called areduced row-echelon
form or echelon form.
Exercise
2.3.12 Reduce
the
following matrices
to echelon
form
1 1 1 −1 1 2 0 −1 1 2 3
(i). 1 2 3 4 (ii). 2 6 −3 −3 (iii). 1 4 2
3 4 5 2 3 10 −6 −5 2 6 5
Exercise
2.3.14
Reduce the
following matrices
to row-echelon
form
1 2 3 14 1 1 1 6 1 2 1 2
(i). 3 1 2 11 (ii). 1 −1 1 2 (iii). 3 6 5 4
2 3 1 11 2 1 −1 1 2 4 3 3
Exercise
2.3.15 Reduce
the
following matrices
toechelon form
5 3 7 4 1 1 1 6 1 1 1 1
(i). 3 26 2 9 (ii). 1 2 3 10 (iii). 1 2 4 λ
7 2 10 5 1 2 λ µ 1 4 10 λ2
Exercise
2.3.16 For given matrix
1 1 0 1 4
2 0 0 4 7
A= 1 1
determine the value of α such that
1 0 5
1 −3 −1 −10 α
all the rows of its corresponding echelon form are non-zero.
Then
1 −1 0 −1 −5 −1
R2 → R2 − 2R1
2 1 −1 −4 1 −1
A ∼ R3 → R3 − R1
1 1 1 −4 −6 3
R4 → R4 − R1
1 4 2 −8 −5 8
1 −1 0 −1 −5 −1
0 1 −2 1 12 −3
∼ R2 → R2 − R3
0 2 1 −3 −1 4
0 5 2 −7 0 9
1 0 −2 0 7 −4
R1 → R1 + R2
0 1 −2 1 12 −3
∼ R3 → R3 − 2R2
0 0 5 −5 −25 10
R4 → R4 − 5R2
0 0 12 −12 −60 24
1 0 −2 0 7 −4
0 1 −2 1 12 −3 R3 → R3 /5
∼
0 0 1 −1 −5 2 R4 → R4 /12
0 0 1 −1 −5 2
1 0 −2 0 7 −4
0 1 −2 1 12 −3
∼ R4 → R4 − R3
0 0 1 −1 −5 2
0 0 0 0 0 0
1 0 0 −2 −3 0
0 1 0 −1 2 1 R1 → R1 + 2R3
∼
0 0 1 −1 −5 2 R2 → R2 + 2R3
0 0 0 0 0 0
Exercise 2.3.20 Reduce the following matrices to reduced row-
echelon form
1 0 1 −1 −3 1 0 −1 −2
0 1 0 −1 1 1 1 6
2 0 1 0 −1
0 0 1 −1 , , 1 −1 1 2
−5 0 0 −1 −1
2 1 −1 1
0 0 0 0 0 0 0 0 1
j = 1, 2, . . . , n. Then,
a1j
a2j
Ri = a1 a2 . . . an and Cj =
..
.
amj
Then,
R1 + R2 = (1 1 3) + (0 2 1) = (1 + 0 1 + 2 3 + 1) = (1 3 4) = R3 .
k1 x1 + k2 x2 + . . . + kp xp = 0
k1 x1 + k2 x2 + . . . + kp xp = 0 ⇒ k1 = k2 = . . . = kp = 0
Suppose that
a(1 2 0) + b(1 0 1) + c(0 1 1) = (0 0 0)
Then, we have
(a + b 2a + c b + c) = (0 0 0)
Thus,
a+b = 0 . . . (1)
2a + c = 0 . . . (2)
b+c = 0 . . . (3)
Subtracting equation (3) from equation (2), we have
2a − b = 0 . . . (4)
Solving equations (1) and (4), we have a = b = 0. Putting the
value of a in equation (2), we have c = 0. Thus, we have
a(1 2 0) + b(1 0 1) + c(0 1 1) = (0 0 0) ⇒ a = b = c = 0
By definition of linearly independent rows, all the rows are lin-
early independent. Similarly, all the columns of A are linearly
independent.
Exercise 2.4.10 Prove that rows of matrix A are linearly inde-
pendent while the columns of A are linearly dependent, where
1 2 0 2
0 1 1 1 .
1 0 1 0
Also determine the maximum number of linearly independent rows
and maximum number of linearly independent columns. Does max-
imum number of linearly independent rows is same as the maximum
number of linearly independent columns?
Exercise 2.4.11 Determine the maximum number of linearly in-
dependent rows and maximum number of linearly independent columns
of
following matrices:
5 3 7 4 1 1 1 6 1 1 1 1
(i)3 26 2 9 (ii) 1 2 3 10 (iii) 1 2 4 λ
7 2 10 5 1 2 λ µ 1 4 10 λ2
2.4. LINEAR DEPENDENCE AND INDEPENDENCE 47
k1 x1 + k2 x2 + . . . + kp xp = 0 ⇔
k1 k2 ki −1 ki +1 kp
x i0 = − x1 − x2 − . . . − 0 xi0 −1 − 0 xi0 +1 − xp
ki0 ki0 ki0 ki0 ki0
Theorem 2.4.20 Row rank and column rank both are invariant
under elementary row and column operations (elementary opera-
tions).
Theorem 2.4.21 Row rank and column rank are same for a given
matrix.
is 3. For
1 2 3 14 1 2 3 14
R → R2 − 3R1
3 1 2 11 ∼ 0 −5 −7 −31 2
R3 → R3 − 2R1
2 3 1 11 0 −1 −5 −17
1 2 3 14
∼ 0 −1 −5 −17 R2 ↔ R3
0 −5 −7 −31
1 2 3 14
∼ 0 −1 −5 −17 R3 → R3 − 5R2
0 0 18 54
2 3 −1 −1
1 −1 −2 −4
Exercise 2.4.27 Find the rank of matrix .
3 1 3 −2
6 3 0 7
2 3 −1 −1
1 −1 −2 −4
Solution: Given that A = .
3 1 3 −2
6 3 0 7
52 CHAPTER 2. RANK OF A MATRIX
Then
2 3 −1 −1
1 −1 −2 −4
A =
3 1 3 −2
6 3 0 7
1 −1 −2 −4
2 3 −1 −1
∼ R1 ↔ R2
3 1 3 −2
6 3 0 7
1 −1 −2 −4
0 R2 → R2 − 2R1
5 3 7
∼
0
R3 → R3 − 3R1
4 9 10
R4 → R4 − 6R1
0 9 12 31
1 −1 −2 −4
0 1 −6 −3
∼
0
R2 → R2 − R3
4 9 10
0 9 12 31
1 −1 −2 −4
0 1 −6 −3 R3 → R3 − 4R2
∼
0 0 33 22 R4 → R4 − 9R2
0 0 66 58
1 −1 −2 −4
0 1 −6 −3 R3 → R3 − 4R2
∼
0 0 33 22 R4 → R4 − 9R2
0 0 0 14
1 2 0 −1
2 6 −3 −3
3 10 −6 −5
2.4. LINEAR DEPENDENCE AND INDEPENDENCE 53
Solution:
1 2 0 −1 1 2 0 −1
2 6 −3 −3 ∼ 0 2 −3 R2 → R2 − 2R1
−1
R3 → R3 − 3R1
3 10 −6 −5 0 4 −6 −2
1 2 0 −1
∼ 0 2 −3 −1 R3 → R3 − 2R2
0 0 0 0
Solution:
1 2 −1 −2 1 2 −1 −2
−1 −1 1 1 ∼ 0 1 0 −1 R2 → R2 + R1
0 1 2 1 0 1 2 1
1 2 −1 −2
∼ 0 1 0 −1 R3 → R3 − R1
0 0 2 2
1 3 4 5
A = 1 2 6 7 .
1 5 0 1
Then
1 3 4 5
A = 1 2 6 7
1 5 0 1
1 3 4 5
R2 → R2 − R1
∼ 0 −1 2 2
R3 → R3 − R1
0 2 −4 −4
1 0 10 11
R1 → R1 + 3R2
∼ 0 −1 2 2
R3 → R3 + 2R2
0 0 0 0
1 0 0 0 C2 → (−1).C2
∼ 0 1 2 2 C3 → C3 − 10C1
0 0 0 0 C4 → C4 − 11C1
1 0 0 0
C3 → C3 − 2C2
∼ 0 1 0 0
C4 → C4 − 2C2
0 0 0 0
1 0 0 0
0 1 0 0
0 0 0 0
Exercise 2.5.3 Reduce the matrix to its normal form and find its
rank, where
0 1 −3 −1
1 0 1 1
A= 3 1 0
.
2
1 1 −2 0
56 CHAPTER 2. RANK OF A MATRIX
Solution:
0 1 −3 −1
1 0 1 1
A =
3
1 0 2
1 1 −2 0
1 0 1 1
0 1 −3 −1
∼
3
R1 ↔ R2
1 0 2
1 1 −2 0
1 0 1 1
0 1 −3 −1 R3 → R3 − 3R1
∼
0 1 −3 −1 R4 → R4 − R1
0 1 −3 −1
1 0 1 1
0 1 −3 −1 R3 → R3 − R2
∼
0 0 0 0 R4 → R4 − R2
0 0 0 0
1 0 0 0
0 1 −3 −1 C3 → C3 − C1
∼
0 0 0 0 C4 → C4 − C1
0 0 0 0
1 0 0 0
0 1 0 0 C3 → C3 + 3C2
∼
0 0 0 0 C4 → C4 + C1
0 0 0 0
Exercise 2.5.4 Reduce the matrix to its normal form and hence
find its rank, where
2 −2 0 0
4 2 0 2
A=
1 −1 0 3
1 −2 1 2
2.5. NORMAL FORM 57
Solution:
2 −2 0 0
4 2 0 2
A =
1 −1 0 3
1 −2 1 2
1 −1 0 3
4 2 0 2
∼ (R1 ←→ R3 )
2 −2 0 0
1 −2 1 2
1 −1 0 3
R2 → R2 − 4R1
0 6 0 −10
∼ R3 → R3 − 2R1
0 0 0 −6
R4 → R4 − R1
0 −1 1 −1
1 −1 0 3
0 0 6 −16
∼ R2 → R2 + 6R4
0 0 0 −6
0 −1 1 −1
1 −1 0 3
0 −1 1 −1
∼ R2 ↔ R4
0 0 0 −6
0 0 6 −16
1 −1 0 3
0 −1 1 −1
∼ R3 ↔ R4
0 0 6 −16
0 0 0 −6
1 −1 0 0
R1 → R1 − 3R4
0 −1 1 0
∼
0
R2 → R2 + R4
0 6 0
R3 → R3 + 16R1
0 0 0 1
1 −1 0 0
0 1 −1 0 R2 → (−1)R2
∼
0 0 1 0 R3 → 61 R3
0 0 0 1
1 −1 0 0
0 1 0 0
∼
0
R2 → R2 + R3
0 1 0
0 0 0 1
58 CHAPTER 2. RANK OF A MATRIX
1 0 0 0
0 1 0 0
∼
0
R1 → R1 + R2
0 1 0
0 0 0 1
∼ I4
1 1/2 0 −1/6
1 0 0 0 1 0 0
0 1 0 0 = −4/3 1/3 0 A 0 1/2 −1/2 −1/2
0 0 0 −1/3
0 0 1 0 −2/3 2/3 −1
0 0 1 0
P = P In = P (AQ) = (P A)Q = In Q = Q
It can be written as
1 2 3 1 0 0
1 3 4 = 0 1 0 A
1 4 4 0 0 1
Exercises
64 CHAPTER 2. RANK OF A MATRIX
2
66 CHAPTER 2. RANK OF A MATRIX
Chapter 3
System of linear
equations
It can be written as
a11 a12 ... a1n x1 b1
a21 a22 ... a2n x2 b2
.. .. = ..
.. .. ..
. . . . . .
am1 am2 . . . amn xn bm
i.e; AX = b, where
a11 a12 . . . a1n x1 b1
a21 a22 . . . a2n x2 b2
A= . .. , X = .. and b = .. .
. .. ..
. . . . . .
am1 am2 . . . amn xn bm
67
68 CHAPTER 3. SYSTEM OF LINEAR EQUATIONS
Definition 3.0.6 Two linear systems which have same sets of so-
lutions are called equivalent.
a11 a12 ... a1n x1 0
a21 a22 ... a2n x2 0
.. .. = ..
.. .. ..
. . . . . .
am1 am2 . . . amn xn 0
i.e; AX = 0, where
a11 a12 ... a1n x1
a21 a22 ... a2n x2
A= . .. , X = ..
.. ..
.. . . . .
am1 am2 . . . amn xn
AY = 0 ⇔ (P A)Y = 0
⇔ EY = 0.
Thus proved. 2
Thus, to find solution of AX = 0, we first reduce A to
its Echelon form E and then solve the equation EX = 0.
solution which is 0.
(ii). If r < n, then (n − r) solutions are linearly independent.
Indeed, in this case, n − r unknowns are free (independent) to take
any values and so it has infinitely many solutions.
x + 3y − 2z = 0
2x − y + 4z = 0
x − 11y + 14z = 0
Now,
1 3 −2 1 3 −2
2 −1 R2 → R2 − 2R1
4 ∼ 0 −7 8
R3 → R3 − R1
1 −11 14 0 −14 16
1 3 −2
∼ 0 −7 8 R3 → R3 − 2R2
0 0 0
This gives
x + 3y − 2z = 0, −7y + 8z = 0 .........(∗)
8c
Take z = c, where c is arbitrary. Then y = 7 and so
8c 10c
x = −3y + 2z = −3 × + 2c = −
7 7
i.e; x = − 10c 8c
7 , y = 7 , z = c is a solution to the given system,
where c is arbitrary. If c = 0 then we have x = y = z = 0 is a
solution. If c = 7 then x = −10, y = 8, z = 7 is a solution. If c = 1
then x = − 10 8
7 , y = 7 , z = 1 is a solution. Indeed, it has infinitely
many solutions which is given by
10c
x − 7
8c
X= y =
7
z c
4k 7k
y=− ,z = −
5 5
Hence
k
Y = − 4k
5
7k
− 10
is also solution for the given system, where k is any scalar.
It is noted that both solutions
10c
− 7
X = 8c 7
c
3.1. SYSTEM OF HOMOGENEOUS LINEAR EQUATIONS 75
and
k
Y = − 4k
5
7k
− 10
10c
are same. Indeed, if we put k = − then Y = X.
7
x+y+z = 0
2x − y − 3z = 0
3x − 5y + 4z = 0
x + 17y + 4z = 0
λx + y + z = 0
x + λy + z = 0
x + y + λz = 0
Now,
λ 1 1
1 λ 1 = λ(λ2 − 1) − (λ − 1) + 1 − λ
1 1 λ
= (λ − 1) [λ(λ + 1) − 1 − 1]
= (λ − 1) λ2 + λ − 2
1 1 1
A = 1 1 1
1 1 1
Since
1 1 1 1 1 1
A = 1 1 1 ∼ 0 0 0 R2 → R2 − R1 , R3 → R3 − R1
1 1 1 0 0 0
Since
−2 1 1 1 1 −2
A = 1 −2 1 ∼ 1 −2 1 R1 ↔ R3
1 1 −2 −2 1 1
1 1 −2
R2 → R2 − R1
∼ 0 −3 3
R3 → R3 + 2R1
80 3 −3
1 1 −2
∼ 0 −3 3 R3 → R3 + R2
0 0 0
Then,
|P A| = 0 (as last row of P A is a zero − row)
|P ||A| = 0 (as |P A| = |P ||A|)
|A| = 0 (as |P | =
6 0)
Thus, if ρ(A) < n then |A| = 0. One may also prove that if
|A| = 0 then ρ(A) < n. We shall use this fact frequently in the
section of eigen values and eigen vectors. Indeed, this section is
applicable for the section eigen values and eigen vectors
Exercise 3.1.11 Discuss for all values of k, the system of equa-
tions
2x + 3ky + (3k + 4)z = 0
x + (k + 4)y + (4k + 2)z = 0
x + 2(k + 1)y + (3k + 4)z = 0
Exercise 3.1.12 Prove that the only value of λ for which the fol-
lowing system of equations has non-zero solution is 6:
x + 2y + 3z = λx
3x + y + 2z = λy
2x + 3y + z = λz
Exercise 3.1.13 Solve the following system of homogeneous linear
equations:
x1 − x2 + x3 = 0, x1 + 2x2 − x3 = 0 2x1 + x2 + 3x3 = 0
Exercise 3.1.14 Solve:
x−y+z =0 x+y−z =0 −x+y+z =0
C1 x 1 + C2 x 2 + . . . + Cn x n = b ......(1)
for each k = 1, 2, . . . , n − r.
Suppose that system is consistent, then there exists scalars pj ,
j = 1, 2, . . . , n, not all zero(since b 6= 0), such that
p1 C1 + p2 C2 + . . . + pn Cn = b ......(3)
b = c1 C1 + c2 C2 + . . . + cr Cr
i.e;
r
X
Cr+k = clk Cl .........(1)
l=1
and
b = c1 C1 + c2 C2 + . . . + cr Cr ....(2)
where c1k , c2k , . . . , crk are not all zero for each k = 1, 2, . . . , n − r
and c − 1, c2 , . . . , cr are zero. By(1), wehave n − r inde-
not all
c11 c12
c21 c22
.. ..
. .
cr1 cr2
pendent solutions X1 = −1 , X2 = 0 ,
0 −1
0 0
. .
.
. .
.
0 0
82 CHAPTER 3. SYSTEM OF LINEAR EQUATIONS
c13
c1(n−r)
c23
c2(n−r)
..
..
.
.
cr3
cr(n−r)
0
X3 = , . . ., Xn−r = 0
0
0
−1
..
0
.
.. 0
.
−1
0
of equation C1 x1 + C2 x2 + . . . + Cn xn = 0, i.e;
AXi = 0 . . . (3)
for all i = 1, 2, . . . , n − r.
It is noted that rank A = r < n, thus by fact, AX = 0 has n − r
linearly independent solutions X1 , X2 , . . . , Xr , say.
By (2),
c1
c2
..
.
cr
X0 = . . . (4)
0
0
.
..
0
is a solution of AX = b. Using Equations (3) and (4), we have
x + 2y + 3z = 14
3x + y + 2z = 11
2x + 3y + z = 11
1 2 3 x 14
Here A = 3 1 2 , X = y and b = 11.
2 3 1 z 11
Hence, augmented matrix is
1 2 3 : 14
(A|b) = 3 1 2 : 11
2 3 1 : 11
Now,
1 2 3 : 14
(A|b) = 3 1 2 : 11
2 3 1 : 11
1 2 3 : 14
R2 → R2 − 3R1
∼ 0 −5 −7 : −31
R3 → R3 − 2R1
0 −1 −5 : −17
1 2 3 : 14
∼ 0 0 18 : 54 R2 → R2 − 5R3
0 −1 −5 : −17
1 2 3 : 14
1
∼ 0 0 1 : 3 R2 → 18 R2
0 −1 −5 : −17
1 2 3 : 14
∼ 0 −1 −5 : −17 R2 ↔ R3
0 0 1 : 3
x + y + z = −3
3x + y − 2z = −2
2x + 4y + 7z = 7
1 1 1 x 6
A = 1 2 3 , X = y and b = 14 .
1 4 7 z 30
Then
1 1 1 : 6
(A|b) = 1 2 3 : 14 .
1 4 7 : 30
Now,
1 1 1 : 6
(A|b) = 1 2 3 : 14
1 4 7 : 30
1 1 1 : 6
R2 → R2 − R1
∼ 0 1 2 : 8
R3 → R3 − R1
0 3 6 : 24
1 1 1 : 6
∼ 0 1 2 : 8 R3 → R3 − 3R2
0 0 0 : 0
3.2. SYSTEM OF NON-HOMO. LINEAR EQUATIONS 87
1 1 1 x 6
A = 1 2 3 , X = y and b = 10 .
1 2 λ z µ
Then augmented matrix is
1 1 1 : 6
(A |b) = 1 2 3 : 10
1 2 λ : µ
88 CHAPTER 3. SYSTEM OF LINEAR EQUATIONS
Now,
1 1 1 : 6
(A |b) = 1 2 3 : 10
1 2 λ : µ
1 1 1 : 6
R2 → R2 − R1
∼ 0 1 2 : 4
R3 → R3 − R1
0 1 λ−1 : µ−6
1 1 1 : 6
∼ 0 1 2 : 4 R3 → R3 − R2
0 0 λ − 3 : µ − 10
x+y+z = 1
x + 2y + 4z = a
x + 4y + 10z = a2
1 1 1 x 1
A = 1 2 4 , X = y and b = a .
1 4 10 z a2
Then augmented matrix is
1 1 1 : 1
(A |b) = 1 2 4 : a
1 4 10 : a2
3.2. SYSTEM OF NON-HOMO. LINEAR EQUATIONS 89
Now,
1 1 1 : 1
(A | b) = 1 2 4 : a
1 4 10 : a2
1 1 1 : 1
R2 → R2 − R1
∼ 0 1 3 : a−1
R3 → R3 − R1
0 3 9 : a2 − 1
1 1 1 : 1
∼ 0 1 3 : a−1 R3 → R3 − 3R2
0 2
0 0 : a − 3a + 2
Exercises
Exercise 3.2.12 Solve the following systems of equations:
1. 2x − y + 3z = 8, −x + 2y + z = 4, 3x + y − 4z = 0.
2. x − 2y + 3z = 2, 2x − 3z = 0, x + y + z = 0.
90 CHAPTER 3. SYSTEM OF LINEAR EQUATIONS
4. x + y + z = 7, x + 2y + 3z = 16, x + 3y + 4z = 22.
5. x − y + 2z = 4, 3x + y + 4z = 6, x + y + z = 1.
−2x + y + z = a
x − 2y + z = b
x + y − 2z = c
2
Chapter 4
91
92 CHAPTER 4. EIGEN VALUES AND EIGEN VECTORS
Thus, all the eigen values of a square matrix A are roots of the
equation det (A − λIn ) = 0, where λ is a parameter.
det (A − xIn ) = 0
exactly two eigen values i, −i. It is also noted that eigen values of
a square matrix may or may not be equal.
If c1 , c2 , . . . , ck are distinct roots of det (A − xIn ) such that
4 4
A − 1.I2 =
1 1
4 4 1
∼ R2 → R2 − R1
0 0 4
x
Hence, eigen vector X = is given by 4x+ 4y = 0, i.e; y = −x.
y
Take x = a, where a 6= 0 is arbitrary scalar, then y = −a.
Hence, eigen vector associated with λ = 1 is
a
X1 = , a 6= 0
−a
and a is arbitrary scalar.
If we take a = 1, then
1
X= .
−1
95
−1 4
A − 6.I2 =
1 −4
−1 4
∼ R2 → R2 + R1
0 0
x
Hence, eigen vector X = is given by −x+ 4y = 0, i.e; x = 4y.
y
Take y = 1, we have x = 4. Hence, eigen vector associated to
λ = 6 is
4
X2 = .
1
Then
−3 − λ 1 −1
|A − λI3 | = −7 5−λ −1
−6 6 −2 − λ
−2 − λ 1 −1
= −2 − λ 5 − λ −1 C1 → C1 + C2
0 6 −2 − λ
1
1 −1
= −(2 + λ) 1 5 − λ −1
0 6 −2 − λ
1
1 −1
= −(2 + λ) 0 4 − λ
0
R2 → R2 − R1
0 6 −2 − λ
= −(2 + λ)2 (λ − 4)
0 1 −1 1
R1 → −6 R1
∼ 1 −1 1
R2 ↔ R3
0 0 0
1 −1 1
∼ 0 1 −1 R1 ↔ R2
0 0 0
1 0 0
∼ 0 1 −1 R1 ↔ R1 + R2
0 0 0
Thus, eigen vector is given by
1 0 0 x 0
0 1 −1 y = 0
0 0 0 z 0
i.e, x = 0, y − z = 0 and so y = z, x = 0.
Take z = 1. Then, eigen vector associated to λ = 6 is given by
0
X2 = 1 .
1
Exercise
4.0.9 Prove that the eigen values of matrix
a h g
0 b 0 are a, b and c.
0 c c
98 CHAPTER 4. EIGEN VALUES AND EIGEN VECTORS
4.1 Properties
Proposition 4.1.1 Corresponding to each eigen vector X of a
square matrix, there exists a unique eigen value of A whereas cor-
responding to each eigen value there exists infinitely many eigen
vectors.
AX = λX (4.1.1)
X θ Aθ = λX θ (4.1.2)
and hence
X θ A = λX θ as Aθ = A
. Thus
(X θ A)X = (λX θ )X
i.e;
X θ (AX) = λX θ X
Using equation 4.1.1, we have λX θ X = λX θ X. But X θ X 6= 0 as
X 6= 0. Hence λ = λ, i.e; λ is purely real. 2
AX1 = λ1 X1 . . . (1)
AX2 = λ2 X2 . . . (2)
Aθ = A . . . (3)
Pre-multiplying (1) by X2θ and (2) by X1θ , we have
Proof: Assume that the result is not true. Then there exists
a natural number i such that {X1 , X2 , . . . , Xi } is linearly inde-
pendent but the set {X1 , X2 , . . . , Xi , Xi+1 } is linearly dependent.
Then there exists scalars c1 , c2 , . . . , ci+1 not all zero such that
AX1 = λ1 X1 . . . (1)
AX2 = λ2 X2 . . . (2)
θ
A A = In . . . (3)
Taking conjugated transpose of (2), we have
X2θ Aθ = λ2 X2θ . . . (4)
Post-multiplying (4) by AX1 , we have
X2θ Aθ AX1 = λ2 X2θ (AX1 ) = λ2 X2θ λ1 X1 = λ2 λ1 X2θ X1
Using (3), we have
X2θ X1 = λ2 λ1 X2θ X1
and so (λ2 λ1 − 1)X2θ X1 = 0, i.e; λ2 (λ1 − λ2 )X2θ X1 = 0 as 1 = λ2 λ2 .
But λ1 6= λ2 , λ2 6= 0, therefore X2θ X1 = 0. Thus, X1 , X2 are
orthogonal. 2
4.2. DIAGONALIZABLE MATRIX 103
+20 + 4λ
= (8 − λ) 5 − 10λ + λ2 − 40 + 40λ
8 −6 2
|A − 0I3 | = −6 7 −4
2 −4 3
104 CHAPTER 4. EIGEN VALUES AND EIGEN VECTORS
0 10 −10
R → R1 − 4R3
∼ 0 −5 5 1
R2 → R2 + 3R3
2 −4 3
0 0 0
∼ 0 −5 5 R1 → R1 + 2R2
2 −4 3
2 −4 3
∼ 0 −5 5 R1 ↔ R3
0 0 0
1 z
y = z, x = (4y − 3z) = ,
2 2
Then
6 − λ −2 2
|A − λI3 | = −2 3 − λ −1
2 −1 3 − λ
4.2. DIAGONALIZABLE MATRIX 107
6 − λ −2 2
= −2 3 − λ −1 R3 → R3 + R2
0 2−λ 2−λ
6 − λ −2 2
= (2 − λ) −2 3 − λ −1
0 1 1
= (2 − λ) [(6 − λ)(3 − λ + 1) + 2(−2 − 2)]
= (2 − λ)(16 − 10λ + λ2 )
= −(2 − λ)2 (λ − 8)
Case I: When λ = 2.
6 − 2 −2 2
A − 2I3 = −2 3 − 2 −1
2 −1 3 − 2
4 −2 2
= −2 1 −1
2 −1 1
4 −2 2
R2 → R2 + 21 R1
∼ 0 0 0
R3 → R3 − 21 R1
0 0 0
6 − 8 −2 2
A − 8I3 = −2 3 − 8 −1
2 −1 3 − 8
−2 −2 2
= −2 −5 −1
2 −1 −5
108 CHAPTER 4. EIGEN VALUES AND EIGEN VECTORS
−2 −2 2
R2 → R2 − R1
∼ 0 −3 −3
R3 → R3 + R1
0 −3 −3
−2 −2 2
∼ 0 −3 −3 R3 → R3 − R2
0 0 0
Thus, eigen value is given by −3y − 3z = 0 and −2x − 2y + 2z =
0. Solving these, we get
y = −z, x = 2z
Take z = 1, then y = −1 and x = 2. Hence, eigen vector is
2
X3 = −1
1
Here eigen vectors X1 , X2 , X3 are linearly independent.
Take
1 1 2
P = [X1 X2 X3 ] = 0 2 −1
−2 0 1
then
2 −1 5
1
P −1 = 2 5 1
12
4 −2 2
Now,
4 −2 −10
1
P −1 A = 4 10 2
12
4 −2 2
and hence
2 0 0
P −1 AP = 0 2 0
0 0 8
Remark 4.2.4 Here X1 , X2 , X3 are 3 linearly independent eigen
vectors associated to eigen values 2, 2, 8 of an 3×3 matrix A. Then,
we have a non-singular matrix P = [X1 X2 X3 ] such that
2 0 0
P −1 AP = 0 2 0
0 0 8
4.2. DIAGONALIZABLE MATRIX 109
D = diag[λ1 , λ2 , . . . , λn ],
P − 1AP = diag[λ1 , λ2 , . . . , λn ],
therefore
AP = P.diag[λ1 , λ2 , . . . , λn ]
AP = A[X1 , X2 , . . . , Xn ]
= [AX1 , AX2 , . . . , AXn ]
= [λ1 X1 , λ2 X2 , . . . , λn Xn ]
λ1 0 0 . . . 0
0 λ2 0 . . . 0
= [X1 , X2 , . . . , Xn ] .
. .. .. .. ..
. . . . .
0 0 0 . . . λn
= PD
This gives, P −1 AP = D which is diagonal, where P is non-singular
because its all columns are linearly independent. Thus, A is diag-
onalizable. 2
2 + ... + B n−1
(A − λIn ) B0 + B 1 λ + B 2 λ n−1 λ
= a0 + a1 λ + a2 λ2 + . . . + an λn In
i.e;
AB0 = a0 In
AB1 − B0 = a1 In
AB2 − B1 = a2 In
.. .. ..
. . .
ABn−1 − Bn−2 = an−1 In
−Bn = an In
AB0 = a0 In
2
A B1 − AB0 = a1 A
A3 B2 − A2 B1 = a2 A2
.. .. ..
. . .
An Bn−1 − An−1 Bn−2 = an−1 An−1
−An Bn = an An
0 = a0 In + a1 A + a2 A2 + . . . + an−1 An−1 + an An
a0 In + a1 A + a2 A2 + . . . + an−1 An−1 + an An = 0
λ2 − 3λ + 2 = 0 (4.3.4)
A2 − 3A + 2 I2 = 0
Given that
1 2
A= ,
0 2
2 1 6 3 6 2 0
then A = , 3A = and 2I2 = and so
0 4 0 6 0 2
2 1 6 3 6 2 0
A − 3A + 2I2 = − +
0 4 0 6 0 2
1−3+2 6−6+0
=
0−0+0 4−6+2
0 0
=
0 0
Thus,
A2 − 3A + 2I2 = 0 (4.3.5)
This shows that A satisfies its characteristic equation λ2 − 3λ + 2 =
0.
Observe that |A| = 2 6= 0. Hence, A is invertible. To obtain
A−1 , we multiply both sides the equation 4.3.5 by A−1 , we have
A − 3I2 + 2A−1 = 0
114 CHAPTER 4. EIGEN VALUES AND EIGEN VECTORS
Thus,
−1 1−3 2−0 −2 2
−2A = A − 3I2 = =
0−0 2−3 0 −1
1 −1
Hence, A−1 =
0 12
Next, by characteristic equation 4.3.5, we have
A2 = 3A − 2I2
Using this recursively, we have
A3 = A(3A − 2I2 )
= 3A2 − 2A
= 3(3A − 2I2 ) − 2A
= 7A − 6I2
A4 = 7A2 − 6A
= 7(3A − 2I2 ) − 6A
= 15A − 14I2
Hence
2A4 − 3A3 + 9A2 − 7A + 5I2 = 2(15A − 14I2 ) − 3(7A − 6I2 )
+9(3A − 2I2 ) − 7A + 5I2
= (30 − 21 + 27 − 7)A
+(−28 + 18 − 18 + 5)I2
= 29A − 23I2
2−λ −1 1
|A − λI3 | = −1 2 − λ −1
1 −1 2 − λ
= (2 − λ) (2 − λ)2 − 1 − (−1) [−(2 − λ) + 1]
+1 [1 − (2 − λ)]
= (2 − λ) 3 − 4λ + λ2 + λ − 1 + λ − 1
= −λ3 + 6λ2 − 9λ + 4
λ3 − 6λ2 + 9λ − 4 = 0
A3 − 6A2 + 9A − 4I3 = 0.
2 −1 1
Now, A = −1 2 −1 , hence
1 −1 2
2 −1 1 2 −1 1 6 −5 5
A2 = −1 2 −1 −1 2 −1 = −5 6 −5 .
1 −1 2 1 −1 2 5 −5 6
6 −5 5 2 −1 1
A3 = A2 A = −5 6 −5 −1 2 −1
5 −5 6 1 −1 2
22 −21 21
= −21 22 −21
21 −21 22
116 CHAPTER 4. EIGEN VALUES AND EIGEN VECTORS
Then,
22 −21 21 6 −5 5
A3 − 6A2 + 9A − 4I3 = −21 22 −21 − 6 −5
6 −5
21 −21 22 5 −5 6
2 −1 1 1 0
+9 −1 2 −1 − 4 0 1
0
1 −1 2 0 0 1
0 0 0
= 0 0 0
0 0 0
Thus, A satisfies its characteristic equation.
Since |A| = 6−1−1 = 4 6= 0, therefore A−1 exists. Multiplying
the characteristic equation by A−1 , we have
1 2
A−1 =
A − 6A + 9I3
4
6 −5 5 12 −6 6 9 0 0
1
= −5 6 −5 − −6 12 −6 + 0 9 0
4
5 −5 6 6 −6 12 0 0 9
3 1 −1
1
= 1 3 1
4
−1 1 3
Exercise 4.3.6 Find the characteristic equation of the matrix
1 0 2
0 2 1
2 0 3
Thus,
−3 0 2
A−1 = −1 1/2 1/2
2 0 −1
Also verify the Cayley -Hamilton theorem and hence find A−1 .
Exercises
Exercise 4.3.11 Find the eigen values and eigen vectors of the
matrix
1 2 0
2 −1 0
0 0 −1
Is it diagonalizable?
4.3. CAYLEY-HAMILTON THEOREM 119
Exercise 4.3.12 Verify the Cayley -Hamilton theorem for the ma-
trix
0 1 2
0 −3 0 .
1 1 −1
Also find A−1 .
2
120 CHAPTER 4. EIGEN VALUES AND EIGEN VECTORS
Chapter 5
Vector Spaces
121
122 CHAPTER 5. VECTOR SPACES
One may easily verify that this operation is well defined. This sum
is called as coordinate-wise addition. Next, let (x, y, z) ∈ R3 and
a ∈ R. Define · on R3 as follows
One may easily verify that this operation is well defined. This mul-
tiplication is called as coordinate-wise multiplication. Let (x1 , y1 , z1 ),
(x2 , y2 , z2 ) and (x3 , y3 , z3 ) be any three elements of R3 . Then,
of (x, y, z).
(3). addition is commutative, i.e;
and
a(x1 , x2 , . . . , xn ) = (ax1 , ax2 , . . . , axn )
for all (x1 , x2 , . . . , xn ), (y1 , y2 , . . . , yn ) ∈ F n and a ∈ F .
a0 + a1 x + a2 x2 + . . . + an xn + . . . ,
a0 + a1 x + a2 x2 + . . . + an xn + . . . + am xm + . . .
b0 + b1 x + b2 x2 + . . . + bn xn + . . . + bm xm + . . .
c0 + c1 x + c2 x2 + . . . + cn xn + . . . + cm xm + . . .
where cn = an + bn for all n = 0, 1, 2, . . .. and
a(a0 + a1 x + a2 x2 + . . . + an xn + . . .)
= aa0 + aa1 x + aa2 x2 + . . . . . . + aan xn + . . . +
a0 + a1 x + a2 x2 + . . . + an xn
Exercise 5.1.7 Prove that the set Pn (x) of all polynomials over
R (or C) in x of degree at most n, forms a vector space over R (or
C) with respect to addition + and · given by
(a0 + a1 x + a2 x2 + . . . + an xn ) + (b0 + b1 x + b2 x2 + . . . + bn xn )
= (a0 + b0 ) + (a1 + b1 )x + (a2 + b2 )x2 + . . . + (an + bn )xn
Exercise 5.1.8 Prove that the set of all real valued functions de-
fined on R forms a vector space over R w.r.t vector addition and
scalar multiplication given by (f +g)(x) = f (x)+g(x) and (af )(x) =
af (x) respectively, where a ∈ R. Also, prove the same for the set
C[a, b] of all real valued continuous functions defined on closed in-
terval [a, b], where a < b and a, b are real numbers.
Exercise 5.1.21 Prove that the set of all real symmetric (skew-
symmetric) matrices form a vector space over the field of real num-
bers.
v = a1 s1 + a2 s2 + . . . + an sn
2 −1 1 : 3
−1 1 1 : −1
(A | b) =
3
1 9 : 0
2 −3 −5 : −1
−1 1 1 : −1
2 −1 1 : 3
∼
3
R1 ↔ R2
1 9 : 0
2 −3 −5 : −1
−1 1 1 : −1
0 R2 → R2 + 2R1
1 3 : 1
∼ R3 → R3 + 3R1
0 4 12 : −3
R4 → R4 + 2R1
0 −1 −3 : −3
−1 1 1 : −1
0 1 3 : 1 R3 → R3 − 4R2
∼
0 0 0 : −7 R4 → R4 + R2
0 0 0 : −2
Since rank A = 2 6= rank (A|b) = 3, therefore the system of equa-
tions AX = b has no solutions. Thus, the given vector (3, −1, 0, −1) ∈
R4 is not a linear combination of vectors (2, −1, 3, 2), (−1, 1, 1, −3)
and (1, 1, 9, −5).
5.2. LINEAR DEPENDENCE AND LINEAR INDEPENDENCE131
a1 v1 + a2 v2 + . . . + an vn = 0 . . . (1)
2a + 0b + 2c = 0 (1)
a + b + 2c = 0 (2)
a + 2b + 3c = 0 (3)
5.2. LINEAR DEPENDENCE AND LINEAR INDEPENDENCE133
a1 v1 + a2 v2 + . . . + an vn = 0
implies a1 = 0, a2 = 0, . . . , an = 0,
i.e; a1 v1 +a2 v2 +. . .+an vn = 0 only if all the coefficients a1 , a2 , . . .,
an are 0.
Example 5.2.18 The set {(1, 0, 0), (0, 1, 0), (0, 0, 1)} is linearly in-
dependent subset of vector space R3 over R. For
Suppose that
a1 v1 + a2 v2 + . . . + an vn = 0 . . . (1)
a1 v1 + a2 v2 + . . . + ak vk = 0 . . . (2)
If k = 1 then a1 v1 = 0, which is a contradiction because a1 6= 0
and v1 6= 0. Thus, 2 ≤ k ≤ n. Now, ak 6= 0 so equation (2) can be
written as
−a1 −a2 −ak−1
vk = v1 + v2 + . . . + vk−1
ak ak ak
Thus proved. 2
av + a1 v1 + a2 v2 + . . . + an vn = 0 (1)
Example 5.3.2 Consider the set B = {(1, 0), (0, 1)}. Suppose that
Then, we have
(a, b) = (0, 0).
By equality of ordered pairs, we have a = 0, b = 0.
This shows that a(1, 0) + b(0, 1) = (0, 0) implies a = 0, b = 0.
Thus, B is linearly independent.
Since every (x, y) in R2 can be expressed as
Example 5.3.4 The vector space C over R has a basis {1, i}.
Example 5.3.5 Consider the set S = {(1, 0, 1), (0, 1, 0), (0, 1, 1)}
of vectors in R3 . We claim that L(S) = R3 . For,
let (x, y, z) ∈ R3 and
Thus, L(S) = R3 .
Next, by (2), it follows that
implies a = b = c = 0.
Thus, S is linearly independent. Since it spans R3 , therefore S
is a basis of R3 .
a1 v1 + a2 v2 + . . . + an vn = v . . . (∗)
is a basis of R3 .
Example 5.3.12 Since {(1, 0, 0), (0, 1, 0), (0, 0, 1)} spans R3 (R),
therefore R3 (R) is a finite dimensional vector spaces.
Example 5.3.14 The vector spaces C(Q) and R(Q) are infinite
dimensional vector spaces (why?).
n≥m (1)
m≥n (2)
Example 5.3.18 The basis {(1, 0, 0), (0, 1, 0), (0, 0, 1)} of R3 con-
tains 3 elements, thus dimF V = 3.
Example 5.3.20 Let SS(n) be the set all n×n real skew-symmetric
matrices over R. Let Eij be n × n matrix whose (i, j)-th entry is
1 and (j, i)-th entry is −1 and all other remaining entries are 0,
where 1 ≤ i, j ≤ n. Then {Eij | i < j, 1 ≤ i, j ≤ n} is a basis of
SS(n) containing n(n−1)
2 elements. Thus, dimension of this space
n(n−1)
is 2 .
5.3. BASES AND DIMENSIONS 141
S1 = {w1 , w2 , . . . , wm ; v1 , v2 , . . . , vn }.
v = a1 v1 + a2 v2 + . . . + an vn . . . (1)
v = b1 v1 + b2 v2 + . . . + bn vn . . . (2)
a1 v1 + a2 v2 + . . . + an vn = b1 v1 + b2 v2 + . . . + bn vn
144 CHAPTER 5. VECTOR SPACES
This gives
a1 = b1 , a2 = b2 , . . . , an = bn .
If x = 0 then
W = {(0, ky) | k ∈ R},
which is the y-axis given by x = 0.
Next, if x 6= 0, then xy will be constant, say m. Then y = mx
and so
W = {(kx, k(mx)) | k ∈ R},
which represents a line passing through origin having slope m. In-
deed, every line passing through origin determines one dimensional
subspaces, therefore there are infinitely many one dimensional vec-
tor subspaces of R2 .
If dimR R2 = 2, then W = R2 .
W1 + W2 = {w1 + w2 | w1 ∈ W1 , w2 ∈ W2 }
W1 + W2 = {(x, 0, z) | x ∈ R}.
a1 v1 + a2 v2 + . . . + ak vk + b1 u1 + b2 u2 + . . .
+ . . . + bl−k ul−k + c1 t1 + c2 t2 + . . . + cm−k tm−k = 0 (1)
Then
a1 v1 + a2 v2 + . . . + ak vk + b1 u1 + b2 u2 + . . . + bl−k ul−k
= −c1 t1 − c2 t2 + . . . − cm−k tm−k (2)
i.e;
c1 t1 + c2 t2 + . . . + cm−k tm−k + d1 v1 + d2 v2 + . . . + dk vk = 0
a1 v1 + a2 v2 + . . . + ak vk + b1 u1 + b2 u2 + . . . bl−k ul−k = 0.
148 CHAPTER 5. VECTOR SPACES
dimF (W1 + W2 ) = k + (l − k) + (m − k) = l + m − k
i.e;
W1 = {(x, y, 0); x, y ∈ R}
and z-axis
W2 = {(0, 0, z); z ∈ R},
i.e; R3 = W1 ⊕ W2 .
w1 + w2 = u1 + u2
B = {w1 , w2 , . . . , wm ; v1 , v2 , . . . , vn−m }
u∼v ⇔u−v ∈W
{u ∈ V ; u ∼ v} = {u ∈ V ; u − v ∈ W }
= {u ∈ V ; u = w + v f or some w ∈ W }
= {w + v; w ∈ W }
v + W = {v + w; w ∈ W }
V /W = {W + v | v ∈ V }
(W + u) + (W + v) = W + (u + v) and a(W + u) = W + au
(W + u) + (W + v) = (W + u0 ) + (W + v 0 )
au − a0 u0 = au − au0 = a(u − u0 ) ∈ W.
152 CHAPTER 5. VECTOR SPACES
(W + u) + (W + v) = W + (u + v) and a(W + u) = W + au
a1 (W + v1 ) + a2 (W + v2 ) + . . . + an−m (W + vn−m ) = W
a1 v1 + a2 v2 + . . . + an−m vn−m = b1 w1 + b2 w2 + . . . + bm wm
Then
a1 v1 + a2 v2 + . . . + an−m vn−m − b1 w1 − b2 w2 − . . . − bm wm = 0
5.6. COORDINATES 153
v = a1 v1 + a2 v2 + . . . + an−m vn−m + b1 w1 + b2 w2 + . . . + bm wm
W + v = a1 (W + v1 ) + a2 (W + v2 ) + . . . + an−m (W + vn−m )
5.6 Coordinates
Definition 5.6.1 A finite sequence v1 , v2 , . . . , vn of vectors in an
n-dimensional vector space V (F ) is called an ordered basis of V .
It is denoted by n-tuple (v1 , v2 , . . . , vn ).
B = (v1 , v2 , . . . , vn )
The scalar xi is P
called the i-th coordinate of v relative to B.
Note that if v = ni=1 xi vi and w = ni=1 yi vi then
P
n
X n
X
av = (axi )vi , v+w = (xi + yi )vi
i=1 i=1
.
154 CHAPTER 5. VECTOR SPACES
n
X
wj = Pij vi = P1j v1 + P2j v2 + . . . + Pnj vn . . . (1)
i=1
Then,
P1j
P2j
[wj ]B = . f or each 1 ≤ j ≤ n
.
.
Pnj
Then
v = y1 w1 + y2 w2 + . . . + yn wn
Xn
= yj wj
j=1
n n
!
X X
= yj Pij vi
j=1 i=1
n
X Xn
= Pij yj vi . . . (3)
i=1 j=1
Now
[v]B = 0 ⇔ xi = 0 f or all 1 ≤ i ≤ n
Xn
⇔ Pij yj = 0 (by Eq.(4))
j=1
n
X n
X
⇔ y1 w1 + y2 w2 + . . . + yn wn = Pij yj vi
i=1 j=1
= 0 (by Eq.(3))
⇔ yj = 0, ∀1 ≤ j ≤ n (as B 0 is linearly independent)
⇔ [v]B0 = 0
[v]B0 = P −1 [v]B
B 0 = BP
v = y1 w1 + y2 w2 + . . . + yn wn
Xn
= yj wj
j=1
n n
!
X X
= yj Pij vi
j=1 i=1
n
X Xn
= Pij yj vi . . . (3)
i=1 j=1
Hence
x01
x1 cos θ + x2 sin θ
=
x02 −x1 sin θ + x2 cos θ
Thus, x01 = x1 cos θ + x2 sin θ and x02 = −x1 sin θ + x2 cos θ
0 0 18
Then coordinates of any vector (a, b, c) relative to this basis is
−1 2 11 −a + 2b + 11c
a 8 a 8
P −1 b = 0 12 16 3
b = b 3c
2 + 16
1 c
c 0 0 8 c 8
Exercises
160 CHAPTER 5. VECTOR SPACES
0 1 4 2
2
162 CHAPTER 5. VECTOR SPACES
Chapter 6
Linear Transformations
Definition 6.0.1 Let U and V be any two vector spaces over same
field F . Then a map T : U → V is called a linear transforma-
tion if it satisfies the following condition:
T (X) = XA.
T (aX + bY ) = (aX + bY )A
= a(XA) + b(Y A) (by right distributive law)
= aT (X) + bT (Y )
163
164 CHAPTER 6. LINEAR TRANSFORMATIONS
T [a(x1 , y1 , z1 ) + b(x2 , y2 , z2 )]
= T [(ax1 , ay1 , az1 ) + (bx2 , by2 , bz2 )]
= T [(ax1 + bx2 , ay1 + by2 , az1 + bz2 )]
= (2[ax1 + bx2 ], 3[ay1 + by2 ])
= (2ax1 + 2bx2 , 3ay1 + 3by2 )
= (2ax1 , 3ay1 ) + (2bx2 + 3by2 )
= a(2x1 , 3y1 ) + b(2x2 , 3y2 )
= aT (x1 , y1 , z1 ) + bT (x2 , y2 , z2 )
T (x, y) = (x − y, x + y, 2y)
T [a(x1 , y1 ) + b(x2 , y2 )]
= T [(ax1 , ay1 ) + (bx2 , by2 )]
= T [(ax1 + bx2 , ay1 + by2 )]
= (ax1 + bx2 − ay1 − by2 , ax1 + bx2 + ay1 + by2 , 2[ay1 + by2 ])
= (a(x1 − y1 ) + b(x2 − y2 ), a(x1 + y1 ) + b(x2 + y2 ), 2ay1 + 2by2 )
= (a(x1 − y1 ), a(x1 + y1 ), 2ay1 ) + (b(x2 − y2 ), b(x2 + y2 ), 2by2 )
= a (x1 − y1 , x1 + y1 , 2y1 ) + b (x2 − y2 , x2 + y2 , 2y2 )
= aT (x1 , y1 ) + bT (x2 , y2 )
pi (x1 , x2 , . . . , xn ) = xi
is a linear transformation.
Exercise 6.0.9 Let Mn×n (R) denotes the vector space of all n ×
n over R. Let B ∈ Mn×n (R) be a non-zero matrix. Prove the
following :
Exercise 6.0.10 Prove that the map T : Pn (x) → Pn+1 (x) defined
by T (p(x)) = p(0) + xp(x) is a linear transformation but the map
T1 : Pn (x) → Pn+1 (x) defined by T1 (p(x)) = 1 + xp(x) is not a
linear transformation.
v = a1 v1 + a2 v2 + . . . + an vn
166 CHAPTER 6. LINEAR TRANSFORMATIONS
a = x, b = y, c = z − x . . . (3)
T (U ) = {T (u) | u ∈ U }
T −1 (W ) = {u ∈ U | T (u) ∈ W }
C1 x1 + C2 x2 + . . . + Cn xn = b
ν(TA ) = n − rank A.
Since
1 0 −1 0
A =
−1 1 0 0
1 0 −1 0
∼ R2 → R2 + R1
0 1 −1 0
Exercise 6.0.24 Find the basis of range space and null space of
the linear transformation T : R4 → R3 given by
T (x, y, z, u) = (x + y, y − z, x + u).
Exercise 6.0.25 Find the basis of range space and null space of
the linear transformation T : R3 → R3 given by
T (x, y, z) = (x + y, y + z, x + z).
T (x, y, z) = (x − y, y + z, x + z) = (a, b, c)
range T = {(a, b, a + b) | a, b ∈ R}
x − y = 0, y + z = 0, x+z =0
⇔ the system
T is surjective AX = 0 is consistent, where A =
1 −1 0 x
0 1 1 and X = y = (x, y, z)t . From above reduction,
1 0 1 z
we see that rank A = 2, hence AX = 0 has 3 − 2 = 1 linearly
independent solution and so ker T 6= {0}. Thus, T is not injective.
i.e;
dim V = ρ(T ) + ν(T )
v = a1 v1 + a2 v2 + . . . + am vm + am+1 vm+1 + . . . + an vn
Then
T (am+1 vm+1 + . . . + an vn ) = 0
Hence am+1 vm+1 + . . . + an vn ∈ ker T . But S is a basis of ker T .
Hence there exists scalars a1 , a2 , . . . , am ∈ F such that
am+1 vm+1 + . . . + an vn = a1 v1 + a2 v2 + . . . + am vm
i.e;
am+1 vm+1 + . . . + an vn − a1 v1 − a2 v2 − . . . − am vm = 0
i.e;
dim V = n = dim T (V ) + dim ker T
But ρ(T ) = dim T (V ) and ν(T ) = dim ker T , therefore we have
v = a1 v1 + a2 v2 + . . . + an vn
T (a1 v1 + a2 v2 + . . . + an vn ) = (a1 , a2 , . . . , an )
Xn n
X
T [a( ai vi ) + b( bi vi )]
i=1 i=1
n
X
= T [ (aai + bbi )vi ]
i=1
= (aa1 + bb1 , aa2 + bb2 , . . . , aan + bbn )
= a(a1 , a2 , . . . , an ) + b(b1 , b2 , . . . , bn )
X n X n
= aT ( ai vi ) + bT ( bi vi )
i=1 i=1
Thus, T is linear.
Hence T is an isomorphism, i.e; V ' F n . 2
Hence
Let a, b ∈ F . Then
" n n
# " n #
X X X
T a. ai vi + b. bi vi = T (aai + bbi )vi
i=1 i=1 i=1
n
X
= (aai + bbi )wi
i=1
n
X n
X
= a. ai wi + b. bi wi
i=1 i=1
" n # " n
#
X X
= aT ai vi + bT bi vi
i=1 i=1
T (vi ) = T 0 + 0 + . . . + vi +0 + . . . + 0
|{z}
i−th place
= 0 + 0 + ... + wi +0 + . . . + 0
|{z}
i−th place
= wi
6.1. RANK-NULLITY THEOREM 179
for each i = 1, 2, . . . , n.
Let U be a linear transformation from V into W such that
U (vi ) = wi for each i = 1, 2, .., n. Then,
n n n n
! !
X X X X
U ai vi = ai U (vi ) = a i wi = T ai vi
i=1 i=1 i=1 i=1
Thus, U = T . 2
Take v3 = (0, 0, 1) ∈
/ L({(1, −1, 1), (1, 1, 1)}). Suppose that T (v3 ) =
(0, 0). Then T is a linear transformation which is described as be-
low:
Let
Then
a+b = x
−a + b = y
a+c = z
i.e;
(1, 1) = −(1, 0) − (0, 1) = (−1, −1)
and so 1 = −1. This is a contradiction. Thus, there exists no linear
transformation such that T (vi ) = wi for all i = 1, 2, 3
and
(cT )(v) = c.T (v)
is a vector space over field F .
(T1 + T2 ) + T3 = T1 + (T2 + T3 )
Similarly we have the following:
m X
n m X
n
" #
X X h i
ik
Aik T (vj ) = Aik T ik (vj )
i=1 k=1 i=1 k=1
m
" n #
X X
= Aik δjk wi (by (1))
i=1 k=1
6.2. ALGEBRA OF LINEAR TRANSFORMATIONS 183
i.e;
m X
n m
" #
X X
ik
Aik T (vj ) = Aij wi . . . (2)
i=1 k=1 i=1
= T (vj ) (by Eq. 6.2.1)
for all j.
By linear independence of B 0 , Aij = 0 for all i and j. Thus,
S is linearly independent subset of L(V, W ). Then S is a basis of
L(V, W ) and dim L(V, W ) = mn. 2
U oT (v) = U (T (v))
1. IU = U = U I;
2. (U T1 )T2 = U (T1 T2 );
3. U (T1 + T2 ) = U T1 + U T2 ; (T1 + T2 )U = T1 U + T2 U ;
f (x1 , x2 , . . . , xn ) = a1 x1 + a2 x2 + . . . + an xn
Then,
n
X
aj fj (vi ) = 0̂(vi ) = 0 ∀i = 1, 2, ..., n
j=1
i.e;
n
X
aj δji = 0̂(vi ) = 0 ∀i = 1, 2, ..., n
j=1
186 CHAPTER 6. LINEAR TRANSFORMATIONS
i.e;
ai = 0 f or all i = 1, 2, ..., n
Thus, eq. (1) holds only if all the coefficients are zero. Hence B ∗
is linearly independent and so B ∗ is a basis of V ∗ . Thus, dim V ∗ =
n = dim V . 2
Proof: Since any two finite dimensional vector spaces over same
field are isomorphic if and only if they have same dimension. Thus,
to prove the result it is sufficient to prove dim V = dim V ∗ . Thus
the result follows from the Theorem 6.3.4. 2
v = fi (v)vi (6.3.1)
Pn
Next, let v ∈ V . Then v = i=1 ai vi and so
n
X n
X
fj (v) = ai fj (vi ) = ai δij = aj
i=1 i=1
This gives
a + b + 2c = x
b + 2c = y
−a + b = z
f1 (x, y, z) = a = x − y
f2 (x, y, z) = b = x − y + z
−x + 2y − z
f3 (x, y, z) = c =
2
Alternate method: We consider the augmented matrix
1 1 2 : x 1 0 0 : x−y
(A|B) = 0 1 2 : y ∼ 0 1 2 : y
−1 1 0 : z −1 1 0 : z
1 0 0 : x−y 1 0 0 : x−y
∼ 0 1 2 : y ∼ 0 0 2 : −x + 2y − z
0 1 0 : x−y+z 0 1 0 : x−y+z
Hence, solution to equation (1) is a = x − y, b = x − y + z, c =
(−x + 2y − z)/2
Example 6.3.10 Consider the vector space R3 (R) with basis {v1 , v2 , v3 },
where v1 = (1, 0, 1), v2 = (0, 1, −2) and v3 = (−1, −1, 0). Suppose
that f is a linear functional such that f (v1 ) = 1, f (v2 ) = −1 and
f (v3 ) = 3 and we wish to determine f (a, b, c). Suppose that
x = 2a − 2b − c, y = a − b − c, z = a − 2b − c . . . (2)
f (a, b, c) = 4a − 7b − 3c
of R3 (R).
Answer: {f1 , f2 , f3 }, where f1 (x, y, z) = −3x−5y−2z, f2 (x, y, z) =
2x + y and f3 (x, y, z) = x + 2y + z.
of R3 (R).
Answer: {f1 , f2 , f3 }, where f1 (x, y, z) = (y + z)/2, f2 (x, y, z) =
(x + z)/2 and f3 (x, y, z) = (x + y)/2.
of R3 (R).
Solution:
1 0 0 : x
(A|B) = −1 1 3 : y
3 −1 −2 : z
1 0 0 : x
∼ 0 1 3 : x+y R2 → R2 + R1 , R3 → R3 − 3R1
0 −1 −2 : −3x + z
190 CHAPTER 6. LINEAR TRANSFORMATIONS
1 0 0 : x
∼ 0 1 3 : x+y R3 → R3 + R1
0 0 1 : −2x + y + z
1 0 0 : x
∼ 0 1 0 : 7x − 2y − 3z R2 → R2 − 3R3
0 0 1 : −2x + y + z
6.4 Annihilators
Definition 6.4.1 If V is a vector space ove field F and S is a
subset of V . Then the annihilator of S is the set of all linear
functionals f on V such that f (s) = 0 for all s ∈ S. It is denoted
by S 0 . Thus,
S 0 = {f ∈ V ∗ | f (s) = 0∀s ∈ S}
{0}0 = {f ∈ V ∗ |f (0) = 0} = V ∗
1. A ⊂ B ⇒ B 0 ⊂ A0 .
Hence proved. 2
As we have observed in the Example 6.1.2 (section of Rank-
Nullity theorem) that every non -zero functional f determines a
hyperspace of dimension n − 1. Thus, hyperspace corresponding to
each fj , m + 1 ≤ j ≤ n is of dimension n − 1.
x1 + 2x2 + 2x3 + x4 = 0
2x2 + x4 = 0
−2x1 − 4x3 + 3x4 = 0
1 2 2 4
Consider the coefficient matrix A = 0 2 0 1. Now
−2 0 −4 3
1 2 2 4
A = 0 2 0 1
−2 0 −4 3
1 2 2 4
∼ 0 2 0 1 R3 → R3 + 2R1
0 4 0 11
1 2 2 4
∼ 0 2 0 1 R3 → R3 − 2R2
0 0 0 9
1 0 2 3
R1 → R1 − R2
∼ 0 2 0 1
R3 → 19 R3
0 0 0 1
1 0 2 0
R1 → R1 − 3R3
∼ 0 2 0 0
R2 → R2 − R3
0 0 0 1
From this it follows that the common solution is x4 = 0, x2 = 0 and
x1 + 2x3 = 0, i.e; x1 = −2x3 . Thus, W = {(−2x3 , 0, x3 , 0) | x3 ∈
R}. 2
From the above example it also follows that
h{f1 , f2 , f3 }i = h{g1 , g2 , g3 }i
where
g1 (x1 , x2 , x3 , x4 ) = x1 + 2x3
g2 (x1 , x2 , x3 , x4 ) = 2x2
g3 (x1 , x2 , x3 , x4 ) = x4
6.4. ANNIHILATORS 195
∗
Solution: Let f ∈ W 0 then f ∈ R5 and so it will be of the
form
f (x1 , x2 , x3 , x4 , x5 ) = c1 x1 + c2 x2 + . . . + c5 x5 (1)
c 0
4 −1 1
2 −2 3
−1 1 2 5 2
c
0
2
c3 = 0
0 0 −1 −2 3
c4 0
1 −1 2 3 0
c5 0
But
2 −2 3 4 −1
−1 1 2 5 2
0 0 −1 −2 3
1 −1 2 3 0
1 −1 2 3 0
−1 1 2 5 2
∼ R1 ↔ R4
0 0 −1 −2 3
2 −2 3 4 −1
1 −1 2 3 0
0 0 4 8 2 R2 → R2 + R1
∼
0 0 −1 −2 3
R4 → R4 − 2R1
0 0 −1 −2 −1
196 CHAPTER 6. LINEAR TRANSFORMATIONS
1 −1 2 3 0
0 0 1 2 21 R2 → 14 R2
∼
0 0 −1 −2 3 R4 → R4 − R3
0 0 0 0 4
1 −1 2 3 0
0 0 1 2 21
∼ R3 → R3 + R2
0 0 0 0 72
0 0 0 0 1
1 −1 2 3 0
0 0 1 2 21
∼
0
R3 → 27 R3
0 0 0 1
0 0 0 0 1
1 −1 2 3 0
0 0 1 2 0 R2 → R2 − 12 R3
∼
0 0 0 0 1 R4 → R4 − R3
0 0 0 0 0
1 −1 0 −1 0
0 0 1 2 0
∼
0
R1 → R1 − 2R2
0 0 0 1
0 0 0 0 0
and
2
6.4. ANNIHILATORS 197
Lv (f ) = f (v)
L(v) = Lv
Then
W 00 = {v ∈ V | f (v) = 0∀ f ∈ W 0 }
Let w ∈ W then f (w) = 0 for all f ∈ W 0 and so w ∈ W 00 . Thus,
W ⊆ W 00
But
dim W + dim W 0 = dim V
dim W 0 + dim W 00 = dim V ∗
and dim V = dim V ∗ . Therefore dim W = dim W 00 . Since W ⊆
W 00 and hence W = W 00 . 2
6.4. ANNIHILATORS 199
(W10 + W20 )0 = W1 ∩ W2
i.e;
W10 + W20 = (W1 ∩ W2 )0
2
Exercises
Exercise 6.4.19 Determine which of the following maps T : R3 →
R3 are linear:
1. T (x, y, z) = (y, z, 0).
2. T (x, y, z) = (x − y, y + z, x).
3. T (x, y, z) = (x − y − z, x + y + z, z).
4. T (x, y, z) = (x − y, z, |x|).
Exercise 6.4.20 Let Mn×n (R) denotes the vector space of all n ×
n over R. Let B ∈ Mn×n (R) be a non-zero matrix. Prove that the
map T : Mn×n (R) → Mn×n (R) defined by T (X) = XB + BX is
a linear transformation.
Exercise 6.4.23 Find the basis of range space and null space of
the linear transformation T : R4 → R3 given by
T (x, y, z, u) = (x + y, y − z, x).
Exercise 6.4.24 Find the basis of range space and null space of
the linear transformation T : R3 → R3 given by
T (x, y, z) = (x − y, y − z, x − z).
Is T invertible? If so, find a rule for T −1 like the one which defines
T.
2
Chapter 7
Matrix Representations
201
202 CHAPTER 7. MATRIX REPRESENTATIONS
Let v = x1 v1 + x2 v2 + . . . + xn vn in V . Then
Xn
T (v) = T xj vj
j=1
n
X
= xj T (vj )
j=1
Xn m
X
= xj Aij wi
j=1 i=1
i.e;
m
X Xn
T (v) = Aij xj wi (7.0.4)
i=1 j=1
Thus,
Pn
A1j xj
Pj=1
nj=1 A2j xj
Pn
[T (v)]B0 = j=1 A3j xj (7.0.5)
..
Pn .
j=1 Amj xj
and so
A11 A12 ... A1n x1
A21 A22 ... A2n x2
[T (v)]B0 = .
.. .. .. ..
.. . . . .
Am1 Am2 . . . Amn xn
i.e;
Hence,
[aT ]B0 = a.[T ]B,B0 [v]B i.e; [aT ]B,B0 = a.[T ]B,B0
i.e;
(T + U )B,B0 = [T ]B,B0 + [U ]B,B0
Thus,
for all a, b ∈ F .
Hence, we have the following:
Hence
1 1
[T (1, 0, 0)]B = , [T (0, 1, 0)]B =
0 −1
and
1
[T (0, 0, 1)]B =
0
Thus, matrix of T relative to B and B 0 is
1 1 1
[T ]B,B0 =
0 −1 0
T (v) = x1 w1 + x2 w2 + . . . + xm wm
for general vector v, i.e; first we determine the general formula for
T (v) in terms of basis B 0 .
After determining these coefficients, we evaluate it for basis
vectors v = vj , 1 ≤ j ≤ n and then determine [T (vj )]B0 . Thus,
−a − 3b + 5c −a + b + c a+b−c
T (x, y, z) = (a, b, c) = v1 + v2 + v3
4 4 2
Hence, we have
17 −3 −1
T (v1 ) = T (1, 0, 1) = (4, −2, 3) = v1 + v2 + v3
4 4 2
35 15 −7
T (v2 ) = T (−1, 2, 1) = (−2, 4, 9) = v1 + v2 + v3
4 4 2
22 −6
T (v3 ) = T (2, 1, 1) = (7, −3, 4) = v1 + v2 + 0.v3
4 4
and so 17 35
4 4
−3 15
[T (v1 )]B =
4
, [T (v2 )]B =
4
,
−1 −7
2 2
and 22
4
−6
[T (v3 )]B =
4
0
206 CHAPTER 7. MATRIX REPRESENTATIONS
Thus,
17 35 22
4 4 4
−3 15 −6
[T ]B = [[T (v1 )]B [T (v2 )]B [T (v3 )]B ] =
4 4 4
−1 −7
2 2 0
Then
1 −1 2 : a
(A|B) = 0 2 1 : b
1 1 1 : c
1 −1 2 : a
∼ 0 2 1 : b R3 → R3 − R1
0 2 −1 : c − a
1 −1 2 : a
∼ 0 2 1 : b R3 → R3 − R2
0 0 −2 : c − a − b
T (x, y) = (−y, x)
T (x1 , x2 , x3 ) = (x1 + x2 , x2 + x3 , x1 + x3 )
T (x, y, z) = (x + z, −2x + y, −x + 2y + z)
This gives
[T ]B P = P [T ]B0
and so
[T ]B0 = P −1 [T ]B P.
Next, let U be a linear operator on V such that U (vjP
) = wj for
all j = 1, 2, .., n, thenPit will be invertible. But wj = ni=1 Pij vi
and so U (vj ) = wj = ni=1 Pij vi . Thus,
[U ]B = P
Then
Hence
1 −5
[T ]B0 = P −1 [T ]B P =
0 −1
Exercises
Exercise 7.0.20 Find the matrix of linear operator T on R3 de-
fined by
with respect to the ordered basis B and also with respect to B 0 , where
(1) B = {(1, 0, 0), (0, 1, 0), (0, 0, 1)},
(2) B 0 = {(1, 1, 1), (1, 1, 0), (1, 0, 0)}.
Exercise 2
7.0.22 If the matrix of T on C relative to its standard
1 1
basis is , what is the matrix of T relative to ordered basis
1 1
B 0 = {(1, 1), (1, −1)}?
2
214 CHAPTER 7. MATRIX REPRESENTATIONS
Chapter 8
.
(ii). (Conjugate - Symmetry).
.
(iii). (Non- negativity).
215
216 CHAPTER 8. INNER PRODUCT SPACES
hu, av + bwi = h av + bw , ui
= ah v, ui + bhw , ui
= ah v, ui + bhw , ui
= ahu, vi + bhu, wi
hx, yi = x1 y1 + x2 y2 + · · · + xn yn
n
X
= x i yi
i=1
hax + by, zi
= h(ax1 + by1 , ax2 + by2 , . . . , axn + byn ) , (z1 , z2 , . . . , zn )i
Xn
= (axi + byi )zi
i=1
Xn n
X
= a xi z i + b yi zi
i=1 i=1
= ahx, zi + bhy, zi
hx, yi = x1 y1 + x2 y2 + 2x3 y3 + x2 y3 + x3 y2
hx, yi = x1 y1 − x1 y2 − x2 y1 + 4 x2 y2 (8.1.7)
Example 8.1.9 Let V denote the complex vector space of all com-
plex valued continuous functions on [0, 1]. Define h , i by
Z 1
h f, gi = f (t)g(t)dt. (8.1.8)
0
ku − vk ≥ 0; ku − vk = 0 ⇔ u = v (8.1.10)
8.1. INNER PRODUCTS 219
and
ku − vk = kv − uk (8.1.11)
Proof: If y = 0, then
and
kxk kyk = kxk k0k = 0
and so equality holds . Assume that y 6= 0 and so kyk 6= 0.
hy, xi
Consider the vector x − αy, where α = hy, yi . By non-negativity
property,
h x − αy, x − αyi ≥ 0
Using linearity property, we have
hy, xi
Putting α = hy, yi in the above equation and using that hy, xi =
hx, yi, we have
h x, yih x, yi
hx, xi − ≥ 0
hy, yi
that is;
h x, yih x, yi
≤ hx, xi
hy, yi
Thus
|hx, yi| ≤ kxk kyk .
If {x, y} is linearly independent, then x − αy 6= 0 for all
α ∈ F and so kx − αyk > 0. Hence
2
Using Cauchy-Schwarz inequality in above examples, we have
several important inequalities (try).
|| x + y || ≤ || x || + || y ||
(kx + yk)2 = hx + y, x + yi
= hx, xi + hx, yi + hy, xi + hy, yi
≤ kxk2 + hx, yi + hx, yi + kyk2
≤ kxk2 + 2 kxk kyk + kyk2
(by Cauchy Schwarz inequality)
= (kxk + kyk)2
|| x + y || ≤ || x || + || y ||
i.e;
Re hx, yi = kxk · kyk
But Rehx, yi ≤ |hx, yi| and so kxk · kyk ≤ |hx, yi|. Using
Cauchy- Schwarz inequality, we have kxk · kyk = |hx, yi|. Thus,
by Theorem 8.1.13, the set {x, y} is linearly dependent. 2
8.2. NOTION OF ANGLE AND ORTHOGONALITY 221
Remark 8.1.15 Converse of the above result is not true; that is,
if {x, y} is linearly dependent then it is not necessary that
|| x + y || = || x || + || y ||
|| u − w || ≤ || u − v || + || v − w || (8.1.12)
<x, y>
cos θ = ||x|| ||y|| .
kx − yk2 = hx − y, x − yi
= hx, xi − hx, yi − hy, xi + hy, yi
= kxk2 + kyk2 − 2 hx, yi
1. (Parallelogram Law)
for all x, y ∈ V .
8.3. ORTHONORMAL SETS AND BESSEL’S INEQUALITY223
2. (Polarization identity). hx, yi = 1
4 kx + yk2 − kx − yk2
hα1 x1 + α2 x2 + · · · + αn xn , xm i = h0, xm i = 0, ∀m = 1, 2, . . . n.
hx, xi i = hα1 x1 + α2 x2 + · · · + αn xn , xi i
= αi , (since hxj , xi i = 0 f or all j 6= i)
n
* +
X
x− ai xi , xj = 0
i=1
n
X
⇒ hx, xj i − ai hxi , xj i = 0
i=1
⇒ hx, xj i − aj = 0
Thus aj = hx, xj i. 2
8.3. ORTHONORMAL SETS AND BESSEL’S INEQUALITY225
for all x ∈ V .
Thus
r
X
2
kxk ≥ |hx, xi i|2
i=1
x = a1 x1 + a2 x2 + · · · + an xn
226 CHAPTER 8. INNER PRODUCT SPACES
n n
* +
X X
x − hx, xi i xi , x − hx, xi i xi = 0
i=1 i=1
Then
n
X
x − hx, x i x i
= 0
i
i=1
and so
n
X
x− hx, xi i xi = 0
i=1
(−2,1,1)
Thus, y2 = √
6
. Now,
i.e; (2, 1, 1) = < (2, 1, 1), y1 > y1 + < (2, 1, 1), y2 > y2
and so q √
1+i 1+i 5 u2
+ 1.1̄ + 1−i
1−i
ku2 k = 2 2 2 2 = 2 . Take w2 = ku2 k .
Then
1+i 1−i
w2 = √ , 1, √ .
5 5
Clearly L({v1 , v2 }) = L({w1 , w2 }). Thus the required orthonormal
basis of subspaces spanned by (1, 0, i), (2, 1, 1+i) is { √12 , 0, √i2 , 1+i
√ , 1, 1−i
5
√
5
}.2
Exercises
1. Show that we can always define an inner product on a finite
dimensional vector space F n (F ), where F = R or C.
Exercise 8.4.16 Given the basis (2, 0, 1), (3, −1, 5) and
(0, 4, 2) for R3 , construct from it by the Gram-Schmidt process an
orthonormal basis of R3 relative to standard inner product.
Chapter 9
f (X, Y ) = X t AY
233
234 CHAPTER 9. BILINEAR AND QUADRATIC FORMS
and so
f (X, Y ) = X t AY
m m
! !
X X
= xi Ai1 y1 + xi Ai2 y2 + . . .
i=1 i=1
m
!
X
+ xi Ain yn
i=1
n m
!
X X
= xi Aij yj
j=1 i=1
n
XX m
= Aij xi yj
j=1 i=1
Xm X n
= Aij xi yj
i=1 j=1
P
f v1 , nj=1 yj wj
P
f v2 , nj=1 yj wj
= (x1 , x2 , . . . , xm )
..
.
Pn
f vm , j=1 yj wj
Pn
f (v1 , wj )yj
Pj=1
n f (v2 , wj )yj
j=1
= (x1 , x2 , . . . , xm )
..
Pn .
j=1 f (vm , wj )yj
Pn
j=1 a1j yj
n a2j yj
P
j=1
= (x1 , x2 , . . . , xm )
..
Pn .
j=1 amj yj
236 CHAPTER 9. BILINEAR AND QUADRATIC FORMS
a11 a12 ... a1n y1
a21 a22 ... a2n y2
= (x1 , x2 , . . . , xm ) .
.. .. .. ..
.. . . . .
am1 am2 . . . amn yn
= [v]tB A[w]B0
aij
= f (vi , wj )
0
0
..
b11 b12
. . . b1n .
b21 b22 0
. . . b2n
= (0, 0, .., 0, |{z}
1 , 0, .., 0) . 1
.. .. ..
.. . . . j−|{z}
i− place place
bm1 bm2 . . . bmn 0
.
. .
0
b1j
b2j
..
.
= (0, 0, . . . , 0, |{z}
1 , 0, . . . , 0)
bij
i−th place
..
.
bmj
= bij
One may easily verify that f is a bilinear form such that f (vi , wj ) =
aij . Next, if g is any other bilinear form such that g(vi , wj ) = aij .
9.1. BILINEAR FORMS 237
Pm Pn
Let v = i=1 xi vi and w = j=1 yj wj . Then
Xm n
X m X
X n
g(v, w) = g xi vi , y j wj = xi g(vi , wj )yj
i=1 j=1 i=1 j=1
m X
X n
= xi aij yj
i=1 j=1
Hence
a11 a12 ... a1n y1
a21 a22 ... a2n y2
g(v, w) = (x1 , x2 , . . . , xm ) .
.. .. .. ..
.. . . . .
am1 am2 . . . amn yn
= [v]tB A[w]B0
f (X, Y ) = X t AY
238 CHAPTER 9. BILINEAR AND QUADRATIC FORMS
[f ]B0 = P t [f ]B P,
[f ]B0 = P t [f ]B P
9(0.0) = 0.
Indeed, aij = f (ei , ej 0 ) is the coefficient of xi yj . Thus, a12 = 2,
a13 = 0, a21 = −1, a22 = 0 and a23 = 9.
Thus, matrix of f relative to standard ordered basis B = {e1 , e2 }
of R2 and B 0 of R3 is
0 2 0
[f ]B =
−1 0 9
f (X, Y ) = 2x1 y2 + x2 y1 + x2 y2
Then
f (e1 , e1 ) = f (1 0)t , (1 0)t = 2.1.0 + 0.0 + 0.0 = 0
Hence
0 2
[f ]B =
1 1
Next,
f (e01 , e01 ) = f (1 1)t , (1 1)t = 2.1.1 + 1.1 + 1.1 = 4
1 12
1 .
2 0
Exercise 9.2.5 Find the real symmetric matrix A for the quadratic
form q(X) = x21 − x3 x4 in variables x1 , x2 , x3 and x4 .
Exercises
1. Find the matrix of the quadratic form
x1 y1 + x2 y2 + 3x2 y1 + 5x2 y1
x1 y1 + 2x2 y2 + x2 y1
f (X, Y ) = X t AY,
ax2 + by 2 + 2hxy.
2
244 CHAPTER 9. BILINEAR AND QUADRATIC FORMS
Bibliography
245
Index
246
INDEX 247
scalar matrix, 4
scalar multiplication, 6
similar, 211
similar matrix, 92
singular, 178
skew-Hermitian matrix, 18
skew-symmetric, 14
subspace, 125
surjective, 163
symmetric, 13, 237
unitary matrix, 23
upper (lower) triangular, 4
zero matrix, 4