0% found this document useful (0 votes)
733 views201 pages

Vector Spaces and Matrices

1. A vector space is defined as a non-empty set V along with two compositions: addition of vectors and scalar multiplication, that satisfy certain properties. 2. For V to be a vector space, addition of vectors must make V an abelian group. Scalar multiplication must distribute over vector addition and multiplication of scalars. 3. Some key properties of vector spaces are that they contain a zero vector, vectors have unique additive inverses, and scalar multiplication distributes over vector addition and multiplication of scalars from the field.

Uploaded by

wecharri
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
733 views201 pages

Vector Spaces and Matrices

1. A vector space is defined as a non-empty set V along with two compositions: addition of vectors and scalar multiplication, that satisfy certain properties. 2. For V to be a vector space, addition of vectors must make V an abelian group. Scalar multiplication must distribute over vector addition and multiplication of scalars. 3. Some key properties of vector spaces are that they contain a zero vector, vectors have unique additive inverses, and scalar multiplication distributes over vector addition and multiplication of scalars from the field.

Uploaded by

wecharri
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 201

Krishna's

TEXT BOOK on

V ector S paces
and M atrices
(For B.A. and B.Sc. IVth Semester students of Kumaun University)

Kumaun University Semester Syllabus w.e.f. 2017-18

By

A.R. Vasishtha A.K. Vasishtha


Retired Head, Dept. of Mathematics M.Sc., Ph.D.
Meerut College, Meerut C.C.S. University, Meerut

Kumaun

KRISHNA Prakashan Media (P) Ltd.


KRISHNA HOUSE, 11, Shivaji Road, Meerut-250001 (U.P.), India
Jai Shri Radhey Shyam

Dedicated
to

Lord

Krishna
Authors & Publishers
P reface
This book on Vector Spaces and Matrices has been specially written according to
the latest Syllabus to meet the requirements of B.A. and B.Sc. Semester-IV Students
of all colleges affiliated to Kumaun University.
The subject matter has been discussed in such a simple way that the students will find
no difficulty to understand it. The proofs of various theorems and examples have been
given with minute details. Each chapter of this book contains complete theory and a fairly
large number of solved examples. Sufficient problems have also been selected from various
university examination papers. At the end of each chapter an exercise containing objective
questions has been given.
We have tried our best to keep the book free from misprints. The authors shall be
grateful to the readers who point out errors and omissions which, inspite of all care, might
have been there.
The authors, in general, hope that the present book will be warmly received by the
students and teachers. We shall indeed be very thankful to our colleagues for their
recommending this book to their students.
The authors wish to express their thanks to Mr. S.K. Rastogi, M.D., Mr. Sugam Rastogi,
Executive Director, Mrs. Kanupriya Rastogi, Director and entire team of KRISHNA
Prakashan Media (P) Ltd., Meerut for bringing out this book in the present nice form.
The authors will feel amply rewarded if the book serves the purpose for which it is
meant. Suggestions for the improvement of the book are always welcome.

— Authors
Syllabus

V e ctor s paces and M atrices


B.A./B.Sc. IV Semester
Kumaun University
Fourth Semester – First Paper
B.A./B.Sc. Paper-I M.M.-60

Vector spaces: Vector space, sub spaces, Linear combinations, linear spans, Sums and
direct sums.

Bases and Dimensions: Linear dependence and independence, Bases and dimensions,
Dimensions and subspaces, Coordinates and change of bases.

Matrices: Idempotent, nilpotent, involutary, orthogonal and unitary matrices, singular and
nonsingular matrices, negative integral powers of a nonsingular matrix; Trace of a matrix.

Rank of a matrix: Rank of a matrix, linear dependence of rows and columns of a matrix,
row rank, column rank, equivalence of row rank and column rank, elementary
transformations of a matrix and invariance of rank through elementary transformations,
normal form of a matrix, elementary matrices, rank of the sum and product of two matrices,
inverse of a non-singular matrix through elementary row transformations; equivalence of
matrices.

Applications of Matrices: Solutions of a system of linear homogeneous equations,


condition of consistency and nature of the general solution of a system of linear non-
homogeneous equations, matrices of rotation and reflection.
B rief C ontents
Dedication.........................................................................(v)
Preface............................................................................(vi)
Syllabus (Kumaun University)............................................(vii)
Brief Contents ................................................................(viii)

Chapter 1: Vector Spaces...............................................................V-01—V-34

Chapter 2: Linear Dependence of Vectors.......................................V-35—V-46

Chapter 3: Basis and Dimensions...................................................V-47—V-64

Chapter 4: Homomorphism of Vector Spaces..................................V-65—V-86

Chapter 5: Matrices.....................................................................V-87—V-108

Chapter 6: Rank of a Matrix........................................................V-109—V-154

Chapter 7: Applications of Matrices............................................V-155—V-196


Krishna’s

VECTOR SPACES AND MATRICES


C hapters

1. Vector Spaces

2. Linear Dependence of Vectors

3. Basis and Dimensions

4. Homomorphism of Vector Spaces


5. Matrices

6. Rank of a Matrix

7. Applications of Matrices
V-3

1
V ector S paces

o far we have studied groups and rings. Now we shall study another important
S algebraic structure known as vector space. Before giving the definition of a vector
space we shall make a distinction between internal and external compositions.
Let A be any set. If a * b ∈ A V a, b ∈ A, and a * b is unique then * is said to be an
internal composition in the set A. Here a and b are both elements of the set A.
Let V and F be any two sets. If a ο α ∈ V for all a ∈ F and for all α ∈V and a ο α is
unique, then ο is said to be an external composition in V over F. Here a is an element
of the set F and α is an element of the set V and the resulting element a ο α is an
element of the set V.

1 Vector Space
(Kumaun 2002; Garhwal 10)
Definition: Let ( F , + , .) be a field. The elements of F will be called scalars. Let V be a
non-empty set whose elements will be called vectors. Then V is a vector space over the field F, if

1. There is defined an internal composition in V called addition of vectors and denoted


by ‘+’. Also for this composition V is an abelian group.
V-4

2. There is an external composition in V over F called scalar multiplication and denoted


multiplicatively i.e., a α ∈ V for all a ∈ F and for all α ∈V. In other words V is closed
with respect to scalar multiplication.
3. The two compositions i.e., scalar multiplication and addition of vectors satisfy the
following postulates :
(i) a (α + β) = a α + a β V a ∈ F and V α, β ∈V.
(ii) (a + b) α = aα + bα V a, b ∈ F and V α ∈V.
(iii) (ab) α = a (bα) V a, b ∈ F and V α ∈V.
(iv) 1α = α V α ∈V and 1 is the unity element of the field F.
When V is a vector space over the field F, we shall say that V ( F ) is a vector space. If
the field F is understood we can simply say that V is a vector space. If F is the field R of
real numbers, V is called a real vector space ; similarly if F is Q or F is C, we speak of
rational vector spaces or complex vector spaces.
In the above definition of a vector space V over the field F, we have denoted the
addition of vectors by the symbol ‘+’. This symbol also denotes the addition
composition of the field F, i.e., addition of scalars. There should be no confusion about
the two compositions though we have used the same symbol to denote each of them.
If α, β ∈V, then α + β represents addition in V i.e., addition of vectors. If a, b ∈ F then
a + b represents addition of scalars i.e., addition in the field F. Similarly there should
be no confusion in multiplication of scalars i.e., multiplication of the elements of F and
in scalar multiplication i.e., multiplication of an element of V by an element of F. If
a, b ∈ F , then ab represents multiplication of F and ab ∈ F . If a ∈ F , and α ∈V, then a α
represents scalar multiplication and a α ∈ V . Since 1∈ F and α ∈V, therefore 1α
represents scalar multiplication. Again a α ∈ V , a β ∈ V , therefore aα + aβ represents
addition of vectors and thus aα + aβ is an element of V. Further a ∈ F and α + β ∈V,
therefore a (α + β) represents scalar multiplication and we have a (α + β) ∈ V .
Note 1: For V to be an abelian group with respect to addition of vectors, we must have
the following conditions satisfied :
(i) α + β ∈V for all α, β ∈V.
(ii) α + β = β + α for all α, β ∈V.
(iii) α + ( β + γ ) = (α + β) + γ for all α, β, γ ∈V.
(iv) There exists an element 0 ∈V such that 0 + α = α V α ∈V.
The element 0 ∈V will be called the zero vector. It is the additive identity in V.
(v) To every vector α ∈V there exists a vector − α ∈ V such that − α + α = 0. Thus
each vector should possess additive inverse. The vector − α is called the negative
of the vector α.
Note 2: Since (V , +) is an abelian group therefore all the properties of an abelian group
will hold in V. A few of them are as follows :
(i) α + β = α + γ ⇒ β = γ (left cancellation law)
(ii) β + α = γ + α ⇒ β = γ (right cancellation law)
(iii) α + β = 0 ⇒ α = − β and β = − α.
(iv) − (α + β) = − α − β where by α − β we mean α + (− β).
(v) The additive identity 0 will be unique.
(vi) The additive inverse of each vector will be unique.
(vii) If α + β = γ , then α + β − γ = 0.
V-5

Note 3: There should also be no confusion about the use of the word vector. Here by
vector we do not mean the vector quantity which we have defined in vector algebra as
a directed line segment. Here we shall call the elements of the set V as vectors.
Note 4: In a vector space we shall be dealing with two types of zero elements. One is
the zero vector and the other is the zero element of the field F i.e., the 0 scalar. To
distinguish between the two, we shall use the zero letter in bold type to represent the
zero vector. However the students may use the same symbol 0 to denote the zero
vector as well as the zero scalar. There will be no confusion in this use. The use of 0 will
itself tell whether it stands for zero vector or for zero scalar.
Note 5: We shall usually use the lower case Greek letters α, β, γ etc. to denote vectors
i.e., the elements of V and the lower case Latin letters, a, b, c etc. to denote the scalars
i.e., the elements of the field F.

Example 1: A field K can be regarded as a vector space over any subfield F of K .


(Garhwal 2007)
Solution: Here K is the set of vectors. Addition of vectors is the addition composition
in the field K. Since K is a field, therefore ( K, +) is an abelian group. Further the
elements of the subfield F constitute the set of scalars. The composition of scalar
multiplication is the multiplication composition in the field K. K is a field, therefore
a α ∈ K V a ∈ F and V α ∈ K because both a and α are elements of K. If 1 is the unity
element of K, then 1 is also the unity element of the subfield F. We make the following
observations.
(i) a (α + β) = a α + aβ V a ∈ F and V α, β ∈ K. This result follows from the left
distributive law in K.
(ii) (a + b) α = aα + bα V a, b ∈ F and V α ∈ K. This result is a consequence of the
right distributive law in K.
(iii) (ab) α = a (bα) V a, b ∈ F and V α ∈ K. This result is a consequence of
associativity of multiplication in K.
(iv) 1α = α V α ∈ K and 1 is the unity element of the subfield F. Since 1 is also the
unity element of the field K, therefore 1α = α V α ∈ K.
Hence K ( F ) is a vector space.
Note 1: If F is any field, then F itself is a vector space over the field F.
Note 2: If C is the field of complex numbers and R is the field of real numbers, then C
is a vector space over R because R is a subfield of C. But R is not a vector space over C.
Here R is not closed with respect to scalar multiplication. For example, 2 ∈R and
3 + 4 i ∈ C and (3 + 4 i) 2 ∉R.

Example 2: The set V of all m × n matrices with their elements as real numbers is a vector space
over the field F of real numbers with respect to addition of matrices as addition of vectors and
multiplication of a matrix by a scalar as scalar multiplication. (Garhwal 2007)
V-6

Solution: As in groups, we can easily prove that V is an abelian group with respect to
addition of matrices. The null matrix O of the type m × n is the additive identity of this
abelian group.
If a ∈ F and α ∈V (i. e., α is a matrix of the type m × n with elements as real numbers),
then aα ∈ V because aα is also a matrix of the type m × n with elements as real
numbers. Therefore V is closed with respect to scalar multiplication. Also from our
study of matrices we observe that
(i) a (α + β) = aα + aβ V a ∈ F and V α, β ∈ V .
(ii) (a + b) α = aα + bα V a, b ∈ F and V α ∈ V .
(iii) (ab) α = a (bα) V a, b ∈ F and V α ∈ V .
(iv) 1α = α V α ∈ V where 1 is the unity element of the field F of real numbers.
Hence V ( F ) is a vector space.
Note: If V is the set of all m × n matrices with their elements as rational numbers and
F is the field of real numbers, then V will not be closed with respect to scalar
multiplication. For √ 7 ∈ F and if α ∈V, then √ 7α ∉V because the elements of the
matrix √ 7α will not be rational numbers. Therefore V ( F ) will not be a vector space.
Example 3: The vector space of all ordered n-tuples over a field F.
Solution: Let F be a field. An ordered set α = (a1, a2 , a3 , …, an) of n elements of F is
called an n-tuple over F. Let V be the totality of all ordered n-tuples over F i.e., let
V = {(a1, a2 , …, an) : a1, a2 , a3 , … , an ∈ F }.
Now we shall give a vector space structure to V over the field F. For this we define
equality of two n-tuples, addition of two n-tuples and multiplication of an n-tuple by a
scalar as follows :
Equality of two n-tuples. Two elements
α = (a1, a2 , …, an)
and β = (b1, b2 , …, bn)
of V are said to be equal if and only if ai = bi for each i = 1, 2 , …, n.
Addition composition in V. We define
α + β = (a1 + b1, a2 + b2 , … , an + bn)
V α = (a1, a2 , … , an), β = (b1, b2 , …, bn) ∈ V .
Since a1 + b1, a2 + b2 , … , an + bn are all elements of F, therefore α + β ∈V and thus V is
closed with respect to addition of n-tuples.
Scalar multiplication in V over F. We define
a α = (aa1, aa2 , …, aan) V a ∈ F , V α = (a1, a2 , …, an) ∈ V .
Since aa1, aa2 , … , aan are all elements of F, therefore a α ∈ V and thus V is closed with
respect to scalar multiplication.
Now we shall see that V is a vector space for these two compositions.
Associativity of addition in V. We have
(a1, a2 , … , an) + [(b1, b2 , … , bn) + (c1, c2 , … , c n)]
= (a1, a2 , … , an) + (b1 + c1, b2 + c2 , … , bn + c n)
= (a1 + [b1 + c1], a2 + [b2 + c2 ], …, an + [bn + c n])
V-7

= ([a1 + b1] + c1, [a2 + b2 ] + c2 , … , [an + bn] + c n)


= (a1 + b1, a2 + b2 , … , an + bn) + (c1, c2 , … , c n)
= [(a1, a2 , …, an) + (b1, b2 , … , bn)] + (c1, c2 , …, c n).
Commutativity of addition in V. We have
(a1, a2 , … , an) + (b1, b2 , … , bn) = (a1 + b1, a2 + b2 , … , an + bn)
= (b1 + a1, b2 + a2 , … , bn + an)
= (b1, b2 , …, bn) + (a1, a2 , … , an).
Existence of additive identity in V. We have
(0 , 0 , …, 0 ) ∈V. Also if (a1, a2 , …, an) ∈ V , then
(a1, a2 ,… , an) + (0 , 0 , … , 0 ) = (a1 + 0 , a2 + 0 , … , an + 0 )
= (a1, a2 , …, an).
∴ (0 , 0 , …, 0 ) is the additive identity in V.
Existence of additive inverse of each element of V. If
(a1, a2 , … , an) ∈ V , then (− a1, − a2 , …, − an) ∈ V .
Also we have (− a1, − a2 , …, − an) + (a1, a2 , …, an)
= (− a1 + a1, − a2 + a2 , … , − an + an) = (0 , 0 , …, 0 ).
∴ (− a1, − a2 , … , − an) is the additive inverse of (a1, a2 , … , an).
Thus V is an abelian group with respect to addition. Further we observe that
1. If a ∈ F and α = (a1, a2 , …, an), β = (b1, b2 , …, bn) ∈ V , then
a (α + β) = a (a1 + b1, a2 + b2 , … , an + bn)
= (a [a1 + b1], a [a2 + b2 ], …, a [an + bn])
= (aa1 + ab1, aa2 + ab2 , … , aan + abn)
= (aa1, aa2 , … , aan) + (ab1, ab2 , … , abn)
= a (a1, a2 , … , an) + a (b1, b2 , … , bn) = aα + aβ.
2. If a, b ∈ F and α = (a1, a2 , … , an) ∈ V , then
(a + b) α = ([a + b] a1, [a + b] a2 , … , [a + b] an)
= (aa1 + ba1, aa2 + ba2 , … , aan + ban)
= (aa1, aa2 , … , aan) + (ba1, ba2 , … , ban)
= a (a1, a2 , … , an) + b (a1, a2 , … , an) = a α + b α.
3. If a, b ∈ F and α = (a1, a2 , … , an) ∈ V , then
(ab) α = ([ab] a1, [ab] a2 , … , [ab] an)
= (a [ba1], a [ba2 ], … , a [ban])
= a (ba1, ba2 , … , ban) = a [b (a1, a2 , … , an)] = a (bα).
4. If 1 is the unity element of F and α = (a1, a2 , … , an) ∈ V , then
1 α = (1 a1, 1 a2 , … , 1 an) = (a1, a2 , … , an) = α.
Hence V is a vector space over F. The vector space of all ordered n-tuples over F will be denoted
by Vn ( F ) . Sometimes we also denote it by F n or F (n) or F n ( F ). Here the zero vector i.e., 0
is the n-tuple (0 , 0 , … , 0 ).
Note: V2 ( F ) = {(a1, a2 ) : a1, a2 ∈ F} is the vector space of all ordered pairs over F.
Similarly V3 ( F ) = {(a1, a2 , a3 ) : a1, a2 , a3 ∈ F } is the vector space of all ordered
triads over F.
V-8

Example 4: The vector space of all polynomials over a field F.


Solution: Let F [ x] denote the set of all polynomials in an indeterminate x over a
field F. Then F [ x] is a vector space over the field F with respect to addition of two
polynomials as addition of vectors and the product of a polynomial by a constant
polynomial (i. e., by an element of F ) as scalar multiplication.
Let f ( x) = Σ ai x i = a0 + a1 x + a2 x2 + a3 x3 + … ,
g ( x) = Σ bi x i = b0 + b1 x + b2 x2 + b3 x3 + …
and h ( x) = Σ c i x i = c0 + c1 x + c2 x2 + c3 x3 + …
be any arbitrary members of F [ x].
Equality of two polynomials. We define f ( x) = g ( x) if and only if ai = bi for each
i = 0 , 1, 2 , … .
Addition composition in F [ x ] . We define
f ( x) + g ( x) = (a0 + b0 ) + (a1 + b1) x + (a2 + b2 ) x 2 + …
= Σ (ai + bi) x i.
Since a0 + b0 , a1 + b1, a2 + b2 , … are all elements of F, therefore
f ( x) + g ( x) ∈ F [ x]
and thus F [ x] is closed with respect to addition of polynomials.
Scalar multiplication in F [ x ] over F. If k is any scalar i. e., k ∈ F , we define
k f ( x) = ka0 + (ka1) x + (ka2 ) x 2 + (ka3 ) x 3 + …
= Σ (kai) x i.
Since ka0 , ka1, ka2 ,… are all elements of F, therefore k f ( x) ∈ F [ x] and thus F [ x] is
closed with respect to scalar multiplication.
Now we shall show that F [ x] is a vector space for these two compositions.
Commutativity of addition in F [ x ] . We have
f ( x) + g ( x) = (a0 + b0 ) + (a1 + b1) x + (a2 + b2 ) x 2 + …
= (b0 + a0 ) + (b1 + a1) x + (b2 + a2 ) x 2 + …
[∵ addition in the field F is commutative]
= g ( x) + f ( x).
Associativity of addition in F [ x ] . We have
[ f ( x) + g ( x)] + h ( x) = Σ (ai + bi) x i + Σ c i x i

= Σ [(ai + bi) + c i] x i = Σ [ai + (bi + c i)] x i


= Σ ai x i + Σ (bi + c i) x i

= f ( x) + [ g ( x) + h ( x)].
Existence of additive identity in F [ x ] . Let 0 denote the zero polynomial over the
field F i. e., 0 = 0 + 0 x + 0 x 2 + 0 x 3 + …
Then 0 ∈ F [ x] and 0 + f ( x) = f ( x).
∴ the zero polynomial 0 is the additive identity.
V-9

Existence of additive inverse of each member of F [ x ] . Let − f ( x) be the


polynomial over the field F defined as
− f ( x) = − a0 + (− a1) x + (− a2 ) x2 + …
= − a0 − a1 x − a2 x2 − … .
Then − f ( x) ∈ F [ x] and we have
− f ( x) + f ( x) = 0 i. e., the zero polynomial.
∴ − f ( x) is the additive inverse of f ( x).
Thus F [ x] is an abelian group with respect to addition of polynomials.
Now for the operation of scalar multiplication we make the following observations.
1. If k ∈ F , then
k [ f ( x) + g ( x)] = k [(a0 + b0 ) + (a1 + b1) x + (a2 + b2 ) x2 + …]
= k (a0 + b0 ) + k (a1 + b1) x + k (a2 + b2 ) x2 + …
= (ka0 + kb0 ) + (ka1 + kb1) x + (ka2 + kb2 ) x2 + …
= [ka0 + (ka1) x + (ka2 ) x2 + …] + [kb0 + (kb1) x + (kb2 ) x2 + …]
= k (a0 + a1 x + a2 x2 + …) + k (b0 + b1 x + b2 x2 + …)
= kf ( x) + kg ( x).
2. If k1, k2 ∈ F , then
(k1 + k2 ) f ( x) = (k1 + k2 ) a0 + [(k1 + k2 ) a1] x + [(k1 + k2 ) a2 ] x2 + …
= (k1 a0 + k2 a0 ) + (k1 a1 + k2 a1) x + ( k1 a2 + k2 a2 ) x2 + …

= [k1 a0 + (k1 a1) x + (k1 a2 ) x2 + …] + [k2 a0 + (k2 a1) x + (k2 a2 ) x2 + …]


= k1 (a0 + a1 x + a2 x2 + …) + k2 (a0 + a1 x + a2 x2 + …)
= k1 f ( x) + k2 f ( x).
3. If k1, k2 ∈ F , then
(k1 k2 ) f ( x) = (k1 k2 ) a0 + [(k1 k2 ) a1] x + [(k1 k2 ) a2 ] x2 + …
= k1 (k2 a0 ) + [k1 (k2 a1)] x + [k1 (k2 a2 )] x2 + …
= k1 [k2 a0 + (k2 a1) x + (k2 a2 ) x2 + …] = k1 [k2 f ( x)].
4. If 1 is the unity element of the field F, then
1 f ( x) = (1a0 ) + (1a1) x + (1a2 ) x2 + …
= a0 + a1 x + a2 x2 + … = f ( x).
Hence F [ x] is a vector space over the field F.
Example 5: The set of all convergent real sequences is a vector space over the field of real
numbers.
Solution: Let V denote the set of all convergent sequences over the field of real
numbers.
Let α = < α1, α2 , … , α n , …> = < α n >, β = < β1, β2 , … , β n , …> = < β n >
and γ = < γ1, γ 2 , … , γ n , …> = < γ n >
be any three convergent sequences.
V-10

1. (V , +) is an abelian group.
(i) We have α + β = < α n > + < β n > = < α n + β n > which is also a convergent
sequence. Therefore V is closed for addition of sequences.
(ii) Commutativity of addition. We have
α + β = < αn > + < βn > = < αn + βn > = < βn + αn >
= < βn > + < αn > = β + α
(iii) Associativity of addition. We have
α + ( β + γ ) = < α n > + [< β n > + < γ n >] = < α n > + < β n + γ n >
= < α n + ( β n + γ n) > = < (α n + β n) + γ n > = < α n + β n > + < γ n >
= [< α n > + < β n >] + < γ n > = (α + β) + γ .
(iv) Existence of additive identity. The zero sequence < 0 > = < 0 , 0 , 0 , …> is the
additive identity.
(v) Existence of additive inverse. For every sequence < α n > there exists a
sequence < − α n > such that
< α n > + < − α n > = < α n − α n > = < 0 > = the additive identity.
∴ (V , +) is an abelian group.
2. V is closed for scalar multiplication.
Let a be any scalar i. e., a be any real number. Then
aα = a < α n > = < aα n > which is also a convergent sequence because
lim aα n = a lim α n.
n→ ∞ n→ ∞

Thus V is closed for scalar multiplication.


3. Laws of scalar multiplication. Let a, b ∈R. We have
(i) a (α + β) = a [< α n > + < β n >] = a < α n + β n >
= < a (α n + β n) > = < aα n + aβ n > = < aα n > + < aβ n >
= a < α n > + a < β n > = aα + aβ.
(ii) (a + b) α = (a + b) < α n > = < (a + b) α n >
= < aα n + bα n > = < aα n > + < bα n >
= a < α n > + b < α n > = aα + bα.
(iii) (ab) α = (ab) < α n > = < (ab) α n > = < a (bα n) >
= a < bα n > = a [b < α n >] = a (bα).
(iv) 1α = 1 < α n > = < 1α n > = < α n > = α.
Thus all the postulates of a vector space are satisfied. Hence V is a vector space over
the field of real numbers.
Example 6: Let V be the set of all pairs ( x, y) of real numbers, and let F be the field of real
numbers. Define
( x, y) + ( x1, y1) = ( x + x1, 0 )
c ( x, y) = (c x, 0 ).
Is V, with these operations, a vector space over the field of real numbers ? (Meerut 2002)
V-11

Solution: If any of the postulates of a vector space is not satisfied, then V will not be
a vector space. We shall show that for the operation of addition of vectors as defined
in this problem the identity element does not exist. Suppose the ordered pair ( x1, y1) is
to be the identity element for the operation of addition of vectors. Then we must have
( x, y) + ( x1, y1) = ( x, y) V x, y ∈ R
⇒ ( x + x1, 0 ) = ( x, y) V x, y ∈ R.
But if y ≠ 0, then we cannot have ( x + x1, 0 ) = ( x, y). Thus there exists no element
( x1, y1) of V such that
( x, y) + ( x1, y1) = ( x, y) V ( x, y) ∈ V .
Therefore the identity element does not exist and V is not a vector space over the
field R.

Example 7: Let V be the set of all pairs ( x, y) of real numbers, and let F be the field of real
numbers. Define
( x, y) + ( x1, y1) = (3 y + 3 y1, − x − x1)
c ( x, y) = (3 cy, − cx).
Verify that V, with these operations, is not a vector space over the field of real numbers.

Solution: Take α = (1, 2), β = (2, 3), γ = (1, 3)


Then (α + β) + γ = {(1, 2) + (2, 3)} + (1, 3)
= (15, − 3) + (1, 3) = (0 , − 16)
and α + (β + γ ) = (1, 2) + {(2, 3) + (1, 3)}
= (1, 2) + (18, − 3) = (− 3, − 19).
Thus we conclude that in general (α + β) + γ ≠ α + (β + γ ). Hence V is not a vector
space over the field of real numbers.

Example 8: How many elements are there in the vector space of polynomials of degree at most n
in which the coefficients are the elements of the field I ( p) over the field I ( p), p being a prime
number ?
Solution: The field I ( p) is the field ({0 , 1, 2 , … , p − 1}, + p , × p ).

The number of distinct elements in the field I ( p) is p.


If f ( x) is a polynomial of degree at most n over the field I ( p), then
f ( x) = a0 + a1 x + a2 x2 + … + an x n, where a0 , a1, a2 ,… , an ∈I ( p).
Now in the polynomial f ( x), the coefficient of each of the n + 1 terms
a0 , a1 x, a2 x2 , … , an x n, can be filled in p ways because any of the p elements of the
field I ( p) can be filled there.
Thus we can have p × p × p × … upto (n + 1) times i. e., p n +1 distinct polynomials of
degree at most n over the field I ( p). Hence if Pn is the vector space of polynomials of
degree at most n in which the coefficients are the elements of the field I ( p) over the
field I ( p), then Pn has p n+1 distinct elements.
V-12

Comprehensive Exercise 1

1. Show that a field F may be considered as a vector space over F if scalar


multiplication is identified with field multiplication.
2. Show that the complex field C is a vector space over the real field R.
(Garhwal 2007)
3. Let V be the set of ordered pairs (z1, z2 ) of complex numbers. Show that V is a
vector space over the real field R with addition in V and scalar multiplication on
V defined by (z1, z2 ) + (w1, w2 ) = (z1 + w1, z2 + w2 ) and a (z1, z2 ) = (az1, az2 ) where
z1, z2 , w1, w2 ∈C and a ∈R .
4. Let F be the field I ( p), p being a prime number. Show that Vn( F ) is a vector space.
How many vectors are there in this vector space ?
5. Let S be any non-empty set and let F be any field. Let V be the set of all functions
from S to F i. e., let
V = { f : f : S → F}
Let us define sum of two elements f and g in V as follows :
( f + g) ( x) = f ( x) + g ( x) V x ∈ S.
Also let us define scalar multiplication of an element f in V by an element c in F as
follows :
(cf ) ( x) = cf ( x) V x ∈ S.
Then V ( F ) is a vector space.
6. The vector space of all real valued continuous functions defined in some interval
[0 , 1].
7. Prove that the set of all vectors in a plane over the field of real numbers is a vector
space.
8. Let V be the set of all pairs ( x, y) of real numbers, and let F be the field of real
numbers. Examine in each of the following cases whether V is a vector space over
the field of real numbers or not :
(i) ( x, y) + ( x1, y1) = ( x + x1, y + y1)
c ( x, y) = (| c | x,| c | y)
(ii) ( x, y) + ( x1, y1) = ( x + x1, y + y1)
c ( x, y) = (0 , c y)
(iii) ( x, y) + ( x1, y1) = ( x + x1, y + y1)
c ( x, y) = (c 2 x, c 2 y)
(iv) ( x, y) + ( x1, y1) = ( x + y1, y + x1)
c ( x, y) = (cx, cy)
9. Show that the axiom 1α = α V α ∈ V and 1∈ F, defining the vector space is not a
consequence of the remaining axioms.
10. Let V be the set of all pairs ( x, y) of real numbers, and let F be the field of real
numbers. Define
( x, y) + ( x1, y1) = ( x + x1, y + y1)
c ( x, y) = (cx, y).
V-13

Show that with these operations V is not a vector space over the field of real
numbers.
11. Let V be the set of ordered pairs ( x, y) of real numbers, Show that V is not a vector
space over R with addition in V and scalar multiplication on V defined by :
(i) ( x, y) + ( x1,
y1) = ( x + x1, y + y1) and c ( x, y) = ( x, y) ;
(ii) ( x, y) + ( x1,
y1) = ( x, y) and c ( x, y) = (cx, cy) ;
(iii) ( x, y) + ( x1,
y1) = (0 , 0 ) and c ( x, y) = (cx, cy) ;
(iv) ( x, y) + ( x1,
y1) = ( xx1, yy1) and c ( x, y) = (cx, cy).
12. Let V be the set of all pairs ( x, y) of real numbers and let F be the field of real
numbers. Define
( x, y) + ( x1, y1) = ( x + x1, y + y1)
c ( x, y) = (c x, 0 ).
Is V a vector space over the field of real numbers with these operations ?
13. Let R be the field of real numbers and let Pn be the set of all polynomials (of degree
at most n) over the field R. Prove that Pn is a vector space over the field R.
14. Let U and W be vector spaces over a field F. Let V be the set of ordered pairs i. e.,
V = {(u, w) : u ∈ U , w ∈ W }.
Show that V is a vector space over F with addition in V and scalar multiplication
on V defined by
(u, w) + (u1, w1) = (u + u1, w + w1)
and k (u, w) = (ku, kw) where u, u1 ∈ U and w, w1 ∈ W and k ∈ F .
(Kumaun 2003)

A nswers 1
4. pn
8. (i) not a vector space (ii) not a vector space
(iii) not a vector space (iv) not a vector space
12. not a vector space

2 General Properties of Vector Spaces


Theorem 1: Let V ( F ) be a vector space and 0 be the zero vector of V. Then
(i) a 0 = 0 V a ∈ F. (Gorakhpur 2015)
(ii) 0 α = 0 V α ∈V. (Kumaun 2006, 08; Gorakhpur 15)
(iii) a (− α) = − (aα) V a ∈ F , V α ∈V. (Garhwal 2007)
(iv) (− a) α = − (aα) V a ∈ F , V α ∈V. (Garhwal 2007)
(v) a (α − β) = aα − aβ V a ∈ F and V α, β ∈V.
(vi) aα = 0 ⇒ a = 0 or α = 0. (Kumaun 2006)
V-14

Proof: (i) We have a 0 = a (0 + 0) [∵ 0 = 0 + 0]


= a 0 + a 0.
∴ 0 + a0 = a0 + a0 [∵ a 0 ∈ V and 0 + a 0 = a 0]
Now V is an abelian group with respect to addition.
Therefore by right cancellation law in V, we get 0 = a 0.
(ii) We have 0 α = (0 + 0 ) α [∵ 0 = 0 + 0 ]
= 0 α + 0 α.
∴ 0 + 0 α = 0 α + 0 α. [∵ 0 α ∈ V and 0 + 0 α = 0 α]
Now V is an abelian group with respect to addition of vectors. Therefore by right
cancellation law in V, we get 0 = 0 α.
(iii) We have a [α + (− α)] = a α + a (− α) ⇒ a 0 = aα + a(− α)
⇒ 0 = aα + a (− α) [∵ a0 = 0]
⇒ a (− α) is the additive inverse of a α ⇒ a (− α) = − (a α).
(iv) We have [a + (− a)] α = aα + (− a) α ⇒ 0 α = aα + (− a) α
⇒ 0 = a α + (− a) α ⇒ (− a) α is the additive inverse of aα
⇒ (− a) α = − (a α).
(v) We have a (α − β) = a [α + (− β)] = aα + a (− β)
= aα + [− (aβ)] [∵ a (− β) = − (aβ)]
= aα − aβ .
(vi) Let aα = 0 and a ≠ 0. Then a −1 exists because a is a non-zero element of the field
F.
∴ a α = 0 ⇒ a − 1 (aα) = a − 1 0 ⇒ (a − 1 a) α = 0 ⇒ 1α = 0 ⇒ α = 0.
Again let aα = 0 and α ≠ 0. Then to prove that a = 0. Suppose a ≠ 0. Then a −1 exists.
∴ aα = 0 ⇒ a − 1 (aα) = a − 1 0 ⇒ (a − 1 a) α = 0 ⇒ 1α = 0 ⇒ α = 0.
Thus we get a contradiction that α must be a zero vector. Therefore a must be equal to
0. Hence α ≠ 0 and a α = 0 ⇒ a = 0.
Theorem 2: Let V ( F ) be a vector space. Then
(i) If a , b ∈ F and α is a non-zero element of V, we have aα = b α ⇒ a = b.
(ii) If α , β ∈V and a is a non-zero element of F, we have aα = a β ⇒ α = β.
Proof: (i) We have aα = bα ⇒ aα − bα = 0 ⇒ (a − b) α = 0.
But α ≠ 0. Therefore (a − b) α = 0 ⇒ a − b = 0 ⇒ a = b.
(ii) We have aα = aβ ⇒ aα − aβ = 0 ⇒ a (α − β) = 0.
But a ≠ 0. Therefore a (α − β) = 0 ⇒ α − β = 0 ⇒ α = β.

3 Vector Subspaces
Definition: Let V be a vector space over the field F and W ⊆ V . Then W is called a subspace
of V if W itself is a vector space over F with respect to the operations of vector addition and scalar
multiplication in V. (Garhwal 2010)
V-15

Remark: Let V ( F ) be any vector space. Then V itself and the subset of V consisting
of zero vector only are always subspaces of V. These two are called improper
subspaces. If V has any other subspace, then it is called a proper subspace. The
subspace of V consisting of zero vector only is called the zero subspace.
Theorem 1: The necessary and sufficient condition for a non-empty subset W of a vector
space V (F ) to be a subspace of V is that W is closed under vector addition and scalar
multiplication in V. (Kumaun 2003, 10, 13)
Proof: If W itself is a vector space over F with respect to vector addition and scalar
multiplication in V, then W must be closed with respect to these two compositions.
Hence the condition is necessary.
The condition is sufficient. Now suppose that W is a non-empty subset of V and
W is closed under vector addition and scalar multiplication in V.
Let α ∈W. If 1 is the unity element of F, then − 1 ∈ F. Now W is closed under scalar
multiplication. Therefore
− 1 ∈ F, α ∈ W ⇒ (− 1) α ∈ W ⇒ − (1α) ∈ W ⇒ − α ∈ W .
[∵ α ∈ W ⇒ α ∈ V and 1 α = α in V .]
Thus the additive inverse of each element of W is also in W.
Now W is closed under vector addition.
Therefore α ∈W, − α ∈ W ⇒ α + (− α) ∈ W ⇒ 0 ∈W ,
where 0 is the zero vector of V.
Hence the zero vector of V is also the zero vector of W . Since the elements of W are
also the elements of V, therefore vector addition will be commutative as well as
associative in W. Hence W is an abelian group with respect to vector addition. Also it
is given that W is closed under scalar multiplication. The remaining postulates of a
vector space will hold in W since they hold in V of which W is a subset.
Hence W itself is a vector space for the two compositions.
∴ W is a subspace of V.
Theorem 2: The necessary and sufficient conditions for a non-empty subset W of a vector
space V ( F ) to be a subspace of V are
(i) α ∈W, β ∈ W ⇒ α − β ∈ W. (Garhwal 2013)
(ii) a ∈ F , α ∈ W ⇒ aα ∈ W.
Proof: The conditions are necessary. If W is a subspace of V, then W is an
abelian group with respect to vector addition. Therefore α ∈W, β ∈ W ⇒ α − β ∈ W.
Also W must be closed under scalar multiplication. Therefore the condition (ii) is also
necessary.
The conditions are sufficient: Now suppose W is a non-empty subset of V
satisfying the two given conditions. From condition (i), we have α ∈W,
α ∈W ⇒ α − α ∈W ⇒ 0 ∈W.
Thus the zero vector of V belongs to W and it will also be the zero vector of W.
Now 0 ∈W, α ∈ W ⇒ 0 − α ∈ W ⇒ − α ∈ W .
Thus the additive inverse of each element of W is also in W.
V-16

Again α ∈W, β ∈ W ⇒ α ∈ W, − β ∈ W
⇒ α − (− β) ∈ W ⇒ α + β ∈ W.
Thus W is closed with respect to vector addition.
Since the elements of W are also the elements of V, therefore vector addition will be
commutative as well as associative in W. Hence W is an abelian group under vector
addition. Also from condition (ii), W is closed under scalar multiplication. The
remaining postulates of a vector space will hold in W since they hold in V which is a
superset of W. Hence W is a subspace of V.
Theorem 3: The necessary and sufficient condition for a non-empty subset W of a vector
space V (F) to be a subspace of V is
a, b ∈ F and α, β ∈W ⇒ aα + b β ∈ W .
(Kumaun 2007,14 ; Garhwal 10 ; Gorakhpur 10)
Proof: The condition is necessary. If W is a subspace of V, then W must be
closed under scalar multiplication and vector addition.
Therefore a ∈ F , α ∈ W ⇒ aα ∈ W
and b ∈ F , β ∈ W ⇒ bβ ∈ W.
Now aα ∈ W , bβ ∈ W ⇒ aα + bβ ∈ W . Hence the condition is necessary.
The condition is sufficient. Now suppose W is a non-empty subset of V satisfying
the given condition i.e.,
a, b ∈ F and α, β ∈ W ⇒ aα + bβ ∈ W.
Taking a = 1, b = 1, we see that if α, β ∈W then 1 α + 1 β ∈W
⇒ α + β ∈W. [∵ α ∈ W ⇒ α ∈ V and 1 α = α in V ]
Thus W is closed under vector addition.
Now taking a = − 1, b = 0, we see that if α ∈W, then
(− 1) α + 0 α ∈ W [in place of β we have taken α]
⇒ − (1α) + 0 ∈ W ⇒ − α ∈ W.
Thus the additive inverse of each element of W is also in W.
Taking a = 0, b = 0, we see that if α ∈W, then
0 α + 0 α ∈W ⇒ 0 + 0 ∈W ⇒ 0 ∈W.
Thus the zero vector of V belongs to W. It will also be the zero vector of W.
Since the elements of W are also the elements of V, therefore vector addition will be
associative as well as commutative in W. Thus W is an abelian group with respect to
vector addition.
Now taking β = 0, we see that if a, b ∈ F
and α ∈W, then
aα + b 0 ∈ W i.e., aα + 0 ∈ W i.e., a α ∈ W .
Thus W is closed under scalar multiplication.
The remaining postulates of a vector space will hold in W since they hold in V of
which W is a subset. Hence W ( F ) is a subspace of V ( F ).
Theorem 4: A non-empty subset W of a vector space V ( F ) is a subspace of V if and only if
for each pair of vectors α, β in W and each scalar a in F the vector aα + β is again in W.
V-17

Proof: The condition is necessary. If W is a subset of V, then W must be closed


with respect to scalar multiplication and as well as with respect to vector addition.
Therefore
a ∈ F , α ∈ W ⇒ aα ∈ W .
Further aα ∈ W , β ∈ W ⇒ aα + β ∈ W .
Hence the condition is necessary.
The condition is sufficient. It is given that W is a non-empty subset of V and
a ∈ F , α , β ∈ W ⇒ aα + β ∈ W . We are to prove that W is a subspace of V.
(i) Since W is non-empty, therefore there is at least one vector in W, say γ. Now
1 ∈ F ⇒ − 1 ∈ F. Therefore taking a = − 1, α = γ , β = γ , we get from the given
condition that
(− 1) γ + γ = − (1 γ ) + γ = − γ + γ = 0 is in W.
(ii) Now let a ∈ F , α ∈ W . Since 0 is in W, therefore taking β = 0 in the given
condition, we get
aα + 0 = aα is in W.
Thus W is closed with respect to scalar multiplication.
(iii) Let α ∈W. Since − 1 ∈ F and W is closed with respect to scalar multiplication,
therefore (− 1) α = − (1α) = − α is in W.
(iv) We have 1∈ F. If α, β ∈W, then 1α + β = α + β is in W. Thus W is closed with
respect to vector addition.
The remaining postulates of a vector space will hold in W since they hold in V of
which W is a subset.
Hence W is a subspace of V.
Note: If we are to prove that a non-empty subset W of a vector space V is a subspace
of V, then either it is sufficient to prove that
a, b ∈ F
and α, β ∈W
⇒ aα + bβ ∈ W
or it is sufficient to prove that
a ∈ F , and α, β ∈ W ⇒ aα + β ∈ W .

Example 9: The set W of ordered triads (a1, a2 , 0 ) where a1, a2 ∈ F is a subspace of V3 ( F ).

Solution: W is non-empty since 0 = (0 , 0 , 0 ) ∈ W . Let α = (a1, a2 , 0 ) and β = (b1, b2 , 0 )


be any two elements of W. Then a1, a2 , b1, b2 ∈ F . If a, b be any two elements of F, we
have
aα + bβ = a (a1, a2 , 0 ) + b (b1, b2 , 0 )
= (aa1, aa2 , 0 ) + (bb1, bb2 , 0 )
= (aa1 + bb1 , aa2 + bb2 , 0 ) ∈ W
since aa1 + bb1, aa2 + bb2 ∈ F and the last co-ordinate of this triad is zero.
Hence W is a subspace of V3 ( F ).
V-18

Example 10: Let V be the vector space of all polynomials in an indeterminate x over a field F.
Let W be a subset of V consisting of all polynomials of degree ≤ n. Then W is a subspace of V.
Solution: Let α and β be any two elements of W. Then α, β are polynomials over F of
degree ≤ n. If a, b are any two elements of F, then aα + bβ will also be a polynomial of
degree ≤ n.
Therefore aα + bβ ∈ W . Hence W is a subspace of V.
Example 11: If a1, a2 , a3 are fixed elements of a field F, then the set W of all ordered triads
( x1, x2 , x3 ) of elements of F, such that a1 x1 + a2 x2 + a3 x3 = 0 , is a subspace of V3 ( F ).
(Kumaun 2007)
Solution: W is non-empty since (0 , 0 , 0 ) ∈W . Let α = ( x1, x2 , x3 ) and
β = ( y1, y2 , y3 ) be any two elements of W. Then x1, x2 , x3 , y1, y2 , y3 are elements of
F and are such that
a1 x1 + a2 x2 + a3 x3 = 0 …(1)
and a1 y1 + a2 y2 + a3 y3 = 0 . …(2)
If a, b be any two elements of F, we have
aα + bβ = a ( x1, x2 , x3 ) + b ( y1, y2 , y3 )
= (ax1, ax2 , ax3 ) + (by1, by2 , by3 )
= (ax1 + by1, ax2 + by2 , ax3 + by3 ).
Now a1 (ax1 + by1) + a2 (ax2 + by2 ) + a3 (ax3 + by3 )
= a (a1 x1 + a2 x2 + a3 x3 ) + b (a1 y1 + a2 y2 + a3 y3 )
= a0 + b0 [by (1) and (2)]
= 0.
∴ a α + bβ = (ax1 + by1, ax2 + by2 , ax3 + by3 ) ∈ W .
Hence W is a subspace of V3 ( F ).
Example 12: Which of the following sets of vectors α = (a1, a2 , … , an) in R n are subspaces of
R n (n ≥ 3) ?
(i) all α such that a1 ≤ 0 ;
(ii) all α such that a3 is an integer ;
(iii) all α such that a2 + 4 a3 = 0 ;
(iv) all α such that a1 + a2 + … + an = k (a given constant).
Solution: (i) Let W = {α : α ∈ R n and a1 ≤ 0}.
If we take a1 = − 3, then a1 < 0 and so
α = (− 3, a2 , … , an) ∈ W .
Now if we take a = − 2, then
aα = (6, − 2 a2 , … , − 2 an).
Since the first coordinate of aα is 6 which is > 0, therefore aα ∉W .
Thus α ∈ W , a ∈ R but aα ∉W . Therefore W is not closed for scalar multiplication and
so W is not a subspace of R n .
(ii) Let W = {α : α ∈ R n and a3 is an integer}.
If we take a3 = 5, then a3 is an integer and so
α = (a1, a2 , 5, … , an) ∈ W .
V-19

1
Now if we take a = , then
2
1 1 5 1 
aα =  a1, a2 , , … , an ⋅
2 2 2 2 
Since the third coordinate of aα is 5 / 2 which is not an integer therefore aα ∉W .
Thus α ∈ W , a ∈ R but aα ∉W . Therefore W is not closed for scalar multiplication and
so W is not a subspace of R n .
(iii) Let W = {α : α ∈ R n and a2 + 4 a3 = 0}.
W is non-empty since 0 = (0 , 0 , … , 0 ) ∈ W .
Let α = (a1, … , an) and β = (b1, … , bn) be any two members of W.
Then a2 + 4 a3 = 0 and b2 + 4 b3 = 0 .
If a, b ∈R, then aα + bβ = (aa1 + bb1, … , aan + bbn).
We have (aa2 + bb2 ) + 4 (aa3 + bb3 ) = a (a2 + 4 a3 ) + b (b2 + 4 b3 ) = a . 0 + b . 0 = 0 .
Thus according to the definition of W , aα + bβ ∈ W .
In this way α, β ∈W and a, b ∈ R ⇒ aα + bβ ∈ W . Hence W is a subspace of R n .
(iv) A subspace if k = 0 and not a subspace if k ≠ 0.
Example 13: Let R be the field of real numbers. Which of the following are subspaces of
V3 (R) ?
(i ) {( x , 2 y , 3 z ) : x , y , z ∈R}.
(ii ) {( x , x , x) : x ∈R}.
(iii) {( x , y , z ) : x , y , z are rational numbers}.
Solution: (i) Let W = {( x , 2 y, 3 z ) : x, y, z ∈ R}.
Let α = ( x1, 2 y1, 3 z1) and β = ( x2 , 2 y2 , 3 z2 ) be any two elements of W. Then
x1, y1, z1, x2 , y2 , z2 are real numbers. If a, b are any two real numbers, then
aα + bβ = a ( x1, 2 y1 , 3 z1) + b ( x2 , 2 y2 , 3 z2 )
= (ax1 + bx2 , 2 ay1 + 2 by2 , 3 az1 + 3 bz2 )
= (ax1 + bx2 , 2 [ay1 + by2 ], 3 [az1 + bz2 ])
∈W, since ax1 + bx2 , ay1 + by2 , az1 + bz2
are real numbers.
Thus a, b ∈R and α, β ∈ W ⇒ aα + bβ ∈ W.
∴ W is a subspace of V3 (R).
(ii) Let W = {( x, x, x) : x ∈ R}.
Let α = ( x1, x1, x1) and β = ( x2 , x2 , x2 ) be any two elements of W. Then x1, x2 are real
numbers. If a, b are any real numbers, then
aα + bβ = a ( x1, x1, x1) + b ( x2 , x2 , x2 )
= (ax1 + bx2 , ax1 + bx2 , ax1 + bx2 ) ∈ W , since ax1 + bx2 ∈R.
Thus W is a subspace of V3 (R).
(iii) Let W = {( x , y , z ) : x , y , z are rational numbers}.
Now α = (3, 4, 5) is an element of W. Also a = √ 7 is an element of R . But
aα = √ 7 (3, 4, 5) = (3 √ 7, 4 √ 7, 5 √ 7) ∉W since 3 √ 7, 4 √ 7, 5 √ 7
are not rational numbers.
Therefore W is not closed under scalar multiplication. Hence W is not a subspace of
V3 (R).
V-20

Example 14: Let V be the vector space of all 2 × 2 matrices over the field R. Show that W is not
a subspace of V, where W contains all 2 × 2 matrices with zero determinants.
a 0 0 0 
Solution: Let A =  , B =  0 b  , where a, b ∈R and a ≠ 0 , b ≠ 0 .
0 0   
We have | A| = 0 ,| B| = 0 so A, B ∈ W .
a 0
Now A + B =   . We have | A + B| = ab ≠ 0 .
0 b 
∴ A + B ∉W .
Thus A ∈ W , B ∈ W but A + B ∉W .
Hence W is not a subspace of V.
Example 15: If a vector space V is the set of all real valued continuous functions over the field of
real numbers R , then show that the set W of solutions of the differential equation.
d2 y dy
2 −9 +2y =0
dx2 dx
is a subspace of V.
 d2 y dy 
Solution: We have W =  y : 2 2 − 9 + 2 y = 0  , where y = f ( x).
 dx dx 
Obviously y = 0 satisfies the given differential equation and as such it belongs to W
and thus W ≠ ∅.
Now let y1, y2 ∈ W . Then
d2 y1 dy1
2 2
−9 + 2 y1 = 0 …(1)
dx dx
d2 y2 dy2
and 2 2
−9 + 2 y2 = 0 . …(2)
dx dx
Let a, b ∈R. If W is to be a subspace then we should show that ay1 + by2 also belongs to
W i. e., it is a solution of the given differential equation.

d2 d
We have 2 2
(ay1 + by2 ) − 9 (ay1 + by2 ) + 2 (ay1 + by2 )
dx dx
d2 y1 d2 y2 dy1 dy
= 2a 2
+ 2b 2
− 9a − 9 b 2 + 2 ay1 + 2 by2
dx dx dx dx
 d y1 2
dy1   d2 y2 dy 
= a 2 2
− 9 + 2 y1  + b 2
  2
− 9 2 + 2 y2 
 dx dx   dx dx 
= a ⋅ 0 + b ⋅ 0, by (1) and (2)
= 0.
Thus ay1 + by2 is a solution of the given differential equation and so it belongs to W.
Hence W is a subspace of V.
V-21

Example 16: Find the solution of system of homogeneous linear equations. Let V ( F ) be the
vector space of all n × 1matrices over the field F. Let A be an m × n matrix over F. Then the set W of
all n × 1matrices X over F such that AX = O is a subspace of V. Here O is a null matrix of the type
m × 1.
Solution: W is non-empty since On× 1 ∈ W . Let X , Y ∈ W . Then X and Y are n × 1
matrices over F such that AX = O, AY = O.
Let a ∈ F . Then aX + Y is also an n × 1 matrix over F.
We have A (aX + Y ) = A (aX ) + AY = a ( AX ) + AY = aO + O = O + O = O
Therefore aX + Y ∈ W . Thus
a ∈ F , X , Y ∈ W ⇒ aX + Y ∈ W .
Hence W is a subspace of V.

Comprehensive Exercise 2

1. Show that the subset {(a, b, c ) : a + b + c = 0 } of R 3 is a subspace of R 3 .


(Gorakhpur 2014)
2. Prove that the set of all solutions (a, b, c ) of the equation a + b + 2 c = 0 is a
subspace of the vector space V3 (R).
3. Show that the set W of the elements of the vector space V3 (R) of the form
( x + 2 y, y, − x + 3 y) where x, y ∈R is a subspace of V3 (R). (Kumaun 2013)
4. Let V be the vector space of all polynomials over the field R. Determine whether
or not W is a subspace of V where (i) W contains all polynomials with integral
coefficients (ii) W consists of all polynomials b0 + b1 x2 + b2 x4 + … + bn x2 n i. e.,
polynomials with only even powers of x.
5. Which of the following sets of vectors α = (a1, … , an) in R n are subspaces of
R n ? (n ≥ 3).
(i) all α such that a1 ≥ 0 ;
(ii) all α such that a1 + 3 a2 = a3 ;
2
(iii) all α such that a2 = a1 ; (Kumaun 2011)
(iv) all α such that a1 a2 = 0 ;
(v) all α such that a2 is rational.
6. Let F be the field of integers modulo 2, V be the set of all 2 × 2 matrices over F.
Show that V ( F ) is a finite vector space. Give two non-trivial subspaces of this
vector space.
7. Let V be the vector space of all n × n square matrices over a field F. Show that W is
a subspace of V if W consists of all matrices which are
(i) symmetric i. e., all matrices A = [aij] for which a ji = aij.
(ii) skew-symmetric.
(iii) upper triangular.
(iv) diagonal.
(v) scalar.
V-22

8. Let V be a vector space of all real n × n matrices. Prove that the set W consisting of
all n × n real matrices which commute with a given matrix T of V forms a subspace
of V.
9. Let V be the vector space of all 2 × 2 matrices over the real field R . Show that the
subset of V consisting of all matrices A for which A2 = A is not a subspace of V.
10. Determine whether or not W is a subspace of R 3 if W consists of those vectors
(a, b, c ) ∈R 3 for which :
(i) a = 2 b (ii) ab = 0
(iii) a ≤ b ≤ c (iv) a = b2
(v) a2 + b2 + c 2 ≤ 1 (vi) k1 a + k2 b + k3 c = 0 , k i ∈ R .
11. Let V be the (real) vector space of all functions f from R into R . Which of the
following sets of functions are sub-spaces of V ?
(i) all f such that f ( x2 ) = [ f ( x)]2 ;
(ii) all f such that f (0 ) = f (1) ;
(iii) all f such that f (3) = 1 + f (− 5) ;
(iv) all f such that f (− 1) = 0;
(v) all f which are continuous.
12. Let V be the vector space of all functions from the real field R into R . Check
whether W is a subspace of V where
(i) W = { f : f (7) = 2 + f (1) };
(ii) W consists of all bounded functions;
(iii) W consists of all integrable functions in the interval [0 , 1].
13. Let V be the vector space of infinite sequences < a1, a2 , … an, … > in a field F. Show
that W is a subspace of V, if
(i) W consists of all sequences with 0 as the first component ;
(ii) W consists of all sequences with only a finite number of non-zero
components.
14. Let AX = B be a non-homogeneous system of linear equations in n unknowns
over a field F. Show that the solution set W of the system is not a subspace of F n.

A nswers 2
4. (i) not a subspace (ii) a subspace
5. (i) not a subspace (ii) a subspace (iii) not a subspace;
(iv) not a subspace (v) not a subspace
10. (i) a subspace (ii) not a subspace
(iii) not a subspace (iv) not a subspace
(v) not a subspace (vi) a subspace
11. (i) not a subspace (ii) a subspace (iii) not a subspace
(iv) a subspace (v) a subspace
12. (i) not a subspace (ii) a subspace (iii) a subspace
V-23

4 Algebra of Subspaces
Theorem 1: The intersection of any two subspaces W1 and W2 of a vector space V ( F ) is also a
subspace of V ( F ).
(Kumaun 2002, 07, 09; Garhwal 07, 12; Gorakhpur 10, 13, 15)
Proof: Since 0 ∈W1 and W2 both, therefore W1 ∩ W2 is not empty.
Let α , β ∈ W1 ∩ W2 and a , b ∈ F .
Now α ∈ W1 ∩ W2 ⇒ α ∈ W1 and α ∈W2
and β ∈ W1 ∩ W2 ⇒ β ∈ W1 and β ∈W2 .
Since W1 is a subspace, therefore a , b ∈ F and α , β ∈ W1 ⇒ aα + bβ ∈ W1.
Similarly a , b ∈ F and α , β ∈ W2 ⇒ aα + bβ ∈ W2 .
Now aα + bβ ∈ W1, aα + bβ ∈ W2 ⇒ aα + bβ ∈ W1 ∩ W2 .
Thus a , b ∈ F and α , β ∈ W1 ∩ W2 ⇒ aα + bβ ∈ W1 ∩ W2 .
Hence W1 ∩ W2 is a subspace of V ( F ).
Note: The union of two subspaces of V ( F ) may not be a subspace of V ( F ). For
example if R be the field of real numbers, then W1 = {(0 , 0 , z ) : z ∈ R} and
W2 = {(0 , y, 0 ) : y ∈ R} are two subspaces of V3 (R).
We have (0 , 0 , 3) ∈W1 and (0, 5, 0) ∈W2 .
But (0 , 0 , 3) + (0 , 5, 0 ) = (0 , 5, 3) ∉W1 ∪ W2 since neither (0 , 5, 3) ∈W1 nor
(0 , 5, 3) ∈W2 . Thus W1 ∪ W2 is not closed under vector addition. Hence W1 ∪ W2 is
not a subspace of V3 (R). (Gorakhpur 2015)
Theorem 2: The union of two subspaces is a subspace if and only if one is contained in the
other. (Kumaun 2003, 07, 11; Garhwal 06, 07, 13)
Proof: Suppose W1 and W2 are two subspaces of a vector space V.
Let W1 ⊆ W2 or W2 ⊆ W1. Then W1 ∪ W2 = W2 or W1. But W1, W2 are subspaces and
therefore, W1 ∪ W2 is also a subspace.
Conversely, suppose W1 ∪ W2 is a subspace.
To prove that W1 ⊆ W2 or W2 ⊆ W1 .
Let us assume that W1 is not a subset of W2 and W2 is also not a subset of W1.
Now W1 is not a subset of W2 ⇒ ∃ α ∈ W1 and α ∉W2 …(1)
and W2 is not a subset of W1 ⇒ ∃ β ∈ W2 and β ∉W1. …(2)
From (1) and (2), we have α ∈ W1 ∪ W2 and β ∈ W1 ∪ W2 .
Since W1 ∪ W2 is a subspace, therefore α + β is also in W1 ∪ W2 .
But α + β ∈ W1 ∪ W2 ⇒ α + β ∈ W1 or W2 .
Suppose α + β ∈W1 . Since α ∈W1 and W1 is a subspace, therefore (α + β) − α = β is in
W1.
But from (2), we have β ∉W1. Thus we get a contradiction. Again suppose that
α + β ∈W2 . Since β ∈W2 and W2 is a subspace, therefore (α + β) − β = α is in W2 . But
from (1), we have α ∉W2 . Thus here also we get a contradiction. Hence either
W1 ⊆ W2 or W2 ⊆ W1 .
V-24

Theorem 3: Arbitrary intersection of subspaces i.e., the intersection of any family of


subspaces of a vector space is a subspace. (Kumaun 2010, 14)
Proof: Let V ( F ) be a vector space and let {Wt : t ∈ T } be any family of subspaces of
V. Here T is an index set and is such that V t ∈ T , Wt is a subspace of V.
Let U = ∩ Wt = { x ∈ V : x ∈ Wt V t ∈ T }
t ∈T
be the intersection of this family of subspaces of V. Then to prove that U is also a
subspace of V.
Obviously U ≠ ∅ , since at least the zero vector 0 of V is in Wt V t ∈ T .
Now let a , b ∈ F and α , β be any two elements of ∩ Wt .
t ∈T
Then α , β ∈ Wt V t ∈ T . Since each Wt is a subspace of V, therefore
aα + bβ ∈ Wt V t ∈ T .
Thus aα + bβ ∈ ∩ Wt .
t ∈T

Thus a , b ∈ F and α , β ∈ ∩ Wt ⇒ aα + bβ ∈ ∩ Wt .
t ∈T t ∈T

Hence ∩ Wt is a subspace of V ( F ).
t ∈T

Note: In the above theorem we cannot replace intersection by union.

5 Smallest Subspace Containing any Subset of V ( F)


Let V ( F ) be a vector space and S be any subset of V. If U is a subspace of V containing
S and is itself contained in every subspace of V containing S, then U is called the
smallest subspace of V containing S. The smallest subspace of V containing S is
also called the subspace of V generated or spanned by S and we shall denote it by the
symbol {S} or by (S ). It can be easily seen that the intersection of all the subspaces of
V ( F ) containing S is the subspace of V ( F ) generated by S. If { S } = V , then we say
that V is spanned by S.

6 Linear Combination of Vectors and Linear Span of a Set


Linear combination:
Definition: Let V ( F ) be a vector space. If α1, α2 , … , α n ∈ V , then any vector
α = a1 α1 + a2 α2 + … + an α n where a1, a2 , … , an ∈ F
is called a linear combination of the vectors α1, α2 , … , α n .
Linear span:
Definition: Let V ( F ) be a vector space and S be any non-empty subset of V. Then the linear
span of S is the set of all linear combinations of finite sets of elements of S and is denoted by L (S).
Thus we have
L (S) = { a1 α1 + a2 α2 + … + an α n : α1, α2 , … , α n
is any arbitrary finite subset of S and a1, a2 , … , an
is any arbitrary finite subset of F }.
(Kumaun 2009; Garhwal 14)
V-25

Theorem 1: The linear span L (S) of any subset S of a vector space V ( F ) is a subspace of V
generated by S i.e., L (S) = { S }. (Kumaun 2009)
Proof: Let α , β be any two elements of L (S).
Then α = a1 α1 + a2 α2 + … + am α m
and β = b1 β1 + b2 β2 + … + bn β n
where the a’s and b’s are elements of F and the α’s and β’s are elements of S.
If a , b be any two elements of F, then
aα + bβ = a (a1 α1 + a2 α2 + … + am α m )
+ b (b1 β1 + b2 β2 + … + bn β n)
= a (a1 α1) + a (a2 α2 ) + … + a (am α m ) + b (b1 β1) + b (b2 β2 ) + …
…+ b (bn β n)
= (aa1) α1 + (aa2 ) α2 + …+ (aam ) α m + (bb1) β1 + (bb2 ) β2 + …
… + (bbn) β n.
Thus aα + bβ has been expressed as a linear combination of a finite set
α1 , α2 ,…, α m , β1, β2 , …, β n of the elements of S. Consequently aα + bβ ∈ L (S ).
Thus a , b ∈ F and α , β ∈ L (S) ⇒ aα + bβ ∈ L (S).
Hence L (S) is a subspace of V ( F ).
Also each element of S belongs to L (S) because if α r ∈ S, then α r = 1 α r and this
implies that α r ∈ L (S). Thus L (S) is a subspace of V and S is contained in L (S).
Now if W is any subspace of V containing S, then each element of L (S) must be in W
because W is to be closed under vector addition and scalar multiplication. Therefore
L (S) will be contained in W.
Hence L (S) = { S} i.e., L (S) is the smallest subspace of V containing S.
Note 1: Important. Suppose S is a non-empty subset of a vector space V ( F ). Then a
vector α ∈V will be in the subspace of V generated by S if it can be expressed as a
linear combination over F of a finite number of vectors belonging to S.
Note 2: If in any case we are to prove that L (S) = V , then we should prove that
V ⊆ L (S) because L (S) ⊆ V since L (S) is a subspace of V. In order to prove that
V ⊆ L (S), we should prove that each element of V can be expressed as a linear
combination of a finite number of elements of S. Then each element of V will also be
an element of L (S) and we shall have V ⊆ L (S).
Finally V ⊆ L (S) and L (S) ⊆ V ⇒ L (S) = V .
Illustration 1: The subset containing a single element (1, 0, 0) of the vector space V3 ( F )
generates the subspace which is the totality of elements of the form (a , 0 , 0 ).
Illustration 2: The subset {(1, 0 , 0 ), (0 , 1, 0 )} of V3 ( F ) generates the subspace which is the
totality of the elements of the form (a, b, 0 ).
Illustration 3: The subset S = {(1, 0 , 0 ), (0 , 1, 0 ), (0 , 0 , 1)} of V3 ( F ) generates or spans the
entire vector space V3 ( F ) i.e., L (S) = V .
V-26

If (a, b, c ) be any element of V, then


(a, b, c ) = a (1, 0 , 0 ) + b (0 , 1, 0 ) + c (0 , 0 , 1).
Thus (a, b, c ) ∈ L (S).
Hence V ⊆ L (S). Also L (S) ⊆ V . Hence L (S) = V .
Illustration 4: Let V be the vector space of all polynomials over the field F. Let S be the subset
of V consisting of the polynomials f0 , f1, f2 , … , defined by f n = x n, n = 0 , 1, 2 , …

Then V = L (S ).

7 Linear Sum of Two Subspaces


Definition: Let W1 and W2 be two subspaces of the vector space V ( F ) . Then the linear sum
of the subspaces W1 and W2 , denoted by W1 + W2 , is the set of all sums α1 + α2 such that
α1 ∈ W1, α2 ∈ W2 .
Thus W1 + W2 = {α1 + α2 : α1 ∈ W1, α2 ∈ W2}.
For any subspace W of a vector space V, we have W + W = W .
Theorem 1: If W1 and W2 are subspaces of the vector space V ( F ) , then
(i) W1 + W2 is a subspace of V ( F ) .
(Kumaun 2004, 12; Garhwal 10; Gorakhpur 12, 14)
(ii) W1 + W2 = {W1 ∪ W2} i. e., L (W1 ∪ W2 ) = W1 + W2 .
(Kumaun 2002; Gorakhpur 11, 14)
Proof: (i) Let α , β be any two elements of W1 + W2 .
Then α = α1 + α2 and β = β1 + β2 where α1, β1 ∈W1 and α2 , β2 ∈W2 . If a, b ∈ F , we
have
aα + bβ = a (α1 + α2 ) + b ( β1 + β2 ) = (aα1 + bβ1) + (aα2 + bβ2 ).
Since W1 is a subspace of V, therefore a, b ∈ F
and α1, β1 ∈ W1 ⇒ aα1 + bβ1 ∈ W1 .
Similarly aα2 + bβ2 ∈ W2 .
Consequently aα + bβ = (aα1 + bβ1) + (aα2 + bβ2 ) ∈ W1 + W2 .
Thus a , b ∈ F and α , β ∈ W1 + W2 ⇒ aα + bβ ∈ W1 + W2 .
Hence W1 + W2 is a subspace of V ( F ).
(ii) Since W2 contains the zero vector, therefore if α1 ∈W1 , then we can write
α1 = α1 + 0 ∈ W1 + W2 .
Thus W1 ⊆ W1 + W2 .
Similarly W2 ⊆ W1 + W2 .
Hence W1 ∪ W2 ⊆ W1 + W2 .
Therefore W1 + W2 is a subspace of V ( F ) containing W1 ∪ W2 .
Now to prove that W1 + W2 = {W1 ∪ W2} we should prove that
W1 + W2 ⊆ L (W1 ∪ W2 ) and L (W1 ∪ W2 ) ⊆ W1 + W2 .
Let α = α1 + β1 be any element of W1 + W2 . Then α1 ∈W1 and β1 ∈W2 . Therefore
α1, β1 ∈ W1 ∈ W2 . We can write
α1 + β1 = 1α1 + 1β1 .
V-27

Thus α1 + β1 is a linear combination of a finite number of elements


α1, β1 ∈ W1 ∪ W2 .
Therefore α1 + β1 ∈ L (W1 ∪ W2 ).
∴ W1 + W2 ⊆ L (W1 ∪ W2 ).
Also L (W1 ∪ W2 ) is the smallest subspace containing W1 ∪ W2 and W1 + W2 is a
subspace containing W1 ∪ W2 . Therefore L (W1 ∪ W2 ) must be contained in
W1 + W2 . Consequently
L (W1 ∪ W2 ) ⊆ W1 + W2 .
Hence W1 + W2 = L (W1 ∪ W2 ) = {W1 ∪ W2}.
Note: If W1, W2 , … , Wk are subspaces of the vector space V, then their linear sum,
denoted by W1 + W2 + … + Wk , is the set of all sums α1 + α2 + … + α k such that
α i ∈ Wi . It can be proved that W1 + W2 + … + Wk is a subspace of V which contains
each of the subspaces Wi and which is spanned by the union of W1, W2 , … , Wk .
(Kumaun 2006)
Illustration: Let V be the vector space of all 2 × 2 matrices over R. If W1 consists of those
matrices in V whose second row is zero and W2 consists of those matrices in V whose second
column is zero i. e.,
 a b    a 0  
W1 =    : a, b ∈ R  , W2 =    : a, c ∈ R 
 0 0    c 0  
then W1 and W2 are subspaces of V. We have
 a b    a 0  
W1 + W2 =    : a, b, c ∈ R  , W1 ∩ W2 =   0 0  : a ∈ R  .
 c 0     

Example 17: If S, T are subsets of V ( F ), then


(i) S ⊆ T ⇒ L (S ) ⊆ L (T ).
(ii) L (S ∪ T ) = L (S ) + L (T ).
(iii) S is a subspace of V ⇔ L (S ) = S.
(iv) L ( L (S )) = L (S ).
Solution: (i) Let α = a1α1 + a2α2 + … + an α n ∈ L (S ) where {α1, α2 , … , α n} is a
finite subset of S. Since S ⊆ T , therefore {α1, α2 , … , α n} is also a finite subset of T . So
α ∈ L (T ).
Thus α ∈ L (S ) ⇒ α ∈ L (T ).
∴ L (S ) ⊆ L (T ).
(ii) Let α be any element of L (S ∪ T ). Then
α = a1α1 + a2 α2 + … + am α m + b1 β1 + b2 β2 + … + bp β p
where {α1, α2 , … , α m , β1, β2 , … , β p} is a finite subset of S ∪ T such that
{α1, α2 , … , α m} ⊆ S and { β1, β2 , … , β p} ⊆ T .
V-28

Now a1α1 + a2 α2 + … + am α m ∈ L (S )
and b1 β1 + b2 β2 + … + bp β p ∈ L (T ).
Therefore α ∈ L (S ) + L (T ).
Consequently L (S ∪ T ) ⊆ L (S ) + L (T ).
Now let γ be any element of L (S ) + L (T ). Then γ = β + δ where β ∈ L (S ) and
δ ∈ L (T ). Now β will be a linear combination of a finite number of elements of S and δ
will be a linear combination of a finite number of elements of T . Therefore β + δ will be
a linear combination of a finite number of elements of S ∪ T .
Thus β + δ ∈ L (S ∪ T ). Consequently
L (S ) + L (T ) ⊆ L (S ∪ T ).
Hence L (S ∪ T ) = L (S ) + L (T ).
(iii) Suppose S is a subspace of V. Then we are to prove that L (S ) = S.
Let α ∈ L (S ). Then α = a1α1 + a2α2 + … + an α n where a1, … , an ∈ F and α1, … , α n ∈ S.
But S is a subspace of V. Therefore it is closed with respect to scalar multiplication and
vector addition. Hence α = a1α1 + … + an α n ∈ S. Thus
α ∈ L (S ) ⇒ α ∈ S.
Therefore L (S ) ⊆ S. Also S ⊆ L (S ). Therefore L (S ) = S.
Converse. Suppose L (S ) = S. Then to prove that S is a subspace of V. We know
that L (S ) is a subspace of V. Since S = L (S ), therefore S is also a subspace of V.
(iv) L ( L (S )) is the smallest subspace of V containing L (S ). But L (S ) is a subspace
of V. Therefore the smallest subspace of V containing L (S ) is L (S ) itself.
Hence L ( L (S )) = L (S ).
Example 18: In the vector space R3 express the vector (1, − 2, 5) as a linear combination of the
vectors (1, 1, 1), (1, 2, 3) and (2, − 1, 1). (Kumaun 2008)

Solution: Let α = (1, − 2, 5), α1 = (1, 1, 1), α2 = (1, 2, 3), α3 = (2, − 1, 1).
We wish to express α as a linear combination of α1, α2 , α3 .
Let α = a1α1 + a2α2 + a3 α3 where a1, a2 , a3 ∈R.
Then (1, − 2, 5) = a1 (1, 1, 1) + a2 (1, 2, 3) + a3 (2, − 1, 1)
= (a1, a1, a1) + (a2 , 2 a2 , 3 a2 ) + (2 a3 , − a3 , a3 )
= (a1 + a2 + 2 a3 , a1 + 2 a2 − a3 , a1 + 3 a2 + a3 )
∴ a1 + a2 + 2 a3 = 1 …(1)
a1 + 2 a2 − a3 = − 2 …(2)
a1 + 3 a2 + a3 = 5 …(3)
Method 1. The augmented matrix of the system of equations (1), (2) and (3) is
1 1 2 ⋮ 1
[ A : B] =  1 2 −1 ⋮ −2  .
 1 3 1 ⋮ 5 
We shall reduce the matrix [ A : B ] to Echelon form by applying the row
transformations only.
V-29

 1 1 2 ⋮ 1
We have [ A : B] ~  0 1 −3 ⋮ −3  .
 0 2 −1 ⋮ 4 
 1 1 2 ⋮ 1
~  0 1 −3 ⋮ −3  .

 0 0 5 ⋮ 10 
Thus we have rank A = rank [ A : B ] i. e., the system of equations (1), (2), (3) is
consistent and so has a solution i. e., α can be expressed as a linear combination of
α1, α2 , α3 . The equations (1), (2) and (3) are equivalent to equations
a1 + a2 + 2 a3 = 1, a2 − 3 a3 = − 3, 5 a3 = 10 .
Solving for unknowns, we get a3 = 2 , a2 = 3, a1 = − 6.
Hence (1, − 2 , 5) = − 6 (1, 1, 1) + 3 (1, 2 , 3) + 2 (2 , − 1, 1).
Method 2. Eliminating a3 between (1) and (2), we get
3 a1 + 5 a2 = − 3 …(4)
Eliminating a3 between (2) and (3), we get
2 a1 + 5 a2 = 3 …(5)
Solving (4) and (5), we get a1 = − 6, a2 = 3.
Putting a1 = − 6, a2 = 3 in (1), we get a3 = 2.
Hence (1, − 2, 5) = − 6 (1, 1, 1) + 3 (1, 2 , 3) + 2 (2 , − 1, 1).
Example 19: For what value of m, the vector (m, 3, 1) is a linear combination of the vectors
(3, 2, 1) and (2, 1, 0 ) ?
Solution: Let α1 = (3, 2, 1) and α2 = (2, 1, 0 ).
Let (m, 3, 1) = aα1 + bα2 where a, b ∈R.
Then (m, 3, 1) = a (3, 2, 1) + b (2, 1, 0 ) = (3 a, 2 a, a) + (2 b, b, 0 )
= (3 a + 2 b, 2 a + b, a)
∴ 3 a + 2 b = m, 2 a + b = 3, a = 1.
These give a = 1, b = 1, m = 5.
Hence the required value of m is 5.
Example 20: In the vector space R4 determine whether or not the vector (3, 9, − 4, − 2) is a
linear combination of the vectors (1, − 2, 0 , 3), (2, 3, 0 , − 1) and (2 , − 1, 2 , 1).
Solution: Let α = (3, 9, − 4, − 2), α1 = (1, − 2 , 0 , 3),
α2 = (2 , 3, 0 , − 1), α3 = (2 , − 1, 2 , 1).
Let α = aα1 + bα2 + cα3 where a, b, c ∈R.
Then (3, 9, − 4, − 2) = a (1, − 2, 0 , 3) + b (2 , 3, 0 , − 1) + c (2, − 1, 2, 1)
or (3, 9, − 4, − 2) = (a + 2 b + 2 c , − 2 a + 3 b − c , 2 c , 3 a − b + c )
∴ a + 2b + 2c = 3 …(1)
− 2a + 3b − c = 9 …(2)
2c = − 4 …(3)
3a − b + c = − 2 …(4)
V-30

Solving (1), (2) and (3), we get a = 1, b = 3, c = − 2. These values satisfy (4). Thus the
system of equations (1), (2), (3) and (4) has a solution i. e., α can be written as a linear
combination of α1, α2 and α3 .
Also we have α = α1 + 3α2 − 2α3 .
Example 21: In the vector space R3 , let α = (1, 2 , 1), β = (3 , 1, 5), γ = (3 , − 4 , 7). Show that
the subspaces spanned by S = {α , β} and T = {α , β, γ } are the same.(Kumaun 2008, 11)
Solution: First we shall show that the vector γ can be expressed as a linear
combination of the vectors α and β. Let
(3, − 4, 7) = a (1, 2 , 1) + b (3, 1, 5).
Then a + 3 b = 3, 2 a + b = − 4, a + 5 b = 7. Solving the first two equations, we get
a = − 3, b = 2 and these values satisfy the third equation also. Therefore we can write
γ = − 3α + 2β.
Now S ⊆ T ⇒ L (S ) ⊆ L (T ).
Further let δ ∈ L (T ). Then δ can be expressed as a linear combination of the vectors
α , β and γ. In this linear combination the vector γ can be replaced by − 3α + 2β. Thus δ
can be expressed as a linear combination of the vectors α and β. Therefore δ ∈ L (S ).
Thus δ ∈ L (T ) ⇒ δ ∈ L (S ). Therefore L (T ) ⊆ L (S ).
Hence L (T ) = L (S ).
Example 22: Show that (1, 1, 1), (0 , 1, 1) and (0 , 1, − 1) generate R3 .

Solution: We shall show that any vector (a, b, c ) ∈R3 is a linear combination of the
given vectors.
Let (a, b, c ) = x (1, 1, 1) + y (0 , 1, 1) + z (0 , 1, − 1), where x, y, z ∈R.
Then (a, b, c ) = ( x , x + y + z , x + y − z ).
∴ x = a, x + y + z = b, x + y − z = c .
b + c − 2a b−c
These give x = a, y = , z= i. e., the above system has a solution. Thus
2 2
the given vectors generate R3 .

Example 23: Let V be the vector space of all functions from R into R ; let Ve be the subset of even
functions, f (− x) = f ( x) ; let Vo be the subset of odd functions, f (− x) = − f ( x).
(i) Prove that Ve and Vo are subspaces of V.
(ii) Prove that Ve + Vo = V .
(iii) Prove that Ve ∩ Vo = {0}.
Solution: (i) Suppose f e and ge ∈ Ve and a is any scalar i. e., a ∈R. Then
(af e + ge ) (− x) = af e (− x) + ge (− x)
= af e ( x) + ge ( x) = (af e + ge) ( x).
Therefore af e + ge is an even function.
Thus a ∈R and f e , ge ∈ Ve ⇒ af e + ge ∈ Ve .
Hence Ve is a subspace of V.
V-31

Again suppose f o and go ∈ Vo and a is any scalar. Then


(af o + go ) (− x) = af o (− x) + go (− x) = a [− f o ( x)] − go ( x)
= − [af o ( x) + go ( x)] = − (af o + go ) ( x).
Therefore af o + go is an odd function.
Thus a ∈R and f o , go ∈ Vo ⇒ af o + go ∈ Vo .
Hence Vo is a subspace of V.
(ii) Since Ve and Vo are subspaces of V, therefore Ve + Vo is also a subspace of V and
consequently Ve + Vo ⊆ V .
Now let f ∈ V . We shall show that f can be expressed as the sum of an even and an
odd function.
1 1
Let f e ( x) = [ f ( x) + f (− x)] and f o ( x) = [ f ( x) − f (− x)].
1 2 1 2
Then obviously f e is an even function and f o is an odd function. We can easily see
1 1
that f e (− x) = f e ( x) and f o (− x) = − f o ( x).
1 1 1 1
1 1
Now f ( x) = [ f ( x) + f (− x)] + [ f ( x) − f (− x)]
2 2
= f e ( x) + f o ( x) = ( f e + f o ) ( x).
1 1 1 1
Therefore f = f e + f o where f e ∈ Ve , f o ∈ Vo .
1 1 1 1
Thus f ∈ V ⇒ f ∈ Ve + Vo . Therefore V ⊆ Ve + Vo .
Hence V = Ve + Vo .
(iii) Let 0 denote the zero function i. e., 0 ( x) = 0 V x ∈ R.
Then 0 ∈Ve and also 0 ∈Vo .
Let f ∈ Ve and also f ∈ Vo .
Then f (− x) = f ( x) = − f ( x).
∴ 2 f ( x) = 0 ⇒ f ( x) = 0 = 0 ( x).
Therefore f = 0 (zero function).
Thus f ∈ Ve ∩ Vo ⇔ f = 0.
Hence Ve ∩ Vo = { 0 }.

Comprehensive Exercise 3

1. For which value of k will the vector (1, − 2, k ) in R 3 be a linear combination of the
vectors (3, 0 , − 2) and (2, − 1, − 5) ?
2. In the vector space R 4 determine whether or not the vector (3, 9, − 4, 2) is a linear
combination of the vectors (1, − 2, 0 , 3), (2, 3, 0 , − 1) and (2, − 1, 2, 1).
3. Is the vector (3, − 1, 0 , − 1) in the subspace of R 4 spanned by the vectors
(2, − 1, 3, 2), (− 1, 1, 1, − 3) and (1, 1, 9, − 5) ?
4. Is the vector (2, − 5, 3) in the subspace of R 3 spanned by the vectors
(1, − 3, 2),(2, − 4, − 1),(1, − 5, 7)? (Kumaun 2009; Garhwal 13)
V-32

3 1
5. Express the matrix E =   as a linear combination of the matrices
 1 −1 
 1 1  0 0 0 2
A=   ,B=  and C =  .
1 0   1 1  0 −1 
6. Write E as a linear combination of
 1 1  1 1  1 −1 
A=   ,B =   and C = 0 0 
 0 −1   −1 0   
 3 −1   2 1
where (i) E =   (ii) E =  .
 1 −2   −1 −2 
7. Express the polynomial f = x2 + 4 x − 3 over R as a linear combination of the
polynomials
f1 = x2 − 2 x + 5, f2 = 2 x2 − 3 x and f3 = x + 3.
8. Express f as a linear combination of the polynomials
f1 = 2 x2 + 3 x − 4, f2 = x2 − 2 x − 3
where (i) f = 3 x2 + 8 x − 5 (ii) f = 4 x2 − 6 x − 1.
9. Find the condition on a, b and c so that (a, b, c ) ∈R 3 is in the space generated by
(2, 1, 0 ), (1, − 1, 2) and (0 , 3, − 4).
10. Find one vector in R 3 which generates the intersection of W1 and W2 where
W1 = {(a, b, 0 ) : a, b ∈ R } and W2 is the space generated by the vectors (1, 2, 3) and
(1, − 1, 1).
11. If α, β and γ are vectors such that α + β + γ = 0, then α and β span the same
subspace as β and γ.
12. Give an example of a subset W of a vector space V which is not a subspace of V but
for which
(i) W + W = W . (ii) W + W ⊂ W .
13. Let V be the vector space of n × n matrices over a field F. Let W1 and W2 be the
subspaces of upper triangular matrices and lower triangular matrices
respectively. Find
(i) W1 + W2 (ii) W1 ∩ W2 .
14. Show that the space U generated by the vectors u1 = (1, 2, − 1, 3),
u2 = (2, 4,1, − 2) and u3 = (3, 6, 3, − 7) and the space V generated by the vectors
v1 = (1, 2, − 4, 11) and v2 = (2, 4, − 5, 14) are equal.

A nswers 3
1. k = −8 2. No 3. No
4. No 5. E = 3 A − 2B − C
6. (i) E = 2 A − B + 2 C (ii) Not possible
7. f = − 3 f1 + 2 f2 + 4 f3 8. (i) f = 2 f1 − f2 (ii) Not possible
9. 2 a − 4 b − 3 c = 0 10. (− 2, 5, 0 )
13. (i) W1 + W2 = V (ii) space of diagonal matrices
V-33

O bjective T ype Q uestions

Multiple Choice Questions


Indicate the correct answer for each question by writing the corresponding letter from (a),
(b), (c) and (d).
1. The zero vector in the vector space R4 is
(a) (0, 0) (b) (0, 0, 0)
(c) (0, 0, 0, 0) (d) None of these.
2. The number of elements in the vector space of polynomials of degree at most n in
which the coefficients are the elements of the field I ( p) over the field I ( p), p being
a prime number, is
(a) p n +1 (b) p n
(c) p n −1 (d) None of these.
3. If (m, 3, 1) is a linear combination of vectors (3, 2 , 1) and (2 , 1, 0 ) in R3 , then the
value of m is
(a) 2 (b) 3
(c) 5 (d) None of these.
4. A subspace of a vector space V over a field F other than V is called.
(a) zero subspace of V (b) trivial subspace of V
(c) proper subspace of V (d) none of these (Garhwal 2007)
5. Let V be a vector space over the field of F. The intersection of any collection of
subspaces of V is
(a) zero subspace of V (b) trivial subspace of V
(c) a subspace of V (d) none of these (Garhwal 2008)
6. If α1 = (1, 1, 1), α2 = (1, 2, 3) and α3 = (2, − 1, 1) be three vectors in R3 , the vector
(1, − 2, 5) ∈ R3 as a linear combination of α1, α2 , α3 will be
(a) − 6α1 + 3α2 + 2α3 (b) 6α1 − 3α2 + 2α3
(c) 6α1 + 3α2 − 2α3 (d) 6α1 + 3α2 + 2α3
7. If W1 and W2 be two subspaces of the vector space V over a field F, then which
one will always be the subspace of V
(a) W1 ∪ W2 (b) W1 ∩ W2
(c) W1 − W2 (d) none of these

Fill in the Blank(s)


Fill in the blanks “……” so that the following statements are complete and correct.
1. The necessary and sufficient condition for a non-empty subset W of a vector
space V ( F ) to be a subspace of V is
a, b ∈ F and α, β ∈ W ⇒ …… ∈ W .
2. The subset containing a single element (1, 0 , 0 ) of the vector space V3 ( F )
generates the subspace which is the totality of the elements of the form …… .
V-34

3. If W1 and W2 be two subspaces of the vector space V ( F ),


then L (W1 ∪ W2 ) = .... . (Garhwal 2006)

True or False
Write ‘T’ for true and ‘F’ for false statement.
1. A subspace of V3 (R), where R is the real field, must always contain the origin.
2. The set of vectors α = ( x, y) ∈ V2 (R) for which x2 = y2 is a subspace of V2 ( R ).
3. The set of ordered triads ( x, y, z ) of real numbers with x > 0 is a subspace of V3 (R).
4. The set of ordered triads ( x, y, z ) of real numbers with x + y = 0 is a subspace of
V3 (R).
5. The union of any two subspaces W1 and W2 of a vector space V ( F ) is always a
subspace of V ( F ).
6. The intersection of two subspaces of a vector space V is also a subspace of V.
7. If A and B are subsets of a vector space, then A ≠ B ⇒ L ( A) ≠ L ( B).

A nswers

Multiple Choice Questions


1. (c) 2. (a) 3. (c) 4. (c) 5. (c)
6. (a) 7. (b)

Fill in the Blank(s)


1. aα + bβ 2. (a, 0 , 0 ) 3. W1 + W2
True or False
1. T 2. F 3. F 4. T
5. F 6. T 7. F

¨
V-35

2
L inear D ependence of V ectors

1 Linear Dependence and Linear Independence of


Vectors
inear dependence:
L
Definition: Let V ( F ) be a vector space. A finite set {α1, α2 , …, α n} of vectors of V is said to
be linearly dependent if there exist scalars a1, a2 , …, an ∈ F not all of them 0 (some of them may
be zero) such that
a1 α1 + a2 α2 + a3 α3 + … + an α n = 0.
Linear independence:
Definition: Let V ( F ) be a vector space. A finite set {α1, α2 , …, α n} of vectors of V is said to
be linearly independent if every relation of the form
a1 α1 + a2 α2 + … + an α n = 0, ai ∈ F , 1≤ i ≤ n
⇒ ai = 0 for each 1≤ i ≤ n.
Any infinite set of vectors of V is said to be linearly independent if its every finite subset is
linearly independent, otherwise it is linearly dependent.
Note: The empty set ∅ is defined as a linearly independent set.
V-36

2 Some Theorems on Linear Dependence and Linear


Independence
Theorem 1: Let V ( F ) be a vector space. If α1, α2 , … , α n are non-zero vectors ∈V then
either they are linearly independent or some α k , 2 ≤ k ≤ n, is a linear combination of the preceding
ones α1, α2 , … , α k − 1 . (Garhwal 2004; Gorakhpur 11; Kumaun 13)
Proof: If α1, α2 , … , α n are linearly independent we are nothing to prove. So let
α1, α2 , … , α n are linearly dependent. Then there exists a relation of the form
a1α1 + a2 α2 + … + an α n = 0, …(1)
where, not all the scalar coefficients a1, a2 , … , an are 0. Let k be the largest integer for
which ak ≠ 0 , i. e., ak + 1 = 0 , ak + 2 = 0 , … , an = 0 and ak ≠ 0. There is no harm in this
assumption because at the most if an ≠ 0 then k = n.
Also 2 ≤ k. Because if a2 = 0 , a3 = 0 , … , an = 0 then a1α1 = 0 and α1 ≠ 0 ⇒ a1 = 0. This
contradicts the fact that not all the a’s are 0.
Now the relation (1) reduces to
a1α1 + a2 α2 + … + ak α k = 0, where ak ≠ 0
or ak α k = − a1α1 − a2 α2 − … − ak − 1 α k − 1
or ak−1 (ak α k ) = ak−1 (− a1 α1 − a2 α2 − … − ak − 1 α k − 1)
or α k = (− ak−1 a1) α1 + (− ak−1 a2 ) α2 + … + (− ak−1 ak − 1) α k − 1 .
Thus α k is a linear combination of its preceding vectors.
Theorem 2: The set of non-zero vectors α1, α2 , … , α n of V ( F ) is linearly dependent if some
α k , 2 ≤ k ≤ n, is a linear combination of the preceding ones. (Gorakhpur 2012)
Proof: If some α k , 2 ≤ k ≤ n, is a linear combination of the preceding ones
α1, α2 , … , α k − 1 then ∃ scalars a1, a2 , … , ak − 1 such that
α k = a1α1 + … + ak − 1 α k − 1
⇒ 1α k − a1α1 − a2α2 − … − ak − 1 α k − 1 = 0
⇒ the set {α1, α2 , … , α k} is linearly dependent.
Hence the set {α1, α2 , … , α n} of which {α1, α2 , … , α k} is a subset, must be linearly
dependent.
Theorem 3: If in a vector space V ( F ), a vector β is a linear combination of the set of vectors
α1, α2 , α3 , … , α n , then the set of vectors β, α1, α2 , … , α n is linearly dependent.
Proof: Since β is a linear combination of α1, α2 , … , α n , therefore there exist scalars
a1, a2 , … , an such that
β = a1α1 + a2 α2 + … + an α n
⇒ 1β − a1α1 − a2 α2 − … − an α n = 0. …(1)
In the relation (1) the scalar coefficient of β is 1 which is ≠ 0. Hence in the relation (1)
not all the scalar coefficients are 0. Therefore the set β, α1, … , α n is linearly dependent.
Theorem 4: Let S be a linearly independent subset of a vector space V. Suppose β is a vector
in V which is not in the subspace spanned by S. Then the set obtained by adjoining β to S is
linearly independent. (Kumaun 2006)
Proof: Suppose α1, α2 , … , α m are distinct vectors in S. Let
c1 α1 + c2 α2 + … + c m α m + bβ = 0. …(1)
V-37

Then b must be zero, for otherwise


 c   c 
β =  − 1  α1 + … +  − m  α m
 b  b
and consequently β is in the subspace spanned by S which is a contradiction.
Putting b = 0 in (1), we get
c1 α1 + c2 α2 + … + c m α m = 0 ⇒ c1 = 0 , c2 = 0 , … , c m = 0 because
the set {α1, α2 , … , α m} is linearly independent since it is a subset of a linearly
independent set S.
Thus the relation (1) implies c1 = 0 , c2 = 0 , … , c m = 0 , b = 0 .
Therefore the set {α1, α2 , … , α m , β} is linearly independent. If S ′ is the set obtained by
adjoining β to S, then we have proved that every finite subset of S ′ is linearly
independent. Hence S ′ is linearly independent.
Theorem 5: The set of non-zero vectors α1, α2 , … , α n of V ( F ) is linearly dependent iff one
of these vectors is a linear combination of the remaining (n − 1) vectors.
Proof: This theorem can be easily proved.

Example 1: Prove that if two vectors are linearly dependent, one of them is a scalar multiple of
the other. (Gorakhpur 2011; Kumaun 13)
Solution: Let α , β be two linearly dependent vectors of the vector space V. Then ∃
scalars a, b not both zero, such that
aα + bβ = 0.
If a ≠ 0, then we get a α = − bβ
 b
⇒ α =  −  β ⇒ α is a scalar multiple of β.
 a
If b ≠ 0, then we get bβ = − aα
 a
⇒ β =  −  α ⇒ β is a scalar multiple of α.
 b
Thus one of the vectors α and β is a scalar multiple of the other.
Example 2: In the vector space Vn ( F ) , the system of n vectors
e1 = (1, 0 , 0 , …, 0 ), e2 = (0 , 1, 0 , …, 0 , 0 ),…, en = (0 , 0 , …, 0 , 1)
is linearly independent where 1 denotes the unity of the field F.
Solution: If a1, a2 , a3 , …, an be any scalars, then
a1 e1 + a2 e2 + … + an en = 0
⇒ a1 (1, 0 , 0 , …, 0 ) + a2 (0 , 1, 0 , … , 0 ) + … + an (0 , 0 , …, 0 , 1) = 0
⇒ (a1, a2 , …, an) = (0 , 0 , …, 0 )
⇒ a1 = 0 , a2 = 0 , …, an = 0 .
Therefore the given set of n vectors is linearly independent.
In particular {(1, 0 , 0 ), (0 , 1, 0 ), (0 , 0 , 1)} is a linearly independent subset of V3 ( F ).
Example 3: If the set S = {α1, α2 , …, α n} of vectors of V ( F ) is linearly independent, then
none of the vectors α1, α2 , …, α n can be zero vector.
V-38

Solution: Let α r be equal to zero vector where 1≤ r ≤ n.


Then 0 α1 + 0 α2 + … + aα r + 0 α r + 1 + … + 0 α n = 0 for any a ≠ 0 in F.
Since a ≠ 0, therefore from this relation we conclude that S is linearly dependent.
Thus we get a contradiction because it is given that S is linearly independent. Hence
none of the vectors α1, α2 , …, α n can be zero vector. We also conclude that a set of
vectors which contains the zero vector is necessarily linearly dependent.
Example 4: Every superset of a linearly dependent set of vectors is linearly dependent.
Solution: Let S = {α1, α2 , …, α n} be a linearly dependent set of vectors. Then there
exist scalars a1, a2 , …, an not all zero such that
a1 α1 + a2 α2 + … + an α n = 0. …(1)
Now let S ′ = {α1, α2 ,…, α n , β1, β2 , …, β m} be a superset of S.
Then we have from (1)
a1 α1 + a2 α2 + … + an α n + 0 β1 + 0 β2 + … + 0 β m = 0. …(2)
Since in the relation (2) the scalar coefficients are not all 0, therefore S ′ is linearly
dependent.
From this we also conclude that any subset of a linearly independent set of vectors is also
linearly independent.
Example 5: A system consisting of a single non-zero vector is always linearly independent.
Solution: Let S = {α} be a subset of a vector space V and let α be not equal to zero
vector. If a is any scalar, then
aα = 0
⇒ a = 0. [Since α is not-zero vector]
∴ the set S is linearly independent.
Example 6: Show that S = {(1, 2, 4), (1, 0 , 0 ), (0 , 1, 0 ), (0 , 0 , 1)} is a linearly dependent subset of
the vector space V3 (R) where R is the field of real numbers.
Solution: We have
1 (1, 2, 4) + (− 1) (1, 0 , 0 ) + (− 2) (0 , 1, 0 ) + (− 4) (0 , 0 , 1)
= (1, 2, 4) + (− 1, 0 , 0 ) + (0 , − 2, 0 ) + (0 , 0 , − 4) = (0 , 0 , 0 )
i.e., the zero vector.
Since in this relation the scalar coefficients 1, − 1, − 2, − 4 are not all zero, therefore the
given system S is linearly dependent.
Example 7: Show that the system of three vectors (1, 3, 2), (1, − 7, − 8), (2 , 1, − 1) of V3 (R) is
linearly dependent. (Garhwal 2010; Kumaun 10)
Solution: Let a, b, c be scalars i. e., real numbers such that
a (1, 3, 2) + b (1, − 7, − 8) + c (2 , 1, − 1) = (0 , 0 , 0 )
i. e., (a + b + 2 c , 3 a − 7 b + c , 2 a − 8 b − c ) = (0 , 0 , 0 ).
Then a + b + 2 c = 0, …(1)
3a − 7b + c = 0, …(2)
and 2a − 8b − c = 0. …(3)
If the equations (1), (2), (3) possess a non-zero solution, then the given vectors are
linearly dependent.
Adding (2) and (3), we get
5 a − 15 b = 0 or a − 3b = 0. …(4)
V-39

Multiplying (3) by 2 and adding to (1), we get


5 a − 15 b = 0 or a − 3 b = 0 . …(5)
The equations (4) and (5) are the same and give a = 3 b.
Putting a = 3 b in (1), we get 2 c + 4 b = 0 or c = − 2 b.
If we take b = 1, we get a = 3, c = − 2 .
Thus a = 3, b = 1, c = − 2 is a non-zero solution of the equations (1), (2) and (3). Hence
the given set of vectors is linearly dependent.
Alternative method. The coefficient matrix A of the system of equations (1), (2)
and (3) is
1 1 2

A =  3 −7 1  .
 2 −8 −1 
1 1 2 1 0 0
We have det A = 3 −7 1 = 3 −10 −5 by C2 − C1 and C3 − 2 C1
2 −8 −1 2 −10 −5
= 50 − 50 = 0.
∴ rank A < 3 i. e., rank A < the number of unknowns a, b, c in the equations (1),
(2) and (3).
Therefore the equations (1), (2) and (3) must possess a non-zero solution. Hence the
given vectors are linearly dependent.
Example 8: In V3 (R) , where R is the field of real numbers, examine each of the following sets of
vectors for linear dependence :
(i) {(2, 1, 2), (8 , 4, 8)}
(ii) {(1, 2, 0), (0, 3, 1) , (− 1, 0 , 1} )
(iii) {(− 1, 2, 1), (3, 0 , − 1), (− 5, 4, 3)} (Gorakhpur 2010)
(iv) {(2, 3, 5), (4, 9, 25)}
(v) {(1, 2, 1), (3, 1, 5), (3, − 4, 7)}.
Solution: (i) We have 4 (2, 1, 2) + (− 1) (8, 4, 8) = (8, 4, 8) + (− 8, − 4, − 8)
= (0 , 0 , 0 ) i.e., the zero vector.
Since in this relation the scalar coefficients 4, − 1are not both zero, therefore the given
set is linearly dependent.
(ii) Let a, b, c be scalars i.e., real numbers such that
a (1, 2, 0 ) + b (0 , 3, 1) + c (− 1, 0 , 1) = (0 , 0 , 0 )
i.e., (a − c , 2 a + 3 b, b + c ) = (0 , 0 , 0 )
i.e., a + 0 b − c = 0, 2 a + 3 b + 0 c = 0 , 0 a + b + c = 0 .
These equations will have a non-zero solution i.e., a solution in which a, b, c are not all
zero if the rank of the coefficient matrix is less than three i.e., the number of unknowns
a, b, c . If the rank is 3, then the zero solution a = 0, b = 0, c = 0 will be the only
solution.
 1 0 −1 
Coefficient matrix A =  2 3 0  .
 0 1 1 
We have | A| = 1 (3 − 0 ) − 2 (0 + 1) = 1 ≠ 0.
V-40

∴ Rank A = 3. Hence a = 0, b = 0, c = 0 is the only solution. Therefore the given


system is linearly independent.
(iii) Let a, b, c be scalars such that
a (− 1, 2, 1) + b (3, 0 , − 1) + c (− 5, 4, 3) = (0 , 0 , 0 )
i.e., (− a + 3 b − 5 c , 2 a + 0 b + 4 c , a − b + 3 c ) = (0 , 0 , 0 )
i.e., − a + 3 b − 5 c = 0, 2 a + 0 b + 4 c = 0 , a − b + 3 c = 0.
The coefficient matrix of these equations is
 −1 3 −5 
A =  2 0 4  .
 1 −1 3 
We have | A| = − 1 (0 + 4) − 2 (9 − 5) + 1 (12 − 0 ) = 0.
∴ Rank A is < 3 i.e., the number of unknowns a, b, c . Therefore the given system of
equations will possess a non-zero solution. For example a = − 2, b = 1, c = 1 is a
non-zero solution. Hence the given system of vectors is linearly dependent.
(iv) Let a, b be scalars i.e., real numbers such that
a (2, 3, 5) + b (4, 9, 25) = (0 , 0 , 0 )
i.e., (2 a + 4 b, 3 a + 9 b, 5 a + 25 b) = (0 , 0 , 0 )
i.e., 2 a + 4 b = 0 , 3 a + 9 b = 0 , 5 a + 25 b = 0 .
2 4
The coefficient matrix of these equations is A =  3 9  .
 
 5 25 
Obviously rank A = 2 i.e., equal to the number of unknowns a and b.
Therefore these equations have the only solution a = 0, b = 0.
Hence the given set of vectors is linearly independent.
(v) Let a, b, c be scalars i. e., real numbers such that
a (1, 2, 1) + b (3, 1, 5) + c (3, − 4, 7) = (0 , 0 , 0 )
i. e., (a + 3 b + 3 c , 2 a + b − 4 c , a + 5 b + 7 c ) = (0 , 0 , 0 )
i. e., a + 3 b + 3 c = 0, …(1)
2a + b − 4c = 0, …(2)
a + 5 b + 7 c = 0. …(3)
Multiplying (1) by 2, we get
2a + 6b + 6c = 0. …(4)
Subtracting (4) from (2), we get
− 5 b − 10 c = 0
or b + 2 c = 0. …(5)
Again subtracting (3) from (1), we get
− 2b − 4c = 0 or b + 2 c = 0. …(6)
The equations (5) and (6) are the same and give b = − 2 c .
Putting b = − 2 c in (1), we get a = 3 c . If we take c = 1, we get b = − 2 and a = 3. Thus
a = 3, b = − 2, c = 1 is a non-zero solution of the equations (1), (2) and (3). Hence the
given set of vectors is linearly dependent.
Example 9: If F is the field of complex numbers, prove that the vectors (a1, a2 ) and (b1, b2 ) in
V2 ( F ) are linearly dependent iff a1 b2 − a2 b1 = 0 . (Meerut 2003)
V-41

Solution: Let x, y ∈ F . Then


x (a1, a2 ) + y (b1, b2 ) = (0 , 0 ) ⇒ ( xa1 + yb1, xa2 + yb2 ) = (0 , 0 ).
Therefore, a1 x + b1 y = 0 
.
and a2 x + b2 y = 0 
The necessary and sufficient condition for these equations to possess a non-zero
solution is that
a1 b1
= 0 i. e., a1 b2 − a2 b1 = 0 .
a2 b2
Hence the given system is linearly dependent iff a1 b2 − a2 b1 = 0 .
Example 10: If α1 and α2 are vectors of V ( F ) , and a, b ∈ F , show that the set
{α1, α2 , aα1 + bα2} is linearly dependent.
Solution: We have (− a) α1 + (− b) α2 + 1 (aα1 + bα2 )
= (− a + a) α1 + (− b + b) α2
= 0 α1 + 0 α2 = 0 i. e., zero vector.
In the above linear combination the scalar coefficient 1 ≠ 0 . Therefore whatever may
be the scalars − a and − b, the given set of vectors is linearly dependent.
Example 11: If α, β, γ are linearly independent vectors of V ( F ) where F is any subfield of the
field of complex numbers, then so also are α + β, β + γ , γ + α .
(Kumaun 2007, 12 ; Garhwal 13)
Solution: Let a, b, c be scalars such that
a (α + β) + b ( β + γ ) + c (γ + α) = 0
i. e., (a + c ) α + (a + b) β + (b + c ) γ = 0. …(1)
But α, β, γ are linearly independent. Therefore (1) implies
a + 0 b + c = 0, a + b + 0 c = 0, 0 a + b + c = 0 .
The coefficient matrix A of these equations is
1 0 1
A =  1 1 0  .
 0 1 1 
We have rank A = 3 i. e., the number of unknowns a, b, c . Therefore a = 0, b = 0, c = 0 is
the only solution of the given equations.
Hence α + β, β + γ , γ + α are also linearly independent.
Example 12: Show that the set {1, x, 1 + x + x2} is a linearly independent set of vectors in the
vector space of all polynomials over the real number field.
Solution: Let a, b, c be scalars (real numbers) such that
a (1) + bx + c (1 + x + x2 ) = 0 .
We have a (1) + bx + c (1 + x + x2 ) = 0
⇒ (a + c ) + (b + c ) x + c x2 = 0
⇒ a + c = 0, b + c = 0, c = 0
⇒ c = 0, b = 0, a = 0.
∴ the vectors 1, x, 1 + x + x2 are linearly independent over the field of real numbers.

Example 13: In the vector space F [ x] of all polynomials over the field F the infinite set
S = {1, x, x2 , x3 , …} is linearly independent.
V-42
m1 m2 mn
Solution: Let S ′ = {x ,x , …, x } be any finite subset of S having n vectors. Here
m1, m2 ,…, mn are some non-negative integers. Let a1, a2 , …, an be scalars such that
a1 x m1 + a2 x m2 + … + an x mn = 0. (i. e., zero polynomial) …(1)
By the definition of equality of two polynomials we have from (1)
a1 = 0 , a2 = 0 , … , an = 0.
Thus every finite subset of S is linearly independent.
Therefore S is linearly independent.
Example 14: Show that the three vectors (1, 1, − 1), (2 , − 3, 5) and (−2 , 1, 4) of R3 are linearly
independent. (Kumaun 2010, 13)
Solution: Let a, b, c be scalars i. e., real numbers such that
a (1, 1, − 1) + b (2 , − 3, 5) + c (−2 , 1, 4) = (0 , 0 , 0 )
i. e., (a + 2 b − 2 c , a − 3 b + c , − a + 5 b + 4 c ) = (0 , 0 , 0 )
i. e., a + 2b − 2c = 0 …(1)
a − 3b + c = 0 …(2)
− a + 5b + 4c = 0 …(3)
Now we shall solve the simultaneous equations (1), (2) and (3).
Multiplying (2) by 2 and adding to (1), we get
3a − 4b = 0. …(4)
Again multiplying (1) by 2 and adding to (3), we get
a + 9 b = 0. …(5)
Multiplying (5) by 3 and subtracting from (4), we get
−31b = 0 or b = 0 .
Putting b = 0 in (5), we get a = 0.
Now putting a = 0 , b = 0 in (1), we get c = 0.
Thus a = 0 , b = 0 , c = 0 is the only solution of the equations (1), (2) and (3).
∴ a (1, 1, − 1) + b (2 , − 3, 5) + c (−2 , 1, 4) = (0 , 0 , 0 )
⇒ a = 0, b = 0, c = 0.
Hence the vectors (1, 1, − 1), (2 , − 3, 5), (−2 , 1, 4) of R3 are linearly independent.

Example 15: Find a maximal linearly independent subsystem of the system of vectors
α1 = (2, − 2, − 4), α2 = (1, 9, 3), α3 = (− 2, − 4, 1)
and α4 = (3, 7, − 1).
 2 −2 −4 
 1 9 3
Solution: Let A denote the matrix   whose rows consist of the vectors
 −2 −4 1
 3 7 −1 

α1, α2 , α3 and α4 .
We shall reduce the matrix A to Echelon form by applying the row transformations.
We have
 1 −1 −2 
 1 9 3
A~  , applying R1 → 1 R1
 − 2 − 4 1  2
 3 7 − 1 
 
V-43

 1 −1 −2 
0 by R2 → R2 − R1
10 5
~ , R3 → R3 + 2 R1
0 −6 −3 
0 R4 → R4 − 3 R1
 10 5 
1
 1 −1 −2  by R2 → R2
0 10
1 12  1
~ , R3 → − R3
0 1 12  6
0 1 12  1
 R4 → R4
10
 1 −1 −2 
0 1 12  by R3 → R3 − R2
~ ,
0 0 0 R4 → R4 − R2
0 0 0 

which is in Echelon form.
We have rank A = the number of non-zero rows in its Echelon form = 2.
∴ the maximum number of linearly independent row vectors in the matrix A
= the rank A = 2.
The vectors α1 and α2 are linearly independent and so {α1, α2} is a maximal linearly
independent subsystem of the given system of vectors. We observe that none of the
given four vectors is a scalar multiple of any of the remaining three vectors. So any two
of the given four vectors form a maximal linearly independent subsystem of the given
system of vectors.

Comprehensive Exercise 1

1. Show that the vectors (1, 1, 2 , 4), (2 , − 1, − 5, 2), (1, − 1, − 4 , 0 ) and (2 , 1, 1, 6) are
linearly dependent in R 4 .
2. Show that the vectors (1, 1, 0 , 0 ), (0 , 1, − 1, 0 ), (0 , 0 , 0 , 3) in R 4 are linearly
independent.
3. Determine whether the following set of vectors in V3 (Q ) is linearly dependent or
independent, Q being the field of rational numbers :
{(− 1, 2, 1), (3, 1, − 2)}.
4. Find a linearly independent subset T of the set S = {α1, α2 , α3 , α4 }
where α1 = (1, 2, − 1), α2 = (− 3, − 6, 3), α3 = (2, 1, 3), α4 = (8, 7, 7) ∈ R 3
which spans the same space as S.
5. If the vectors (0 , 1, x),( x, 1, 0 ),(1, x, 1) of the vector space R 3 are linearly dependent,
then find the value of x. (Kumaun 2012, 14)
6. Show that the vectors (0 , 2, − 4) , (1, − 2, − 1), (1, − 4, 3) in R 3 are linearly
dependent. Also express (0 , 2, − 4) as a linear combination of (1, − 2, − 1) and
(1, − 4, 3).
V-44

7. Determine if the vectors (1, − 2, 1), (2, 1, − 1), (7, − 4, 1) in R 3 are linearly
independent or linearly dependent.
8. Let α1, α2 , α3 be the vectors of V ( F ), a, b ∈ F . Show that the set {α1, α2 , α3 } is
linearly dependent if the set {α1 + aα2 + bα3 , α2 , α3 } is linearly dependent.
9. If α, β, γ are linearly independent vectors of V ( F ) where F is the field of complex
numbers, then so also are α + β, α − β, α − 2β + γ .
10. Show that the set {1, x, x (1 − x)} is a linearly independent set of vectors in the
space of all polynomials over the real number field.
11. Find whether the vectors
2 x3 + x2 + x + 1, x3 + 3 x2 + x − 2 and x3 + 2 x2 − x + 3 of R [ x] ,
the vector space of all polynomials over the real number field, are linearly
independent or not.
12. Prove that a set of vectors which contains the zero vector is linearly dependent.
13. Let V be the vector space of 2 × 2 matrices over R. Determine whether the
matrices A, B, C ∈ V are linearly dependent, where
 1 2  3 −1   1 −5 
A=   ,B=  ,C =  .
 3 1 2 2  −4 0  (Garhwal 2013)
14. Prove that the three non-zero non-coplanar vectors are linearly independent.
15. Show that any three non-zero coplanar vectors are linearly dependent in R 3 .
16. Show that the set of functions { x, | x |} is L.I. in vector space of continuous
functions defined in (− 1, 1).
17. Show that the vectors α = (1 + i, 2 i), β = (1, 1 + i) in C2 (C) are L.D. but in C2 (R)
are L.I.
18. If V ( F ) is a vector space and α1, α2 , … , α n ∈ V , λ 2 , λ 3 , … , λ n ∈ F
such that {α2 + λ 2 α1, α3 + λ 3 α1, … , α n + λ n α1}
is L.D. then show that {α1, α2 , … , α n} is L.D.
19. Let V be the vector space of functions from R into R. Show that f , g, h ∈ V are
linearly independent where
(i) f ( x) = e2 x , g ( x) = x2 , h ( x) = x
(ii) f ( x) = sin x, g ( x) = cos x, h ( x) = x.
20. Test the vectors (0 , 1, 0 , 1, 1), (1, 0 , 1, 0 , 1), (0 , 1, 0 , 1, 1),(1, 1, 1, 1, 1) in V5 over the field
of rational numbers for linear independence.
21. Prove that the four vectors (1, 0 , 0 ), (0 , 1, 0 ), (0 , 0 , 1), (1, 1, 1) in V3 (C) form a
linearly dependent set but any three of them are linearly independent.

A nswers 1

3. linearly independent 4. {α1, α3 } 5. 0 , ± √ 2


6. (0 , 2, − 4) = (1, − 2, − 1) − (1, − 4, 3) 7. linearly dependent
11. linearly independent 13. Yes 20. linearly dependent
V-45

O bjective T ype Q uestions

Multiple Choice Questions


Indicate the correct answer for each question by writing the corresponding letter from (a),
(b), (c) and (d).
1. The singleton set {x} is linearly independent if and only if
(a) x = 0 (b) x ≠ 0
(c) may or may not be zero (d) none of these
(Garhwal 2006, 10)
2. The set of vectors {(1, 2, 0 ), (0 , 3, 1), (− 1, 0 , 1)} are
(a) linearly independent (b) linearly dependent
(c) neither linearly independent nor dependent
(d) none of these
3. The vectors (a1, a2 ) and (b1, b2 ) are linearly dependent if
(a) a1b1 + a2 b2 = 0 (b) a1b2 + a2 b1 = 0
(c) a1b2 − a2 b1 = 0 (d) a1b1 − a2 b2 = 0
4. Every superset of a linearly dependent set of vectors is
(a) linearly dependent (b) linearly independent
(c) neither linearly dependent nor independent
(d) none of these
5. Three non-zero coplanar vectors are
(a) linearly independent
(b) linearly dependent
(c) neighter dependent nor independent
(d) none of these

Fill in the Blank(s)


Fill in the blanks “……” so that the following statements are complete and correct.
1. Any set of vectors containing the zero vector as a member is linearly …… .
2. Intersection of two linearly independent subsets of a vector space will be linearly
…… .
3. The subset {(1, 0 , 0 ), (0 , 1, 0 ), (0 , 0 , 1)} of the vector space R3 is linearly ……
4. A system consisting of a single non-zero vector is always linearly …… .
5. In the vector space R3 the vectors (1, 0 , 1), (3, 7, 0 ) and (− 1, 0 , − 1) are linearly …… .

True or False
Write ‘T’ for true and ‘F’ for false statement.
1. A set containing a linearly independent set of vectors is itself linearly
independent.
2. Union of two linearly independent subsets of a vector space is linearly
independent.
3. The vectors (2, − 3, 7), (0 , 0 , 0 ), (3, − 1, − 4) are linearly independent in R 3.
4. The vectors (1, − 3, 7), (2, 0 , − 6), (3, − 1, − 1), (2, 4, − 5) are linearly dependent in R3 .
V-46

A nswers

Multiple Choice Questions


1. (b) 2. (a) 3. (c) 4. (a)
5. (b)

Fill in the Blank(s)


1. dependent 2. independent
3. independent 4. independent
5. dependent

True or False
1. F 2. F 3. F 4. T

¨
V-47

3
B asis and D imensions

1 Basis of A Vector Space

D efinition: A subset S of a vector space V ( F ) is said to be a basis of V ( F ) , if

(i) S consists of linearly independent vectors,


(ii) S generates V ( F ) i.e., L (S) = V i.e., each vector in V is a linear combination of a finite
number of elements of S. (Garhwal 2010; Gorakhpur 10)

Example 1: A system S consisting of n vectors,


e1 = (1, 0 , 0 , …, 0 ), e2 = (0 , 1, 0 , …, 0 ), …, en = (0 , 0 , …, 0 , 1)
is a basis of Vn ( F ). (Kumaun 2009)
Solution: First we should show that S is a linearly independent set of vectors. We
have proved it in one of the previous examples.
Now we should prove that L (S) = Vn ( F ). We have always L (S) ⊆ Vn ( F ). So we
should prove that Vn ( F ) ⊆ L (S) i.e., each vector in Vn ( F ) is a linear combination of
elements of S.
V-48

Let α = (a1, a2 , …, an) be any vector in Vn ( F ).


We can write
(a1, a2 , …, an) = a1 (1, 0 , …, 0 ) + a2 (0 , 1, …, 0 ) + … + an (0 , 0 , …, 0 , 1)
i.e., α = a1 e1 + a2 e2 + … + an en.
Hence S is a basis of Vn ( F ). We shall call this particular basis the standard basis of
Vn ( F ).
Note: The set {(1, 0 ), (0 ,1)} is a basis of V2 ( F ). The set {(1, 0 , 0 ), (0 , 1, 0 ), (0 , 0 , 1}
) is a
basis of V3 ( F ). As a particular case a basis of F ( F ) is the set consisting of only the
unity element of F.
Example 2: Show that the infinite set S = {1, x , x2 , …, x n,…} is a basis of the vector space
F [ x] of polynomials over the field F.
Solution: First we should prove that S is a linearly independent set of vectors. For
proof refer some previous example.
Now we should show that S spans F [ x] i.e., each polynomial in F [ x] can be expressed
as a linear combination of a finite number of elements of S.
Let f ( x) = a0 + a1 x + a2 x 2 + … + at x t be a polynomial of degree t.
Then f ( x) = (a0 ) 1 + a1 x + a2 x 2 + … + at x t . Hence S is a basis of F [ x].
Note: The vector space F [ x] has no finite basis. If we take any finite set W of
polynomials, we can find a polynomial of degree greater than that of each of them.
Such a polynomial cannot at any cost be expressed as a linear combination of the
elements of W.

2 Finite Dimensional Vector Space


(Garhwal 2003, 07)
Definition: The vector space V ( F ) is said to be finite dimensional or finitely
generated if there exists a finite subset S of V such that V = L (S) .
The vector space Vn ( F ) of n-tuples is a finite dimensional vector space.
The vector space F [ x] of all polynomials over a field F is not finite dimensional. There
exists no finite subset S of F [ x] which spans F [ x]. A vector space which is not finitely
generated may be referred to as an infinite dimensional space. Thus the vector space F [ x] of
all polynomials over a field F is infinite dimensional.
Note: The dimension of the zero space {0} is defined to be zero. It is generated by ∅.

3 Existence of Basis of a Finite Dimensional Vector Space


Theorem: There exists a basis for each finite dimensional vector space.
(Garhwal 2006, 10, 12, 14; Kumaun 06; Gorakhpur 10)
Proof: Let V ( F ) be a finitely generated vector space.
Let S = {α1, α2 , … , α m}
be a finite subset of V such that L (S ) = V . We may suppose that no member of S is 0.
If S is linearly independent, then S itself is a basis of V.
V-49

If S is linearly dependent, then there exists a vector α i ∈ S which can be expressed as a


linear combination of the preceding vectors α1, α2 , … , α i − 1.
If we omit this vector α i from S, then the remaining set S ′ of m − 1 vectors
α1, α2 , … , α i − 1, α i + 1, … , α m
also generates V i. e., V = L (S ′ ). For if α is any element of V, then L (S ) = V implies
that α can be written as a linear combination of α1, α2 , … , α m . Let
α = a1α1 + a2 α2 + … + ai − 1 α i − 1 + ai α i + ai + 1 α i + 1 + … + am α m .
But α i can be expressed as a linear combination of α1, … , α i − 1. Let
α i = b1α1 + … + bi − 1 α i − 1.
Putting this value of α i in the expression for α, we get
α = a1α1 + … + ai−1 α i − 1 + ai (b1α1 + … + bi − 1 α i − 1) + ai + 1 α i + 1
+ … + am α m .
Thus α has been expressed as a linear combination of the vectors
α1, α2 , … , α i − 1, α i+1, … , α m .
In this way α ∈ V ⇒ α can be expressed as a linear combination of the vectors
belonging to the set S ′ .
Thus S ′ generates V i. e., L (S ′ ) = V .
If S ′ is linearly independent, then S ′ will be a basis of V. If S ′ is linearly dependent,
then proceeding as above we shall get a new set of n − 2 vectors which generates V.
Continuing this process, we shall, after finite number of steps, obtain a linearly
independent subset of S which generates V and which is therefore a basis of V.
At the most it may happen that we shall be left with a subset of S which contains only
one non-zero vector and which spans V. We know that a set containing a single
non-zero vector is definitely linearly independent and so it will form a basis of V.
Note: The above theorem may also be stated as below :
If a finite set S of vectors spans a finite dimensional vector space V ( F ), there exists a subset of S
which forms a basis of V.

4 Invariance of Number of Elements in the Basis of a Finite


Dimensional Vector Space
Dimension theorem for vector spaces: If V ( F ) is a finite dimensional vector space,
then any two bases of V have the same number of elements.
(Garhwal 2004; Gorakhpur 11, 14)
Proof: Suppose V ( F ) is a finite dimensional vector space. Then V definitely
possesses a basis. Let
S1 = {α1, α2 , … , α m}
and S2 = { β1, β2 , … , β n}
be two bases of V. We shall prove that m = n.
Since V = L (S1) and β1 ∈V, therefore β1 can be expressed as a linear combination of
α1, α2 , … , α m . Consequently the set S3 = { β1, α1, α2 , … , α m } which also obviously
generates V ( F ) is linearly dependent. Therefore there exists a member α i ≠ β1 of this
set S3 such that α i is a linear combination of the preceding vectors
β1, α1, α2 , … , α i − 1.
V-50

If we omit the vector α i from S3 then V is also generated by the remaining set
S4 = { β1, α1, α2 , … , α i − 1, α i + 1, … , α m }.
Since V = L (S4 ) and β2 ∈V, therefore β2 can be expressed as a linear combination of
the vectors belonging to S4 . Consequently the set
S5 = { β2 , β1, α1, α2 , … , α i − 1, α i + 1, … , α m }
is linearly dependent. Therefore there exists a member α j of this set S5 such that α j is a
linear combination of the preceding vectors. Obviously α j will be different from β1 and
β2 since { β1, β2} is a linearly independent set. If we exclude the vector α j from S5 , then
the remaining set will generate V ( F ).
We may continue to proceed in this manner. Here each step consists in the exclusion of
an α and the inclusion of a β in the set S1.
Obviously the set S1 of α’s cannot be exhausted before the set S2 of β’s otherwise V ( F )
will be a linear span of a proper subset of S2 and thus S2 will become linearly
dependent. Therefore we must have
m< | n.
Interchanging the roles of S1 and S2 , we shall get that
n <| m.
Hence n = m.
Illustration: For the vector space V3 ( F ), the sets
S1 = {(1, 0 , 0 ), (0 , 1, 0 ), (0 , 0 , 1)}
and S2 = {(1, 0 , 0 ), (1, 1, 0 ), (1, 1, 1)}
are bases as can easily be seen. Both these bases contain the same number of elements
i. e., 3.

5 Dimension of a Finitely Generated Vector Space


Definition: The number of elements in any basis of a finite dimensional vector space V ( F ) is
called the dimension of the vector space V ( F ) and will be denoted by dim V.
(Garhwal 2010)
The vector space Vn ( F ) is of dimension n. The vector space V3 ( F ) is of dimension 3.
If a field F is regarded as a vector space over F, then F will be of dimension 1 and the
set S = {1} consisting of unity element of F alone is a basis of F. In fact every non-zero
element of F will form a basis of F.

6 Some Properties of Finite Dimensional Vector Spaces


Theorem 1: Extension theorem. Every linearly independent subset of a finitely
generated vector space V ( F ) forms a part of a basis of V.
Or
Every linearly independent subset of a finitely generated vector space V ( F ) is either a basis of V
or can be extended to form a basis of V. (Meerut 2003)
Proof: Let S = {α1, α2 , … , α m} be a linearly independent subset of a finite
dimensional vector space V ( F ). If dim V = n, then V has a finite basis, say
{ β1, β2 , … , β n}. Consider the set
S1 = {α1, α2 , … , α m , β1, β2 , … , β n}.
V-51

Obviously L (S1) = V . Since the α’s can be expressed as linear combinations of the β’s
therefore the set S1 is linearly dependent.
Therefore there is some vector of S1 which is a linear combination of its preceding
vectors. This vector cannot be any of the α’s since the α’s are linearly independent.
Therefore this vector must be some β, say β i . Now omit the vector β i from S1 and
consider the set
S2 = {α1, α2 , … , α m , β1, β2 , … , β i − 1, β i + 1, … , β n}.
Obviously L (S2 ) = V . If S2 is linearly independent, then S2 will be a basis of V and it is
the required extended set which is a basis of V. If S2 is not linearly independent, then
repeating the above process a finite number of times, we shall get a linearly
independent set containing α1, α2 , … , α m and spanning V. This set will be a basis of V
and it will contain S. Since each basis of V contains the same number of elements,
therefore exactly n − m elements of set of β’s will be adjoined to S so as to form a basis
of V.
Theorem 2: Each set of (n + 1) or more vectors of a finite dimensional vector space V ( F ) of
dimension n is linearly dependent.
Proof: Let V ( F ) be a finite dimensional vector space of dimension n. Let S be a
linearly independent subset of V containing (n + 1) or more vectors. Then S will form a
part of basis of V. Thus we shall get a basis of V containing more than n vectors. But
every basis of V will contain exactly n vectors. Hence our assumption is wrong.
Therefore if S contains (n + 1) or more vectors, then S must be linearly dependent.
Theorem 3: Let V be a vector space which is spanned by a finite set of vectors β1, β2 , … , β m .
Then any linearly independent set of vectors in V is finite and contains no more than m vectors.
(Garhwal 2009; Kumaun 14)
Proof: Let S = { β1, β2 , … , β m}.
Since L (S ) = V , therefore V has a finite basis and dim V ≤ m. Hence every subset S ′ of
V which contains more than m vectors is linearly dependent. This proves the theorem.
Theorem 4: If V ( F ) is a finite dimensional vector space of dimension n, then any set of n
linearly independent vectors in V forms a basis of V.
Proof: Let {α1, α2 , … , α n} be a linearly independent subset of a finite dimensional
vector space V ( F ) of dimension n. If S is not a basis of V, then it can be extended to
form a basis of V. Thus we shall get a basis of V containing more than n vectors.
But every basis of V must contain exactly n vectors. Therefore our
assumption is wrong and S must be a basis of V.
Theorem 5: If a set S of n vectors of a finite dimensional vector space V ( F ) of dimension n
generates V ( F ), then S is a basis of V.
Proof: Let V ( F ) be a finite dimensional vector space of dimension n. Let
S = {α1, α2 , … , α n} be a subset of V such that L (S ) = V . If S is linearly independent,
then S will form a basis of V. If S is not linearly independent, then there will exist a
proper subset of S which will form a basis of V. Thus we shall get a basis of V
containing less than n elements. But every basis of V must contain exactly n elements.
Hence S cannot be linearly dependent and so S must be a basis of V.
Note: If V is a finite dimensional vector space of dimension n, then V cannot be generated by
fewer than n vectors.
V-52

Dimension of a Subspace:
Theorem 6: Each subspace W of a finite dimensional vector space V ( F ) of dimension n is a
finite dimensional space with dim m ≤ n.
Also V = W iff dim V = dim W . (Garhwal 2008; Kumaun 10, 12)
Proof: Let V ( F ) be a finite dimensional vector space of dim n. Let W be a subspace
of V. Any subset of W containing (n + 1) or more vectors is also a subset of V and any
(n + 1) vectors in V are linearly dependent. Therefore any linearly independent set of
vectors in W can contain, at the most n vectors. Let
S = {α1, α2 , … , α m}
be a linearly independent subset of W with maximum number of elements. We claim
that S is a basis of W. The proof is as follows :
(i) S is a linearly independent subset of W.
(ii) L (S ) = W . Let α be any element of W.
Then the (m + 1) vectors α, α1, α2 , … , α m belonging to W are linearly dependent
because we have supposed that the largest independent subset of W contains m
vectors.
Now {α1, α2 , … , α m , α} is a linearly dependent set. Therefore there exists a vector
belonging to it which can be expressed as a linear combination of the preceding
vectors. Since α1, α2 , … , α m are linearly independent, therefore this vector cannot be
any of these m vectors. So it must be α itself. Thus α can be expressed as a linear
combination of α1, α2 , … , α m . Hence L (S ) = W .
∴ S is a basis of W.
∴ dim W = m and m ≤ n.
Now if V = W , then every basis of V is also a basis of W.
Hence dim V = dim W = n.
Conversely let dim W = dim V = n. Then to prove that W = V .
Let S be a basis of W. Then L (S ) = W and S contains n vectors. Since S is also a subset
of V and S contains n linearly independent vectors, therefore S will also be a basis of V.
Therefore L (S ) = V . Hence W = V . We thus conclude :
If W is a proper subspace of a finite-dimensional vector space V, then W is finite dimensional and
dim W < dim V. (Garhwal 2008)

Theorem 7: If W is a subspace of a finite-dimensional vector space V, every linearly


independent subset of W is finite and is part of a ( finite) basis for W.
Proof: Let dim V = n. Let W be a subspace of V. Let S0 be a linearly independent
subset of W. Let S be a linearly independent subset of W which contains S0 and which
has maximum number of elements. Then S is also a linearly independent subset of V.
So S will have at the most n elements. Therefore S is finite and consequently S0 is
finite.
Now our claim is that S is a basis of W. The proof is as follows :
(i) S is a linearly independent subset of W.
(ii) L (S ) = W . Because if β ∈W, then β must be in the linear span of S. If β is not in the
linear span of S, then the subset of W obtained by adjoining β to S will be linearly
independent. Thus S will not remain the maximum linearly independent subset of W
containing S0 . Hence β ∈ W ⇒ β ∈ L (S ). Thus W ⊆ L (S ). Since S is a subset of the
subspace W, therefore L (S ) ⊆ W .
V-53

Hence L (S ) = W .
Thus S is a finite basis of W and S0 ⊆ S.
Theorem 8: Let S = {α1, α2 , … , α n} be a basis of a finite dimensional vector space V ( F ) of
dimension n. Then every element α of V can be uniquely expressed as
α = a1α1 + a2 α2 + … + an α n where a1, a2 , … , an ∈ F .
Proof: Since S is a basis of V, therefore L (S ) = V . Therefore any vector α ∈V can be
expressed as
α = a1α1 + a2 α2 + … + an α n .
To show uniqueness let us suppose that
α = b1α1 + b2 α2 + … + bn α n .
Then we must show that a1 = b1, a2 = b2 , … , an = bn .
We have a1α1 + a2 α2 + … + an α n = b1α1 + b2 α2 + … + bn α n
⇒ (a1 − b1) α1 + (a2 − b2 ) α2 + … + (an − bn) α n = 0
⇒ a1 − b1 = 0 , a2 − b2 = 0 , … , an − bn = 0 ,
since α1, α2 , … , α n are linearly independent
⇒ a1 = b1, a2 = b2 , … , an = bn .
Hence the theorem.
Theorem 9: If W1, W2 are two subspaces of a finite dimensional vector space V ( F ), then
dim (W1 + W2 ) = dim W1 + dim W2 − dim (W1 ∩ W2 ).
(Kumaun 2004, 13; Garhwal 08, 13; Gorakhpur 15)
Proof: Let dim (W1 ∩ W2 ) = k and let the set
S = { γ1, γ 2 , γ 3 , … , γ k}
be a basis of W1 ∩ W2 . Then S ⊆ W1 and S ⊆ W2 .
Since S is linearly independent and S ⊆ W1, therefore S can be extended to form a
basis of W1. Let
{ γ1, γ 2 , … , γ k , α1, α2 , … , α m}
be a basis of W1. Then dim W1 = k + m. Similarly let
{ γ1, γ 2 , … , γ k , β1, β2 , … , β t}
be a basis of W2 . Then dim W2 = k + t.
∴ dim W1 + dim W2 − dim (W1 ∩ W2 ) = (m + k ) + (k + t) − k
= k + m + t.
∴ to prove the theorem we must show that
dim (W1 + W2 ) = k + m + t.
We claim that the set
S1 = { γ1, γ 2 , … , γ k , α1, α2 , … , α m , β1, β2 , … , β t}
is a basis of W1 + W2 .
First we show that S1 is linearly independent. Let
c1γ1 + c2 γ 2 + … + c k γ k + a1α1 + a2 α2 + … + am α m
+ b1 β1 + b2 β2 + … + bt β t = 0 …(1)

⇒ b1 β1 + b2 β2 + … + bt β t = − (c1 γ1 + … + c k γ k + a1α1 + … + am α m ) …(2)


V-54

Now − (c1γ1 + … + c k γ k + a1α1 + … + am α m ) ∈ W1 since it is a linear combination


of elements of a basis of W1 . Again
b1 β1 + b2 β2 + … + bt β t ∈ W2
since it is a linear combination of elements belonging to a basis of W2 .
Also by virtue of the equality (2), b1 β1 + … + bt β t ∈ W1. Therefore
b1 β1 + b2 β2 + … + bt β t ∈ W1 ∩ W2 .
Therefore it can be expressed as a linear combination of elements of the basis of
W1 ∩ W2 . Thus we have a relation of the form
b1 β1 + b2 β2 + … + bt β t = d1 γ1 + … + dk γ k
⇒ b1 β1 + b2 β2 + … + bt β t − d1 γ1 − d2 γ 2 − … − dk γ k = 0.
But β1, β2 , … , β t , γ1, … , γ k are linearly independent vectors.
Therefore we must have
b1 = 0 , b2 = 0 , … , bt = 0 .
Putting these values of b’s in (1), it reduces to
c1γ1 + c2 γ 2 + … + c k γ k + a1α1 + a2 α2 + … + am α m = 0
⇒ c1 = 0 , c2 = 0 , … , c k = 0 , a1 = 0 , a2 = 0 , … , am = 0
since the vectors γ1, γ 2 , … , γ k , α1, α2 , … , α m are linearly independent.
Thus the relation (1) implies that
c1 = 0 , c2 = 0 , … , c k = 0 , a1 = 0 , … , am = 0 , b1 = 0 , … , bt = 0 .
Therefore the set S1 of vectors
γ1, … , γ k , α1, … , α m , β1 , … , β t
is linearly independent.
Now to show that L (S1) = W1 + W2 .
Since W1 + W2 is a subspace of V and each element of S1 belongs to W1 + W2 ,
therefore L (S1) ⊆ W1 + W2 .
Again let α be any element of W1 + W2 . Then
α = some element of W1 + some element of W2
= a linear combination of elements of basis of W1
+ a linear combination of elements of basis of W2
= a linear combination of elements of S1.
∴ α ∈ L (S1). Hence W1 + W2 ⊆ L (S1).
∴ L (S1) = W1 + W2 .
∴ S1 is a basis of W1 + W2 and consequently
dim (W1 + W2 ) = k + m + t.
Hence the theorem.

Example 3: Let V be the vector space of all 2 × 2 matrices over the field F. Prove that V has
dimension 4 by exhibiting a basis for V which has 4 elements.
(Garhwal 2006; Kumaun 11)
V-55

1 0 0 1 0 0  0 0 
Solution: Let α =   ,β =   ,γ =   and δ =  
0 0  0 0  1 0 0 1
be four elements of V.
The subset S = {α, β, γ , δ } of V is linearly independent because
aα + bβ + c γ + d δ = 0
 1 0 0 1 0 0  0 0  0 0 
⇒ a  +b  +c  +d = 
0 0  0 0  1 0 0 1 0 0 
 a b  0 0 
⇒  c d = 0 0 
   
⇒ a = 0, b = 0, c = 0, d = 0.
a b
Also L (S) = V because if   is any vector in V, then we can write
 c d
a b
 c d  = aα + bβ + cγ + dδ.
 
Therefore S is a basis of V. Since the number of elements in S is 4, therefore
dim V = 4.
Example 4: Show that if S = {α , β, γ } is a basis of C 3 (C), then the set
S ′ = {α + β, β + γ , γ + α}
is also a basis of C 3 (C). (Kumaun 2011)
Solution: The vector space C 3 (C) is of dimension 3. Any subset of C 3 having three
linearly independent vectors will form a basis of C 3 . We have shown in one of the
previous examples that if S is a linearly independent subset of C 3 , then S ′ is also a
linearly independent subset of C 3 .
Therefore S ′ is also a basis of C 3 .

Example 5: Let V be the vector space of ordered pairs of complex numbers over the real field R
i. e., let V be the vector space C 2 (R). Show that the set
S = {(1, 0 ), (i, 0 ), (0 , 1), (0 , i)}
is a basis for V.
Solution: S is linearly independent. We have
a (1, 0 ) + b (i, 0 ) + c (0 , 1) + d (0 , i) = (0 , 0 ) where a, b, c , d ∈R
⇒ (a + ib, c + id ) = (0 , 0 )
⇒ a + ib = 0 , c + id = 0
⇒ a = 0, b = 0, c = 0, d = 0.
Therefore S is linearly independent.
Now we shall show that L (S ) = V . Let any ordered pair (a + ib, c + id ) ∈ V where
a, b, c , d ∈R. Then as shown above we can write
(a + ib, c + id ) = a (1, 0 ) + b (i, 0 ) + c (0 , 1) + d (0 , i).
Thus any vector in V is expressible as a linear combination of elements of S. Therefore
L (S ) = V and so S is a basis for V.
V-56

Example 6: Show that the vectors (1, 2, 1), (2, 1, 0 ), (1, − 1, 2) form a basis of R3 .
(Garhwal 2010; Kumaun 14)
Solution: We know that the set {(1, 0 , 0 ), (0 , 1, 0 ), (0 , 0 , 1)} forms a basis for R3 .
Therefore dim R3 = 3. If we show that the set S = {(1, 2, 1), (2, 1, 0 ), (1, − 1, 2)} is linearly
independent, then this set will also form a basis for R3 .
[See theorem 4 of article 3]
We have a1 (1, 2, 1) + a2 (2, 1, 0 ) + a3 (1, − 1, 2) = (0 , 0 , 0 )
⇒ (a1 + 2 a2 + a3 , 2 a1 + a2 − a3 , a1 + 2 a3 ) = (0 , 0 , 0 ).
∴ a1 + 2 a2 + a3 = 0 …(1)
2 a1 + a2 − a3 = 0 …(2)
a1 + 2 a3 = 0 . …(3)
Now we shall solve these equations to get the values of a1, a2 , a3 . Multiplying the
equation (2) by 2, we get
4 a1 + 2 a2 − 2 a3 = 0 . …(4)
Subtracting (4) from (1), we get
− 3 a1 + 3 a3 = 0
or − a1 + a3 = 0. …(5)
Adding (3) and (5), we get 3 a3 = 0 or a3 = 0 . Putting a3 = 0 in (3), we get a1 = 0 . Now
putting a3 = 0 and a1 = 0 in (1), we get a2 = 0 .
Thus solving the equations (1), (2) and (3), we get a1 = 0 , a2 = 0 , a3 = 0 . Therefore the
set S is linearly independent. Hence it forms a basis for R3 .
Example 7: Determine whether or not the following vectors form a basis of R3 :
(1, 1, 2), (1, 2, 5), (5, 3, 4). (Kumaun 2013)
Solution: We know that dim R3 = 3. If the given set of vectors is linearly
independent, it will form a basis of R3 otherwise not. We have
a1 (1, 1, 2) + a2 (1, 2, 5) + a3 (5, 3, 4) = (0 , 0 , 0 )
⇒ (a1 + a2 + 5 a3 , a1 + 2 a2 + 3 a3 , 2 a1 + 5 a2 + 4 a3 ) = (0 , 0 , 0 ).
a1 + a2 + 5 a3 = 0  ...(1)

∴ a1 + 2 a2 + 3 a3 = 0  . ...(2)
2 a1 + 5 a2 + 4 a3 = 0  ...(3)
Now we shall solve these equations to get the values of a1, a2 , a3 . Subtracting (2) from
(1), we get
− a2 + 2 a3 = 0. …(4)
Multiplying (1) by 2, we get
2 a1 + 2 a2 + 10 a3 = 0 . …(5)
Subtracting (5) from (3), we get
3 a2 − 6 a3 = 0
or a2 − 2 a3 = 0 . …(6)
We see that the equations (4) and (6) are the same and give a2 = 2 a3 . Putting
a2 = 2 a3 in (1), we get a1 = − 7 a3 . If we put a3 = 1, we get a2 = 2 and a1 = − 7. Thus
a1 = − 7, a2 = 2, a3 = 1 is a non-zero solution of the equations (1), (2) and (3). Hence
the given set is linearly dependent so it does not form a basis of R3 .
V-57

Example 8: Show that the vectors α1 = (1, 0 , − 1), α2 = (1, 2 , 1), α3 = (0 , − 3, 2) form a basis
for R3 . Express each of the standard basis vectors as a linear combination of α1, α2 , α3 .
(Gorakhpur 2013)
Solution: Let S = {α1, α2 , α3}.
First we shall show that the set S is linearly independent. Let a, b, c be scalars i. e., real
numbers such that
aα1 + bα2 + cα3 = 0
i. e., a (1, 0 , − 1) + b (1, 2 , 1) + c (0 , − 3, 2) = (0 , 0 , 0 )
i. e., (a + b + 0 c , 0 a + 2 b − 3 c , − a + b + 2 c ) = (0 , 0 , 0 )
i. e., a+ b=0 …(1)
2b − 3c = 0 …(2)
− a + b + 2 c = 0. …(3)
Adding both sides of the equations (1) and (3), we get
2b + 2c = 0 …(4)
Subtracting (2) from (4), we get
5 c = 0 or c = 0.
Putting c = 0 in (2), we get b = 0 and then putting b = 0 in (1), we get a = 0.
Thus a = 0 , b = 0 , c = 0 is the only solution of the equations (1), (2) and (3) and so
aα1 + bα2 + cα3 = 0
⇒ a = 0, b = 0, c = 0.
∴ The vectors α1, α2 , α3 are linearly independent.
We have dim R3 = 3. Therefore the vectors α1, α2 , α3 form a basis of R3 .
Now let γ = ( p, q, r) be any vector in R3 and let
γ = ( p, q, r) = xα1 + yα2 + zα3 , x, y, z ∈ R
= x (1, 0 , − 1) + y (1, 2 , 1) + z (0 , − 3, 2). …(5)
Then x+ y=p …(6)
2 y − 3z = q …(7)
− x + y + 2z = r …(8)
Adding both sides of equations (6) and (8), we get
2 y + 2 z = p + r. …(9)
Subtracting (7) from (9), we get
5z = p + r − q
1 1 1
or z = p − q + r.
5 5 5
Then from (9), we get
1 1
y = − z + p+ r
2 2
3 1 3
= p+ q + r
10 5 10
and from (6), we get
x = p− y
V-58

3 1 3
= p− p− q − r
10 5 10
7 1 3
= p− q − r.
10 5 10
Thus, the relation (5) expresses the vector γ = ( p, q, r) as a linear combination of α1, α2
and α3 where x , y, z ∈R are as found above.
The standard basis vectors are
e1 = (1, 0 , 0 ), e2 = (0 , 1, 0 ) and e3 = (0 , 0 , 1).
If γ = e1, then p = 1, q = 0 , r = 0 and so
7 3 1
x= , y= and z = ⋅
10 10 5
7 3 1
∴ e1 = α1 + α2 + α3 .
10 10 5
If γ = e2 , then p = 0 , q = 1, r = 0 and so
1 1 1
x = − , y = and z = − ⋅
5 5 5
1 1 1
∴ e2 = − α1 + α2 − α3 .
5 5 5
Finally if γ = e3 , then p = 0 , q = 0 , r = 1 and so
3 3 1
x=− , y= and z = ⋅
10 10 5
3 3 1
∴ e3 = − α1 + α2 + α3 .
10 10 5

Example 9: Show that a system X consisting of the vectors


α1 = (1, 0 , 0 , 0 ), α2 = (0 , 1, 0 , 0 ), α3 = (0 , 0 , 1, 0 ) and α4 = (0 , 0 , 0 , 1)
is a basis set of R4 (R).
Solution: First we show that the set X is a linearly independent set of vectors.
If a1, a2 , a3 , a4 be any scalars i. e., elements of the field R , then
a1 α1 + a2 α2 + a3 α3 + a4 α4 = zero vector
⇒ a1 (1, 0 , 0 , 0 ) + a2 (0 , 1, 0 , 0 ) + a3 (0 , 0 , 1, 0 ) + a4 (0 , 0 , 0 , 1) = (0 , 0 , 0 , 0 )
⇒ (a1, a2 , a3 , a4 ) = (0 , 0 , 0 , 0 )
⇒ a1 = 0 , a2 = 0 , a3 = 0 , a4 = 0 .
Therefore the given set X of four vectors is linearly independent.
Now we shall show that X generates R4 i. e., each vector of R4 can be expressed as a
linear combination of the vectors of X .
Let (a, b, c , d ) be any vector in R4 . We can write
(a, b, c , d ) = a (1, 0 , 0 , 0 ) + b (0 , 1, 0 , 0 ) + c (0 , 0 , 1, 0 ) + d (0 , 0 , 0, 1)
= aα1 + bα2 + cα3 + dα4 .
Thus (a, b, c , d ) has been expressed as a linear combination of the vectors of X and so
X generates R4 .
Since X is a linearly independent subset of R4 and it also generates R4 , therefore it is a
basis of R4 .
V-59

Example 10: Show that the set S = {1, x, x2 , … , x n} of n + 1polynomials in x is a basis of the
vector space Pn (R), of all polynomials in x (of degree at most n) over the field of real numbers.
Solution: Pn (R) is the vector space of all polynomials in x (of degree at most n) over
the field R of real numbers.
S = {1, x, x2 , … , x n} is a subset of Pn consisting of n + 1polynomials. To prove that S is a
basis of the vector space Pn (R).
First we show that the vectors in the set S are linearly independent over the field R.
The zero vector of the vector space Pn (R) is the zero polynomial. Let
a0 , a1, a2 , … , an ∈R be such that
a0 (1) + a1 x + a2 x2 + … + an x n
= 0 i. e., zero polynomial.
Now by the definition of the equality of two polynomials, we have
a0 + a1 x + a2 x2 + … + an x n
=0
⇒ a0 = 0 , a1 = 0 , a2 = 0 , … , an = 0 .
∴ the vectors 1, x, x2 , … , x n of the vector space Pn (R) are linearly independent.
Now we shall show that the set S generates the whole vector space Pn (R).
Let α = a0 + a1 x + a2 x2 + … + an x n be any arbitrary member of Pn , where
a0 , a1, … , an ∈R. Then α is a linear combination of the polynomials 1, x, x2 , … , x n over
the field R. Therefore S generates Pn (R).
Since S is a linearly independent subset of Pn (R) and it also generates Pn (R), therefore
it is a basis of Pn (R).

Example 11: Show that a finite subset W of a vector space V ( F ) is linearly dependent if and
only if some element of W can be expressed as a linear combination of the others.
Solution: Let W = {α1, α2 , … , α n} be a finite subset of a vector space V ( F ).
First suppose that W is linearly dependent. Then to show that some vector of W can
be expressed as a linear combination of the others.
Since the vectors α1, α2 , … , α n are linearly dependent, therefore there exist scalars
a1, a2 , … , an not all zero such that
a1α1 + a2 α2 + … + an α n = 0. …(1)
Suppose ar ≠ 0 , 1 ≤ r ≤ n.
Then from (1), we have
ar α r = − a1 α1 − a2 α2 − … − ar − 1 α r − 1 − ar + 1 α r + 1 − … − an α n
or ar −1 (ar α r ) = ar −1 (− a1α1 − a2 α2 − … − ar − 1 α r − 1 − ar + 1 α r + 1
− … − an α n)
−1 −1 −1
or α r = (− ar a1) α1 + (− ar a2 ) α2 + … + (− ar ar − 1) α r − 1

+ (− ar −1 ar + 1) α r + 1 + … + (− ar −1 an) α n .
Thus α r is a linear combination of the other vectors of the set W.
Conversely suppose that some vector of W is a linear combination of the others. Then
to prove that W is linearly dependent.
V-60

Without loss of generality suppose α1 is a linear combination of the vectors α2 , … , α n .


Let α1 = b2α2 + b3α3 + … + bn α n ,
where b2 , b3 , … , bn ∈ F .
Then α1 − b2α2 − b3α3 − … − bn α n = 0 . …(2)
Since in the linear relation (2) among the vectors α1, α2 , … , α n , the scalar coefficient
of α1 is 1 which is ≠ 0, therefore the vectors α1, α2 , … , α n are linearly dependent.

Example 12: If n vectors span a vector space V containing r linearly independent vectors, then
show that n ≥ r. (Kumaun 2013)
Solution: Suppose a subset S of V containing n vectors spans V. Then there exists a
linearly independent subset T of S which also spans V. This subset T of S will form a
basis of V. Suppose the number of vectors in T is m. Then dim V = m ≤ n.
Since dim V = m, therefore any subset of V containing more than m vectors will be
linearly dependent.
Hence if S1 is a linearly independent subset of V and S1 contains r vectors, we must
have
r≤m ⇒ r ≤ n. [ ∵ m ≤ n]
Example 13: Let W1 and W2 be distinct subspaces of V and that dim W1 = 4,
dim W2 = 4 and dim V = 6. Find the possible dimensions of W1 ∩ W2 .
Solution: Since W1 and W2 are distinct, so W1 + W2 properly contains W1 and W2 i. e.,
dim (W1 + W2 ) > 4. Also dim V = 6 ⇒ dim (W1 + W2 ) cannot be greater than 6. Thus
we have two possibilities i. e., dimension of (W1 + W2 ) must be either 5 or 6.
Using theorem dim (W1 + W2 ) = dim W1 + dim W2 − dim (W1 ∩ W2 ), we have
(i) 5 = 4 + 4 − dim (W1 ∩ W2 ) or dim W1 ∩ W2 = 3
(ii) 6 = 4 + 4 − dim (W1 ∩ W2 ) or dim W1 ∩ W2 = 2.
Hence the possible dimensions of W1 ∩ W2 are 2, 3.

Comprehensive Exercise 1

1. Give a basis for each of the following vector spaces over the indicated fields.
(i) R (√ 2) over R (ii) C over R (iii) Q (21 /4 ) over Q
where Q, R, C are fields of rational, real and complex numbers respectively.
2. Find the dimension of Q (√ 2, √ 3) over Q.
3. Find bases for the subspaces
W1 = {(a, b, 0 ): a, b, ∈ R }, W2 = {(0 , b, c ): b, c ∈ R } of R 3 .
Find the dimensions of W1, W2 and W1 ∩ W2 .
4. Show that the real field R is a vector space of infinite dimension over the rational
field Q.
5. Let V be the vector space of all 2 × 2 symmetric matrices over F. Show that
dimV = 3.
V-61

6. Prove that the space of all m × n matrices over the field F has dimension mn, by
exhibiting a basis for this space.
7. If {α1, α2 , α3 } is a basis of V3 (R), show that {α1 + α2 , α2 + α3 , α3 + α1} is also a
basis of V3 (R).
8. If W1 and W2 are finite-dimenisonal subspaces with the same dimension, and if
W1 ⊆ W2, then W1 = W2 .

9. In the vector space R 3 , let α = (1, 2, 1), β = (3, 1, 5), γ = (3, − 4, 7). Show that there
exists more than one basis for the subspace spanned by the set S = {α, β, γ }.
10. For the 3-dimensional space R 3 over the field of real numbers R, determine if the
set {(2, − 1, 0 ), (3, 5, 1), (1, 1, 2)} is a basis.

11. Show that the set {(1, i, 0 ), (2 i, 1, 1), (0 , 1 + i, 1 − i)} is a basis for V3 (C).
(Kumaun 2011)
3
12. Find three vectors in R which are linearly dependent, and are such that any two
of them are linearly independent.
13. Select a basis, if any, of R 3 (R) from the set {α1, α2 , α3 , α4 }, where
α1 = (1, − 3, 2), α2 = (2, 4, 1), α3 = (3, 1, 3), α4 = (1, 1, 1).

14. If W is a subspace of a finite dimensional vector space V, prove that any basis of
W can be extended to form a basis of V.
15. Prove that any finite set S of vectors, not all the zero vectors, contains a linearly
independent subset T which spans the same space as S.
16. Let V be a vector space. Let W be a subspace of V generated by the vectors
α1, …, α s . Prove that W is spanned by a linearly independent subset of α1, …, α s .
17. If a vector space V is spanned by a finite set of m vectors, then show that any
linearly independent set of vectors in V has at most m elements.
18. In a vector space V over the field F, let B = {α1, α2 , …, α n} span V. Prove that the
following two statements are equivalent :
(i) B is linearly independent
n
(ii) If α ∈V, then the expression α = ∑ ai α i with ai ∈ F is unique.
i =1

19. Determine a basis of the subspace spanned by the vectors


α1 = (1, 2, 3), α2 = (2, 1, − 1), α3 = (1, − 1, − 4), α4 = (4, 2, − 2).

20. Let W be the subspace of R 4 generated by the vectors (1, − 2, 5, − 3), (2, 3, 1, − 4)
and (3, 8, − 3, − 5). Find a basis and the dimension of W. Extend the basis of W to a
basis of the whole space R 4 .
21. Let E be a subfield of a field F and F a subfield of a field K i. e., E ⊂ F ⊂ K. Let K be
of dimension n over F and F is of dimension m over E. Show that K is of dimension
mn over E.
22. Let W1 and W2 be the subspaces of R 4 :
W1 = {(a, b, c , d) : b + c + d = 0 }, W2 = {(a, b, c , d): a + b = 0 , c = 2 d}.
Find the dimension of W1 ∩ W2 .
V-62

23. Let W1 and W2 be subspaces of V and that dim W1 = 4, dim W2 = 5 and dim
V = 7. Find the possible dimensions of W1 ∩ W2 .
24. Let W1 and W2 be the subspaces of R4 generated by the sets
B1 = {(1, 1, 0, − 1), (1, 2, 3, 0 ), (2, 3, 3, − 1)}
and B2 = {(1, 2, 2, − 2),(2, 3, 2, − 3),(1, 3, 4, − 3)} respectively.
Find (i) dim (W1 + W2 ) (ii) dim (W1 ∩ W2 ).

A nswers 1
1. (i) {1, √ 2} (ii) {1, i} (iii) {1, 21 /4 }
2. 4
3. Basis for W1 = {(1, 0 , 0 ), (0 , 1, 0 )}; dim W1 = 2
Basis for W2 = {(0 , 1, 0 ), (0 , 0 , 1)}; dim W2 = 2 ; dim (W1 ∩ W2 ) = 1
10. Yes 12. (1, 0 , 0 ), (0 , 1, 0 ), (1, 1, 0 )
13. {α1, α2 , α4 } 19. {α1, α2 }
20. A basis of W = {(1, − 2, 5, − 3), (0 , 7, − 9, 2)} ; dim W = 2
A basis of R 4 is {(1, − 2, 5, − 3), (0 , 7, − 9, 2), (0 , 0 , 1, 0 ), (0 , 0 , 0 , 1)}
22. 1 23. 2, 3 or 4
24. (i) 3 (ii) 1

O bjective T ype Q uestions

Multiple Choice Questions


Indicate the correct answer for each question by writing the corresponding letter from (a),
(b), (c) and (d).
1. The vector space V3 (R) is of dimension
(a) 1 (b) 3
(c) 2 (d) None of these.
2. Let V be the vector space of all 2 × 2 matrices over the field F. The dimension of V
is
(a) 2 (b) 3
(c) 4 (d) None of these. (Garhwal 2007)
3. The dimension of the vector space C 2 (R) is
(a) 2 (b) 4
(c) 1 (d) None of these.
4. The dimension of the vector space C (R) is
(a) 1 (b) 2
(c) 4 (d) None of these.
V-63

5. Let W1 and W2 be finite dimensional subspaces of a vector space V. If dim W1 = 2 ,


dim W2 = 2 , dim (W1 + W2 ) = 3 then dim (W1 ∩ W2 ) is
(a) 1 (b) 2
(c) 3 (d) 4.
6. The dimension of the vector space spanned by (1, − 2, 3, − 1) and (1, 1, − 2 , 3) is
(a) 1 (b) 2
(c) 4 (d) None of these.
7. The dimension of the vector space spanned by (3, − 6, 3, − 9) and (− 2 , 4, − 2 , 6) is
(a) 1 (b) 2
(c) 4 (d) None of these.
8. In the vector space R3 , the coordinate vector of (3, 5, − 2) relative to the ordered
basis {(1, 0 , 0 ), (0 , 1, 0 ), (0 , 0 , 1)} is
(a) (5, 3, − 2) (b) (3, 5, − 2)
(c) (− 2 , 5, 3) (d) (− 2 , 3, 5).

Fill in the Blank(s)


Fill in the blanks “……” so that the following statements are complete and correct.
1. Each set of (n + 1) or more vectors of a finite dimensional vector space V ( F ) of
dimension …… is linearly dependent.
2. If W1, W2 are two subspaces of a finite dimensional vector space V ( F ), then
dim (W1 + W2 ) + dim (W1 ∩ W2 ) = …… .

3. The set S = …… of n + 1polynomials in x is a basis of the vector space Pn (R), of all


polynomials in x (of degree at most n) over the field of real numbers.
4. Let V be the vector space of m × n matrices over a field F. Then dim V = …… .
5. If W is a proper subspace of a finite dimensional vector space V, then W is finite
dimensional and ...... . (Garhwal 2008)

6. The number of members in any basis of a vector space is called .... .


(Garhwal 2010, 12)

True or False
Write ‘T’ for true and ‘F’ for false statement.
1. If a subset of an n-dimensional vector space V consists of n non-zero vectors, then
it will be a basis of V.
2. If A and B are subspaces of a vector space, then dim A < dim B ⇒ A ⊂ B.
3. If A and B are subspaces of a vector space, then
A ≠ B ⇒ dim A ≠ dim B.
4. If M and N are finite dimensional subspaces with the same dimension, and if
M ⊂ N , then M = N .
5. In an n-dimensional space any subset consisting of n linearly independent vectors
will form a basis.
V-64

A nswers

Multiple Choice Questions


1. (b) 2. (c) 3. (b) 4. (b)
5. (a) 6. (b) 7. (a) 8. (b)

Fill in the Blank(s)


1. n 2. dim W1 + dim W2
3. {1, x, x2 , … , x n} 4. mn
5. dim W < dim V 6. dimension

True or False
1. F 2. F 3. F 4. T
5. T

¨
V-65

4
H omomorphism of
V ector S paces

1 Homomorphism of Vector Spaces or Linear


Transformation

efinition: Let U ( F ) and V ( F ) be two vector spaces. Then a mapping


D f :U → V
is called a homomorphism or a linear transformation of U into V, if
(i) f (α + β) = f (α) + f ( β), V α, β ∈ U
and (ii) f (aα) = af (α) V a ∈ F , V α ∈U.
The conditions (i) and (ii) can be combined into a single condition
f (aα + bβ) = af (α) + bf ( β) V a, b ∈ F and V α, β ∈ U .
If f is a homomorphism of U onto V, then V is called a homomorphic image of U.
Theorem: If f is a homomorphism of U ( F ) into V ( F ), then
(i) f (0) = 0 ′ where 0 and 0′ are the zero vectors of U and V respectively.
(ii) f (− α) = − f (α) V α ∈ U .
V-66

Proof: (i) Let α ∈U. Then f (α) ∈ V . Since 0′ is the zero vector of V, therefore
f (α) + 0 ′ = f (α) = f (α + 0) = f (α) + f (0).
Now V is an abelian group with respect to addition of vectors.
∴ f (α) + 0 ′ = f (α) + f (0)
⇒ 0 ′ = f (0) by left cancellation law.
(ii) If α ∈U, then − α ∈ U. Also we have
0 ′ = f (0) = f [α + (− α)] = f (α) + f (− α).
Now f (α) + f (− α) = 0 ′ ⇒ f (− α) = the additive inverse of f (α)
⇒ f (− α) = − f (α).

Example 1: Show that the mapping f : V3 ( F ) → V2 ( F ) defined by


f (a1, a2 , a3 ) = (a1, a2 )
is a homomorphism of V3 ( F ) onto V2 ( F ). (Kumaun 2011, 14)
Solution: Let α = (a1, a2 , a3 ) and β = (b1, b2 , b3 ) be any two elements of V3 ( F ). Also
let a, b be any two elements of F. We have
f (aα + bβ) = f [a (a1, a2 , a3 ) + b (b1, b2 , b3 )]
= f [(aa1 + bb1, aa2 + bb2 , aa3 + bb3 )]
= (aa1 + bb1, aa2 + bb2 )
= a (a1, a2 ) + b (b1, b2 )
= a f (a1, a2 , a3 ) + b f (b1, b2 , b3 )
= a f (α) + b f ( β).
∴ f is a linear transformation.
To show that f is onto V2 ( F ). Let (a1, a2 ) be any element of V2 ( F ). Then
(a1, a2 , 0 ) ∈ V3 ( F ) and we have f (a1, a2 , 0 ) = (a1, a2 ). Therefore f is onto V2 ( F ).
Therefore f is a homomorphism of V3 ( F ) onto V2 ( F ).

Comprehensive Exercise 1

1. Define linear transformation of a vector space V ( F ) into a vector space W ( F ).


Show that the mapping
T : (a, b) → (a + 2, b + 3)
of V2 (R ) into itself is not a linear transformation.
2. Let f be a linear transformation from a vector space U into a vector space V. If S is
a subspace of U, prove that f (S) will be a subspace of V.
3. If V is finite dimensional and f is a homomorphism of V into itself which is not
onto prove that there is some α ≠ 0 in V such that f (α) = 0.
V-67

2 Quotient Space
(Garhwal 2007, 10, 12; Gorakhpur 10)
Right and Left Cosets.
Let W be any subspace of a vector space V ( F ). Let α be any element of V. Then the set
W + α = {γ + α : γ ∈ W }
is called a right coset of W in V generated by α. Similarly the set
α + W = {α + γ : γ ∈ W }
is called a left coset of W in V generated by α.
Obviously W + α and α + W are both subsets of V. Since addition in V is
commutative, therefore we have W + α = α + W . Hence we shall call W + α as simply
a coset of W in V generated by α.
The following results about cosets are both to be remembered :
(i) We have 0 ∈V and W + 0 = W . Therefore W itself is a coset of W in V.
(ii) α ∈ W ⇒ W + α = W.
Proof: First we shall prove that W + α ⊆ W .
Let γ + α be any arbitrary element of W + α. Then γ ∈W.
Now W is a subspace of V. Therefore
γ ∈W, α ∈W
⇒ γ + α ∈W.
Thus every element of W + α is also an element of W. Hence
W + α ⊆ W.
Now we shall prove that W ⊆ W + α.
Let β ∈W. Since W is a subspace, therefore
α ∈ W ⇒ − α ∈ W.
Consequently β ∈ W , − α ∈ W ⇒ β − α ∈ W . Now we can write
β = ( β − α) + α ∈ W + α since β − α ∈W.
Thus β ∈ W ⇒ β ∈ W + α. Therefore W ⊆ W + α.
Hence W = W + α.
(iii) If W + α and W + β are two cosets of W in V, then
W + α = W + β ⇔ α − β ∈W.
Proof: Since 0 ∈W, therefore 0 + α ∈ W + α. Thus
α ∈ W + α.
Now W +α =W +β
⇒ α ∈W + β
⇒ α − β ∈ W + ( β − β)
⇒ α − β ∈W + 0
⇒ α − β ∈W.
Conversely,
α − β ∈W
⇒ W + (α − β) = W
V-68

⇒ W + [(α − β) + β] = W + β
⇒ W + α = W + β.
Let V / W denote the set of all cosets of W in V i. e., let
V / W = {W + α : α ∈ V }.
We have just seen that if α − β ∈W, then W + α = W + β. Thus a coset of W in V can
have more than one representation.
Now if V ( F ) is a vector space, then we shall give a vector space structure to the set
V / W over the same field F. For this we shall have to define addition in V / W i. e.,
addition of cosets of W in V and multiplication of a coset by an element of F i. e., scalar
multiplication.
Theorem: If W is any subspace of a vector space V ( F ), then the set V / W of all cosets
W + α where α is any arbitrary element of V, is a vector space over F for the addition and scalar
multiplication compositions defined as follows :
(W + α) + (W + β) = W + (α + β) V α, β ∈ V
and a (W + α) = W + aα ; a ∈ F , α ∈ V .
Proof: We have α, β ∈ V ⇒ α + β ∈ V .
Also a ∈ F , α ∈ V ⇒ aα ∈ V .
Therefore W + (α + β) ∈ V / W and also W + aα ∈ V / W . Thus V / W is closed with
respect to addition of cosets and scalar multiplication as defined above. Now first of
all we shall show that these two compositions are well defined i. e., are independent of
the particular representation chosen to denote a coset.
Let W + α = W + α ′ , α, α ′ ∈ V
and W + β = W + β ′ , β, β ′ ∈ V .
We have W + α = W + α′ ⇒ α − α′ ∈W
and W + β = W + β′ ⇒ β − β′ ∈W.
Now W is a subspace, therefore
α − α ′ ∈W, β − β′ ∈W
⇒ (α − α ′ ) + ( β − β ′ ) ∈W
⇒ (α + β) − (α ′ + β ′ ) ∈W
⇒ W + (α + β) = W + (α ′ + β ′ )
⇒ (W + α) + (W + β) = (W + α ′ ) + (W + β ′ ).
Therefore addition in V /W is well defined.
Again a ∈ F, α − α ′ ∈ W
⇒ a (α − α ′ ) ∈ W
⇒ aα − aα ′ ∈ W ⇒ W + aα = W + aα ′ .
∴ scalar multiplication in V /W is also well defined.
Commutativity of addition. Let W + α , W + β be any two elements of V / W .
Then
(W + α) + (W + β) = W + (α + β)
= W + ( β + α)
= (W + β) + (W + α).
V-69

Associativity of addition. Let W + α , W + β, W + γ be any three elements of


V / W . Then
(W + α) + [(W + β) + (W + γ )] = (W + α) + [W + ( β + γ )]
= W + [α + ( β + γ )]
= W + [(α + β) + γ ]
= [W + (α + β)] + (W + γ )
= [(W + α) + (W + β)] + (W + γ ).
Existence of additive identity. If 0 is the zero vector of V, then
W + 0 = W ∈V /W.
If W + α is any element of V / W , then
(W + 0) + (W + α) = W + (0 + α) = W + α .
∴ W + 0 = W is the additive identity.
Existence of additive inverse. If W + α is any element of V / W , then
W + (− α) = W − α ∈ V / W .
Also we have
(W + α) + (W − α) = W + (α − α)
=W + 0
= W.
∴ W − α is the additive inverse of W + α.
Thus V / W is an abelian group with respect to addition composition. Further we
observe that if
a, b ∈ F and W + α, W + β ∈ V / W , then
1. a [(W + α) + (W + β)] = a [W + (α + β)]
= W + a (α + β)
= W + (aα + aβ)
= (W + aα) + (W + aβ)
= a (W + α) + a (W + β).
2. (a + b) (W + α) = W + (a + b) α
= W + (aα + bα)
= (W + aα) + (W + bα)
= a (W + α) + b (W + α).
3. (ab) (W + α) = W + (ab) α
= W + a (bα)
= a (W + bα)
= a [b (W + α)].
4. 1 (W + α) = W + 1α
= W + α.
∴ V /W is a vector space over F for these two compositions. The vector space V / W is
called the Quotient Space of V relative to W. The coset W is the zero vector of this
vector space.
V-70

3 Dimension of a Quotient Space


Theorem: If W be a subspace of a finite dimensional vector space V ( F ), then
dim V / W = dim V − dim W .
(Garhwal 2006, 07, 08, 10, 14; Gorakhpur 10, 13)
Proof: Let m be the dimension of the subspace W of the vector space V ( F ).
Let S = {α1, α2 , … , α m}
be a basis of W. Since S is a linearly independent subset of V, therefore it can be
extended to form a basis of V. Let
S ′ = {α1, α2 , … , α m , β1, β2 , … , β l }
be a basis of V. Then dim V = m + l.
∴ dim V − dim W = (m + l ) − m = l.
So we should prove that dim V / W = l.
We claim that the set of l cosets
S1 = {W + β1, W + β2 , … , W + β l}
is a basis of V / W .
First we show that S1 is linearly independent. The zero vector of V / W is W.
Let a1 (W + β1) + a2 (W + β2 ) + … + al (W + β l) = W
⇒ (W + a1 β1) + (W + a2 β2 ) + … + (W + al β l) = W
⇒ W + (a1 β1 + a2 β2 + … + al β l) = W + 0
⇒ a1 β1 + a2 β2 + … + al β l ∈ W
⇒ a1 β1 + a2 β2 + … + al β l = b1 α1 + b2 α2 + … + bm α m
[∵ any vector in W can be expressed as a linear
combination of its basis vectors]
⇒ a1 β1 + a2 β2 + … + al β l − b1 α1 − b2 α2 − … − bm α m = 0
⇒ a1 = 0 , a2 = 0 , …, al = 0 since the vec tors
β1, β2 , … β l , α1, α2 , … , α m are linearly independent.
∴ The set S1 is linearly independent.
Now to show that L (S1) = V / W . Let W + α be any element of V / W . The vector α ∈V
can be expressed as
α = c1 α1 + c2 α2 + … + c m α m + d1 β1 + d2 β2 +… + dl β l
= γ + d1 β1 + d2 β2 + … + dl β l
where γ = c1 α1 + c2 α2 + … + c m α m ∈ W .
So W + α = W + (γ + d1 β1 + d2 β2 + … + dl β l)
= (W + γ ) + d1 β1 + d2 β2 + … + dl β l
= W + (d1 β1 + d2 β2 + … + dl β l) [∵ γ ∈W ⇒ W + γ = W ]
= (W + d1 β1) + (W + d2 β2 ) + … + (W + dl β l)
= d1 (W + β1) + d2 (W + β2 ) + … + dl (W + β l).
Thus any element W + α of V / W can be expressed as a linear combination of elements
of S1.
∴ V / W = L (S1).
∴ S1 is a basis of V / W .
∴ dim V / W = l. Hence the theorem.
V-71

4 Direct Sum of Spaces


(Garhwal 2006, 10)
Vector space as a direct sum of subspaces.
Definition: Let V ( F ) be a vector space and let W1, W2 , … , Wm be subspaces of V. Then V
is said to be the direct sum of W1, W2 , … , Wm if every element α ∈V can be written in one and
only one way as α = α1 + α2 + … + α m where
α1 ∈ W1, α2 ∈ W2 , … , α m ∈ Wm .
If a vector space V ( F ) is a direct sum of its two subspaces W1 and W2 then we should
have not only V = W1 + W2 but also that each vector of V can be uniquely expressed as
sum of an element of W1 and an element of W2 . Symbolically the direct sum is
represented by the notation V = W1 ⊕ W2 .
Illustration 1: Let V2 ( F ) be the vector space of all ordered pairs of F. Then
W1 = {(a, 0 ) : a ∈ F } and W2 = {(0 , b) : b ∈ F } are two subspaces of V2 ( F ). Obviously any
element ( x, y) ∈ V2 ( F ) can be uniquely expressed as the sum of an element of W1 and an element
of W2 . The unique expression is ( x, y) = ( x, 0 ) + (0 , y). Thus V2 ( F ) is the direct sum of W1 and
W2 .Also we observe that the only element common to both W1 and W2 is the zero vector (0 , 0 ).
Illustration 2: Let W1 = {(a, b, 0 ) : a, b ∈ R } and W2 = {(0 , 0 , c ) : c ∈ R} be two subspaces of
R3 .
Now any vector (a, b, c ) ∈R3 can be uniquely expressed as the sum of a vector in W1
and a vector in W2 : (a, b, c ) = (a, b, 0 ) + (0 , 0 , c ).
Thus R3 is the direct sum of W1 and W2 .
Also we observe that the only element common to both W1 and W2 is the zero vector
(0 , 0 , 0 ).
Illustration 3: Let W1 = {(a, b, 0 ) : a, b ∈ R } and W2 = {(0 , b, c ) : b, c ∈ R } be two subspaces of
R3 . Then R3 = W1 + W2 since every vector in R3 is the sum of a vector in W1 and a vector in W2 .
However, R3 is not the direct sum of W1 and W2 since such sums are not unique.
For example, (4, 6, 8) = (4, 5, 0 ) + (0 , 1, 8).
Also (4, 6, 8) = (4, − 1, 0 ) + (0 , 7, 8).
Here we observe that W1 ∩ W2 ≠ { 0 }.

5 Disjoint Subspaces
Definition: Two subspaces W1 and W2 of the vector space V ( F ) are said to be disjoint if
their intersection is the zero subspace i. e., if W1 ∩ W2 = { 0 }.
Theorem: The necessary and sufficient conditions for a vector space V ( F ) to be a direct sum
of its two subspaces W1 and W2 are that
(i) V = W1 + W2
(ii) W1 ∩ W2 = { 0 } i. e., W1 and W2 are disjoint. (Garhwal 2010, 14; Gorakhpur 11)
V-72

Proof: The conditions are necessary.


Let V be direct sum of its two subspaces W1 and W2 . Then each element of V is
expressible uniquely as sum of an element of W1 and an element of W2 . Therefore we
have
V = W1 + W2 .
Let, if possible 0 ≠ α ∈ W1 ∩ W2 . Then α ∈ W1, α ∈ W2 . Also α ∈V and we can write
α = 0 + α where 0 ∈ W1, α ∈ W2
and α = α + 0 where α ∈ W1, 0 ∈ W2 .
Thus α ∈V can be expressed in at least two different ways as sum of an element of W1
and an element of W2 . This contradicts the fact that V is direct sum of W1 and W2 .
Hence 0 is the only vector common to both W1 and W2 i. e., W1 ∩ W2 = { 0 }. Thus the
conditions are necessary.
The conditions are sufficient.
Let V = W1 + W2 and W1 ∩ W2 = { 0 }. Then to show that V is direct sum of W1 and W2 .
V = W1 + W2 ⇒ that each element of V can be expressed as sum of an element of W1
and an element of W2 . Now to show that this expression is unique.
Let, if possible,
α = α1 + α 2 , α ∈ V , α1 ∈ W1, α 2 ∈ W2 ,
and α = β1 + β2 , β1 ∈ W1, β2 ∈ W2 .
Then to show that α1 = β1 and α2 = β2 .
We have α1 + α 2 = β1 + β2
⇒ α1 − β1 = β2 − α 2 .
Since W1 is a subspace, therefore
α1 ∈ W1, β1 ∈ W1 ⇒ α1 − β1 ∈ W1.
Similarly, β2 − α2 ∈W2 .
∴ α1 − β1 = β2 − α2 ∈ W1 ∩ W2 .
But 0 is the only vector which belongs to W1 ∩ W2 . Therefore
α1 − β1 = 0 ⇒ α1 = β1.
Also β 2 − α2 = 0 ⇒ α2 = β 2 .
Thus each vector α ∈V is uniquely expressible as sum of an element of W1 and an
element of W2 . Hence V = W1 ⊕ W2 .

6 Dimension of a Direct Sum


Theorem 1: If a finite dimensional vector space V ( F ) is a direct sum of two subspaces W1 and
W2 , then
dim V = dim W1 + dim W2 . (Gorakhpur 2012, 14)
Proof: Let dim W1 = m and dim W2 = l. Also let the sets of vectors
S1 = {α1, α2 , … , α m} and S2 = { β1, β2 , … , β l}
be the bases of W1 and W2 respectively.
We have dim W1 + dim W2 = m + l.
V-73

In order to prove that dim V = dim W1 + dim W2 , we should therefore prove that dim
V = m + l. We claim that the set
S = S1 ∪ S2 = {α1, α2 , … , α m , β1, β2 , … , β l }
is a basis of V.
First we show that the set S is linearly independent. Let
a1α1 + a2 α2 + … + am α m + b1 β1 + b2 β2 + … + bl β l = 0.
⇒ a1α1 + a2 α2 + … + am α m = − (b1 β1 + b2 β2 + … + bl β l).
Now a1α1 + a2 α2 + … + am α m ∈ W1
and − (b1 β1 + b2 β2 + … + bl β l) ∈ W2 .
Therefore a1α1 + a2 α2 + … + am α m ∈ W1 ∩ W2
and − (b1 β1 + b2 β2 + … + bl β l) ∈ W1 ∩ W2 .
But V is the direct sum of W1 and W2 . Therefore 0 is the only vector belonging to
W1 ∩ W2 . Then we have
a1α1 + a2 α2 + … + am α m = 0, b1 β1 + b2 β2 + … + bl β l = 0.
Since both the sets {α1, α2 , … , α m} and { β1, β2 , … , β l} are linearly independent,
therefore we have
a1 = 0 , a2 = 0 , … , am = 0 , b1 = 0 , b2 = 0 , … , bl = 0 .
Therefore S is linearly independent.
Now we shall show that L (S ) = V . Let α be any element of V. Then
α = an element of W1 + an element of W2 .
= a linear combination of S1 + a linear combination of S2
= a linear combination of elements of S.
∴ L (S ) = V
∴ S is a basis of V. Therefore dim V = m + l.
Hence the theorem.
A sort of converse of this theorem is true. It has been proved in the following theorem.
Theorem 2: Let V be a finite dimensional vector space and let W1, W2 be subspaces of V such
that V = W1 + W2 and dim V = dim W1 + dim W2 .
Then V = W1 ⊕ W2 .
Proof: Let dim W1 = l and dim W2 = m. Then
dim V = l + m.
Let S1 = {α1, α2 , … , α l }
be a basis of W1 and
S2 = { β1, β2 , … , β m}
be a basis of W2 . We shall show that S1 ∪ S2 is a basis of V.
Let α ∈V. Since V = W1 + W2 , therefore we can write
α = γ + δ where γ ∈ W1, δ ∈ W2 .
V-74

Now γ ∈W1 can be expressed as a linear combination of the elements of S1 and δ ∈W2
can be expressed as a linear combination of the elements of S2 . Therefore α ∈V can be
expressed as a linear combination of the elements of S1 ∪ S2 . Therefore
V = L (S1 ∪ S2 ). Since dim V = l + m and L (S1 ∪ S2 ) = V , therefore the number of
distinct elements in S1 ∪ S2 cannot be less than l + m. Thus S1 ∪ S2 has l + m distinct
elements and therefore S1 ∪ S2 is a basis of V. Therefore the set
{α1, α2 , … , α l , β1, β2 , … , β m}
is linearly independent.
Now we shall show that
W1 ∩ W2 = { 0 }.
Let α ∈ W1 ∩ W2 .
Then α ∈ W1, α ∈ W2 .
Therefore α = a1α1 + a2 α2 + … + al α l
and α = b1 β1 + b2 β2 + … + bm β m
for some a’s and b’s ∈ F.
∴ a1α1 + a2 α2 + … + al α l = b1 β1 + b2 β2 + … + bm β m
⇒ a1α1 + a2 α2 + … + al α l − b1 β1 − b2 β2 − … − bm β m = 0
⇒ a1 = 0 , a2 = 0 , … , al = 0 , b1 = 0 , b2 = 0 , … , bm = 0
⇒ α = 0.
∴ W1 ∩ W2 = { 0 }.

7 Complementary Subspaces
Definition: Let V ( F ) be a vector space and W1, W2 be two subspaces of V. Then the
subspace W2 is called the complement of W1 in V if V is the direct sum of W1 and W2 .

8 Existence of Complementary Subspaces


Theorem: Corresponding to each subspace W1 of a finite dimensional vector space V ( F ), there
exists a subspace W2 such that V is the direct sum of W1 and W2 .
Proof: Let dim W1 = m. Let the set
S1 = {α1, α2 , … , α m}
be a basis of W1.
Since S1 is a linearly independent subset of V, therefore S1 can be extended to form a
basis of V. Let the set
S = {α1, α2 , … , α m , β1, β2 , … , β l }
be a basis of V.
Let W2 be the subspace of V generated by the set
S2 = { β1, β2 , … , β l }.
We shall prove that V is the direct sum of W1 and W2 . For this we shall prove that
V = W1 + W2 and W1 ∩ W2 = { 0 }.
V-75

Let α be any element of V. Then we can express


α = a linear combination of elements of S
= a linear combination of S1 + a linear combination of S2
= an element of W1 + an element of W2 .
∴ V = W1 + W2 .
Again let β ∈ W1 ∩ W2 . Then β can be expressed as a linear combination of S1 and also
as a linear combination of S2 . So we have
β = a1α1 + a2 α2 + … + am α m = b1 β1 + b2 β2 + … + bl β l .
⇒ a1α1 + a2 α2 + … + am α m − b1 β1 − b2 β2 − … − bl β l = 0
⇒ a1 = 0 , a2 = 0 , … , am = 0 , b1 = 0 , b2 = 0 , … , bl = 0
since α1, α2 , … , α m , β1, β2 , … , β l are linearly independent.
∴ β = 0 (zero vector).
Thus W1 ∩ W2 = {0}.
Hence V is the direct sum of W1 and W2 .

9 Direct Sum of Several Subspaces


Now we shall have some discussion on the direct sum of several subspaces. For this
purpose we shall first define the concept of independence of subspaces analogous to
the disjointness condition on two subspaces.
Definition: Suppose W1, W2 , … , Wk are subspaces of the vector space V. We shall say that
W1, W2 , … , Wk are independent, if
α1 + α2 + … + α k = 0, α i ∈ Wi implies that each α i = 0.
Theorem 1: Suppose V ( F ) is a vector space. Let W1, W2 , … , Wk be subspaces of V and let
W = W1 + W2 + … + Wk . The following are equivalent :
(i) W1, W2 , … , Wk are independent.
(ii) Each vector α in W can be uniquely expressed in the form
α = α1 + α2 + … + α k with α i in Wi , i = 1, … , k .
(iii) For each j, 2 ≤ j ≤ k , the subspace Wj is disjoint from the sum
(W1 + … + Wj − 1).
Proof: In order to prove the equivalence of the three statements we shall prove that
(i) ⇒ (ii), (ii) ⇒ (iii) and (iii) ⇒ (i).
(i) ⇒ (ii). Suppose W1, … , Wk are independent. Let α ∈W.
Since W = W1 + … + Wk , therefore we can write
α = α1 + … + α k with α i in Wi .
Suppose that also α = β1 + … + β k with β i in Wi .
Then α1 + … + α k = β1 + … + β k
⇒ (α1 − β1) + … + (α k − β k ) = 0 with α i − β i in Wi as Wi is a subspace
⇒ αi − βi = 0 [ ∵ W1, … , Wk are independent]
⇒ α i = β i , i = 1, … , k .
Therefore the α i’s are uniquely determined by α.
V-76

(ii) ⇒ (iii). Let α ∈ Wj ∩ (W1 + … + Wj−1).


Then α ∈Wj and α ∈ W1 + … + Wj − 1 .
Now α ∈ W1 + … + Wj − 1 implies that there exist vectors α1, … , α j − 1 with α i in Wi such
that
α = α1 + … + α j − 1 .
Also α ∈Wj .
Therefore we get two expressions for α as a sum of vectors, one in each Wi . These are
α = α1 + … + α j − 1 + 0 + … + 0
in which the vector belonging to Wj is 0
and α = 0 +…+ 0 + α +…+ 0
in which the vector belonging to Wj is α.
Since the expression for α is given to be unique, therefore we must have
α1 = … = α j − 1 = 0 = α.
Thus Wj ∩ (W1 + … + Wj − 1) = {0}.
(iii) ⇒ (i).
Let α1 + … + α k = 0 where α i ∈ Wi , i = 1, … , k . …(1)
Then we are to prove that each α i = 0.
Suppose that for some i we have α i ≠ 0.
Let j be the largest integer i between 1 and k such that α i ≠ 0. Obviously j must be ≥ 2
and at the most j can be equal to k. Then (1) reduces to
α1 + … + α j = 0, α j ≠ 0
⇒ α j = − α1 − … − α j − 1
⇒ α j ∈ W1 + … + Wj − 1 [ ∵ − α1 − … − α j − 1 ∈ W1 + … + Wj − 1]
⇒ α j ∈ Wj ∩ (W1 + … + Wj − 1) [ ⇒ α j ∈ Wj ]
⇒ α j = 0.
Thus we get a contradiction. Hence each α i = 0.
Note: If any (and hence all) of the three conditions of theorem 1 hold for W1, … , Wk ,
then we shall say that W is the direct sum of W1, … , Wk and we write
W = W1 ⊕ … ⊕ Wk .
Theorem 2: Let V ( F ) be a vector space. Let W1, … , Wn be subspaces of V. Suppose that
V = W1 + … + Wn
and that Wi ∩ (W1 + … + Wi − 1 + Wi + 1 + … + Wn) = {0}
for every i = 1, 2, … , n. Prove that V is the direct sum of W1, … , Wn .
Proof: In order to prove that V is the direct sum of W1, … , Wn , we should prove that
each vector α ∈V can be uniquely expressed as
α = α1 + … + α n where α i ∈ Wi , i = 1, … , n.
Since V = W1 + … + Wn , therefore any vector α in V can be written as
α = α1 + … + α n where α i ∈ Wi . …(1)
To show that α1 , … , α n are unique.
Let α = β1 + … + β n where β i ∈ Wi . …(2)
V-77

From (1) and (2), we get


α1 + … + α n = β1 + … + β n
(α1 − β1) + … + (α i − 1 − β i − 1) + (α i − β i)
+ (α i + 1 − β i + 1) + … + (α n − β n) = 0. …(3)
Now each Wi is a subspace of V. Therefore α i − β i and also its additive inverse
β i − α i ∈ Wi , i = 1, … , n.
From (3), we get
(α i − β i) = ( β1 − α1) + … + ( β i − 1 − α i − 1)
+ ( β i + 1 − α i + 1) + … + ( β n − α n). …(4)
Now the vector on the right hand side of (4) and consequently the vector α i − β i is in
W1 + … + Wi − 1 + Wi + 1 + … + Wn .
Also α i − β i ∈ Wi .
∴ α i − β i ∈ Wi ∩ (W1 + … + Wi − 1 + Wi + 1 + … + Wn).
But for every i = 1, … , n, it is given that
Wi ∩ (W1 + … + Wi − 1 + Wi + 1 + … + Wn) = {0}.
Therefore α i − β i = 0, i = 1, … , n
⇒ α i = β i , i = 1, … , n
⇒ the expression (1) for α is unique.
Hence V is the direct sum of W1, … , Wn .
Theorem 3: Let V ( F ) be a finite dimensional vector space and let W1, … , Wk be subspaces
of V. Then the following two statements are equivalent.
(i) V is the direct sum of W1, … , Wk .
(ii) If Bi is a basis of Wi, i = 1, … , k , then the union
k
B = ∪ Bi is also a basis for V.
i =1

Proof: Let Bi = {α1i, α2 i, … , α ni i } be a basis for Wi .


Here ni = dim Wi = number of vectors in Bi . Also let B be the union of the bases Bi .
(i) ⇒ (ii). It is given that V is the direct sum of W1, … , Wk , therefore for any α ∈V, we
can write
α = α1 + … + α k for α i ∈ Wi , i = 1, … , k .
Now α i can be expressed as a linear combination of the vectors in Bi which is a basis of
k
Wi . Therefore α can be expressed as a linear combination of the elements of B = ∪ Bi.
i =1

Therefore L ( B) = V i. e., B spans V.


Now to show that B is linearly independent. Let
k
Σ (a1i α1i + a2 i α2 i + … + ani i α ni i) = 0. …(1)
i =1

Since V is the direct sum of W1, … , Wk , therefore 0 ∈V can be uniquely expressed as a


sum of vectors one in each Wi . This unique expression is
0 = 0 + … + 0 where 0 ∈ Wi , i = 1, … , k .
V-78

Now a1i α1i + … + ani i α ni i ∈ Wi . Therefore from (1) which is an expression for 0 ∈V as
a sum of vectors one in each Wi , we get
a1i α1i + … + ani i α ni i = 0, i = 1, … , k
a1i = … = ani i = 0
since {α1i, … , α ni i} is linearly independent being a basis for Wi .
k
Therefore B = ∪ Bi is linearly independent. Therefore B is a basis of V.
i =1
k
(ii) ⇒ (i). It is given that B = ∪ Bi is a basis of V.
i =1

Therefore for any α ∈V, we can write


k
α= Σ (a1i α1i + a2 i α2 i + ...+ ani i α ni i)
i =1

= α1 + α2 + … + α k …(2)
where α i = a1i α1i + … + ani i α ni i ∈ Wi .
Thus each vector in V can be expressed as a sum of vectors one in each Wi .
Now V will be the direct sum of W1, … , Wk if the expression (2) for α is unique. Let
α = β1 + β2 + … + β k …(3)
where β i = b1i α1i + … + bni i α ni i ∈ Wi .
From (2) and (3), we get
α1 + … + α k = β1 + …+ β k
⇒ (α1 − β1) + … + (α i − β i) + … + (α k − β k ) = 0
k
⇒ Σ (α i − β i) = 0
i =1

k
⇒ Σ [(a1i − b1i) α1i + … + (ani i − bni i) α ni i ] = 0
i =1

⇒ a1i − b1i = … = ani i − bni i = 0 , i = 1, … , k


k
[∵ ∪ Bi is linearly independent being a basis of V ]
i =1
i i i i
⇒ a1 = b1 , … , ani = bni , i = 1, … , k
⇒ α i = β i , i = 1, … , k
⇒ the expression (2) for α is unique.
Hence V is the direct sum of W1, … , Wk .

Note: While proving this theorem we have proved that if a finite dimensional vector
space V ( F ) is the direct sum of its subspaces W1, … , Wk , then
dim V = dim W1 + … + dim Wk .
V-79

Example 2: Find dimension of V / W where V = R3 , W = {(a, 0 , 0 ) : a ∈ R}.


Solution: Here V = R3 . dim V = 3.

W = {(a, 0 , 0 ) : a ∈ R}. {(1, 0 , 0 )} is a basis of W so dim W = 1.


Now dim V / W = dim V − dim W = 3 − 1 = 2.
Example 3: If W1 = {( x, 0 , z ) : x, z ∈ R}, W2 = {(0 , y, z ) : y, z ∈ R} be two subspaces of R3
then show that R3 = W1 + W2 but R3 ≠ W1 ⊕ W2 .

Solution: Any ( x, y, z ) ∈R3 can be written as a sum of a vector of W1 and a vector of


W2 as shown below.
 1   1   1 
We have ( x, y, z ) =  x, 0 , z  + 0 , y, z  , where  x, 0 , z  ∈ W1
 2   2   2 
 1 
and 0 , y, z  ∈ W2 .
 2 
Thus R3 = W1 + W2 .

 1   3   1 
Also we have ( x, y, z ) =  x, 0 , z  + 0 , y, z  where  x, 0 , z  ∈ W1 and
 4   4   4 
 3 
0 , y, z  ∈ W2 .
 4 
It follows that representation of ( x, y, z ) as a sum of a vector of W1 and a vector of W2 is
not unique.
Hence R3 ≠ W1 ⊕ W2 .

Example 4: If W1 = {(a, b, c ) : a = b = c }, W2 = {(a, b, c ) : a = 0 } are two subspaces of R3 ,


show that R3 = W1 ⊕ W2 .

Solution: We claim that R3 = W1 + W2 .


For if α = (a, b, c ) ∈ R3 , then α = (a, a, a) + (0 , b − a, c − a) where (a, a, a) ∈ W1 and
(0 , b − a, c − a) ∈ W2 .
Also we have W1 ∩ W2 = {0}.
For α = (a, b, c ) ∈ W1 ∩ W2 ⇒ a = b = c and a = 0
⇒ a = 0 , b = 0 , c = 0 i. e., α = 0.
Thus R3 = W1 + W2 and W1 ∩ W2 = {0}.

Hence R3 = W1 ⊕ W2 .
Example 5: Construct three subspaces W1, W2 , W3 of a vector space V so that
V = W1 ⊕ W2 = W1 ⊕ W3 but W2 ≠ W3 .
V-80

Solution: Take the vector space V = R2 .


Obviously W1 = {(a, 0 ) : a ∈ R},
W2 = {(0 , a) : a ∈ R}, and W3 = {(a, a) : a ∈ R},
2
are three subspaces of R .
We have V = W1 + W2 and W1 ∩ W2 = {(0 , 0 )}. ∴ V = W1 ⊕ W2 .
Also it can be easily shown that
V = W1 + W3 and W1 ∩ W3 = {(0 , 0 )}. ∴ V = W1 ⊕ W3 .
Thus V = W1 ⊕ W2 = W1 ⊕ W3 but W2 ≠ W3 .

Comprehensive Exercise 2

1. Let V be the vector space of square matrices of order n over the field R. Let W1 and
W2 be the subspaces of symmetric and antisymmetric matrices respectively.
Show that V = W1 ⊕ W2 .
2. Let V be the vector space of all functions from the real field R into R. Let U be the
subspace of even functions and W the subspace of odd functions. Show that
V = U ⊕ W.
3. Let W1, W2 and W3 be the following subspaces of R 3 :
W1 = {(a, b, c ) : a + b + c = 0 }, W2 = {(a, b, c ) : a = c },
W3 = {(0 , 0 , c ) : c ∈ R }.
Show that (i) R 3 = W1 + W2 ; (ii) R 3 = W1 + W3 ; (iii) R 3 = W2 + W3 .
When is the sum direct ?
4. Let W be a subspace of a vector space V over a field F. Show that α ∈ β + W iff
α − β ∈W.
5. Let W1 and W2 be two subspaces of a finite dimensional vector space V. If dim
V = dim W1 + dim W2 and W1 ∩ W2 = {0}, prove that V = W1 ⊕ W2 .

A nswers 2
3. The sum is direct in (ii) and (iii)

10 Co-ordinates
Let V ( F ) be a finite dimensional vector space. Let
B = {α1, α2 , … , α n}
be an ordered basis for V. By an ordered basis we mean that the vectors of B have been
enumerated in some well-defined way i. e., the vectors occupying the first, second,
… , nth places in the set B are fixed.
V-81

Let α ∈V. Then there exists a unique n-tuple ( x1, x2 , … , xn) of scalars such that
n
α = x1α1 + x2 α2 + … + xn α n = Σ xi α i .
i =1

The n-tuple ( x1, x2 , … , xn) is called the n-tuple of co-ordinates of α relative to the
ordered basis B. The scalar xi is called the i th coordinate of α relative to the
ordered basis B. The n × 1 matrix
 x1 
x 
X= 
2
 ...
x 
 n
is called the coordinate matrix of α relative to the ordered basis B. We shall use the
symbol [α]B for the coordinate matrix of the vector α relative to the ordered basis B.
(Garhwal 2008)
It should be noted that for the same basis set B, the coordinates of the vector α are
unique only with respect to a particular ordering of B. The basis set B can be ordered in
several ways. The coordinates of α may change with a change in the ordering of B.
Note: If V ( F ) is a vector space of dimension n, then coordinate vector of 0 ∈V
relative to any basis of V is always the zero n-tuple (0 , 0 , … , 0 ).

11 Change of Basis
Suppose V is an n-dimensional vector space over the field F. Let B and B′ be two
ordered bases for V. If α is any vector in V, then we are now interested to know what is
the relation between its coordinates with respect to B and its coordinates with respect
to B′ . (Garhwal 2010)
Theorem 1: Let V ( F ) be an n-dimensional vector space and let B and B ′ be two ordered
bases for V. Then there is a unique necessarily invertible, n × n matrix A with entries in F such
that
(1) [α]B = A [α]B ′ (Garhwal 2008)
(2) [α]B ′ = A−1 [α]B for every vector α in V.
Proof: Let B = {α1, α2 , … , α n} and B′ = { β1, β2 , … , β n}.
Then there exists a unique linear transformation T from V into V such that
T (α j) = β j , j = 1, 2, … , n. …(1)
Since T maps a basis B onto a basis B ′ , therefore T is necessarily invertible. The
matrix of T relative to B i. e., [T ]B will be a unique n × n matrix with elements in F.
Also this matrix will be invertible because T is invertible.
Let [T ]B = A = [aij]n × n . Then
n
T (α j) = Σ aij α i, j = 1, 2, … , n. …(2)
i =1
Let x1, x2 , … , xn be the coordinates of α with respect to B and y1, y2 , … , yn be the
coordinates of α with respect to B′. Then
n
α = y1 β1 + y2 β2 + … + yn β n = Σ y j β j
j =1
V-82
n
= Σ y j T (α j) [From (1)]
j =1
n n
= Σ y j Σ aij α i [From (2)]
j =1 i =1
n  n 
= Σ  Σ aij y j α i .
i =1  j =1 
n
Also α = Σ xi α i .
i =1
n
∴ xi = Σ aij y j
j =1

because the expression for α as a linear combination of elements of B is unique.


Now [α]B is a column matrix of the type n × 1. Also [α]B ′ is a column matrix of the type
n × 1. The product matrix A [α]B ′ will also be of the type n × 1.
n
The ith entry of [α]B = xi = Σ aij y j = ith entry of A [α]B ′ .
j =1

∴ [α]B = A [α]B ′ ⇒ A−1 [α]B = A−1 A [α]B ′


⇒ A−1 [α]B = I [α]B ⇒ A−1 [α]B = [α]B ′ .
Note: The matrix A = [T ]B is called the transition matrix from B to B ′. It expresses
the coordinates of each vector in V relative to B in terms of its coordinates relative to
B′ .
How to write the transition matrix from one basis to another ?
Let B = {α1, α2 , … , α n} and B′ = { β1, β2 , … , β n} be two ordered bases for the
n-dimensional vector space V ( F ). Let A be the transition matrix from the basis B to
the basis B′ . Let T be the linear transformation from V into V which maps the basis B
onto the basis B ′ . Then A is the matrix of T relative to B i. e., A = [T ]B . So in order
to find the matrix A, we should first express each vector in the basis B ′ as a linear
combination over F of the vectors in B. Thus we write the relations
β1 = a11 α1 + a21 α 2 + … + an1 α n
β2 = a12 α1 + a22 α 2 + … + an2 α n
… … … … …
… … … … …
β n = a1n α1 + a2 n α 2 + … + ann α n .
Then the matrix A = [aij]n × n i. e., A is the transpose of the matrix of coefficients in the
above relations. Thus
 a11 a12 ... a1n
a a ... a2 n
 21 22 
A =  ... ... ... ... ⋅
 ... ... ... ...
 
 an1 an2 ... ann
V-83

Now suppose α is any vector in V. If [α]B is the coordinate matrix of α relative to the
basis B and [α]B ′ its coordinate matrix relative to the basis B ′ then
[α]B = A [α]B ′
and [α]B ′ = A−1 [α]B .
Theorem 2: Let B = {α1, … , α n} and B′ = { β1, … , β n} be two ordered bases for an
n
n-dimensional vector space V ( F ). If ( x1, … , xn) is an ordered set of n scalars, let α = Σ xi α i
i =1
n
and β = Σ xi β i .
i =1

Then show that T (α) = β, where T is the linear operator on V defined by


T (α i) = β i , i = 1, 2, … , n.
 n  n
Proof: We have T (α) = T  Σ xi α i = Σ xi T (α i) [∵ T is linear]
 i =1  i =1
n
= Σ xi β i = β.
i =1

Example 6: Show that the set


S = {(1, 0 , 0 ), (1, 1, 0 ), (1, 1, 1)}
is a basis of R 3 (R) where R is the field of real numbers. Hence find the coordinates of the vector
(a, b, c ) with respect to the above basis. (Kumaun 2010)
3
Solution: The dimension of the vector space R (R) is 3. If the set S is linearly
independent, then S will form a basis of R 3 (R) because S contains 3 vectors. Let
x, y, z be scalars in R such that
x (1, 0 , 0 ) + y (1, 1, 0 ) + z (1, 1, 1) = 0 = (0 , 0 , 0 )
⇒ ( x + y + z , y + z , z ) = (0 , 0 , 0 )
⇒ x + y + z = 0, y + z = 0, z = 0
⇒ x = 0, y = 0, z = 0
⇒ the set S is linearly independent.
∴ S is a basis of R 3 (R).
Now to find the coordinates of (a, b, c ) with respect to the ordered basis S. Let p, q, r be
scalars in R such that
(a, b, c ) = p (1, 0 , 0 ) + q (1, 1, 0 ) + r (1, 1, 1)
⇒ (a, b, c ) = ( p + q + r, q + r, r)
⇒ p + q + r = a, q + r = b, r = c
⇒ r = c , q = b − c , p = a − b.
Hence the coordinates of the vector (a, b, c ) are ( p, q, r) i. e., (a − b, b − c , c ).
V-84

Example 7: Find the coordinates of the vector (2, 1, − 6) of R3 relative to the basis
α1 = (1, 1, 2), α2 = (3, − 1, 0 ), α3 = (2, 0 , − 1).
Solution: To find the coordinates of the vector (2 , 1, − 6) relative to the ordered basis
{α1, α2 , α3}, we shall express the vector (2 , 1, − 6) as a linear combination of the vectors
α1, α2 , α3 . Let p, q, r be scalars in R such that
(2 , 1, − 6) = pα1 + qα2 + rα3
⇒ (2 , 1, − 6) = p (1, 1, 2) + q (3, − 1, 0 ) + r (2 , 0 , − 1)
⇒ (2 , 1, − 6) = ( p + 3 q + 2 r, p − q, 2 p − r).
∴ p + 3 q + 2 r = 2 , p − q = 1, …(1)
and 2 p − r = − 6.
Solving the equations (1), we have
p = − 7 / 8, q = − 15 / 8
and r = 17 / 4.
Hence the required coordinates of the vector (2 , 1, − 6) relative to the ordered basis
{α1, α2 , α3} are ( p, q, r) i. e., (− 7 / 8, − 15 / 8, 17 / 4).

Comprehensive Exercise 3

1. Show that the set S = {(1, 0 , 0 ), (1, 1, 0 ), (1, 1, 1)} is a basis of C3 (C) where C is the
field of complex numbers. Hence find the coordinates of the vector
(3 + 4 i, 6 i, 3 + 7 i) in C3 with respect to the above basis.
2. Find the co-ordinate vector of (3, 5, − 2) relative to the basis { e1, e2 , e3 } where
e1 = (1, 1, 1), e2 = (0 , 2, 3), e3 = (0 , 2, − 1).
3. Let B = {α1, α2 , α3 } be an ordered basis for R 3 , where
α1 = (1, 0 , − 1), α2 = (1, 1, 1), α3 = (1, 0 , 0 ).
Obtain the coordinates of the vector (a, b, c ) in the ordered basis B.
4. Let V be the vector space of all polynomial functions of degree less than or equal
to two from the field of real numbers R into itself. For a fixed t ∈R, let
g1 ( x) = 1, g2 ( x) = x + t, g3 ( x) = ( x + t)2 .
Prove that { g1, g2 , g3 } is a basis for V and obtain the coordinates of
c0 + c1 x + c2 x2 in this ordered basis.

A nswers 3
1. (3 − 2 i, − 3 − i, 3 + 7 i) 2. (3, − 1, 2)
3. (b − c , b, a − 2 b + c ) 4. (c0 − c1 t + c2 t2 , c1 − 2 c2 t, c2 )
V-85

O bjective T ype Q uestions

Multiple Choice Questions


Indicate the correct answer for each question by writing the corresponding letter from (a),
(b), (c) and (d).

1. If W be a subspace of a finite dimensional vector space V ( F ) , then


dim V / W =
(a) dim V − dim W (b) dim V + dim W
(c) (dim V ) / (dim W ) (d) None of these.
(Garhwal 2010, 12)
2. The set S = {(1, 0 , 0 ), (1, 1, 0 ), (1, 1, 1)} is a basis of R 3 (R ), then coordinates of the
vector (a b, c ) with respect to the basis are
(a) (a + b, b − c , c ) (b) (a − b, b − c , c )
(c) (a − b, b + c , c ) (d) (a − b, b − c , − c )
V 3
3. The dimension of , where V = R , W = { (a, 0 , 0 ) : a ∈ R } is
W
(a) 1 (b) 3
(c) 2 (d) 4

Fill in the Blank(s)


Fill in the blanks “……” so that the following statements are complete and correct.

1. Let V be the vector space of all functions from the real field R into R. Let W1 be the
subspace of even functions and W2 be the subspace of odd functions. Then
V = …… .
2. If W is an m-dimensional subspace of an n-dimensional vector space V, then the
V
dimension of the quotient space is …… .
W (Garhwal 2006)
3. Two subspaces W1 and W2 of the vector space V ( F ) are said to be disjoint if
W1 ∩ W2 = …… .
4. If a finite dimensional vector space V ( F ) is a direct sum of two subspaces W1
and W2 , then dim V = …… .

True or False
Write ‘T’ for true and ‘F’ for false statement.
1. If a finite dimensional vector space V ( F ) is a direct sum of two subspaces W1 and
W2 , then dim V = dim W1 + dim W2 .
V,
2. The dimension of where V = R 3 , W = { (a, 0 , 0 ) : a ∈ R } is 3.
W
3. If W1 = {( x, 0 , z ) : x, z ∈ R }, W2 = {(0 , y, z ) : y, z ∈ R } be two subspaces
of R 3 then R 3 ≠ W1 + W2 .
V-86

A nswers

Multiple Choice Questions


1. (a) 2. (b) 3. (c)

Fill in the Blank(s)


1. W1 ⊕ W2 2. n − m
3. {0} 4. dim W1 + dim W2

True or False
1. T 2. F 3. F

¨
5
Matrices

1 Idempotent, Involutary, Nilpotent and Periodic Matrices


dempotent Matrix. A matrix such that A 2 = A is called idempotent matrix.
I
Example,

 2 −2 −4
 
A = −1 3 4
 1 −2 −3

Involutary Matrix. A matrix A is said to be an involutary matrix if A2 = I ( unit


matrix). Example,
−5 −8 0
 
A= 3 5 0
 1 2 −1

Since I2 = I always, we conclude that a Unit matrix is always involutary.


V-88

Nilpotent Matrix. A matrix is said to be a nilpotent matrix if A k = 0 if ( null matrix)


where k is a +ve integer; if however k is the least + ve integer for which A k = 0, then k is
called as the index of the nilpotent matrix. Examples,

 ab b2 
A=  is nilpotent of index 2.
2
− a − ab 

 1 −3 4
 
A = −1 3 4 is nilpotent of index 2.
 1 −3 −4

Periodic Matrix. A matrix A is said to be a periodic matrix if A k + 1 = A, where k is a


+ve integer. If k is the least +ive integer for which A k + 1 = A then k is said to be period
of A. If we choose k = 1, then A2 = A and we call it to be an idempotent matrix.

Example 1: A matrix A such that A 2 = A is called Idempotent. Determine all the


idempotent diagonal matrices of order n.

Solution: Let A = diag. [d1, d2 , d3 , …, dn] be an idempotent matrix of order n.

Then A 2 = A .
d1 0 0 ... 0 d1 0 0 ... 0
   
0 d2 0 ... 0 0 d2 0 ... 0
Now A2 = 
... ... ... ... 0 ... ... ... ... 0
   
 0 0 0 ... dn  0 0 0 ... dn

d 2 0 0 ... 0
1 2

0 d2 0 ... 0
= .
 ... ... ... ... 0
0 0 0 ... dn 
2

∴ A 2 = A gives
2 2 2
d1 = d1, d2 = d2 , … , dn = dn

i. e., d1 = 0 , 1; d2 = 0 , 1 ; … ; dn = 0, 1.

Hence Diag. [d1 , d2 ..., dn], d1, d2 , …, dn = 0, 1 is the required idempotent diagonal
matrix of order n.
V-89

Example 2: If B is an idempotent matrix, show that A = I − B is also idempotent and that


AB = BA = O.

Solution: Since B is idempotent, therefore B 2 = B.


Now A 2 = (I − B) 2= (I − B) (I − B)

= I2 − IB − BI + B 2

= I − B − B + B2 [∵ IB = B = BI]
=I− B− B+ B [∵ B2 = B]
=I− B [ ∵ − B + B = O]
= A.

Since A 2 = A, therefore A is idempotent.


Again AB = (I − B) B = IB − B 2 = B − B = O.
Similarly BA = B (I − B) = BI − B 2 = B − B = O.
−5 −8 0
Example 3: Show that the matrix A =  3 5 0  is involutory.
 1 2 −1 

Solution: The matrix A is involutory if A 2 = I.


−5 −8 0 −5 −8 0
We have A 2 = AA =  3 5 0  3
 5 0 
 1 2 −1  1 2 −1 
 25 − 24 + 0 40 − 40 + 0 0 +0 +0 
= −15 + 15 + 0 −24 + 25 + 0 0 +0 +0 

 −5 + 6 − 1 −8 + 10 − 2 0 + 0 +1 
1 0 0
= 0 1 0  = I.
0 0 1 
∴ the matrix A is involutory.

Example 4: Show that the matrix

 1 1 3
A =  5 2 6  is nilpotent and find its index.
−2 −1 −3 
Solution: We have A 2 = AA
 1 1 3  1 1 3
=  5 2 6  ×  5 2 6 
−2 −1 −3  −2 −1 −3 
V-90

0 0 0
=  3 3 9  ⋅
−1 −1 −3 
 1 1 3 0 0 0
Again A = AA =  5
3 2
2 6   3
 3 9 
−2 −1 −3  −1 −1 −3 
0 0 0
= 0 0 0  = O.
0 0 0 

Thus 3 is the least positive integer such that A 3 = O. Hence the matrix A is nilpotent of
index 3.

2 Determinant of a Square Matrix


Let A be any square matrix. The determinant formed by the elements of A is said to be
the determinant of matrix A. This is denoted by| A|or det A. Since in a determinant the
number of rows is equal to the number of columns, therefore only square matrices
can have determinants.

12 0 13 12 0 13


   
Hence, if A = 15 12 11, then det A = | A| = 15 12 11 .
13 11 14 13 11 14
 

Difference between a matrix and a determinant.


(i) A matrix “A” cannot be reduced to a number whereas the determinant can be
reduced to a number.

(ii) The number of rows may or may not be equal to number of columns in a matrix
while in a determinant the number of rows is equal to the number of columns.

(iii) Interchanging the rows and columns, a different matrix is formed while in a
determinant, an interchange of rows and columns does not change the value of
the determinant.

3 Non-singular and Singular Matrices


Definition: A square matrix A is said to be non-singular or singular according as
| A| ≠ 0 or | A| = 0 .
V-91

4 Negative Integral Powers of a Non-singular Matrix


The negative integral powers are defined only for non-singular matrices.

Definition: Let A and B be non-singular matrices. If p is any positive integer, then we define
(A − p ) = (A p )−1.

A −1 is the unique inverse of A .

If A is non-singular and its negative integral powers are considered, then for any
integers r, s, we have

(i) A r A s = A r +s (ii) (A r )s = A rs .

For two non-singular matrices A and B of the same order, the index law
(AB)m = A mB m fails. However it becomes valid if A and B commute with respect to
multiplication.

Theorem 1: If A is any n-rowed non-singular matrix, then


(A − k ) = (A −1)k , for all k ∈ N .

Proof: If A be a non-singular matrix and k ∈ N , by definition, we have


A − k = (A k )−1

= (A ⋅ A ⋅ .... k times)−1

= (A −1 ⋅ A −1 … k times)

= (A −1)k .

Hence, A − k = (A k )−1 = (A −1)k .

Theorem 2: Let A , B be invertible matrices, then

(i) (A −1)−1 = A

(ii) (kA )−1 = k −1A −1, 0 ≠ k ∈ C

(iii) (AB)−1 = B −1 A −1 (Reversal Law for the Inverse of a Product)

(iv) (A ′ )−1 = (A −1)′

(v) (A θ )−1 = (A −1)θ .

Proof: (i) Since A is an invertible matrix therefore there exists a matrix X = A −1 such
that AX = XA = I.
V-92

By definition of inverse, we have


X −1 = A .

Hence, (A −1)−1 = A .

(ii) Let X = k −1A −1. Then, we have


(kA ) X = (kA ) (k −1A −1)

= kk −1AA −1

= I.

Similarly, X (kA ) = I.
∴ (kA ) X = X (kA ) = I.

Hence, (kA )−1 = X = k −1A −1.

(iii) Let X = B −1A −1. Then, we have


(AB) X = (AB) (B −1A −1)

= A [B (B −1 A −1)]

= A (BB −1 A −1)

= A (IA −1)

= AA −1

= I.

Similarly, X (AB) = I.
∴ (AB) X = X (AB) = I.

Hence, (AB)−1 = X = B −1A −1.

(iv) By definition of inverse, we have


AA −1 = A −1A = I.

Taking transpose, we get


(AA −1)′ = (A −1A )′ = I ′
or (A −1)′ A ′ = A ′ (A −1)′ = I [∵ (AB)′ = B ′ A ′ and I ′ = I]

This equality shows that (A −1)′ is the inverse of the matrix A ′.


Hence, (A ′ )−1 = (A −1)′.

(v) The proof is similar to that of part (iv). Take transposed conjugate instead of
transpose.
V-93

5 Trace of a Matrix
Definition: Let A be a square matrix of order n. The sum of the elements of A lying along the
principal diagonal is called the trace of A . We shall write the trace of A as tr A . Thus if
n
A = [ai] n× n , then tr A = Σ aii = a11 + a22 + … + ann.
i =1

In the following theorem we have given some fundamental properties of the trace
function.

Theorem: Let A and B be two square matrices of order n and λ be a scalar. Then
(1) tr ( λ A ) = λ tr A ;
(2) tr ( A + B) = tr A + tr B;
(3) tr ( AB) = tr (BA).

Proof: Let A = [aij ] n × n and B = [bij ] n × n.

(1) We have λ A = [λaij ] n× n by def. of multiplication of a matrix by a scalar.


n n
∴ tr ( λ A ) = Σ λaii = λ Σ aii = λ tr A .
i =1 i =1

(2) We have A + B = [aij + bij ] n × n.


n
∴ tr ( A + B) = Σ ( aii + bii).
i =1

n n
= Σ aii + Σ bii
i =1 i =1

= tr A + tr B.
n
(3) We have AB = [cij ] n × n , where cij = Σ aik bkj .
k =1
n
Also BA = [d ij ] n × n , where d ij = Σ bik akj .
k =1
n  n n 
Now tr ( AB) = Σ cii = Σ  Σ aik bki
i =1 i = 1 k = 1 
n n
= Σ Σ aik bki , interchanging the order of summation in
k =1 i =1 the last sum
n  n  n
= Σ  Σ bki aik  = Σ d kk
k = 1 i = 1  k =1

= d11 + d22 + … + dnn = tr (BA ).


V-94

6 Transpose of a Matrix
Definition: Let A = [aij]m × n. Then the n × m matrix obtained from A by changing its rows
into columns and its columns into rows is called the transpose of A and is denoted by the symbol A′
or AT .

The operation of interchanging rows with columns is called transposition.


Symbolically if

A = [aij]m × n ,

then A′ = [b ji]n × m , where b ji = aij,

i. e., the ( j, i)th element of A′ is the (i, j)th element of A.

For example, the transpose of the 3 × 4 matrix

1 2 3
1 2 3 4  
  2 3 4
A = 2 3 4 1 is the 4 × 3 matrix A′ =  .
3 3 4 2
 4 2 13 × 4  
4 1 14 × 3

The first row of A is the first column of A′. The second row of A is the second column of
A′ . The third row of A is the third column of A′.

Theorems: If A′ and B′ be the transposes of A and B respectively, then

(i) (A′ )′ = A;

(ii) (A + B)′ = A ′ + B ′ , A and B being of the same size.

(iii) (kA)′ = kA ′ , k being any complex number.

(iv) (AB)′ = B ′ A ′ , A and B being conformable to multiplication.

The above law (iv) is called the reversal law for transposes i. e., the transpose of the product
is the product of the transposes taken in the reverse order.

7 Orthogonal Matrix
Definition: A square matrix A is said to be orthogonal if A ′ A = I . (Kanpur 2009)

If A is an orthogonal matrix, then A ′ A = I

⇒ | A ′ A| = | I | ⇒ | A ′| ⋅| A| = 1 [∵det (AB) = ( det A). (det B)]

⇒ | A| ⋅| A| = 1 [∵| A ′| = | A|]
V-95

⇒ | A|2 = 1 ⇒ | A| = ± 1 ⇒ | A| ≠ 0

⇒ A is invertible.

Also then A ′ A = I ⇒ A ′ = A −1 which in turn implies AA ′ = I.

Thus A is an orthogonal matrix if and only if

A ′ A = I = AA ′ .

Theorem: If A, B be n-rowed orthogonal matrices,AB and BA are also orthogonal matrices.

Proof: Since A and B are both n-rowed square matrices, therefore AB is also an n-rowed
square matrix.

Since | AB | = | A| ⋅| B | and | A| ≠ 0 , also | B|1 ≠ 0 , therefore |AB | ≠ 0. Hence, AB is a


non-singular matrix.

Now (AB)′ = B ′ A ′.

∴ (AB)′ (AB) = (B ′ A ′ ) (AB)

= B ′ (A ′ A) B

= B ′ IB [∵ A ′ A = I]

= B′ B

= I. [∵ B ′ B = I]

∴ AB is orthogonal. Similarly we can prove that BA is also orthogonal.

0 2β γ
 
Example 5: Determine the values of α, β, γ when α β − γ  is orthogonal.
α −β γ 

0 2β γ
 
Solution : Let A = α β − γ .
α −β γ 

 0 α α
 
Then A′ = 2β β −β .
 γ −γ γ 

If A is orthogonal, then AA ′ = I .
V-96

0 2β γ  0 α α  1 0 0
     
α β − γ 2β β −β =  0 1 0
α −β γ   γ −γ γ  0 0 1
 

 4β2 + γ 2 2β2 − γ 2 −2β2 + γ 2  1 0 0


 2 2 2 2 2 2 2 2
  
or  2β − γ α +β + γ α − β − γ  = 0 1 0
  0
2
−2β + γ
2 2
α −β − γ 2 2
α 2 + β2 + γ 2   0 1

Now equating the corresponding elements, we get

4β2 + γ 2 = 1 …(1)

2β2 − γ 2 = 0 …(2)

α2 + β2 + γ 2 = 1. …(3)
1 1
From (1) and (2), we get β = ± ,γ = ± .
6 3
1
From (3), we get α = ± .
2
1 1 1
Hence α=± ,β = ± ,γ = ± .
2 6 3

Example 6: Show that the matrix

 cos θ sin θ
  is orthogonal.
− sin θ cos θ

 cos θ sin θ 
Solution: Let A=  .
− sin θ cos θ

cos θ − sin θ
Then A′ =  .
sin θ cos θ

 cos θ sin θ  cos θ − sin θ


We have AA′ =    
− sin θ cos θ sin θ cos θ 

1 0
= 
0 1

= I2 .

Hence the matrix A is orthogonal.


V-97

Example 7: Verify that the matrix

 1 2 2
1 
 2 1 −2 is orthogonal.
3
−2 2 −1

 1 2 2
1 
Solution: Let A=  2 1 −2 .
3
−2 2 −1

1 2 −2
1  
Then A′ = 2 1 2
3 2
 −2 −1

 1 2 2 1 2 −2
1   
We have AA′ =  2 1 −2 2 1 2
9
−2 2 −1 2
 −2 −1

9 0 0
1  
= 0 9 0 .
9 0
 0 9

1 0 0
 
= 0 1 0  = I3 .
0 0 1

Hence, the matrix A is orthogonal.

8 Conjugate of a Matrix
If i = (−1 ), then z = x + iy is called a complex number where x and y are any real
numbers. If z = x + iy, then z = x − iy is called the conjugate of the complex number z.

We have zz = ( x + iy) ( x − iy) = x2 + y2 i. e., is real.

Also if z = z , then x + iy = x − iy i. e., 2 iy = 0 i. e., y =0 i. e., z is real.

Conversely, if z is real then z = z .

If z = x + iy, then z = x − iy. ∴ (z ) = x + iy = z .

If z1 and z2 are two complex numbers, then it can be easily seen that

(i) z1 + z2 = z1 + z2 and (ii) z1z2 = (z1) (z2 ).


V-98

Conjugate of a Matrix
Definition: The matrix obtained from any given matrix A on replacing its elements by the
corresponding conjugate complex numbers is called the conjugate of A and is denoted by A.

Thus if A = [aij]m × n, then A = [aij]m × n where aij denotes the conjugate complex of aij.

If A be a matrix over the field of real numbers, then obviously A coincides with A.

2 + 3 i 4 − 7i 8  2 − 3 i 4 + 7i 8 
Illustration: If A =  , then A =  .
 −i 6 9 + i  i 6 9 − i

Theorem: If A and B be the conjugates of A and B respectively, then

(i) (A) = A

(ii) (A + B) = A + B, A and B being of the same size

(iii) (k A) = k A, k being any complex number

(iv) (AB) = A B, A and B being conformable to multiplication.

9 Transposed Conjugate of a Matrix


Definition: The transpose of the conjugate of a matrix A is called transposed conjugate of A and
is denoted byA θ or by A *.

Obviously the conjugate of the transpose of A is the same as the transpose of the
conjugate of A i. e.,

( A ′) = (A)′ = A θ .

If A = [aij]m × n, then A θ = [b ji]n × m

where b ji = aij i. e., the ( j, i)th element of A θ = the conjugate complex of the (i, j)th
element of A.

1 + 2 i 2 − 3i 3 + 4 i
 
Illustration: If A = 4 − 5 i 5 + 6i 6 − 7 i ,
 8 7 + 8i 7 

1 + 2 i 4 − 5i 8 
 
then A′ = 2 − 3 i 5 + 6i 7 + 8 i
3 + 4 i 6 − 7i 7 

V-99

1 − 2 i 4 + 5i 8 
θ 
and (A ′ ) = A = 2 + 3 i 5 − 6i 7 − 8 i .
3 − 4 i 6 + 7i 7 

Theorem: If A θ and B θ be the transposed conjugates of A and B respectively, then

(i) (A θ )θ = A

(ii) (A + B)θ = A θ + B θ , A and B being of the same size

(iii) (kA)θ = k A θ , k being any complex number

(iv) (AB)θ = B θ A θ , A and B being conformable to multiplication.

10 Invertible Matrices

Inverse or reciprocal of a matrix.


Definition: Let A be any n-rowed square matrix. Then a matrix B, if it exists, such that
AB = BA = In is called inverse of A.

Note: For the products AB, BA to be both defined and be equal, it is necessary that A
and B are both square matrices of the same order. Thus non-square matrices cannot
possess inverse.

Existence of the Inverse. Theorem. The necessary and sufficient condition for a square
matrix A to possess the inverse is that | A| ≠ 0 .
1
Important: If A be an invertible matrix, then the inverse of A is Adj. A. It is usual to
| A|
denote the inverse of A by A −1 .
1
Thus, A −1 = Adj . A, provided | A| ≠ 0 .
| A|

Thus the necessary and sufficient condition for a matrix to be invertible is that it is
non-singular.

11 Reversal Law for the Inverse of a Product


Theorem 1: If A, B be two n-rowed non-singular matrices, then AB is also non-singular and
(AB) −1 = B −1 A −1, i. e., the inverse of a product is the product of the inverses taken in the reverse
order.
V-100

Theorem 2: If A be an n × n non-singular matrix, then (A ′ )−1 = (A −1)′ i. e., the operations of


transposing and inverting are commutative.

Theorem 3: If A be an n × n non-singular matrix, then (A −1)θ = (A θ )−1.

Example 8: Find the :

(i) Transpose (ii) Adjoint (iii) Inverse of the matrix :

 2 −1 3
 
A = −5 3 1 .
−3 2 3

 2 −5 −3
T  
Solution: (i) We have the transpose of A = A = −1 3 2 .
 3 1 3

2 −1 3
(ii) We have | A| = −5 3 1
−3 2 3

= 2 (9 − 2) − (− 1) (− 15 + 3) + 3 (− 10 + 9)

= 14 − 12 − 3 = − 1.

Let A ij denote the cofactor of the element aij of | A| .

3 1 −5 1
Then A11 = = 7, A12 = − = 12,
2 3 −3 3

−5 3
A13 = = − 1,
−3 2

−1 3 2 3
A21 = − = 9, A22 = = 15,
2 3 −3 3

2 −1
A23 = − = − 1,
−3 2

−1 3
A31 = = − 10 ,
3 1
V-101

2 3
A32 = − = − 17,
−5 1

2 −1
A33 = = 1.
−5 3

∴ The matrix B formed of the cofactors of the elements of |A| is,

 7 12 −1
 
B= 9 15 −1 .
−10 −17 1

 7 9 −10 
 
Now adj. A (= the transpose of the matrix B) = 12 15 −17 .
 −1 −1 1

 7 9 −10 
−1 1 1  
(iii) We have A = adj. A = 12 15 −17
| A| (− 1)  −1
 1 1

 7 9 −10 
 
= (−1) 12 15 −17
 −1 −1 1

 −7 −9 10 
 
= −12 −15 17 .
 1 1 −1

Note: To check that the answer is correct the students should verify that AA −1 = I.

12 Unitary Matrix
Definition: A square matrix A is said to be unitary if A θ A = I .

Since | A θ| = | A| and | A θ A| = | A θ|| A|, therefore if A θ A = I, we have | A|| A| = 1.

Thus the determinant of a unitary matrix is of unit modulus.

If A is a unitary matrix, then | A|| A θ| = 1 and so | A| ≠ 0 i. e., A is non-singular and so


invertible.

Hence A θ A = I implies AA θ = I .
V-102

Thus A is a unitary matrix iff A θ A = I = AA θ .

Theorem: If A, B be n-rowed unitary matrices, AB and BA are also unitary matrices.

Proof: Since A and B are both n-rowed square matrices, therefore AB is also an n-rowed
square matrix.

Since | AB| = | A|| B| and | A| ≠ 0 , also |B| ≠ 0 , therefore | AB| ≠ 0 .

Hence AB is a non-singular matrix.

Now (AB)θ = B θ A θ .

∴ (AB)θ (AB) = (B θ A θ ) (AB)

= B θ (A θ A) B

= B θ IB [∵ A θ A = I]

= Bθ B

= I. [∵ B θ B = I]

∴ AB is unitary. Similarly we can prove that BA is also unitary.

Example 9: Prove that

1  1 1+ i
B=   is unitary.
3 1 − i −1  (Bundelkhand 2006; Garhwal 11)

1  1 1− i
Solution: We have B=  .
3 1 + i −1 

1  1 1+ i
∴ B θ = (B)′ =  .
3 1 − i −1 

1 1 1 + i  1 1+ i
Now Bθ B =    
3 1 − i −1  1 − i −1 

1  1.1 + (1 + i) (1 − i) 1. (1 + i) + (1 + i) (−1) 
=  
3 (1 − i).1 + (−1).(1 − i) (1 − i) (1 + i) + (−1) (−1)
V-103

1 3 0
=  
3 0 3

1 0
=  = I.
0 1

∴ B is unitary.

α + iγ −β + iδ
Example 10: Show that the matrix A =   is a unitary matrix, if
β + iδ α − iγ 
α2 + β2 + γ 2 + δ2 = 1.
α + iγ −β + iδ
Solution: We have A= 
β + iδ α − iγ 
 α + iγ β + iδ
∴ A′ =  
− β + iδ α − iγ 

 α − iγ β − iδ
Aθ = A′ =  .
− β − iδ α + iγ 

A square matrix A is said to be unitary if AA θ = I.

α + iγ −β + iδ  α − iγ β − iδ 1 0
∴   = 
β + iδ α − iγ  − β − iδ α + iγ  0 1

 α2 + β2 + γ 2 + δ2
or 
αβ − iβγ + iαδ + γδ − αβ − iαδ + iβγ − δγ

αβ − iαδ + iβγ + γδ − αβ − i β α + iαδ − δγ  1 0


=  
β2 + δ2 + α2 + γ 2  0 1

α2 + β2 + γ 2 + δ2 0  1 0
or  = 
 0 α2 + β2 + γ 2 + δ2  0 1

if α2 + β2 + γ 2 + δ2 = 1 .

Example 11: If A is a unitary matrix, show that A −1 is also unitary.

Solution: We have AA θ = A θ A = 1, since A is a unitary matrix.


⇒ (AA θ )−1 = (A θ . A)−1 = (I)−1 [Taking inverse]
⇒ (A θ )−1. A −1 = A −1 (A θ )−1 = I
⇒ (A −1)θ . A −1 = A −1 (A −1)θ = I.
Hence A −1 is a unitary matrix.
V-104

13 Matrices of Rotation and Reflection


A rotation about the origin O by an angle θ is denoted by Rot (θ) and a reflection about
a line L through the origin which makes an angle θ with the x-axis is denoted by Ref (θ).
If these rotations and reflections operate on all points on the plane and if these points
be represented by position vectors, then a rotation is represented by a matrix.
cos θ − sin θ
Rot (θ) =  
sin θ cos θ 

and a reflection is represented by a matrix


cos 2θ sin 2θ 
Ref. (θ) =  .
sin 2θ − cos 2θ

Rot (θ) rotates column vectors by means of the following matrix multiplication
 x ′  cos θ − sin θ  x 
 =   .
 y ′ sin θ cos θ  y

Thus the coordinates ( x ′ , y ′ ) of the point ( x, y) after rotation are

x ′ = x cos θ − y sin θ,

y ′ = x sin θ + y cos θ.

The direction of vector rotation is positive if angle θ is measured counter clockwise


(e.g. 90°), and negative if angle θ is measured clockwise (e.g. − 90 °).

The clockwise rotation is represented by a matrix


 cos θ sin θ 
Rot (− θ) =  .
 − sin θ cos θ

Reflection transformation matrix is the matrix used to make reflection transformation


of a figure.

The following relations hold true for rotation and reflection

(i) Rot (θ) Rot (φ) = Rot (θ + φ)

(ii) Ref (θ) Ref (φ) = Rot (2θ − 2 φ)

 1 
(iii) Rot (θ) Ref (φ) = Ref  φ + θ
 2 

 1 
(iv) Ref (φ) Rot (θ) = Ref  φ − θ .
 2 

These relations can be proved using matrix multiplication and application of


trigonometric identities.
V-105

The set of all reflections in lines through the origin and rotations about the origin,
together with the operation of composition of reflections and rotations, forms a group
Rots (0) is the identity of the group. There exists an inverse Rot (− ϕ) for every rotation
(Rot ϕ). Every reflection Ref. (θ) is its own inverse. Matrix multiplication being
associative, the composition has closure and is associative. Ref (θ) and Rot (θ) both
being represented by orthogonal matrices have a determinant whose absolute value is
unity. Rotation matrices have a determinant of value +1 and reflection matrices have
a determinant of value −1 .

Note : The set of all two dimensional matrices together with matrix multiplication
form the orthogonal group O(2).

Rotation Matrix Transformations

1 0 Identity matrix. x  x 
   y →  y
0 1 Right remains right, up remains up.    

 0 1 Rotation by 90° clockwise. x   y


   y → − x
 −1 0 Right has become down, up has become right.    

0 −1 Rotation by 90° anticlockwise. x   − y


   y →  x
1 0 Right has become up, up has become left.    

 −1 0 Rotation by 180°. x  − x 
   y →  − y
 0 −1 Right has become left, up has become down.    

Rotation by 270° anticlockwise.


 0 1 Right has become down, up has become right x   y
   y →  − x 
 −1 0
A positive 90° rotation (left) after a 180°    
rotation gives a negative 90° rotation (right).

Reflection Matrix Transformations

1 0 Reflection in the x-axis. x  − x 


   y →  − y
0 −1 Right remains right, up has become down.    

 −1 0 Reflection in the y-axis. x  − x


   y →  y 
 0 1 Right has become left, up remains up.    
V-106

0 1 Reflection in the line y = x. x   y


   y → x 
1 0 Right has become up, up has become right.    

 0 −1 Reflection in the line y = − x. x   − y


   y → − x 
 −1 0 Right has become down, up has become left.    

 −1 0 Reflection about the origin. x  − x 


   y →  − y
 0 −1 Right has become left, up has become down.    

Note: Combinations of these matrices give multiple transformations. Two


reflections generate a rotations.

Comprehensive Exercise 1

 2 −2 −4
1. Show that the matrix A = −1 3 4 is idempotent.
 1 −2 −3

2. If AB = A and BA = B, show that A and B are idempotent.

3. A matrix A such that A 2 = I is called involutory. Show that A is involutory if and


only if (I + A) (I − A) = O.

 ab b2  2
4. Show that the matrix A =   is nilpotent such that A = O .
2
− a − ab

−1 2 2
1 
5. Show that the matrix  2 −1 2 is orthogonal.
3
 2 2 −1

 1 1 1
 −
3 6 2
 1 2 
6. Verify that the matrix  − 0  is orthogonal.
 3 6 
 1 1 1 
 3 6 2 

1  1 i
7. Show that the matrix =   is unitary.
2 − i −1 
V-107

1 + i − 1 + i
 2  is unitary.
8. Prove that the matrix  2
1+ i 1− i 
 
 2 2  (Rohilkhand 2010)

9. If (lr , mr , nr ), r = 1, 2, 3 be the direction cosines of three mutually perpendicular


lines referred to an orthogonal cartesian co-ordinate system, then prove that

 l1 m1 n1 
 
l2 m2 n2  is an orthogonal matrix.
l m3 n3 
3

10. Show that if A is an orthogonal matrix, then A ′ and A −1 are also orthogonal.
 0 1 + 2 i −1
11. If N =   is a matrix, then show that (I − N) (I + N) is a unitary
− 1 + 2 i 0 
matrix, where I is an identity matrix.

Objective Type Questions

Multiple Choice Questions


Indicate the correct answer for each question by writing the corresponding letter from (a),
(b), (c) and (d).
 ab b2 
1. A matrix A =   is nilpotent matrix of index
2
− a − ab
(a) 1 (b) 2
(c) 3 (d) 4
 6 9
2. A matrix A =   is
−4 −6
(a) unitary (b) orthogonal
(c) idempotent (d) nilpotent

3. A square matrix A is said to be orthogonal if


(a) A ′A = A (b) A ′A = I
(c) A ′A = O (d) A ′A = A 2

4. The necessary and sufficient condition for a matrix to be invertible is that it is


(a) singular (b) non-singular
(c) zero matrix (d) none of these (Rohilkhand 2007)
V-108

Fill in the Blank(s)


Fill in the blanks “……” so that the following statements are complete and correct.

1. A matrix A is said to be involutary if .......

2. A square matrix A is said to be idempotent if .......

3. A square matrix A is said to be unitary if .......

True or False
Write ‘T’ for true and ‘F’ for false statement.

1. If A is a unitary matrix then A −1 is also unitary.

2. If A is an orthogonal matrix then A ′ is also orthogonal.

 cos θ sin θ 
3. A matrix A =   is unitary matrix.
 − sin θ cos θ

A nswers

Multiple Choice Questions


1. (b) 2. (d) 3. (b) 4. (b)

Fill in the Blank(s)


1. A2 = I 2. A2 = A 3. Aθ A =I

True or False
1. T 2. T 3. F

o
V-109

6
Rank of a Matrix

1 Submatrix of a Matrix
uppose A is any matrix of the type m × n. Then a matrix obtained by leaving some
S rows and columns from A is called a submatrix of A. In particular the matrix A
itself is a sub-matrix of A because it is obtained from A by leaving no rows or columns.

Minors of a matrix. We know that every square matrix possesses a determinant. If A


be an m × n matrix, then the determinant of every square sub-matrix of A is called a
minor of the matrix A. If we leave m − p rows and n − p columns from A, we shall get a
square submatrix of A of order p. The determinant of this square submatrix is called a
p-rowed minor of A.

For example, let

2 4 1 9 1
 
0 5 2 5 2
A=
1 9 7 3 4
 
3 −2 8 1 84 × 5.

In a determinant the number of rows is equal to the number of columns. Therefore


there can be no 5-rowed minor of A.
V-110

If we leave any columns from A, we shall get a square sub-matrix of A of order 4.

Thus 2 4 1 9 2 4 9 1
0 5 2 5 0 5 5 2
, , etc.
1 9 7 3 1 9 3 4
3 −2 8 1 3 −2 1 8

are 4-rowed minors of A.

If we leave two columns and one row from A, we shall get a square submatrix of A of
order 3.

Thus 2 4 1 4 1 9 5 2 5
0 5 2 , 5 2 5 , 9 7 3 , etc.
1 9 7 9 7 3 −2 8 1

are 3-rowed minors of A.

If we leave three columns and two rows from A , we shall get a square submatrix of A of
order 2.

Thus 2 4 4 1 5 2
, , , etc. are 2-rowed minors of A.
0 5 5 2 9 7

The numbers 2, 4, 1, 9, 1, 0, 5 etc. are all 1-rowed minors of A.

2 Rank of a Matrix
(Lucknow 2006)

Definition: A number r is said to be the rank of a matrix A if it possesses the following two
properties :

(i) There is at least one square submatrix of A of order r whose determinant is not equal to zero.

(ii) If the matrix A contains any square submatrix of order r + 1, then the determinant of every
square submatrix of A of order r + 1 should be zero.

In short the rank of a matrix is the order of any highest order non-vanishing minor of the matrix.

Thus the rank of a matrix A is the order of any highest order square submatrix of A whose
determinant is not equal to zero.

We shall denote the rank of a matrix A by the symbol ρ (A).


V-111

It is obvious that the rank r of an (m × n) matrix can at most be equal to the smaller of
the numbers m and n , but it may be less.

If there is a matrix A which has at least one non-zero minor of order n and there is no
minor of A of order n + 1, then the rank of A is n. Thus the rank of every non-singular
matrix of order n is n. The rank of a square matrix A of order n can be less than n if and
only if A is singular i. e., |A| = 0.

Note 1: Since the rank of every non-zero matrix is ≥ 1, we agree to assign the rank, zero, to
every null matrix :

Note 2: Every (r + 1)-rowed minor of a matrix can be expressed as a linear combination


of its r-rowed minors. Therefore if all the r-rowed minors of a matrix are equal to zero,
then obviously all its (r + 1)-rowed minors will also be equal to zero.

Important : The following two simple results will help us very much in finding the
rank of a matrix:

(i) The rank of a matrix is ≤ r, if all (r + 1) - rowed minors of the matrix vanish.

(ii) The rank of a matrix is ≥ r, if there is at least one r-rowed minor of the matrix which is not
equal to zero.

Illustration:

1 0 0
 
(a) Let A = I3 = 0 1 0  be a unit matrix of order 3.
0 0 1

We have |A| = 1. Therefore A is a non-singular matrix. Hence rank A = 3. In


particular, the rank of a unit matrix of order n is n.

0 0 0
 
(b) Let A = 0 0 0 .
0 0 0 

Since A is a null matrix, therefore rank A = 0.

1 2 3
 
(c) Let A = 2 3 4 .
0 2 2

We have | A| = 1 (6 − 8) − 2 (4 − 6) = 2 ≠ 0 . Thus A is a non-singular matrix. Therefore


rank A = 3.
V-112

1 2 3
 
(d) Let A = 3 4 5 .
4 5 6

We have | A| = 1 (24 − 25) − 2 (18 − 20 ) + 3 (15 − 16) = 0

Therefore the rank of A is less than 3. Now there is at least one minor of A of order 2,
1 2
namely which is not equal to zero. Hence rank A = 2.
3 4
3 1 2
 
(e) Let A = 6 2 4 .
3 1 2

We have | A| = 0 , since the first two columns are identical.

Also each 2-rowed minor of A is equal to zero. But A is not a null matrix. Hence rank
A = 1.
2 4 3 2
(f) Let A =  .
3 5 1 4

2 4
Here we see that there is at least one minor of A of order 2 i. e., which is not
3 5
equal to zero. Also there is no minor of A of order greater than 2. Hence the rank of
A = 2.

3 Echelon Form of a Matrix


Definition: A matrix A is said to be in Echelon form if :

(i) Every row of A which has all its entries 0 occurs below every row which has a non-zero entry.

(ii) The number of zeros before the first non-zero element in a row is less than the number of such
zeros in the next row.

Important result: The rank of a matrix in Echelon form is equal to the number of non-zero
rows of the matrix.
0 1 2 3
 
Illustration: Find the rank of the matrix 0 0 1 −1 .
0 0 0 0 

V-113

The matrix A has one zero row. We see that it occurs below every non-zero row.

Further the number of zeros before the first non-zero element in the first row is one.
The number of zeros before the first non-zero element in the second row is two. Thus
the number of zeros before the first non-zero element in any row is less than the
number of such zeros in the next row.

Thus the matrix A is in Echelon form.

∴ rank A = the number of non-zero rows of A = 2.

4 Theorem
The rank of the transpose of a matrix is the same as that of the original matrix.

Proof: Let A be any matrix and A′ be the transpose of the matrix A. Let rank A = r and
rank A′ = s. Then to prove that r = s.

We have rank A = r

⇒ there exists at least one r-rowed square submatrix, say R, of A such that

| R ′| = | R| ≠ 0

⇒ there exists at least one r-rowed square submatrix R′ of A ′ such that

| R ′| = | R| ≠ 0

⇒ rank A′ ≥ r ⇒ s ≥ r. ...(1)

Since (A ′ )′ = A, therefore interchanging the roles of A and A′ in the above result (1), we
have

r ≥ s. …(2)

From (1) and (2), we conclude that r = s. Hence the result.

Example 1: Find the ranks of the following matrices :

1 2 3
  1 2 3
(i) 2 1 0 , (ii)  
0 2 4 5
 1 2 (Bundelkhand 2009)
V-114

1 2 3
 
Solution : (i) Let A = 2 1 0 .
0 1 2

We have | A| = 1 (2 − 0 ) − 2 (4 − 0 ) + 3 (2 − 0 ), expanding along the first row

= 2 − 8 + 6 = 0.

1 2
But there is at least one minor of order 2 of the matrix A, namely which is not
2 1
equal to zero. Hence rank A = 2.

1 2 3
(ii) Let A= .
2 4 5

1 3
Here there is at least one minor of order 2 of the matrix A, namely which is not
2 5
equal to 0. Also there is no minor of the matrix A or order greater than 2. Hence rank
A = 2.

Example 2: Under what conditions the rank of the following matrix is 3

2 4 2
 
A = 2 1 2 ?
1 0 x

Solution: The rank of the given matrix A = 3.

∴ The minor of order 3 of matrix A ≠ 0

2 4 2
i. e., 2 1 2 ≠0
1 0 x

or 2 ( x − 0 ) − 4 (2 x − 2) + 2 (0 − 1) ≠ 0

or 2x − 8x + 8 − 2 ≠ 0

or − 6x + 6 ≠ 0

or x ≠ 1.
V-115

1 5 4
 
Example 3: For which value of ‘b’ the rank of the matrix A = 0 3 2 is 2?
b 13 10 

Solution: The rank of the given matrix A = 2.

∴ The minor of order 3 of matrix A = 0

1 5 4
i. e., 0 3 2 =0
b 13 10

or 1 (30 − 26) − 5 (0 − 2 b) + 4 (0 − 3 b) = 0

or 4 + 10 b − 12 b = 0

or 4 − 2b = 0

or b = 2.

Example 4: Find the values of a so that rank (A) < 3, where A is the matrix

3 a − 8 3 3 
 
A= 3 3a − 8 3 .
 3 3 3 a − 8

Solution: The rank of the given matrix A < 3

∴ The minor of order 3 of matrix A = 0

3 a − 8 3 3 
 
i. e.,  3 3a − 8 3  =0
 3 3 3 a − 8

Applying R1 → R1 + R2 + R3 , we get

3a − 2 3a − 2 3a − 2
3 3a − 8 3 =0
3 3 3a − 8

1 1 1
or (3 a − 2) 3 3a − 8 3 =0
3 3 3a − 8
V-116

Applying C2 → C2 − C1, C3 → C3 − C1, we get

1 0 0
(3 a − 2) 3 3 a − 11 0 =0
3 0 3 a − 11

or (3 a − 2) (3 a − 11) (3 a − 11) = 0

2 11 11
or a= , , .
3 3 3

Example 5: Prove that the points ( x1, y1), ( x2 , y2 ) and ( x3 , y3 ) are collinear if and only if the
rank of the matrix.
 x1 y1 1
 
A =  x2 y2 1 is less than three.
x y3 1
 3 (Kanpur 2001; Bundelkhand 07)

Solution: The condition is necessary:

Given that the points ( xi, yi); i = 1, 2, 3 are collinear.

 x1 y1 1
 
We are to prove that the rank of matrix A =  x2 y2 1 is less than three.
x y3 1
 3

The given points are collinear

⇒ The area of the triangle formed by these points is zero.

x1 y1 1
1
⇒ x2 y2 1 =0
2
x3 y3 1

x1 y1 1
⇒ x2 y2 1 =0
x3 y3 1

x1 y1 1
⇒ The rank of matrix x2 y2 1 is less than 3.
x3 y3 1

Hence the condition is necessary.


V-117

The condition is sufficient:

x1 y1 1
The rank of the given matrix x2 y2 1 is less than 3.
x3 y3 1

We are to prove that the points ( xi, yi); i = 1, 2, 3 are collinear.

rank A < 3

x1 y1 1
⇒ x2 y2 1 =0
x3 y3 1

x1 y1 1
1
⇒ x2 y2 1 = 0.
2
x3 y3 1

The given points are collinear.

Hence the condition is sufficient.

Example 6: A is a non-zero column and B a non-zero row matrix, show that rank (AB) = 1 .

 a11 
 
 a21 
Solution : Let A =  a31  and B = [b11 b12 b13 … b1n]
 
…
 
am1

be two non-zero column and row matrices respectively.

 a11b11 a11b12 a11b13 … a11b1n 


 
 a21b11 a21b12 a21b13 … a21b1n
We have AB =  .
… … … … … 
 
am1b11 am1b12 am1b13 … am1b1n

Since A and B are non-zero matrices, therefore the matrix AB will also be non-zero. The
matrix AB will have at least one non-zero element obtained by multiplying
corresponding non-zero elements of A and B .

All the two-rowed minors of A obviously vanish. But A is a non-zero matrix. Hence
rank A = 1.
V-118

Comprehensive E xercise 1

1. Determine the rank of each of the following matrices :


1 0 0 0  1 1 1 1
   
0 1 0 0 1 1 1 1
(i)  (ii)
0 0 1 0 1 1 1 1
   
0 0 0 1 1 1 1 1
0 0  5 10  (Meerut 2006B)
(iii)   (iv)  
0 0  3 6

1 2 3 4 1 2 −7 5
   
(v) 2 4 6 8 (vi) 0 5 0 8
3 6 9 12 0 0 0 −3
 
1 2 3
(vii)  .
2 4 5

2. Show that the rank of a matrix is ≥ the rank of every sub-matrix thereof.
3. Show that the rank of a matrix does not alter on affixing any number of
additional rows or columns of zeros.
0 1 0 0
 
0 0 1 0
4. If A =  , find the rank of A and A2 .
0 0 0 1
 
0 0 0 0 

5. Under what conditions the rank of the following matrix is 3 ? Is it possible for the
2 4 2
 
rank to be 1? Why ? A = 3 1 4.
1 0 x

(Kanpur 2010)
6. A is an n-rowed square matrix of rank (n − 1) , show that Adj. A is not a null matrix.

A nswers 1

1. (i) 4 (ii) 1 (iii) 0 (iv) 1 (v) 1


(vi) 3 (vii) 2
V-119

4. Rank A = 3, rank A2 = 2
7
5. x≠ ; No, because one minor of order 2 of A is non-zero
5

5 Elementary Operations or Elementary Transformations


of a Matrix
An elementary transformation (or, an E-transformation) is an operation of any one of
the following types :

1. The interchange of any two rows(or columns).

2. The multiplication of the elements of any row (or column) by any non-zero number.

3. The addition to the elements of any other row (or column) the corresponding elements of any
other row (or column) multiplied by any number.

An elementary transformation is called a row transformation or a column


transformation according as it applies to rows or columns.

6 Symbols to be Employed for the Elementary


Transformations
The following notation will be used to denote the six elementary transformations :

1. The interchange of ith and j th rows will be denoted by Ri ↔ Rj.

2. The multiplication of the ith row by a non-zero number k will be denoted by


Ri → k Ri.

3. The addition of k times the j th row to the ith row will be denoted by Ri → Ri + kRj.

The corresponding column transformations will be denoted by writing C, in place of R


i. e., by Ci ↔ C j, Ci → k Ci, Ci → Ci + kC j respectively.

Important: It is quite obvious that if a matrix B is obtained from A by an elementary


transformation, A can also be obtained from B by an elementary transformation of the
same type.
For example, let
1 4 2 9
 
A = 2 5 1 3
3 7 8 43 × 4.

V-120

The elementary transformation R2 → R2 + 2 R3 transforms A into a matrix B,where

1 4 2 9
 
B = 8 19 17 11
3 7 8 4 3 × 4.

Now if we apply the elementary transformation R2 → R2 − 2 R3 to the matrix B, we


see that the matrix B transforms to the matrix A.

Again suppose we apply the elementary transformation R2 → 3 R2 to the matrix A.

Then A transforms into a matrix C, where

1 4 2 9
 
C = 6 15 3 9 .
3 7 8 4

1
Now if we apply the elementary transformation R2 → R2 to the matrix C , we see that
3
the matrix C transforms back to the matrix A.

7 Elementary Matrices
Definition: A matrix obtained from a unit matrix by a single elementary transformation is
called an elementary matrix (or E-matrix). For example,

0 0 1 1 0 0 1 2 0
     
0 1 0 , 0 4 0 , 0 1 0
1 0 0  0 0 1 0 0 1
  

are the elementary matrices obtained from I3 by subjecting it to the elementary


operations C1 ↔ C3 , R2 → 4 R2 , R1 → R1 + 2 R2 respectively.

It may be worthwhile to note that an E-matrix can be obtained from I by subjecting it


to a row transformation or a column transformation. We shall use the following
symbols to denote elementary matrices of different types :

(i) Eij will denote the E-matrix obtained by interchanging the ith and j th rows of a unit
matrix. The students can easily see that the matrices obtained by interchanging
the ith and j th rows or the ith and j th columns of a unit matrix are the same.
Therefore Eij will also denote the elementary matrix obtained by interchanging
the ith and j th columns of a unit matrix.
V-121

(ii) Ei (k ) will denote the E-matrix obtained by multiplying the ith row of a unit matrix
by a non-zero number k. It can be easily seen that the matrices obtained by
multiplying the ith row or the ith column of a unit matrix by k are the same.
Therefore Ei (k ) will also denote the elementary matrix obtained by multiplying
the ith column of a unit matrix by a non-zero number k.

(iii) Eij (m) will denote the elementary matrix, obtained by adding to the elements of
the ith row of a unit matrix, the products by any number m of the corresponding
elements of the j th row. It may be easily seen that the E-matrix, Eij (m) can also be
obtained by adding to the elements of the j th column of a unit matrix, the
products by m of the corresponding elements of the ith column.

It can be easily seen that

| Eij| = − 1, | Ei (k )| = k ≠ 0 , | Eij (m)| = 1.

Thus all the elementary matrices are non-singular.

Therefore each elementary matrix possesses inverse.

Now it is very interesting to note that the elementary transformations of a matrix can
also be obtained by algebraic operations on the same by the corresponding elementary
matrices. In this connection we have the following theorem.

8 Theorem
Every elementary row (column) transformation of a matrix can be obtained by pre-multiplication
(post-multiplication) with the corresponding elementary matrix.

We shall first prove that every elementary row transformation of a product AB of two
matrices A and B can be obtained by subjecting the pre-factor A to the same
elementary row transformation. Similarly every elementary column transformation
of a product AB of two matrices A and B can be obtained by subjecting the post-factor B
to the same elementary column transformation.

Let A = [aij] and B = [b jk ] be two m × n and n × p matrices respectively so that the


product AB is defined.

Let R1, R2 , R3 ,…, R m denote the row vectors of the matrix A and C1, C2 , …, Cp denote
the column vectors of the matrix B.
V-122

 R1 
 
 R2 
We can then write, A =  R3 , B = [C1 C2 C3 … Cp ].
 
…
 
R m 

 R1C1 R1C2 R1C3 … R1Cp 


 
 R2C1 R2C2 R2C3 … R2Cp 
∴ AB =  … … … … … 
 
 … … … … … 
 
R mC1 R mC2 R mC3 … R mCp 

Now if σ denotes any elementary row transformation, it is quite obvious from the
above representation that (σ A) B = σ (AB). For example, if σ denotes the elementary
row transformation R1 ↔ R2 , it is quite obvious that (σA) B = σ (AB).

Similarly it is quite obvious that if the columns C1, C2 ,…, Cp of B be subjected to any
elementary column transformation, the columns ofAB are also subjected to the same
elementary column transformation. Hence the result.

Now to prove our main theorem, if A be an m × n matrix, we can write

A = Im A.

If σ denotes any elementary row transformation, we have

σ A = σ (I m A) = (σ Im ) A = EA,

where E is the E-matrix corresponding to the same row transformation σ.

Similarly, we can write A = A In.

If σ denotes any elementary column transformation, we have

σ (A) = σ (AIn) = Aσ (In) = AE1,

where E1 is the E-matrix corresponding to the same column transformation σ.

1 4 2
 
Illustration : Let A = 2 7 1 .
3 8 4

V-123

The E-transformation R1 → R1 + 2 R3 transforms A into B, where

7 20 10 
 
B = 2 7 1 .
3 8 4 

Also if we apply the row transformation R1 → R1 + 2 R3 to the unit matrix

1 0 0
 
I3 = 0 1 0 ,
0 0 1

1 0 2
 
then the E-matrix E thus obtained is E = 0 1 0 .
0 0 1

1 0 2 1 4 2
   
Now EA = 0 1 0 2 7 1
0 0 1 3 8 4
 

 1⋅ 1 + 0 ⋅ 2 + 2 ⋅ 3 1⋅ 4 + 0 ⋅ 7 + 2 ⋅ 8 1 ⋅ 2 + 0 ⋅ 1 + 2 ⋅ 4
 
= 0 ⋅ 1 + 1 ⋅ 2 + 0 ⋅ 3 0 ⋅ 4 + 1⋅ 7 + 0 ⋅ 8 0 ⋅ 2 + 1 ⋅ 1 + 0 ⋅ 4
0 ⋅ 1 + 0 ⋅ 2 + 1 ⋅ 3 0 ⋅ 4 + 0 ⋅ 7 + 1⋅ 8 0 ⋅ 2 + 0 ⋅ 1 + 1 ⋅ 4

7 20 10 
 
= 2 7 1  = B.
3 8 4 

Similarly we can see that a column transformation of A can be effected by


post-multiplying A with the corresponding elementary matrix.

Non-singularity and Inverses of the Elementary Matrices


(i) The elementary matrix corresponding to the E-operation R1 ↔ Rj is its own inverse.

Let Eij denote the elementary matrix obtained by interchanging the ith and j th rows of a
unit matrix.

The interchange of the ith and j th rows of Eij will transform Eij to the unit matrix. But
every elementary row transformation of a matrix can be brought about by
pre-multiplication with the corresponding elementary matrix. Therefore the row
transformation which changes Eij to I can be effected on pre-multiplication by Eij.
V-124

Thus Eij Eij = I or (Eij)−1 = Eij.

Hence Eij is its own inverse.

Similarly, we can show that the elementary matrix corresponding to the E-operation Ci ↔ C j is
its own inverse.

(ii) The inverse of the E-matrix corresponding to the E-operation Ri → kRi, (k ≠ 0 ), is the
E-matrix corresponding to the E-operation Ri → k −1 Ri.

Let Ei (k ) denote the elementary matrix obtained by multiplying the elements of the ith
row of a unit matrix I by a non-zero number k.

The operation of the multiplication of the ith row of Ei (k ), by k −1 will transform E i (k )


to the unit matrix I. This row transformation of E i (k ) can be effected on
pre-multiplication by the corresponding elementary matrix E i (k −1).

Thus Ei (k −1) Ei (k ) = I or { Ei (k )} −1 = Ei (k −1).

Similarly, we can show that the inverse of the E-matrix corresponding to the E-operation
Ci → kCi, k ≠ 0 , is the E-matrix corresponding to the E-operation Ci → k −1 Ci.

(iii) The inverse of the E-matrix corresponding to the E-operation Ri → Ri + kRj is the E-matrix
corresponding to the E-operation Ri → Ri − kRj.

Let Eij (k ) denote the elementary matrix obtained by adding to the elements of the ith
row of a unit matrix I, the products by any number k of the corresponding elements of
the j th row of I.

If we add to elements of the ith row of Eij (k ), the products by −k of the corresponding
elements of its j th row, then this row operation will transform Eij(k ) to the unit matrix I.
Now this row transformation of Eij(k ) can be effected on pre-multiplication by the
corresponding elementary matrix Eij (− k ).

Therefore Eij (− k ) Eij (k ) = I or { Eij (k )} −1 = Eij (− k ).

Similarly, we can show that the inverse of the E-matrix corresponding to the E-operation
Ci → Ci + kC j is the E-matrix corresponding to the E-operation Ci → Ci − kC j.

From the above theorem, we thus conclude that the inverse of an elementary matrix
is also an elementary matrix of the same type.
V-125

9 Invariance of Rank Under Elementary Transformations


Theorem: Elementary transformations do not change the rank of a matrix.
(Lucknow 2005, 06)

Proof: Let A = [aij] be an m × n matrix of rank r. We shall prove the theorem in three
stages.

Case I. Interchange of a pair of rows does not change the rank.

Let B be the matrix obtained from the matrix A by the E-transformation Rp ↔ Rq . Let r
be the rank of A and s be the rank of B. Then to prove that r = s.

We have rank A = r ⇒ there exists at least one r-rowed square submatrix, say R, of A
such that | R| ≠ 0 .

Let S be the r-rowed square submatrix of B which has the same rows as are in R though
they may be in different relative positions.

Then either |S| = | R| or |S| = − | R|.

∴ | R | ≠ 0 ⇒ |S| ≠ 0 .

∴ rank B ≥ r ⇒ S ≥ r.

Again, as A can also be obtained from B by an interchange of rows, we have r ≥ s.

Hence r = s.

Case II. Multiplication of the elements of a row by a non-zero number does not change the rank.

Let B be the matrix obtained from the matrix A by the E-transformation


Rp → k Rp , (k ≠ 0 ), and let s be the rank of the matrix B .

Now if| B0| be any (r + 1) -rowed minor of B there exists a uniquely determined minor
| A0| of A such that

| B0| = | A0| ( this happens if the pth row of B is one of those rows which

are struck off to obtain B0 from B),

or | B0| = k | A0|

(this happens when pth row is retained while obtaining B0 from B).
V-126

Since the matrix A is of rank r, therefore every (r + 1)-rowed minor of A vanishes i. e.,
| A0| = 0 . Hence | B0 | = 0 . Thus we see that every (r + 1)-rowed minor of B also vanishes.
Therefore, s (the rank of B) cannot exceed r ( the rank of A).

∴ s ≤ r.

Also, since A can be obtained from B by E-transformation of the same type i. e.,
Rp → (1 / k ) Rp , therefore, by interchanging the roles of A and B we find that

r ≤ s.

Thus r = s.

Case III. Addition to the elements of a row the products by any number k of the corresponding
elements of any other row does not change the rank .

Let B be the matrix obtained from the matrix A by the E-transformation


Rp → Rp + kRq and let s be the rank of the matrix B.

Let B0 be any (r + 1)-rowed square sub-matrix of B and A0 be the correspondingly


placed sub-matrix of A.

The transformation Rp → Rp + kRq has changed only the pth row of the matrix A. Also
the value of a determinant does not change if we add to the elements of any row the
corresponding elements of any other row multiplied by some number k.

Therefore, if no row of the sub-matrix A0 is part of the pth row of A, or if two rows of A0
are parts of the pth and q th rows of A, then | B0| = | A0|.

Since the rank of A is r, therefore | A0| = 0 , and consequently | B0| = 0 .

Again, if a row of A0 is a part of the pth row of A, but no row is a part of the q th row, then
| B0| = | A0| + k |C0| , where C0 is an (r + 1) -rowed square matrix which can be obtained
from A0 by replacing the elements of A0 in the row which corresponds to the pth row of
A by the corresponding elements in the q th row of A. Obviously all the r + 1rows of the
matrix C0 are exactly the same as the rows of some (r + 1)-rowed square sub-matrix of A,
though arranged in some different order. Therefore|C0| is ± 1times some (r + 1)-rowed
minor of A. Since the rank of A is r, therefore, every (r + 1)-rowed minor of A is zero, so
that | A0| = 0 ,|C0| = 0 , and consequently | B0| = 0 .

Thus we see that every (r + 1)-rowed minor of B also vanishes. Hence, s (the rank of B)
cannot exceed r ( the rank of A).

∴ s ≤ r.
V-127

Also, since A can be obtained from B by an E-transformation of the same type i. e.,
Rp → Rp − kRq , therefore, interchanging the roles of A and B, we have r ≤ s.

Thus r = s.

We have thus shown that rank of a matrix remains unaltered by any E-row
transformation. Therefore we can also say that the rank of a matrix remains unaltered
by a series of elementary row transformations.

Similarly we can show that the rank of a matrix remains unaltered by a series of
elementary column transformations.

Finally, we conclude that the rank of a matrix remains unaltered by a finite chain of elementary
operations.

Corollary: We have already proved that every elementary row (column)


transformation of a matrix can be effected by pre-multiplication (post-multiplication)
with the corresponding elementary matrix. Combining this theorem with the theorem
just established, we conclude the following important results :

The pre-multiplication or post-multiplication by any elementary matrix, and as such by any


series of elementary matrices, does not alter the rank of a matrix.

10 Reduction to Normal Form


 Ir O
Theorem: Every m × n matrix of rank r can be reduced to the form   by a finite chain of
O O
E-operations where Ir , is the r-rowed unit matrix.

Proof: Let A = [aij] be an m × n matrix of rank r. If A is a zero matrix, then r is equal to


zero and we have nothing to prove. So let us take A as a non-zero matrix.

Since A is a non-zero matix, therefore A has at least one element different from zero,
say apq = k ≠ 0.

By interchanging the pth row with the first row and the q th column with the first
column respectively, we obtain a matrix B whose leading element is equal to k which is
not equal to zero.

Multiplying the elements of the first row of the matrix B by 1/k, we obtain a matrix C
whose leading element is equal to unity.
V-128

 1 c12 c13 … c1n 


 
 c21 c22 c23 … c2 n 
Let C=  .
… … … … …
 
c m1 c m2 c m3 … c mn

Subtracting suitable multiples of the first column of C from the remaining columns,
and suitable multiples of the first row from the remaining rows, we obtain a matrix D in
which all elements of the first row and first column except the leading element are
equal to zero.

1 0 0 0 … 0
 
0 
0 A1 
Let D=  ,
0 
 
… 
 
0 

where A1 is an (m − 1) × (n − 1) matrix.

If now, A1 be a non-zero matrix, we can deal with it as we did with A. If the elementary
operations applied to A1 for this purpose be applied to D, they will not affect the first
row and the first column of D . Continuing this process, we shall finally obtain a matrix
M, such that

 Ik O
M=  .
O O

The matrix M has the rank k. Since the matrix M has been obtained from the matrix A
by elementary transformations and elementary transformations do not alter the rank,
therefore we must have k = r.

 Ir O
Hence every m × n matrix of rank r can be reduced to the form   by a finite chain
O O
of elementary transformations.

Note: The above form is usually called the first canonical form or normal form of a
matrix.

Corollary 1: The rank of an m × n matrix A is r if and only if (iff) it can be reduced to the form
 Ir O
  by a finite chain of E-operations.
O O
V-129

The condition is necessary. The proof has been given in the above theorem.

The condition is also sufficient. The matrix A has been transformed into the form
 Ir O
  by elementary transformations which do not alter the rank of the matrix.
O O
 Ir O
Since the rank of the matrix   is r, therefore the rank of the matrix A must also
O O
be r.

Corollary 2: If A be an m × n matrix of rank r, there exist non-singular matrices P and Q such


that

 Ir O
PAQ =  .
O O

 Ir O
Proof: If A be an m × n matrix of rank r, it can be transformed into the form   by
O O
elementary operations. Since E-row (column) operations are equivalent to
pre-(post)-multiplication by the corresponding elementary matrices, we have the
following result :

If A be an m × n matrix of rank r, there exist E-matrices P1, P2 ,… Ps , Q1, Q2 , …, Qt such


that

 Ir O
Ps Ps − 1 … P1A Q1Q2 …Qt =  .
O O

Now each elementary matrix is non-singular and the product of non-singular matrices
is also non-singular. Therefore if P = Ps Ps − 1 … P2 P1 and Q = Q1 Q2 , …Qt , then P and Q
are non-singular matrices. Hence

 Ir O
PAQ =  .
O O

11 Equivalence of Matrices
Definition: If B be an m × n matrix obtained from an m × n matrix A by finite number of
elementary transformations of A then A is called equivalent to B. Symbolically, we write
A ~ B, which is read as ‘A is equivalent to B.’

The following three properties of the relation ‘~’ in the set of all m × n matrices are
quite obvious :
V-130

(i) Reflexivity: If A is any m × n matrix, then A ~ A. Obviously A can be obtained from


A by the elementary transformation Ri → kRi , where k = 1.

(ii) Symmetry: If A ~ B, then B ~ A. If B can be obtained from A by a finite number of


elementary transformations of A, then A can also be obtained from B by a finite
number of elementary transformations of B.

(iii) Transitivity: If A ~ B, B ~ C, then A ~ C.

If B can be obtained from A by a series of elementary transformations of A and C can be


obtained from B by a series of elementary transformations of B, then C can also be
obtained from A by a series of elementary transformations of A.

Therefore the relation ‘~’ in the set of all m × n matrices is an equivalence relation.

Example 7: If A and B be two equivalent matrices, then show that rank A = rank B.

Solution : If A ~ B, then B can be obtained from A by a finite number of elementary


transformations of A. Now the elementary transformations do not change the rank of
a matrix.

∴ If A ~ B, then rank A = rank B.

Example 8: Show that if two matrices A and B have the same size and the same rank, they are
equivalent.

Solution: Let A and B be two m × n matrices of the same rank r. Then by article 10, we
have

 Ir O  Ir O
A~  and also B ~  .
O O O O

By the symmetry of the equivalence relation,

 Ir O  Ir O
B~  implies   ~ B.
O O O O

Now by the transitivity of the equivalence relation

 Ir O  Ir O
A~  and   ~ B implies A ~ B.
O O O O
V-131

Example 9: (i) Use elementary transformations to reduce the following matrix A to triangular
form and hence find rank A :

5 3 14 4
 
A = 0 1 2 1.
1 −1 2 0 
 (Bundelkhand 2010)

 8 1 3 6
 
(ii) Find the rank of the matrix  0 3 2 2.
−8 −1 −3 4

Solution : (i) We have the matrix

1 −1 2 0
 
A ~ 0 1 2 1 by R1 ↔ R3
5 3 14 4

1 −1 2 0
 
~ 0 1 2 1 by R3 → R3 − 5 R1
0 8 4 4

1 −1 2 0
 
~ 0 1 2 1 by R3 → R3 − 8 R2 .
0 0 −12 −4

The last equivalent matrix is in Echelon form (or in triangular form). The number of
non-zero rows in this matrix is 3. Therefore its rank is 3. Hence rank A = 3.

(ii) Let us denote the given matrix by A. To find the rank of A, we shall reduce it to
1
Echelon form. Performing the column operation C1 → C1, we get
8
 1 1 3 6  1 1 3 6
   
A~ 0 3 2 2 ~ 0 3 2 2 by R3 → R3 + R1.
−1 −1 −3 4 0 0 0 10 
   

The last equivalent matrix is in Echelon form. The number of non-zero rows in this
matrix is 3. Therefore its rank is 3. Hence rank A = 3.

 1 2 1
 
Example 10: Is the matrix −1 0 2 equivalent to I3 ?
 2 1 −3
 (Meerut 2008)
V-132

 1 2 1
 
Solution: Let A = −1 0 2.
 2 1 −3

1 2 1 1 2 1
We have | A | = −1 0 2 = 0 2 3 R2 + R1, R3 − 2 R1
2 1 −3 0 −3 −5

= − 10 + 9 = − 1 i. e., ≠ 0.

Thus the matrix A is non-singular. Hence it is of rank 3. The rank of I3 is also 3. Since A
and I3 are matrices of the same size and the same rank, therefore A ~ I3 .

1 −1 2 −3
 
4 1 0 2  Ir O
Example 11: Reduce the matrix, A =  to the nor mal form   and
0 3 0 4 O O
 
 0 1 0 2
hence determine its rank. (Meerut 2001, 09B, 10)

Solution: We have the matrix A

1 0 0 0
 
4 5 −8 14
~ by C2 → C2 + C1, C3 → C3 − 2 C1, C4 → C4 + 3 C1
0 3 0 4
 
0 1 0 2

1 0 0 0
 
0 5 −8 14
~ by R2 → R2 − 4 R1
0 3 0 4
 
0 1 0 2

1 0 0 0
 
0 1 0 2
~ by R2 ↔ R4
0 3 0 4
 
0 5 −8 14

1 0 0 0
 
0 1 0 0
~ by C4 → C4 − 2 C2
0 3 0 −2
 
0 5 −8 4
V-133

1 0 0 0
 
0 1 0 0
~ by R3 → R3 − 3 R2 , R4 → R4 − 5 R2
0 0 0 −2
 
0 0 −8 4

1 0 0 0
 
0 1 0 0
~ by C3 ↔ C4
0 0 −2 0
 
0 0 4 −8

1 0 0 0
 
0 1 0 0 1 1
~ by C3 → − C3 , C4 → − C4
0 0 1 0 2 8
 
0 0 −2 1

1 0 0 0
 
0 1 0 0
~ by R4 → R4 + 2 R3
0 0 1 0
 
0 0 0 1

~ I4 . Hence the matrix A is of rank 4.

Example 12: Determine the rank of the following matrices :

2 −1 3 4
 
0 3 4 1
(i) 2 3 7 5
 
2 5 11 6 (Kanpur 2010)

−2 −1 −3 −1
 
 1 2 3 −1
(ii)  1 0 1 1
 
 0 1 1 −1 (Meerut, 2003, 09 B, 10B; Garhwal 14)

Solution : (i) Let us denote the given matrix by A. Performing the elementary
operations R3 → R3 − R1, R4 → R4 − R1, we see that
V-134

2 −1 3 4 2 −1 3 4
   
0 3 4 1 0 3 4 1
A~ ~ by R4 → R4 − 2 R2
0 4 4 1 0 4 4 1
   
0 6 8 2 0 0 0 0 

2 −1 3 4
 
0 12 16 4
~ by R2 → 4 R2 , R3 → 3 R3
0 12 12 3
 
0 0 0 0 

2 −1 3 4
 
0 12 16 4
~ by R3 → R3 − R2 .
0 0 −4 −1
 
0 0 0 0 

The last equivalent matrix is in Echelon form. The number of non-zero rows in this
matrix is 3. Therefore its rank is 3. Hence rank A = 3.

(ii) Let us denote the given matrix by A. Performing the elementary operation
R1 ↔ R2 , we see that

 1 2 3 −1
 
−2 −1 −3 −1
A~
1 0 1 1
 
 0 1 1 −1

1 2 3 −1
 
0 3 3 −3
~ by R2 → R2 + 2 R1, R3 → R3 − R1
0 −2 −2 2
 
0 1 1 −1

1 2 3 −1
 
0 1 1 −1 1 1
~  by R2 → R2 , R3 → − R3
0 1 1 −1 3 2
 
0 1 1 
−1

1 2 3 −1
 
0 1 1 −1
~ by R3 → R3 − R2 , R4 → R4 − R2 .
0 0 0 0
 
0 0 0 0 
V-135

The last equivalent matrix is in Echelon form. The number of non-zero rows in this
matrix is 2. Therefore rank A = 2.

2 −2 0 6
 
4 2 0 2
Example 13: Find the rank of the matrix A =  by reducing it to normal form.
1 −1 0 3
 
 1 −2 1 2
(Meerut 2008; Avadh 08; Rohilkhand 09)

1 1
Solution : Performing the operation R1 → R1, R2 → R2 , we see that
2 2

1 −1 0 3  1 −1 0 3
   
2 1 0 1 0 3 0 −5
A~ ~
1 −1 0 3 0 0 0 0
   
 1 −2 1 2 0 −1 1 −1

by R2 → R2 − 2 R1, R3 → R3 − R1, R4 → R4 − R1

1 0 0 0
 
0 3 0 −5
~ by C2 → C2 + C1, C4 → C4 − 3 C1
0 0 0 0
 
0 −1 1 −1

1 0 0 0
 
0 −1 1 −1
~ by R2 ↔ R4
0 0 0 0
 
0 3 0 −5

1 0 0 0
 
0 1 −1 1
~ by R2 → (−1) R2
0 0 0 0
 
0 3 0 −5

1 0 0 0
 
0 1 −1 1
~ by R4 → R4 − 3 R2
0 0 0 0
 
0 0 3 −8
V-136

1 0 0 0
 
0 1 0 0
~ by C3 → C3 + C2 , C4 → C4 − C2
0 0 0 0
 
0 0 3 −8

1 0 0 0
 
0 1 0 0
~ by R4 ↔ R3
0 0 3 −8
 
0 0 0 0 

1 0 0 0
 
0 1 0 0 1 1
~  by C3 → C3 , C4 → − C4
0 0 1 1 3 8
 
0 0 0 
0

1 0 0 0
 
0 1 0 0
~ by C4 → C4 − C3 ,
0 0 1 0
 
0 0 0 0 

I3 O
which is the normal form  . Hence rank A = 3.
O O

Example 14: Find two non-singular matrices P and Q such that PAQ is in the normal form where

1 1 1
 
A = 1 −1 −1 .
3 1 1

Also find the rank of the matrix A. (Garhwal 2006, 10; Kumaun 10)

Solution: We write A = I3 AI3 i. e.,

1 1 1  1 0 0  1 0 0
     
1 −1 −1 = 0 1 0  A 0 1 0 ⋅
3 1 1 0 0 1 0 0 1

Now we go on applying E-operations on the matrix A (the left hand member of the
above equation) until it is reduced to the normal form. Every E-row operation will also
be applied to the pre-factor I3 (or its transform) of the product on the right hand
member of the above equation and every E-column operation to the post-factor I3 (or
its transform).
V-137

Performing R2 → R2 − R1, R3 → R3 − 3 R1, we get

1 1 1  1 0 0  1 0 0
     
0 −2 −2 =  −1 1 0  A 0 1 0 .
0 −2 −2 −3 0 1 0 0 1

Performing C2 → C2 − C1, C3 → C3 − C1, we get

1 0 0  1 0 0  1 −1 −1
     
0 −2 −2 =  −1 1 0  A 0 1 0 .
0 −2 −2 −3 0 1 0 0 1

1
Performing R2 → − R2 , we get
2

1 0 0  1 0 0  1 −1 −1
   1 1   
0 1 1 =  − 0  A 0 1 0 .
0 2 2
 −2 −2 −3 0 1 0 0 1

Performing R3 → R3 + 2 R2 , we get

1 0 0  1 0 0  1 −1 −1
   1 1   
0 1 1 =  − 0  A 0 1 0 .
0 2 2
 0 0  −2 −1 1 0 0 1

Performing C3 → C3 − C2 , we get

1 0 0  1 0 0  1 −1 0
   1 1   
0 1 0 =  − 0  A 0 1 −1 .
0 2 2
 0 0  −2 −1 1 0 0 1

I2 O
∴ PAQ =  ,
O O

 1 0 0 1 −1 0
 1 1   
where P= − 0 , Q = 0 1 −1 .
−2 2
−1 1 0 1
 2  0

I2 O
Since A~ , therefore rank A = 2.
O O
V-138

Comprehensive Exercise 2

Determine the rank of the following matrices :


1 2 3 0  6 1 3 8
   
2 4 3 2  4 2 6 −1
1. 3 2. 
2 1 3 10 3 9 7
   
6 8 7 5 16 4 12 15
(Bundelkhand 2008;
Rohilkhand 08; Avadh 05; Lucknow 11)
2 3 −1 −1
 
1 −1 −2 −4
3. 3 1 3 −2
 
6 3 0 −7 (Avadh 2006, 09; Purvanchal 06)
1 a b 0
  1 3 4 3
0 c d 1  
4. 1  (Lucknow 2007) 5. 3 9 12 9
a b 0 1 3
 
 4 1
 0 c d 1 (Rohilkhand 2007)
1 2 1 2
 1 2 −1 4  
   1 3 2 2
6.  2 4 3 5 7. 
−1 2 4 3 4
 −2 6 −7  
3 7 4 6 (Lucknow 08)
3 −2 0 −1
  1 2 −4 5
0 2 2 1  
8. 1  9. 2 −1 3 6
−2 −3 2 8
 
 1 9 7
0 1 2 1

8 0 0 1
1 −1 3 6  
  1 0 8 1
10. 1 3 −3 −4 11. 
5 0 0 1 8
 3 3 11  
0 8 1 8 (Kumaun 2008)
0 1 −3 −1
 
1 0 1 1
12. 3 1 0 2
 
 1 1 −2 0  (Garhwal 2006, 13; Avadh 10)
V-139

1 1 1
1 −3 4 7  
13.   (Kumaun 2008) 14. 2 2 2
9 1 2 0 3
 3 3

2 1 3 4 5 6
   
15. 4 7 13 (Kumaun 2012) 16. 5 6 7
4 −3 −1 7 8 9
 
 1 0 2 1
  1 1 1
 0 1 −2 1  
17.  1 18.  a b c
−1 4 0 a3
 
 b3 c 3 
−2 2 8 0  (Meerut 2011; Kumaun 12)
19. With the help of elementary transformations find the rank of the following
matrix:
1 1 2 3
 
1 3 0 3
1 −2 −3 −3
 
1 1 2 3 (Rohilkhand 2010)
20. Reduce the following matrix to its Echelon form and find its rank :
 1 3 4 5
 
 3 9 12 9
−1 −3 −4 −3
 (Meerut 2004B; Rohilkhand 06, 10)
1 2 3
 
21. Find the rank of the matrix A = 2 3 4 after reducing it to normal form.
3 5 7
 (Garhwal 2015)
0 1 3 −1
 
1 0 1 1
22. Reduce the matrix  to normal form and find its rank.
3 1 0 2
 
 1 1 −2 0 

(Meerut 2009; Bundelkhand 11; Avadh 11)


9 7 3 6
 
23. (i) Reduce the matrix 5 −1 4 1 to normal form and find its rank.
6 8 2 4
 (Agra 2007)
(ii) Reduce the following matrix A into normal form and hence find its rank:
0 1 2 −2
 
A = 4 0 2 6 ⋅
2 1 3 1
 (Kanpur 2011)
V-140

24. Are the following pairs of matrices equivalent?


2 −1 3 4  1 0 −5 6
   
0 3 4 1 3 −2 1 2
(i)  ,
2 3 7 5 5 −2 −9 14
   
2 5 11 5 4 −2 −4 8

4 0 2 3 9 0 2
   
(ii) 3 1 0 , 7 −2 0 1
5 0 0  8 1 1 5
 
25. Find the ranks of A, B, A + B, AB and BA where
1 1 −1 −1 −2 −1
   
A = 2 −3 4, B =  6 12 6 .
3 −2 3  5 10 5
 
26. Show that if A and B are equivalent matrices, then there exist non-singular
matrices P and Q such that B = PAQ.
27. Show that the rank of a matrix is not altered if a column of it is multiplied by a
non-zero scalar.
28. (i) What is the rank of a non-singular matrix of order n?
(ii) What is the rank of an elementary matrix?

A nswers 2
1. 3 2. 2 3. 3 4. 2 5. 2 6. 2 7. 3 8. 4 9. 3
10. 3 11. 4 12. 2 13. 2 14. 1 15. 2 16. 2 17. 3
18. rank (A ) = 3 if a ≠ b ≠ c and a + b + c ≠ 0; rank (A ) = 2 if a ≠ b ≠ c and
a + b + c = 0; Also rank (A ) = 2 if a = b ≠ c ; and rank (A ) = 1 if a = b = c
19. 3 20. 2 21. 2 22. 3 23. (i) 3 (ii) 2
24. (i) No, since the rank of the first matrix is 4 and that of the second matrix is 2.
(ii) No, since the matrices are not of the same type
25. Rank A = 2; Rank B = 1; Rank A + B = 2;
Rank AB = 0; Rank BA = 1
28. (i) n. (ii) Equal to the order of the matrix
V-141

12 Row and Column Equivalence of Matrices


Definition: A matrix A is said to be row equivalent to B if B is obtainable from A by a finite
R
number of E-row transformations of A. Symbolically, we then write A ~ B. Similarly a matrix A
is said to be column equivalent to B if B is obtainable from A by a finite number of E-column
C
transformations of A. Symbolically, we then write A ~ B.

13 Employment of Only Row Transformations


Theorem: If A be an m × n matrix of rank r, then there exists a non-singular matrix P such
G
that PA =  , where G is an r × n matrix of rank r and O is (m − r) × n.
O

Proof: Since A is an m × n matrix of rank r, therefore there exist non-singular


matrices P and Q, such that

 Ir O
PAQ =  . …(1)
O O

Now every non-singular matrix can be expressed as the product of elementary


matrices. So let

Q = Q1 Q2 ...... Qt where Q1, Q2 , ........, Qt are elementary matrices.

Thus the relation (1) can be written as

 Ir O
PAQ1 Q2 ..... Qt =  ⋅ …(2)
O O

Now everyE-column transformation of a matrix, is equivalent to post-multiplication


with the corresponding elementary matrix. Since no column transformation can affect
the last (m − r) rows of the right hand side of (2), therefore post-multiplying the L. H.S.
of (2) by the elementary matrices Q−t 1, Q−t 1− 1, .... , Q2−1, Q1−1 successively and affecting
the corresponding column transformations in the right hand side of (2), we get a
relation of the form

G
PA =   ⋅
O
V-142

Since elementary transformations do not alter the rank, therefore the rank of the
G
matrix PA is the same as that of the matrix A which is r. Thus the rank of the matrix  
O
is r and therefore the rank of the matrix G is also r as the matrix G has r rows and last
G
m − r rows of the matrix   consist of zeros only.
O

14 Employment of Only Column Transformations


Theorem: If A be an m × n matrix of rank r, then there exists a non-singular matrix Q such
that AQ = [H O] , where H is an m × n matrix of rank r and O is m × (n − r).

Proof: Since A is an m × n matrix of rank r, therefore there exist non-singular


matrices P and Q such that

 Ir O
PAQ =  ⋅ …(1)
O O

Now every non-singular matrix can be expressed as the product of elementary


matrices. So let

P = P1 P2 ...... Ps where P1, P2 , ....., Ps are elementary matrices.

Thus the relation (1) can be written as

 Ir O
P1 P2 .... Ps AQ =  . …(2)
O O

Now every E-row transformation of a matrix is equivalent to pre-multiplication with


the corresponding elementary matrix. Again no row transformation can affect the last
 Ir O
n − r columns of  ⋅
O O
Therefore pre-multiplying the L.H.S. of (2) by the elementary matrices.

P1−1, P2−1, ..... Ps−1


successively and affecting the corresponding row transformations in the R.H.S. of (2),
we get a relation of the form AQ = [H O].

Now elementary transformations do not alter the rank. Therefore the rank of the
matrix AQ is the same as that of A which is r. Thus the rank of the matrix [H O] is r and
therefore the rank of the matrix H is also r as the matrix H has r columns and the last
n − r columns of the matrix [H O] consists of zero only.
V-143

15 The Rank of a Product


Theorem: The rank of a product of two matrices cannot exceed the rank of either matrix.
(Lucknow 2009; Kumaun 11)

Proof: Let A and B be two m × n and n × p matrices respectively. Let r1, r2 be the ranks
of A and B respectively and let r be the rank of the product AB.

To prove r ≤ r1 and r ≤ r2 .

Since A is an m × n matrix of rank r1, therefore there exists a non-singular matrix P such
G
that PA =  , where G is an r1 × n matrix of rank r1 and O is (m − r1) × n.
O

G
∴ PAB =   B.
O

Since the rank of a matrix does not alter by multiplying it with a non-singular matrix,
therefore

Rank (PAB) = Rank (AB) = r.

G
∴ Rank of the matrix   B = r.
O

G
Since the matrix G has only r1 non-zero rows, therefore the matrix   B cannot have
O
more than r1 non-zero rows which arise by multiplying the r1 non-zero rows of G with
the columns of B.

G
∴ Rank of the matrix   B is ≤ r1 i. e., t ≤ r1
O

i. e., Rank (AB) ≤ Rank of the prefactor A .

Again r = Rank (AB) = Rank (AB)′ = Rank (B ′ A ′ ) ≤ Rank of the prefactor B′

= Rank B [∵ Rank B = Rank B′]

= r2 .

∴ r ≤ r2 i. e., Rank (AB) ≤ Rank of the post-factor B .


V-144

16 Theorem
Every non-singular matrix is row equivalent to a unit matrix.

Proof: We shall prove the theorem by induction on n, the order of the matrix. If the
matrix be of order 1 i. e., if A =| a11|, the theorem obviously holds.

Let us assume that the theorem holds for all non-singular matrices of order n − 1.

Let A = [aij] be an n × n non-singular matrix. The first column of the matrix A has at
least one element different from zero, for otherwise we shall have |A | = 0 and the
matrix A will not be non-singular.

Let ap1 = k ≠ 0 .

By interchanging the p th row with the first row (if necessary), we obtain a matrix B
whose leading element is equal to k which is not equal to zero.

Multiplying the elements of the first row of the matrix B by 1 / k, we obtain a matrix C
whose leading element is equal to unity.

1 c12 c13 .... c1n 


 
c21 c22 c23 .... c2 n
Let C = c31 c32 c33 .... c3 n ⋅
 
.... .... .... .... .... 
 
c n1 c n2 c n3 .... c nn

Subtracting suitable multiples of the first row of C from the remaining rows, we obtain
a matrix D in which all elements of the first column except the leading element are
equal to zero.

1 d12 d13 ... d1n


 
0 
Let D = 0 A1 .
 
... 
 
0 

where A1 is an (n − 1) × (n − 1) matrix. The matrix A1 is non-singular, for otherwise


| A1| = 0 and so |D | is also equal to zero. Thus the matrix D will be non-singular, and
therefore A, which is row equivalent to D, will also not be non-singular.
V-145

By the inductive hypothesis, A1 can be transformed to In −1 by E-row operations. If


these elementary row operations be applied to D, they will not affect the first row and
the first column of D and we shall obtain a matrix M such that

1 d12 d13 .... d1n


 
0 1 0 .... 0 
M = 0 0 1 .... 0 ⋅
 
.... .... .... .... .... 
 
0 0 0 .... 1 

By adding suitable multiples of the second, third, …, n th rows to the first row of M, we
obtain the matrix In.
Thus the matrix A has been reduced to In by E-row operations only.
The proof is now complete by induction.

Corollary 1: If A be an n-rowed non-singular matrix, there exist E-matrices E1, E2 , ...., Et such
that

Et Et − 1.... E2 E1A = In.

If A be an n-rowed non-singular matrix, it can be reduced to In by E-row operations


only. Since every E-row operation is equivalent to pre-multiplication by the
corresponding E-matrix, therefore we can say that if A be an n -rowed non-singular
matrix, there exist E-matrices E1, E2 , ...., Et such that

Et Et − 1 ..... E2 E1A = In.

Corollary 2: Every non-singular matrix A is expressible as the product of elementary


matrices.
If A be an n-rowed non-singular matrix, there exist E-matrices E1, E2 , ...., Et such that

Et Et − 1 ..... E2 E1A = In. …(1)

Pre-multiplying both sides of the relation (1) by (Et Et − 1 ...... E2 E1)−1, we get

(Et Et − 1..... E2 E1)−1 (Et Et − 1...... E2 E1)A = (Et Et − 1.... .. E2 E1)−1 In

or InA = E1−1E2−1E3−1..... Et − 1−1E−t 1In

−1
or A = E1−1E2−1 ..... Et − 1E−t 1.

Since the inverse of an elementary matrix is also an elementary matrix of the same type
hence we get the result.
V-146

Corollary 3: The rank of a matrix does not alter by pre-multiplication or post-multiplication


with a non-singular matrix.

Every non-singular matrix can be expressed as the product of elementary matrices.


Also E-row (column) transformations are equivalent to pre-(post)-multiplication with
the corresponding elementary matrices and elementary transformations do not alter
the rank of a matrix. Hence we get the result.

17 Use of Elementary Transformations to Find the Inverse of


a Non-Singular Matrix
Let A be a non-singular matrix of order n. It can be easily shown that A can be reduced
to the unit matrix In by a finite number of E-row transformations only. Now each
E-row transformation of a matrix is equivalent to pre-multiplication by the
corresponding E-matrix. Therefore there exist elementary matrices, say, E1, E2 , ....., Et
such that (Et Et − t ...... E2 E1) A = In.

Post-multiplying both sides by A −1, we get

(Et Et − 1...... E2 E1)AA −1 = InA −1

or (Et Et − 1...... E2 E1)In = A −1 [∵ AA −1 = In, InA −1 = A −1]

or A −1 = (Et Et − 1...... E2 E1)In.

Hence we get the following result :

If a non-singular matrix A of order n is reduced to the unit matrix In by a sequence of E-row


transformations only, then the same sequence of E-row transformations applied to the unit matrix
In gives the inverse of A (i. e., A −1).

18 Working Rule for Finding the Inverse of a Non-Singular


Matrix by E-Row Transformations
Suppose A is a non-singular matrix of order n. Then we write A = InA.

Now we go on applying E-row transformations only to the matrix A and the pre-factor
In of the product InA till we reach the result In = BA.

Then obviously B is the inverse of A.


V-147

1 2 1
 
Example 15: Find the inverse of the matrix A = 3 2 3 by using E-transformations.
1 1 2

(Avadh 2006, 08; Purvanchal 09; Rohilkhand 09; Lucknow 09; Garhwal 10, 12)

1 2 1 1 0 0
   
Solution : We write A = I3 A, i. e., 3 2 3 = 0 1 0  A.
1 1 2 0 0 1

Now we go on applying E-row transformations to the matrix A (the left hand member
of the above equation) until it is reduced to the form I3 . Every E-row transformation
will also be applied to the prefactor I3 (or its transform) of the product on the right
hand side of the above equation.

Applying R2 → R2 − 3 R1, R3 → R3 − R1, we get

1 2 1  1 0 0
   
0 −4 0  = −3 1 0  A.
0 −1 1  −1 0 1

Now we should try to make 1 in the place of the second element of the second row of
1
the matrix on the left hand side. So applying R2 → − R2 , we get
4

1 2 1  1 0 0
  3 1 
0 1 0 =  − 0  A.
0 4 4
 −1 1 −1 0 1

Now we shall make zeros in the place of the second elements of the first and third rows
with the help of the second row. So applying R1 → R1 − 2 R2 , R3 → R3 + R2 , we get

 1 1 
− 0
1 0 1  2 2
   3 1 
0 1 0 =  − 0  A.
0  4 4 
 0 1  1 1
− − 1
 4 4 
V-148

Now the third element of the third row is already 1. So to make the third element of
the first row zero, we apply R1 → R1 − R3 , and we get

 1 3 
− −1
1 0 1  3 4
   3 1 
0 1 0 =  − 0  A.
0  4 4 
 0 1  1 1
− − 1
 4 4 

 1 3 
− 4 4
−1
 3 1 
Thus I3 = BA, where B =  − 0 .
 4 4 
− 1 −
1
1
 4 4 

 1 3 
− 4 4
−1
 3 1 
∴ A −1 =B= − 0 ⋅
 4 4 
− 1 −
1
1
 4 4 

0 1 2 2
 
1 1 2 3
Example 16: Find the inverse of the matrix A =  by using E-transformations.
2 2 2 3
 
2 3 3 3

0 1 2 2 1 0 0 0
   
1 1 2 3 0 1 0 0
Solution : We write 2 = A.
2 2 3 0 0 1 0
   
2 3 3 3 0 0 0 1

Performing R1 ↔ R2 , we get

1 1 2 3 0 1 0 0
   
0 1 2 2 1 0 0 0
2 = A.
2 2 3 0 0 1 0
   
2 3 3 3 0 0 0 1
V-149

Performing R3 → R3 − 2 R1, R4 → R4 − 2 R1, we get

1 1 2 3  0 1 0 0
   
0 1 2 2  1 0 0 0
0 0 = A.
−2 −3 0 −2 1 0
   
0 1 −1 −3 0 −2 0 1
Performing R4 → R4 − R2 , R1 → R1 − R2 , we get

1 0 0 1  − 1 1 0 0
   
0 1 2 2  1 0 0 0
0 = A.
0 −2 −3  0 −2 1 0
   
0 1 −3 −5 −1 −2 0 1

Performing R4 → R4 − R2 , R1 → R1 − R2 , we get

1 0 0 1  − 1 1 0 0
   
0 1 2 2  1 0 0 0
0 = A.
0 −2 −3  0 −2 1 0
   
0 0 −3 −5 −1 −2 0 1

1
Performing R3 → − R3 , we get
2

1 0 0 1  − 1 1 0 0
   
0 1 2 2  1 0 0 0
0 3 =  1 A.
0 1 0 1 − 0
 2  2 
0 0 −3 −5 −1 −2 0 1

Performing R2 → R2 − 2 R3 , R4 → R4 + 3 R3 , we get

1 0 0 1 −1 1 0 0
   
0 1 0 −1  1 −2 1 0
0 3 =  1
0 1 0 1 − 0  A.
 2  2 
 1  3 
0 0 0 −  −1 1 − 1
2  2 
Performing R4 → − 2 R4 , we get

1 0 0 1 −1 1 0 0
   
0 1 0 −1  1 −2 1 0
0 3 =  1 A.
0 1 0 1 − 0
 2  2 
0 0 0 1  2 −2 3 −2
V-150

3
Performing R3 → R3 − R4 , R2 → R2 + R4 , R1 → R1 − R4 we get
2

1 0 0 0  −3 3 −3 2
   
0 1 0 0  3 −4 4 −2
0 = A.
0 1 0  −3 4 −5 3
   
0 0 0 1  2 −2 3 −2

−3 3 −3 2
 
 3 −4 4 −2
∴ A −1 = ⋅
−3 4 −5 3
 
 2 −2 3 −2

Comprehensive Exercise 3

 1 2 1
 
1. Reduce the matrix A = −1 0 2 to I3 by E-row transformations only.
 2 1 3

2. Compute the inverse of the following matrices by using elementary


row-transformations :
1 2 3  i −1 2 i
   
(i) 2 4 5 (ii)  2 0 2
3 5 6 −1 0 1
 

−1 −3 3 −1
  1 3 3
 1 1 −1 0  
(iii)  (iv) 1 4 3
2 −5 2 −3 1
 
 3 4
−1 1 0 1
(Avadh 2007; Garhwal 15)

0 1 2
 
(v) 1 2 3
3 1 1
 (Rohilkhand 2011; Bundelkhand 08)
V-151

A nswers 3

 1 1
0 − 
 1 −3 2 4 2
1    3i i
2. (i) −3 3 −1 (ii) −1 
157   4 2
−2 −1 0  0 1 1
 4 2 

0 2 1 3
   7 −3 −3
 1 1 −1 −2  
(iii)  (iv) −1 1 0
1 2 0 1 −1
 
 0 1
−1 1 2 6

 1 1 1
 2 −
2 2
(v) −4 3 −1
 5 3 1
 − 
 2 2 2

Objective Type Questions

Multiple Choice Questions


Indicate the correct answer for each question by writing the corresponding letter from (a),
(b) , (c) and (d).

1. If A is a matrix such that there exists a square submatrix of order r which is


non-singular and every square submatrix of order r + 1or more is singular, then

(a) rank A = r + 1

(b) rank A = r

(c) rank A > r

(d) rank A ≥ r + 1
V-152

2. Let A = [aij]m × n be a matrix such that aij = 1 for all i, j. Then

(a) rank A > 1

(b) rankA = 1
(c) rank A = m
(d) rank A = n

1 2 3 0
 
2 4 3 2
3. The rank of the matrix  is
3 2 1 3
 
6 8 7 5

(a) 1 (b) 2

(c) 3 (d) 4

4. If A and B are two matrices such that rank of A = m and rank of B = n, then

(a) rank (A B) = m n

(b) rank (A B) ≥ rank A

(c) rank (A B) ≥ rank B

(d) rank (A B) ≤ min (rank A, rank B) (Agra 2007)


5. If A is an invertible matrix and B is a matrix, then

(a) rank (A B) = rank A

(b) rank (A B) = rank B

(c) rank (A B) > rank A

(d) rank (A B) > rank B


1 2 0 0
 
6. The rank of the matrix A = 2 3 0 0  is
0 0 0 0 

(a) 0 (b) 1

(c) 2 (d) 3 (Bundelkhand 2001; Meerut 03)


V-153

Fill in the Blank(s)


Fill in the blanks “……”, so that the following statements are complete and correct.
1. The rank of a matrix is the …… of any highest order non-vanishing minor of the
matrix.
2. The rank of every non-singular matrix of order n is ……………
(Bundelkhand 2005)
3. The rank of a matrix in Echelon form is equal to the number of non-zero ……… of
the matrix.
(Bundelkhand 2008)
4. A is a non-zero column matrix and B is a non-zero row matrix, then rank
(AB) = ..... .

5. If a matrix B is obtained from A by an elementary transformation, A can also be


obtained from B by an elementary transformation of the …… type.

6. A matrix obtained from a …… by a single elementary transformation is called an


elementary matrix.

7. If a matrix A of order m × n can be expressed as :


 Ir O
A~ , then rank of A is ………
O O (Meerut 2001)

8. If A be an m × n matrix of rank r, there exist non-singular matrices P and Q such


that PAQ = ..... .

9. If B be an m × n matrix obtained from an m × n matrix A by finite number of


elementary transformations of A, then A is called ……… to be B.

10. If two matrices A and B have the same size and the same rank, they are ……… .
1 0
11. The rank of a matrix   is ………
0 1 (Agra 2008)

True or False
Write ‘T’ for true and ‘F’ for false statement.
1. The rank of the transpose of a matrix is the same as that of the original matrix.

2. Every elementary row transformation of a matrix can be obtained by


post-multiplication with the corresponding elementary matrix.

3. Elementary transformations change the rank of a matrix.


V-154

4. If A and B are two equivalent matrices, then rank A = rank B.

5. The rank of a matrix does not alter by pre-multiplication or post-multiplication


with a non-singular matrix.

A nswers

Multiple Choice Questions


1. (b) 2. (b) 3. (c)
4. (d) 5. (b) 6. (c)

Fill in the Blank(s)


1. order 2. n 3. rows
4. 1 5. same 6. unit matrix
 Ir O
7. r 8.   9. equivalent
O O
10. equivalent 11. 2

True or False
1. T 2. F 3. F
4. T 5. T

o
V-155

7
A pplications of M atrices

1 Vectors
efinition: Any ordered n-tuple of numbers is called an n-vector. By an ordered n-tuple
D we mean a set consisting of n numbers in which the place of each number is fixed.
If x1, x2 , ...., xn be any n numbers, then the ordered n -tuple X = ( x1, x2 , ..., xn) is called an
n-vector. The ordered triad ( x1, x2 , x3 ) is called a 3-vector. Similarly (1, 0 , 1, − 1) and
(1, 8, − 5, 7) are 4-vectors. The n numbers x1, x2 , ..., xn are called components of the
n-vector X = ( x1, x2 , ..., xn). A vector may be written either as a row vector or as a column
vector. If A be a matrix of the type m × n, then each row of A will be an n-vector and each
column of A will be an m-vector. A vector whose components are all zero is called a zero
vector and will be denoted by O.

If k be any number and X be any vector, then relative to the vector X, k is called a scalar.

Algebra of vectors. Since an n-vector is nothing but a row matrix or a column


matrix, therefore we can develop an algebra of vectors in the same manner as the
algebra of matrices.
V-156

Equality of two vectors. Two n-vectors X and Y where X = ( x1, x2 ,..., xn) and
Y = ( y1, y2 , ..., yn) are said to be equal if and only if their corresponding components
are equal i. e., if xi = yi, for all i = 1, 2,.., n.

For example if

X = (1, 4, 7) and Y = (1, 4, 7),

then X = Y.

But if X = (1, 4, 7) and Y = (4, 1, 7), then X ≠ Y.

Addition of two vectors. If X = ( x1, x2 , …, xn) and Y = ( y1, y2 , …, yn) then by


definition X + Y = ( x1 + y1, x2 + y2 , …, xn + yn).

Thus X + Y is an n-vector whose components are the sums of corresponding


components of X and Y.

If X = (2, 4, − 7) and Y = (1, − 3, 5),

then X + Y = (2 + 1, 4 − 3, − 7 + 5) = (3, 1, − 2).

Multiplication of a vector by a scalar (number).

If k be any number and

X = ( x1, x2 , …, xn),

then by definition

kX = (kx1, kx2 , ..., kxn).

The vector k X is called the scalar multiple of the vector X by the scalar k.

If X = (1, 3, 8), then 4 X = (4, 12, 32)

and 0 X = (0 , 0 , 0 ).

Properties of addition and scalar multiplication of vectors. If X, Y, Z be any


three n-vectors and p, q be any two numbers, then obviously

(i) X + Y = Y + X. (ii) X + (Y + Z) = (X + Y) + Z.

(iii) p (X + Y) = pX + pY (iv) ( p + q)X = pX + qX.

(v) p (qX) = ( pq) X.


V-157

2 Linear Dependence and Linear Independence of Vectors


(Agra 2005)

Linearly dependent set of vectors. Definition:

A set of r n-vectors X1, X2 , …, X r is said to be linearly dependent if there exist r scalars (numbers)
k1, k2 , ..., k r , not all zero, such that

k1X1 + k2 X2 + .. + k r X r = O,

where, O denotes the n-vector whose components are all zero.

Linearly independent set of vectors. Definition:

A set of r n-vectors X1, X2 , ..., X r is said to be linearly independent if every relation of the type

k1X1 + k2 X2 + … + k r X r = O

implies k1 = k2 = k3 … = k r = 0 .

Example 1: Show that the vectors X1 = (1, 2, 4), X2 = (3, 6, 12) are linearly dependent.

Solution: By a little inspection, we see that

3 X1 + (− 1) X 2 = (3, 6, 12) + (− 3, − 6, − 12) = (0 , 0 , 0 ) = O.

Thus there exist numbers k1 = 3, k2 = − 1 which are not all zero such that

k1X1 + k2 X2 = O.

Hence the vectors X1 and X2 are linearly dependent.

Example 2: Show that the set consisting only of the zero vector, O, is linearly dependent.

Solution : Let X = (0 , 0 , 0 , ...., 0 ) be an n-vector whose components are all zero. Then
the relation kX = O is true for some non-zero value of the number k. For example,1X = O
and 1 ≠ 0.

Hence the vector O is linearly dependent.

Example 3: Show that the vectors X1 = (1, 2, 3) and X 2 = (4, − 2, 7) are linearly independent.

Solution: Let k1 and k2 be two numbers such that k1X1 + k2 X2 = O,

i. e., k1 (1, 2, 3) + k2 (4, − 2, 7) = (0 , 0 , 0 )


V-158

i.e., (k1 + 4 k2 , 2 k1 − 2 k2 , 3 k1 + 7 k2 ) = (0 , 0 , 0 ).
Equating the corresponding components, we get

k1 + 4 k2 = 0 , 2 k1 − 2 k2 = 0 , 3 k1 + 7 k2 = 0 .

The only common values of k1 and k2 which satisfy these equations are k1 = 0 , k2 = 0 .

Thus k1X1 + k2 X2 = O, iff k1 = k2 = 0.

Hence the vectors X1 and X2 are linearly independent.

Example 4: Show that the set of three 3-vectors

X1 = (1, 0 , 0 ), X2 = (0 , 1, 0 ), X3 = (0 , 0 , 1)

is linearly independent.
Solution : Let k1, k2 , k3 be three numbers such that

k1X1 + k2 X2 + k3 X3 = O,

i. e., k1(1, 0 , 0 ) + k2 (0 , 1, 0 ) + k3 (0 , 0 , 1) = (0 , 0 , 0 ),

i. e., (k1, 0 , 0 ) + (0 , k2 , 0 ) + (0 , 0 , k3 ) = (0 , 0 , 0 ),

i. e., (k1, k2 , k3 ) = (0 , 0 , 0 ).

Obviously this relation is true if and only if k1 = 0 , k2 = 0 , k3 = 0.

Hence the vectors X1, X2 , X3 are linearly independent.

3 A Vector as a Linear Combination of Vectors


Definition: A vector X which can be expressed in the form

X = k1X1 + k2 X2 + … + k r X r ,

is said to be a linear combination of the set of vectors X1, X2 , ..., X r .

Here k1, k2 , …, k r are any numbers.

The following two results are quite obvious :

(i) If a set of vectors is linearly dependent, then at least one member of the set can be expressed as
a linear combination of the remaining members.

(ii) If a set of vectors is linearly independent then no member of the set can be expressed as a linear
combination of the remaining members.
V-159

4 The n-Vector Space


The set of all n vectors of a field F is called the n-vector space over F. It is usually
denoted by Vn( F ) or simply by Vn if the field is understood. Similarly the set of all
3-vectors is a vector space which is usually denoted by V3 . The elements of the field F
are known as scalars relatively to the vectors.

5 Sub-space of an n-Vector SpaceVn


Definition: A non-empty set, S, of vectors of Vn is called a vector sub-space of Vn, if
a + b belongs to S whenever a, b belong to S and k a belongs to S,where k is any scalar.

It is important to note that every sub-space of Vn contains the zero vector, being the
scalar product of any vector with the scalar zero.

Illustration: If a = (a1, a2 , a3 ) is any non-zero vector of V3 , then the set S of vectors ka


is a subspace of V3 , where k is a variable scalar which can take any value. The sum of any
two members k1a, k2 a of S, is the vector k1a + k2 a

i. e., (k1a1, k1a2 , k1a3 ) + (k2 a1, k2 a2 , k2 a3 )

i. e., (k1 + k2 ) a

which is also a member of S. Also the scalar multiple by any scalar x of any vector k1a of
S is the vector ( xk1)a which is again a member of S.

Hence the set S of vectors ka is a subspace of V3 .

6 Vector Subspace Spanned by a given System of Vectors


Let a, b, c be any three vectors of V3 . The set S of all vectors of the form xa + yb + zc,
where x, y, z are any scalars, is a subspace of V3 . For, if x1a + y1 b + z1c, x2 a + y2 b + z2 c
be any two members of S, then

( x1a + y1 b + z1c) + ( x2 a + y2 b + z2 c) = ( x1 + x2 )a + ( y1 + y2 ) b + (z1 + z2 )c,

which is also a member of S. Also if k be any scalar, then

k ( x1a + y1 b + z1c) = (kx1)a + (ky1) b + (kz1)c,

which is again a member of S. Thus S is a vector subspace and we say that S is a spanned
by the vectors a, b and c. More generally, if a1, a2 , ..., a r be a set of r fixed vectors of Vn,
then the set S of all n-vectors of the form p1a1 + p2 a2 + … pr a r where p1, p2 , ...., pr are
any scalars is a vector subspace of Vn.
V-160

This vector space is said to be spanned by the vectors a 1, a 2, . . . . , a r . Thus a vector


space which arises as a set of all linear combinations of any given set of vectors,
is said to be spanned by the given set of vectors.

7 Basis and Dimension of a Subspace


A set of vectors a1, a2 , a3 ,..., a k belonging to the subspace S is said to be a basis of S,if

(i) the subspace S is spanned by the set a1, a2 , ..., a k and

(ii) the vectors a1, a2 , ..., a k are linearly independent.

It can be easily shown that every subspace, S, of Vn possesses a basis.

It can be easily shown that the vectors

e1 = (1, 0 , 0 , ..., 0 ), e2 = (0 , 1, 0 , ..., 0 ), e3 = (0 , 0 , 1, 0 , ..., 0), ...,

en = (0 , 0 , 0 , ..., 1)

constitute a basis of Vn.

We have already shown that these vectors are linearly independent. Moreover any
vector a = (a1, a2 , ..., an) of Vn is expressible as a = a1e1 + a2 e2 + a3 e3 + … + anen. Hence
the vectors e1, e2 , e3 , ..., en constitute a basis of Vn.

Theorem: A basis of a subspace, S, can always be selected out of a set of vectors which span S.

Proof: Let a1, a2 , ..., a r be a set of vectors which spans a subspace S. If these vectors are
linearly independent, they already constitute a basis of S as they span S. In case they
are linearly dependent, some member of the set is a linear combination of the
members. Deleting this member we obtain another set which also spans.

Continuing in this manner we shall ultimately, in a finite number to steps arrive at a


basis of S.

A vector subspace may (and in fact does) possess several bases

For example, if

a1 = (1, 0 , 0 ), a2 = (0 , 1, 0 ), a3 = (0 , 0 , 1), b1 = (1, 1, 1), b2 = (1, 1, 0 ), b3 = (1, 0 , 0 ),

then a1 , a2 , a3 constitute a basis of V3 . Also b1, b2 , b3 constitute a basis of V3 .

But it is important to note that the number of members in any one basis of a subspace is the
same as in any other basis. This number is called the dimension of the subspace.
V-161

We have already shown that one basis of Vn possesses n members. Therefore every
basis of Vn must possess n members. Thus Vn is of dimension n. In a particular the
dimension of V3 is 3.

Also it can be easily shown that if r be the dimension of a subspace S and if a1, a2 ,… a k
be a linearly independent set of vectors belonging to S, then we can always find vectors
a k + 1, a k + 2 , .... a r such that the vectors a1, a2 , . . . , a k , a k + 1,..., a r constitute a basis of S.
In other words we can say that every linearly independent set of vectors belonging to a
subspace S can always be extended so as to constitute a basis of S.

Moreover if r be the dimension of a subspace S, then every set of more than r members
of S will be linearly dependent.

Intersection of subspaces. If S and T be two subspaces of Vn, then the vectors


common to both S and T also constitute a subspace. This subspace is called the
intersection of the subspaces S and T .

8 Row Rank and Column Rank of a Matrix


Row rank of a matrix. Let A = [aij] be any m × n matrix. Each of the m rows of A
consists of n elements. Therefore the row vectors of A are n-vectors. These row vectors
of A will span a subspace R of Vn. This subspace R is called the row space of the matrix
A. The dimension r of R is called the row rank of A. In other words the row rank of a
matrix A is equal to the maximum number of linearly independent rows of A.

Row rank of a matrix. Definition: The maximum number of linearly independent rows
of a matrix A is said to be the row rank of the matrix A.

Left nullity of a matrix. Suppose X is an m-vector written in the form of a row


vector.Then the matrix product XA is defined. The subspace S of Vm generated by the
row vectors X belonging to Vm such that XA = O is called the row null space of the matrix
A. The dimension s of Sis called the left nullity or row nullity of the matrix A.

We shall now prove that the sum of the row rank and the row nullity of a matrix is equal to the
number of rows, i. e.,

r + s = m.

Proof: Since the row space R of A is spanned by the row vectors of A, therefore it will
be a set of all vectors of the form

x1(a11, a12 , a13 , ..., a1n) + x2 (a21, a22 , a23 , ...., a2 n) + …

… + xm (am1, am2 , am3 , ..., amn)


V-162

i. e., of the form ( x1a11 + x2 a21 + … + xm am1, x1a12 + x2 a22 + .... + xm am2 ,

x1a13 + x2 a23 + … + xm am3 , ..., x1a1n + x2 a2 n + … + xm amn)

i. e., of the form XA, where X = ( x1, x2 , ..., xm ) is an m-vector.

Let u1, u2 , ..., u s be a basis of the subspace S of Vm generated by all vectors X such that
XA = O. Then, we have

u1A = u2 A = … = u s A = O.

Since the vectors u1, u2 , ..., u s belong to Vm and form a linearly independent set,
therefore we can find vectors u s + 1, u s + 2 , …, u m in Vm such that the vectors
u1, u2 , ..., u s , u s + 1, ..., u m constitute a basis of Vm . Then every vector X belonging to Vm
can be expressed in the form

X = h1u1 + h2 u2 + .... + hm u m .

Now every member of the subspace R is expressible as XA

i. e., as (h1u1 + h2 u2 + … + hm u m ) A

i. e., as h1u1A + h2 u2 A + … + hs u s A + hs + 1u s + 1A + hs + 2 u s + 2 A + … + hm u m A

i. e., as hs + 1u s + 1A + hs + 2 u s + 2 A + … + hm u m A.

Therefore the m − s n-vectors u s + 1A, u s + 2 A, ..., u m A span R. In fact, these vectors


form a basis of R. For any relation of the form

k s + 1u s + 1A + k s + 2 u s + 2 A + … + k m u m A = O

implies (k s + 1u s + 1 + k s + 2 u s + 2 + … + k m u m )A = O,

which shows that k s + 1u s + 1 + k s + 2 u s + 2 + … + k m u m is a member of the subspace S and


as such it can be linearly expressed in terms of the basis u1, u2 , ..., u s of S.

But the vectors u1, u2 , ..., u m are linearly independent. Therefore a relation of the form
k s + 1u s + 1 + k s + 2 u s + 2 + … + k m u m = p1u1 + p2 u2 + .....+ ps u s will exist if and only if
k s + 1 = k s + 2 = … = k m = 0 . Hence the vectors u s + 1A, u s + 2 A , ..., u m A are linearly
independent and form a basis of R. Thus the dimension of R is m − s.

Hence r=m−s or r + s = m.
V-163

9 Column Rank of a Matrix


Let A = [aij] be any m × n matrix. Each of the n columns of A consists of m elements.
Therefore the column vectors of A are m-vectors. These column vectors of A will span a
subspace C of Vm . This subspace C is called the column space of the matrix A. The
dimension c of C is called the column rank of A. In other words the column rank of a
matrix A is equal to the maximum number of linearly independent columns of A.

Column rank of a matrix. Definition. The maximum number of linearly independent


columns of a matrix A is said to be the column rank of the matrix A.

Right nullity of a matrix. Suppose Y is an n-vector written in the form of a column


vector. Then the matrix product AY is defined. The subspace T of Vn generated by the
column vectors Y belonging to Vn such that AY = O is called the column null space of the
matrix A.The dimension t of T is called the right nullity or column nullity of the matrix A.

As in article 8, we can show that c + t = n.

10 Invariance of Row Rank under E-row Operations


Theorem: Row equivalent matrices have the same row rank.

Proof: Let A be any given m × n matrix. Let B be a matrix row equivalent to A. Since B
is obtainable from A by a finite chain of E-row operations and every E-row operation is
equivalent to pre-multiplication by the corresponding E-matrix, there exist E-matrices
E1, E2 , ..., Ek each of the type m × m such that

B = Ek Ek − 1 … E2 E1A, i. e., B = PA ,

where P = Ek Ek − 1 … E2 E1 is a non-singular matrix of the type m × m.

Let us write

 p11 p12 ... p1m   R1 


   
 p21 p22 ... p2 m   R2 
B = PA =  ... ... ... ...   ...  …(1)
   
 ... ... ... ...   ... 
   
 pm1 pm2 ... pmm  R m 

where the matrix A has been expressed as a matrix of its row sub-matrices
R1, R2 , ..., R m .
V-164

From the product of the matrices on the R.H.S. of (1), we observe that the rows of the
matrix B are

p11R1 + p12 R2 + … + p1m R m ,

p21R1 + p22 R2 + … + p2 m R m ,

… … … … …

… … … … …

pm1R1 + pm2 R2 + … + pmm R m .

Thus we see that the rows of B are all linear combinations of the rows R1, R2 , ...., R m of
A. Therefore every member of the row space of B is also a member of the row space of A.

Similarly by writing A = P−1B and giving the same reasoning we can prove that every
member of the row space of A is also a member of the row space of B. Therefore the row
spaces of A and B are identical.

Thus we see that elementary row operations do not alter the row space of a matrix.
Hence the row rank of a matrix remains invariant under E-row transformations.

Note: From the above theorem we also conclude that pre-multiplication by a


non-singular matrix does not alter the row rank of a matrix.

11 Invariance of Column Rank under E-column Operations


Theorem: Column equivalent matrices have the same column rank.

Or

Post-multiplication by a non-singular matrix does not alter the column rank of a matrix.

Proof: Proceeding in the same way as in article 10, we can show that
post-multiplication with a non-singular matrix does not alter the column space and
therefore the column rank of a matrix.

Note: Since every n-rowed E-matrix is obtained from In by single E-operation (row or
column operation as may be desired), therefore the row rank and column rank of an
E-matrix are equal to n.
V-165

12 Invariance of Column Rank under E-row Operations


Theorem: Row equivalent matrices have the same column rank.

Proof: Let A be any given m × n matrix and let B be a matrix row equivalent to A. Then
there exists a non-singular matrix P such that B = PA.

For every column vector X such that AX = O, we have

BX = (PA) X = P (AX) = PO = O.

Since B = PA, therefore A = P−1B.

Therefore for every vector X such that BX = O, we have

AX = (P−1B) X = P−1 (BX) = P−1O = O.

Thus we see that the matrices A and B have the same right nullities and consequently
their column ranks are equal.

Similarly we can prove that column equivalent matrices have the same row rank.

13 Theorem
If r be the row rank of an m × n matrix A then there exists a non-singular matrix, P such that

K 
PA =   ,
O 

where K is an r × n matrix consisting of a set of r linearly independent rows of A.

Proof: If the row rank r of A is zero, we have nothing to prove. Therefore let us
assume that r > 0. The matrix A has then r linearly independent rows. By elementary
row operations on A we can bring these linearly independent rows in the first r places.
Since the last m − r rows are now linear combinations of the first r rows, they can be
made zero by E-row operations without altering the first r rows.

Thus we see that the matrix A is row equivalent to a matrix B such that

K 
B= ,
O 

where K is an r × n matrix consisting of a set of r linearly independent rows of A.


V-166

Since every elementary row operation is equivalent to pre-multiplication by the


corresponding E-matrix and the product of E-matrices is a non-singular matrix,
therefore there exists a non-singular matrix P such that

K
PA =   ⋅
O

Similarly considering column transformations instead of row transformations, we can


show that if c be the column rank of a matrix A , then there exists a non-singular matrix R such
that

AR = [L O],

where L is an m × c matrix consisting of, c, linearly independent columns of A.

14 Equality of Row Rank, Column Rank and Rank


Theorem 1: The row rank of a matrix is the same as its rank.

Proof: Let, s be the row rank and r, the rank of an m × n matrix A.Since the matrix A is of
row rank s, therefore by article 13, there exists a non-singular matrix P such that
K 
PA =  , where K is an s × n matrix.
O 

Now we know that pre-multiplication by a non-singular matrix does not alter the rank
of a matrix.

∴ Rank (PA) = Rank A = r.

But each minor of order (s + 1) of the matrix PA involves at least one row of zeros.

∴ Rank (PA) ≤ s.

∴ r≤s

Again, since the rank of the matrix A is r, therefore by article 13 of chapter 6 there exists
a non-singular matrix R such that

G 
RA =   ,
O 

where G is an r × n matrix.

Now we know that pre-multiplication by a non-singular matrix does not alter the row
rank of a matrix.
V-167

∴ Row rank (RA) = Row rank A = s.

But the matrix RA has only r non-zero rows. Therefore the row rank of RA can, at the
most, be equal to r.

∴ s≤ r

Hence r = s.

Theorem 2: The column rank of a matrix is the same as its rank.

Proof: Let the matrix A′ be the transpose of the matrix A. Then the columns of A are
the rows of A′ .

∴ the column rank of A = the row rank of A′ = the rank of A′ = the rank of A.

Thus from theorems 1 and 2, we conclude that the rank, row rank and column rank of a
matrix are all equal. In other words we can say that the maximum number of linearly
independent rows of a matrix is equal to the maximum number of its linearly
independent columns and is equal to the rank of the matrix.

Thus we have proved that the row rank, the column rank and the rank of a matrix are
all equal. Therefore sometimes the rank of a matrix is also defined as the maximum number of
linearly independent row vectors or column vectors.

We shall now first consider systems of linear homogeneous equations and then
proceed to discuss systems of non-homogeneous linear equations.

15 Homogeneous Linear Equations


a11 x1 + a12 x2 + … + a1n xn = 0 , 

a21 x1 + a22 x2 + … + a2 n xn = 0 ,

Suppose … … … … … … …  …(1)

… … … … … … … 
am1 x1 + am2 x2 + … + amn xn = 0 

is a system of m homogeneous equations in n unknowns x1, x2 , … , xn. Let

 a11 a12 ... a1n   x1  0 


     
 a21 a22 ... a2 n   x2  0 
a a32 ... a3 n  x  0 
31 3
A=  X=  O= 
 ... ... ... ...   ... ...
     
 ... ... ... ...   ... ...
     
am1 am2 ... amn m × n,  x n  n × 1,  0  m × 1,
V-168

where A, X, O are m × n, n × 1, m × 1matrices respectively. Then obviously we can write


the system of equations (1) in the form of a single matrix equation

AX = O. …(2)

The matrix A is called the coefficient matrix of the system of equations (1).

Obviously x1 = 0 , x2 = 0 , …, xn = 0 i. e., X = O is a solution of (1). It is a trivial


(self-obvious) solution of (1).

Again suppose X1 and X2 are two solutions of (2). Then their linear combination k1X1 + k2 X2 ,
where k1 and k2 are any arbitrary numbers, is also a solution of (2).

We have A(k1X1 + k2 X2 ) = k1(AX1) + k2 (AX2 )

= k1O + k2O = O. [∵AX1 = O and AX2 = O]

Hence k1X1 + k2 X2 is also a solution of (2).

Therefore the collection of all the solutions of the system of equations AX = O forms a sub-space of
the n-vector space Vn.

Theorem: The number of linearly independent solutions of the system of m homogeneous


linear equations in n variables, AX = O, is (n − r), where r is the rank of the matrix A.

Proof:

 a11 a12 ... a1n   x1 


   
 a21 a22 ... a2 n   x2 
a a32 ... a3 n  x 
31 3
Let A=  and X =  
 ... ... ... ...   ...
   
 ... ... ... ...   ...
   
am1 am2 ... amn m × n ,  xn n × 1.

Since the rank of the coefficient matrix A is r, therefore it has r linearly independent
columns. Without loss of generality we can suppose that the first r columns from the
left of the matrix A are linearly independent, because it amounts only to renaming the
components of X.

The matrix A can be written as A = [C1, C2 , …, Cr , … Cn] 1 × n , where C1, C2 , ..., Cn are the
column vectors of the matrix A each of them being an m-vector.

The equation AX = O can now be written as the vector equation

x1C1 + x2C2 + … + xrCr + xr + 1Cr + 1 + … + xnCn = O . …(1)


V-169

Since each of the vectors C r + 1 , C r + 2 , … , C n is a linear combination of the vectors


C1, C2 , ..., Cr , therefore we have relations of the type

Cr + 1 = p11C1 + p12C2 + … + p1rCr , 



Cr + 2 = p21C1 + p22C2 + … + p2 rCr , 
 ...(2)
… … … … … … 
Cn = pk1C1 + pk2C2 + … + pkrCr , where k = n − r

The relations (2) can be written in the form

p11C1 + p12C2 + … + p1rCr − 1. Cr + 1 + 0 . Cr + 2 + … + 0 . Cn = O,



p21C1 + p22C2 + … + p2 rCr + 0 . Cr + 1 − 1. Cr + 2 + … + 0 . Cn = O, 
 ...(3)
… … … … … … … … … … … … 
pk1C1 + pk2C2 + … + pkrCr + 0 . Cr + 1 + 0 . Cr + 2 + ..... − 1. Cn = O 

Comparing (1) and (3), we find that the vectors

 p11  p11   p k1


     
 p12   p22   pk2 
 ...   ...   ... 
     
 ...   ...   ... 
     
 p1r   p2 r   pkr 
     
X1 =  −1  , X 2 =  0  ,..., X n − r =  0 
0   −1  0 
     
0   0  0 
     
 ...   ...   ... 
 ...   ...   ... 
     
0    0 
   0   

are (n − r) solutions of the equation AX = O.

The vectors X1, X2 , ..., X n − r form a linearly independent set. For, if we have a relation
of type

l1X1 + l2 X2 + … + ln − r X n − r = O, …(4)

then comparing the (r + 1)th, (r + 2)th, … , nth components on both sides of (4), we get

− l1 = 0 , − l2 = 0 , ..., − ln − r = 0 ,

i. e., the vectors X1, X2 , ..., X n − r are linearly independent.


V-170

It can now be easily seen that every solution of the equation AX = O is some suitable
linear combination of these n − r solutions X1, X2 , ...., X n − r .

Suppose the vector X, with components x1, x2 , ..., xn is any solution of the equation
AX = O. The the vector

X + xr + 1X1 + xr + 2 X2 + … + xnX n − r , …(5)

which, being a linear combination of solutions, is also a solution. It is quite obvious


that the last n − r components of the vector (5) are all equal to zero. Let z1, z2 , ..., z r be
the first r components of the vector (5). Then the vector whose components are
(z1, z2 , ..., z r , 0 , 0 , ..., 0 ) is a solution of the equation

AX = O.

Therefore from (1), we have

z1C1 + z2C2 + … + z rCr = O.

But the vectors C1, C2 , ..., Cr

are linearly independent. Therefore, we have z1 = 0 , z2 = 0 , ..., z r = 0 . Hence (5) is a


zero vector.

Therefore

X = − xr + 1X1 − xr + 2 X2 − … − xnX n − r .

Thus every solution X is a linear combination of the n − r linearly independent


solutions X1, X2 , …, X n − r .

Therefore the set of solutions

{ X1, X2 , …, X n − r }

forms a basis of the vector space of all the solutions of the system of equations AX = O.

16 Some Important Conclusions About the Nature of


Solutions of the Equation AX = O
Suppose we have m equations in n unknowns. Then the coefficient matrix A will be of
the type m × n. Let r be the rank of the matrix A. Obviously r cannot be greater than n
(the number of columns of the matrix A). Therefore we have either r = n or r < n.
V-171

Case I. If r = n, the equation AX = O will have n − n i. e., no linearly independent


solutions. In this case the zero solution will be the only solution. We know that zero
vector forms a linearly dependent set.

Case II. If r < n, we shall have n − r linearly independent solutions. Any linear
combination of these n − r solutions will also be a solution of AX = O. Thus in this case
the equation AX = O will have an infinite number of solutions.

Case III. Suppose m < n i. e., the number of equations is less than the number of
unknowns. Since r ≤ m, therefore r is definitely less than n. Hence in this case the given
system of equations must possess a non-zero solution. The number of solutions of the
equation AX = O will be infinite.

17 Fundamental Set of Solutions of the Equation AX = O


Suppose the rank r of the coefficient matrix A is less than the number of the unknowns
n. In this case the given equations have a set of n − r linearly independent solutions and
every possible solution is a linear combination of these n − r solutions. This set of n − r
solutions is called a fundamental set of solutions of the equation AX = O.

Definition: A set of linearly independent solutions X1, X2 , ..., X k of the system of


homogeneous equations AX = O is called the fundamental system of solutions of AX = O, if every
solution X of AX = O can be written as a linear combination of these vectors i. e., in the form

X = c1X1 + c2 X2 + … + c k X k ,

where c1, c2 , ..., c k are suitable numbers.

18 Working Rule for Finding the Solutions of the Equation


AX = O

Reduce the coefficient matrix A to Echelon form by applying elementary row


transformations only. This Echelon form will help us to know the rank of the matrix
A. Suppose the matrix A is of the type m × n and its rank comes out to be r. If r < m, then
in the process of reducing the matrix A to Echelon form, (m − r) equations will be
eliminated. The given system of m equations will thus be replaced by an equivalent
system of r equations. Solving these r equations (by Cramer’s rule or otherwise), we
can express the values of some r unknowns in terms of the remaining n − r unknowns.
These n − r unknowns can be given any arbitrarily chosen values.

If r = n, the zero solution (trivial solution) will be the only solution. If r < n, there will
be an infinity of solutions.
V-172

Example 5: Does the following system of equations possess a common non-zero solution?

x + 2 y + 3 z = 0, 3 x + 4 y + 4 z = 0 , 7 x + 10 y + 12 z = 0 . (Lucknow 2005)

Solution : The given system of equations can be written in the form of the single matrix
equation

1 2 3  x  0 
     
AX = 3 4 4  y = 0  = O.
7 10 12  z  0 
    

We shall start reducing the coefficient matrix A to triangular form by applying only
E-row transformations on it. Applying R2 → R2 − 3 R1, R3 → R3 − 7 R1, the given
system of equations is equivalent to

1 2 3  x
   
0 −2 −5  y = O.
0 −4 −9 z 
  

Here we find that the determinant of the matrix on the left hand side of this equation is
not equal to zero. Therefore the rank of this matrix is 3. So there is no need of further
applying E-row transformations on the coefficient matrix. The rank of the coefficient
matrix A is 3, i. e., equal to the number of unknowns. Therefore the given system of
equations does not possess any linearly independent solution. The zero solution, i. e.,
x = y = z = 0 is the only solution of the given system of equations.

Example 6: Solve completely the system of equations

x + 3 y − 2 z = 0 , 2 x − y + 4 z = 0 , x − 11 y + 14 z = 0 . (Meerut 2005B)

Solution : The given system of equations is equivalent to the single matrix equation

1 3 −2  x
   
AX = 2 −1 4  y = O.
1 −11 14 z 
  

We shall reduce the coefficient matrix A to Echelon form by applying only E-row
operations on it. Performing R2 → R2 − 2 R1, R3 → R3 − R1, we have
V-173

1 3 −2  x
   
0 −7 8  y = O.
0 −14 1 z 
  

Performing R3 → R3 − 2 R2 , we have

1 3 −2  x
   
0 −7 8  y = O.
0 0 0  z 
  

The coefficient matrix is now triangular. The coefficient matrix being of rank 2, the
given system of equations possess 3 − 2 = 1 linearly independent solution. We shall
assign arbitrary values to n − r = 3 − 2 = 1 variable and the remaining r = 2 variables
shall be found in terms of these. The given system of equations is equivalent to

x + 3 y − 2 z = 0, − 7 y + 8 z = 0 .

8 10
Thus y= z, x = − z.
7 7

Choose z = c.

8 10
Then y= c, x = − c.
7 7

10 8
Hence x=− c, y = c, z = c
7 7

constitute the general solution of the given system, where c is an arbitrary parameter.

Remark: In matrix form, the general solution can be expressed as

 10   10 
− c − 7 
 x  7
   8   8 
 y =  c =c , where c is an arbitrary number.
z   7   7 
   c   1 
   

Example 7: Does the following system of equations possess a common non-zero solution?

x + y + z = 0 , 2 x − y − 3 z = 0 , 3 x − 5 y + 4 z = 0 , x + 17 y + 4 z = 0 .

Solution: The given system of equations is equivalent to the single matrix equation
V-174

1 1 1 
   x
2 −1 −3   
AX =    y = O.
3 −5 4 z 
 
 1 17 4   

We shall first find the rank of the coefficient matrix A by reducing it to Echelon form
by applying elementary row transformations only.

Applying R2 → R2 − 2 R1, R3 → R3 − 3 R1, R4 → R4 − R1, we get

1 1 1  1 1 1
   
0 −3 −5 0 −3 −5
A~ ~ by R3 → 3 R3 , R4 → 3 R4
0 −8 1 0 −24 3
   
 0 16 3 0 48 9

1 1 1
 
0 −3 −5
~ by R3 → R3 − 8 R2 , R4 → R4 + 16 R2
0 0 43
 
 0 0 −71

1 1 1
 
0 −3 −5 71
~ by R4 → R4 + R3 .
0 0 43 43
 
 0 0 0 

Above is the Echelon form of the coefficient matrix A. We have rank A = the number
of non-zero rows in this Echelon form = 3. The number of unknowns is also 3. Since
rank A is equal to the number of unknowns, therefore the given system of equations
does not possess any linearly independent solution. Thus the given system of
equations possesses no non-zero solution. Hence the zero solution i. e., x = y = z = 0 is
the only solution of the given system of equations.

Example 8: Find all the solutions of the following system of equations :

3 x + 4 y − z − 6w = 0, 2 x + 3 y + 2z − 3w = 0,

2 x + y − 14 z − 9 w = 0 , x + 3 y + 13 z + 3 w = 0 .

Solution: The given system of equations is equivalent to the single matrix equation
V-175

3 4 −1 −6  x
   
2 3 2 −3  y
AX =   z  = O.
2 1 −14 −9
   
 1 3 13 3  w

We shall first find the rank of the coefficient matrix A by reducing it to Echelon form
by applying E-row transformations only.

Applying R4 ↔ R1, we get

1 3 13 3
 
2 3 2 −3
A~
2 1 −14 −9
 
3 4 −1 −6

1 3 13 3
 
0 −3 −20 −9
~
0 −5 −40 −15
 
0 −5 −40 −15

By R2 → R2 − 2 R1, R3 → R3 − 2 R1, R4 → R4 − 3 R1

1 3 13 3
 
0 1 8 3
~
0 1 8 3
 
0 1 8 3

1 1 1
by R2 → − R2 , R3 → − R3 , R4 → − R4
3 5 5

1 3 13 3
 
0 1 8 3
~ by R3 → R3 − R2 , R4 → R4 − R2 .
0 0 0 0
 
0 0 0 0 

The rank of A is obviously 2 which is less than the number of unknowns 4. Therefore
the given system of equations possesses 4 − 2, i. e., 2 linearly independent solutions.
The given system of equations is equivalent to the equation
V-176

1 3 13 3  x  0 
     
0 1 8 3  y 0 
0 0 0 0  z  = 0  ⋅
     
0 0 0 0   w 0 

Thus the given system of four equations is equivalent to the system of two equations,
i. e.,
x + 3 y + 13 z + 4 w = 0 

⋅
y + 8z + 3w = 0 

From these equations, we get

y = − 8 z − 3 w, x = − 3 (− 8 z − 3 w) − 13 z − 3 w

i. e., y = − 8 z − 3 w, x = 11z + 6 w.

Hence x = 11c1 + 6 c2 , y = − 8 c1 − 3 c2 , z = c1, w = c2 constitute the general solution of


the given system of equations, where c1 and c2 are arbitrary numbers. Since we can give
any arbitary values to c1 and c2 , therefore the given system of equations have an infinite
number of solutions.

Remark: In matrix form, the general solution of the given system of equations can
be expressed as

 x   11c1 + 6 c2   11  6
       
 y − 8 c1 − 3 c2  − 8 − 3
 z  =  1c + 0 c  = c1  1  + c2  0 .
   1 2    
 w  0 c1 + 1c2   0   1 

 x  11  x −6
       
 y − 8  y −3
Here =
 z   1 and  z  =  0
       
 w  0  w  1

are two linearly independent solutions of the given system of equations and all their
linear combinations will also be the solutions of the given system of equations.

Example 9: Show that the only real value of λ for which the following equations have non-zero
solutions is 6 :

x + 2 y + 3 z = λx, 3 x + y + 2 z = λy, 2 x + 3 y + z = λz .

(Kanpur 2011; Bundelkhand 10; Lucknow 10)


V-177

Solution: The given system of equations is equivalent to the single matrix equation

1 − λ 2 3   x
   
AX =  3 1− λ 2   y = O.
 2 3 1 − λ  z 
  

If the given system of equations is to possess a non-zero solution, the coefficient


matrix A must be of rank less than 3. If the matrix A is to be of rank less than 3 its
determinant must be equal to zero. Thus we have

1 − λ 2 3 
 
 3 1− λ 2  =0
 2 3 1 − λ 

6 − λ 6−λ 6 − λ
 
or  3 1− λ 2  = 0,
 2 3 1 − λ 

adding the second and the third rows to the first

1 1 1 
 
or (6 − λ ) 3 1− λ 2  =0
2 3 1 − λ 

1 1 0 
 
or (6 − λ ) 3 −λ − 2 −1  = 0 , C2 − C1, C3 − C1
1 1 − λ − 1

or (6 − λ ) [(λ + 2) (λ + 1) + 1] = 0

or (6 − λ ) [λ 2 + 3 λ + 3] = 0 .

− 3 ± (9 − 12)
The roots of the equation λ 2 + 3 λ + 3 = 0 are λ = i. e.,are imaginary.
2

Hence the only real value of λ for which the system of equations is to have a non-zero
solution is 6.
V-178

Comprehensive E xercise 1

Find all the solutions of the following system of linear homogeneous equations :
1. 2 x − 3 y + z = 0 , x + 2 y − 3 z = 0 , 4 x − y − 2 z = 0 .
2. x + y − 3 z + 2 w = 0 , 2 x − y + 2 z − 3 w = 0 , 3 x − 2 y + z − 4 w = 0 ,
− 4 x + y − 3z + w = 0.
3. x + y + z = 0 , 2 x + 5 y + 7 z = 0 , 2 x − 5 y + 3 z = 0 .
4. x + 2 y + 3 z = 0 , 2 x + 3 y + 4 z = 0 , 7 x + 13 y + 19 z = 0 .
5. 4 x + 2 y + z + 3 u = 0 , 6 x + 3 y + 4 z + 7 u = 0 , 2 x + y + u = 0 .
6. 2 x − 2 y + 5 z + 3 w = 0 , 4 x − y + z + w = 0 , 3 x − 2 y + 3 z + 4 w = 0 ,
x − 3 y + 7 z + 6 w = 0.
7. x − 2 y + z − w = 0 , x + y − 2 z + 3 w = 0 , 4 x + y − 5 z + 8 w = 0 ,
5 x − 7 y + 2z − w = 0

A nswers 1
1. x = 0 , y = 0 , z = 0 2. x = 0 , y = 0 , z = 0 , w = 0
3. x = 0 , y = 0 , z = 0 4. x = 2 c , y = − 2 c , z = c
5. x = c1, u = c2 , y = − 2 c1 − c2 , z = − c2
5 7
6. x = c , y = 4 c , z = c , w = c
9 9
5 4
7. x = c1 − c2 , y = c1 − c2 , z = c1, w = c2
3 3

19 System of Linear Non-homogeneous Equations


Sometimes we think that we can solve every two simultaneous equations of the type
a1 x + b1 y = c1 
.
a2 x + b2 y = c2 

But it is not so. For example, consider the simultaneous equations


3x + 4 y = 5 
.
6 x + 8 y = 13

There is no set of values of x and y which satisfies both these equations. Such
equations are said to be inconsistent.
V-179

Let us take another example. Consider the simultaneous equations

3x + 4 y = 5 
.
6 x + 8 y = 10 

These equations are consistent since there exist values of x and y which satisfy both of
4 5
these equations. We see that x = − c + , y = c constitute a solution of these
3 3
equations, where c is arbitrary. Thus these equations possess an infinite number of
solutions.

Now we shall discuss the nature of solutions of a system of non-homogeneous linear


equations :

a11 x1 + a12 x2 + … + a1n xn = b1 


a21 x1 + a22 x2 + … + a2 n xn = b2 

Let … … … … … … … …  …(1)
… … … … … … … … 

am1 x1 + am2 x2 + .. + amn xn = bm 

be a system of m non-homogeneous equations in n unknown x1, x2 , ..., xn.

If we write

 a11 a12 … a1n   x1   b1 


     
 a21 a22 … a2 n   x2   b2 
A= , X = ,B =
… … … … … … 
     
am1 am2 … amn m × n  xn bm 
n ×1 m ×1

where A, X, B are m × n, n × 1and m × 1 matrices respectively, the above equations can


be written in the form of a single matrix equation AX = B.

Any set of values of x1, x2 , ..., xn which simultaneously satisfy all these equations is called a
solution of the system (1). When the system of equations has one or more solutions, the equations
are said to be consistent otherwise they are said to be inconsistent.

The matrix

 a11 a12 … a1n b1 


 
 a21 a22 … a2 n b2 
[A B] = 
… … … … …
 
am1 am2 … amn bm 

is called the augmented matrix of the given system of equations.


V-180

20 Condition for Consistency


Theorem: The system of equations AX = B is consistent i. e., possesses a solution, if any only if
the coefficient matrix A and the augmented matrix [A B] are of the same rank.

(Lucknow 2005; Meerut 07B; Kumaun 10, 12, 13; Kanpur 11)

Proof: Let C1, C2 , ..., C n denote the column vectors of the matrix A. The equation
AX = B is then equivalent to

 x1 
 
 x2 
[C1, C2 , ...., Cn] … = B i. e., x1C1 + x2C2 + … + xnCn = B. …(1)
… 
 
 xn

Let now r be the rank of the matrix A. The matrix A has then r linearly independent
columns and without loss of generality, we can suppose that the first r columns
C1, C2 , ..., Cr form a linearly independent set so that each of the remaining n − r
columns is a linear combination of these r columns.

The condition is necessary. If the given system of equations is consistent, there


must exist n scalars (numbers) k1, k2 , ..., k n such that

k1C1 + k2C2 + …… + k nCn = B. …(2)

Since each of the n − r columns Cr + 1, Cr + 2 , ...., Cn is a linear combination of first r


columns C1, C2 , ..., Cr it is obvious from (2) that B is also a linear combination of
C1, C2 , ..., Cr . Thus the maximum number of linearly independent columns of the
matrix [A B] is also r. Therefore the matrix [A B] is also of rank r. Hence the matrices A
and [A B] are of the same rank.

The condition is sufficient. Now suppose that the matrices A and [A B] are of the
same rank r. The maximum number of linearly independent columns of the matrix
[A B] is then r. But the first r columns C1, C2 , ..., Cr of the matrix [A B] already form a
linearly independent set. Therefore the column B should be expressed as a linear
combination of the columns C1, C2 , ...., Cr .

Thus there exist r scalars k1, k2 , ..., k r such that

k1C1 + k2C2 + … + k rCr = B. …(3)

Now (3) may be written as

k1C1 + k2C2 + … + k rCr + 0 . Cr + 1 + 0 . Cr + 2 + … + 0 . Cn = B. …(4)


V-181

Comparing (1) and (4), we see that

x1 = k1, x2 = k2 , ...., xr = k r , xr + 1 = 0 , xr + 2 = 0 , ..., xn = 0

constitute a solution of the equation AX = B.

Therefore the given system of equations is consistent.

21 Condition for a System of n Equations in n Unknowns to


have a Unique Solution
Theorem: If A be an n-rowed non-singular matrix, X be an n × 1matrix,B be an n × 1matrix,
the system of equations AX = B has a unique solution.

Proof: If A be an n-rowed non-singular matrix, the ranks of the matrices A and [A B]


are both n. Therefore the system of equations AX = B is consistent i. e., possesses a
solution.

Pre-multiplying both sides of AX = B by A −1, we have

A −1AX = A −1B i. e., IX = A −1B i. e., X = A −1B

is a solution of the equation AX = B.

To show that the solution is unique, let us suppose that X1 and X2 be two solutions of
AX = B.

Then AX1 = B, AX 2 = B ⇒ AX 1 = AX 2 ⇒ A −1AX1 = A −1AX2

⇒ IX 1 = IX 2 ⇒ X 1 = X 2.

Hence the solution is unique.

22 Working Rule for Finding the Solution of the Equation


AX = B

Suppose the coefficient matrix A is of the type m × n, i. e., we have m equations in n


unknowns. Write the augmented matrix [A B] and reduce it to Echelon form by
applying only E-row transformations on it. This Echelon form will enable us to know
the ranks of the augmented matrix [A B] and the coefficient matrix A. Then the
following different cases arise :

Case I. Rank A < Rank [A B] .

In this case the equations AX = B are inconsistent i. e., they have no solution.
V-182

Case II. Rank A = Rank [A B] = r (say).

In this case the equations AX = B are consistent i. e., they possess a solution. If r < m,
then in the process of reducing the matrix [A B] to Echelon form, (m − r) equations
will be eliminated. The given system of m equations will then be replaced by an
equivalent system of r equations. From these r equations we shall be able to express
the values of some r unknowns in terms of the remaining n − r unknowns which can be
given any arbitrarily chosen values.

If r = n, then n − r = 0, so that no variable is to be assigned arbitrary values and


therefore in this case there will be a unique solution.

If r < n, then n − r variables can be assigned arbitrary values. So in this case there will be
an infinite number of solutions. Only n − r + 1 solutions will be linearly independent
and the rest of the solutions will be linear combinations of them.

If m < n, then r ≤ m < n. Thus in this case n − r > 0. Therefore when the number of
equations is less than the number of unknowns, the equations will always have an
infinite number of solutions, provided they are consistent.

Example 10: Show that the equations

x + y + z = − 3, 3 x + y − 2 z = − 2, 2 x + 4 y + 7 z = 7

are not consistent. (Avadh 2008; Purvanchal 07)

Solution : The given system of equations is equivalent to the single matrix equation

1 1 1  x −3
     
AX = 3 1 −2  y = −2 = B.
2 4 7  z   7
    

1 1 1 : −3
 
The augmented matrix [A B] = 3 1 −2 : −2.
2 4 7 : 7

We shall reduce the augmented matrix [A B] to Echelon form by applying E-row


transformations only. Applying

R2 → R2 − 3 R1, R3 → R3 − 2 R1, we get


V-183

1 1 1: −3
 
[A B] ~ 0 −2 −5 : 7
0 2 5 : 13

1 1 1: −3
 
~ 0 −2 −5 : 7, applying R3 → R3 + R2 .
0 0 0 : 20 

Above is the Echelon form of the matrix [A B]. We have rank [A B] = the number of
non-zero rows in this Echelon form = 3

Also by the same E-row transformations, we get

1 1 1
 
A ~ 0 −2 −5.
0 0 0 

Obviously rank A = 2.

Since Rank A ≠ Rank [A B], therefore the given equations are inconsistent i. e., they
have no solution.

Remark: The inconsistency of the given equations can also be shown as below :

The given system of equations is equivalent to the matrix equation

1 1 1  x   −3
     
0 −2 −5  y =  7.
0 0 0   z  20 
    

This matrix equation is equivalent to the system of equations

x + y + z = −3 

0 x − 2 y − 5z = 7  .
0 x + 0 y + 0 z = 20 

The last equation shows that 0 = 20 , which is not possible. Hence the given equations
are inconsistent.

Example 11: Show that the equations

x + y + z = 6, x + 2 y + 3 z = 14, x + 4 y + 7 z = 30

are consistent and solve them. (Meerut 2010 B; Avadh 09)


V-184

Solution : The given system of equations is equivalent to the single matrix equation

1 1 1  x   6
     
AX = 1 2 3  y = 14 = B.
1 4 7  z  30 
    

1 1 1 ⋮ 6
 
The augmented matrix [A B] = 1 2 3 ⋮ 14 .
1 4 7 ⋮ 30 

We shall reduce the augmented matrix [A B] to Echelon form by applying elementary


row transformations only. Applying R2 → R2 − R1, R3 → R3 − R1, we get

1 1 1⋮ 6  1 1 1⋮ 6
   
[A B] ~ 0 1 2⋮ 8  ~ 0 1 2⋮ 8, by R3 → R3 − 3 R2 .
0 3 6⋮ 24 0 0 0 ⋮ 0 

Above is the Echelon form of the matrix [A B]. We have rank [A B] = the number of
non-zero rows in this Echelon form = 2 .

1 1 1
 
By the same elementary transformations, we get A ~ 0 1 2.
0 0 0 

Obviously rank A = 2. Since rank A = rank [A B], therefore the given equations are
consistent. Here the number of unknowns is 3. Since rank A is less than the number of
unknowns therefore the given system will have an infinite number of solutions. We
see that the given system of equations is equivalent to the matrix equation

1 1 1  x  6
     
0 1 2  y = 8.
0 0 0   z  0 
    

This matrix equation is equivalent to the system of equations

x + y + z = 6
.
y + 2z = 8 

∴ y = 8 − 2 z , x = 6 − y − z = 6 − (8 − 2 z ) − z = z − 2.

Taking z = c , we see that x = c − 2, y = 8 − 2 c , z = c constitute the general solution of


the given system, where c is an arbitrary constant.
V-185

Example 12: Apply the test of rank to examine if the following equations are consistent :

2 x − y + 3 z = 8, − x + 2 y + z = 4, 3 x + y − 4 z = 0

and if consistent, find the complete solution. (Meerut 2007)

Solution : The given system of equations is equivalent to the single matrix equation

 2 −1 3  x  8
     
AX = −1 2 1  y = 4 = B.
 3 1 −4  z  0 
    

The augmented matrix

2 −1 3 ⋮ 8
 
[A B] = −1 2 1⋮ 4.
3 1 −4 ⋮ 0 

We shall reduce the augmented matrix to Echelon form by applying elementary row
transformations only. Applying R1 ↔ R2 , we get

−1 2 1⋮ 4
 
[A B] ~  2 −1 3 ⋮ 8
 3 1 −4 ⋮ 0 

−1 2 1⋮ 4
 
~ 0 3 5 ⋮ 16, by R2 → R2 + 2 R1, R3 → R3 + 3 R1
0 7 −1 ⋮ 12

−1 2 1 ⋮ 4
 
0 3 5 ⋮ 16, by R3 → 3 R3
0 21 −3 ⋮ 36

−1 2 1 ⋮ 4
 
~ 0 3 5 ⋮ 16, by R3 → R3 − 7 R2
0 0 −38 ⋮ −76

−1 2 1 ⋮ 4
  1
~ 0 3 5 ⋮ 16, by R3 → − R3 .
0  38
 0 1 ⋮ 2
V-186

Above is the Echelon form of the matrix [A B]. We have the rank [A B] = the number
of non-zero rows in this Echelon form = 3.

−1 2 1
 
By the same transformations, we get A ~  0 3 5 ⋅
0 0 1

Obviously rank A = 3. Since rank A = rank [A B], therefore the given equations are
consistent.Here the number of unknowns is 3. Since rank A is equal to the number of
unknowns, therefore the given equations have a unique solution. We see that the
given equations are equivalent to the matrix equation

−1 2 1  x   4
     
0 3 5  y = 16.
0 0 1  z   2
    

This matrix equation is equivalent to the equations

− x + 2 y + z = 4, 3 y + 5 z = 16, z = 2.

These give z = 2, y = 2, x = 2.

Example 13: Show that the equations

x + 2 y − z = 3, 3 x − y + 2 z = 1, 2 x − 2 y + 3 z = 2, x − y + z = − 1

are consistent and solve them. (Meerut 2006B, 09; Rohilkhand 06; Garhwal 15)

Solution: The given system of equations is equivalent to the single matrix equation

1 2 −1  3
  x 
3 −1 2    1
AX =  y =   = B.
2 −2 3    2
  z   
 1 −1 1 −1

The augmented matrix

1 2 −1 ɺ: 3
 
3 −1 2 ɺ: 1
[A B] =  .
2 −2 3 ɺ: 2
 
 1 −1 1 ɺ: −1

Performing R2 → R2 − 3 R1, R3 → R3 − 2 R1, R4 → R4 − R1, we get


V-187

1 2 −1 ⋮ 3
 
0 −7 5 ⋮ −8
[A B] ~ 
0 −6 5 ⋮ −4
 
0 −3 2 ⋮ −4

1 2 −1 ⋮ 3
 
0 −1 0 ⋮ −4
~ , by R2 → R2 − R3
0 −6 5 ⋮ −4
 
0 −3 2 ⋮ −4

1 2 −1 ⋮ 3
 
0 −1 0 ⋮ −4
~ , by R3 → R3 − 6 R2 , R4 → R4 − 3 R2
0 0 5 ⋮ 20 
 
0 0 2 ⋮ 8

1 2 −1 ⋮ 3
 
0 −1 0 ⋮ −4 1 1
~ , by R3 → R3 , R4 → R4
0 0 1 ⋮ 4 5 2
 
0 0 1 ⋮ 4

1 2 −1 ⋮ 3
 
0 −1 0 ⋮ −4
~ , by R4 → R4 − R3 .
0 0 1 ⋮ 4
 
0 0 0 ⋮ 0 

Thus the matrix [A B] has been reduced to Echelon form. We have rank [A B] = the
number of non-zero rows in this Echelon form = 3. Also

1 2 −1
 
0 −1 0
A ~ .
0 0 1
 
0 0 0 

We have rank A = 3. Since rank [A B] = rank A, therefore the given equations are
consistent. Since rank A = 3 = the number of unknowns, therefore the given equations
have unique solution. The given equations are equivalent to the equations

x + 2 y − z = 3, − y = − 4, z = 4.

These give z = 4, y = 4, x = − 1.
V-188

Example 14: State the conditions under which a system of non-homogeneous equations will have
(i) no solution (ii) a unique solution (iii) infinity of solutions. (Lucknow 2008)

Solution : Let AX = B be a system of linear non-homogeneous equations, where A, X, B


are m × n, n × 1, m × 1 matrices respectively.

(i) These equations will have no solution if the coefficient matrix A and the
augmented matrix [A B] are not of the same rank.

(ii) These equations will possess a unique solution if the matrices A and [A B] are of
the same rank and the rank is equal to the number of variables. In particular if A is
a square matrix, these equations will possess a unique solution if and only if the
matrix A is non-singular.

(iii) These equations will have infinity of solutions if the matrices A and [A B] are of
the same rank and the rank is less than the number of variables.

Example 15: By the use of matrices, solve the equations :

x + y + z = 9, 2 x + 5 y + 7 z = 52, 2 x + y − z = 0 .

(Bundelkhand 2011; Avadh 10; Rohilkhand 10)

Solution : The given system of equations is equivalent to the single matrix equation

1 1 1   x   9
 
AX = 2 5 7   y = 52 = B.
2 1 −1 z   0 

The augmented matrix

1 1 1 ⋮ 9
 
[A B] = 2 5 7 ⋮ 52
2 1 −1 ⋮ 0 

1 1 1 ⋮ 9
 
~ 0 3 5 ⋮ 34, by R2 → R2 − 2 R1, R3 → R3 − 2 R1
0 −1 −3 ⋮ −18

1 1 1 ⋮ 9
 
~ 0 −1 −3 ⋮ −18 , by R2 ←→ R3
0 3 5 ⋮ 34

V-189

1 1 1 ⋮ 9
 
~ 0 −1 −3 ⋮ −18 , by R3 → R3 + 3 R2 .
0 0 −4 ⋮ −20 

Above is the Echelon form of the matrix [A B]. We have rank [A B] = the number of
non-zero rows in this Echelon form = 3.

Also by the same E-row transformations, we get

1 1 1
 
A ~ 0 −1 −3 .
0 0 −4

∴ rank A = 3.

Since rank A = rank [A B], therefore the given equations are consistent. Also rank
A = 3 and the number of unknowns is also 3. Hence the given equations will have a
unique solution. To find the solution we see that the given system of equations is
equivalent to the matric equation

1 1 1  x   9
  
0 −1 −3  y =  −18.
0 0 −4  z  −20 

This matrix equation is equivalent to the system of equations

x + y + z = 9, − y − 3 z = − 18, − 4 z = − 20 .

Solving these, we get z = 5, y = 3, x = 1.

Example 16: Investigate for what values of λ , µ the simultaneous equations

x + y + z = 6, x + 2 y + 3 z = 10 , x + 2 y + λz = µ

have (i) no solution, (ii) a unique solution, (iii) an infinite number of solutions.

(Meerut 2006, 09B; Garhwal 06, 11; Bundelkhand 09; Rohilkhand 07; Kanpur 09)

Solution : The matrix form of the given system of equations is

1 1 1  x   6 
    
AX = 1 2 3   y = 10  = B .
1 2 λ   z   µ 

V-190

1 1 1 ⋮ 6
 
The augmented matrix [A B] = 1 2 3 ⋮ 10 
1 2 λ ⋮ µ 

1 1 1 ⋮ 6 
 
~ 0 1 2 ⋮ 4  by R2 → R2 − R1, R3 → R3 − R1
0 1 λ −1 ⋮ µ − 6

1 1 1 ⋮ 6 
 
~ 0 1 2 ⋮ 4  by R3 → R3 − R2 .
0 0 λ −3 ⋮ µ − 10 

If λ ≠ 3, we have rank [A B] = 3 = rank A.

So in this case the given system of equations is consistent. Since rank A = the number
of unknowns, therefore the given system of equations possesses a unique solution.
Thus if λ ≠ 3,the given system of equations possesses a unique solution for any value of
µ.

If λ = 3 and µ ≠ 10, we have rank [A B] = 3 and rank A = 2. Thus in this case rank
[A B] ≠ rank A and so the given system of equations is inconsistent i. e., possesses no
solution.

If λ = 3 and µ = 10, we have rank [A B] = 2 = rank A.

So in this case the given system of equations is again consistent. Since rank A < the
number of unknowns, therefore in this case the given system of equations possesses an
infinite number of solutions.

Example 17: For what values of η the equations

x + y + z = 1, x + 2 y + 4 z = η, x + 4 y + 10 z = η2 ,

have a solution and solve them completely in each case.

(Meerut 2011; Agra 07; Kanpur 06, 08, 10; Kumaun 10, 13)

Solution : The matrix form of the given system is

1 1 1   x 1
    
1 2 4   y =  η.
1 4 10   z  η2 
  
V-191

Performing R2 → R2 − R1, R3 → R3 − R1, we get

1 1 1  x   1 
    
0 1 3  y =  η − 1 .
0 3 9  z  η2 − 1
  

Performing R3 → R3 − 3 R2 , we have

1 1 1  x   1 
    
0 1 3  y =  η−1  …(1)
0 0 0   z  η2 − 3 η + 2
 

Now the given equations will be consistent if and only if

η2 − 3 η + 2 = 0 , i. e., iff (η − 2) (η − 1) = 0 , i. e., iff η = 2 or η = 1.

Case I. If η = 2, the equation (1) becomes

1 1 1  x  1 
 
0 1 3  y = 1 .
 
0 0  z 
 0 0 

The above system of equations is equivalent to

y + 3 z = 1, x + y + z = 1.

∴ y = 1 − 3z , x = 2z .

Thus, x = 2 k , y = 1 − 3 k , z = k constitute the general solution where k is an arbitrary


constant.

Case II. If η = 1, the equation (1) becomes

1 1 1  x  1
     
0 1 3  y = 0  .
0 0 0   z  0 
    

The above system of equations is equivalent to

y + 3 z = 0 , x + y + z = 1.

∴ y = − 3z , x = 1 + 2z .

Thus x = 1 + 2 c , y = − 3 c , z = c constitute the general solution, where c is an arbitrary


constant.
V-192

Comprehensive Exercise 2

1. Use the test of rank to show that the following equations are not consistent :
2 x − y + z = 4, 3 x − y + z = 6, 4 x − y + 2 z = 7, − x + y − z = 9.
(Rohilkhand 2009; Kumaun 11)

2. Show that the equations −2 x + y + z = a, x − 2 y + z = b, x + y − 2 z = c have no


solution unless a + b + c ≠ 0 (Kanpur 2009)
3. Apply the test of rank to examine if the following system of equations is
consistent and if consistent, find the complete solution :
x + y + z = 6, x + 2 y + 3 z = 10 , x + 2 y + 4 z = 1.
4. Use matrix method to solve the equations :
2 x − y + 3 z = 9, x + y + z = 6, x − y + z = 2.
(Meerut 2005; Avadh 06,11; Garhwal 12)
5. Solve completely the equations
x + 2 y + 3 z = 6, 2 x + 4 y + z = 7, 3 x + 2 y + 9 z = 4.
(Kumaun 2014)
6. Show that the equations x + 2 y − 5 z = − 9, 3 x − y + 2 z = 5, 2 x + 3 y − z = 3,
4 x − 5 y + z = − 3 are consistent and solve the same.
7. Show that the equations
x − 3 y − 8 z + 10 = 0 , 3 x + y − 4 z = 0 , 2 x + 5 y + 6 z − 13 = 0
are consistent and solve the same. (Bundelkhand 2007)
8. Solve completely the equations 2 x + 3 y + z = 9, x + 2 y + 3 z = 6, 3 x + y + 2 z = 8.
9. Solve completely the equations
x + y + z = 1, x + 2 y + 3 z = 4, x + 3 y + 5 z = 7, x + 4 y + 7 z = 10 .
(Lucknow 2009)
10. Express the following system of equations into the matrix equation form AX = B:
x + 2 y + z = − 1, 6 x + y + z = − 4, 2 x − 3 y − z = 0 , − x − 7 y − 2 z = 7, x − y = 1.
Determine if this system of equations is consistent and if so find its solution.
11. Examine if the system of equations :
x + y + 4 z = 6, 3 x + 2 y − 2 z = 9, 5 x + y + 2 z = 13
is consistent. Find also the solution if it is consistent.
12. Prove, without actually solving, that the following system of equations has a
unique solution :
5 x + 3 y + 14 z = 4, y + 2 z = 1, x − y + 2 z = 0 .
V-193

13. For what values of the parameter λ will the following equations fail to have a
unique solution
3 x − y + λz = 1, 2 x + y + z = 2, x + 2 y − λz = − 1?
Will the equations have any solutions for these values of λ?
(Garhwal 2008)
14. Solve the equations
λx + 2 y − 2 z − 1 = 0 , 4 x + 2 λy − z − 2 = 0 , 6 x + 6 y + λz − 3 = 0 ,
considering specially the case when λ = 2.

15. Discuss for all values of λ, the system of equations


x + y + 4 z = 6, x + 2 y − 2 z = 6, λx + y + z = 6,
as regards existence and nature of solutions.
16. Solve the following equations by matrix method :
5 x + 3 y + 7 z = 4, 3 x + 26 y + 2 z = 9, 7 x + 2 y + 10 z = 5.

17. Solve the following system of linear equations by matrix method :


5 x − 6 y + 4 z = 15, 7 x + 4 y − 3 z = 19, 2 x + y + 6 z = 46.
18. Solve the following equations by matrix method :
x − y + 2 z = 4, 3 x + y + 4 z = 6, x + y + z = 1.
(Garhwal 2014)
19. Solve the following equations by matrix method :
x − 2 y + 3 z = 6, 3 x + y − 4 z = − 7, 5 x − 3 y + 2 z = 5. (Meerut 2010; Garhwal 13)

A nswers 2
3. x = − 7, y = 22, z = − 9

4. x = 1, y = 2, z = 3

5. x = 1, y = 1, z = 1
1 3 5
6. x = , y = ,z =
2 2 2

7. x = − 1 + 2 c , y = 3 − 2 c , z = c , where c is arbitrary
35 29 5
8. x = , y= ,z =
18 18 18
9. x = c − 2, y = 3 − 2 c , z = c

10. Consistent; x = − 1, y = − 2, z = 4
V-194

1
11. Consistent ; x = 2, y = 2, z =
2
7 7
13. λ ≠ , the solution is unique; λ = − , no solution
2 2

14. In case λ = 2, the general solution of the given system of equations is given by
1
x = − c, y = c, z = 0
2
7 7
15. λ ≠ , the solution is unique; λ = − , no solution
10 10
7 16 3 1
16. x = − c, y = + c, z = c
11 11 11 11

17. x = 3, y = 4, z = 6
5 3 3 1
18. x = − c, y = − + c, z = c
2 2 2 2

−8 5 25 13
19. x = + c, y = − + c, z = c
7 7 7 7

Objective Type Questions

Multiple Choice Questions


Indicate the correct answer for each question by writing the corresponding letter from (a),
(b), (c) and (d).

1. A set of r n-vectors X1, X2 , ........., X r is said to be linearly independent if every


relation of the type k1 X1 + k2 X2 + ......... + k r X r = O implies
(a) k1 + k2 + ......... + k r = 0
(b) k1 = k2 = ......... = k r = 0

(c) k1 + k2 + ......... + k r = 1

(d) none of these

2. Consider the system of equations


a1 x + b1 y + c1z = 0 , a2 x + b2 y + c2 z = 0 , a3 x + b3 y + c3 z = 0 .
a1 b1 c1
If a b2 c2 = 0 , then the system has
2 
a3 b3 c3
V-195

(a) more than two solutions

(b) only trivial solution

(c) no solution
(d) none of these
3. The system of linear equations x + y + z = 2, 2 x + y − z = 3, 3 x + 2 y + kz = 4 has
a unique solution if

(a) k ≠ 0

(b) −1 < k < 1


(c) −2 < k < 2
(d) k = 0
4. The system of equations x + 2 y + 3 z = 1, 2 x + y + 3 z = 2, 5 x + y + 9 z = 4 has

(a) only one solution


(b) infinitely many solutions

(c) no solution

(d) none of these


(Meerut 2003)

Fill in the Blank(s)


Fill in the blanks ‘’.........,’’ so that the following statements are complete and correct.

1. The set consisting only of the zero vector, O, is linearly ......... .

2. The set of three vectors X1 = (1, 0 , 0 ), X2 = (0 , 1, 0 ), X 3 = (0 , 0 , 1) is linearly ........

3. The number of linearly independent solutions of the system of m homogeneous


linear equations in n variables, AX = O, is ......., where r is the rank of the matrix A.

4. When the system of equations has one or more solutions, the equations are said
to be ....., otherwise they are said to be .......

5. The system of m non-homogeneous linear equations in n unknowns, AX = B, is


....... if and only if the coefficient matrix A and the ungmented matrix [A B] are
of the same rank. (Meerut 2001)

6. If A be an n-rowed non-singular matrix, X be an n × 1matrix, B be an n × 1matrix,


the system of equations A X = B, has a ...... solution.

7. If the number of equations is less than the number of unknowns, the equations
will aways have an ....... number of solutions, provided they are consistent.
V-196

True or False
Write ‘T’ for true and ‘F’ for false statement.

1. A set of r n-vectors X1, X2 , ......., X r , is said to be linearly dependent if there exist r


scalars k1, k2 , ........., k r , not all zero, such that k1X1 + k2 X2 + ......... + k r X r = O,
where O denotes the n-vector whose components are all zero.

2. If a set of vectors is linearly independent, then at least one member of the set can
be expressed as a linear combination of the remaining members.

3. The zero solution i. e., x = y = z = 0 is the only solution of the system of equations

x + 2 y + 3 z = 0, 3 x + 4 y + 4 z = 0 , 7 x + 10 y + 12 z = 0 .

4. Let AX = B be a system of m non-homogeneous linear equations in n unknowns.


These equations will have no solution if the coefficient matrix A and the
augmented matrix [A B] are not of the same rank.

5. The system of m non-homogeneous linear equations in n unknowns, AX = B will


possess a unique solution if the matrices A and [A B] are of the same rank and the
rank is less than the number of variables.

A nswers

Multiple Choice Questions


1. (b) 2. (a) 3. (a) 4. (a)

Fill in the Blank(s)


1. dependent 2. independent
3. (n − r) 4. consistent; inconsistent
5. consistent 6. unique
7. infinite

True or False
1. T 2. F 3. T
4. T 5. F

You might also like