Math 3412: Advanced Linear Algebra Chapter 1: Real Vector Spaces
Math 3412: Advanced Linear Algebra Chapter 1: Real Vector Spaces
Math 3412: Advanced Linear Algebra Chapter 1: Real Vector Spaces
1
5) (α + β)v = αv + βv for all α, β ∈ R and v ∈ V, in Rn , we say u = v to mean that αi = βi for all i. Given any γ ∈ R as
6) α(v + w) = αv + αw for all α ∈ R and v, w ∈ V, well, we define u + v and γu by
7) (αβ)v = α(βv) for all α, β ∈ R and v ∈ V, and
α 1 + β1 γα1
8) 1v = v for all v ∈ V.
α 2 + β2 γα2
u+v =
.. ;
γu =
.. .
Thus we define a vector space in such a way that addition and scalar . .
multiplication are intrinsic. Furthermore, the properties listed here char- α n + βn γαn
acterize the fundamental aspects of these operations.
We will refer to a finite collection of vectors, such as {u, v, w}, many Using this arithmetic, one verifies that Rn satisfies Properties (1)–(8)
times in this book. We call such a collection a family to mean that it is an listed above. Thus Rn is a real vector space.
ordered k-tuple of vectors. A family differs from a set in two ways. First,
vectors may occur more than once in a family and we do not discard any Example 1.1.2. Let Zn consist of all polynomials with real coefficients
repetitions; otherwise we alter the family. Thus, for example, the family and having degree at most n, subject to the arithmetic described below.
{u, v, v} is not the same as {u, v}. Second, if we reorder the vectors in a Thus
family we consider the resulting family distinct. Thus, for example, the Zn = {α0 + α1 x + · · · + αn xn : α0 , α1 , . . . , αn ∈ R}
families {u, v, w} and {w, u, v} are distinct. We will also refer to a finite
subject to the arithmetic we presently describe. Note that Zn includes
collection of scalars many times in this book. We call such a collection a
the polynomial with αi = 0 for i = 0, 1, . . ., n. We call this the zero
family to mean that it is an ordered k-tuple of scalars. Pn
polynomial and usually denote it by 0. In Zn , given p(x) = i=0 αi xi
Pn
and q(x) = i=0 βi xi , we say that p = q to mean that αi = βi for i = 0,
Example 1.1.1. Let n be a positive integer. Then Rn denotes the set 1, . . ., n. Given any γ ∈ R as well, we define p + q and γp by
of all columns of real numbers of length n, subject to the arithmetic
described below. Thus p + q (x) = (α0 + β0 ) + (α1 + β1 )x + · · · + (αn + βn )xn ;
γp (x) = γα0 + γα1 x + · · · + γαn xn
α1
α2
R =
n
. : α1 , α2 , . . . , αn ∈ R
. Using this arithmetic, one verifies that Zn satisfies Properties (1)–(8)
. listed above. Thus Zn is a real vector space.
αn
Example 1.1.3. Let Rm×n consist of all m×n matrices with real entries,
subject to the arithmetic we now describe. Given
subject to the arithmetic described below. Thus
α1 β1
α11 α12 · · · α1n
α2 β2
u=
... , v=
.. Rm×n
α
= 21
α22 · · · α2n
ij
: α ∈ R for all i and j
.
····················
αn βn αm1 αm2 · · · αmn
2 3
subject to the arithmetic we now describe. Given Definition 1.2.1. Let A = {v1 , v2 , . . . , vk } be a family of vectors in a
real vector space V. A linear combination of vectors in A is an expression
α11 α12 · · · α1n β11 β12 · · · β1n of the form α1 v1 + α2 v2 + · · · + αn vn , where α1 , α2 , . . . , αk ∈ R.
α α22 · · · α2n β β22 · · · β2n
A = 21 , B = 21 ,
···················· ···················· Thus linear combinations are produced by scaling each vi ∈ A by some αi
αm1 αm2 · · · αmn βm1 βm2 · · · βmn and summing the results. Note that some or even all of the αi can be 0.
For example, 0v1 + 0v2 + · · · + 0vk = 0 is a linear combination of vectors
we say A = B to mean that αij = βij for all i and j. Given any γ ∈ R
in A. Also, say, 0v1 + 1v2 + · · · + 0vk = v2 is such a linear combination.
as well, we define A + B and γA by
Furthermore, the sum of two linear combinations is a linear combination:
α11 + β11 α12 + β12 · · · α1n + β1n
(α1 v1 + α2 v2 + · · · + αk vk ) + (β1 v1 + β2 v2 + · · · + βk vk )
α + β21 α22 + β22 · · · α2n + β2n
A + B = 21 , = (α1 + β1 )v1 + (α2 + β2 )v2 + · · · + (αk + βk )vk
·········································
αm1 + βm1 αm2 + βm2 · · · αmn + βmn
Similarly, a scalar multiple of a linear combination is a linear combination:
γα11 γα12 · · · γα1n
γα21 γα22 · · · γα2n γ(α1 v1 + α2 v2 + · · · + αk vk ) = (γα1 )v1 + (γα2 )v2 + · · · + (γαk )vk
γA =
························
γαm1 γαm2 · · · γαmn
One particularly valuable use of linear combinations is to establish a
Using this arithmetic, one verifies that R m×n
satisfies Properties (1)–(8) coordinate system in a vector space V. To see how this may be done, we
listed above. Thus R m×n
is a real vector space. consider in some detail the system of standard coordinates in Rn . Let n
be a positive integer and define
Example 1.1.4. Let C[0, 1] consist of all continuous functions mapping
1 0 0
[0, 1] → R. Given f, g ∈ C[0, 1], define f + g by f + g (x) = f (x) + g(x) 0 1 0
for all x. Given γ ∈ R as well, define γf by (γf (x) = γ f (x) . Using e1 =
... , e2 =
... , ..., en =
... ,
this arithmetic, one verifies that C[0, 1] satisfies Properties (1)–(8) listed
0 0 1
above. Thus C[0, 1] is a real vector space.
Thus each ej is the column of length n whose jth entry is 1 and whose
1.2. The Standard Basis in Rn other entries are 0. Define E = {e1 , e2 , . . . , en }. We call E the standard
basis for Rn . This set has two essential properties which allow it to be
In any real vector space, there are two basic arithmetical operations the basis for a coordinate system.
to be found: vector addition and scalar multiplication. The essential First, in Rn let
properties of these operations are given in Properties (1)-(8) listed at the
α1
beginning of Section 1.1. These operations are fundamental to vector α2
x= ...
(1.1)
spaces and lead to the basic notion of a linear combination of a set of
vectors. αn
4 5
Then, using the vector arithmetic described in Example 1.1.1, Exercise Set 1.2
1) Let A = {v1 , v2 , . . . , vk } and B = {w1 , w2 , . . . , wp } be families of
α1 0 0 1 0 0
0 α2 0 0 1 0 vectors in a real vector space V. Let also v ∈ V and suppose v is a
x=
... + ... + · · · + .. = α1
... + α2 ... + · · · + αn ...
. linear combination of vectors in A. If each vi ∈ A is itself a linear
0 0 αn 0 0 1 combination of vectors in B, show that v is a linear combination of
vectors in B.
That is, x = α1 e1 + α2 e2 + · · · + αn en . Thus each x ∈ Rn can be written
as a linear combination of vectors in E in at least one way.
1.3. Spanning and Independent Sets
On the other hand, suppose x ∈ Rn and
Let B be a family of vectors in a real vector space V. We consider in
x = α1 e1 + α2 e2 + · · · + αn en , x = β1 e1 + β2 e2 + · · · + βn en more detail the properties B must have in order to serve as the basis for
a coordinate system in V. We also give examples of families of vectors
for some scalars α1 , α2 , . . ., αn , and β1 , β2 , . . ., βn . Rewriting each ej in a real vector space V which do or do not have some of the requisite
as a column of length n and using the arithmetic in Rn , one obtains properties. We begin with a definition.
α1 β1
Definition 1.3.1. Let B = {v1 , v2 , . . . , vk } be a family of vectors in a
α2 β2
x=
... , x=
.. real vector space V. The span of B, denoted by span B, is the set of all
. linear combinations of vectors in B. Thus
αn βn
Thus αi = βi for all i. It follows that each x ∈ Rn can be written as a span B = {α1 v1 + α2 v2 + · · · + αk vk : α1 α2 , . . . , αk ∈ R}
linear combination of vectors in E in at most one way
In summary, each x as in (1.1) has a unique representation as a linear
Example 1.3.2. In R3 , let
combination of vectors in E. This is x = α1 e1 + α2 e2 + · · · + αn en . We
call the scalars α1 , α2 , . . ., αn the coordinates of x relative to E. Clearly,
1 −1 −1 2
these n numbers (in the order listed) completely characterize x. v1 = 2 , v2 = 0 , v3 = 4 , v = −6
In general, given a real vector space V we may seek a family of vectors 0 1 3 −5
B = {v1 , v2 , . . . , vn } for use in establishing a coordinate system in V. It
is natural to require of B that each v ∈ V have a unique expression as a Setting B = {v1 , v2 , v3 }, we determine whether v ∈ span B. Thus we
linear combination of its vectors. In this case, if v = α1 v1 + α2 v2 + · · · + seek α1 , α2 , α3 such that α1 v1 + α2 v2 + α3 v3 = v. That is, we seek α1 ,
αn vn , we call α1 , α2 , . . ., αn the coordinates of v relative to B. We also α2 , α3 such that
say that B is the basis for this coordinate system.
1 −1 −1 2
α1 2 + α2 0 + α3 4 = −6
0 1 3 −5
6 7
Using the arithmetic of vectors in R3 , this holds if and only if for a coordinate system in V we must first be able to write each v ∈ V
as a linear combination of the vectors in B in at least one way. That is,
α1 − α2 − α3 = 2
we must have v ∈ span B for all v ∈ V. In this case, we say that B is a
2α1 + 4α3 = −6
spanning set for V or simply that B spans V.
α2 + 3α3 = −5
For B to be suitable as a basis for V, we’ve also seen that each v ∈ V
Using Gauss-Jordan elimination, one obtains the solutions must be expressible as a linear combination of vectors in B in at most
one way. Thus if v ∈ span B and, say,
α1 −3 −2
α2 = −5 + t −3 ,
α3 0 1 v = α 1 v1 + α 2 v2 + · · · + α k vk , v = β1 v1 + β2 v2 + · · · + βk vk , (1.2)
where t ∈ R is an arbitrary parameter. Thus v ∈ span B. In fact, v we must have αi = βi for all i. We now consider ways of reformulating
can be represented as a linear combination of vectors in B in infinitely this property.
many ways, one for each choice of t. For example (choosing t = 0), Suppose first that at least one v ∈ span B can be written as a linear
v = −3v1 − 5v2 + 0v3 . Also (choosing t = 1), v = −5v1 − 8v2 + v3 . combination as in (1.2), and furthermore αi 6= βi for at least one index i.
Then, subtracting equations,
Example 1.3.3. In Z2 , let
0 = (α1 − β1 )v1 + (α2 − β2 )v2 + · · · + (αk − βk )vk
p1 (x) = 2 − 3x + x2 , p2 (x) = −1 + x + 2x2 , q(x) = −8 + 11x + 8x2 .
Setting B = {p1 , p2 }, we determine whether q ∈ span B. Thus we seek In this case at least one coefficient, namely αi − βi , is nonzero. However,
scalars α1 and α2 such that α1 p1 + α2 p2 = q. That is, we seek α1 and α2 it is obvious that 0 = 0v1 +0v2 +· · ·+0vk as well. Thus 0 can be written
such that as a linear combination in more than one way. In other words, if some
v ∈ span B has more than one representation as a linear combination of
α1 (2 − 3x + x2 ) + α2 (−1 + x + 2x2 ) = −8 + 11x + 8x2 vectors in B then 0 has more than one such representation as well. It
(2α1 − α2 ) + (−3α1 + α2 )x + (α1 + 2α2 )x2 = −8 + 11x + 8x2 follows that if the representation for 0 is unique then the representation
for each v ∈ span B is unique. This motivates the following definition.
Equating coefficients, this last equation holds if and only if
2α1 − α2 = −8 Definition 1.3.4. Let B = {v1 , v2 , . . . , vk } be a family of vectors in a
−3α1 + α2 = 11 real vector space V. We say B is linearly independent to mean that the
α1 + 2α2 = 8 only solution to the equation α1 v1 + α2 v2 + · · · + αk vk = 0 is the trivial
solution α1 = 0, α2 = 0, . . ., αk = 0.
Using Gauss-Jordan elimination, one determines that this system has no
solution. Thus q ∈
/ span{p1 , p2 }.
Proposition 1.3.5. Let B be a linearly independent family of vectors in a
Suppose now that V is a real vector space and B = {v1 , v2 , . . . , vk } real vector space V. Then each v ∈ span B can be written as a linear
is a family of vectors in B. We’ve seen that for B to be suitable as a basis combination of vectors in B in only one way.
8 9
This follows immediately from Definition 1.3.4 and the discussion preced- We determine whether A is linearly independent. To do this we seek all
ing it. solutions to the equation β1 B1 + β2 B2 + β3 B3 + β4 B4 = O, where O
Given a family B = {v1 , v2 , . . . , vk } in a real vector space V, we denotes the 2 × 2 zero matrix. Thus we seek β1 , β2 , β3 , β4 such that
say that B is linearly dependent to mean that it is not independent. In
1 0 1 1 1 1 1 1 0 0
this case, the equation α1 v1 + α2 v2 + · · · + αk vk = 0 has at least one β1 + β2 + β3 + β4 =
0 0 0 0 1 0 1 1 0 0
nontrivial solution. In fact, it has infinitely many in this case.
Using the standard arithmetic of matrices, one obtains the linear system
Example 1.3.6. In R3 , let A = {v1 , v2 , v3 }, where of scalar equations
1 4 7 β1 + β2 + β3 + β4 = 0
v1 = 2 , v2 = 5 , v3 = 8 β2 + β3 + β4 = 0
3 6 9
β3 + β4 = 0
β4 = 0
We determine whether A is linearly independent. To do this we seek all
solutions to the equation α1 v1 + α2 v2 + α3 v3 = 0. That is, we seek α1 , Since this has unique solution β1 = β2 = β3 = β4 = 0, A is linearly
α2 , α3 such that independent.
Exercise Set 1.3
1 4 7 0
α1 2 + α2 5 + α3 8 = 0 4
1) In R , let B = {b1 , b2 , . . . , b5 }, where b1 , b2 , b3 , b4 , and b5 are
3 6 9 0 the respective vectors
This leads to the system of equations 1 0 2 −3 1
−2 1 0 4 5
, , , , and .
α1 + 4α2 + 7α3 = 0 3 2 14 −4 −1
2α1 + 5α2 + 8α3 = 0 2 0 4 −1 −8
3α1 + 6α2 + 9α3 = 0
a) Determine whether B is linearly independent.
The solution set is b) Determine whether x1 = [1, 4, −3, −8]T is in span B. If so, de-
α1 1
α2 = t −2 termine whether its representation as a linear combination of
α3 1 the vectors in B is unique.
c) Determine whether x2 = [4, −1, 6, 2]T is in span B. If so, deter-
Since there are infinitely many solutions, A is linearly dependent.
mine whether its representation as a linear combination of the
Example 1.3.7. In R2×2 let A = {B1 , B2 , B3 , B4 }, where vectors in B is unique.
10 11
Determine whether B is linearly independent. Proof. (⇒) Suppose that A is linearly dependent. Then the equation
3) Let A = {v1 , v2 , v3 } be the set vectors given in Example 1.3.6.
α 1 v1 + α 2 v2 + · · · + α k vk = 0
a) Write 0 as a linear combination of the vectors in A in two dis-
tinct ways. has at least one nontrivial solution. That is, there is some solution with
b) Define v = v1 + 2v2 − 3v3 . Write v as a linear combination at least one coefficient, say αi , nonzero. For this solution,
of the vectors in A in another way. What are all other ways of
−αi vi = α1 v1 + · · · + αi−1 vi−1 + αi+1 vi+1 + · · · + αk vk
writing v as a linear combination of the vectors in A?
α1 αi−1 αi+1 αk
4) Let B = {v1 , v2 , . . . , vk } be a family of vectors in a real vector vi = − v1 − · · · − vi−1 − vi+1 − · · · − vk
αi αi αi αi
space V.
a) If x, y ∈ span B, show that x + y ∈ span B. Thus vi ∈ span A0 , where A0 = {v1 , . . . , vi−1 , vi+1 , . . . , vk }.
(⇐) Suppose there is an index i such that vi is in the span of the
b) If x ∈ span B and γ ∈ R, show that γx ∈ span B.
remaining vectors in A. Thus there are scalars α1 , . . ., αi−1 , αi+1 , . . ., αk
5) Let {v1 , v2 } be a family in a real vector space V, and define w1 = such that
v1 + v2 and w2 = v1 − v2 . If {v1 , v2 } is linearly independent, show
that {w1 , w2 } is linearly independent. vi = α1 v1 + · · · + αi−1 vi−1 + αi+1 vi+1 + · · · + αn vn
6) Let V be a real vector space and let A = {v1 , v2 , . . . , vk } be a
Subtracting vi from both sides,
family of vectors in V. If vj = 0 for some j show that A is linearly
dependent. 0 = α1 v1 + · · · + αi−1 vi−1 − vi + αi+1 vi+1 + · · · + αn vn .
Since not all coefficients on the right side are 0 (the coefficient of vi is −1),
1.4. More on Spanning and Independent Sets
A is linearly dependent.
Let A = {v1 , v2 , . . . , vk } be a family of vectors in a real vector Corollary 1.4.2. Let A = {v1 , v2 , . . . , vk } be a linearly independent
space V. We’ve seen that the span of A is the set of all v ∈ V that can be family of vectors in a real vector space V. Let also v ∈ V and suppose
written as a linear combination of vectors of A in at least one way; A is v∈/ span A. Then B = A ∪ {v} is linearly independent.
linearly independent if each v ∈ span A can be written so in at most one
way. Since the notions of span and independence both arise from that of Proof. Suppose α1 , α2 , . . ., αk , β are scalars such that
a linear combination, they are related to each other in subtle ways. In
this section, we give results establishing some of these relationships. α1 v1 + α2 v2 + · · · + αk vk + βv = 0. (1.3)
12 13
In particular, v ∈ span A. Since this is impossible, β = 0. Thus (1.3) Adding (1.4) and (1.5), one obtains
reduces to α1 v1 + α2 v2 + · · ·+ αk vk = 0. Since A is linearly independent,
the only solution here is α1 = 0, α2 = 0, . . ., αk = 0. Therefore all v = (α1 + β1 )v1 + · · · + (αi−1 + βi−1 )vi−1
coefficients in (1.3) must be 0, and B is linearly independent. + (αi+1 + βi+1 )vi+1 + · · · + (αk + βk )vk
Theorem 1.4.3. Let A = {v1 , v2 , . . . , vk } be a family of vectors in a real Thus v ∈ span A0 , and hence span A ⊆ span A0 . It follows that span A0 =
vector space V. If A is linearly dependent then A contains some subset span A.
A0 = {v1 , . . . , vi−1 , vi+1 , . . . , vk } such that span A0 = span A.
Before giving our last results on independent and spanning sets, we
Proof. Suppose A is linearly dependent. By Proposition 1.4.1, some introduce a result we will use presently.
vector, say vi , is in the span of the others. Define
Proposition 1.4.4. Let A ∈ Rm×n . If n > m then the homogeneous
0
A = {v1 , . . . , vi−1 , vi+1 , . . . , vk }, system Ax = 0 has at least one nontrivial solution. That is, there is
some x0 ∈ Rn with x0 6= 0 such that Ax0 = 0.
so that vi ∈ span A0 . We show that span A0 = span A.
This result follows from the row-reduction algorithm for solving linear
First, let v ∈ span A0 . Then there are scalars α1 , . . ., αi−1 , αi+1 ,
systems of equations. An alternative, somewhat more formal proof can
. . ., αk such that
be given by induction on m, the number of rows of A. We omit the
v = α1 v1 + · · · + αi−1 vi−1 + αi+1 vi+1 + · · · + αk vk . details.
Then, obviously, Theorem 1.4.5. Let A = {w1 , w2 , . . . , wp } be a spanning set for a real
vector space V. If B = {v1 , v2 , . . . , vk } is a family in V and k > p then
v = α1 v1 + · · · + αi−1 vi−1 + 0vi + αi+1 vi+1 + · · · + αk vk B is linearly dependent.
as well. Thus v ∈ span A, and hence span A0 ⊆ span A. Proof. Since A spans V and each vj is a vector in V, each vj is a linear
Suppose now that v ∈ span A. Then there are scalars α1 , α2 , . . ., αk combination of the vectors in A. Thus there are scalars αij with i = 1, 2,
such that v = α1 v1 + α2 v2 + · · · + αk vk . Thus . . . p and j = 1, 2, . . ., k such that
v − αi vi = α1 v1 + · · · + αi−1 vi−1 + αi+1 vi+1 + · · · + αk vk (1.4)
v1 = α11 w1 + α21 w2 + · · · + αp1 wp
v2 = α12 w1 + α22 w2 + · · · + αp2 wp
Furthermore, since vi ∈ span A0 one has αi vi ∈ span A0 as well. (See also
·································
Problem 2 in Exercise Set 1.3.) Thus there are scalars β1 , . . ., βi−1 , βi+1 , vk = α1k w1 + α2k w2 + · · · + αpk wp
. . ., βk such that
(Note that the indices on the right sides above do not follow the usual
αi vi = β1 v1 + · · · + βi−1 vi−1 + βi+1 vi+1 + · · · + βk vk . (1.5) pattern for matrices.) Using these coefficients (but in the “transposed”
14 15
order), produce the homogeneous system of equations Corollary 1.4.6. Let {w1 , w2 , . . . , wp } be a spanning set for a real vector
space V. If {v1 , v2 , . . . , vk } is a linearly independent family in V then
α11 x1 + α12 x2 + · · · + α1k xk = 0 k ≤ p.
α21 x1 + α22 x2 + · · · + α2k xk = 0
······························
Exercise Set 1.4
αp1 x1 + αp2 x2 + · · · + αpk xk = 0 1) Let A = {v1 , v2 , . . . , vk } be a linearly independent family of vectors
in a real vector space V. Let also v ∈ V and define B = A ∪ {v}. If
Since k > p, our last proposition implies that this system has at least B is linearly dependent, show that v ∈ span A.
one nontrivial solution. Let β1 , β2 , . . ., βk be such a solution. We show
that β1 v1 + β2 v2 + · · · + βk vk = 0. Since at least one βj is nonzero, this 2) Let A = {v1 , v2 , . . . , vk } be a family of vectors in a real vector
proves that B is linearly dependent. space V, let α1 , α2 , . . ., αk be nonzero scalars, define wi = αi vi for
Note that all i, and set B = {w1 , w2 , . . . , wk }.
a) Show that A is linearly independent if and only if B is linearly
β1 v1 + β2 v2 + · · · + βk vk = β1 (α11 w1 + α21 w2 + · · · + αp1 wp )
independent.
+ β2 (α12 w1 + α22 w2 + · · · + αp2 wp )
b) Show that span A = span B.
+ ·······························
+ βk (α1k w1 + α2k w2 + · · · + αpk wp ) 3) Let A = {v1 , v2 , . . . , vm } be a family of vectors in a real vector
space V, let k < m, and set A0 = {v1 , v2 , . . . , vk }.
Regrouping terms on the right side above, a) If A is linearly independent, show that A0 is linearly indepen-
dent.
β1 v1 + β2 v2 + · · · + βk vk = (α11 β1 + α12 β2 + · · · + α1k βk )w1
+ (α21 β1 + α22 β2 + · · · + α2k βk )w2 b) If A0 is linearly dependent, show that A is linearly dependent.
16 17
1.5. Basis and Dimension B = {v1 , v2 , . . . , vn } it has infinitely many such bases. (We will prove
this in Section 1.6.) Thus there are infinitely many coordinate systems
Let V be a real vector space. Clearly, any family which is linearly
that may be produced in V. Despite this abundance of bases (and hence
independent and spans V has particularly desirable properties. In partic-
of coordinate systems), however, there is one feature common to all of
ular, it may serve as a basis for a coordinate system in V. In this section,
them: if any basis contains, say, n vectors then all of them contain n
we consider such sets in more detail. We also formalize our notion of
vectors.
coordinates.
Definition 1.5.1. Let B be a family of vectors in a real vector space V. Theorem 1.5.4. Let V be a real vector space and let B = {v1 , v2 , . . . , vn }
We say B is a basis for V to mean that B is linearly independent and and D = {w1 , w2 , . . . , wm } be bases for V. Then n = m.
spans V.
Proof. By definition, B and D are each linearly independent families
Theorem 1.5.2. Let B = {v1 , v2 , . . . , vn } be a basis for a real vector
that span V. Since B is linearly independent and D spans V, n ≤ m by
space V. For every v ∈ V, there are unique scalars α1 , α2 , . . ., αn such
Corollary 1.4.6. Since at the same time D is linearly independent and B
that
spans V, we also have m ≤ n. Therefore, n = m.
v = α 1 v1 + α 2 v2 + · · · + α n vn . (1.6)
18 19
Similarly, let C = {w1 , w2 , . . . , wp } be a spanning set for V. Since B is fewer vectors than the linearly independent set B. Since this is impossible
linearly independent, p ≥ n by Corollary 1.4.6. (see Corollary 1.4.6), A is linearly independent.
We now give two more results concerning properties of bases. [(2) and (3) ⇒ (1)] Suppose A is linearly independent and spans V.
Then A is a basis for V. Since B is also a basis for V and all bases contain
Theorem 1.5.6. Let V be a real n-dimensional vector space, let A = the same number of vectors, k = n.
{w1 , w2 , . . . , wk } be a family of vectors in V, and consider the following
three statements: This result is useful for verifying that some family of vectors forms a
basis for a vector space V. Suppose, for example, that V is n-dimensional
1) k = n;
and A = {v1 , v2 , . . . , vn } is a family of vectors in V. Thus, in particular,
2) A is linearly independent;
A contains n vectors. If A is shown to be linearly independent then by
3) A spans V.
the foregoing result A also spans V. On the other hand, if A spans V
If any two of these statements are true then so is the third. then by the foregoing result A is also linearly independent. In either case,
A forms a basis for V.
Proof. Note first that since V is n-dimensional it has some basis con- Suppose now that V is an n-dimensional vector space and A =
taining n vectors. Let B = {v1 , v2 , . . . , vn } be such a basis. {v1 , v2 , . . . , vk } is a family in V. If A is linearly independent then k ≤ n.
[(1) and (2) ⇒ (3)] Suppose A contains n vectors and is linearly If k = n then A spans V and hence is a basis for V. Otherwise, k < n
independent. Suppose also that v ∈ V. Since B is a spanning set for V and it is impossible for A to span V. On the other hand, suppose A is a
containing n vectors and A∪{v} contains n+1 vectors, A∪{v} is linearly spanning set for V. Then k ≥ n. If k = n then A is linearly independent
dependent by Theorem 1.4.5. Thus there are scalars α1 , α2 , . . ., αn , β and hence a basis for V. Otherwise, k > n and it is impossible for A to
with at least one of them nonzero such that be linearly independent. Our last result elaborates further on these facts.
20 21
n vectors, but when i ≤ n − k the set {v1 , v2 , . . . , vk+i−1 } contains fewer also v, w ∈ V and suppose
than n vectors. Thus a vector vk+i ∈ / span{v1 , v2 , . . . , vk+i−1 } may be
chosen as indicated. Furthermore, since vk+1 ∈ / span A and A is linearly β1 δ1
β2 δ2
independent, the set {v1 , v2 , . . . , vk+1 } is linearly independent as well. v=
.. , w=
... .
Similarly, the sets {v1 , v2 , . . . , vk+2 } and so on, up to {v1 , v2 , . . . , vn }, .
are all linearly independent. Since this last set is linearly independent βn B δn B
and contains n vectors, it is a basis for V.
Show that, if γ ∈ R as well,
(2) We prove by induction on k − n. If k − n = 1 then, in particular,
k > n. Thus A is linearly dependent. By Theorem 1.4.3, there is some
β1 + δ1 γβ1
set A0 = {v1 , . . . , vi−1 , vi+1 , . . . , vk } such that span A0 = span A = V. β2 + δ2 γβ2
Furthermore, A0 contains n vectors. Thus, by Theorem 1.5.6, A0 is a v+w =
.. ,
γv =
.. .
. .
basis for V. βn + δn γβn B
B
Suppose now that m is a natural number and the conclusion holds
whenever k − n = m. If k − n = m + 1 then, as before, k > n and hence
A is linearly dependent. By Theorem 1.4.3, there is some set 1.6. Changing Basis
A0 = {v1 , . . . , vi−1 , vi+1 , . . . , vn+m+1 }. Let V be a real n-dimensional vector space. Thus V has some basis,
say B = {b1 , b2 , . . . , bn }, containing n vectors. We have mentioned in
such that span A0 = span A = V. By our induction hypothesis, we may Section 1.5 that in this case V has infinitely many bases. Thus there are
discard m more vectors from A0 in such a way that the remaining vec- infinitely many coordinate systems for V as well. In this section, given a
tors form a basis for V. Since, in this case, the total number of vectors basis B for V, we characterize all other bases for V. We also show how
discarded is m + 1, this proves our result. . to change from one coordinate system to another.
We begin by showing that V has infinitely many bases.
Exercise Set 1.5
Theorem 1.6.1. Let B = {v1 , v2 , . . . , vn } be a basis for a real vector
1) What is the dimension of Rn ? [Justify your answer.] space V and let D = {w1 , w2 , . . . , wm } be a family of vectors in V. Then
2) Let n ∈ N and let Zn be as in Example 1.1.2. Find a basis for Zn and D is a basis for V if and only if m = n and there is an invertible matrix
show that your answer is correct. Hence determine the dimension A = [αij ]ni,j=1 such that wj = α1j v1 + α2j v2 + · · · + αnj vn for all j.
of Zn . Before proving this result, we recall the following fact usually proved in
an introductory lineary algebra course. A square matrix A is invertible if
3) Let m, n ∈ N and let Rm×n be as in Example 1.1.3. Find a basis
and only if the matrix-vector equation Ax = 0 has unique solution x = 0.
for Rm×n and show that your answer is correct. Hence determine
the dimension of Rm×n .
Proof. (⇒) Suppose D is a basis for V. Then m = n by Theorem 1.5.4.
4) Let B = {b1 , b2 , . . . , bn } be a basis for a real vector space V. Let Also, since B is a basis for V and wj ∈ V for all j, there are scalars αij
22 23
with i, j = 1, 2, . . ., n such that wj = α1j v1 + α2j v2 + · · · + αnj vn for Regrouping on the right side above,
all j. Define A = [αij ]ni,j=1 . To show that A is invertible we show that
the equation Ax = 0 has unique solution x = 0. 0 = (α11 β1 + α12 β2 + · · · + α1n βn )v1
Clearly, x = 0 is a solution. Suppose now that x = [β1 , β2 , . . . , βn ]T + (α21 β1 + α22 β2 + · · · + α2n βn )v2
is also a solution. Thus + ·······························
αi1 β1 + αi2 β2 + · · · + αin βn = 0 + (αn1 β1 + αn2 β2 + · · · + αnn βn )vn
for all i. Then Since B is linearly independent, all coefficients on the right side above
0 = 0v1 + 0v2 + · · · + 0vn are 0. That is, αi1 β1 + αi2 β2 + · · · + αin βn = 0 for i = 1, 2, . . ., n.
Equivalently,
= (α11 β1 + α12 β2 + · · · + α1n βn )v1
β1 0
+ (α21 β1 + α22 β2 + · · · + α2n βn )v2 β2 0
A
.. = ...
+ ······························· .
+ (αn1 β1 + αn2 β2 + · · · + αnn βn )vn βn 0
Regrouping terms on the right side above, Since A is invertible, βi = 0 for all i. Thus D is linearly independent.
Since D also contains n vectors, D spans V by Theorem 1.5.6. Thus D
0 = β1 (α11 v1 + α21 v2 + · · · + αn1 vn ) is a basis for V.
+ β2 (α12 v1 + α22 v2 + · · · + αn2 vn )
Suppose now that V is a real n-dimensional vector space and B =
+ ······························· {v1 , v2 , . . . , vn } and D = {w1 , w2 , . . . , wn } are bases for V. Then for
+ βn (α1n v1 + α2n v2 + · · · + αnn vn ) each v ∈ V, there are scalars β1 , β2 , . . ., βn and δ1 , δ2 , . . ., δn such that
= β1 w1 + β2 w2 + · · · + βn wn
v = β1 v 1 + β2 v 2 + · · · + βn v n , v = δ1 w1 + δ2 w2 + · · · + δn wn .
Since D is a basis for V, it is linearly independent. Thus βi = 0 for all i.
Since the equation Ax = 0 has unique solution x = 0, A is invertible.
Alternatively, we may write
(⇒) Suppose now that m = n, there are scalars αij such that wj =
α1j v1 + α2j v2 + · · · + αnj vn for all j, and A = [αij ] is invertible. We
β1 δ1
show first that D = {w1 , w2 , . . . , wn } is linearly independent. Suppose β2 δ2
β1 , β2 , . . ., βn are scalars such that β1 w1 + β2 w2 + · · · + βn wn = 0. Then v=
.. , v=
... (1.7)
.
0 = β1 (α11 v1 + α21 v2 + · · · + αn1 vn ) βn B δn D
+ β2 (α12 v1 + α22 v2 + · · · + αn2 vn ) We now show how these columns of coordinates are related.
+ ······························· Since B is a basis for V it is, among other things, a spanning set
+ βn (α1n v1 + α2n v2 + · · · + αnn vn ) for V. Since wj ∈ V for all j, there are scalars αij with i, j = 1, 2, . . ., n
24 25
such that so that x and y are ordinary columns in Rn . Then the columns of
coordinates for v relative to B and D, respectively, satisfy the relation
α11 α12 α1n
α21 α22 α2n S −1 y = x. Alternatively, Sx = y. We call S the change-of-basis matrix
w1 =
... , w2 =
... , ..., wn =
... (1.8) from the B- to the D-system of coordinates, while S −1 is the change-of-
αn1 αn2 αnn basis matrix from the D- to the B-system.
B B B
To summarize the foregoing discussion, we have the following al-
Let v ∈ V be a vector with B- and D-system coordinates as in (1.7). As
gorithm for changing from one system of coordinates to another. Let
noted above, the latter relation means that
B = {v1 , v2 , . . . , vn } be a given basis for V and let D = {w1 , w2 , . . . , wn }
δ1 w1 + δ2 w2 + · · · + δn wn = v. be another basis. To produce the change-of-basis matrix from the B- to
the D-system of coordinates:
Representing w1 , w2 , . . ., wn , and v in their B-system coordinates,
1) For j = 1, 2, . . . , n, produce the family of coordinates α1j , α2j ,
α11 α12 α1n β1
α21 α22 α2n β2 . . ., αnj for wj relative to B. Thus each wj may be represented as
δ1
... + δ2 ... + · · · + δn ... = ..
in (1.8);
.
αn1 αn2 αnn βn 2) Define S −1 = [αij ]n×n . (This is the change-of-basis matrix for pass-
B B B B
ing from the D- to the B-system of coordinates.)
Following the arithmetic established in Problem 1 of Exercise Set 1.5, we
obtain 3) Compute S = (S −1 )−1 . This is the desired change-of-basis matrix.
α11 δ1 + α12 δ2 + · · · + α1n δn β1
α21 δ1 + α22 δ2 + · · · + α2n δn β2 Exercise Set 1.6
· · · · · · · · · · · · · · · · · · · · · · · · · · · · = ..
. 2
1) In R let B = {v1 , v2 }, where
αn1 δ1 + αn2 δ2 + · · · + αnn δn B βn B
Since the coordinates of v relative to B are unique, this yields the simul- −2 4
v1 = , v2 =
3 −3
taneous scalar equations
α11 δ1 + α12 δ2 + · · · + α1n δn = β1 a) Find S, the change-of-basis matrix from the E- to the B-system
of coordinates.
α21 δ1 + α22 δ2 + · · · + α2n δn = β2
······························ b) Find the B-coordinates for x and y, where
αn1 δ1 + αn2 δ2 + · · · + αnn δn = βn 5 7
x= , y=
Define S −1 = [αij ]n×n . Note that S −1 is, indeed, invertible by our last 3 −4
result. Define also
2) In R3 let B = {v1 , v2 , v3 }, where
β1 δ1
β2 δ2
x= y= −2 −1 2
.. , ... , v1 = 1 , v2 = 0 , v3 = 1
.
βn δn 0 1 1
26 27
a) Find S, the change-of-basis matrix from the E- to the B-system
of coordinates.
b) Find the B-coordinates for x and y, where
1 2
x = 2, y = 1
3 1
1 + x − x2 , 1 − x + x2 , −1 + x + x2 .
and set D = {q0 , q1 , q2 }. Then B and D are each bases for Z2 , and
you may assume so here.
a) Find S, the change-of-basis matrix from the B- to the D-system
of coordinates.
b) Let r(x) = 3 + 4x + 5x2 . Hence find the coordinates of r relative
to D.
28 29