Linear Algebra: Answers To Exercises
Linear Algebra: Answers To Exercises
LINEAR ALGEBRA
Jim Hefferon
http://joshua.smcvt.edu/linearalgebra
Notation
R, R+ , Rn real numbers, positive reals, n-tuples of reals
N, C natural numbers {0, 1, 2, . . . }, complex numbers
(a .. b), [a .. b] open interval, closed interval
h. . .i sequence (a list in which order matters)
hi,j row i and column j entry of matrix H
V, W, U vector spaces
~v, ~0, ~0V vector, zero vector, zero vector of a space V
Pn , Mn×m space of degree n polynomials, n×m matrices
[S] span of a set
hB, Di, β, ~ ~δ basis, basis vectors
En = h~e1 , . . . , ~en i standard basis for Rn
V= ∼W isomorphic spaces
M⊕N direct sum of subspaces
h, g homomorphisms (linear maps)
t, s transformations (linear maps from a space to itself)
RepB (~v), RepB,D (h) representation of a vector, a map
Zn×m or Z, In×n or I zero matrix, identity matrix
|T | determinant of the matrix
R(h), N (h) range space, null space of the map
R∞ (h), N∞ (h) generalized range space and null space
These are answers to the exercises in Linear Algebra by J Hefferon. An answer labeled
here as One.II.3.4 is for the question numbered 4 from the first chapter, second section,
and third subsection. The Topics are numbered separately.
If you have an electronic version of this file then save it in the same directory as the
book. That way, clicking on the question number in the book takes you to its answer
and clicking on the answer number takes you to the question,
I welcome bug reports and comments. Contact information is on the book’s home
page http://joshua.smcvt.edu/linearalgebra.
Jim Hefferon
Saint Michael’s College, Colchester VT USA
2014-Mar-17
Contents
to conclude that the system has a unique solution if and only if b 6= 0 (we use the case
assumption that c 6= 0 to get a unique x in back substitution). But — where a = 0
and c 6= 0 — the condition “b 6= 0” is equivalent to the condition “ad − bc 6= 0”. That
finishes the second case.
Finally, for the third case, if both a and c are 0 then the system
0x + by = j
0x + dy = k
might have no solutions (if the second equation is not a multiple of the first) or it
might have infinitely many solutions (if the second equation is a multiple of the first
then for each y satisfying both equations, any pair (x, y) will do), but it never has a
unique solution. Note that a = 0 and c = 0 gives that ad − bc = 0.
One.I.1.30 Recall that if a pair of lines share two distinct points then they are the same
line. That’s because two points determine a line, so these two points determine each
of the two lines, and so they are the same line.
Thus the lines can share one point (giving a unique solution), share no points
(giving no solutions), or share at least two points (which makes them the same line).
One.I.1.31 For the reduction operation of multiplying ρi by a nonzero real number k,
we have that (s1 , . . . , sn ) satisfies this system
a1,1 x1 + a1,2 x2 + · · · + a1,n xn = d1
..
.
kai,1 x1 + kai,2 x2 + · · · + kai,n xn = kdi
..
.
am,1 x1 + am,2 x2 + · · · + am,n xn = dm
if and only if a1,1 s1 + a1,2 s2 + · · · + a1,n sn = d1 and . . . kai,1 s1 + kai,2 s2 + · · · +
kai,n sn = kdi and . . . am,1 s1 + am,2 s2 + · · · + am,n sn = dm by the definition of
‘satisfies’. Because k 6= 0, that’s true if and only if a1,1 s1 + a1,2 s2 + · · · + a1,n sn = d1
and . . . ai,1 s1 +ai,2 s2 +· · ·+ai,n sn = di and . . . am,1 s1 +am,2 s2 +· · ·+am,n sn = dm
(this is straightforward canceling on both sides of the i-th equation), which says that
(s1 , . . . , sn ) solves
a1,1 x1 + a1,2 x2 + · · · + a1,n xn = d1
..
.
ai,1 x1 + ai,2 x2 + · · · + ai,n xn = di
..
.
am,1 x1 + am,2 x2 + · · · + am,n xn = dm
as required.
6 Linear Algebra, by Hefferon
The i-th row unchanged because of the i 6= j restriction. Because the i-th row is
unchanged, the operation −kρi + ρj returns the j-th row to its original state.
(Observe that the i = j conditino on the kρi + ρj is needed, or else this could
happen
2ρ1 +ρ1 −2ρ1 +ρ1
3x + 2y = 7 −→ 9x + 6y = 21 −→ −9x − 6y = −21
and so the result wouldn’t hold.)
One.I.1.35 Let p, n, and d be the number of pennies, nickels, and dimes. For variables
that are real numbers, this system
p + n + d = 13 −ρ1 +ρ2 p + n + d = 13
−→
p + 5n + 10d = 83 4n + 9d = 70
has more than one solution; in fact, it has infinitely many of them. However, it has a
limited number of solutions in which p, n, and d are non-negative integers. Running
through d = 0, . . . , d = 8 shows that (p, n, d) = (3, 4, 6) is the only solution using
natural numbers.
One.I.1.36 Solving the system
(1/3)(a + b + c) + d = 29
(1/3)(b + c + d) + a = 23
(1/3)(c + d + a) + b = 21
(1/3)(d + a + b) + c = 17
we obtain a = 12, b = 9, c = 3, d = 21. Thus the second item, 21, is the correct
answer.
One.I.1.37 This is how the answer was given in the cited source. A comparison of
the units and hundreds columns of this addition shows that there must be a carry
from the tens column. The tens column then tells us that A < H, so there can be no
carry from the units or hundreds columns. The five columns then give the following
five equations.
A+E=W
2H = A + 10
H=W+1
H + T = E + 10
A+1=T
The five linear equations in five unknowns, if solved simultaneously, produce the
unique solution: A = 4, T = 5, H = 7, W = 6 and E = 2, so that the original example
in addition was 47474 + 5272 = 52746.
8 Linear Algebra, by Hefferon
One.I.1.38 This is how the answer was given in the cited source. Eight commissioners
voted for B. To see this, we will use the given information to study how many voters
chose each order of A, B, C.
The six orders of preference are ABC, ACB, BAC, BCA, CAB, CBA; assume they
receive a, b, c, d, e, f votes respectively. We know that
a + b + e = 11
d + e + f = 12
a + c + d = 14
from the number preferring A over B, the number preferring C over A, and the number
preferring B over C. Because 20 votes were cast, we also know that
c+d+ f=9
a+b+c=8
b+ e+ f=6
from the preferences for B over A, for A over C, and for C over B.
The solution is a = 6, b = 1, c = 1, d = 7, e = 4, and f = 1. The number of
commissioners voting for B as their first choice is therefore c + d = 1 + 7 = 8.
Comments. The answer to this question would have been the same had we known
only that at least 14 commissioners preferred B over C.
The seemingly paradoxical nature of the commissioner’s preferences (A is preferred
to B, and B is preferred to C, and C is preferred to A), an example of “non-transitive
dominance”, is not uncommon when individual choices are pooled.
One.I.1.39 This is how the answer was given in the cited source. We have not used
“dependent” yet; it means here that Gauss’s Method shows that there is not a
unique solution. If n > 3 the system is dependent and the solution is not unique.
Hence n < 3. But the term “system” implies n > 1. Hence n = 2. If the equations are
ax + (a + d)y = a + 2d
(a + 3d)x + (a + 4d)y = a + 5d
then x = −1, y = 2.
0 0 0 0 0
ends with z and w free. We
have
this solution
set.
0 −1 −1
0 0 −1
{ + z + w | z, w ∈ R}
0 1 0
0 0 1
(d) Gauss’s Method done in this way
! !
1 2 3 1 −1 1 −3ρ1 +ρ2 1 2 3 1 −1 1
−→
3 −1 1 1 1 3 0 −7 −8 −2 4 0
ends with c, d, and e free. Solving for b shows that b = (8c + 2d − 4e)/(−7)
and then substitution a + 2(8c + 2d − 4e)/(−7) + 3c + 1d − 1e = 1 shows that
a = 1 − (5/7)c − (3/7)d − (1/7)e and we have the solution set.
1 −5/7 −3/7 −1/7
0 −8/7 −2/7 4/7
{ 0 + 1 c + 0 d + 0 e | c, d, e ∈ R}
0 0 1 0
0 0 0 1
12 Linear Algebra, by Hefferon
One.I.2.20 For each problem we get a system of linear equations by looking at the
equations of components.
(a) k = 5
(b) The second components show that i = 2, the third components show that j = 1.
(c) m = −4, n = 2
One.I.2.21 For each problem we get a system of linear equations by looking at the
equations of components.
(a) Yes; take k = −1/2.
(b) No; the system with equations 5 = 5 · j and 4 = −4 · j has no solution.
(c) Yes; take r = 2.
(d) No. The second components give k = 0. Then the third components give j = 1.
But the first components don’t check.
One.I.2.22 (a) Let c be the number of acres of corn, s be the number of acres of soy,
and a be the number of acres of oats.
c + s + a = 1200 −20ρ1 +ρ2 c + s + a = 1200
−→
20c + 50s + 12a = 40 000 30s − 8a = 16 000
To describe the solution set we can parametrize using a.
c 20 000/30 −38/30
{ s = 16 000/30 + 8/30 a | a ∈ R}
a 0 1
(b) There are many answers possible here. For instance we can take a = 0 to get
c = 20 000/30 ≈ 666.66 and s = 16000/30 ≈ 533.33. Another example is to take
a = 20 000/38 ≈ 526.32, giving c = 0 and s = 7360/38 ≈ 193.68.
(c) Plug your answers from the prior part into 100c + 300s + 80a.
One.I.2.23 This system has one equation. The leading variable is x1 , the other variables
are free.
−1 −1
1 0
{ . x2 + · · · + .
xn | x2 , . . . , xn ∈ R }
.. ..
0 1
One.I.2.24 (a) Gauss’s Method here gives
1 2 0 −1 a 1 2 0 −1 a
−2ρ1 +ρ2
2 0 1 0 b −ρ−→ 0 −4 1 2 −2a + b
1 +ρ3
1 1 0 2 c 0 −1 0 3 −a + c
1 2 0 −1 a
−(1/4)ρ2 +ρ3
−→ 0 −4 1 2 −2a + b
0 0 −1/4 5/2 −(1/2)a − (1/4)b + c
Answers to Exercises 13
One.I.2.29 On plugging in the five pairs (x, y) we get a system with the five equations
and six unknowns a, . . . , f. Because there are more unknowns than equations, if no
inconsistency exists among the equations then there are infinitely many solutions (at
least one variable will end up free).
But no inconsistency can exist because a = 0, . . . , f = 0 is a solution (we are only
using this zero solution to show that the system is consistent — the prior paragraph
shows that there are nonzero solutions).
14 Linear Algebra, by Hefferon
One.I.2.30 (a) Here is one — the fourth equation is redundant but still OK.
x+y− z+ w=0
y− z =0
2z + 2w = 0
z+ w=0
(b) Here is one.
x+y−z+w=0
w=0
w=0
w=0
(c) This is one.
x+y−z+w=0
x+y−z+w=0
x+y−z+w=0
x+y−z+w=0
One.I.2.31 This is how the answer was given in the cited source. My solution was
to define the numbers of arbuzoids as 3-dimensional vectors, and express all possible
elementary transitions as such vectors, too:
R: 13 −1 −1 2
G: 15 Operations: −1, 2 , and −1
B: 17 2 −1 −1
Now, it is enough to check whether the solution to one of the following systems of
linear equations
exists:
13 −1 −1 2 0 0 45
15 + x −1 + y 2 + −1 = 0 (or 45 or 0 )
17 2 −1 −1 45 0 0
Solving
−1 −1 2 −13 −1 −1 2 −13
−ρ1 +ρ2 ρ2 +ρ3
−1 2 −1 −15 −→ −→ 0 3 −3 −2
2ρ1 +ρ3
2 −1 −1 28 0 0 0 0
gives y + 2/3 = z so if the number of transformations z is an integer then y is not.
The other two systems give similar conclusions so there is no solution.
One.I.2.32 This is how the answer was given in the cited source.
(a) Formal solution of the system yields
a3 − 1 −a2 + a
x= 2 y= 2 .
a −1 a −1
If a + 1 6= 0 and a − 1 6= 0, then the system has the single solution
a2 + a + 1 −a
x= y= .
a+1 a+1
Answers to Exercises 15
If a = −1, or if a = +1, then the formulas are meaningless; in the first instance we
arrive at the system
−x + y = 1
x−y=1
which is a contradictory system. In the second instance we have
x+y=1
x+y=1
which has an infinite number of solutions (for example, for x arbitrary, y = 1 − x).
(b) Solution of the system yields
a4 − 1 −a3 + a
x= 2 y= 2 .
a −1 a −1
Here, is a2 − 1 6= 0, the system has the single solution x = a2 + 1, y = −a. For
a = −1 and a = 1, we obtain the systems
−x + y = −1 x+y=1
x−y= 1 x+y=1
both of which have an infinite number of solutions.
One.I.2.33 This is how the answer was given in the cited source. Let u, v, x, y, z be
the volumes in cm3 of Al, Cu, Pb, Ag, and Au, respectively, contained in the sphere,
which we assume to be not hollow. Since the loss of weight in water (specific gravity
1.00) is 1000 grams, the volume of the sphere is 1000 cm3 . Then the data, some of
which is superfluous, though consistent, leads to only 2 independent equations, one
relating volumes and the other, weights.
u+ v+ x+ y+ z = 1000
2.7u + 8.9v + 11.3x + 10.5y + 19.3z = 7558
Clearly the sphere must contain some aluminum to bring its mean specific gravity
below the specific gravities of all the other metals. There is no unique result to
this part of the problem, for the amounts of three metals may be chosen arbitrarily,
provided that the choices will not result in negative amounts of any metal.
If the ball contains only aluminum and gold, there are 294.5 cm3 of gold and
705.5 cm3 of aluminum. Another possibility is 124.7 cm3 each of Cu, Au, Pb, and Ag
and 501.2 cm3 of Al.
(c) Yes.
One.I.3.17 Gauss’s Method on the associated homogeneous system
1 −1 0 1 0 1 −1 0 1 0
−2ρ1 +ρ2
2 3 −1 0 0 −→ 0 5 −1 −2 0
0 1 1 1 0 0 1 1 1 0
1 −1 0 1 0
−(1/5)ρ2 +ρ3
−→ 0 5 −1 −2 0
0 0 6/5 7/5 0
gives this is the solution to the homogeneous problem.
−5/6
1/6
{ w | w ∈ R}
−7/6
1
(a) That vector is indeed a particular solution, so the required general solution is
this.
0 −5/6
0 1/6
{ + w | w ∈ R}
0 −7/6
4 1
(b) That vector is a particular solution so the required general solution is this.
−5 −5/6
1 1/6
{ + w | w ∈ R}
−7 −7/6
10 1
(c) That vector is not a solution of the system since it does not satisfy the third
equation. No such general solution exists.
One.I.3.18 The first is nonsingular while the second is singular. Just do Gauss’s Method
and see if the echelon form result has non-0 numbers in each entry on the diagonal.
One.I.3.19 (a) Nonsingular: !
−ρ1 +ρ2 1 2
−→
0 1
ends with each row containing a leading entry.
(b) Singular: !
3ρ1 +ρ2 1 2
−→
0 0
ends with row 2 without a leading entry.
20 Linear Algebra, by Hefferon
Linear Geometry
2 −1/2 −1/2
P = { 0 + y · 1 + z · 0 | y, z ∈ R }
0 0 1
instead it is
2 −0.5 −0.5 1
0 + 1 · 1 + 0 · 1 = 1
0 0 1 1
which adds the parameters.
One.II.1.9 The “if” half is straightforward. If b1 − a1 = d1 − c1 and b2 − a2 = d2 − c2
then q q
(b1 − a1 )2 + (b2 − a2 )2 = (d1 − c1 )2 + (d2 − c2 )2
so they have the same lengths, and the slopes are just as easy:
b2 − a2 d2 − c2
=
b1 − a1 d1 − a1
(if the denominators are 0 they both have undefined slopes).
For “only if”, assume that the two segments have the same length and slope (the
case of undefined slopes is easy; we will do the case where both segments have a slope
m). Also assume, without loss of generality, that a1 < b1 and that c1 < d1 . The first
segment is (a1 , a2 )(b1 , b2 ) = {(x, y) | y = mx + n1 , x ∈ [a1 ..b1 ] } (for some intercept
n1 ) and the second segment is (c1 , c2 )(d1 , d2 ) = { (x, y) | y = mx + n2 , x ∈ [c1 ..d1 ] }
(for some n2 ). Then the lengths of those segments are
q q
(b1 − a1 )2 + ((mb1 + n1 ) − (ma1 + n1 ))2 = (1 + m2 )(b1 − a1 )2
and, similarly, (1 + m2 )(d1 − c1 )2 . Therefore, |b1 − a1 | = |d1 − c1 |. Thus, as we
p
~ ∈ Rn have components
One.II.2.18 In each item below, assume that the vectors ~u,~v, w
u 1 , . . . , u n , v 1 , . . . , wn .
Answers to Exercises 27
un vn wn
u 1 + v1 w1
.. • ..
=
. .
u n + vn wn
= (u1 + v1 )w1 + · · · + (un + vn )wn
= (u1 w1 + · · · + un wn ) + (v1 w1 + · · · + vn wn )
= ~u • w
~ + ~v • w
~
(b) Dot product is also left distributive: w
~ • (~u + ~v) = w~ • ~u + w
~ • ~v. The proof is
just like the prior one.
(c) Dot product commutes.
u1 v1 v1 u1
.. • .. .. • ..
. . = u 1 v1 + · · · + u n vn = v1 u 1 + · · · + vn u n = . .
un vn vn un
(d) Because ~u ~v is a scalar, not a vector, the expression (~u ~v) w
• • ~ makes no sense;
•
But that’s equivalent to the assertion that one of the two vectors ~u and ~v is a scalar
multiple of the other, as desired.
One.II.2.20 No. These give an example.
! ! !
1 1 1
~u = ~v = ~ =
w
0 0 1
One.II.2.21 We prove that a vector has length zero if and only if all its components are
zero.
Let ~u ∈ Rn have components u1 , . . . , un . Recall that the square of any real
number is greater than or equal to zero, with equality only when that real is zero.
Thus |~u |2 = u1 2 + · · · + un 2 is a sum of numbers greater than or equal to zero, and
so is itself greater than or equal to zero, with equality if and only if each ui is zero.
Hence |~u | = 0 if and only if all the components of ~u are zero.
One.II.2.22 We can easily check that
x1 + x2 y1 + y2
,
2 2
is on the line connecting the two, and is equidistant from both. The generalization is
obvious.
One.II.2.23 Assume that ~v ∈ Rn has components v1 , . . . , vn . If ~v 6= ~0 then we have this.
v
u !2 !2
u v1 vn
t p + ··· + p
v1 2 + · · · + vn 2 v1 2 + · · · + vn 2
s
v1 2 vn 2
= + · · · +
v1 2 + · · · + vn 2 v1 2 + · · · + vn 2
=1
If ~v = ~0 then ~v/|~v | is not defined.
One.II.2.24 For the first question, assume that ~v ∈ Rn and r > 0, take the root, and
factor. q q
|r~v | = (rv1 )2 + · · · + (rvn )2 = r2 (v1 2 + · · · + vn 2 = r|~v |
For the second question, the result is r times as long, but it points in the opposite
direction in that r~v + (−r)~v = ~0.
One.II.2.25 Assume that ~u,~v ∈ Rn both have length 1. Apply Cauchy-Schwarz: |~u • ~v| 6
|~u | |~v | = 1.
To see that ‘less than’ can happen, in R2 take
! !
1 0
~u = ~v =
0 1
and note that ~u • ~v = 0. For ‘equal to’, note that ~u • ~u = 1.
Answers to Exercises 29
One.II.2.26 Write
u1 v1
. .
~u = .. ~v = ..
un vn
and then this computation works.
|~u + ~v |2 + |~u − ~v |2 = (u1 + v1 )2 + · · · + (un + vn )2
+ (u1 − v1 )2 + · · · + (un − vn )2
= u1 2 + 2u1 v1 + v1 2 + · · · + un 2 + 2un vn + vn 2
+ u1 2 − 2u1 v1 + v1 2 + · · · + un 2 − 2un vn + vn 2
= 2(u1 2 + · · · + un 2 ) + 2(v1 2 + · · · + vn 2 )
= 2|~u |2 + 2|~v |2
One.II.2.27 We will prove this demonstrating that the contrapositive statement holds: if
~x 6= ~0 then there is a ~y with ~x • ~y 6= 0.
Assume that ~x ∈ Rn . If ~x 6= ~0 then it has a nonzero component, say the i-th one
xi . But the vector ~y ∈ Rn that is all zeroes except for a one in component i gives
~x • ~y = xi . (A slicker proof just considers ~x • ~x.)
One.II.2.28 Yes; we can prove this by induction.
Assume that the vectors are in some Rk . Clearly the statement applies to one
vector. The Triangle Inequality is this statement applied to two vectors. For an
inductive step assume the statement is true for n or fewer vectors. Then this
|~u1 + · · · + ~un + ~un+1 | 6 |~u1 + · · · + ~un | + |~un+1 |
follows by the Triangle Inequality for two vectors. Now the inductive hypothesis,
applied to the first summand on the right, gives that as less than or equal to |~u1 | +
· · · + |~un | + |~un+1 |.
One.II.2.29 By definition
~u • ~v
= cos θ
|~u | |~v |
where θ is the angle between the vectors. Thus the ratio is |cos θ|.
One.II.2.30 So that the statement ‘vectors are orthogonal iff their dot product is zero’
has no exceptions.
One.II.2.31 We can find the angle between (a) and (b) (for a, b 6= 0) with
ab
arccos( √ √ ).
a2 b2
If a or b is zero then the angle is π/2 radians. Otherwise, if a and b are of opposite
signs then the angle is π radians, else the angle is zero radians.
30 Linear Algebra, by Hefferon
000 0
0
gives 00 0
0
*
0
(we ignore the case where ~z1 and ~z2 are the zero vector).
The ~z1 + ~z2 = ~0 case is easy. For the rest, by the definition of angle, we will be
finished if we show this.
~z1 • (~z1 + ~z2 ) ~z2 • (~z1 + ~z2 )
=
|~z1 | |~z1 + ~z2 | |~z2 | |~z1 + ~z2 |
But distributing inside each expression gives
~z1 • ~z1 + ~z1 • ~z2 ~z2 • ~z1 + ~z2 • ~z2
|~z1 | |~z1 + ~z2 | |~z2 | |~z1 + ~z2 |
and ~z1 • ~z1 = |~z1 |2 = |~z2 |2 = ~z2 • ~z2 , so the two are equal.
One.II.2.37 We can show the two statements together. Let ~u,~v ∈ Rn , write
u1 v1
.. ..
~u = . ~v = .
un vn
and calculate.
ku1 v1 + · · · + kun vn k ~u · ~v ~u • ~v
cos θ = q = =±
p |k| |~
u | |~
v | |~
u | |~v |
(ku1 )2 + · · · + (kun )2 b1 2 + · · · + bn 2
Answers to Exercises 31
One.II.2.38 Let
u1 v1 w1
. . .
~u = .. , ~v = .. ~ = ..
w
un vn wn
and then
u1 kv1 mw1
. . .
~u • w = .. • .. + ..
k~v + m~
un kvn mwn
u1 kv1 + mw1
. ..
= .. •
.
un kvn + mwn
= u1 (kv1 + mw1 ) + · · · + un (kvn + mwn )
= ku1 v1 + mu1 w1 + · · · + kun vn + mun wn
= (ku1 v1 + · · · + kun vn ) + (mu1 w1 + · · · + mun wn )
= k(~u • ~v) + m(~u • w
~)
as required.
One.II.2.39 For x, y ∈ R+ , set
√ ! √ !
x y
~u = √ ~v = √
y x
so that the Cauchy-Schwarz inequality asserts that (after squaring)
√ √ √ √ √ √ √ √ √ √ √ √
( x y + y x)2 6 ( x x + y y)( y y + x x)
√ √
(2 x y)2 6 (x + y)2
√ x+y
xy 6
2
as desired.
One.II.2.40 (a) For instance, a birthday of October 12 gives this.
! !
7 10
•
12 12 214
θ = arccos( ! ! ) = arccos( √ √ ) ≈ 0.17 rad
7 10 244 193
| |·| |
12 12
(b) Applying the same equation to (9 19) gives about 0.09 radians.
(c) The angle will measure 0 radians if the other person is born on the same day. It
will also measure 0 if one birthday is a scalar multiple of the other. For instance, a
person born on Mar 6 would be harmonious with a person born on Feb 4.
32 Linear Algebra, by Hefferon
Given a birthday, we can get Sage to plot the angle for other dates. This
example shows the relationship of all dates with July 12.
sage: plot3d(lambda x, y: math.acos((x*7+y*12)/(math.sqrt(7**2+12**2)*math.sqrt(x**2+y**2))),
(1,12),(1,31))
The result is
For (7, 12) worst case 0.95958064648 rads, date (12, 1)
That is 54.9799211457 degrees
Answers to Exercises 33
A more conceptual approach is to consider the relation of all points (month, day)
to the point (7, 12). The picture below makes clear that the answer is either Dec 1
or Jan 31, depending on which is further from the birthdate. The dashed line
bisects the angle between the line from the origin to Dec 1, and the line from the
origin to Jan 31. Birthdays above the line are furthest from Dec 1 and birthdays
below the line are furthest from Jan 31.
30
20
10
J F MAM J J A S O N D
One.II.2.41 This is how the answer was given in the cited source. The actual velocity
~v of the wind is the sum of the ship’s velocity and the apparent velocity of the wind.
Without loss of generality we may assume a ~ and ~b to be unit vectors, and may write
a = ~v2 + t~b
~v = ~v1 + s~
where s and t are undetermined scalars. Take the dot product first by a
~ and then by
34 Linear Algebra, by Hefferon
~b to obtain
s − t~a • ~b = a
~ • (~v2 − ~v1 )
a • ~b − t = ~b • (~v2 − ~v1 )
s~
(d) A row swap in the second step makes the arithmetic easier.
2 −1 0 −1 2 −1 0 −1
−(1/2)ρ1 +ρ2
1 3 −1 5 −→ 0 7/2 −1 11/2
0 1 2 5 0 1 2 5
2 −1 0 −1 2 −1 0 −1
ρ2 ↔ρ3 −(7/2)ρ2 +ρ3
−→ 0 1 2 5 −→ 0 1 2 5
0 7/2 −1 11/2 0 0 −8 −12
1 −1/2 0 −1/2 1 −1/2 0 −1/2
(1/2)ρ1 −2ρ3 +ρ2
−→ 0 1 2 5 −→ 0 1 0 2
−(1/8)ρ2
0 0 1 3/2 0 0 1 3/2
1 0 0 1/2
(1/2)ρ2 +ρ1
−→ 0 1 0 2
0 0 1 3/2
(b) As in the prior problem, the reduced echelon form is all zeroes but for a diagonal
of ones.
1 3 1 1 3 1
−2ρ1 +ρ2 −(1/6)ρ2
−→ 0 −6 2 −→ 0 1 −1/3
ρ1 +ρ3 −(1/2)ρ3
0 0 −2 0 0 1
1 3 0 1 0 0
(1/3)ρ3 +ρ2 −3ρ2 +ρ1
−→ 0 1 0 −→ 0 1 0
−ρ3 +ρ1
0 0 1 0 0 1
(c) There are more columns than rows so we must get more than just a diagonal of
ones.
1 0 3 1 2 1 0 3 1 2
−ρ1 +ρ2 −ρ2 +ρ3
−→ 0 4 −1 0 3 −→ 0 4 −1 0 3
−3ρ1 +ρ3
0 4 −1 −2 −4 0 0 0 −2 −7
1 0 3 1 2 1 0 3 0 −3/2
(1/4)ρ2 −ρ3 +ρ1
−→ 0 1 −1/4 0 3/4 −→ 0 1 −1/4 0 3/4
−(1/2)ρ3
0 0 0 1 7/2 0 0 0 1 7/2
Answers to Exercises 37
(of course, we could omit the zero vector from the description).
(d) The “Jordan” half
!
−(1/7)ρ2 1 2 3 1 −1 1
−→
0 1 8/7 2/7 −4/7 0
!
−2ρ2 +ρ1 1 0 5/7 3/7 1/7 1
−→
0 1 8/7 2/7 −4/7 0
ends with this solution set.
1 −5/7 −3/7 −1/7
0 −8/7 −2/7 4/7
{ 0 + 1 c + 0 d + 0 e | c, d, e ∈ R }
0 0 1 0
0 0 0 1
One.III.1.11 Routine Gauss’s Method gives one:
2 1 1 3 2 1 1 3
−3ρ1 +ρ2 −(9/2)ρ2 +ρ3
−→ 0 1 −2 −7 −→ 0 1 −2 −7
−(1/2)ρ1 +ρ3
0 9/2 1/2 7/2 0 0 19/2 35
and any cosmetic change, such as multiplying the bottom row by 2,
2 1 1 3
0 1 −2 −7
0 0 19 70
gives another.
One.III.1.12 In the cases listed below, we take a, b ∈ R. Thus, some canonical forms
listed below actually include infinitely many cases. In particular, they includes the
cases a = 0 !and b = 0.
! ! !
0 0 1 a 0 1 1 0
(a) , , ,
0 0 0 0 0 0 0 1
! ! ! ! ! !
0 0 0 1 a b 0 1 a 0 0 1 1 0 a 1 a 0
(b) , , , , , ,
0 0 0 0 0 0 0 0 0 0 0 0 0 1 b 0 0 1
!
0 1 0
0 0 1
0 0 1 a 0 1 1 0
(c) 0 0, 0 0 , 0 0, 0 1
0 0 0 0 0 0 0 0
0 0 0 1 a b 0 1 a 0 1 0 0 0 1 1 0 a
(d) 0 0 0, 0 0 0 , 0 0 0 , 0 0 1, 0 0 0, 0 1 b,
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
Answers to Exercises 39
1 a 0 1 0 0
0 0 1, 0 1 0
0 0 0 0 0 1
One.III.1.13 A nonsingular homogeneous linear system has a unique solution. So a
nonsingular matrix must reduce to a (square) matrix that is all 0’s except for 1’s down
the upper-left to lower-right diagonal, such as these.
! 1 0 0
1 0
0 1 0
0 1
0 0 1
One.III.1.14 It is an equivalence relation. To prove that we must check that the relation
is reflexive, symmetric, and transitive.
Assume that all matrices are 2×2. For reflexive, we note that a matrix has the
same sum of entries as itself. For symmetric, we assume A has the same sum of
entries as B and obviously then B has the same sum of entries as A. Transitivity is no
harder — if A has the same sum of entries as B and B has the same sum of entries as
C then A has the same as C.
One.III.1.15 (a) For instance,
! ! !
1 2 −ρ1 +ρ1 0 0 ρ1 +ρ1 0 0
−→ −→
3 4 3 4 3 4
leaves the matrix changed.
(b) This operation
.. ..
. .
i,1 · · · ai,n ···
a ai,1 ai,n
. kρi +ρj ..
. −→
. .
aj,1 · · · aj,n · · · kai,n + aj,n
kai,1 + aj,1
.. ..
. .
leaves the i-th row unchanged because of the i 6= j restriction. Because the i-th row
is unchanged, this operation
..
.
ai,1 ··· ai,n
−kρi +ρj .
.
−→
.
−kai,1 + kai,1 + aj,1 · · · −kai,n + kai,n + aj,n
..
.
returns the j-th row to its original state.
40 Linear Algebra, by Hefferon
One.III.1.17 For each we must check the three conditions of reflexivity, symmetry, and
transitivity.
(a) Any matrix clearly has the same product down the diagonal as itself, so the
relation is reflexive. The relation is symmetric because if A has the same product
down its diagonal as does B, if a1,1 · a2,2 = b1,1 · b2,2 , then B has the same product
as does A.
Transitivity is similar: suppose that A’s product is r and that it equals B’s
product. Suppose also that B’s product equals C’s. Then all three have a product
of r, and A’s equals C’s.
There is an equivalence class for each real number, namely the class contains all
2×2 matrices whose product down the diagonal is that real.
(b) For reflexivity, if the matrix A has a 1 entry then it is related to itself while if it
does not then it is also related to itself. Symmetry also has two cases: suppose that
A and B are related. If A has a 1 entry then so does B, and thus B is related to A.
If A has no 1 then neither does B, and again B is related to A.
For transitivity, suppose that A is related to B and B to C. If A has a 1 entry
then so does B, and because B is related to C, therefore so does C, and hence A is
related to C. Likewise, if A has no 1 then neither does B, and consequently neither
does C, giving the conclusion that A is related to C.
There are exactly two equivalence classes, one containing any 2×2 matrix that
has at least one entry that is a 1, and the other containing all the matrices that
have no 1’s.
One.III.1.18 (a) This relation is not reflexive. For instance, any matrix with an upper-
left entry of 1 is not related to itself.
(b) This relation is not transitive. For these three, A is related to B, and B is related
to C, but A is not related to C.
! ! !
0 0 4 0 8 0
A= , B= , C= ,
0 0 0 0 0 0
Answers to Exercises 41
One.III.2.11 First, the only matrix row equivalent to the matrix of all 0’s is itself (since
row operations have no effect).
Second, the matrices that reduce to
!
1 a
0 0
have the form !
b ba
c ca
(where a, b, c ∈ R, and b and c are not both zero).
Next, the matrices that reduce to
!
0 1
0 0
have the form !
0 a
0 b
(where a, b ∈ R, and not both are zero).
Finally, the matrices that reduce to
!
1 0
0 1
are the nonsingular matrices. That’s because a linear system for which this is the matrix
of coefficients will have a unique solution, and that is the definition of nonsingular.
(Another way to say the same thing is to say that they fall into none of the above
classes.)
One.III.2.12 (a) They have the form
!
a 0
b 0
where a, b ∈ R.
(b) They have this form (for a, b ∈ R).
!
1a 2a
1b 2b
(c) They have the form !
a b
c d
(for a, b, c, d ∈ R) where ad − bc 6= 0. (This is the formula that determines when a
2×2 matrix is nonsingular.)
Answers to Exercises 43
Other Computer Algebra Systems have similar commands. These Maple commands
> A:=array( [[40,15],
[-50,25]] );
> u:=array([100,50]);
> linsolve(A,u);
A Maple session
> A:=array( [[2,2],
[1,-4]] );
> u:=array([5,0]);
> linsolve(A,u);
sage: var('x,y')
(x, y)
sage: system = [-1*x + y == 1,
....: x + y == 2]
sage: solve(system, x,y)
[[x == (1/2), y == (3/2)]]
(c) This system has infinitely many solutions. In the first subsection, with z as a
parameter, we got x = (43 − 7z)/4 and y = (13 − z)/4. Sage gets the same.
sage: var('x,y,z')
(x, y, z)
sage: system = [x - 3*y + z == 1,
....: x + y + 2*z == 14]
sage: solve(system, x,y)
[[x == -7/4*z + 43/4, y == -1/4*z + 13/4]]
Similarly, When the array A and vector u are given to Maple and it is asked to
linsolve(A,u), it returns no result at all; that is, it responds with no solutions.
2 The reduction
−3ρ1 +ρ2 x + 2y = 3
−→
−8y = −7.992
gives (x, y) = (1.002, 0.999). So for this system a small change in the constant produces
only a small change in the solution.
4 (a) For the first one, first, (2/3) − (1/3) is .666 666 67 − .333 333 33 = .333 333 34
and so (2/3) + ((2/3) − (1/3)) = .666 666 67 + .333 333 34 = 1.000 000 0.
For the other one, first ((2/3) + (2/3)) = .666 666 67 + .666 666 67 = 1.333 333 3
and so ((2/3) + (2/3)) − (1/3) = 1.333 333 3 − .333 333 33 = .999 999 97.
(b) The first equation is .333 333 33 · x + 1.000 000 0 · y = 0 while the second is
.666 666 67 · x + 2.000 000 0 · y = 0.
1 (a) The total resistance is 7 ohms. With a 9 volt potential, the flow will be
9/7 amperes. Incidentally, the voltage drops will then be: 27/7 volts across the
3 ohm resistor, and 18/7 volts across each of the two 2 ohm resistors.
(b) One way to do this network is to note that the 2 ohm resistor on the left has
a voltage drop of 9 volts (and hence the flow through it is 9/2 amperes), and the
remaining portion on the right also has a voltage drop of 9 volts, and so we can
analyze it as in the prior item. We can also use linear systems.
−→ −→
i0 i2
i1 ↓
i3
←−
Kirchoff’s Voltage law, applied to the loop in the top right, and to the loop in the
bottom right, gives these.
i3 r3 − ig rg − i1 r1 = 0
i4 r4 − i2 r2 + ig rg = 0
Assuming that ig is zero gives that i1 = i2 , that i3 = i4 , that i1 r1 = i3 r3 , and that
i2 r2 = i4 r4 . Then rearranging the last equality
i2 r2 i3 r3
r4 = ·
i4 i1 r1
and cancelling the i’s gives the desired conclusion.
4 (a) An adaptation is: in any intersection the flow in equals the flow out. It does
seem reasonable in this case, unless cars are stuck at an intersection for a long time.
(b) We can label the flow in this way.
Shelburne St
Willow Jay Ln
west
east
Winooski Ave
Because 50 cars leave via Main while 25 cars enter, i1 − 25 = i2 . Similarly Pier’s
in/out balance means that i2 = i3 and North gives i3 + 25 = i1 . We have this
system.
i1 − i2 = 25
i2 − i3 = 0
−i1 + i3 = −25
(c) The row operations ρ1 + ρ2 and rho2 + ρ3 lead to the conclusion that there are
infinitely many solutions. With i3 as the parameter,
25 + i3
{ i3 | i3 ∈ R}
i3
of course, since the problem is stated in number of cars, we might restrict i3 to be
a natural number.
(d) If we picture an initially-empty circle with the given input/output behavior, we
can superimpose a z3 -many cars circling endlessly to get a new solution.
(e) A suitable restatement might be: the number of cars entering the circle must
equal the number of cars leaving. The reasonableness of this one is not as clear.
Over the five minute time period we could find that a half dozen more cars entered
than left, although the problem statement’s into/out table does satisfy this property.
In any event, it is of no help in getting a unique solution since for that we would
need to know the number of cars circling endlessly.
Answers to Exercises 53
5 (a) Here is a variable for each unknown block; each known block has the flow shown.
65 55
i1
75 40
i2 i3
i4
5 50
80 30
i5 i7 i6
70
We apply Kirchhoff’s principle that the flow into the intersection of Willow and
Shelburne must equal the flow out to get i1 + 25 = i2 + 125. Doing the intersections
from right to left and top to bottom gives these equations.
i1 − i2 = 10
−i1 + i3 = 15
i2 + i4 = 5
−i3 − i4 + i6 = −50
i5 − i7 = −10
−i6 + i7 = 30
The row operation ρ1 + ρ2 followed by ρ2 + ρ3 then ρ3 + ρ4 and ρ4 + ρ5 and finally
ρ5 + ρ6 result in this system.
i1 − i2 = 10
−i2 + i3 = 25
i3 + i4 − i5 = 30
−i5 + i6 = −20
−i6 + i7 = −30
0= 0
Since the free variables are i4 and i7 we take them as parameters.
i6 = i7 − 30
i5 = i6 + 20 = (i7 − 30) + 20 = i7 − 10
i3 = −i4 + i5 + 30 = −i4 + (i7 − 10) + 30 = −i4 + i7 + 20
i2 = i3 − 25 = (−i4 + i7 + 20) − 25 = −i4 + i7 − 5
i1 = i2 + 10 = (−i4 + i7 − 5) + 10 = −i4 + i7 + 5
Obviously i4 and i7 have to be positive, and in fact the first equation shows
that i7 must be at least 30. If we start with i7 , then the i2 equation shows that
0 6 i4 6 i7 − 5.
(b) We cannot take i7 to be zero or else i6 will be negative (this would mean cars
going the wrong way on the one-way street Jay). We can, however, take i7 to be as
small as 30, and then there are many suitable i4 ’s. For instance, the solution
(i1 , i2 , i3 , i4 , i5 , i6 , i7 ) = (35, 25, 50, 0, 20, 0, 30)
results from choosing i4 = 0.
Chapter Two
Chapter Two: Vector
Spaces
We must also check conditions (6)-(10), those for scalar multiplication. For (6),
the condition that the space be closed under scalar multiplication, suppose that r is a
real number and a + bx ∈ P (so that a − 2b = 0), then r(a + bx) = (ra) + (rb)x is an
element of P because it is a linear polynomial with real number coefficients satisfying
that (ra) − 2(rb) = r(a − 2b) = 0. Condition (7) holds for the same reason that it
holds in the first item of this exercise, because (r+s)(a+bx) = r(a+bx) +s(a+bx)
is true from the distributive property for real number multiplication. Condition (8)
is also unchanged from the first item: r((a + bx) + (c + dx)) = r((a + c) + (b + d)x) =
r(a + c) + r(b + d)x = (ra + rc) + (rb + rd)x = r(a + bx) + r(c + dx). So
is (9): (rs)(a + bx) = (rsa) + (rsb)x = r(sa + sbx) = r(s(a + bx)). Finally,
so is condition (10): 1(a + bx) = (1a) + (1b)x = a + bx.
Two.I.1.20 Use Example 1.3 as a guide. (Comment. Because many of the conditions
are quite easy to check, sometimes a person can feel that they must have missed
something. Keep in mind that easy to do, or routine, is different from not necessary
to do.)
(a) Here are three elements.
! ! !
1 2 −1 −2 0 0
, ,
3 4 −3 −4 0 0
For (1), the sum of 2×2 real matrices is a 2×2 real matrix. For (2) we consider
the sum of two matrices
! ! !
a b e f a+e b+f
+ =
c d g h c+g d+h
and apply commutativity of real number addition
! ! !
e+a f+b e f a b
= = +
g+c h+d g h c d
to verify that the addition of the matrices is commutative. The verification for
condition (3), associativity of matrix addition, is similar to the prior verification:
! ! ! !
a b e f i j (a + e) + i (b + f) + j
+ + =
c d g h k l (c + g) + k (d + h) + l
while
! ! ! !
a b e f i j a + (e + i) b + (f + j)
+ + =
c d g h k l c + (g + k) d + (h + l)
and the two are the same entry-by-entry because real number addition is associative.
For (4), the zero element of this space is the 2×2 matrix of zeroes. Condition (5)
holds because for any 2×2 matrix A the additive inverse is the matrix whose entries
are the negative of A’s, the matrix −1 · A.
58 Linear Algebra, by Hefferon
Condition (6) holds because a scalar multiple of a 2×2 matrix is a 2×2 matrix.
For condition (7) we have this.
! !
a b (r + s)a (r + s)b
(r + s) =
c d (r + s)c (r + s)d
! ! !
ra + sa rb + sb a b a b
= =r +s
rc + sc rd + sd c d c d
Condition (8) goes the same way.
! ! ! !
a b e f a+e b+f ra + re rb + rf
r + =r =
c d g h c+g d+h rc + rg rd + rh
! ! ! !
a b e f a b e f
=r +r =r +
c d g h c d g h
For (9) we have this.
! ! ! !
a b rsa rsb sa sb a b
(rs) = =r =r s
c d rsc rsd sc sd c d
Condition (10) is just as easy.
! ! !
a b 1·a 1·b sa sb
1 = =
c d 1·c 1·d sc sd
(b) This differs from the prior item in this exercise only in that we are restricting to
the set T of matrices with a zero in the second row and first column. Here are three
elements of T . ! ! !
1 2 −1 −2 0 0
, ,
0 4 0 −4 0 0
Some of the verifications for this item are the same as for the first item in this
exercise, and below we’ll just do the ones that are different.
For (1), the sum of 2×2 real matrices with a zero in the 2, 1 entry is also a 2×2
real matrix with a zero in the!2, 1 entry. ! !
a b e f a+e b+f
+
0 d 0 h 0 d+h
The verification for condition (2) given in the prior item works in this item also.
The same holds for condition (3). For (4), note that the 2×2 matrix of zeroes is
an element of T . Condition (5) holds because for any 2×2 matrix A the additive
inverse is the matrix −1 · A and so the additive inverse of a matrix with a zero in
the 2, 1 entry is also a matrix with a zero in the 2, 1 entry.
Condition 6 holds because a scalar multiple of a 2×2 matrix with a zero in the
2, 1 entry is a 2×2 matrix with a zero in the 2, 1 entry. Condition (7)’s verification
is the same as in the prior item. So are condition (8)’s, (9)’s, and (10)’s.
Answers to Exercises 59
is also a member of L, which is true because it satisfies the criteria for membership
in L: (a + e) + (b + f) − (c + g) + (d + h) = (a + b − c + d) + (e + f − g + h) = 0 + 0.
The verifications for conditions (2), (3), and (5) are similar to the ones in the first
part of this exercise. For condition (4) note that the vector of zeroes is a member
of L because its first component plus its second, minus its third, and plus its fourth,
totals to zero.
Condition (6), closure of scalar multiplication, is similar: where the vector is an
element of L,
a ra
b rb
r =
c rc
d rd
is also an element of L because ra + rb − rc + rd = r(a + b − c + d) = r · 0 = 0.
The verification for conditions (7), (8), (9), and (10) are as in the prior item of this
exercise.
Two.I.1.22 In each item the set is called Q. For some items, there are other correct
ways to show that Q is not a vector space.
(a) It is not closed under addition; it fails to meet condition (1).
1 0 1
0 , 1 ∈ Q 1 6∈ Q
0 0 0
(b) It is not closed under addition.
1 0 1
0 , 1 ∈ Q 1 6∈ Q
0 0 0
(c) It is not closed under addition.
! ! !
0 1 1 1 1 2
, ∈Q 6∈ Q
0 0 0 0 0 0
(d) It is not closed under scalar multiplication.
1 + 1x + 1x2 ∈ Q − 1 · (1 + 1x + 1x2 ) 6∈ Q
(e) It is empty, violating condition (4).
Two.I.1.23 The usual operations (v0 + v1 i) + (w0 + w1 i) = (v0 + w0 ) + (v1 + w1 )i and
r(v0 + v1 i) = (rv0 ) + (rv1 )i suffice. The check is easy.
Two.I.1.24 No, it is not closed under scalar multiplication since, e.g., π · (1) is not a
rational number.
Answers to Exercises 61
(b) No; the same calculation as the prior answer shows a condition in the definition
of a vector space that is violated. Another example of a violation of the conditions
for a vector space is that 1 · (0, 1) 6= (0, 1).
Two.I.1.32 It is not a vector space since it is not closed under addition, as (x2 )+(1+x−x2 )
is not in the set.
Two.I.1.33 (a) 6
(b) nm
(c) 3
(d) To see that the answer is 2, rewrite it as
!
a 0
{ | a, b ∈ R }
b −a − b
so that there are two parameters.
Two.I.1.34 A vector space (over R) consists of a set V along with two operations ‘+ ~’
and ‘~·’ subject to these conditions. Where ~v, w ~ ∈ V, (1) their vector sum ~v + w ~ ~ is an
element of V. If ~u,~v, w ~ ∈ V then (2) ~v +~w ~ =w ~+~ ~v and (3) (~v +
~w ~ ~u = ~v +
~ )+ ~ (~w+ ~ ~u).
(4) There is a zero vector ~0 ∈ V such that ~v + ~ ~0 = ~v for all ~v ∈ V. (5) Each ~v ∈ V
has an additive inverse w ~ ∈ V such that w ~ +~ ~v = ~0. If r, s are scalars, that is,
members of R), and ~v, w ~ ∈ V then (6) each scalar multiple r · ~v is in V. If r, s ∈ R
and ~v, w~ ∈ V then (7) (r + s) · ~v = r · ~v + ~ s · ~v, and (8) r~· (~v + w~ ) = r~· ~v + r~· w
~ , and
(9) (rs)~· ~v = r~· (s~· ~v), and (10) 1~· ~v = ~v.
Two.I.1.35 (a) Let V be a vector space, assume that ~v ∈ V, and assume that w ~ ∈V
is the additive inverse of ~v so that w ~
~ + ~v = 0. Because addition is commutative,
~0 = w ~ + ~v = ~v + w~ , so therefore ~v is also the additive inverse of w~.
~
(b) Let V be a vector space and suppose ~v,~s, t ∈ V. The additive inverse of ~v is −~v
so ~v + ~s = ~v + ~t gives that −~v + ~v + ~s = −~v + ~v + ~t, which says that ~0 + ~s = ~0 + ~t
and so ~s = ~t.
Two.I.1.36 Addition is commutative, so in any vector space, for any vector ~v we have
that ~v = ~v + ~0 = ~0 + ~v.
Two.I.1.37 It is not a vector space since addition of two matrices of unequal sizes is not
defined, and thus the set fails to satisfy the closure condition.
Two.I.1.38 Each element of a vector space has one and only one additive inverse.
For, let V be a vector space and suppose that ~v ∈ V. If w ~ 2 ∈ V are both
~ 1, w
additive inverses of ~v then consider w ~ 1 + ~v + w
~ 2 . On the one hand, we have that
it equals w~ 1 + (~v + w ~ 1 + ~0 = w
~ 2) = w ~ 1 . On the other hand we have that it equals
w1 + ~v) + w
(~ ~ 2 = ~0 + w
~2 =w ~ 2 . Therefore, w~1 =w ~ 2.
Answers to Exercises 63
Two.I.1.39 (a) Every such set has the form { r · ~v + s · w ~ | r, s ∈ R } where either or
both of ~v, w~ may be ~0. With the inherited operations, closure of addition (r1~v +
~ ) + (r2~v + s2 w
s1 w w and scalar multiplication c(r~v + s~
~ ) = (r1 + r2 )~v + (s1 + s2 )~ w) =
(cr)~v + (cs)~ w are easy. The other conditions are also routine.
(b) No such set can be a vector space under the inherited operations because it does
not have a zero element.
Two.I.1.40 Assume that ~v ∈ V is not ~0.
(a) One direction of the if and only if is clear: if r = 0 then r · ~v = ~0. For the other
way, let r be a nonzero scalar. If r~v = ~0 then (1/r) · r~v = (1/r) · ~0 shows that ~v = ~0,
contrary to the assumption.
(b) Where r1 , r2 are scalars, r1~v = r2~v holds if and only if (r1 − r2 )~v = ~0. By the
prior item, then r1 − r2 = 0.
(c) A nontrivial space has a vector ~v 6= ~0. Consider the set { k · ~v | k ∈ R }. By the
prior item this set is infinite.
(d) The solution set is either trivial, or nontrivial. In the second case, it is infinite.
Two.I.1.41 Yes. A theorem of first semester calculus says that a sum of differentiable
functions is differentiable and that (f + g)0 = f0 + g0 , and that a multiple of a
differentiable function is differentiable and that (r · f)0 = r f0 .
Two.I.1.42 The check is routine. Note that ‘1’ is 1 + 0i and the zero elements are
these.
2
(a) (0 + 0i) + (0 + 0i)x
! + (0 + 0i)x
0 + 0i 0 + 0i
(b)
0 + 0i 0 + 0i
Two.I.1.43 Notably absent from the definition of a vector space is a distance measure.
Two.I.1.44 (a) A small rearrangement does the trick.
(~v1 + (~v2 + ~v3 )) + ~v4 = ((~v1 + ~v2 ) + ~v3 ) + ~v4
= (~v1 + ~v2 ) + (~v3 + ~v4 )
= ~v1 + (~v2 + (~v3 + ~v4 ))
= ~v1 + ((~v2 + ~v3 ) + ~v4 )
Each equality above follows from the associativity of three vectors that is given as
a condition in the definition of a vector space. For instance, the second ‘=’ applies
the rule (~ w1 + w ~ 2) + w
~3 =w ~ 1 + (~ ~ 3 ) by taking w
w2 + w ~ 1 to be ~v1 + ~v2 , taking w
~2
to be ~v3 , and taking w ~ 3 to be ~v4 .
(b) The base case for induction is the three vector case. This case ~v1 + (~v2 + ~v3 ) =
(~v1 + ~v2 ) + ~v3 is one of the conditions in the definition of a vector space.
For the inductive step, assume that any two sums of three vectors, any two
sums of four vectors, . . . , any two sums of k vectors are equal no matter how we
64 Linear Algebra, by Hefferon
parenthesize the sums. We will show that any sum of k + 1 vectors equals this one
((· · · ((~v1 + ~v2 ) + ~v3 ) + · · · ) + ~vk ) + ~vk+1 .
Any parenthesized sum has an outermost ‘+’. Assume that it lies between ~vm
and ~vm+1 so the sum looks like this.
(· · · ~v1 · · ·~vm · · · ) + (· · · ~vm+1 · · ·~vk+1 · · · )
The second half involves fewer than k + 1 additions, so by the inductive hypothesis
we can re-parenthesize it so that it reads left to right from the inside out, and in
particular, so that its outermost ‘+’ occurs right before ~vk+1 .
= (· · · ~v1 · · · ~vm · · · ) + ((· · · (~vm+1 + ~vm+2 ) + · · · + ~vk ) + ~vk+1 )
Apply the associativity of the sum of three things
= (( · · · ~v1 · · · ~vm · · · ) + ( · · · (~vm+1 + ~vm+2 ) + · · · ~vk )) + ~vk+1
and finish by applying the inductive hypothesis inside these outermost parenthesis.
Two.I.1.45 Let ~v be a member of R2 with components v1 and v2 . We can abbreviate
the condition that both components have the same sign or are 0 by v1 v2 > 0.
To show the set is closed under scalar multiplication, observe that the components
of r~v satisfy (rv1 )(rv2 ) = r2 (v1 v2 ) and r2 > 0 so r2 v1 v2 > 0.
To show the set is not closed under addition we need only produce one example.
The vector with components −1 and 0, when added to the vector with components 0
and 1 makes a vector with mixed-sign components of −1 and 1.
Two.I.1.46 (a) We outline the check of the conditions from Definition 1.1.
Additive closure holds because if a0 + a1 + a2 = 0 and b0 + b1 + b2 = 0 then
(a0 + a1 x + a2 x2 ) + (b0 + b1 x + b2 x2 ) = (a0 + b0 ) + (a1 + b1 )x + (a2 + b2 )x2
is in the set since (a0 + b0 ) + (a1 + b1 ) + (a2 + b2 ) = (a0 + a1 + a2 ) + (b0 + b1 + b2 )
is zero. The second through fifth conditions are easy.
Closure under scalar multiplication holds because if a0 + a1 + a2 = 0 then
r · (a0 + a1 x + a2 x2 ) = (ra0 ) + (ra1 )x + (ra2 )x2
is in the set as ra0 + ra1 + ra2 = r(a0 + a1 + a2 ) is zero. The remaining conditions
here are also easy.
(b) This is similar to the prior answer.
(c) Call the vector space V. We have two implications: left to right, if S is a subspace
then it is closed under linear combinations of pairs of vectors and, right to left, if a
nonempty subset is closed under linear combinations of pairs of vectors then it is
a subspace. The left to right implication is easy; we here sketch the other one by
assuming S is nonempty and closed, and checking the conditions of Definition 1.1.
First, to show closure under addition, if ~s1 ,~s2 ∈ S then ~s1 + ~s2 ∈ S as ~s1 + ~s2 =
1 · ~s1 + 1 · ~s2 . Second, for any ~s1 ,~s2 ∈ S, because addition is inherited from V,
the sum ~s1 + ~s2 in S equals the sum ~s1 + ~s2 in V and that equals the sum ~s2 + ~s1
Answers to Exercises 65
in V and that in turn equals the sum ~s2 + ~s1 in S. The argument for the third
condition is similar to that for the second. For the fourth, suppose that ~s is in the
nonempty set S and note that 0 · ~s = ~0 ∈ S; showing that the ~0 of V acts under
the inherited operations as the additive identity of S is easy. The fifth condition is
satisfied because for any ~s ∈ S closure under linear combinations shows that the
vector 0 · ~0 + (−1) · ~s is in S; showing that it is the additive inverse of ~s under the
inherited operations is routine.
The proofs for the remaining conditions are similar.
(d) The a = 2b − c gives that the set {(2b − c) + bx + cx3 | b, c ∈ R } equals the
set {b(2 + x) + c(−1 + x3 ) | b, c ∈ R }. So the subspace is the span of the set
{ 2 + x, −1 + x3 }.
(e) The set {a + bx + cx2 | a + 7b + 49c = 0 } can be parametrized as
{b(−7 + x) + c(−49 + x2 ) | b, c ∈ R }
and so has the spanning set { −7 + x, −49 + x2 }.
Two.I.2.26 (a) We canparametrize
in this
way
x 1 0
{ 0 | x, z ∈ R} = {x 0 + z 0 | x, z ∈ R }
z 0 1
giving this for a spanning set.
1 0
{ 0 , 0 }
0 1
(b) Here is a parametrization,
and the
associated spanning
set.
−2/3 −1/3 −2/3 −1/3
{ y 1 + z 0 | y, z ∈ R } { 1 , 0 }
0 1 0 1
1 −1/2
−2 0
(c) { , }
1 0
0 1
(d) Parametrize the description as { −a1 + a1 x + a3 x2 + a3 x3 | a1 , a3 ∈ R} to get
{ −1 + x, x2 + x3 }.
(e) {1, x, x2!, x3 , x4 } ! ! !
1 0 0 1 0 0 0 0
(f) { , , , }
0 0 0 0 1 0 0 1
Two.I.2.27 Technically, no. Subspaces of R3 are sets of three-tall vectors, while R2 is a
2 3
set of two-tall vectors. Clearly though, R is “just like” this subspace of R .
x
{ y | x, y ∈ R }
0
Two.I.2.28 Of course, the addition and scalar multiplication operations are the ones
inherited from the enclosing space.
(a) This is a subspace. It is not empty as it contains at least the two example
functions given. It is closed because if f1 , f2 are even and c1 , c2 are scalars then we
have this.
(c1 f1 + c2 f2 ) (−x) = c1 f1 (−x) + c2 f2 (−x) = c1 f1 (x) + c2 f2 (x) = (c1 f1 + c2 f2 ) (x)
Answers to Exercises 69
(b) This is also a subspace; the check is similar to the prior one.
Two.I.2.29 It can be improper. If ~v = ~0 then this is a trivial subspace. At the opposite
extreme, if the vector space is R1 and ~v 6= ~0 then the subspace is all of R1 .
Two.I.2.30 No, such a set is not closed. For one thing, it does not contain the zero
vector.
Two.I.2.31 (a) This nonempty subset of M2×2 is not a subspace.
! !
1 2 5 6
A={ , }
3 4 7 8
One reason that it is not a subspace of M2×2 is that it does not contain the zero
matrix. (Another reason is that it is not closed under addition, since the sum of
the two is not an element of A. It is also not closed under scalar multiplication.)
(b) This set of two vectors does not span R2 .
! !
1 3
{ , }
1 3
No linear combination of these two can give a vector whose second component is
unequal to its first component.
Two.I.2.32 No. The only subspaces of R1 are the space itself and its trivial subspace.
Any subspace S of R that contains a nonzero member ~v must contain the set of all of
its scalar multiples {r · ~v | r ∈ R }. But this set is all of R.
Two.I.2.33 Item (1) is checked in the text.
Item (2) has five conditions. First, for closure, if c ∈ R and ~s ∈ S then c · ~s ∈ S
as c · ~s = c · ~s + 0 · ~0. Second, because the operations in S are inherited from V, for
c, d ∈ R and ~s ∈ S, the scalar product (c + d) · ~s in S equals the product (c + d) · ~s
in V, and that equals c · ~s + d · ~s in V, which equals c · ~s + d · ~s in S.
The check for the third, fourth, and fifth conditions are similar to the second
condition’s check just given.
Two.I.2.34 An exercise in the prior subsection shows that every vector space has only
one zero vector (that is, there is only one vector that is the additive identity element
of the space). But a trivial space has only one element and that element must be this
(unique) zero vector.
Two.I.2.35 As the hint suggests, the basic reason is the Linear Combination Lemma
from the first chapter. For the full proof, we will show mutual containment between
the two sets.
The first containment [[S]] ⊇ [S] is an instance of the more general, and obvious,
fact that for any subset T of a vector space, [T ] ⊇ T .
70 Linear Algebra, by Hefferon
For the other containment, that [[S]] ⊆ [S], take m vectors from [S], namely
c1,1~s1,1 + · · · + c1,n1~s1,n1 , . . . , c1,m~s1,m + · · · + c1,nm~s1,nm , and note that any linear
combination of those
r1 (c1,1~s1,1 + · · · + c1,n1~s1,n1 ) + · · · + rm (c1,m~s1,m + · · · + c1,nm~s1,nm )
is a linear combination of elements of S
= (r1 c1,1 )~s1,1 + · · · + (r1 c1,n1 )~s1,n1 + · · · + (rm c1,m )~s1,m + · · · + (rm c1,nm )~s1,nm
and so is in [S]. That is, simply recall that a linear combination of linear combinations
(of members of S) is a linear combination (again of members of S).
Two.I.2.36 (a) It is not a subspace because these are not the inherited operations. For
one thing, in this space,
x 1
0 · y = 0
z 0
while this does not, of course, hold in R3 .
(b) We can combine the argument showing closure under addition with the argument
showing closure under scalar multiplication into one single argument showing closure
under linear combinations of two vectors. If r1 , r2 , x1 , x2 , y1 , y2 , z1 , z2 are in R then
x1 x2 r1 x1 − r1 + 1 r2 x2 − r2 + 1
r1 y1 + r2 y2 = r1 y1 + r2 y2
z1 z2 r1 z1 r2 z2
r1 x1 − r1 + r2 x2 − r2 + 1
= r1 y1 + r2 y2
r1 z1 + r2 z2
(note that the definition of addition in this space is that the first components
combine as (r1 x1 − r1 + 1) + (r2 x2 − r2 + 1) − 1, so the first component of the last
vector does not say ‘ + 2’). Adding the three components of the last vector gives
r1 (x1 − 1 + y1 + z1 ) + r2 (x2 − 1 + y2 + z2 ) + 1 = r1 · 0 + r2 · 0 + 1 = 1.
Most of the other checks of the conditions are easy (although the oddness of the
operations keeps them from being routine). Commutativity of addition goes like
this.
x1 x2 x1 + x2 − 1 x2 + x1 − 1 x2 x1
y + y = y + y = y + y = y + y
1 2 1 2 2 1 2 1
z1 z2 z1 + z2 z 2 + z1 z2 z1
Associativity of addition has
x1 x2 x3 (x1 + x2 − 1) + x3 − 1
(y1 + y2 ) + y3 = (y1 + y2 ) + y3
z1 z2 z3 (z1 + z2 ) + z3
Answers to Exercises 71
while
x1 x2 x3 x1 + (x2 + x3 − 1) − 1
y1 + (y2 + y3 ) = y1 + (y2 + y3 )
z1 z2 z3 z1 + (z2 + z3 )
and they are equal. The identity element with respect to this addition operation
works this way
x 1 x+1−1 x
y + 0 = y + 0 = y
z 0 z+0 z
and the additive inverse is similar.
x −x + 2 x + (−x + 2) − 1 1
y + −y = y−y = 0
z −z z−z 0
The conditions on scalar multiplication are also easy. For the first condition,
x (r + s)x − (r + s) + 1
(r + s) y = (r + s)y
z (r + s)z
while
x x rx − r + 1 sx − s + 1
r y + s y = ry + sy
z z rz sz
(rx − r + 1) + (sx − s + 1) − 1
= ry + sy
rz + sz
and the two are equal. The second condition compares
x1 x2 x1 + x2 − 1 r(x1 + x2 − 1) − r + 1
r · (y1 + y2 ) = r · y1 + y2 = r(y1 + y2 )
z1 z2 z1 + z2 r(z1 + z2 )
with
x1 x2 rx1 − r + 1 rx2 − r + 1
r y1 + r y2 = ry1 + ry2
z1 z2 rz1 rz2
(rx1 − r + 1) + (rx2 − r + 1) − 1
= ry1 + ry2
rz1 + rz2
72 Linear Algebra, by Hefferon
Two.I.2.47 Call the subset S. By Lemma 2.9, we need to check that [S] is closed under
linear combinations. If c1~s1 + · · · + cn~sn , cn+1~sn+1 + · · · + cm~sm ∈ [S] then for any
p, r ∈ R we have
Two.I.2.48 For this to happen, one of the conditions giving the sensibleness of the
addition and scalar multiplication operations must be violated. Consider R2 with
these operations.
! ! ! ! !
x1 x2 0 x 0
+ = r =
y1 y2 0 y 0
The set R2 is closed under these operations. But it is not a vector space.
! !
1 1
1· 6=
1 1
76 Linear Algebra, by Hefferon
Linear Independence
has more unknowns than equations, and so Gauss’s Method must end with at
least one variable free (there can’t be a contradictory equation because the system
is homogeneous, and so has at least the solution of all zeroes). To exhibit a
combination, we can do the reduction
9 2 3 12 0
−ρ1 +ρ2 (1/2)ρ2 +ρ3
−→ −→ 0 −2 2 0 0
0 0 −3 −1 0
and take, say, c4 = 1. Then we have that c3 = −1/3, c2 = −1/3, and c1 = −31/27.
Two.II.1.21 In the cases of independence, you must prove that it is independent. Other-
wise, you must exhibit a dependence. (Here we give a specific dependence but others
are possible.)
(a) This set is independent. Setting up the relation c1 (3 − x + 9x2 ) + c2 (5 − 6x +
3x2 ) + c3 (1 + 1x − 5x2 ) = 0 + 0x + 0x2 gives a linear system
3 5 1 0 3 5 1 0
(1/3)ρ1 +ρ2 3ρ2 −(12/13)ρ2 +ρ3
−1 −6 1 0 −→ −→ −→ 0 −13 4 0
−3ρ1 +ρ3
9 3 −5 0 0 0 −128/13 0
with only one solution: c1 = 0, c2 = 0, and c3 = 0.
(b) This set is independent. We can see this by inspection, straight from the definition
of linear independence. Obviously neither is a multiple of the other.
(c) This set is linearly independent. The linear system reduces in this way
2 3 4 0 2 3 4 0
−(1/2)ρ1 +ρ2 −(17/5)ρ2 +ρ3
1 −1 0 0 −→ −→ 0 −5/2 −2 0
−(7/2)ρ1 +ρ3
7 2 −3 0 0 0 −51/5 0
to show that there is only the solution c1 = 0, c2 = 0, and c3 = 0.
(d) This set is linearly dependent. The linear system
8 0 2 8 0
3 1 2 −2 0
3 2 2 5 0
must, after reduction, end with at least one variable free (there are more variables
than equations, and there is no possibility of a contradictory equation because the
system is homogeneous). We can take the free variables as parameters to describe
the solution set. We can then set the parameter to a nonzero value to get a nontrivial
linear relation.
Two.II.1.22 Let Z be the zero function Z(x) = 0, which is the additive identity in the
vector space under discussion.
78 Linear Algebra, by Hefferon
(a) This set is linearly independent. Consider c1 · f(x) + c2 · g(x) = Z(x). Plugging
in x = 1 and x = 2 gives a linear system
c1 · 1 + c2 · 1 = 0
c1 · 2 + c2 · (1/2) = 0
with the unique solution c1 = 0, c2 = 0.
(b) This set is linearly independent. Consider c1 · f(x) + c2 · g(x) = Z(x) and plug
in x = 0 and x = π/2 to get
c1 · 1 + c2 · 0 = 0
c1 · 0 + c2 · 1 = 0
which obviously gives that c1 = 0, c2 = 0.
(c) This set is also linearly independent. Considering c1 · f(x) + c2 · g(x) = Z(x) and
plugging in x = 1 and x = e
c1 · e + c2 · 0 = 0
c1 · e e + c2 · 1 = 0
gives that c1 = 0 and c2 = 0.
Two.II.1.23 In each case, if the set is independent then you must prove that and if it is
dependent then you must exhibit a dependence.
(a) This set is dependent. The familiar relation sin2 (x) + cos2 (x) = 1 shows that
2 = c1 · (4 sin2 (x)) + c2 · (cos2 (x)) is satisfied by c1 = 1/2 and c2 = 2.
(b) This set is independent. Consider the relationship c1 ·1+c2 ·sin(x)+c3 ·sin(2x) = 0
(that ‘0’ is the zero function). Taking three suitable points such as x = π, x = π/2,
x = π/4 gives a system
c1 =0
c1 + c2 =0
√
c1 + ( 2/2)c2 + c3 = 0
whose only solution is c1 = 0, c2 = 0, and c3 = 0.
(c) By inspection, this set is independent. Any dependence cos(x) = c · x is not
possible since the cosine function is not a multiple of the identity function (we are
applying Corollary 1.18).
(d) By inspection, we spot that there is a dependence. Because (1 + x)2 = x2 + 2x + 1,
we get that c1 · (1 + x)2 + c2 · (x2 + 2x) = 3 is satisfied by c1 = 3 and c2 = −3.
(e) This set is dependent. The easiest way to see that is to recall the trigonometric
relationship cos2 (x) − sin2 (x) = cos(2x). (Remark. A person who doesn’t recall
this, and tries some x’s, simply never gets a system leading to a unique solution,
and never gets to conclude that the set is independent. Of course, this person might
wonder if they simply never tried the right set of x’s, but a few tries will lead most
people to look instead for a dependence.)
(f) This set is dependent, because it contains the zero object in the vector space,
the zero polynomial.
Answers to Exercises 79
Two.II.1.24 No, that equation is not a linear relationship. In fact this set is independent,
as the system arising from taking x to be 0, π/6 and π/4 shows.
Two.II.1.25 No. Here are two members of the plane where the second is a multiple of
the first.
1 2
0 , 0
0 0
(Another reason that the answer is “no” is the the zero vector is a member of the
plane and no set containing the zero vector is linearly independent.)
Two.II.1.26 We have already showed this: the Linear Combination Lemma and its
corollary state that in an echelon form matrix, no nonzero row is a linear combination
of the others.
Two.II.1.27 (a) Assume that {~u,~v, w ~ } is linearly independent, so that any relationship
~ = ~0 leads to the conclusion that d0 = 0, d1 = 0, and d2 = 0.
d0 ~u + d1~v + d2 w
Consider the relationship c1 (~u) + c2 (~u + ~v) + c3 (~u + ~v + w ~ ) = ~0. Rewrite it to
get (c1 + c2 + c3 )~u + (c2 + c3 )~v + (c3 )~ ~
w = 0. Taking d0 to be c1 + c2 + c3 , taking
d1 to be c2 + c3 , and taking d2 to be c3 we have this system.
c1 + c2 + c3 = 0
c2 + c3 = 0
c3 = 0
Conclusion: the c’s are all zero, and so the set is linearly independent.
(b) The second set is dependent
1 · (~u − ~v) + 1 · (~v − w w − ~u) = ~0
~ ) + 1 · (~
whether or not the first set is independent.
Two.II.1.28 (a) A singleton set {~v } is linearly independent if and only if ~v 6= ~0. For the
‘if’ direction, with ~v 6= ~0, we can apply Lemma 1.5 by considering the relationship
c · ~v = ~0 and noting that the only solution is the trivial one: c = 0. For the ‘only if’
direction, just recall that Example 1.11 shows that {~0 } is linearly dependent, and
so if the set {~v } is linearly independent then ~v 6= ~0.
(Remark. Another answer is to say that this is the special case of Lemma 1.14
where S = ∅.)
(b) A set with two elements is linearly independent if and only if neither member is
a multiple of the other (note that if one is the zero vector then it is a multiple of
the other). This is an equivalent statement: a set is linearly dependent if and only
if one element is a multiple of the other.
The proof is easy. A set {~v1 ,~v2 } is linearly dependent if and only if there is a
relationship c1~v1 + c2~v2 = ~0 with either c1 6= 0 or c2 6= 0 (or both). That holds if
and only if ~v1 = (−c2 /c1 )~v2 or ~v2 = (−c1 /c2 )~v1 (or both).
80 Linear Algebra, by Hefferon
Two.II.1.29 This set is linearly dependent set because it contains the zero vector.
Two.II.1.30 Lemma 1.19 gives the ‘if’ half. The converse (the ‘only if’ statement) does
not hold. An example is to consider the vector space R2 and these vectors.
! ! !
1 0 1
~x = , ~y = , ~z =
0 1 1
Two.II.1.31 (a) The linear system arising from
1 −1 0
c1 1 + c2 2 = 0
0 0 0
has the unique solution c1 = 0 and c2 = 0.
(b) The linear system arising from
1 −1 3
c1 1 + c2 2 = 2
0 0 0
has the unique solution c1 = 8/3 and c2 = −1/3.
(c) Suppose that S is linearly independent. Suppose that we have both ~v = c1~s1 +
· · · + cn~sn and ~v = d1~t1 + · · · + dm~tm (where the vectors are members of S). Now,
c1~s1 + · · · + cn~sn = ~v = d1~t1 + · · · + dm~tm
can be rewritten in this way.
c1~s1 + · · · + cn~sn − d1~t1 − · · · − dm~tm = ~0
Possibly some of the ~s ’s equal some of the ~t ’s; we can combine the associated
coefficients (i.e., if ~si = ~tj then · · · + ci~si + · · · − dj~tj − · · · can be rewritten as
· · · + (ci − dj )~si + · · · ). That equation is a linear relationship among distinct (after
the combining is done) members of the set S. We’ve assumed that S is linearly
independent, so all of the coefficients are zero. If i is such that ~si does not equal
any ~tj then ci is zero. If j is such that ~tj does not equal any ~si then dj is zero. In
the final case, we have that ci − dj = 0 and so ci = dj .
Therefore, the original two sums are the same, except perhaps for some 0 · ~si or
0 · ~tj terms that we can neglect.
(d) This set is not linearly independent:
! !
1 2
S={ , } ⊂ R2
0 0
and these two linear combinations give the same result
! ! ! ! !
0 1 2 1 2
=2· −1· =4· −2·
0 0 0 0 0
Thus, a linearly dependent set might have indistinct sums.
Answers to Exercises 81
In fact, this stronger statement holds: if a set is linearly dependent then it must
have the property that there are two distinct linear combinations that sum to the
same vector. Briefly, where c1~s1 + · · · + cn~sn = ~0 then multiplying both sides of the
relationship by two gives another relationship. If the first relationship is nontrivial
then the second is also.
Two.II.1.32 In this ‘if and only if’ statement, the ‘if’ half is clear — if the polynomial is
the zero polynomial then the function that arises from the action of the polynomial
must be the zero function x 7→ 0. For ‘only if’ we write p(x) = cn xn + · · · + c0 .
Plugging in zero p(0) = 0 gives that c0 = 0. Taking the derivative and plugging in
zero p0 (0) = 0 gives that c1 = 0. Similarly we get that each ci is zero, and p is the
zero polynomial.
Two.II.1.33 The work in this section suggests that we should define an n-dimensional
non-degenerate linear surface as the span of a linearly independent set of n vectors.
Two.II.1.34 (a) For any a1,1 , . . . , a2,4 ,
! ! ! ! !
a1,1 a1,2 a1,3 a1,4 0
c1 + c2 + c3 + c4 =
a2,1 a2,2 a2,3 a2,4 0
yields a linear system
a1,1 c1 + a1,2 c2 + a1,3 c3 + a1,4 c4 = 0
a2,1 c1 + a2,2 c2 + a2,3 c3 + a2,4 c4 = 0
that has infinitely many solutions (Gauss’s Method leaves at least two variables
free). Hence there are nontrivial linear relationships among the given members of
R2 .
(b) Any set five vectors is a superset of a set of four vectors, and so is linearly
dependent.
With three vectors from R2 , the argument from the prior item still applies, with
the slight change that Gauss’s Method now only leaves at least one variable free
(but that still gives infinitely many solutions).
(c) The prior item shows that no three-element subset of R2 is independent. We
know that there are two-element subsets of R2 that are independent — one is
! !
1 0
{ , }
0 1
and so the answer is two.
Two.II.1.35 Yes; here is one.
1 0 0 1
{ 0 , 1 , 0 , 1 }
0 0 1 1
82 Linear Algebra, by Hefferon
Two.II.1.36 Yes. The two improper subsets, the entire set and the empty subset, serve
as examples.
Two.II.1.37 In R4 the biggest linearly independent set has four vectors. There are many
examples of such sets, this is one.
1 0 0 0
0 1 0 0
{ , , , }
0 0 1 0
0 0 0 1
To see that
no set
with five or
more vectors
can be
independent,
set up
a1,1 a1,2 a1,3 a1,4 a1,5 0
a a a a a 0
2,1 2,2 2,3 2,4 2,5
c1 + c2 + c3 + c4 + c5 =
a3,1 a3,2 a3,3 a3,4 a3,5 0
a4,1 a4,2 a4,3 a4,4 a4,5 0
and note that the resulting linear system
a1,1 c1 + a1,2 c2 + a1,3 c3 + a1,4 c4 + a1,5 c5 = 0
a2,1 c1 + a2,2 c2 + a2,3 c3 + a2,4 c4 + a2,5 c5 = 0
a3,1 c1 + a3,2 c2 + a3,3 c3 + a3,4 c4 + a3,5 c5 = 0
a4,1 c1 + a4,2 c2 + a4,3 c3 + a4,4 c4 + a4,5 c5 = 0
has four equations and five unknowns, so Gauss’s Method must end with at least one c
variable free, so there are infinitely many solutions, and so the above linear relationship
among the four-tall vectors has more solutions than just the trivial solution.
The smallest linearly independent set is the empty set.
The biggest linearly dependent set is R4 . The smallest is {~0 }.
Two.II.1.38 (a) The intersection of two linearly independent sets S ∩ T must be linearly
independent as it is a subset of the linearly independent set S (as well as the linearly
independent set T also, of course).
(b) The complement of a linearly independent set is linearly dependent as it contains
the zero vector.
(c) A simple example in R2 is these two ! sets. !
1 0
S={ } T ={ }
0 1
2
A somewhat subtler example, again
! in R , is these
! two.!
1 1 0
S={ } T ={ , }
0 0 1
2
(d) We must produce an example. One, ! in R , is !
1 2
S={ } T ={ }
0 0
since the linear dependence of S1 ∪ S2 is easy to see.
Answers to Exercises 83
Two.II.1.39 (a) Lemma 1.5 requires that the vectors ~s1 , . . . ,~sn , ~t1 , . . . , ~tm be distinct.
But we could have that the union S ∪ T is linearly independent with some ~si equal
to some ~tj .
(b) One example in R2 is these two.
! ! !
1 1 0
S={ } T ={ , }
0 0 1
(c) An example from R2 is these sets.
! ! ! !
1 0 1 1
S={ , } T ={ , }
0 1 0 1
(d) The union of two linearly independent sets S∪T is linearly independent if and only
if their spans of S and T − (S ∩ T ) have a trivial intersection [S] ∩ [T − (S ∩ T )] = {~0 }.
To prove that, assume that S and T are linearly independent subsets of some vector
space.
For the ‘only if’ direction, assume that the intersection of the spans is trivial
[S] ∩ [T − (S ∩ T )] = {~0 }. Consider the set S ∪ (T − (S ∩ T )) = S ∪ T and consider the
linear relationship c1~s1 + · · · + cn~sn + d1~t1 + · · · + dm~tm = ~0. Subtracting gives
c1~s1 + · · · + cn~sn = −d1~t1 − · · · − dm~tm . The left side of that equation sums to
a vector in [S], and the right side is a vector in [T − (S ∩ T )]. Therefore, since the
intersection of the spans is trivial, both sides equal the zero vector. Because S is
linearly independent, all of the c’s are zero. Because T is linearly independent so
also is T − (S ∩ T ) linearly independent, and therefore all of the d’s are zero. Thus,
the original linear relationship among members of S ∪ T only holds if all of the
coefficients are zero. Hence, S ∪ T is linearly independent.
For the ‘if’ half we can make the same argument in reverse. Suppose that
the union S ∪ T is linearly independent. Consider a linear relationship among
members of S and T − (S ∩ T ). c1~s1 + · · · + cn~sn + d1~t1 + · · · + dm~tm = ~0 Note
that no ~si is equal to a ~tj so that is a combination of distinct vectors, as required
by Lemma 1.5. So the only solution is the trivial one c1 = 0, . . . , dm = 0.
Since any vector ~v in the intersection of the spans [S] ∩ [T − (S ∩ T )] we can write
~v = c1~s1 + · · · + cn~sn = −d1~t1 − · · · − dm~tm , and it must be the zero vector because
each scalar is zero.
Two.II.1.40 (a) We do induction on the number of vectors in the finite set S.
The base case is that S has no elements. In this case S is linearly independent
and there is nothing to check — a subset of S that has the same span as S is S itself.
For the inductive step assume that the theorem is true for all sets of size n = 0,
n = 1, . . . , n = k in order to prove that it holds when S has n = k + 1 elements.
If the k + 1-element set S = {~s0 , . . . ,~sk } is linearly independent then the theorem
is trivial, so assume that it is dependent. By Corollary 1.18 there is an ~si that is
84 Linear Algebra, by Hefferon
we can clear
down the first column.
1 b/a c/a 0 1 b/a c/a 0
(1/a)ρ1 −dρ1 +ρ2
−→ d e f 0 −→ 0 (ae − bd)/a (af − cd)/a 0
−gρ1 +ρ3
g h i 0 0 (ah − bg)/a (ai − cg)/a 0
Then we get a 1 in the second row, second column entry. (Assuming for the moment
that ae − bd 6= 0, in order to do the row reduction step.)
1 b/a c/a 0
(a/(ae−bd))ρ2
−→ 0 1 (af − cd)/(ae − bd) 0
0 (ah − bg)/a (ai − cg)/a 0
Then, under the assumptions, we perform the row operation ((ah − bg)/a)ρ2 + ρ3
to get this.
1 b/a c/a 0
0 1 (af − cd)/(ae − bd) 0
0 0 (aei + bgf + cdh − hfa − idb − gec)/(ae − bd) 0
Therefore, the original system is nonsingular if and only if the above 3, 3 entry is
nonzero (this fraction is defined because of the ae − bd 6= 0 assumption). It equals
zero if and only if the numerator is zero.
We next worry about the assumptions. First, if a 6= 0 but ae − bd = 0 then we
swap
1 b/a c/a 0
0 0 (af − cd)/a 0
0 (ah − bg)/a (ai − cg)/a 0
1 b/a c/a 0
ρ2 ↔ρ3
−→ 0 (ah − bg)/a (ai − cg)/a 0
0 0 (af − cd)/a 0
and conclude that the system is nonsingular if and only if either ah − bg = 0 or
af − cd = 0. That’s the same as asking that their product be zero:
ahaf − ahcd − bgaf + bgcd = 0
ahaf − ahcd − bgaf + aegc = 0
a(haf − hcd − bgf + egc) = 0
(in going from the first line to the second we’ve applied the case assumption that
ae − bd = 0 by substituting ae for bd). Since we are assuming that a 6= 0, we have
that haf − hcd − bgf + egc = 0. With ae − bd = 0 we can rewrite this to fit the
form we need: in this a 6= 0 and ae − bd = 0 case, the given system is nonsingular
when haf − hcd − bgf + egc − i(ae − bd) = 0, as required.
The remaining cases have the same character. Do the a = 0 but d 6= 0 case
and the a = 0 and d = 0 but g 6= 0 case by first swapping rows and then going on
86 Linear Algebra, by Hefferon
Two.III.1: Basis
Two.III.1.18 (a) This is a basis for P2 . To show that it spans the space we consider
a generic a2 x2 + a1 x + a0 ∈ P2 and look for scalars c1 , c2 , c3 ∈ R such that
a2 x2 + a1 x + a0 = c1 · (x2 − x + 1) + c2 · (2x + 1) + c3 (2x − 1). Gauss’s Method on
the linear system
c1 = a2
2c2 + 2c3 = a1
c2 − c3 = a0
shows that given the ai ’s we can compute the cj ’s as c1 = a2 , c2 = (1/4)a1 +(1/2)a0 ,
and c3 = (1/4)a1 − (1/2)a0 . Thus each element of P2 is a combination of the given
three.
To prove that the set of the given three is linearly independent we can set up
the equation 0x2 + 0x + 0 = c1 · (x2 − x + 1) + c2 · (2x + 1) + c3 (2x − 1) and solve,
and it will give that c1 = 0, c2 = 0, and c3 = 0. Or, we can instead observe that
the solution in the prior paragraph is unique, and cite Theorem 1.12.
(b) This is not a basis. It does not span the space since no combination of the two
c1 · (x + x2 ) + c2 · (x − x2 ) will sum to the polynomial 3 ∈ P2 .
Two.III.1.19 By Theorem 1.12, each is a basis if and only if we can express each vector
in the space in a unique way as a linear combination of the given vectors.
(a) Yes this is a basis. Therelation
1 3 0 x
c1 2 + c2 2 + c3 0 = y
3 1 1 z
gives
1 3 0 x 1 3 0 x
−2ρ1 +ρ2 −2ρ2 +ρ3
2 2 0 y −→ −→ 0 −4 0 −2x + y
−3ρ1 +ρ3
3 1 1 z 0 0 1 x − 2y + z
which has the unique solution c3 = x−2y+z, c2 = x/2−y/4, and c1 = −x/2+3y/4.
88 Linear Algebra, by Hefferon
Two.III.1.21 A natural basis is h1, x, x2 i. There are bases for P2 that do not contain
any polynomials of degree one or degree zero. One is h1 + x + x2 , x + x2 , x2 i. (Every
basis has at least one polynomial of degree two, though.)
Two.III.1.22 The reduction ! !
1 −4 3 −1 0 −2ρ1 +ρ2 1 −4 3 −1 0
−→
2 −8 6 −2 0 0 0 0 0 0
gives that the only condition is that x1 = 4x2 − 3x3 + x4 . The solution set is
4x2 − 3x3 + x4
x2
{ | x2 , x3 , x4 ∈ R}
x3
x4
4 −3 1
1 0 0
= {x2 + x3 + x4 | x2 , x3 , x4 ∈ R}
0 1 0
0 0 1
and so the obvious candidate forthe basis
isthis.
4 −3 1
1 0 0
h , , i
0 1 0
0 0 1
We’ve shown that this spans the space, and showing it is also linearly independent is
routine.
Two.III.1.23 There are many bases. This is a natural one.
! ! ! !
1 0 0 1 0 0 0 0
h , , , i
0 0 0 0 1 0 0 1
Two.III.1.24 For each item, many answers are possible.
(a) One way to proceed is to parametrize by expressing the a2 as a combination of
the other two a2 = 2a1 + a0 . Then a2 x2 + a1 x + a0 is (2a1 + a0 )x2 + a1 x + a0 and
{(2a1 + a0 )x2 + a1 x + a0 | a1 , a0 ∈ R}
= {a1 · (2x2 + x) + a0 · (x2 + 1) | a1 , a0 ∈ R}
suggests h2x2 + x, x2 + 1i. This only shows that it spans, but checking that it is
linearly independent is routine.
90 Linear Algebra, by Hefferon
Thus, the only quadratic polynomials a0 + a1 x + a2 x2 with associated c’s are the
ones such that 0 = (−4/3)a0 − (4/3)a1 + a2 . Hence the span is this.
{ (−a1 + (3/4)a2 ) + a1 x + a2 x2 | a1 , a2 ∈ R }
Parametrizing gives {a1 · (−1 + x) + a2 · ((3/4) + x2 ) | a1 , a2 ∈ R}, which suggests
h−1 + x, (3/4) + x2 i (checking that it is linearly independent is routine).
Two.III.1.27 (a) The subspace is this.
{ a0 + a1 x + a2 x2 + a3 x3 | a0 + 7a1 + 49a2 + 343a3 = 0 }
Rewriting a0 = −7a1 − 49a2 − 343a3 gives this.
{(−7a1 − 49a2 − 343a3 ) + a1 x + a2 x2 + a3 x3 | a1 , a2 , a3 ∈ R}
On breaking out the parameters, this suggests h−7 + x, −49 + x2 , −343 + x3 i for
the basis (it is easily verified).
(b) The given subspace is the collection of cubics p(x) = a0 + a1 x + a2 x2 + a3 x3
such that a0 + 7a1 + 49a2 + 343a3 = 0 and a0 + 5a1 + 25a2 + 125a3 = 0. Gauss’s
Method
a0 + 7a1 + 49a2 + 343a3 = 0 −ρ1 +ρ2 a0 + 7a1 + 49a2 + 343a3 = 0
−→
a0 + 5a1 + 25a2 + 125a3 = 0 −2a1 − 24a2 − 218a3 = 0
gives that a1 = −12a2 − 109a3 and that a0 = 35a2 + 420a3 . Rewriting (35a2 +
420a3 )+(−12a2 −109a3 )x+a2 x2 +a3 x3 as a2 ·(35−12x+x2 )+a3 ·(420−109x+x3 )
suggests this for a basis h35 − 12x + x2 , 420 − 109x + x3 i. The above shows that
it spans the space. Checking it is linearly independent is routine. (Comment. A
worthwhile check is to verify that both polynomials in the basis have both seven
and five as roots.)
(c) Here there are three conditions on the cubics, that a0 + 7a1 + 49a2 + 343a3 = 0,
that a0 + 5a1 + 25a2 + 125a3 = 0, and that a0 + 3a1 + 9a2 + 27a3 = 0. Gauss’s
Method
a0 + 7a1 + 49a2 + 343a3 = 0 a0 + 7a1 + 49a2 + 343a3 = 0
−ρ1 +ρ2 −2ρ2 +ρ3
a0 + 5a1 + 25a2 + 125a3 = 0 −→ −→ −2a1 − 24a2 − 218a3 = 0
−ρ1 +ρ3
a0 + 3a1 + 9a2 + 27a3 = 0 8a2 + 120a3 = 0
yields the single free variable a3 , with a2 = −15a3 , a1 = 71a3 , and a0 = −105a3 .
The parametrization is this.
{(−105a3 ) + (71a3 )x + (−15a3 )x2 + (a3 )x3 | a3 ∈ R }
= {a3 · (−105 + 71x − 15x2 + x3 ) | a3 ∈ R}
Therefore, a natural candidate for the basis is h−105 + 71x − 15x2 + x3 i. It spans
the space by the work above. It is clearly linearly independent because it is a
one-element set (with that single element not the zero object of the space). Thus,
any cubic through the three points (7, 0), (5, 0), and (3, 0) is a multiple of this one.
(Comment. As in the prior question, a worthwhile check is to verify that plugging
seven, five, and three into this polynomial yields zero each time.)
92 Linear Algebra, by Hefferon
(d) This is the trivial subspace of P3 . Thus, the basis is empty hi.
Remark. Alternatively, we could have derived the polynomial in the third item by
multiplying out (x − 7)(x − 5)(x − 3).
Two.III.1.28 Yes. Linear independence and span are unchanged by reordering.
Two.III.1.29 No linearly independent set contains a zero vector.
Two.III.1.30 (a) To show that it is linearly independent, note that if d1 (c1 β ~ 1) +
~ 2 ) + d3 (c3 β
d2 (c2 β ~ 3 ) = ~0 then (d1 c1 )β
~ 1 + (d2 c2 )β
~ 2 + (d3 c3 )β
~ 3 = ~0, which in turn
implies that each di ci is zero. But with ci 6= 0 that means that each di is zero.
Showing that it spans the space is much the same; because hβ ~ 1, β
~ 2, β
~ 3 i is a basis,
and so spans the space, we can for any ~v write ~v = d1 β ~ 1 + d2 β
~ 2 + d3 β~ 3 , and then
~ ~ ~
~v = (d1 /c1 )(c1 β1 ) + (d2 /c2 )(c2 β2 ) + (d3 /c3 )(c3 β3 ).
If any of the scalars are zero then the result is not a basis, because it is not
linearly independent.
(b) Showing that h2β ~ 1, β
~1 + β~ 2, β
~1 + β
~ 3 i is linearly independent is easy. To show
that it spans the space, assume that ~v = d1 β ~ 1 + d2 β~ 2 + d3 β
~ 3 . Then, we can
~ ~ ~ ~ ~
represent the same ~v with respect to h2β1 , β1 + β2 , β1 + β3 i in this way ~v =
(1/2)(d1 − d2 − d3 )(2β ~ 1 ) + d2 ( β
~1 + β
~ 2 ) + d3 ( β
~1 + β
~ 3 ).
Two.III.1.31 Each forms a linearly independent set if we omit ~v. To preserve linear
independence, we must expand the span of each. That is, we must determine the
span of each (leaving ~v out), and then pick a ~v lying outside of that span. Then to
finish, we must check that the result spans the entire given space. Those checks are
routine.
(a) Any vector that is not a multiple of the given one, that is, any vector that is not
on the line y = x will do here. One is ~v = ~e1 .
(b) By inspection, we notice that the vector ~e3 is not in the span of the set of the
two given vectors. The check that the resulting set is a basis for R3 is routine.
(c) For any member of the span { c1 · (x) + c2 · (1 + x2 ) | c1 , c2 ∈ R}, the coefficient
of x2 equals the constant term. So we expand the span if we add a quadratic
without this property, say, ~v = 1 − x2 . The check that the result is a basis for P2 is
easy.
Two.III.1.32 To show that each scalar is zero, simply subtract c1 β ~ 1 + · · · + ck β
~k −
~ ~ ~
ck+1 βk+1 − · · · − cn βn = 0. The obvious generalization is that in any equation
involving only the β’s,~ and in which each β ~ appears only once, each scalar is zero.
For instance, an equation with a combination of the even-indexed basis vectors (i.e.,
~ 2, β
β ~ 4 , etc.) on the right and the odd-indexed basis vectors on the left also gives the
conclusion that all of the coefficients are zero.
Two.III.1.33 No; no linearly independent set contains the zero vector.
Answers to Exercises 93
Two.III.1.34 Here is a subset of R2 that is not a basis, and two different linear combina-
tions of its elements
! that!sum to the ! same vector.
! ! !
1 2 1 2 1 2
{ , } 2· +0· =0· +1·
2 4 2 4 2 4
Thus, when a subset is not a basis, it can be the case that its linear combinations are
not unique.
But just because a subset is not a basis does not imply that its combinations must
be not unique. For instance, this set
!
1
{ }
2
does have the property that
! !
1 1
c1 · = c2 ·
2 2
implies that c1 = c2 . The idea here is that this subset fails to be a basis because it
fails to span the space; the proof of the theorem establishes that linear combinations
are unique if and only if the subset is linearly independent.
Two.III.1.35 (a) Describing the vector space as
!
a b
{ | a, b, c ∈ R }
b c
suggests this for a basis.
! ! !
1 0 0 0 0 1
h , , i
0 0 0 1 1 0
Verification is easy.
(b) This is one possible basis.
1 0 0 0 0 0 0 0 0 0 1 0 0 0 1 0 0 0
h0 0 0 , 0 1 0 , 0 0 0 , 1 0 0 , 0 0 0 , 0 0 1i
0 0 0 0 0 0 0 0 1 0 0 0 1 0 0 0 1 0
(c) As in the prior two questions, we can form a basis from two kinds of matrices.
First are the matrices with a single one on the diagonal and all other entries zero
(there are n of those matrices). Second are the matrices with two opposed off-
diagonal entries are ones and all other entries are zeros. (That is, all entries in M
are zero except that mi,j and mj,i are one.)
Two.III.1.36 (a) Any four vectors from R3 are linearly related because the vector
equation
x1 x2 x3 x4 0
c1 y1 + c2 y2 + c3 y3 + c4 y4 = 0
z1 z2 z3 z4 0
94 Linear Algebra, by Hefferon
Two.III.2: Dimension
Two.III.2.16 One basis is h1, x, x2 i, and so the dimension is three.
Two.III.2.17 The solution set is
4x2 − 3x3 + x4
x2
{ | x2 , x3 , x4 ∈ R }
x3
x4
so a natural basis is this
4 −3 1
1 0 0
h , , i
0 1 0
0 0 1
(checking linear independence is easy). Thus the dimension is three.
Two.III.2.18 For this space
!
a b
{ | a, b, c, d ∈ R}
c d
! !
1 0 0 0
= {a · + ··· + d · | a, b, c, d ∈ R}
0 0 0 1
this is a natural basis.
! ! ! !
1 0 0 1 0 0 0 0
h , , , i
0 0 0 0 1 0 0 1
The dimension is four.
96 Linear Algebra, by Hefferon
Two.III.2.19 (a) As in the prior exercise, the space M2×2 of matrices without restriction
has this basis ! ! ! !
1 0 0 1 0 0 0 0
h , , , i
0 0 0 0 1 0 0 1
and so the dimension is four.
(b) For this space
!
a b
{ | a = b − 2c and d ∈ R }
c d
! ! !
1 1 −2 0 0 0
= {b · +c· +d· | b, c, d ∈ R}
0 0 1 0 0 1
this is a natural basis.
! ! !
1 1 −2 0 0 0
h , , i
0 0 1 0 0 1
The dimension is three.
(c) Gauss’s Method applied to the two-equation linear system gives that c = 0 and
that a = −b. Thus, we have this description
! ! !
−b b −1 1 0 0
{ | b, d ∈ R } = { b · +d· | b, d ∈ R }
0 d 0 0 0 1
and so this is a natural basis.
! !
−1 1 0 0
h , i
0 0 0 1
The dimension is two.
Two.III.2.20 The bases for these spaces are developed in the answer set of the prior
subsection.
(a) One basis is h−7 + x, −49 + x2 , −343 + x3 i. The dimension is three.
(b) One basis is h35 − 12x + x2 , 420 − 109x + x3 i so the dimension is two.
(c) A basis is {−105 + 71x − 15x2 + x3 }. The dimension is one.
(d) This is the trivial subspace of P3 and so the basis is empty. The dimension is
zero.
Two.III.2.21 First recall that cos 2θ = cos2 θ − sin2 θ, and so deletion of cos 2θ from this
set leaves the span unchanged. What’s left, the set {cos2 θ, sin2 θ, sin 2θ }, is linearly
independent (consider the relationship c1 cos2 θ + c2 sin2 θ + c3 sin 2θ = Z(θ) where
Z is the zero function, and then take θ = 0, θ = π/4, and θ = π/2 to conclude that
each c is zero). It is therefore a basis for its span. That shows that the span is a
dimension three vector space.
Answers to Exercises 97
combination of the other two. That’s simply an assertion that the three-element set
is linearly independent. By Corollary 2.15, that’s equivalent to an assertion that the
set is a basis for R3 (more precisely, any sequence made from the set’s elements is a
basis).
Two.III.2.31 Let the space V be finite dimensional. Let S be a subspace of V.
(a) The empty set is a linearly independent subset of S. By Corollary 2.13, it can be
expanded to a basis for the vector space S.
(b) Any basis for the subspace S is a linearly independent set in the superspace V.
Hence it can be expanded to a basis for the superspace, which is finite dimensional.
Therefore it has only finitely many members.
~
Two.III.2.32 It ensures that we exhaust the β’s. That is, it justifies the first sentence of
the last paragraph.
Two.III.2.33 Let BU be a basis for U and let BW be a basis for W. Consider the
concatenation of the two basis sequences. If there is a repeated element then the
intersection U ∩ W is nontrivial. Otherwise, the set BU ∪ BW is linearly dependent
as it is a six member subset of the five-dimensional space R5 . In either case some
member of BW is in the span of BU , and thus U ∩ W is more than just the trivial
space {~0 }.
Generalization: if U, W are subspaces of a vector space of dimension n and if
dim(U) + dim(W) > n then they have a nontrivial intersection.
Two.III.2.34 First, note that a set is a basis for some space if and only if it is linearly
independent, because in that case it is a basis for its own span.
(a) The answer to the question in the second paragraph is “yes” (implying “yes”
answers for both questions in the first paragraph). If BU is a basis for U then BU is
a linearly independent subset of W. Apply Corollary 2.13 to expand it to a basis
for W. That is the desired BW .
The answer to the question in the third paragraph is “no”, which implies a “no”
answer to the question of the fourth paragraph. Here is an example of a basis for a
superspace with no sub-basis forming a basis for a subspace: in W = R2 , consider
the standard basis E2 . No sub-basis of E2 forms a basis for the subspace U of R2
that is the line y = x.
(b) It is a basis (for its span) because the intersection of linearly independent sets is
linearly independent (the intersection is a subset of each of the linearly independent
sets).
It is not, however, a basis for the intersection of the spaces. For instance, these
are bases for R2 : ! ! ! !
1 0 2 0
B1 = h , i and B2 = h[ir] ,
0 1 0 2
Answers to Exercises 99
and R2 ∩ R2 = R2 , but B1 ∩ B2 is empty. All we can say is that the ∩ of the bases
is a basis for a subset of the intersection of the spaces.
(c) The ∪ of bases need not be a basis: in R2
! ! ! !
1 1 1 0
B1 = h , i and B2 = h , i
0 1 0 2
B1 ∪ B2 is not linearly independent. A necessary and sufficient condition for a ∪ of
two bases to be a basis
B1 ∪ B2 is linearly independent ⇐⇒ [B1 ∩ B2 ] = [B1 ] ∩ [B2 ]
it is easy enough to prove (but perhaps hard to apply).
(d) The complement of a basis cannot be a basis because it contains the zero vector.
Two.III.2.35 (a) A basis for U is a linearly independent set in W and so can be expanded
via Corollary 2.13 to a basis for W. The second basis has at least as many members
as the first.
(b) One direction is clear: if V = W then they have the same dimension. For the
converse, let BU be a basis for U. It is a linearly independent subset of W and so
can be expanded to a basis for W. If dim(U) = dim(W) then this basis for W has
no more members than does BU and so equals BU . Since U and W have the same
bases, they are equal.
(c) Let W be the space of finite-degree polynomials and let U be the subspace of
polynomials that have only even-powered terms.
{a0 + a1 x2 + a2 x4 + · · · + an x2n | a0 , . . . , an ∈ R}
Both spaces have infinite dimension, but U is a proper subspace.
Two.III.2.36 The possibilities for the dimension of V are 0, 1, n − 1, and n.
To see this, first consider the case when all the coordinates of ~v are equal.
z
z
~v =
..
.
z
Then σ(~v) = ~v for every permutation σ, so V is just the span of ~v, which has dimension
0 or 1 according to whether ~v is ~0 or not.
Now suppose not all the coordinates of ~v are equal; let x and y with x 6= y be
among the coordinates of ~v. Then we can find permutations σ1 and σ2 such that
x y
y x
a a
σ1 (~v) = and σ2 (~v) =
3
3
.. ..
. .
an an
100 Linear Algebra, by Hefferon
is in V. That is, ~e2 − ~e1 ∈ V, where ~e1 , ~e2 , . . . , ~en is the standard basis for Rn .
Similarly, ~e3 − ~e2 , . . . , ~en − ~e1 are all in V. It is easy to see that the vectors ~e2 − ~e1 ,
~e3 − ~e2 , . . . , ~en − ~e1 are linearly independent (that is, form a linearly independent
set), so dim V > n − 1.
Finally, we can write
Two.III.3.17 (a) Yes. To see if there are c1 and c2 such that c1 ·(2 1)+c2 ·(3 1) = (1 0)
we solve
2c1 + 3c2 = 1
c1 + c2 = 0
solution.
0 −1 −1 1 1 0 2 1
ρ1 ↔ρ2 −3ρ1 +ρ2 ρ2 +ρ3
1 0 2 1 −→ −→ −→ 0 −1 −1 1
3 1 7 1 0 0 0 −1
Thus, the vector is not in the row space.
! c1 , c2 ∈!R such!that
Two.III.3.18 (a) No. To see if there are
1 1 1
c1 + c2 =
1 1 3
we can use Gauss’s Method on the resulting linear system.
c1 + c2 = 1 −ρ1 +ρ2 c1 + c2 = 1
−→
c1 + c2 = 3 0=2
There is no solution and so the vector is not in the column space.
(b) Yes. From this relationship
1 3 1 1
c1 2 + c2 0 + c3 4 = 0
1 −3 3 0
we get a linear
system that, when
we apply Gauss’s
Method,
1 3 1 1 1 3 1 1
−2ρ1 +ρ2 −ρ2 +ρ3
2 0 4 0 −→ −→ 0 −6 2 −2
−ρ1 +ρ3
1 −3 −3 0 0 0 −6 1
yields a solution. Thus, the vector is in the column space.
Two.III.3.19 (a) Yes; we are asking if!there are!scalars c!1 and c2 such that
2 1 1
c1 + c2 =
2 5 −3
which gives rise to a linear system
2c1 + c2 = 1 −ρ1 +ρ2 2c1 + c2 = 1
−→
2c1 + 5c2 = −3 4c2 = −4
and Gauss’s Method produces c2 = −1 and c1 = 1. That is, there is indeed such a
pair of scalars and so the vector is indeed in the column space of the matrix.
(b) No; we are asking if there are scalars
! c1 and
! c2 such
! that
4 −8 0
c1 + c2 =
2 −4 1
and one way to proceed is to consider the resulting linear system
4c1 − 8c2 = 0
2c1 − 4c2 = 1
that is easily seen to have no solution. Another way to proceed is to note that any
linear combination of the columns on the left has a second component half as big
as its first component, but the vector on the right does not meet that criterion.
102 Linear Algebra, by Hefferon
(c) Yes; we can simply observe that the vector is the first column minus the second.
Or, failing that, setting up the relationship among the columns
1 −1 1 2
c1 1 + c2 1 + c3 −1 = 0
−1 −1 1 0
and considering the resulting linear system
c1 − c2 + c3 = 2 c1 − c2 + c3 = 2 c1 − c2 + c3 = 2
−ρ1 +ρ2 ρ2 +ρ3
c1 + c2 − c3 = 0 −→ 2c2 − 2c3 = −2 −→ 2c2 − 2c3 = −2
ρ1 +ρ3
−c1 − c2 + c3 = 0 −2c2 + 2c3 = 2 0= 0
gives the additional information (beyond that there is at least one solution) that
there are infinitely many solutions. Parametrizing gives c2 = −1 + c3 and c1 = 1,
and so taking c3 to be zero gives a particular solution of c1 = 1, c2 = −1, and
c3 = 0 (which is, of course, the observation made at the start).
Two.III.3.20
A routine Gaussian
reduction
2 0 3 4 2 0 3 4
0 1 1 −1 0 1 1 −1
−(3/2)ρ1 +ρ3 −ρ2 +ρ3 −ρ3 +ρ4
−→ −→ −→
3 1 0 2 −(1/2)ρ1 +ρ4 0 0 −11/2 −3
1 0 −4 −1 0 0 0 0
suggests this basis h(2 0 3 4), (0 1 1 −1), (0 0 −11/2 −3)i.
Another procedure, perhaps more convenient, is to swap rows first,
1 0 −4 −1
ρ1 ↔ρ4 −3ρ1 +ρ3 −ρ2 +ρ3 −ρ3 +ρ4 0 1 1 −1
−→ −→ −→ −→
0 0 11 6
−2ρ1 +ρ4
0 0 0 0
leading to the basis h(1 0 −4 −1), (0 1 1 −1), (0 0 11 6)i.
Two.III.3.21 (a) This reduction
2 1 3
−(1/2)ρ1 +ρ2 −(1/3)ρ2 +ρ3
−→ −→ 0 −3/2 1/2
−(1/2)ρ1 +ρ3
0 0 4/3
shows that the row rank, and hence the rank, is three.
(b) Inspection of the columns shows that the others are multiples of the first (inspec-
tion of the rows shows the same thing). Thus the rank is one.
Alternatively, the reduction
1 −1 2 1 −1 2
−3ρ1 +ρ2
3 −3 6 −→ 0 0 0
2ρ1 +ρ3
−2 2 −4 0 0 0
shows the same thing.
Answers to Exercises 103
Two.III.3.23 Only the zero matrices have rank of zero. The only matrices of rank one
have the form
k1 · ρ
..
.
km · ρ
where ρ is some nonzero row vector, and not all of the ki ’s are zero. (Remark. We
can’t simply say that all of the rows are multiples of the first because the first row
might be the zero row. Another Remark. The above also applies with ‘column’
replacing ‘row’.)
Two.III.3.24 If a 6= 0 then a choice of d = (c/a)b will make the second row be a multiple
of the first, specifically, c/a times the first. If a = 0 and b = 0 then any non-0 choice
for d will ensure that the second row is nonzero. If a = 0 and b 6= 0 and c = 0 then
any choice for d will do, since the matrix will automatically have rank one (even with
the choice of d = 0). Finally, if a = 0 and b 6= 0 and c 6= 0 then no choice for d will
suffice because the matrix is sure to have rank two.
Two.III.3.25 The column rank is two. One way to see this is by inspection — the column
space consists of two-tall columns and so can have a dimension of at least two, and we
can easily find two columns that together form a linearly independent set (the fourth
and fifth columns, for instance). Another way to see this is to recall that the column
rank equals the row rank, and to perform Gauss’s Method, which leaves two nonzero
rows.
Two.III.3.26 We apply Theorem 3.13. The number of columns of a matrix of coefficients
A of a linear system equals the number n of unknowns. A linear system with at
least one solution has at most one solution if and only if the space of solutions of
the associated homogeneous system has dimension zero (recall: in the ‘General =
Particular + Homogeneous’ equation ~v = ~p + ~h, provided that such a ~p exists, the
solution ~v is unique if and only if the vector ~h is unique, namely ~h = ~0). But that
means, by the theorem, that n = r.
Two.III.3.27 The set of columns must be dependent because the rank of the matrix is
at most five while there are nine columns.
Two.III.3.28 There is little danger of their being equal since the row space is a set of
row vectors while the column space is a set of columns (unless the matrix is 1×1, in
which case the two spaces must be equal).
Remark. Consider !
1 3
A=
2 6
Answers to Exercises 105
and note that the row space is the set of all multiples of (1 3) while the column space
consists of multiples of !
1
2
so we also cannot argue that the two spaces must be simply transposes of each other.
Two.III.3.29 First, the vector space is the set of four-tuples of real numbers, under the
natural operations. Although this is not the set of four-wide row vectors, the difference
is slight — it is “the same” as that set. So we will treat the four-tuples like four-wide
vectors.
With that, one way to see that (1, 0, 1, 0) is not in the span of the first set is to
note that thisreduction
1 −1 2 −3 1 −1 2 −3
−ρ1 +ρ2 −ρ2 +ρ3
1 1 2 0 −→ −→ 0 2 0 3
−3ρ1 +ρ3
3 −1 6 −6 0 0 0 0
and this one
1 −1 2 −3 1 −1 2 −3
1 1 2 0
ρ3 ↔ρ4 0 2 0 3
−ρ1 +ρ2 −ρ2 +ρ3
−→ −→ −→
3 −1 6 −6 0 0 −1 3/2
−3ρ 1 +ρ3 −(1/2)ρ2 +ρ4
−ρ1 +ρ4
1 0 1 0 0 0 0 0
yield matrices differing in rank. This means that addition of (1, 0, 1, 0) to the set of the
first three four-tuples increases the rank, and hence the span, of that set. Therefore
(1, 0, 1, 0) is not already in the span.
Two.III.3.30 It is a subspace because it is the column
space of the matrix
3 2 4
1 0 −1
2 2 5
of coefficients. To finda basis
forthe column space,
3 2 4
{c1 1 + c2 0 + c3 −1 | c1 , c2 , c3 ∈ R }
2 2 5
we take the three vectors
from the spanning set, transpose,
reduce,
3 1 2 3 1 2
−(2/3)ρ1 +ρ2 −(7/2)ρ2 +ρ3
2 0 2 −→ −→ 0 −2/3 2/3
−(4/3)ρ1 +ρ3
4 −1 5 0 0 0
and transpose back to get this.
3 0
h1 , −2/3i
2 2/3
106 Linear Algebra, by Hefferon
A yields the same nonzero rows as C, which yields the same nonzero rows as D,
which yields the same nonzero rows as B.
Two.III.3.33 It cannot be bigger.
Two.III.3.34 The number of rows in a maximal linearly independent set cannot exceed
the number of rows. A better bound (the bound that is, in general, the best possible)
is the minimum of m and n, because the row rank equals the column rank.
Two.III.3.35 Because the rows of a matrix A are the columns of AT the dimension of the
row space of A equals the dimension of the column space of AT . But the dimension
of the row space of A is the rank of A and the dimension of the column space of AT
is the rank of AT . Thus the two ranks are equal.
Two.III.3.36 False. The first is a set of columns while the second is a set of rows.
This example, however,
! 1 4
1 2 3
A= , AT = 2 5
4 5 6
3 6
indicates that as soon as we have a formal meaning for “the same”, we can apply it
here: ! ! !
1 2 3
Columnspace(A) = [{ , , }]
4 5 6
while
Rowspace(AT ) = [{(1 4), (2 5), (3 6) }]
are “the same” as each other.
Two.III.3.37 No. Here, Gauss’s Method does not change the column space.
! !
1 0 −3ρ1 +ρ2 1 0
−→
3 1 0 1
Two.III.3.38 A linear system
~ n = ~d
~ 1 + · · · + cn a
c1 a
has a solution if and only if ~d is in the span of the set { a ~ n }. That’s true if and
~ 1, . . . , a
only if the column rank of the augmented matrix equals the column rank of the matrix
of coefficients. Since rank equals the column rank, the system has a solution if and
only if the rank of its augmented matrix equals the rank of its matrix of coefficients.
Two.III.3.39 (a) Row rank equals column rank so each is at most the minimum of the
number of rows and columns. Hence both can be full only if the number of rows
equals the number of columns. (Of course, the converse does not hold: a square
matrix need not have full row rank or full column rank.)
108 Linear Algebra, by Hefferon
(b) If A has full row rank then, no matter what the right-hand side, Gauss’s Method
on the augmented matrix ends with a leading one in each row and none of those lead-
ing ones in the furthest right column (the “augmenting” column). Back substitution
then gives a solution.
On the other hand, if the linear system lacks a solution for some right-hand side
it can only be because Gauss’s Method leaves some row so that it is all zeroes to
the left of the “augmenting” bar and has a nonzero entry on the right. Thus, if A
does not have a solution for some right-hand sides, then A does not have full row
rank because some of its rows have been eliminated.
(c) The matrix A has full column rank if and only if its columns form a linearly
independent set. That’s equivalent to the existence of only the trivial linear
relationship among the columns, so the only solution of the system is where each
variable is 0.
(d) The matrix A has full column rank if and only if the set of its columns is linearly
independent, and so forms a basis for its span. That’s equivalent to the existence of
a unique linear representation of all vectors in that span. That proves it, since any
linear representation of a vector in the span is a solution of the linear system.
Two.III.3.40 Instead of the row spaces being the same, the row space of B would be a
subspace (possibly equal to) the row space of A.
Two.III.3.41 Clearly rank(A) = rank(−A) as Gauss’s Method allows us to multiply all
rows of a matrix by −1. In the same way, when k 6= 0 we have rank(A) = rank(kA).
Addition is more interesting. The rank of a sum can be smaller than the rank of
the summands. ! ! !
1 2 −1 −2 0 0
+ =
3 4 −3 −4 0 0
The rank of a sum can be bigger ! than the rank
! of the summands.
!
1 2 0 0 1 2
+ =
0 0 3 4 3 4
But there is an upper bound (other than the size of the matrices). In general,
rank(A + B) 6 rank(A) + rank(B).
To prove this, note that we can perform Gaussian elimination on A + B in either
of two ways: we can first add A to B and then apply the appropriate sequence of
reduction steps
step1 stepk
(A + B) −→ ··· −→ echelon form
or we can get the same results by performing step1 through stepk separately on A
and B, and then adding. The largest rank that we can end with in the second case is
clearly the sum of the ranks. (The matrices above give examples of both possibilities,
rank(A + B) < rank(A) + rank(B) and rank(A + B) = rank(A) + rank(B), happening.)
Answers to Exercises 109
Two.III.4.39 It happens when at least one of W1 , W2 is trivial. But that is the only way
it can happen.
To prove this, assume that both are non-trivial, select nonzero vectors w ~ 1, w
~ 2 from
each, and consider w ~ 1 +w~ 2 . This sum is not in W1 because w~ 1 +w ~ 2 = ~v ∈ W1 would
imply that w ~ 2 = ~v − w~ 1 is in W1 , which violates the assumption of the independence
of the subspaces. Similarly, w ~1+w ~ 2 is not in W2 . Thus there is an element of V that
is not in W1 ∪ W2 .
Two.III.4.40 Yes. The left-to-right implication is Corollary 4.13. For the other direc-
tion, assume that dim(V) = dim(W1 ) + · · · + dim(Wk ). Let B1 , . . . , Bk be bases
for W1 , . . . , Wk . As V is the sum of the subspaces, we can write any ~v ∈ V as
~v = w ~ 1 + ··· + w ~ k and expressing each w
~ i as a combination of vectors from the
_ _
associated basis Bi shows that the concatenation B1 · · · Bk spans V. Now, that
concatenation has dim(W1 ) + · · · + dim(Wk ) members, and so it is a spanning set of
size dim(V). The concatenation is therefore a basis for V. Thus V is the direct sum.
Two.III.4.41 No. The standard basis for R2 does not split into bases for the complemen-
tary subspaces the line x = y and the line x = −y.
Two.III.4.43 (a) They are equal because for each, V is the direct sum if and only if we
can write each ~v ∈ V in a unique way as a sum ~v = w
~1+w ~ 2 and ~v = w
~2+w ~ 1.
(b) They are equal because for each, V is the direct sum if and only if we can write
each ~v ∈ V in a unique way as a sum of a vector from each ~v = (~ w1 + w~ 2) + w
~3
and ~v = w~ 1 + (~ ~ 3 ).
w2 + w
(c) We can decompose any vector in R3 uniquely into the sum of a vector from each
axis.
(d) No. For an example, in R2 take W1 to be the x-axis, take W2 to be the y-axis,
and take W3 to be the line y = x.
(e) In any vector space the trivial subspace acts as the identity element with respect
to direct sum.
(f) In any vector space, only the trivial subspace has a direct-sum inverse (namely,
itself). One way to see this is that dimensions add, and so increase.
Topic: Fields
1 Going through the five conditions shows that they are all familiar from elementary
mathematics.
2 As with the prior question, going through the five conditions shows that for both of
these structures, the properties are familiar.
3 The integers fail condition (5). For instance, there is no multiplicative inverse for
2 — while 2 is an integer, 1/2 is not.
4 We can do these checks by listing all of the possibilities. For instance, to verify the
first half of condition (2) we must check that the structure is closed under addition and
that addition is commutative a + b = b + a, we can check both of these for all possible
pairs a and b because there are only four such pairs. Similarly, for associativity, there
are only eight triples a, b, c, and so the check is not too long. (There are other ways
to do the checks; in particular, you may recognize these operations as arithmetic
modulo 2. But an exhaustive check is not onerous)
5 These will do.
+ 0 1 2 · 0 1 2
0 0 1 2 0 0 0 0
1 1 2 0 1 0 1 2
2 2 0 1 2 0 2 1
As in the prior item, we could verify that they satisfy the conditions by listing all of
the cases.
Answers to Exercises 115
Topic: Crystals
1 Each fundamental unit is 3.34 × 10−10 cm, so there are about 0.1/(3.34 × 10−10 )
such units. That gives 2.99 × 108 , so there are something like 300, 000, 000 (three
hundred million) regions.
2 (a) We solve
! ! !
1.42 1.23 5.67 1.42c1 + 1.23c2 = 5.67
c1 + c2 = =⇒
0 0.71 3.14 0.71c2 = 3.14
So this point is in the next column of hexagons over, and either one hexagon up or
two hexagons up, depending on how you count them.
(c) This second basis
! !
1.42 0
h , i
0 1.42
(we get c2 ≈ 2.21 and c1 ≈ 3.99), but it doesn’t seem to have to do much with the
physical structure that we are studying.
3 In terms of the basis the locations of the corner atoms are (0, 0, 0), (1, 0, 0), . . . ,
(1, 1, 1). The locations of the face atoms are (0.5, 0.5, 1), (1, 0.5, 0.5), (0.5, 1, 0.5),
(0, 0.5, 0.5), (0.5, 0, 0.5), and (0.5, 0.5, 0). The locations of the atoms a quarter of
the way down from the top are (0.75, 0.75, 0.75) and (0.25, 0.25, 0.25). The atoms a
quarter of the way up from the bottom are at (0.75, 0.25, 0.25) and (0.25, 0.75, 0.25).
Converting to Ångstroms is easy.
4 (a) 195.08/6.02 × 1023 = 3.239 × 10−22
(b) Each platinum atom in the middle of each face is split between two cubes, so
that is 6/2 = 3 atoms so far. Each atom at a corner is split among eight cubes, so
that makes an additional 8/8 = 1 atom, so the total is 4.
(c) 4 · 3.239 × 10−22 = 1.296 × 10−21
(d) 1.296 × 10−21 /21.45 = 6.042 × 10−23 cubic centimeters
−8
× 10 centimeters.
(e) 3.924
3.924 × 10−8 0 0
(f) h 0 , 3.924 × 10−8 , 0 i
−8
0 0 3.924 × 10
do the computation for a voter from the second row, and a voter from the third row.
For a positive spin voter in the second row,
c1 − c2 − c3 = 1 c1 − c2 − c3 = 1
−ρ1 +ρ2 (−1/2)ρ2 +ρ3
c1 + c2 = 1 −→ −→ 2c2 + c3 = 0
−ρ1 +ρ3
c1 + c3 = −1 (3/2)c3 = −2
gives c3 = −4/3, c2 = 2/3, and c1 = 1/3. For a positive spin voter in the third row,
c1 − c2 − c3 = 1 c1 − c2 − c3 = 1
−ρ1 +ρ2 (−1/2)ρ2 +ρ3
c1 + c2 = −1 −→ −→ 2c2 + c3 = −2
−ρ1 +ρ3
c1 + c3 = 1 (3/2)c3 = 1
gives c3 = 2/3, c2 = −4/3, and c1 = 1/3.
3 The mock election corresponds to the table on page 153 in the way shown in the
first table, and after cancellation the result is the second table.
All three come from the same side of the table (the left), as the result from this Topic
says must happen. Tallying the election can now proceed, using the canceled numbers
D D D D
−1 1 1 −1 1 1 7 5
3· T R
+4· T R
+6· T R
= T R
1 1 −1 1
3· T R
+4· T R
+8· T R
= T R
1 1 −1 −1
T R
+ T R
+ T R
= T R
a b −c a+b−c
5 (a) A two-voter election can have a majority cycle in two ways. First, the two
voters could be opposites, resulting after cancellation in the trivial election (with
the majority cycle of all zeroes). Second, the two voters could have the same spin
but come from different rows, as here.
D D D D
−1 1 1 −1 1 1 0 0
1· T R
+1· T R
+0· T R
= T R
1 1 −1 2
(b) There are two cases. An even number of voters can split half and half into
opposites, e.g., half the voters are D > R > T and half are T > R > D. Then
cancellation gives the trivial election. If the number of voters is greater than one
and odd (of the form 2k + 1 with k > 0) then using the cycle diagram from the
proof,
D D D D
−a a b −b c c −a + b + c a−b+c
T R
+ T R
+ T R
= T R
a b −c a+b−c
we can take a = k and b = k and c = 1. Because k > 0, this is a majority cycle.
6 It is nonempty because it contains the zero vector. To see that it is closed under
linear combinations of two of its members, suppose that ~v1 and ~v2 are in U⊥ and
consider c1~v1 + c2~v2 . For any ~u ∈ U,
(c1~v1 + c2~v2 ) • ~u = c1 (~v1 • ~u) + c2 (~v2 • ~u) = c1 · 0 + c2 · 0 = 0
and so c1~v1 + c2~v2 ∈ U⊥ .
As to whether it holds if U is a subset but not a subspace, the answer is yes.
Taking p1 as parameter to express the torque gives this description of the solution
set.
1
−2
{ p1 | p1 ∈ R }
−5/3
−1
Denoting the torque by τ, the rotation rate by r, the volume of air by V, and the
density of air by d we have that Π1 = τr−2 V −5/3 d−1 , and so the torque is r2 V 5/3 d
times a constant.
4 (a) These are the dimensional formulas.
dimensional
quantity formula
speed of the wave v L1 M0 T −1
separation of the dominoes d L1 M0 T 0
height of the dominoes h L1 M0 T 0
acceleration due to gravity g L1 M0 T −2
(b) The relationship
(L1 M0 T −1 )p1 (L1 M0 T 0 )p2 (L1 M0 T 0 )p3 (L1 M0 T −2 )p4 = (L0 M0 T 0 )
gives this linear system.
p1 + p2 + p3 + p4 = 0
ρ1 +ρ4 p1 + p2 + p3 + p4 = 0
0 = 0 −→
p2 + p3 − p4 = 0
−p1 − 2p4 = 0
Taking p3 and p4 as parameters,
we can describe
the solution set in this way.
0 −2
−1 1
{ p3 + p 4 | p 3 , p 4 ∈ R }
1 0
0 1
That gives {Π1 = h/d, Π2 = dg/v } as a complete set.
2
(c) Buckingham’s Theorem says that v2 = dg · f̂(h/d) and so, since g is a constant,
√
if h/d is fixed then v is proportional to d .
5 Checking the conditions in the definition of a vector space is routine.
6 (a) The dimensional formula of the circumference is L, that is, L1 M0 T 0 . The
dimensional formula of the area is L2 .
(b) One is C + A = 2πr + πr2 .
(c) One example is this formula relating the the length of arc subtended by an angle
to the radius and the angle measure in radians: ` − rθ = 0. Both terms in that
formula have dimensional formula L1 . The relationship holds for some unit systems
(inches and radians, for instance) but not for all unit systems (inches and degrees,
for instance).
Chapter Three
Chapter Three: Maps
Between Spaces
Isomorphisms
(b) Denote the map from Example 1.2 by f. To show that it is one-to-one, assume
that f(a0 +a1 x+a2 x2 ) = f(b0 +b1 x+b2 x2 ). Then by the definition of the function,
a0 b0
a =
1 b 1
a2 b2
by f, namely, a + bx + cx2 .
As for structure, this shows that f preserves addition
f (a0 + a1 x + a2 x2 ) + (b0 + b1 x + b2 x2 )
f1 ( (c1 x + c2 y + c3 z) + (d1 x + d2 y + d3 z) )
= f1 ( (c1 + d1 )x + (c2 + d2 )y + (c3 + d3 )z )
= (c1 + d1 ) + (c2 + d2 )x + (c3 + d3 )x2
= (c1 + c2 x + c3 x2 ) + (d1 + d2 x + d3 x2 )
= f1 (c1 x + c2 y + c3 z) + f1 (d1 x + d2 y + d3 z)
Three.I.1.15 (a) No; this map is not one-to-one. In particular, the matrix of all zeroes
is mapped to the same image as the matrix of all ones.
(b) Yes, this is an isomorphism.
It is one-to-one:
! !
a1 b1 a2 b2
if f( ) = f( )
c1 d1 c2 d 2
a1 + b1 + c1 + d1 a2 + b2 + c2 + d2
a +b +c a +b +c
1 1 1 2 2 2
then =
a1 + b1 a2 + b2
a1 a2
gives that a1 = a2 , and that b1 = b2 , and that c1 = c2 , and that d1 = d2 .
It is onto, since this shows
x !
y w z−w
= f( )
z y−z x−y
w
that any four-tall vector is the image of a 2×2 matrix.
Answers to Exercises 127
c1 = c2
d1 + c1 = d2 + c2
b1 + a1 = b2 + a2
a1 = a2
We can check that f preserves structure by using item (2) of Lemma 1.11.
! !
a1 b1 a2 b2
f(r1 · + r2 · )
c1 d 1 c2 d 2
!
r1 a1 + r2 a2 r1 b1 + r2 b2
= f( )
r1 c1 + r2 c2 r1 d1 + r2 d2
= (r1 c1 + r2 c2 ) + (r1 d1 + r2 d2 + r1 c1 + r2 c2 )x
+ (r1 b1 + r2 b2 + r1 a1 + r2 a2 )x2 + (r1 a1 + r2 a2 )x3
= r1 · c1 + (d1 + c1 )x + (b1 + a1 )x2 + a1 x3
f(a0 + a1 x + a2 x2 + a3 x3 + a4 x4 + a5 x5 )
= a0 + a1 (x − 1) + a2 (x − 1)2 + a3 (x − 1)3
+ a4 (x − 1)4 + a5 (x − 1)5
= a0 + a1 (x − 1) + a2 (x2 − 2x + 1)
+ a3 (x3 − 3x2 + 3x − 1)
+ a4 (x4 − 4x3 + 6x2 − 4x + 1)
+ a5 (x5 − 5x4 + 10x3 − 10x2 + 5x − 1)
= (a0 − a1 + a2 − a3 + a4 − a5 )
+ (a1 − 2a2 + 3a3 − 4a4 + 5a5 )x
+ (a2 − 3a3 + 6a4 − 10a5 )x2
+ (a3 − 4a4 + 10a5 )x3
+ (a4 − 5a5 )x4 + a5 x5
This map is a correspondence because it has an inverse, the map p(x) 7→ p(x + 1).
To finish checking that it is an isomorphism we apply item (2) of Lemma 1.11 and
show that it preserves linear combinations of two polynomials. Briefly, f(c · (a0 +
a1 x + · · · + a5 x5 ) + d · (b0 + b1 x + · · · + b5 x5 )) equals this
(ca0 − ca1 + ca2 − ca3 + ca4 − ca5 + db0 − db1 + db2 − db3 + db4 − db5 )
+ · · · + (ca5 + db5 )x5
which equals c · f(a0 + a1 x + · · · + a5 x5 ) + d · f(b0 + b1 x + · · · + b5 x5 ).
Three.I.1.24 No vector space has the empty set underlying it. We can take ~v to be the
zero vector.
Three.I.1.25 Yes; where the two spaces are { a
~ } and { ~b }, the map sending a~ to ~b is
clearly one-to-one and onto, and also preserves what little structure there is.
Three.I.1.26 A linear combination of n = 0 vectors adds to the zero vector and so
Lemma 1.10 shows that the three statements are equivalent in this case.
Answers to Exercises 131
Three.I.1.27 Consider the basis h1i for P0 and let f(1) ∈ R be k. For any a ∈ P0 we
have that f(a) = f(a · 1) = af(1) = ak and so f’s action is multiplication by k. Note
that k 6= 0 or else the map is not one-to-one. (Incidentally, any such map a 7→ ka is
an isomorphism, as is easy to check.)
Three.I.1.28 In each item, following item (2) of Lemma 1.11, we show that the map
preserves structure by showing that the it preserves linear combinations of two
members of the domain.
(a) The identity map is clearly one-to-one and onto. For linear combinations the
check is easy.
id(c1 · ~v1 + c2 · ~v2 ) = c1~v1 + c2~v2 = c1 · id(~v1 ) + c2 · id(~v2 )
(b) The inverse of a correspondence is also a correspondence (as stated in the
appendix), so we need only check that the inverse preserves linear combinations.
Assume that w~ 1 = f(~v1 ) (so f−1 (~
w1 ) = ~v1 ) and assume that w ~ 2 = f(~v2 ).
f−1 (c1 · w ~ 2 ) = f−1 c1 · f(~v1 ) + c2 · f(~v2 )
~ 1 + c2 · w
= f−1 ( f c1~v1 + c2~v2 )
= c1~v1 + c2~v2
= c1 · f−1 (~
w1 ) + c2 · f−1 (~
w2 )
(c) The composition of two correspondences is a correspondence (as stated in the
appendix), so we need only check that the composition map preserves linear combi-
nations.
g ◦ f c1 · ~v1 + c2 · ~v2 = g f(c1~v1 + c2~v2 )
= g c1 · f(~v1 ) + c2 · f(~v2 )
= c1 · g f(~v1 )) + c2 · g(f(~v2 )
= c1 · g ◦ f (~v1 ) + c2 · g ◦ f (~v2 )
Three.I.1.29 One direction is easy: by definition, if f is one-to-one then for any w ~ ∈W
at most one ~v ∈ V has f(~v ) = w ~ , and so in particular, at most one member of V is
mapped to ~0W . The proof of Lemma 1.10 does not use the fact that the map is a
correspondence and therefore shows that any structure-preserving map f sends ~0V to
~0W .
For the other direction, assume that the only member of V that is mapped to ~0W is
~0V . To show that f is one-to-one assume that f(~v1 ) = f(~v2 ). Then f(~v1 ) − f(~v2 ) = ~0W
and so f(~v1 −~v2 ) = ~0W . Consequently ~v1 −~v2 = ~0V , so ~v1 = ~v2 , and so f is one-to-one.
Three.I.1.30 We will prove something stronger — not only is the existence of a depen-
dence preserved by isomorphism, but each instance of a dependence is preserved, that
132 Linear Algebra, by Hefferon
is,
~vi = c1~v1 + · · · + ci−1~vi−1 + ci+1~vi+1 + · · · + ck~vk
⇐⇒ f(~vi ) = c1 f(~v1 ) + · · · + ci−1 f(~vi−1 ) + ci+1 f(~vi+1 ) + · · · + ck f(~vk ).
The =⇒ direction of this statement holds by item (3) of Lemma 1.11. The ⇐=
direction holds by regrouping
f(~vi ) = c1 f(~v1 ) + · · · + ci−1 f(~vi−1 ) + ci+1 f(~vi+1 ) + · · · + ck f(~vk )
= f(c1~v1 + · · · + ci−1~vi−1 + ci+1~vi+1 + · · · + ck~vk )
and applying the fact that f is one-to-one, and so for the two vectors ~vi and c1~v1 +
· · · + ci−1~vi−1 + ci+1 f~vi+1 + · · · + ck f(~vk to be mapped to the same image by f, they
must be equal.
Three.I.1.31 (a) This map is one-to-one because if ds (~v1 ) = ds (~v2 ) then by definition
of the map, s · ~v1 = s · ~v2 and so ~v1 = ~v2 , as s is nonzero. This map is onto as any
~ ∈ R2 is the image of ~v = (1/s) · w
w ~ (again, note that s is nonzero). (Another way
to see that this map is a correspondence is to observe that it has an inverse: the
inverse of ds is d1/s .)
To finish, note that this map preserves linear combinations
ds (c1 · ~v1 + c2 · ~v2 ) = s(c1~v1 + c2~v2 ) = c1 s~v1 + c2 s~v2 = c1 · ds (~v1 ) + c2 · ds (~v2 )
and therefore is an isomorphism.
(b) As in the prior item, we can show that the map tθ is a correspondence by noting
that it has an inverse, t−θ .
That the map preserves structure is geometrically easy to see. For instance,
adding two vectors and then rotating them has the same effect as rotating first
and then adding. For an algebraic argument, consider polar coordinates: the map
tθ sends the vector with endpoint (r, φ) to the vector with endpoint (r, φ + θ).
Then the familiar trigonometric formulas cos(φ + θ) = cos φ cos θ − sin φ sin θ and
sin(φ + θ) = sin φ cos θ + cos φ sin θ show how to express the map’s action in the
usual rectangular coordinate system.
! ! ! !
x r cos φ tθ r cos(φ + θ) x cos θ − y sin θ
= 7−→ =
y r sin φ r sin(φ + θ) x sin θ + y cos θ
Now the calculation for preservation of addition is routine.
! !
x1 + x2 tθ (x1 + x2 ) cos θ − (y1 + y2 ) sin θ
7−→
y1 + y2 (x1 + x2 ) sin θ + (y1 + y2 ) cos θ
! !
x1 cos θ − y1 sin θ x2 cos θ − y2 sin θ
= +
x1 sin θ + y1 cos θ x2 sin θ + y2 cos θ
The calculation for preservation of scalar multiplication is similar.
Answers to Exercises 133
f`
φ 7−→
θ φ − (θ − φ)
Three.I.1.33 (a) For the ‘only if’ half, let f : R1 → R1 to be an isomorphism. Consider
the basis h1i ⊆ R1 . Designate f(1) by k. Then for any x we have that f(x) =
f(x · 1) = x · f(1) = xk, and so f’s action is multiplication by k. To finish this half,
just note that k 6= 0 or else f would not be one-to-one.
For the ‘if’ half we only have to check that such a map is an isomorphism when
k 6= 0. To check that it is one-to-one, assume that f(x1 ) = f(x2 ) so that kx1 = kx2
and divide by the nonzero factor k to conclude that x1 = x2 . To check that it is
onto, note that any y ∈ R1 is the image of x = y/k (again, k 6= 0). Finally, to check
that such a map preserves combinations of two members of the domain, we have
this.
f(c1 x1 + c2 x2 ) = k(c1 x1 + c2 x2 ) = c1 kx1 + c2 kx2 = c1 f(x1 ) + c2 f(x2 )
(b) By the prior item, f’s action is x 7→ (7/3)x. Thus f(−2) = −14/3.
(c) For the ‘only if’ half, assume that f : R2 → R2 is an automorphism. Consider the
standard basis E2 for R2 . Let ! !
a b
f(~e1 ) = and f(~e2 ) = .
c d
Then the action of f on any vector is determined by by its action on the two basis
vectors.
! ! ! !
x a b ax + by
f( ) = f(x · ~e1 + y · ~e2 ) = x · f(~e1 ) + y · f(~e2 ) = x · +y· =
y c d cx + dy
To finish this half, note that if ad − bc = 0, that is, if f(~e2 ) is a multiple of f(~e1 ),
then f is not one-to-one.
For ‘if’ we must check that the map is an isomorphism, under the condition
that ad − bc 6= 0. The structure-preservation check is easy; we will here show that
f is a correspondence.
! For the
! argument that the map!is one-to-one, assume ! this.
x1 x2 ax1 + by1 ax2 + by2
f( ) = f( ) and so =
y1 y2 cx1 + dy1 cx2 + dy2
Then, because ad − bc 6= 0, the resulting system
a(x1 − x2 ) + b(y1 − y2 ) = 0
c(x1 − x2 ) + d(y1 − y2 ) = 0
has a unique solution, namely the trivial one x1 − x2 = 0 and y1 − y2 = 0 (this
follows from the hint).
The argument that this map is onto is closely related — this system
ax1 + by1 = x
cx1 + dy1 = y
has a solution for any x and y if and only
! if this
! set
a b
{ , }
c d
Answers to Exercises 135
spans R2 , i.e., if and only if this set is a basis (because it is a two-element subset of
R2 ), i.e., if and only if ad − bc 6= 0.
(d)
! ! ! ! ! ! ! !
0 1 1 1 1 2 0 2
f( ) = f( − ) = f( ) − f( )= − =
−1 3 4 3 4 −1 1 −2
Three.I.1.34 There are many answers; two are linear independence and subspaces.
First we show that if a set {~v1 , . . . ,~vn } is linearly independent then its image
{f(~v1 ), . . . , f(~vn ) } is also linearly independent. Consider a linear relationship among
members of the image set.
0 = c1 f(~v1 ) + · · · + cn f(v~n ) = f(c1~v1 ) + · · · + f(cn v~n ) = f(c1~v1 + · · · + cn v~n )
Because this map is an isomorphism, it is one-to-one. So f maps only one vector from
the domain to the zero vector in the range, that is, c1~v1 + · · · + cn~vn equals the zero
vector (in the domain, of course). But, if {~v1 , . . . ,~vn } is linearly independent then all
of the c’s are zero, and so { f(~v1 ), . . . , f(~vn ) } is linearly independent also. (Remark.
There is a small point about this argument that is worth mention. In a set, repeats
collapse, that is, strictly speaking, this is a one-element set: {~v,~v }, because the things
listed as in it are the same thing. Observe, however, the use of the subscript n in
the above argument. In moving from the domain set {~v1 , . . . ,~vn } to the image set
{f(~v1 ), . . . , f(~vn ) }, there is no collapsing, because the image set does not have repeats,
because the isomorphism f is one-to-one.)
To show that if f : V → W is an isomorphism and if U is a subspace of the domain V
then the set of image vectors f(U) = { w ~ ∈W|w ~ = f(~u) for some ~u ∈ U} is a subspace
of W, we need only show that it is closed under linear combinations of two of its
members (it is nonempty because it contains the image of the zero vector). We have
c1 · f(~u1 ) + c2 · f(~u2 ) = f(c1 ~u1 ) + f(c2 ~u2 ) = f(c1 ~u1 + c2 ~u2 )
and c1 ~u1 + c2 ~u2 is a member of U because of the closure of a subspace under
combinations. Hence the combination of f(~u1 ) and f(~u2 ) is a member of f(U).
Three.I.1.35 (a) The association
c1
RepB (·)
~ ~ ~
~p = c1 β1 + c2 β2 + c3 β3 7−→ c2
c3
is a function if every member ~p of the domain is associated with at least one member
of the codomain, and if every member ~p of the domain is associated with at most
one member of the codomain. The first condition holds because the basis B spans
~
the domain — every ~p can be written as at least one linear combination of β’s. The
second condition holds because the basis B is linearly independent — every member
~p of the domain can be written as at most one linear combination of the β’s. ~
136 Linear Algebra, by Hefferon
~1 +
(b) For the one-to-one argument, if RepB (~p) = RepB (~q), that is, if RepB (p1 β
~ 2 + p3 β
p2 β ~ 3 ) = RepB (q1 β
~ 1 + q2 β ~ +q β ~ 3 ) then
2 3
p1 q1
p2 = q2
p3 q3
and so p1 = q1 and p2 = q2 and p3 = q3 , which gives the conclusion that ~p = ~q.
Therefore this map is one-to-one.
For onto, we can just note that
a
b
c
equals RepB (aβ ~ 1 + bβ
~ 2 + cβ~ 3 ), and so any member of the codomain R3 is the
image of some member of the domain P2 .
(c) This map respects addition and scalar multiplication because it respects combina-
tions of two members of the domain (that is, we are using item (2) of Lemma 1.11):
where ~p = p1 β~ 1 + p2 β~ 2 + p3 β
~ 3 and ~q = q1 β
~ 1 + q2 β
~ 2 + q3 β
~ 3 , we have this.
RepB (c · ~p + d · ~q) = RepB ( (cp1 + dq1 )β~ 1 + (cp2 + dq2 )β ~ 2 + (cp3 + dq3 )β ~3 )
cp1 + dq1
= cp2 + dq2
cp3 + dq3
p1 q1
= c · p2 + d · q2
p3 q3
= RepB (~p) + RepB (~q)
(d) Use any basis B for P2 whose first two members are x + x2 and 1 − x, say
B = hx + x2 , 1 − x, 1i.
Three.I.1.36 See the next subsection.
Three.I.1.37 (a) Most of the conditions in the definition of a vector space are routine.
We here sketch the verification of part (1) of that definition.
For closure of U × W, note that because U and W are closed, we have that
~u1 + ~u2 ∈ U and w ~ 1 +w ~ 2 ∈ W and so (~u1 + ~u2 , w ~ 1 +w ~ 2 ) ∈ U × W. Commutativity
of addition in U × W follows from commutativity of addition in U and W.
~ 1 )+(~u2 , w
(~u1 , w ~ 2 ) = (~u1 +~u2 , w
~ 1 +w
~ 2 ) = (~u2 +~u1 , w
~ 2 +w ~ 1 ) = (~u2 , w
~ 2 )+(~u1 , w
~ 1)
The check for associativity of addition is similar. The zero element is (0U , 0W ) ∈ ~ ~
U × W and the additive inverse of (~u, w ~ ) is (−~u, −~w).
The checks for the second part of the definition of a vector space are also
straightforward.
Answers to Exercises 137
because there is one and only one way to represent any member of P2 × R2 with
respect to this set; here is an example.
! ! ! ! ! !
2 5 0 0 2 0 1 0
(3+2x+x , ) = 3·(1, )+2·(x, )+(x , )+5·(1, )+4·(1, )
4 0 0 0 0 1
f( c1 · (~u1 , w
~ 1 ) + c2 · (~u2 , w
~ 2 ) ) = f( (c1 ~u1 + c2 ~u2 , c1 w
~ 1 + c2 w
~ 2) )
= c1 ~u1 + c2 ~u2 + c1 w
~ 1 + c2 w
~2
= c1 ~u1 + c1 w
~ 1 + c2 ~u2 + c2 w
~2
= c1 · f( (~u1 , w
~ 1 ) ) + c2 · f( (~u2 , w
~ 2) )
and so it is an isomorphism.
Three.I.2.9 Each pair of spaces is isomorphic if and only if the two have the same
dimension. We can, when there is an isomorphism, state a map, but it isn’t strictly
necessary.
(a) No, they have different dimensions.
(b) No, they have different dimensions.
138 Linear Algebra, by Hefferon
(c) Yes, they have the same dimension. One isomorphism is this.
! a
a b c ..
7→ .
d e f
f
(d) Yes, they have the same dimension. This is an isomorphism.
!
5 a b c
a + bx + · · · + fx 7→
d e f
(e) Yes, both have dimension 2k.
! ! !
5 0 −1
Three.I.2.10 (a) RepB (3 − 2x) = (b) (c)
−2 2 1
Three.I.2.11 They have different dimensions.
Three.I.2.12 Yes, both are mn-dimensional.
Three.I.2.13 Yes, any two (nondegenerate) planes are both two-dimensional vector
spaces.
Three.I.2.14 There are many answers, one is the set of Pk (taking P−1 to be the trivial
vector space).
Three.I.2.15 False (except when n = 0). For instance, if f : V → Rn is an isomorphism
then multiplying by any nonzero scalar, gives another, different, isomorphism. (Be-
tween trivial spaces the isomorphisms are unique; the only map possible is ~0V 7→ 0W .)
Three.I.2.16 No. A proper subspace has a strictly lower dimension than it’s superspace;
if U is a proper subspace of V then any linearly independent subset of U must have
fewer than dim(V) members or else that set would be a basis for V, and U wouldn’t
be proper.
Three.I.2.17 Where B = hβ ~ 1, . . . , β
~ n i, the inverse is this.
c1
.. ~ 1 + · · · + cn β
~n
. 7→ c1 β
cn
Three.I.2.18 All three spaces have dimension equal to the rank of the matrix.
Three.I.2.19 We must show that if a ~ = ~b then f(~ a) = f(~b). So suppose that a1 β
~1 +
~ n = b1 β
· · · + an β ~ 1 + · · · + bn β ~ n . Each vector in a vector space (here, the domain
space) has a unique representation as a linear combination of basis vectors, so we can
conclude that a1 = b1 , . . . , an = bn . Thus,
a1 b1
.. ..
a) = . = . = f(~b)
f(~
an bn
and so the function is well-defined.
Answers to Exercises 139
Three.I.2.22 One direction is easy: if the two are isomorphic via f then for any basis
B ⊆ V, the set D = f(B) is also a basis (this is shown in Lemma 2.4). The check
that corresponding vectors have the same coordinates: f(c1 β ~ 1 + · · · + cn β
~ n) =
~ ~ ~ ~
c1 f(β1 ) + · · · + cn f(βn ) = c1 δ1 + · · · + cn δn is routine.
For the other half, assume that there are bases such that corresponding vectors have
the same coordinates with respect to those bases. Because f is a correspondence, to
show that it is an isomorphism, we need only show that it preserves structure. Because
RepB (~v ) = RepD (f(~v )), the map f preserves structure if and only if representations
preserve addition: RepB (~v1 + ~v2 ) = RepB (~v1 ) + RepB (~v2 ) and scalar multiplication:
RepB (r · ~v ) = r · RepB (~v ) The addition calculation is this: (c1 + d1 )β~ 1 + · · · + (cn +
~ n = c1 β
dn ) β ~ 1 + · · · + cn β
~ n + d1 β
~ 1 + · · · + dn β
~ n , and the scalar multiplication
calculation is similar.
Three.I.2.24 Yes.
First, f̂ is well-defined because every member of V has one and only one represen-
tation as a linear combination of elements of B.
Second we must show that f̂ is one-to-one and onto. It is one-to-one because every
member of W has only one representation as a linear combination of elements of
D, since D is a basis. And f̂ is onto because every member of W has at least one
representation as a linear combination of members of D.
Finally, preservation of structure is routine to check. For instance, here is the
140 Linear Algebra, by Hefferon
Homomorphisms
Three.II.1: Definition
Three.II.1.18 (a) Yes. The verification is straightforward.
x1 x2 c1 x1 + c2 x2
h(c1 · y1 + c2 · y2 ) = h(c1 y1 + c2 y2 )
z1 z2 c1 z 1 + c2 z 2
!
c1 x1 + c2 x2
=
c1 x1 + c2 x2 + c1 y1 + c2 y2 + c1 z1 + c2 z2
! !
x1 x2
= c1 · + c2 ·
x1 + y1 + z1 c2 + y 2 + z 2
x1 x2
= c1 · h(y1 ) + c2 · h(y2 )
z1 z2
Answers to Exercises 141
(b) This argument is similar to the prior one. Let ~v ∈ V be represented with
respect to the basis as ~v = c1 β ~ 1 + · · · + cn β
~ n . Then h(c1 β ~ 1 + · · · + cn β
~ n) =
~ ~ ~ ~
c1 h(β1 ) + · · · + cn h(βn ) = c1 β1 + · · · + cn βn = ~v.
(c) As above, only c1 h(β ~ 1 ) + · · · + cn h(β~ n ) = c1 rβ
~ 1 + · · · + cn rβ
~ n = r(c1 β~1 +···+
~ n ) = r~v.
cn β
Three.II.1.27 That it is a homomorphism follows from the familiar rules that the
logarithm of a product is the sum of the logarithms ln(ab) = ln(a) + ln(b) and that
the logarithm of a power is the multiple of the logarithm ln(ar ) = r ln(a). This map
is an isomorphism because it has an inverse, namely, the exponential map, so it is a
correspondence, and therefore it is an isomorphism.
Three.II.1.28 Where x̂ = x/2 and ŷ = y/3, the image set is
! !
x̂ (2x̂)2 (3ŷ)2 x̂
{ | 4 + 9 = 1} = { | x̂2 + ŷ2 = 1 }
ŷ ŷ
the unit circle in the x̂ŷ-plane.
Three.II.1.29 The circumference function r 7→ 2πr is linear. Thus we have 2π · (rearth +
6) − 2π · (rearth ) = 12π. Observe that it takes the same amount of extra rope to raise
the circle from tightly wound around a basketball to six feet above that basketball as
it does to raise it from tightly wound around the earth to six feet above the earth.
Three.II.1.30 Verifying that it is linear is routine.
x1 x2 c1 x1 + c2 x2
h(c1 · y1 + c2 · y2 ) = h(c1 y1 + c2 y2 )
z1 z2 c1 z 1 + c2 z 2
= 3(c1 x1 + c2 x2 ) − (c1 y1 + c2 y2 ) − (c1 z1 + c2 z2 )
= c1 · (3x1 − y1 − z1 ) + c2 · (3x2 − y2 − z2 )
x1 x2
= c1 · h(y1 ) + c2 · h(y2 )
z1 z2
The natural guess at a generalization is that for any fixed ~k ∈ R3 the map ~v 7→ ~v • ~k
is linear. This statement is true. It follows from properties of the dot product we
have seen earlier: (~v + ~u) • ~k = ~v • ~k + ~u • ~k and (r~v) • ~k = r(~v • ~k). (The natural guess
at a generalization of this generalization, that the map from Rn to R whose action
consists of taking the dot product of its argument with a fixed vector ~k ∈ Rn is linear,
is also true.)
Three.II.1.31 Let h : R1 → R1 be linear. A linear map is determined by its action on a
basis, so fix the basis h1i for R1 . For any r ∈ R1 we have that h(r) = h(r · 1) = r · h(1)
and so h acts on any argument r by multiplying it by the constant h(1). If h(1) is
Answers to Exercises 145
not zero then the map is a correspondence — its inverse is division by h(1) — so any
nontrivial transformation of R1 is an isomorphism.
This projection map is an example that shows that not every transformation of
Rn acts via multiplication by a constant
when n> 1, including when n = 2.
x1 x1
x2 0
. 7→ .
. .
. .
xn 0
Three.II.1.32 (a) Where c and d are scalars, we have this.
x1 y1
.. ..
h(c · . + d · . )
xn yn
cx1 + dy1
..
= h( )
.
cxn + dyn
a1,1 (cx1 + dy1 ) + · · · + a1,n (cxn + dyn )
..
=
.
am,1 (cx1 + dy1 ) + · · · + am,n (cxn + dyn )
a1,1 x1 + · · · + a1,n xn a1,1 y1 + · · · + a1,n yn
.. ..
=c· +d·
. .
am,1 x1 + · · · + am,n xn am,1 y1 + · · · + am,n yn
x1 y1
.. ..
= c · h( . ) + d · h( . )
xn yn
(b) Each power i of the derivative operator is linear because of these rules familiar
from calculus.
di di di di di
i
( f(x) + g(x) ) = i
f(x) + i g(x) i
r · f(x) = r · i f(x)
dx dx dx dx dx
Thus the given map is a linear transformation of Pn because any linear combination
of linear maps is also a linear map.
Three.II.1.33 (This argument has already appeared, as part of the proof that isomor-
phism is an equivalence.) Let f : U → V and g : V → W be linear. The composition
preserves linear combinations
g ◦ f(c1 ~u1 + c2 ~u2 ) = g( f(c1 ~u1 + c2 ~u2 ) ) = g( c1 f(~u1 ) + c2 f(~u2 ) )
= c1 · g(f(~u1 )) + c2 · g(f(~u2 )) = c1 · g ◦ f(~u1 ) + c2 · g ◦ f(~u2 )
146 Linear Algebra, by Hefferon
Three.II.1.35 Recall that the entry in row i and column j of the transpose of M is the
entry mj,i from row j and column i of M. Now, the check is routine. Start with the
transpose of the combination.
.. .. T
. .
[r ·
· · · ai,j · · · + s · · · · bi,j · · ·]
.. ..
. .
Combine and take the transpose.
.. T ..
. .
= · · · rai,j + sbi,j · · ·
= · · · raj,i + sbj,i
· · ·
.. ..
. .
Answers to Exercises 147
(a) The range is nonempty because V is nonempty. To finish we need to show that it
is closed under combinations. A combination of range vectors has the form, where
~v1 , . . . ,~vn ∈ V,
Three.II.1.40 No; the set of isomorphisms does not contain the zero map (unless the
space is trivial).
~ 1, . . . , β
Three.II.1.41 If hβ ~ n i doesn’t span the space then the map needn’t be unique.
For instance, if we try to define a map from R2 to itself by specifying only that ~e1
maps to itself, then there is more than one homomorphism possible; both the identity
map and the projection map onto the first component fit this condition.
If we drop the condition that hβ ~ 1, . . . , β
~ n i is linearly independent then we risk
an inconsistent specification (i.e, there could be no such map). An example is if
we consider h~e2 , ~e1 , 2~e1 i, and try to define a map from R2 to itself that sends ~e2 to
itself, and sends both ~e1 and 2~e1 to ~e1 . No homomorphism can satisfy these three
conditions.
Answers to Exercises 149
no member of the domain that when multiplied by x gives the constant polynomial
p(x) = 7.
(d) The polynomial 12x−0.5x3 ∈ P3 is not in the null space because h(12x−0.5x3 ) =
12x2 − 0.5x4 . The polynomial 12x − 0.5x3 ∈ P4 is in the range space because it is
the image of 12 − 0.5x2 .
(e) The polynomial 1+3x2 −x3 ∈ P3 is not in the null space because h(1+3x2 −x3 ) =
x + 3x3 − x4 . The polynomial 1 + 3x2 − x3 ∈ P4 is not in the range space because
of the constant term.
Three.II.2.22 (a)!The null space is !
a 0
N (h) = { ∈ R | a + ax + ax + 0x = 0 + 0x + 0x + 0x } = {
2 2 3 2 3
| b ∈ R}
b b
while the range space is
R(h) = {a + ax + ax2 ∈ P3 | a, b ∈ R} = {a · (1 + x + x2 ) | a ∈ R}
and so the nullity is one and the rank is one.
(b) The null space is this. ! !
a b −d b
N (h) = { | a + d = 0} = { | b, c, d ∈ R }
c d c d
The range space
R(h) = {a + d | a, b, c, d ∈ R }
is all of R (we can get any real number by taking d to be 0 and taking a to be the
desired number). Thus, the nullity is three and the rank is one.
(c) The null space is ! !
a b −b − c b
N (h) = { | a + b + c = 0 and d = 0 } = { | b, c ∈ R}
c d c 0
while the range space is R(h) = {r + sx2 | r, s ∈ R}. Thus, the nullity is two and
the rank is two.
(d) The null space is all of R3 so the nullity is three. The range space is the trivial
subspace of R4 so the rank is zero.
Three.II.2.23 For each, use the result that the rank plus the nullity equals the dimension
of the domain.
(a) 0 (b) 3 (c) 3 (d) 0
Three.II.2.24 Because
d
(a0 + a1 x + · · · + an xn ) = a1 + 2a2 x + 3a3 x2 + · · · + nan xn−1
dx
we have this.
d
N ( ) = {a0 + · · · + an xn | a1 + 2a2 x + · · · + nan xn−1 = 0 + 0x + · · · + 0xn−1 }
dx
= {a0 + · · · + an xn | a1 = 0, and a2 = 0, . . . , an = 0 }
= {a0 + 0x + 0x2 + · · · + 0xn | a0 ∈ R}
Answers to Exercises 151
2x + y = −3 2x + y = 1
Three.II.2.33 (a) One direction is obvious: if the homomorphism is onto then its range
is the codomain and so its rank equals the dimension of its codomain. For the other
direction assume that the map’s rank equals the dimension of the codomain. Then
the map’s range is a subspace of the codomain, and has dimension equal to the
dimension of the codomain. Therefore, the map’s range must equal the codomain,
and the map is onto. (The ‘therefore’ is because there is a linearly independent
subset of the range that is of size equal to the dimension of the codomain, but any
such linearly independent subset of the codomain must be a basis for the codomain,
and so the range equals the codomain.)
(b) By Theorem 2.20, a homomorphism is one-to-one if and only if its nullity is
zero. Because rank plus nullity equals the dimension of the domain, it follows that
a homomorphism is one-to-one if and only if its rank equals the dimension of its
domain. But this domain and codomain have the same dimension, so the map is
one-to-one if and only if it is onto.
Three.II.2.40 By assumption, h is not the zero map and so a vector ~v ∈ V exists that
is not in the null space. Note that hh(~v)i is a basis for R, because it is a size-one
linearly independent subset of R. Consequently h is onto, as for any r ∈ R we have
r = c · h(~v) for some scalar c, and so r = h(c~v).
Thus the rank of h is one. Because the nullity is n, the dimension of the domain
of h, the vector space V, is n + 1. We can finish by showing {~v, β ~ n } is linearly
~ 1, . . . , β
independent, as it is a size n+1 subset of a dimension n+1 space. Because { β ~n}
~ 1, . . . , β
is linearly independent we need only show that ~v is not a linear combination of the
other vectors. But c1 β~ 1 + · · · + cn β ~ n = ~v would give −~v + c1 β
~ 1 + · · · + cn β
~ n = ~0 and
applying h to both sides would give a contradiction.
Three.II.2.41 Fix a basis hβ~ 1, . . . , β
~ n i for V. We shall prove that this map
~
h(β1 )
Φ .
h 7−→ ..
~ n)
h(β
∗ n
is an isomorphism from V to R .
To see that Φ is one-to-one, assume that h1 and h2 are members of V ∗ such that
Φ(h1 ) = Φ(h2 ). Then
~ 1 ) h2 (β
h1 ( β ~ 1)
.. ..
. = .
h1 (β~ n) ~ n)
h2 ( β
and consequently, h1 (β~ 1 ) = h2 (β
~ 1 ), etc. But a homomorphism is determined by its
action on a basis, so h1 = h2 , and therefore Φ is one-to-one.
To see that Φ is onto, consider
x1
..
.
xn
for x1 , . . . , xn ∈ R. This function h from V to R
~ 1 + · · · + cn β
c1 β ~ n 7−h→ c1 x1 + · · · + cn xn
is linear and Φ maps it to the given vector in Rn , so Φ is onto.
The map Φ also preserves structure: where
~ 1 + · · · + cn β
c1 β ~ n 7−h→1 ~ 1 ) + · · · + cn h 1 ( β
c1 h 1 ( β ~ n)
c1 β ~ n 7−h→
~ 1 + · · · + cn β 2 ~ 1 ) + · · · + cn h 2 ( β
c1 h 2 ( β ~ n)
we have
~ 1 + · · · + cn β
(r1 h1 + r2 h2 )(c1 β ~ n)
~ 1 ) + r2 h2 (β
= c1 (r1 h1 (β ~ 1 )) + · · · + cn (r1 h1 (β~ n ) + r2 h2 (β~ n ))
~ 1 ) + · · · + cn h 1 ( β
= r1 (c1 h1 (β ~ n )) + r2 (c1 h2 (β
~ 1 ) + · · · + cn h2 (β
~ n ))
156 Linear Algebra, by Hefferon
and so
2 0 ! 2v1
v1
RepE3 ( h(~v) ) = 2 1 = 2v1 + v2
v2
0 −1 −v2
is the desired representation.
Three.III.1.17 (a) We must first find the image of each vector from the domain’s
basis, and then represent that image with respect to the codomain’s basis.
0 1 0
d1 0 dx 0 d x2 2
RepB ( ) = RepB ( ) = RepB ( )=
dx 0 dx 0 dx 0
0 0 0
0
d x3 0
RepB ( )=
dx 3
0
Those representations are then adjoined to make the matrix representing the
map.
0 1 0 0
d 0 0 2 0
RepB,B ( ) =
dx 0 0 0 3
0 0 0 0
(b) Proceeding as in the prior item, we represent the images of the domain’s
Answers to Exercises 159
basis vectors
0 1 0
d1 0 dx 0 d x2 1
RepD ( )= RepD ( )= RepD ( )=
dx 0 dx 0 dx 0
0 0 0
0
3
dx 0
RepD ( )=
dx 1
0
Three.III.1.18 For each, we must find the image of each of the domain’s basis vectors,
represent each image with respect to the codomain’s basis, and then adjoin those
representations to get the matrix.
(a) The basis vectors from the domain have these images
1 7→ 0 x 7→ 1 x2 7→ 2x ...
and these images are represented with respect to the codomain’s basis in this way.
0 1 0
0 0 2
0 0 0
RepB (0) =
..
RepB (1) =
..
RepB (2x) =
..
. . .
0
0
0
n−1
... RepB (nx )=
..
.
n
0
160 Linear Algebra, by Hefferon
The matrix
0 1 0 ... 0
0 0 2 ... 0
d ..
RepB,B ( ) = .
dx
0 0 0 ... n
0 0 0 ... 0
has n + 1 rows and columns.
(b) Once the images under this map of the domain’s basis vectors are determined
1 7→ x x 7→ x2 /2 x2 7→ x3 /3 ...
then they can be represented with respect to the codomain’s basis
0 0
1 0
0 RepB 1/2
2
RepBn+1 (x) = n+1
(x /2) =
.. ..
. .
0
0
n+1 0
... RepBn+1 (x /(n + 1)) =
..
.
1/(n + 1)
and put together to make the matrix.
0 0 ... 0 0
Z
1 0 ... 0 0
0 1/2 ... 0 0
RepBn ,Bn+1 ( ) =
..
.
0 0 ... 0 1/(n + 1)
(c) The images of the basis vectors of the domain are
1 7→ 1 x 7→ 1/2 x2 7→ 1/3 ...
and they are represented with respect to the codomain’s basis as
RepE1 (1) = 1 RepE1 (1/2) = 1/2 ...
so the matrix is
Z
RepB,E1 ( ) = 1 1/2 · · · 1/n 1/(n + 1)
so that representing the images with respect to the codomain’s basis and adjoining
those column vectors together gives this.
1 0 0 0
0 0 1 0
RepB,B (trans) =
0 1 0 0
0 0 0 1 B,B
Three.III.1.21 (a) With respect to the basis of the codomain, the images of the members
of the basis of the domain are represented as
0 0 0 0
1 0 0 0
RepB (β ~ 2 ) = RepB (β
~ 3 ) = RepB (β
~ 4 ) = RepB (~0) =
0 1 0 0
0 0 1 0
and consequently, the matrix representing the transformation is this.
0 0 0 0
1 0 0 0
0 1 0 0
0 0 1 0
0 0 0 0
1 0 0 0
(b)
0 0 0 0
0 0 1 0
0 0 0 0
1 0 0 0
(c)
0 1 0 0
0 0 0 0
Three.III.1.22 (a) The picture of ds : R2 → R2 is this.
ds (~u)
~u ds
−→ ds (~v)
~v
This map’s effect on the vectors in the standard basis for the domain is
! ! ! !
1 ds s 0 ds 0
7−→ 7 →
−
0 0 1 s
and those images are represented with respect to the codomain’s basis (again, the
standard basis) by themselves.
! ! ! !
s s 0 0
RepE2 ( )= RepE2 ( )=
0 0 s s
Answers to Exercises 163
f`
7−→
Some calculation (see Exercise I.31) shows that when the line has slope k
! ! ! !
1 f` (1 − k2 )/(1 + k2 ) 0 f` 2k/(1 + k2 )
7−→ 7 →
−
0 2k/(1 + k2 ) 1 −(1 − k2 )/(1 + k2 )
(the case of a line with undefined slope is separate but easy) and so the matrix
representing reflection is this.
!
1 1 − k2 2k
RepE2 ,E2 (f` ) = ·
1 + k2 2k −(1 − k2 )
Three.III.1.23 Call the map t : R2 → R2 .
(a) To represent this map with respect to the standard bases, we must find, and
then represent, the images of the vectors ~e1 and ~e2 from the domain’s basis. The
image of ~e1 is given.
One way to find the image of ~e2 is by eye — we can see this.
! ! ! ! ! !
1 1 0 t 2 −1 3
− = 7−→ − =
1 0 1 0 0 0
A more systematic way to find the image of ~e2 is to use the given information
to represent the transformation, and then use that representation to determine the
image. Taking this for a basis,
! !
1 1
C=h , i
1 0
the given information says this.
!
2 −1
RepC,E2 (t)
0 0
As !
1
RepC (~e2 ) =
−1
C
we have that ! ! !
2 −1 1 3
RepE2 (t(~e2 )) = =
0 0 −1 0
C,E2 C E2
164 Linear Algebra, by Hefferon
and consequently we know that t(~e2 ) = 3 · ~e1 (since, with respect to the standard
basis, this vector is represented by itself). Therefore, this is the representation of t
with respect to E2 , E2 .
!
−1 3
RepE2 ,E2 (t) =
0 0
E2 ,E2
(b) To use the matrix developed in the prior item, note that
! !
0 0
RepE2 ( )=
5 5
E2
and so we have this is the representation, with respect to the codomain’s basis, of
the image of the given vector.
! ! ! !
0 −1 3 0 15
RepE2 (t( )) = =
5 0 0 5 0
E2 ,E2 E2 E2
Because the codomain’s basis is the standard one, and so vectors in the codomain
are represented by themselves, we have this.
! !
0 15
t( )=
5 0
(c) We first find the image of each member of B, and then represent those images
with respect to D. For the first step, we can use the matrix developed earlier.
! ! ! ! ! !
1 −1 3 1 −4 1 −4
RepE2 ( )= = so t( )=
−1 0 0 −1 0 −1 0
E2 ,E2 E2 E2
Actually, for the second member of B there is no need to apply the matrix because
the problem statement gives its image.
! !
1 2
t( )=
1 0
Now representing those images with respect to D is routine.
! ! ! !
−4 −1 2 1/2
RepD ( )= and RepD ( )=
0 2 0 −1
D D
Thus, the matrix is this.
!
−1 1/2
RepB,D (t) =
2 −1
B,D
(d) We know the images of the members of the domain’s basis from the prior item.
! ! ! !
1 −4 1 2
t( )= t( )=
−1 0 1 0
Answers to Exercises 165
We can compute the representation of those images with respect to the codomain’s
basis. ! ! ! !
−4 −2 2 1
RepB ( )= and RepB ( )=
0 −2 0 1
B B
Thus this is the matrix. !
−2 1
RepB,B (t) =
−2 1
B,B
Three.III.1.24 (a) The images of the members of the domain’s basis are
~ 1 7→ h(β
β ~ 1) β
~ 2 7→ h(β
~ 2) . . . β
~ n 7→ h(β
~ n)
and those images are represented with respect to the codomain’s basis in this way.
1 0
0 1
~ 1 ) ) = . Reph(B) ( h(β
Reph(B) ( h(β ~ 2) ) = .
. .
. .
0 0
0
0
... ~ n) ) = .
Reph(B) ( h(β .
.
1
Hence, the matrix is the identity.
1 0 ... 0
0 1 0
RepB,h(B) (h) = ..
.
0 0 1
(b) Using the matrix in the prior item, the representation is this.
c1
..
Reph(B) ( h(~v) ) = .
cn h(B)
Three.III.1.25 The product
0
h1,1 ... h1,i ... h1,n . h1,i
.
.
h ... h2,i ... h2,n h2,i
2,1
.. 1 = .
.
.
.. .
hm,1 ... hm,i ... h1,n . hm,i
0
gives the i-th column of the matrix.
166 Linear Algebra, by Hefferon
d/dx
Three.III.1.26 (a) The images of the basis vectors for the domain are cos x 7−→ − sin x
d/dx
and sin x 7−→ cos x. Representing those with respect to the codomain’s basis (again,
B) and adjoining the representations gives this matrix.
!
d 0 1
RepB,B ( ) =
dx −1 0
B,B
d/dx d/dx
(b) The images of the vectors in the domain’s basis are ex 7−→ ex and e2x 7−→ 2e2x .
Representing with respect to the codomain’s basis and adjoining gives this matrix.
!
d 1 0
RepB,B ( ) =
dx 0 2
B,B
d/dx d/dx
(c) The images of the members of the domain’s basis are 1 7−→ 0, x 7−→ 1,
d/dx d/dx
ex 7−→ ex , and xex 7−→ ex + xex . Representing these images with respect to B
and adjoining gives this matrix.
0 1 0 0
d 0 0 0 0
RepB,B ( ) =
dx 0 0 1 1
0 0 0 1 B,B
Three.III.1.27 (a) It is the set of vectors of the codomain represented with respect to
the codomain’s basis in this way.
! ! !
1 0 x x
{ | x, y ∈ R} = { | x, y ∈ R }
0 0 y 0
As the codomain’s basis is E2 , and so each vector is represented by itself, the range
of this transformation is the x-axis.
(b) It is the set of vectors of the codomain represented in this way.
! ! !
0 0 x 0
{ | x, y ∈ R} = { | x, y ∈ R }
3 2 y 3x + 2y
With respect to E2 vectors represent themselves so this range is the y axis.
(c) The set of vectors represented with respect to E2 as
! ! !
a b x ax + by
{ | x, y ∈ R } = { | x, y ∈ R}
2a 2b y 2ax + 2by
!
1
= { (ax + by) · | x, y ∈ R}
2
is the line y = 2x, provided either a or b is not zero, and is the set consisting of
just the origin if both are zero.
Answers to Exercises 167
and also represents a ĥ : P1 → R2 with respect to h1, xi and E2 that acts in this way.
! !
1 2
1 7→ and x → 7
3 4
The second reason is that, even if the domain and codomain of h and ĥ coincide,
different bases produce different maps. An example is the 2×2 identity matrix
!
1 0
I=
0 1
which represents the identity map on R2 with respect to E2 , E2 . However, with respect
to E2 for the domain but the basis D = h~e2 , ~e1 i for the codomain, the same matrix I
represents the map that swaps the first and second components
! !
x y
7→
y x
(that is, reflection about the line y = x).
Three.III.1.29 We mimic Example 1.1, just replacing the numbers with letters.
~ 1, . . . , β
Write B as hβ ~ n i and D as h~δ1 , . . . , ~δm i. By definition of representation of
a map with respect to bases, the assumption that
h1,1 . . . h1,n
.. ..
RepB,D (h) = . .
hm,1 . . . hm,n
means that h(β~ i ) = hi,1~δ1 + · · · + hi,n~δn . And, by the definition of the representation
of a vector with respect to a basis, the assumption that
c1
..
RepB (~v) = .
cn
168 Linear Algebra, by Hefferon
~ 1 + · · · + cn β
means that ~v = c1 β ~ n . Substituting gives
~ 1 + · · · + cn · β
h(~v) = h(c1 · β ~ n)
~ 1 ) + · · · + cn · β
= c1 · h(β ~n
= c1 · (h1,1~δ1 + · · · + hm,1~δm ) + · · · + cn · (h1,n~δ1 + · · · + hm,n~δm )
= (h1,1 c1 + · · · + h1,n cn ) · ~δ1 + · · · + (hm,1 c1 + · · · + hm,n cn ) · ~δm
and so h(~v) is represented as required.
Three.III.1.30 (a) The picture is this.
~ 1 + · · · + ck · β
so that ~v = c1 · β ~ k then
c1
.
..
c
k
RepBV (~v) =
0
..
.
0
~ 1 + · · · + ck · β
because ~v = c1 · β ~k + 0 · β
~ k+1 + · · · + 0 · β
~ n.
(b) We must first decide what the question means. Compare h : V → W with its
restriction to the subspace h U : U → W. The range space of the restriction is a
subspace of W, so fix a basis Dh(U) for this range space and extend it to a basis
DV for W. We want the relationship between these two.
So, for this first one, we are asking whether there are scalars such that
! ! ! !
1 1 3 1
c1 + c2 + c3 =
0 1 4 3
that is, whether the vector is in the column space of the matrix.
(a) Yes. We can get this conclusion by setting up the resulting linear system and
applying Gauss’s Method, as usual. Another way to get it is to note by inspection
of the equation of columns that taking c3 = 3/4, and c1 = −5/4, and c2 = 0 will
do. Still a third way to get this conclusion is to note that the rank of the matrix is
two, which equals the dimension of the codomain, and so the map is onto — the
range is all of R2 and in particular includes the given vector.
(b) No; note that all of the columns in the matrix have a second component that is
twice the first, while the vector does not. Alternatively, the column space of the
matrix is ! ! ! !
2 0 3 1
{c1 + c2 + c3 | c1 , c2 , c3 ∈ R} = {c | c ∈ R}
4 0 6 2
(which is the fact already noted, but we got it by calculation rather than inspiration),
and the given vector is not in this set.
Three.III.2.14 (a) The first member of the basis
! !
0 1
=
1 0
B
maps to !
1/2
−1/2
D
which is this member of the codomain.
! ! !
1 1 1 1 0
· − · =
2 1 2 −1 1
(b) The second member of the basis maps
! ! !
1 0 (1/2
= 7 →
0 1 1/2
B D
to this member of the codomain. ! ! !
1 1 1 1 1
· + · =
2 1 2 −1 0
(c) Because the map that the matrix represents is the identity map on the basis,
it must be the identity on all members of the domain. We can come to the same
conclusion in another way by considering
! !
x y
=
y x
B
172 Linear Algebra, by Hefferon
which maps to !
(x + y)/2
(x − y)/2
D
which represents this member of R2 .
! ! !
x+y 1 x−y 1 x
· + · =
2 1 2 −1 y
Three.III.2.15 A general member of the domain, represented with respect to the domain’s
basis as !
a
a cos θ + b sin θ =
a+b
B
maps to !
0
representing 0 · (cos θ + sin θ) + a · (cos θ)
a
D
and so the linear map represented by the matrix with respect to these bases
a cos θ + b sin θ 7→ a cos θ
is projection onto the first component.
Three.III.2.16 Denote the given basis of P2 by B. Application of the linear map is
represented by matrix-vector multiplication. Thus the first vector in E3 maps to the
element of P2 represented with respect to B by
1 3 0 1 1
0 1 0 0 = 0
1 0 1 0 1
and that element is 1 + x. Calculate the other two images of basis vectors in the same
way.
1 3 0 0 3 1 3 0 0 0
0 1 0 1 = 1 = RepB (4 + x2 ) 0 1 0 0 = 0 = RepB (x)
1 0 1 0 0 1 0 1 1 1
So the range of h is the span of three polynomials 1 + x, 4 + x2 , and x. We can thus
decide if 1 + 2x is in the range of the map by looking for scalars c1 , c2 , and c3 such
that
c1 · (1 + x) + c2 · (4 + x2 ) + c3 · (x) = 1 + 2x
and obviously c1 = 1, c2 = 0, and c3 = 1 suffice. Thus 1 + 2x is in the range, since it
is the image of this vector.
1 0 0
1 · 0 + 0 · 1 + 1 · 0
0 0 1
Answers to Exercises 173
Matrix Operations
regroups
= ((g1,1 + h1,1 )v1 + · · · + (g1,1 + h1,n )vn ) · ~δ1
+ · · · + ((gm,1 + hm,1 )v1 + · · · + (gm,n + hm,n )vn ) · ~δm
to the entry-by-entry sum of the representation of g(~v) and the representation of
h(~v).
(b) The representation of (r · h) (~v) = r · h(~v)
r· (h1,1 v1 + h1,2 v2 + · · · + h1,n vn )~δ1
+ · · · + (hm,1 v1 + hm,2 v2 + · · · + hm,n vn )~δm
(b) As with the prior answer, except that here we apply that vector space addition
is associative.
(c) As before, except that here we note that g(~v) + z(~v) = g(~v) + ~0 = g(~v).
(d) Apply that 0 · g(~v) = ~0 = z(~v).
(e) Apply that (r + s) · g(~v) = r · g(~v) + s · g(~v).
(f) Apply the prior two items with r = 1 and s = −1.
(g) Apply that r · (g(~v) + h(~v)) = r · g(~v) + r · h(~v).
(h) Apply that (rs) · g(~v) = r · (s · g(~v)).
Three.IV.1.11 For any V, W with bases B, D, the (appropriately-sized) zero matrix
represents this map.
~ 1 7→ 0 · ~δ1 + · · · + 0 · ~δm · · · β
β ~n →
7 0 · ~δ1 + · · · + 0 · ~δm
This is the zero map.
There are no other matrices that represent only one map. For, suppose that H is
not the zero matrix. Then it has a nonzero entry; assume that hi,j 6= 0. With respect
to bases B, D, it represents h1 : V → W sending
~ j 7→ h1,j~δ1 + · · · + hi,j~δi + · · · + hm,j~δm
β
and with respect to B, 2 · D it also represents h2 : V → W sending
~ j 7→ h1,j · (2~δ1 ) + · · · + hi,j · (2~δi ) + · · · + hm,j · (2~δm )
β
(the notation 2 · D means to double all of the members of D). These maps are easily
seen to be unequal.
Three.IV.1.12 Fix bases B and D for V and W, and consider RepB,D : L(V, W) → Mm×n
associating each linear map with the matrix representing that map h 7→ RepB,D (h).
From the prior section we know that (under fixed bases) the matrices correspond
to linear maps, so the representation map is one-to-one and onto. That it preserves
linear operations is Theorem 1.4.
Three.IV.1.13 Fix bases and represent the transformations with 2×2 matrices. The
space of matrices M2×2 has dimension four, and hence the above six-element set is
linearly dependent. By the prior exercise that extends to a dependence of maps. (The
misleading part is only that there are six transformations, not five, so that we have
more than we need to give the existence of the dependence.)
Three.IV.1.14 That the trace of a sum is the sum of the traces holds because both
trace(H + G) and trace(H) + trace(G) are the sum of h1,1 + g1,1 with h2,2 + g2,2 , etc.
For scalar multiplication we have trace(r · H) = r · trace(H); the proof is easy. Thus
the trace map is a homomorphism from Mn×n to R.
Three.IV.1.15 (a) The i, j entry of (G + H)T is gj,i + hj,i . That is also the i, j entry of
GT + HT .
(b) The i, j entry of (r · H)T is rhj,i , which is also the i, j entry of r · HT .
Answers to Exercises 179
Three.IV.1.16 (a) For H+HT , the i, j entry is hi,j +hj,i and the j, i entry of is hj,i +hi,j .
The two are equal and thus H + HT is symmetric.
Every symmetric matrix does have that form, since we can write H = (1/2) ·
(H + HT ).
(b) The set of symmetric matrices is nonempty as it contains the zero matrix.
Clearly a scalar multiple of a symmetric matrix is symmetric. A sum H + G of two
symmetric matrices is symmetric because hi,j + gi,j = hj,i + gj,i (since hi,j = hj,i
and gi,j = gj,i ). Thus the subset is nonempty and closed under the inherited
operations, and so it is a subspace.
Three.IV.1.17 (a) Scalar multiplication leaves the rank of a matrix unchanged except
that multiplication by zero leaves the matrix with rank zero. (This follows from
the first theorem of the book, that multiplying a row by a nonzero scalar doesn’t
change the solution set of the associated linear system.)
(b) A sum of rank n matrices can have rank less than n. For instance, for any matrix
H, the sum H + (−1) · H has rank zero.
A sum of rank n matrices can have rank greater than n. Here are rank one
matrices that sum to a rank two!matrix. ! !
1 0 0 0 1 0
+ =
0 0 0 1 0 1
which, after expanding and regrouping about the y’s yields this.
(h1,1 g1,1 + h1,2 g2,1 + h1,3 g3,1 )y1 + (h1,1 g1,2 + h1,2 g2,2 + h1,3 g3,2 )y2 = d1
(h2,1 g1,1 + h2,2 g2,1 + h2,3 g3,1 )y1 + (h2,1 g1,2 + h2,2 g2,2 + h2,3 g3,2 )y2 = d2
We can express the starting system and the system used for the substitutions in matrix
language, as
! x x1 !
1
h1,1 h1,2 h1,3 d1
x2 = H x2 =
h2,1 h2,2 h2,3 d2
x3 x3
and
g1,1 g1,2 ! ! x1
y1 y1
g2,1 g2,2 =G = x2
y2 y2
g3,1 g3,2 x3
and with this, the substitution is ~d = H~x = H(G~y) = (HG)~y.
Three.IV.2.19 Technically, no. The dot product operation yields a scalar while the
matrix product yields a 1×1 matrix. However, we usually will ignore the distinction.
Three.IV.2.20 The action of d/dx on B is 1 7→ 0, x 7→ 1, x2 7→ 2x, . . . and so this is its
(n + 1)×(n + 1) matrix representation.
0 1 0 0
0 0 2 0
d .
RepB,B ( ) = . .
dx
0 0 0 n
0 0 0 0
The product of this matrix with itself is defined because the matrix is square.
2 0 0 2 0 0
0 1 0 0 0 0 0 6
0 0 2 0
0
..
.. .
. =
0 0 0 n(n − 1)
0 0 0 n
0 0 0 0
0 0 0 0
0 0 0 0
The map so represented is the composition
d
d p dx d
d2 p
p 7−dx
→ 7−→
dx dx2
which is the second derivative operation.
Three.IV.2.21 (a) iii
(b) iv
(c) None
(d) None (or (i) if we allow multiplication from the left)
Answers to Exercises 181
Rotating rx first and then ry is different than rotating ry first and then rx . In
particular, rx (~e3 ) = −~e2 so ry ◦ rx (~e3 ) = −~e2 , while ry (~e3 ) = ~e1 so rx ◦ ry (~e3 ) = ~e1 ,
and hence the maps do not commute.
Three.IV.2.28 It doesn’t matter (as long as the spaces have the appropriate dimensions).
For associativity, suppose that F is m×r, that G is r×n, and that H is n×k. We
can take any r dimensional space, any m dimensional space, any n dimensional space,
and any k dimensional space — for instance, Rr , Rm , Rn , and Rk will do. We can take
any bases A, B, C, and D, for those spaces. Then, with respect to C, D the matrix H
represents a linear map h, with respect to B, C the matrix G represents a g, and with
respect to A, B the matrix F represents an f. We can use those maps in the proof.
The second half is similar, except that we add G and H and so we must take them
to represent maps with the same domain and codomain.
Three.IV.2.29 (a) The product of rank n matrices can have rank less than or equal to
n but not greater than n.
To see that the rank can fall, consider the maps πx , πy : R2 → R2 projecting
onto the axes. Each is rank one but their composition πx ◦ πy , which is the zero
map, is rank zero. That translates over to matrices representing those maps in this
way. ! ! !
1 0 0 0 0 0
RepE2 ,E2 (πx ) · RepE2 ,E2 (πy ) = =
0 0 0 1 0 0
To prove that the product of rank n matrices cannot have rank greater than n,
we can apply the map result that the image of a linearly dependent set is linearly
dependent. That is, if h : V → W and g : W → X both have rank n then a set in
the range R(g ◦ h) of size larger than n is the image under g of a set in W of size
larger than n and so is linearly dependent (since the rank of h is n). Now, the
image of a linearly dependent set is dependent, so any set of size larger than n in
the range is dependent. (By the way, observe that the rank of g was not mentioned.
See the next part.)
(b) Fix spaces and bases and consider the associated linear maps f and g. Recall
that the dimension of the image of a map (the map’s rank) is less than or equal to
the dimension of the domain, and consider the arrow diagram.
f g
V 7−→ R(f) 7−→ R(g ◦ f)
First, the image of R(f) must have dimension less than or equal to the dimension of
R(f), by the prior sentence. On the other hand, R(f) is a subset of the domain of
g, and thus its image has dimension less than or equal the dimension of the domain
of g. Combining those two, the rank of a composition is less than or equal to the
minimum of the two ranks.
The matrix fact follows immediately.
Answers to Exercises 183
while
d/dx s
a0 + a1 x + · · · + an xn 7−→ a1 + · · · + nan xn−1 7−→ a1 x + · · · + an xn
Three.IV.2.37 (a) Tracing through the remark at the end of the subsection gives that
the i, j entry of (FG)H is this
X
s X
r
X
s X
r X
s X
r
fi,k gk,t ht,j = (fi,k gk,t )ht,j = fi,k (gk,t ht,j )
t=1 k=1 t=1 k=1 t=1 k=1
X
r X
s X
r X
s
= fi,k (gk,t ht,j ) = fi,k gk,t ht,j
k=1 t=1 k=1 t=1
(the first equality comes from using the distributive law to multiply through the
h’s, the second equality is the associative law for real numbers, the third is the
commutative law for reals, and the fourth equality follows on using the distributive
law to factor the f’s out), which is the i, j entry of F(GH).
(b) The k-th component of h(~v) is
X
n
hk,j vj
j=1
X
r X
n
Xr X
n X
r X
n
gi,k hk,j vj = gi,k hk,j vj = (gi,k hk,j )vj
k=1 j=1 k=1 j=1 k=1 j=1
X
n X
r X
n X r
= (gi,k hk,j )vj = ( gi,k hk,j ) vj
j=1 k=1 j=1 k=1
(the first equality holds by using the distributive law to multiply the g’s through,
the second equality represents the use of associativity of reals, the third follows by
commutativity of reals, and the fourth comes from using the distributive law to
factor the v’s out).
186 Linear Algebra, by Hefferon
Three.IV.3.28 The product is the identity matrix (recall that cos2 θ + sin2 θ = 1). An
explanation is that the given matrix represents, with respect to the standard bases, a
rotation in R2 of θ radians while the transpose represents a rotation of −θ radians.
The two cancel.
Three.IV.3.29 (a) The adjacency matrix is this (e.g, the first row shows that there is
only one connection including Burlington, the road to Winooski).
0 0 0 0 1
0 0 1 1 1
0 1 0 1 0
0 1 1 0 0
1 1 0 0 0
(b) Because these are two-way roads, any road connecting city i to city j gives a
connection between city j and city i.
(c) The square of the adjacency matrix tells how cities are connected by trips
involving two roads.
Three.IV.3.30 The set of diagonal matrices is nonempty as the zero matrix is diagonal.
Clearly it is closed under scalar multiples and sums. Therefore it is a subspace. The
dimension is n; here is a basis.
1 0 ... 0 0 ...
0 0 0 0
{ }
,...,
.. ..
. .
0 0 0 0 0 1
Three.IV.3.31 No. In P1 , with respect to the unequal bases B = h1, xi and D =
h1 + x, 1 − xi, the identity transformation is represented by this matrix.
!
1/2 1/2
RepB,D (id) =
1/2 −1/2
B,D
Three.IV.3.32 For any scalar r and square matrix H we have (rI)H = r(IH) = rH =
r(HI) = (Hr)I = H(rI).
There are no other such matrices; here is an argument for 2×2 matrices that is
easily extended to n×n. If a matrix commutes with all others then it commutes with
this unit matrix. ! ! ! ! ! !
0 a a b 0 1 0 1 a b c d
= = =
0 c c d 0 0 0 0 c d 0 0
From this we first conclude that the upper left entry a must equal its lower right
entry d. We also conclude that the lower left entry c is zero. The argument for the
upper right entry b is similar.
188 Linear Algebra, by Hefferon
Three.IV.3.39 Perhaps the easiest way is to show that each n×m matrix is a linear
combination of unit matrices in one and only one way:
1 0 ... 0 0 ... a1,1 a1,2 . . .
. ..
c1 0 0 + · · · + cn,m .. = .
..
. 0 ... 1 an,1 . . . an,m
has the unique solution c1 = a1,1 , c2 = a1,2 , etc.
Three.IV.3.40 Call that matrix F. We have
! ! !
2 2 1 3 2 5 3
F = F3 = F = 4
1 1 2 1 3 2
In general, !
fn+1 fn
Fn =
fn fn−1
where fi is the i-th Fibonacci number fi = fi−1 + fi−2 and f0 = 0, f1 = 1, which we
verify by induction, based on this equation.
! ! !
fi−1 fi−2 1 1 fi fi−1
=
fi−2 fi−3 1 0 fi−1 fi−2
Three.IV.3.41 Chapter Five gives a less computational reason — the trace of a
matrix is the second coefficient in its characteristic polynomial — but for now we
can use indices. We have
Tr(GH) = (g1,1 h1,1 + g1,2 h2,1 + · · · + g1,n hn,1 )
+ (g2,1 h1,2 + g2,2 h2,2 + · · · + g2,n hn,2 )
+ · · · + (gn,1 h1,n + gn,2 h2,n + · · · + gn,n hn,n )
while
Tr(HG) = (h1,1 g1,1 + h1,2 g2,1 + · · · + h1,n gn,1 )
+ (h2,1 g1,2 + h2,2 g2,2 + · · · + h2,n gn,2 )
+ · · · + (hn,1 g1,n + hn,2 g2,n + · · · + hn,n gn,n )
and the two are equal.
Three.IV.3.42 A matrix is upper triangular if and only if its i, j entry is zero whenever
i > j. Thus, if G, H are upper triangular then hi,j and gi,j are zero when i > j. An
entry in the product pi,j = gi,1 h1,j + · · · + gi,n hn,j is zero unless at least some of the
terms are nonzero, that is, unless for at least some of the summands gi,r hr,j both
i 6 r and r 6 j. Of course, if i > j this cannot happen and so the product of two
upper triangular matrices is upper triangular. (A similar argument works for lower
triangular matrices.)
190 Linear Algebra, by Hefferon
Three.IV.3.43 The sum along the i-th row of the product is this.
pi,1 + · · · + pi,n = (hi,1 g1,1 + hi,2 g2,1 + · · · + hi,n gn,1 )
+ (hi,1 g1,2 + hi,2 g2,2 + · · · + hi,n gn,2 )
+ · · · + (hi,1 g1,n + hi,2 g2,n + · · · + hi,n gn,n )
= hi,1 (g1,1 + g1,2 + · · · + g1,n )
+ hi,2 (g2,1 + g2,2 + · · · + g2,n )
+ · · · + hi,n (gn,1 + gn,2 + · · · + gn,n )
= hi,1 · 1 + · · · + hi,n · 1
=1
Three.IV.3.44 Matrices representing (say, with respect to E2 , E2 ⊂ R2 ) the maps that
send
~ 1 7−h→ β
β ~1 β ~ 2 7−h→ ~0
and
~ 1 7−g→ β
β ~2 β ~ 2 7−g→ ~0
will do.
Three.IV.3.45 The combination is to have all entries of the matrix be zero except for
one (possibly) nonzero entry in each row and column. We can write such a matrix as
the product of a permutation
matrix
and a diagonal
matrix, e.g.,
0 4 0 0 1 0 4 0 0
2 0 0 = 1 0 0 0 2 0
0 0 −5 0 0 1 0 0 −5
and its action is thus to rescale the rows and permute them.
Three.IV.3.46 (a) Each entry pi,j = gi,1 h1,j + · · · + g1,r hr,1 takes r multiplications
and there are m · n entries. Thus there are m · n · r multiplications.
(b) Let H1 be 5×10, let H2 be 10×20, let H3 be 20×5, let H4 be 5×1. Then, using
the formula from the prior part,
this association uses this many multiplications
((H1 H2 )H3 )H4 1000 + 500 + 25 = 1525
(H1 (H2 H3 ))H4 1000 + 250 + 25 = 1275
(H1 H2 )(H3 H4 ) 1000 + 100 + 100 = 1200
H1 (H2 (H3 H4 )) 100 + 200 + 50 = 350
H1 ((H2 H3 )H4 ) 1000 + 50 + 50 = 1100
shows which is cheapest.
(c) This is an improvement by S. Winograd of a formula due to V. Strassen: let
w = aA − (a − c − d)(A − C + D) ! and then ! !
a b A B α β
=
c d C D γ δ
Answers to Exercises 191
and, since bk+1 A~xk+1 + · · · + br1 A~xr1 ∈ N (A) we get a contradiction unless it is ~0
(clearly it is in R(A), but A~x1 , . . . , A~xk is a basis for R(A) ∩ N (A)).
Hence dim(R(A2 )) = r1 − k = dim(R(A)) − dim(R(A) ∩ N (A)).
Three.IV.4: Inverses
Another point is that just because H and G each has an inverse doesn’t mean H + G
has an inverse; here is an example. ! !
1 0 −1 0
0 1 0 −1
Still a third point is that, even if the two matrices have inverses, and the sum has
an inverse, doesn’t imply that the equation holds:
!−1 !−1 !−1 !−1
2 0 1/2 0 3 0 1/3 0
= =
0 2 0 1/2 0 3 0 1/3
but !−1 !−1
5 0 1/5 0
=
0 5 0 1/5
and (1/2)+ (1/3) does not equal 1/5.
Three.IV.4.18 Yes: T k (T −1 )k = (T T · · · T ) · (T −1 T −1 · · · T −1 ) = T k−1 (T T −1 )(T −1 )k−1 =
· · · = I.
Three.IV.4.19 Yes, the inverse of H−1 is H.
Three.IV.4.20 One way to check that the first is true is with the angle sum formulas
from trigonometry.
!
cos(θ1 + θ2 ) − sin(θ1 + θ2 )
sin(θ1 + θ2 ) cos(θ1 + θ2 )
!
cos θ1 cos θ2 − sin θ1 sin θ2 − sin θ1 cos θ2 − cos θ1 sin θ2
=
sin θ1 cos θ2 + cos θ1 sin θ2 cos θ1 cos θ2 − sin θ1 sin θ2
! !
cos θ1 − sin θ1 cos θ2 − sin θ2
=
sin θ1 cos θ1 sin θ2 cos θ2
Checking the second equation in this way is similar.
Of course, the equations can be not just checked but also understood by recalling
that tθ is the map that rotates vectors about the origin through an angle of θ radians.
Three.IV.4.21 There are two cases. For the first case
! we assume that a is nonzero. Then
!
−(c/a)ρ1 +ρ2 a b 1 0 a b 1 0
−→ =
0 −(bc/a) + d −c/a 1 0 (ad − bc)/a −c/a 1
shows that the matrix is invertible (in this a 6= 0 case) if and only if ad − bc 6= 0. To
find the inverse, we finish with the Jordan half of the reduction. !
(1/a)ρ1 1 b/a 1/a 0
−→
(a/ad−bc)ρ2 0 1 −c/(ad − bc) a/(ad − bc)
!
−(b/a)ρ2 +ρ1 1 0 d/(ad − bc) −b/(ad − bc)
−→
0 1 −c/(ad − bc) a/(ad − bc)
196 Linear Algebra, by Hefferon
The other case is the a = 0 case. We swap to get c into the 1, 1 position.
!
ρ1 ↔ρ2 c d 0 1
−→
0 b 1 0
This matrix is nonsingular if and only if both b and c are nonzero (which, under the
case assumption that a = 0, holds if and only if ad − bc 6= 0). To find the inverse we
do the Jordan half.
! !
(1/c)ρ1 1 d/c 0 1/c −(d/c)ρ2 +ρ1 1 0 −d/bc 1/c
−→ −→
(1/b)ρ2 0 1 1/b 0 0 1 1/b 0
Three.IV.4.22 With H a 2×3 matrix, in looking for a matrix G such that the combination
HG acts as the 2×2 identity we need G to be 3×2. Setting up the equation
! m n !
1 0 1 1 0
p q =
0 1 0 0 1
r s
m +r =1
n +s = 0
p =0
q =1
gives rise to a linear system with nine equations and four unknowns.
a =1
b =0
a =0
c =0
d =1
c =0
e =0
f=0
e =1
This system is inconsistent (the first equation conflicts with the third, as do the
seventh and ninth) and so there is no left inverse.
Three.IV.4.23 With respect to the standard bases we have
1 0
RepE2 ,E3 (ι) = 0 1
0 0
and setting up the equation to find the matrix inverse
! 1 0 !
a b c 1 0
0 1 = = RepE2 ,E2 (id)
d e f 0 1
0 0
gives rise to a linear system.
a =1
b =0
d =0
e =1
There are infinitely many solutions in a, . . . , f to this system because two of these
variables are entirely unrestricted
a 1 0 0
b 0 0 0
c 0 1 0
{ = + c · + f · | c, f ∈ R }
d 0 0 0
e 1 0 0
f 0 0 1
and so there are infinitely many solutions to the matrix equation.
!
1 0 c
{ | c, f ∈ R }
0 1 f
198 Linear Algebra, by Hefferon
With the bases still fixed at E2 , E2 , for instance taking c = 2 and f = 3 gives a matrix
representing this map.
x !
f2,3 x + 2z
y 7−→
y + 3z
z
The check that f2,3 ◦ ι is the identity map on R2 is easy.
Three.IV.4.24 By Lemma 4.2 it cannot have infinitely many left inverses, because a
matrix with both left and right inverses has only one of each (and that one of each is
one of both — the left and right inverse matrices are equal).
Three.IV.4.25 (a) True, It must be linear, as the proof from Theorem II.2.20 shows.
(b) False. It may be linear, but it need not be. Consider the projection map
π : R3 → R2 described at the start of this subsection. Define η : R2 → R3 in this
way.
! x
x
7→ y
y
1
It is a right inverse of π because π ◦ η does this.
! x !
x x
7→ y 7→
y y
1
It is not linear because it does not map the zero vector to the zero vector.
Three.IV.4.26 The associativity of matrix multiplication gives H−1 (HG) = H−1 Z = Z
and also H−1 (HG) = (H−1 H)G = IG = G.
Three.IV.4.27 Multiply both sides of the first equation by H.
Three.IV.4.28 Checking that when I − T is multiplied on both sides by that expression
(assuming that T 4 is the zero matrix) then the result is the identity matrix is easy.
The obvious generalization is that if T n is the zero matrix then (I − T )−1 = I + T +
T 2 + · · · + T n−1 ; the check again is easy.
Three.IV.4.29 The powers of the matrix are formed by taking the powers of the diagonal
entries. That is, D2 is all zeros except for diagonal entries of d1,1 2 , d2,2 2 , etc. This
suggests defining D0 to be the identity matrix.
Three.IV.4.30 Assume that B is row equivalent to A and that A is invertible. Because
they are row-equivalent, there is a sequence of row steps to reduce one to the other.
We can do that reduction with matrices, for instance, A can change by row operations
to B as B = Rn · · · R1 A. This equation gives B as a product of invertible matrices and
by Lemma 4.4 then, B is also invertible.
Three.IV.4.31 (a) See the answer to Exercise 29.
Answers to Exercises 199
(b) We will show that both conditions are equivalent to the condition that the two
matrices be nonsingular.
As T and S are square and their product is defined, they are equal-sized, say
n×n. Consider the T S = I half. By the prior item the rank of I is less than or
equal to the minimum of the rank of T and the rank of S. But the rank of I is n, so
the rank of T and the rank of S must each be n. Hence each is nonsingular.
The same argument shows that ST = I implies that each is nonsingular.
Three.IV.4.32 Inverses are unique, so we need only show that it works. The check
appears above as Exercise 34.
Three.IV.4.33 (a) See the answer for Exercise 26.
(b) See the answer for Exercise 26.
T T
(c) Apply the first part to I = AA−1 to get I = IT = (AA−1 ) = (A−1 ) AT .
(d) Apply the prior item with AT = A, as A is symmetric.
Three.IV.4.34 For the answer to the items making up the first half, see Exercise 31. For
the proof in the second half, assume that A is a zero divisor so there is a nonzero
matrix B with AB = Z (or else BA = Z; this case is similar), If A is invertible then
A−1 (AB) = (A−1 A)B = IB = B but also A−1 (AB) = A−1 Z = Z, contradicting that
B is nonzero.
Three.IV.4.35 Here are four solutions to H2 = I.!
±1 0
0 ±1
Three.IV.4.36 It is not reflexive since, for instance,
!
1 0
H=
0 2
is not a two-sided inverse of itself. The same example shows that it is not transitive.
That matrix has this two-sided inverse !
1 0
G=
0 1/2
and while H is a two-sided inverse of G and G is a two-sided inverse of H, we know
that H is not a two-sided inverse of H. However, the relation is symmetric: if G is a
two-sided inverse of H then GH = I = HG and therefore H is also a two-sided inverse
of G.
Three.IV.4.37 This is how the answer was given in the cited source. Let A be m×m,
non-singular, with the stated property. Let B be its inverse. Then for n 6 m,
Xm X
m X m X
m X m X
m
1= δnr = bns asr = bns asr = k bns
r=1 r=1 s=1 s=1 r=1 s=1
(A is singular if k = 0).
200 Linear Algebra, by Hefferon
Change of Basis
(a) This matrix is nonsingular and so changes bases. Finding to what basis E2 is
changed means finding D such that
!
5 0
RepE2 ,D (id) =
0 4
and by the definition of how a matrix represents a linear map, we have this.
! !
5 0
RepD (id(~e1 )) = RepD (~e1 ) = RepD (id(~e2 )) = RepD (~e2 ) =
0 4
Where ! !
x1 x2
D=h , i
y1 y2
we can either solve the system
! ! ! ! ! !
1 x1 x2 0 x1 x2
=5 +0 =0 +4
0 y1 y1 1 y1 y1
or else just spot the answer (thinking of the proof of Lemma 1.5).
! !
1/5 0
D=h , i
0 1/4
(b) Yes, this matrix is nonsingular and so changes bases. To calculate D, we proceed
as above with ! !
x1 x2
D=h , i
y1 y2
to solve
! ! ! ! ! !
1 x1 x2 0 x1 x2
=2 +3 and =1 +1
0 y1 y1 1 y1 y1
and get this. ! !
−1 1
D=h , i
3 −2
(c) No, this matrix does not change bases because it is singular.
(d) Yes, this matrix changes bases because it is nonsingular. The calculation of the
changed-to basis is as above.
! !
1/2 1/2
D=h , i
−1/2 1/2
Three.V.1.12 This question has many different solutions. One way to proceed is to
make up any basis B for any space, and then compute the appropriate D (necessarily
for the same space, of course). Another, easier, way to proceed is to fix the codomain
202 Linear Algebra, by Hefferon
as R3 and the codomain basis as E3 . This way (recall that the representation of any
vector with respect to the standard
basisis just
the vector itself), we have this.
3 1 4
B = h2 , −1 , 1i D = E3
0 0 4
Three.V.1.13 Checking that B = h2 sin(x) + cos(x), 3 cos(x)i is a basis is routine. Call
the natural basis D. To compute the change of basis matrix RepB,D (id) we must find
RepD (2 sin(x) + cos(x)) and RepD (3 cos(x)), that is, we need x1 , y1 , x2 , y2 such that
these equations hold.
x1 · sin(x) + y1 · cos(x) = 2 sin(x) + cos(x)
x2 · sin(x) + y2 · cos(x) = 3 cos(x)
Obviously this is the answer. !
2 0
RepB,D (id) =
1 3
For the change of basis matrix in the other direction we could look for RepB (sin(x))
and RepB (cos(x)) by solving these.
w1 · (2 sin(x) + cos(x)) + z1 · (3 cos(x)) = sin(x)
w2 · (2 sin(x) + cos(x)) + z2 · (3 cos(x)) = cos(x)
An easier method is to find the inverse
!−1 of the matrix!found above. !
2 0 1 3 0 1/2 0
RepD,B (id) = = · =
1 3 6 −1 2 −1/6 1/3
Three.V.1.14 We start by taking the inverse of the matrix, that is, by deciding what is
the inverse to the map of interest. !
−1 1 − cos(2θ) − sin(2θ)
RepD,E2 (id)RepD,E2 (id) = ·
− cos2 (2θ) − sin2 (2θ) − sin(2θ) cos(2θ)
!
cos(2θ) sin(2θ)
=
sin(2θ) − cos(2θ)
This is more tractable than the representation the other way because this matrix is
the concatenation of these two column ! vectors !
cos(2θ) sin(2θ)
RepE2 (~δ1 ) = RepE2 (~δ2 ) =
sin(2θ) − cos(2θ)
and representations with respect to E!2 are transparent. !
~δ1 = cos(2θ) ~δ2 = sin(2θ)
sin(2θ) − cos(2θ)
This pictures the action of the map that transforms D to E2 (it is, again, the inverse
of the map that is the answer to this question). The line lies at an angle θ to the
x axis.
Answers to Exercises 203
~δ1 = cos(2θ) ~e2
sin(2θ)
f
7−→ ~e1
~δ2 = sin(2θ)
− cos(2θ)
This map reflects vectors over that line. Since reflections are self-inverse, the answer
to the question is: the original map reflects about the line through the origin with
angle of elevation θ. (Of course, it does this to any basis.)
Three.V.1.15 The appropriately-sized identity matrix.
Three.V.1.16 Each is true if and only if the matrix is nonsingular.
Three.V.1.17 What remains is to show that left multiplication by a reduction matrix
represents a change from another basis to B = hβ ~ 1, . . . , β
~ n i.
Application of a row-multiplication matrix Mi (k) translates a representation with
~ 1 , . . . , kβ
respect to the basis hβ ~ i, . . . , β
~ n i to one with respect to B, as here.
~ 1 + · · · + ci · (kβ
~v = c1 · β ~ i ) + · · · + cn · β
~ n 7→ c1 · β
~ 1 + · · · + (kci ) · β
~ i + · · · + cn · β
~ n = ~v
Apply a row-swap matrix Pi,j to translates a representation with respect to the
~ 1, . . . , β
basis hβ ~ j, . . . , β
~ i, . . . , β
~ n i to one with respect to hβ ~ 1, . . . , β
~ i, . . . , β
~ j, . . . , β
~ n i.
Finally, applying a row-combination matrix Ci,j (k) changes a representation with
respect to hβ ~ 1, . . . , β~ i + kβ ~ j, . . . , β
~ j, . . . , β
~ n i to one with respect to B.
~ 1 + · · · + ci · (β
~v = c1 · β ~ i + kβ ~ j ) + · · · + cj β
~ j + · · · + cn · β
~n
7→ c1 · β ~ 1 + · · · + ci · β~ i + · · · + (kci + cj ) · β~ j + · · · + cn · β
~ n = ~v
(As in the part of the proof in the body of this subsection, the various conditions on
the row operations, e.g., that the scalar k is nonzero, assure that these are all bases.)
Three.V.1.18 Taking H as a change of basis matrix H = RepB,En (id), its columns are
h1,i
.. ~ i )) = RepE (β
~ i)
. = RepEn (id(β n
hn,i
and, because representations with respect to the standard basis are transparent, we
have this.
h1,i
.. ~
. = βi
hn,i
That is, the basis is the one composed of the columns of H.
204 Linear Algebra, by Hefferon
Three.V.1.19 (a) We can change the starting vector representation to the ending one
through a sequence of row operations. The proof tells us what how the bases change.
We start by swapping the first and second rows of the representation with respect
to B to get a representation with respect to a new basis B1 .
1
0
RepB1 (1 − x + 3x2 − x3 ) = B1 = h1 − x, 1 + x, x2 + x3 , x2 − x3 i
1
2 B
1
We next add −2 times the third row of the vector representation to the fourth row.
1
0
RepB3 (1 − x + 3x2 − x3 ) = B2 = h1 − x, 1 + x, 3x2 − x3 , x2 − x3 i
1
0 B
2
(The third element of B2 is the third element of B1 minus −2 times the fourth
element of B1 .) Now we can finish by doubling the third row.
1
0
RepD (1 − x + 3x2 − x3 ) = D = h1 − x, 1 + x, (3x2 − x3 )/2, x2 − x3 i
2
0 D
(b) Here are three different approaches to stating such a result. The first is the
assertion: where V is a vector space with basis B and ~v ∈ V is nonzero, for any
nonzero column vector ~z (whose number of components equals the dimension of V)
there is a change of basis matrix M such that M · RepB (~v) = ~z. The second possible
statement: for any (n-dimensional) vector space V and any nonzero vector ~v ∈ V,
where ~z1 , ~z2 ∈ Rn are nonzero, there are bases B, D ⊂ V such that RepB (~v) = ~z1
and RepD (~v) = ~z2 . The third is: for any nonzero ~v member of any vector space (of
dimension n) and any nonzero column vector (with n components) there is a basis
such that ~v is represented with respect to that basis by that column vector.
The first and second statements follow easily from the third. The first follows
because the third statement gives a basis D such that RepD (~v) = ~z and then
RepB,D (id) is the desired M. The second follows from the third because it is just a
doubled application of it.
A way to prove the third is as in the answer to the first part of this question.
Here is a sketch. Represent ~v with respect to any basis B with a column vector ~z1 .
This column vector must have a nonzero component because ~v is a nonzero vector.
Use that component in a sequence of row operations to convert ~z1 to ~z. (We could
fill out this sketch as an induction argument on the dimension of V.)
Three.V.1.20 This is the topic of the next subsection.
Answers to Exercises 205
Three.V.1.21 A change of basis matrix is nonsingular and thus has rank equal to the
number of its columns. Therefore its set of columns is a linearly independent subset
of size n in Rn and it is thus a basis. The answer to the second half is also ‘yes’; all
implications in the prior sentence reverse (that is, all of the ‘if . . . then . . . ’ parts of
the prior sentence convert to ‘if and only if’ parts).
Three.V.1.22 In response to the first half of the question, there are infinitely many such
matrices. One of them represents with respect to E2 the transformation of R2 with
this action.
! ! ! !
1 4 0 0
7→ 7→
0 0 1 −1/3
The problem of specifying two distinct input/output pairs is a bit trickier. The fact
that matrices have a linear action precludes some possibilities.
(a) Yes, there is such a matrix. These conditions
! ! ! ! ! !
a b 1 1 a b 2 −1
= =
c d 3 1 c d −1 −1
can be solved
a + 3b = 1
c + 3d = 1
2a − b = −1
2c − d = −1
to give this matrix.
!
−2/7 3/7
−2/7 3/7
t
R2wrt B̂ −−−−→ R2wrt D̂
T̂
Three.V.2.23 They are closed under nonzero scalar multiplication since a nonzero scalar
multiple of a matrix has the same rank as does the matrix. They are not closed under
addition, for instance, H + (−H) has rank zero.
Three.V.2.24 (a) We have !
1 −1
RepB,E2 (id) =
2 −1
and !−1 !
−1 1 −1 −1 1
RepE2 ,B (id) = RepB,E2 (id) = =
2 −1 −2 1
and thus the answer is this.
! ! ! !
1 −1 1 1 −1 1 −2 0
RepB,B (t) = =
2 −1 3 −1 −2 1 −5 2
As a quick check, we can take a vector at random
!
4
~v =
5
giving ! ! ! !
4 1 1 4 9
RepE2 (~v) = = = t(~v)
5 3 −1 5 7
while the calculation with respect to B, B
! ! ! !
1 −2 0 1 −2
RepB (~v) = =
−3 −5 2 −3 −11
B,B B B
yields the same result.
! ! !
1 −1 9
−2 · − 11 · =
2 −1 7
(b) We have
t
R2wrt E2 −−−−→ R2wrt E2
T
RepB,B (t) = RepE2 ,B (id) · T · RepB,E2 (id)
idy idy
t
R2wrt B −−−−→ R2wrt B
T̂
and, as in the first item of this question
RepB,E2 (id) = β ~1 ··· β ~n RepE2 ,B (id) = RepB,E2 (id)−1
so, writing Q for the matrix whose columns are the basis vectors, we have that
RepB,B (t) = Q−1 T Q.
210 Linear Algebra, by Hefferon
h
Vwrt B2 −−−−→ Wwrt D
Ĥ
Since there is no need to change bases in W (or we can say that the change
of basis matrix P is the identity), we have RepB2 ,D (h) = RepB1 ,D (h) · Q where
Q = RepB2 ,B1 (id).
(b) Here, this is the arrow diagram.
h
Vwrt B −−−−→ Wwrt D1
H
Q
idy idyP
h
Vwrt B −−−−→ Wwrt D2
Ĥ
We have that RepB,D2 (h) = P · RepB,D1 (h) where P = RepD1 ,D2 (id).
Three.V.2.26 (a) Here is the arrow diagram, and a version of that diagram for inverse
functions.
h h−1
Vwrt B −−−−→ Wwrt D Vwrt B ←−−−− Wwrt D
H H−1
Q Q
idy idyP idy idyP
h h−1
Vwrt B̂ −−−−→ Wwrt D̂ Vwrt B̂ ←−−−− Wwrt D̂
Ĥ Ĥ−1
Yes, the inverses of the matrices represent the inverses of the maps. That is, we
can move from the lower right to the lower left by moving up, then left, then down.
In other words, where Ĥ = PHQ (and P, Q invertible) and H, Ĥ are invertible then
Ĥ−1 = Q−1 H−1 P−1 .
(b) Yes; this is the prior part repeated in different terms.
(c) No, we need another assumption: if H represents h with respect to the same
starting as ending bases B, B, for some B then H2 represents h ◦ h. As a specific
example, these two matrices are both rank one and so they are matrix equivalent
! !
1 0 0 0
0 0 1 0
but the squares are not matrix equivalent — the square of the first has rank one
while the square of the second has rank zero.
(d) No. These two are not matrix equivalent but have matrix equivalent squares.
! !
0 0 0 0
0 0 1 0
Answers to Exercises 211
Projection
Three.VI.1.13 Any vector in Rn is the projection of some other into a line, provided
that the dimension n is greater than one. (Clearly, any vector is the projection of
itself into a line containing itself; the question is to produce some vector other than ~v
that projects to ~v.)
Suppose that ~v ∈ Rn with n > 1. If ~v 6= ~0 then we consider the line ` = {c~v | c ∈ R }
and if ~v = ~0 we take ` to be any (non-degenerate) line at all (actually, we needn’t
distinguish between these two cases — see the prior exercise). Let v1 , . . . , vn be the
components of ~v; since n > 1, there are at least two. If some vi is zero then the vector
~ = ~ei is perpendicular to ~v. If none of the components is zero then the vector w
w ~
whose components are v2 , −v1 , 0, . . . , 0 is perpendicular to ~v. In either case, observe
that ~v + w~ does not equal ~v, and that ~v is the projection of ~v + w ~ into `.
~ ) • ~v
(~v + w ~v • ~v w~ • ~v ~v • ~v
· ~v = + · ~v = · ~v = ~v
~v ~v
• ~v ~v
• ~v ~v
• ~v • ~v
We can dispose of the remaining n = 0 and n = 1 cases. The dimension n = 0 case
is the trivial vector space, here there is only one vector and so it cannot be expressed
as the projection of a different vector. In the dimension n = 1 case there is only one
(non-degenerate) line, and every vector is in it, hence every vector is the projection
only of itself.
the distance squared from the point to the line is this (we write a vector dotted with
itself w
~ •w~ as w~ 2 ).
~v • ~s ~v • ~s ~v • ~s ~v • ~s
k~v − · ~s k2 = ~v • ~v − ~v • ( · ~s) − ( · ~s ) • ~v + ( · ~s )2
~s • ~s ~s • ~s ~s • ~s ~s • ~s
~v • ~s ~v • ~s
= ~v • ~v − 2 · ( ) · ~v • ~s + ( ) · ~s • ~s
~s • ~s ~s • ~s
(~v • ~v ) · (~s • ~s ) − 2 · (~v • ~s )2 + (~v • ~s )2
=
~s • ~s
(~v • ~v )(~s • ~s ) − (~v • ~s )2
=
~s • ~s
Three.VI.1.16 Because square root is a strictly increasing function, we can minimize
d(c) = (cs1 − v1 )2 + (cs2 − v2 )2 instead of the square root of d. The derivative is
dd/dc = 2(cs1 − v1 ) · s1 + 2(cs2 − v2 ) · s2 . Setting it equal to zero 2(cs1 − v1 ) · s1 +
2(cs2 − v2 ) · s2 = c · (2s21 + 2s22 ) − (v1 s1 + v2 s2 ) = 0 gives the only critical point.
v 1 s1 + v 2 s2 ~v • ~s
c= 2 2
=
s1 + s2 ~s • ~s
Now the second derivative with respect to c
d2 d
= 2s1 2 + 2s2 2
dc2
is strictly positive (as long as neither s1 nor s2 is zero, in which case the question is
trivial) and so the critical point is a minimum.
The generalization to Rn is straightforward. Consider dn (c) = (cs1 − v1 )2 + · · · +
(csn − vn )2 , take the derivative, etc.
Three.VI.1.17 The Cauchy-Schwarz inequality |~v • ~s | 6 k~v k · k~s k gives that this fraction
~v • ~s ~v • ~s |~v • ~s | |~v • ~s |
k · ~s k = | | · k~s k = 2
· k~s k =
~s ~s
• ~s ~s
• k~s k k~s k
when divided by k~v k is less than or equal to one. That is, k~v k is larger than or equal
to the fraction.
Three.VI.1.18 Write c~s for ~q, and calculate: (~v • c~s/c~s • c~s ) · c~s = (~v • ~s/~s • ~s ) · ~s.
Three.VI.1.19 (a) Fixing !
1
~s =
1
as the vector whose span is the line, the formula gives this action,
! !
x 1
! • ! ! !
x y 1 1 x+y 1 (x + y)/2
7→ ! !· = · =
y 1 1 1 2 1 (x + y)/2
•
1 1
216 Linear Algebra, by Hefferon
(b)
!
0
~κ1 =
1
! !
−1 0
! ! ! • !
−1 −1 −1 3 1 0
~κ2 = − proj[~κ1 ] ( )= − ! ! ·
3 3 3 0 0 1
•
1 1
! ! !
−1 3 0 −1
= − · =
3 1 1 0
(c)
!
0
~κ1 =
1
! !
−1 0
! ! ! • !
−1 −1 −1 0 1 0
~κ2 = − proj[~κ1 ] ( )= − ! ! ·
0 0 0 0 0 1
•
1 1
! ! !
−1 0 0 −1
= − · =
0 1 1 0
Three.VI.2.15 If that set is not linearly independent, then we get a zero vector. Otherwise
(if our set is linearly independent but does not span the space), we are doing Gram-
Schmidt on a set that is a basis for a subspace and so we get an orthogonal basis for a
subspace.
Three.VI.2.17 (a) The argument is as in the i = 3 case of the proof of Theorem 2.7.
The dot product
~κi • ~v − proj[~κ1 ] (~v ) − · · · − proj[~vk ] (~v )
can be written as the sum of terms of the form −~κi • proj[~κj ] (~v ) with j 6= i, and
the term ~κi • (~v − proj[~κi ] (~v )). The first kind of term equals zero because the
~κ’s are mutually orthogonal. The other term is zero because this projection is
orthogonal (that is, the projection definition makes it zero: ~κi • (~v − proj[~κi ] (~v )) =
~κi • ~v − ~κi • ((~v • ~κi )/(~κi • ~κi )) · ~κi equals, after all of the cancellation is done, zero).
(b) The vector ~v is in black and the vector proj[~κ1 ] (~v ) + proj[~v2 ] (~v ) = 1 · ~e1 + 2 · ~e2
is in gray.
The vector ~v − (proj[~κ1 ] (~v ) + proj[~v2 ] (~v )) lies on the dotted line connecting the
black vector to the gray one, that is, it is orthogonal to the xy-plane.
(c) We get this diagram by following the hint.
The dashed triangle has a right angle where the gray vector 1 · ~e1 + 2 · ~e2 meets
the vertical dashed line ~v − (1 · ~e1 + 2 · ~e2 ); this is what first item of this question
proved. The Pythagorean theorem then gives that the hypotenuse — the segment
from ~v to any other vector — is longer than the vertical dashed line.
More formally, writing proj[~κ1 ] (~v ) + · · · + proj[~vk ] (~v ) as c1 · ~κ1 + · · · + ck · ~κk ,
Answers to Exercises 223
consider any other vector in the span d1 · ~κ1 + · · · + dk · ~κk . Note that
Three.VI.2.18 One way to proceed is to find a third vector so that the three together
make a basis for R3 , e.g.,
1
~
β3 = 0
0
(the second vector is not dependent on the third because it has a nonzero second
component, and the first is not dependent on the second and third because of its
nonzero third component), and then apply the Gram-Schmidt process. The first
element of the new basis is this.
1
~κ1 = 5
−1
The result ~κ3 is orthogonal to both ~κ1 and ~κ2 . It is therefore orthogonal to every
vector in the span of the set {~κ1 , ~κ2 }, including the two vectors given in the question.
that ~κ1 • ~κ1 = 1 and ~κ2 • ~κ2 = 1 because these vectors are members of an orthonormal
set.
Three.VI.2.21 It is true, except for the zero vector. Every vector in Rn except the zero
vector is in a basis, and that basis can be orthogonalized.
Three.VI.2.22 The 3×3 case gives the idea. The set
a b c
{ d , e , f }
g h i
is orthonormal if and only if these nine conditions all hold
a b c
(a d g) • d = 1 (a d g) • e = 0 (a d g) • f = 0
g h i
a b c
(b e h) d = 0 (b e h) e = 1 (b e h) f = 0
•
•
•
g h i
a b c
(c f i) • d = 0 (c f i) • e = 0 (c f i) • f = 1
g h i
(the three conditions in the lower left are redundant but nonetheless correct). Those,
in turn, hold if and only if
a d g a b c 1 0 0
b e h d e f = 0 1 0
c f i g h i 0 0 1
as required.
This is an example, the inverse of this matrix is its transpose.
√ √
1/ 2 1/ 2 0
√ √
−1/ 2 1/ 2 0
0 0 1
Three.VI.2.23 If the set is empty then the summation on the left side is the linear
combination of the empty set of vectors, which by definition adds to the zero vector.
In the second sentence, there is not such i, so the ‘if . . . then . . . ’ implication is
vacuously true.
Three.VI.2.24 (a) Part of the induction argument proving Theorem 2.7 checks that ~κi
~ 1, . . . , β
is in the span of hβ ~ i i. (The i = 3 case in the proof illustrates.) Thus, in
the change of basis matrix RepK,B (id), the i-th column RepB (~κi ) has components
i + 1 through k that are zero.
Answers to Exercises 227
(b) One way to see this is to recall the computational procedure that we use to
find the inverse. We write the matrix, write the identity matrix next to it, and
then we do Gauss-Jordan reduction. If the matrix starts out upper triangular then
the Gauss-Jordan reduction involves only the Jordan half and these steps, when
performed on the identity, will result in an upper triangular inverse matrix.
Three.VI.2.25 For the inductive step, we assume that for all j in [1..i], these three
conditions are true of each ~κj : (i) each ~κj is nonzero, (ii) each ~κj is a linear combination
~ 1, . . . , β
of the vectors β ~ j , and (iii) each ~κj is orthogonal to all of the ~κm ’s prior to it
(that is, with m < j). With those inductive hypotheses, consider ~κi+1 .
By the inductive assumption (ii) we can expand each ~κj into a linear combination of
~ 1, . . . , β
β ~j
~ i+1 • ~κ1
β
~ i+1 −
=β ~1
·β
~κ1 • ~κ1
~ i+1 • ~κ2
β
− · linear combination of β ~ 1, β
~2
~κ2 • ~κ2
~ i+1 • ~κi
β
− ··· − · linear combination of β~ 1, . . . , β
~i
~κi • ~κi
m < i + 1 (which is zero because by the hypothesis (iii) the vectors ~κj and ~κm are
orthogonal).
228 Linear Algebra, by Hefferon
Three.VI.3.11 As in Example 3.5, we can simplify the calculation by just finding the
space of vectors perpendicular to all the the vectors in M’s basis.
(a) Parametrizing to get !
−1
M = {c · | c ∈ R}
1
gives that ! ! ! !
u u −1 u
M⊥ { |0= • }={ | 0 = −u + v }
v v 1 v
Parametrizing the one-equation linear system gives this description.
!
1
M⊥ = {k · | k ∈ R}
1
(b) As in the answer to the prior part, we can describe M as a span
! !
3/2 3/2
M = {c · | c ∈ R} BM = h i
1 1
and then M⊥ is the set of vectors perpendicular to the one vector in this basis.
! !
u −2/3
M⊥ = { | (3/2) · u + 1 · v = 0 } = {k · | k ∈ R}
v 1
(c) Parametrizing the linear requirement in the description of M gives this basis.
! !
1 1
M = {c · | c ∈ R} BM = h i
1 1
Now, M⊥ is the set of vectors perpendicular to (the one vector in) BM .
! !
u −1
M ={
⊥
| u + v = 0 } = {k · | k ∈ R}
v 1
(By the way, this answer checks with the first item in this question.)
(d) Every vector in the space is perpendicular to the zero vector so M⊥ = Rn .
(e) The appropriate description and basis for M are routine.
! !
0 0
M = {y · | y ∈ R} BM = h i
1 1
Then ! !
u 1
M ={⊥
| 0 · u + 1 · v = 0} = {k · | k ∈ R}
v 0
and so (y-axis)⊥ = x-axis.
(f) The description of M is easy to find by parametrizing.
3 1 3 1
M = { c · 1 + d · 0 | c, d ∈ R } BM = h1 , 0i
0 1 0 1
230 Linear Algebra, by Hefferon
Finding M⊥ here just requires solving a linear system with two equations
3u + v =0 −(1/3)ρ1 +ρ2 3u + v =0
−→
u +w=0 −(1/3)v + w = 0
and parametrizing.
−1
M⊥ = { k · 3 | k ∈ R}
1
(g) Here, M is one-dimensional
0 0
M = {c · −1 | c ∈ R} BM = h−1i
1 1
and as a result, M⊥ is two-dimensional.
u 1 0
M⊥ = { v | 0 · u − 1 · v + 1 · w = 0 } = {j · 0 + k · 1 | j, k ∈ R}
w 0 1
Three.VI.3.12 (a) Parametrizing the equation leads to this basis for P.
1 0
BP = h0 , 1i
3 2
(b) Because R3 is three-dimensional and P is two-dimensional, the complement P⊥
must be a line. Anyway, the calculation as in Example 3.5
x ! x !
1 0 3 0
P⊥ = { y | y = }
0 1 2 0
z z
gives this basis for P⊥ .
3
BP⊥ = h 2 i
−1
1 1 0 3
(c) 1 = (5/14) · 0 + (8/14) · 1 + (3/14) · 2
2 3 2 −1
1 5/14
(d) projP (1) = 8/14
2 31/14
Answers to Exercises 231
Three.VI.3.20 True; the only vector orthogonal to itself is the zero vector.
Three.VI.3.21 This is immediate from the statement in Lemma 3.7 that the space is the
direct sum of the two.
Three.VI.3.22 The two must be equal, even only under the seemingly weaker condition
that they yield the same result on all orthogonal projections. Consider the subspace
M spanned by the set {~v1 ,~v2 }. Since each is in M, the orthogonal projection of ~v1
into M is ~v1 and the orthogonal projection of ~v2 into M is ~v2 . For their projections
into M to be equal, they must be equal.
Three.VI.3.23 (a) We will show that the sets are mutually inclusive, M ⊆ (M⊥ )⊥ and
(M⊥ )⊥ ⊆ M. For the first, if m ~ ∈ M then by the definition of the perp operation,
~ is perpendicular to every ~v ∈ M⊥ , and therefore (again by the definition of
m
the perp operation) m ~ ∈ (M⊥ )⊥ . For the other direction, consider ~v ∈ (M⊥ )⊥ .
Lemma 3.7’s proof shows that Rn = M ⊕ M⊥ and that we can give an orthogonal
basis for the space h~κ1 , . . . , ~κk , ~κk+1 , . . . , ~κn i such that the first half h~κ1 , . . . , ~κk i is
a basis for M and the second half is a basis for M⊥ . The proof also checks that each
vector in the space is the sum of its orthogonal projections into the lines spanned
by these basis vectors.
~v = proj[~κ1 ] (~v ) + · · · + proj[~κn ] (~v )
Because ~v ∈ (M⊥ )⊥ , it is perpendicular to every vector in M⊥ , and so the projections
in the second half are all zero. Thus ~v = proj[~κ1 ] (~v ) + · · · + proj[~κk ] (~v ), which is a
linear combination of vectors from M, and so ~v ∈ M. (Remark. Here is a slicker
way to do the second half: write the space both as M ⊕ M⊥ and as M⊥ ⊕ (M⊥ )⊥ .
Because the first half showed that M ⊆ (M⊥ )⊥ and the prior sentence shows that
the dimension of the two subspaces M and (M⊥ )⊥ are equal, we can conclude that
M equals (M⊥ )⊥ .)
(b) Because M ⊆ N, any ~v that is perpendicular to every vector in N is also
perpendicular to every vector in M. But that sentence simply says that N⊥ ⊆ M⊥ .
(c) We will again show that the sets are equal by mutual inclusion. The first direction
is easy; any ~v perpendicular to every vector in M + N = { m ~ +n ~ |m ~ ∈ M, n ~ ∈ N}
is perpendicular to every vector of the form m ~ + ~0 (that is, every vector in M) and
every vector of the form ~0 + n ~ (every vector in N), and so (M + N)⊥ ⊆ M⊥ ∩ N⊥ .
The second direction is also routine; any vector ~v ∈ M⊥ ∩N⊥ is perpendicular to any
vector of the form cm+d~
~ n because ~v • (cm+d~
~ ~
n) = c·(~v • m)+d·(~ ~ ) = c·0+d·0 = 0.
v•n
Answers to Exercises 235
respect to the standard bases and taking ~h to be the column vector gotten by
transposing the one row of that matrix representation.
(c) Of course, !
1 2 3
RepE3 ,E2 (f) =
4 5 6
and so the null space is this
set.
v1 ! v
1
!
1 2 3 0
N (f){ v2 | v2 = }
4 5 6 0
v3 v3
That description makes clear that
1 4
2 , 5 ∈ N (f)⊥
3 6
and since N (f) is a subspace of R , the span of the two vectors is a subspace of
⊥ n
(d) As above, generalizing from the specific case is easy: for any f : Rn → Rm the
matrix H representing the map with respect to the standard bases describes the
action
h1,1 v1 + h1,2 v2 + · · · + h1,n vn
v1
.. f ..
. 7−→
.
vn hm,1 v1 + hm,2 v2 + · · · + hm,n vn
and the description of the null space gives that on transposing the m rows of H
h1,1 hm,1
h1,2 hm,2
~h1 = . , . . . ~hm = .
. .
. .
h1,n hm,n
we have N (f)⊥ = [{ ~h1 , . . . , ~hm }]. ([Strang 93] describes this space as the transpose
of the row space of H.)
Three.VI.3.25 (a) First note that if a vector ~v is already in the line then the orthogonal
projection gives ~v itself. One way to verify this is to apply the formula for projection
into the line spanned by a vector ~s, namely (~v • ~s/~s • ~s) · ~s. Taking the line as
{ k · ~v | k ∈ R } (the ~v = ~0 case is separate but easy) gives (~v • ~v/~v • ~v) · ~v, which
simplifies to ~v, as required.
Now, that answers the question because after once projecting into the line, the
result proj` (~v) is in that line. The prior paragraph says that projecting into the
same line again will have no effect.
(b) The argument here is similar to the one in the prior item. With V = M ⊕ N,
the projection of ~v = m ~ +n ~ is projM,N (~v ) = m.
~ Now repeating the projection will
give projM,N (m) ~ = m, ~ as required, because the decomposition of a member of M
into the sum of a member of M and a member of N is m ~ =m ~ + ~0. Thus, projecting
twice into M along N has the same effect as projecting once.
(c) As suggested by the prior items, the condition gives that t leaves vectors in
the range space unchanged, and hints that we should take β ~ 1, . . . , β
~ r to be basis
vectors for the range, that is, that we should take the range space of t for M (so
that dim(M) = r). As for the complement, we write N for the null space of t and
we will show that V = M ⊕ N.
To show this, we can show that their intersection is trivial M ∩ N = {~0 } and
that they sum to the entire space M + N = V. For the first, if a vector m ~ is in the
range space then there is a ~v ∈ V with t(~v) = m, ~ and the condition on t gives that
~ = (t ◦ t) (~v) = t(~v) = m,
t(m) ~ while if that same vector is also in the null space
then t(m) ~
~ = 0 and so the intersection of the range space and null space is trivial.
For the second, to write an arbitrary ~v as the sum of a vector from the range space
and a vector from the null space, the fact that the condition t(~v) = t(t(~v)) can be
rewritten as t(~v − t(~v)) = ~0 suggests taking ~v = t(~v) + (~v − t(~v)).
To finish we taking a basis B = hβ ~ 1, . . . , β
~ n i for V where hβ
~ 1, . . . , β
~ r i is a basis
for the range space M and hβ ~ r+1 , . . . , β
~ n i is a basis for the null space N.
(d) Every projection (as defined in this exercise) is a projection into its range space
and along its null space.
(e) This also follows immediately from the third item.
T
Three.VI.3.26 For any matrix M we have that (M−1 ) = (MT )−1 , and for any two
matrices M, N we have that MNT = NT MT (provided, of course, that the inverse and
product are defined). Applying these two gives that the matrix equals its transpose.
T T T
A(AT A)−1 AT = (AT )( (AT A)−1 )(AT )
T T −1 T
= (AT )( (AT A) )(AT ) = A(AT AT )−1 AT = A(AT A)−1 AT
1 As with the first example discussed above, we are trying to find a best m to “solve”
this system.
8m = 4
16m = 9
24m = 13
32m = 17
40m = 20
Projecting into the linear subspace gives this
4 8
9 16
13 24
•
17 32 8 8
16 16
20 40 1832
· 24 = · 24
8 8 3520
16 16 32 32
40 40
24 24
•
32 32
40 40
so the slope of the line of best fit is approximately 0.52.
238 Linear Algebra, by Hefferon
20
15
10
0
0 10 20 30 40
280
260
240
220
1850 1900 1950 2000
250
240
230
220
210
200
1900 1920 1940 1960 1980 2000
5 These are the equations of the lines for men’s and women’s mile (the vertical intercept
term of the equation for the women’s mile has been adjusted from the answer above,
to zero it at the year 0, because that’s how the men’s mile equation was done).
y = 994.8276974 − 0.3871993827x
y = 3125.6426 − 1.445753225x
Obviously the lines cross. A computer program is the easiest way to do the arith-
metic: MuPAD gives x = 2012.949004 and y = 215.4150856 (215 seconds is 3 minutes
and 35 seconds). Remark. Of course all of this projection is highly dubious — for one
thing, the equation for the women is influenced by the quite slow early times — but it
is nonetheless fun.
240 Linear Algebra, by Hefferon
380
360
340
320
300
280
260
240
220
1850 1900 1950 2000
6 Sage gives the line of best fit as toll = −0.05 · dist + 5.63.
sage: dist = [2, 7, 8, 16, 27, 47, 67, 82, 102, 120]
sage: toll = [6, 6, 6, 6.5, 2.5, 1, 1, 1, 1, 1]
sage: var('a,b,t')
(a, b, t)
sage: model(t) = a*t+b
sage: data = zip(dist,toll)
sage: fit = find_fit(data, model, solution_dict=True)
sage: model.subs(fit)
t |--> -0.0508568169130319*t + 5.630955848442933
sage: p = plot(model.subs(fit), (t,0,120))+points(data,size=25,color='red')
sage: p.save('bridges.pdf')
But the graph shows that the equation has little predictive value.
20 40 60 80 100 120
Apparently a better model is that (with only one intermediate exception) crossings in
the city cost roughly the same as each other, and crossings upstate cost the same as
each other.
7 (a) A computer algebra system like MAPLE or MuPAD will give an intercept of
b = 4259/1398 ≈ 3.239628 and a slope of m = −71/2796 ≈ −0.025393419 Plugging
x = 31 into the equation yields a predicted number of O-ring failures of y = 2.45
(rounded to two places). Plugging in y = 4 and solving gives a temperature of
x = −29.94◦ F.
Answers to Exercises 241
0
40 50 60 70 80
8 (a) The plot is nonlinear.
20
15
10
0
0 2 4 6
(b) Here is the plot.
0.5
−0.5
0 2 4 6
There is perhaps a jog up between planet 4 and planet 5.
(c) This plot seems even more linear.
1
0.5
−0.5
0 2 4 6 8
(d) With this input
1 1 −0.40893539
1 2 −0.1426675
1 3 0
A = 1 4 b = 0.18184359
1 6 0.71600334
1 7 0.97954837
1 8 1.2833012
MuPAD gives that the intercept is b = −0.6780677466 and the slope is m =
0.2372763818.
(e) Plugging x = 9 into the equation y = −0.6780677466 + 0.2372763818x from the
prior item gives that the log of the distance is 1.4574197, so the expected distance
is 28.669472. The actual distance is about 30.003.
(f) Plugging x = 10 into the same equation gives that the log of the distance is
1.6946961, so the expected distance is 49.510362. The actual distance is about
39.503.
√ √ ! √
! !
ρ1 +ρ2 2/2 2/2 (2/ 2)ρ1 1 1 −ρ2 +ρ1 1 0
−→ √ −→
√ −→
0 2 (1/ 2)ρ2 0 1 0 1
! √ ! !
1 −1 2/ 2 0 1 0
√ H=I
0 1 0 1/ 2 1 1
! √ ! !
1 0 2/2 0 1 1
H= √ I
−1 1 0 2 0 1
| {z }
P
gives the desired factorization of H (here, the partial identity is I, and Q is trivial,
that is, it is also an identity matrix).
(d) Reading the composition from right to left (and ignoring the identity matrices
as trivial) gives that H has the same effect as first performing this skew
x
7→
x + y
~u h(~u)
y y
~v −→ h(~v)
√
followed by a dilation that multiplies all first components by 2/2 (this is a “shrink”
√ √
in that 2/2 ≈ 0.707) and all second components by 2, followed by another skew.
~u x
7→
x
~v y −x + y h(~u)
−→
h(~v)
For instance, the effect of H on the unit vector whose angle with the x-axis is π/6
is this.
244 Linear Algebra, by Hefferon
√ √
3/2 ( 3 + 1)/2
x
7→
x + y
1/2 1/2
y y
−→
√ √
√
2( 3 +√1)/2
x ( 2/2)x
7→ √ 2/2
y 2y
−→
x
7→
x
y −x + y
−→
√ √
√2( 3 +
√1)/4
2(1 − 3)/4
Verifying that the resulting vector has unit length and forms an angle of −π/12
with the x-axis is routine.
2 We will first represent the map with a matrix H, perform the row operations and,
if needed, column operations to reduce it to a partial-identity matrix. We will then
translate that into a factorization H = PBQ. Substituting into the general matrix
!
cos θ − sin θ
RepE2 ,E2 (rθ )
sin θ cos θ
gives this representation.
√ !
−1/2 − 3/2
RepE2 ,E2 (r2π/3 ) √
3/2 −1/2
Gauss’s Method is routine.
√ √ ! √ ! √
!
3ρ1 +ρ2 −1/2 − 3/2 −2ρ1 1 3 − 3ρ2 +ρ1 1 0
−→ −→ −→
0 −2 (−1/2)ρ2 0 1 0 1
That translates to a matrix equation in this way.
√ ! ! ! √ !
1 − 3 −2 0 1 0 −1/2 − 3/2
√ √ =I
0 1 0 −1/2 3 1 3/2 −1/2
Taking inverses to solve for H yields this factorization.
√ ! ! ! √ !
−1/2 − 3/2 1 0 −1/2 0 1 3
√ = √ I
3/2 −1/2 − 3 1 0 −2 0 1
Answers to Exercises 245
f(x)
x f−1 (f(x))
1 (a) The sum of the entries of M is the sum of the sums of the three rows.
(b) The constraints on entries of M involving the center entry make this system.
m2,1 + m2,2 + m2,3 = s
m1,2 + m2,2 + m3,2 = s
m1,1 + m2,2 + m3,3 = s
m1,3 + m2,2 + m3,1 = s
Adding those four equations counts each matrix entry once and only once, except
that we count the center entry four times. Thus the left side sums to 3s + 3m2,2
while the right sums to 4s. So 3m2,2 = s.
Answers to Exercises 247
(c) The second row adds to s so m2,1 + m2,2 + m2,3 = 3m2,2 , giving that (1/2) ·
(m2,1 + m2,3 ) = m2,2 . The same goes for the column and the diagonals.
(d) By the prior exercise either both m2,1 and m2,3 are equal to m2,2 or else one is
greater while one is smaller. Thus m2,2 is the median of the set {m2,1 , m2,2 , m2,3 }.
The same reasoning applied to the second column shows that Thus m2,2 is the
median of the set { m1,2 , m2,1 , m2,2 , m2,3 , m3,2 }. Extending to the two diagonals
shows it is the median of the set of all entries.
2 For any k we have this.
1 1 0 0 s 1 1 0 0 s
0 0 1 1 s 0 0 1 1 s
1 0 1 0 s 0 −1 1 0 0
−ρ1 +ρ3
−→
0 1 0 1 s −ρ1 +ρ5 0 1 0 1 s
1 0 0 1 s 0 −1 0 1 0
0 1 1 0 s 0 1 1 0 s
1 1 0 0 s 1 1 0 0 s
0 1 1 0 s 0 1 1 0 s
0 −1 1 0 0 0 0 2 0 s
−ρ2 ↔ρ6 −ρ2 +ρ3
−→ −→
0 1 0 1 s 0 1 −1 1 0
−ρ2 +ρ4
ρ2 +ρ5
0 −1 0 1 0 0 0 1 1 s
0 0 1 1 s 0 0 1 1 s
The unique solution is a = b = c = d = s/2.
3 By the prior exercise the only member is Z2×2 .
4 (a) Where M, N ∈ Mn×n we have Tr(cM + dN) = (cm1,1 + dn1,1 ) + · · · + (cmn,n +
dnn,n ) = (cm1,1 +· · ·+cmn,n )+(dn1,1 +· · ·+dnn,n ) = c·Tr(M)+d·Tr(N) where
all numbers are real, so the trace preserves linear combinations. The argument for
Tr∗ is similar.
(b) It preserves linear combinations: where all numbers are real, θ(cM + dN) =
(Tr(cM + dN), Tr∗ (cM + dN)) = (c · Tr(M) + d · Tr(N), c · Tr∗ (M) + d · Tr∗ (N)) =
c · θ(M) + d · θ(N).
(c) Where h1 , . . . , hn : V → W are linear then so is g : V → W n given by g(~v) =
(h1 (~v), . . . , hn (~v)). The proof just follows the proof of the prior item.
5 (a) The sum of two semimagic squares is semimagic, as is a scalar multiple of
a semimagic square. (b) As with the prior item, a linear combination of two
semimagic squares with magic number zero is also such a matrix.
6 (a) Consider the matrix C ∈ Hn that has all entries zero except that the four
corners are c1,1 = cn,n = 1 and c1,n = cn,1 = −1. Also consider the matrix
248 Linear Algebra, by Hefferon
D ∈ Hn with all entries zero except that d1,1 = d2,2 = 1 and d1,2 = d2,1 = −1.
We have
2
!
2 −1 if n = 3
θ(C) = θ(D) =
−2 2 if n > 3
0
and so the image of θ includes a basis for R2 and thus θ is onto. With that, because
for any linear map the dimension of the domain equals its rank plus its nullity we
conclude that dim(Hn ) = 2 + dim(Mn,0 ), as desired.
(b) We claim that φ : Hn,0 → M(n−1)×(n−1) . is one-to-one and onto.
To show that it is one-to-one we will show that the only member of Hn,0 mapped
to the zero matrix Z(n−1)×(n−1) is the zero matrix Zn×n . Suppose that M ∈ Hn×n
and φ(M) = Z(n−1)×(n−1) . On all but the final row and column φ is the identity
so the entries in M in all but the final row and column are zero: mi,j = 0 for
i, j ∈ {1 . . . n − 1}. The first row of M adds to zero and hence the final entry in
that row m1,n is zero. Similarly the final entry in each row i ∈ {1 . . . n − 1} and
column j ∈ {1 . . . n − 1} is zero. Then, the final column adds to zero so mn,n = 0.
Therefore M is the zero matrix Zn×n and the restriction of φ is one-to-one.
(c) Consider a member M̂ of the codomain M(n−1)×(n−1) . We will produce a matrix
M from the domain Hn,0 that maps to it. The function φ is the identity on all but
the final row and column of M so for i, j ∈ {1 . . . n − 1} the entries
are mi,j = m̂i,j .
m̂1,1 m̂1,2 ... m̂1,n−1 m1,n
.. ..
M=
. .
m̂n−1,1 m̂n−1,2 . . . m̂n−1,n−1 mn−1,n
mn,1 mn,2 ... mn,n−1 mn,n
The first row of M must add to zero so we take m1,n to be −(m̂1,1 + · · · + m̂1,n−1 ).
In the same way we get the final entries mi,n = −(m̂i,1 +· · ·+ m̂i,n−1 ) in all the rows
but the bottom i ∈ {1 . . . n − 1}, and the final entries mn,j = −(m̂1,j +· · ·+ m̂n−1,j )
in all the columns but the last j ∈ {1 . . . n − 1}. The entry remaining is the
one in the lower right mn,n . The final column adds to zero so we set it to
−(m1,n + · · · + mn−1,n ) but we must check that the final row now also adds to
zero. We have mn,n = −m1,n − · · · − mn−1,n and expanding each of the mi,n
as −m̂1,1 − · · · − m̂1,n−1 gives that we have defined mn,n to be the sum of all
the entries of M̂. The sum of the all the entries but the last in the final row is
m1,n + m2,n + · · · + mn−1,n and expanding each mn,j = −m̂1,j − · · · − m̂n−1,j
verifies that the sum of the final row is zero. Thus M is semimagic with magic
number zero and so φ is onto.
(d) Theorem Two.II.2.14 says that for any linear map the dimension of the domain
equals its rank plus its nullity. Because φ : Hn → M(n−1)×(n−1) is one-to-one its
nullity is zero. Because it is onto its rank is dim(M(n−1)×(n−1) ) = (n − 1)2 . Thus
the domain of φ, the subspace Hn,0 of semimagic squares with magic number zero,
has dimension (n − 1)2 .
(e) We have that dim Mn = dim Mn,0 +1 = (dim Hn −2)+1 = (n−1)2 −1 = n2 −n
when n > 3.
0.00000
0.25000
0.00000
0.50000
0.00000
0.25000
This continued for too many steps to list here.
octave:26> v24=coin(p,v23)
v24 =
0.39600
0.00276
0.00000
0.00447
0.00000
0.59676
(b) Using these formulas
p1 (n + 1) = 0.5 · p2 (n) p2 (n + 1) = 0.5 · p1 (n) + 0.5 · p3 (n)
p3 (n + 1) = 0.5 · p2 (n) + 0.5 · p4 (n) p5 (n + 1) = 0.5 · p4 (n)
and these initial conditions
p0 (0) 0
p1 (0) 0
p (0) 0
2
=
p3 (0) 1
p4 (0) 0
p5 (0) 0
we will prove by induction that when n is odd then p1 (n) = p3 (n) = 0 and when
n is even then p2 (n) = p4 (n) = 0. Note first that this is true in the n = 0 base
case by the initial conditions. For the inductive step, suppose that it is true in the
n = 0, n = 1, . . . , n = k cases and consider the n = k + 1 case. If k + 1 is odd then
the two
p1 (k + 1) = 0.5 · p2 (k) = 0.5 · 0 = 0
p3 (k + 1) = 0.5 · p2 (k) + 0.5 · p4 (k) = 0.5 · 0 + 0.5 · 0 = 0
follow from the inductive hypothesis that p2 (k) = p4 (k) = 0 since k is even. The
case where k + 1 is even is similar.
(c) We can use, say, n = 100. This Octave session
octave:1> B=[1,.5,0,0,0,0;
> 0,0,.5,0,0,0;
> 0,.5,0,.5,0,0;
> 0,0,.5,0,.5,0;
> 0,0,0,.5,0,0;
Answers to Exercises 251
> 0,0,0,0,.5,1];
octave:2> B100=B**100
B100 =
1.00000 0.80000 0.60000 0.40000 0.20000 0.00000
0.00000 0.00000 0.00000 0.00000 0.00000 0.00000
0.00000 0.00000 0.00000 0.00000 0.00000 0.00000
0.00000 0.00000 0.00000 0.00000 0.00000 0.00000
0.00000 0.00000 0.00000 0.00000 0.00000 0.00000
0.00000 0.20000 0.40000 0.60000 0.80000 1.00000
octave:3> B100*[0;1;0;0;0;0]
octave:4> B100*[0;1;0;0;0;0]
octave:5> B100*[0;0;0;1;0;0]
octave:6> B100*[0;1;0;0;0;0]
yields these outputs.
starting with: $1 $2 $3 $4
s0 (100) 0.80000 0.60000 0.40000 0.20000
s1 (100) 0.00000 0.00000 0.00000 0.00000
s2 (100) 0.00000 0.00000 0.00000 0.00000
s3 (100) 0.00000 0.00000 0.00000 0.00000
s4 (100) 0.00000 0.00000 0.00000 0.00000
s5 (100) 0.20000 0.40000 0.60000 0.80000
2 (a) From these equations
s1 (n)/6 + 0s2 (n) + 0s3 (n) + 0s4 (n) + 0s5 (n) + 0s6 (n) = s1 (n + 1)
s1 (n)/6 + 2s2 (n)/6 + 0s3 (n) + 0s4 (n) + 0s5 (n) + 0s6 (n) = s2 (n + 1)
s1 (n)/6 + s2 (n)/6 + 3s3 (n)/6 + 0s4 (n) + 0s5 (n) + 0s6 (n) = s3 (n + 1)
s1 (n)/6 + s2 (n)/6 + s3 (n)/6 + 4s4 (n)/6 + 0s5 (n) + 0s6 (n) = s4 (n + 1)
s1 (n)/6 + s2 (n)/6 + s3 (n)/6 + s4 (n)/6 + 5s5 (n)/6 + 0s6 (n) = s5 (n + 1)
s1 (n)/6 + s2 (n)/6 + s3 (n)/6 + s4 (n)/6 + s5 (n)/6 + 6s6 (n)/6 = s6 (n + 1)
We get this transition matrix.
1/6 0 0 0 0 0
1/6 2/6 0 0 0 0
1/6 1/6 3/6 0 0 0
1/6 1/6 1/6 4/6 0 0
1/6 1/6 1/6 1/6 5/6 0
1/6 1/6 1/6 1/6 1/6 6/6
(b) This is the Octave session, with outputs edited out and condensed into the table
at the end.
octave:1> F=[1/6, 0, 0, 0, 0, 0;
> 1/6, 2/6, 0, 0, 0, 0;
> 1/6, 1/6, 3/6, 0, 0, 0;
252 Linear Algebra, by Hefferon
(c) This is a continuation of the Octave session from the prior item.
octave:7> p0=[.0000;.6522;.3478;.0000;.0000]
octave:8> p1=M*p0
octave:9> p2=M*p1
octave:10> p3=M*p2
octave:11> p4=M*p3
yields this plot. There is no threshold value — no probability above which the curve
rises sharply.
Answers to Exercises 255
0.4
line 1
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
(a) We can adapt the script from the end of this Topic.
# Octave script file to compute chance of World Series outcomes.
function w = markov(p,v)
q = 1-p;
A=[0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 0-0
p,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 1-0
q,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 0-1_
0,p,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 2-0
0,q,p,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 1-1
0,0,q,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 0-2__
0,0,0,p,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 3-0
0,0,0,q,p,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 2-1
0,0,0,0,q,p, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 1-2_
0,0,0,0,0,q, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 0-3
0,0,0,0,0,0, p,0,0,0,1,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 4-0
0,0,0,0,0,0, q,p,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 3-1__
0,0,0,0,0,0, 0,q,p,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 2-2
0,0,0,0,0,0, 0,0,q,p,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0; # 1-3
0,0,0,0,0,0, 0,0,0,q,0,0, 0,0,1,0,0,0, 0,0,0,0,0,0; # 0-4_
0,0,0,0,0,0, 0,0,0,0,0,p, 0,0,0,1,0,0, 0,0,0,0,0,0; # 4-1
0,0,0,0,0,0, 0,0,0,0,0,q, p,0,0,0,0,0, 0,0,0,0,0,0; # 3-2
0,0,0,0,0,0, 0,0,0,0,0,0, q,p,0,0,0,0, 0,0,0,0,0,0; # 2-3__
0,0,0,0,0,0, 0,0,0,0,0,0, 0,q,0,0,0,0, 1,0,0,0,0,0; # 1-4
0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,p,0, 0,1,0,0,0,0; # 4-2
0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,q,p, 0,0,0,0,0,0; # 3-3_
0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,q, 0,0,0,1,0,0; # 2-4
0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,p,0,1,0; # 4-3
0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,0,0,0,0, 0,0,q,0,0,1]; # 3-4
v7 = (A**7) * v;
w = v7(11)+v7(16)+v7(20)+v7(23)
endfunction
When the American League has a p = 0.55 probability of winning each game then
their probability of winning the series is 0.60829. When their probability of winning
any one game is p = 0.6 then their probability of winning the series is 0.71021.
(b) From this Octave session and its graph
octave:1> v0=[1;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0;0];
octave:2> x=(.01:.01:.99)';
octave:3> y=(.01:.01:.99)';
octave:4> for i=.01:.01:.99
> y(100*i)=markov(i,v0);
> endfor
octave:5> z=[x, y];
octave:6> gplot z
by eye we judge that if p > 0.7 then the team is close to assured of the series.
1
0.9 line 1
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
7 (a) They must satisfy this condition because the total probability of a state transition
(including back to the same state) is 100%.
(b) See the answer to the third item.
(c) We will do the 2×2 case; bigger-sized cases are just notational problems. This
product
! ! !
a1,1 a1,2 b1,1 b1,2 a1,1 b1,1 + a1,2 b2,1 a1,1 b1,2 + a1,2 b2,2
=
a2,1 a2,2 b2,1 b2,2 a2,1 b1,1 + a2,2 b2,1 a2,1 b1,2 + a2,2 b2,2
has these two column sums
(a1,1 b1,1 +a1,2 b2,1 )+(a2,1 b1,1 +a2,2 b2,1 ) = (a1,1 +a2,1 )·b1,1 +(a1,2 +a2,2 )·b2,1
= 1 · b1,1 + 1 · b2,1 = 1
and
(a1,1 b1,2 +a1,2 b2,2 )+(a2,1 b1,2 +a2,2 b2,2 ) = (a1,1 +a2,1 )·b1,2 +(a1,2 +a2,2 )·b2,2
= 1 · b1,2 + 1 · b2,2 = 1
as required.
1 (a) Yes.
(b) No, the columns do not have length one.
(c) Yes.
2 Some of!these are nonlinear, because they!involve!a nontrivial translation.
√ !
x x · cos(π/6) − y · sin(π/6) 0 x · ( 3/2) − y · (1/2) + 0
(a) 7→ + = √
y x · sin(π/6) + y · cos(π/6) 1 x · (1/2) + y · cos( 3/2) + 1
√
(b) The line y = 2x makes an angle of arctan(2/1) with the x-axis. Thus sin θ = 2/ 5
√
and cos θ = 1/ 5. ! √ √ !
x x · (1/ 5) − y · (2/ 5)
7→ √ √
y x · (2/ 5) + y · (1/ 5)
! √ √ ! ! √ √ !
x x · (1/ 5) − y · (−2/ 5) 1 x/ 5 + 2y/ 5 + 1
(c) 7→ √ √ + = √ √
y x · (−2/ 5) + y · (1/ 5) 1 −2x/ 5 + y/ 5 + 1
3 (a) Let f be distance-preserving and consider f−1 . Any two points in the codomain
can be written as f(P1 ) and f(P2 ). Because f is distance-preserving, the distance
from f(P1 ) to f(P2 ) equals the distance from P1 to P2 . But this is exactly what is
required for f−1 to be distance-preserving.
Answers to Exercises 259
(b) Any plane figure F is congruent to itself via the identity map id : R2 → R2 , which
is obviously distance-preserving. If F1 is congruent to F2 (via some f) then F2 is
congruent to F1 via f−1 , which is distance-preserving by the prior item. Finally, if
F1 is congruent to F2 (via some f) and F2 is congruent to F3 (via some g) then F1
is congruent to F3 via g ◦ f, which is easily checked to be distance-preserving.
4 The first two components of each are ax + cy + e and bx + dy + f.
5 (a) The Pythagorean Theorem gives that three points are collinear if and only if (for
some ordering of them into P1 , P2 , and P3 ), dist(P1 , P2 )+dist(P2 , P3 ) = dist(P1 , P3 ).
Of course, where f is distance-preserving, this holds if and only if dist(f(P1 ), f(P2 ))+
dist(f(P2 ), f(P3 )) = dist(f(P1 ), f(P3 )), which, again by Pythagoras, is true if and
only if f(P1 ), f(P2 ), and f(P3 ) are collinear.
The argument for betweeness is similar (above, P2 is between P1 and P3 ).
If the figure F is a triangle then it is the union of three line segments P1 P2 , P2 P3 ,
and P1 P3 . The prior two paragraphs together show that the property of being a
line segment is invariant. So f(F) is the union of three line segments, and so is a
triangle.
A circle C centered at P and of radius r is the set of all points Q such that
dist(P, Q) = r. Applying the distance-preserving map f gives that the image
f(C) is the set of all f(Q) subject to the condition that dist(P, Q) = r. Since
dist(P, Q) = dist(f(P), f(Q)), the set f(C) is also a circle, with center f(P) and
radius r.
(b) Here are two that are easy to verify: (i) the property of being a right triangle,
and (ii) the property of two lines being parallel.
(c) One that was mentioned in the section is the ‘sense’ of a figure. A triangle whose
vertices read clockwise as P1 , P2 , P3 may, under a distance-preserving map, be sent
to a triangle read P1 , P2 , P3 counterclockwise.
Chapter Four
Chapter Four:
Determinants
Definition
Four.I.1: Exploration
Four.I.1.1 (a) 4 (b) 3 (c) −12
Four.I.1.2 (a) 6 (b) 21 (c) 27
Four.I.1.3 For the first, apply the formula in this section, note that any term with a d,
g, or h is zero, and simplify. Lower-triangular matrices work the same way.
Four.I.1.4 (a) Nonsingular, the determinant is −1.
(b) Nonsingular, the determinant is −1.
(c) Singular, the determinant is 0.
Four.I.1.5 (a) Nonsingular, the determinant is 3.
(b) Singular, the determinant is 0.
(c) Singular, the determinant is 0.
Four.I.1.6 (a) det(B) = det(A) via −2ρ1 + ρ2
(b) det(B) = − det(A) via ρ2 ↔ ρ3
(c) det(B) = (1/2) · det(A) via (1/2)ρ2
Four.I.1.7 Using the formula for the determinant of a 3×3 matrix we expand the left
side
1 · b · c2 + 1 · c · a2 + 1 · a · b2 − b2 · c · 1 − c2 · a · 1 − a2 · b · 1
and by distributing we expand the right side.
(bc − ba − ac + a2 ) · (c − b) = c2 b − b2 c − bac + b2 a − ac2 + acb + a2 c − a2 b
Now we can just check that the two are equal. (Remark. This is the 3×3 case of
Vandermonde’s determinant which arises in applications).
262 Linear Algebra, by Hefferon
a = 0, d = 0, and g = 0 case is easy — that matrix is singular since the columns form
a linearly dependent set, and the determinant comes out to be zero.
Four.I.1.10 Figuring the determinant and doing some algebra gives this.
0 = y1 x + x2 y + x1 y2 − y2 x − x1 y − x2 y1
(x2 − x1 ) · y = (y2 − y1 ) · x + x2 y1 − x1 y2
y2 − y1 x2 y1 − x1 y2
y= ·x+
x2 − x1 x2 − x1
Note that this is the equation of a line (in particular, in contains the familiar expression
for the slope), and note that (x1 , y1 ) and (x2 , y2 ) satisfy it.
Four.I.1.11 (a) The comparison with the formula given in the preamble to this section
is easy.
(b) While it holds for 2×2 matrices
!
h1,1 h1,2 h1,1
= h1,1 h2,2 + h1,2 h2,1
h2,1 h2,2 h2,1
−h2,1 h1,2 − h2,2 h1,1
= h1,1 h2,2 − h1,2 h2,1
it does not hold for 4×4 matrices. An example is that this matrix is singular because
the second and third rows are equal
1 0 0 1
0 1 1 0
0 1 1 0
−1 0 0 1
but following the scheme of the mnemonic does not give zero.
1 0 0 1 1 0 0
0 1 1 0 0 1 1
=1+0+0+0
0 1 1 0 0 1 1
−(−1) − 0 − 0 − 0
−1 0 0 1 −1 0 0
Four.I.1.12 The determinant is (x2 y3 − x3 y2 )~e1 + (x3 y1 − x1 y3 )~e2 + (x1 y2 − x2 y1 )~e3 .
To check perpendicularity, we check that the dot product with the first vector is zero
x1 x2 y3 − x3 y2
x2 x3 y1 − x1 y3 = x1 x2 y3 −x1 x3 y2 +x2 x3 y1 −x1 x2 y3 +x1 x3 y2 −x2 x3 y1 = 0
•
x3 x1 y2 − x2 y1
and the dot product with the second vector is also zero.
y1 x2 y3 − x3 y2
y2 x3 y1 − x1 y3 = x2 y1 y3 −x3 y1 y2 +x3 y1 y2 −x1 y2 y3 +x1 y2 y3 −x2 y1 y3 = 0
•
y3 x1 y2 − x2 y1
264 Linear Algebra, by Hefferon
Four.I.1.13 (a) Plug and chug: the determinant of the product is this
! ! !
a b w x aw + by ax + bz
det( ) = det( )
c d y z cw + dy cx + dz
= acwx + adwz + bcxy + bdyz
−acwx − bcwz − adxy − bdyz
while the product of the determinants
! is !
this.
a b w x
det( ) · det( ) = (ad − bc) · (wz − xy)
c d y z
Verification that they are equal is easy.
(b) Use the prior item.
That similar matrices have the same determinant is immediate from the above two:
det(PT P−1 ) = det(P) · det(T ) · det(P−1 ).
Four.I.1.14 One way is to count these areas
B
A
y2 D
y1 C
F
E
x2 x1
by taking the area of the entire rectangle and subtracting the area of A the upper-left
rectangle, B the upper-middle triangle, D the upper-right triangle, C the lower-left
triangle, E the lower-middle triangle, and F the lower-right rectangle (x1 + x2 )(y1 +
y2 ) − x2 y1 − (1/2)x1 y1 − (1/2)x2 y2 − (1/2)x2 y2 − (1/2)x1 y1 − x2 y1 . Simplification
gives the determinant formula.
This determinant is the negative of the one above; the formula distinguishes
whether the second column is counterclockwise from the first.
Four.I.1.15 The computation for 2×2 matrices, using the formula quoted in the preamble,
is easy. It does also hold for 3×3 matrices; the computation is routine.
Four.I.1.16 No. Recall that ! constants come out!one row at a time. !
2 4 1 2 1 2
det( ) = 2 · det( ) = 2 · 2 · det( )
2 6 2 6 1 3
This contradicts linearity (here we didn’t need S, i.e., we can take S to be the matrix
of zeros).
Four.I.1.17 Bring out the c’s one row at a time.
Four.I.1.18 There are no real numbers θ that make the matrix singular because the
determinant of the matrix cos2 θ + sin2 θ is never 0, it equals 1 for all θ. Geometrically,
with respect to the standard basis, this matrix represents a rotation of the plane
through an angle of θ. Each such map is one-to-one — for one thing, it is invertible.
Answers to Exercises 265
Four.I.1.19 This is how the answer was given in the cited source. Let P be the sum
of the three positive terms of the determinant and −N the sum of the three negative
terms. The maximum value of P is
9 · 8 · 7 + 6 · 5 · 4 + 3 · 2 · 1 = 630.
9 · 6 · 1 + 8 · 5 · 2 + 7 · 4 · 3 = 218.
Any change in P would result in lowering that sum by more than 4. Therefore 412
the maximum value for the determinant and one form for the determinant is
9 4 2
3 8 6 .
5 1 7
Four.I.2.23 A matrix with only rational entries reduces with Gauss’s Method to an
echelon form matrix using only rational arithmetic. Thus the entries on the diagonal
must be rationals, and so the product down the diagonal is rational.
Four.I.2.24 This is how the answer was given in the cited source. The value (1−a4 )3
of the determinant is independent of the values B, C, D. Hence operation (e) does
not change the value of the determinant but merely changes its appearance. Thus
the element of likeness in (a), (b), (c), (d), and (e) is only that the appearance
of the principle entity is changed. The same element appears in (f) changing the
name-label of a rose, (g) writing a decimal integer in the scale of 12, (h) gilding the
lily, (i) whitewashing a politician, and (j) granting an honorary degree.
= −5
Four.I.3.19 Following
Example 3.6 gives this.
t
1,1 t1,2 t1,3
t2,1 t2,2 t2,3 = t1,1 t2,2 t3,3 |Pφ1 | + t1,1 t2,3 t3,2 |Pφ2 |
t3,1 t3,2 t3,3 + t1,2 t2,1 t3,3 |Pφ3 | + t1,2 t2,3 t3,1 |Pφ4 |
+ t1,3 t2,1 t3,2 |Pφ5 | + t1,3 t2,2 t3,1 |Pφ6 |
= t1,1 t2,2 t3,3 (+1) + t1,1 t2,3 t3,2 (−1)
+ t1,2 t2,1 t3,3 (−1) + t1,2 t2,3 t3,1 (+1)
+ t1,3 t2,1 t3,2 (+1) + t1,3 t2,2 t3,1 (−1)
Four.I.3.20 This is all of the permutations where φ(1) = 1
φ1 = h1, 2, 3, 4i φ2 = h1, 2, 4, 3i φ3 = h1, 3, 2, 4i
φ4 = h1, 3, 4, 2i φ5 = h1, 4, 2, 3i φ6 = h1, 4, 3, 2i
the ones where φ(1) = 1
φ7 = h2, 1, 3, 4i φ8 = h2, 1, 4, 3i φ9 = h2, 3, 1, 4i
φ10 = h2, 3, 4, 1i φ11 = h2, 4, 1, 3i φ12 = h2, 4, 3, 1i
the ones where φ(1) = 3
φ13 = h3, 1, 2, 4i φ14 = h3, 1, 4, 2i φ15 = h3, 2, 1, 4i
φ16 = h3, 2, 4, 1i φ17 = h3, 4, 1, 2i φ18 = h3, 4, 2, 1i
and the ones where φ(1) = 4.
φ19 = h4, 1, 2, 3i φ20 = h4, 1, 3, 2i φ21 = h4, 2, 1, 3i
φ22 = h4, 2, 3, 1i φ23 = h4, 3, 1, 2i φ24 = h4, 3, 2, 1i
Four.I.3.21 Each of these is easy to check.
(a) permutation φ1 φ2
inverse φ1 φ2
(b) permutation φ1 φ2 φ3 φ4 φ5 φ6
inverse φ1 φ2 φ3 φ5 φ4 φ6
Four.I.3.22 For the ‘if’ half, the first condition of Definition 3.2 follows from taking
k1 = k2 = 1 and the second condition follows from taking k2 = 0.
The ‘only if’ half also routine. From f(~ρ1 , . . . , k1~v1 + k2~v2 , . . . , ~ρn ) the first
condition of Definition 3.2 gives = f(~ρ1 , . . . , k1~v1 , . . . , ~ρn ) + f(~ρ1 , . . . , k2~v2 , . . . , ~ρn )
and the second condition, applied twice, gives the result.
Answers to Exercises 271
in two of the columns of the derived determinant are proportional, so the determinant
vanishes. That is,
2 1 x − 4 1 x − 3 −1 x − 2 −1 −2
4 2 x − 3 = 2 x − 1 −2 = x + 1 −2 −4 = 0.
6 3 x − 10 3 x − 7 −3 x − 4 −3 −6
Four.I.3.37 This is how the answer was given in the cited source. Let
a b c
d e f
g h i
have magic sum N = S/3. Then
N = (a + e + i) + (d + e + f) + (g + e + c)
− (a + d + g) − (c + f + i) = 3e
and S = 9e. Hence,
adding rows
and columns,
a b c a b c a b 3e a b e
D = d e f = d e f = d e 3e = d e e S.
g h i 3e 3e 3e 3e 3e 9e 1 1 1
Four.I.3.38 This is how the answer was given in the cited source. Denote by Dn
the determinant in question and by ai,j the element in the i-th row and j-th column.
Then from the law of formation of the elements we have
ai,j = ai,j−1 + ai−1,j , a1,j = ai,1 = 1.
Subtract each row of Dn from the row following it, beginning the process with the
last pair of rows. After the n − 1 subtractions the above equality shows that the
element ai,j is replaced by the element ai,j−1 , and all the elements in the first column,
except a1,1 = 1, become zeroes. Now subtract each column from the one following
it, beginning with the last pair. After this process the element ai,j−1 is replaced by
ai−1,j−1 , as shown in the above relation. The result of the two operations is to replace
ai,j by ai−1,j−1 , and to reduce each element in the first row and in the first column
to zero. Hence Dn = Dn+i and consequently
Dn = Dn−1 = Dn−2 = · · · = D2 = 1.
As with the 3×3 expansions described in the subsection, the permutation matrices
from corresponding terms are transposes (although this is disguised by the fact that
each is self-transpose).
Four.I.4.12 Each of these is easy to check.
(a) permutation φ1 φ2
inverse φ1 φ2
(b) permutation φ1 φ2 φ3 φ4 φ5 φ6
inverse φ1 φ2 φ3 φ5 φ4 φ6
Four.I.4.13 (a) sgn(φ1 ) = +1, sgn(φ2 ) = −1
(b) sgn(φ1 ) = +1, sgn(φ2 ) = −1, sgn(φ3 ) = −1, sgn(φ4 ) = +1, sgn(φ5 ) = +1,
sgn(φ6 ) = −1
Four.I.4.14 To get a nonzero term in the permutation expansion we must use the 1, 2
entry and the 4, 3 entry. Having fixed on those two we must also use the 2, 1 entry
and the 3, 4 entry. The signum of h2, 1, 4, 3i is +1 because from
0 1 0 0
1 0 0 0
0 0 0 1
0 0 1 0
the two row swaps ρ1 ↔ ρ2 and ρ3 ↔ ρ4 will produce the identity matrix.
Four.I.4.15 The pattern is this.
i 1 2 3 4 5 6 ...
sgn(φi ) +1 −1 −1 +1 +1 −1 ...
So to find the signum of φn! , we subtract one n! − 1 and look at the remainder on
division by four. If the remainder is 1 or 2 then the signum is −1, otherwise it is +1.
For n > 4, the number n! is divisible by four, so n! − 1 leaves a remainder of −1 on
division by four (more properly said, a remainder or 3), and so the signum is +1. The
n = 1 case has a signum of +1, the n = 2 case has a signum of −1 and the n = 3 case
has a signum of −1.
Four.I.4.16 (a) We can view permutations as maps φ : {1, . . . , n} → { 1, . . . , n} that are
one-to-one and onto. Any one-one and onto map has an inverse.
(b) If it always takes an odd number of swaps to get from Pφ to the identity, then it
always takes an odd number of swaps to get from the identity to Pφ (any swap is
reversible).
(c) This is the first question again.
Answers to Exercises 275
Four.I.4.17 If φ(i) = j then φ−1 (j) = i. The result now follows on the observation that
Pφ has a 1 in entry i, j if and only if φ(i) = j, and Pφ−1 has a 1 in entry j, i if and
only if φ−1 (j) = i,
Four.I.4.18 This does not say that m is the least number of swaps to produce an identity,
nor does it say that m is the most. It instead says that there is a way to swap to the
identity in exactly m steps.
Let ιj be the first row that is inverted with respect to a prior row and let ιk be
the first row giving that inversion. We have this interval of rows.
.
.
.
ιk
ι
r1
.
. j < k < r1 < · · · < rs
.
ιrs
ιj
..
.
Swap.
.
.
.
ιj
ι
r1
.
.
.
ιrs
ιk
..
.
The second matrix has one fewer inversion because there is one fewer inversion in
the interval (s vs. s + 1) and inversions involving rows outside the interval are not
affected.
Proceed in this way, at each step reducing the number of inversions by one with
each row swap. When no inversions remain the result is the identity.
The contrast with Corollary 4.5 is that the statement of this exercise is a ‘there
exists’ statement: there exists a way to swap to the identity in exactly m steps. But
the corollary is a ‘for all’ statement: for all ways to swap to the identity, the parity
(evenness or oddness) is the same.
Four.I.4.19 (a) First, g(φ1 ) is the product of the single factor 2 − 1 and so g(φ1 ) = 1.
Second, g(φ2 ) is the product of the single factor 1 − 2 and so g(φ2 ) = −1.
(b) permutation φ φ1 φ2 φ3 φ4 φ5 φ6
g(φ) 2 −2 −2 2 2 −2
276 Linear Algebra, by Hefferon
Geometry of Determinants
Four.II.1.9 Solving
3 2 1 4
c1 3 + c2 6 + c3 0 = 1
1 1 5 2
gives the unique solution c3 = 11/57, c2 = −40/57 and c1 = 99/57. Because c1 > 1,
the vector is not in the box.
Four.II.1.10 Move the parallelepiped to start at the origin, so that it becomes the box
formed by
! !
3 2
h , i
0 1
and now the absolute value of this determinant is easily computed as 3.
3 2
=3
0 1
Four.II.1.12 Express each transformation with respect to the standard bases and find
the determinant.
Answers to Exercises 277
we get a system
x + z=0 −2ρ1 +ρ2 x +z=0
−→
2x − 3y + 3z = 0 −3y + z = 0
with this solution set.
−1
{ 1/3 z | z ∈ R},
1
A solution of length one is this.
−1
1
1/3
p
19/9
1
Thus the area of the triangle is the absolute value of this determinant.
1 2 √
−3/ 19 √
√
0 −3 1/ 19 = −12/ 19
√
1 3 3/ 19
Four.II.1.23 (a) Because the image of a linearly dependent set is linearly dependent,
if the vectors forming S make a linearly dependent set, so that |S| = 0, then the
vectors forming t(S) make a linearly dependent set, so that |T S| = 0, and in this
case the equation holds.
kρi +ρj
(b) We must check that if T −→ T̂ then d(T ) = |T S|/|S| = |T̂ S|/|S| = d(T̂ ). We
can do this by checking that combining rows first and then multiplying to get T̂ S
gives the same result as multiplying first to get T S and then combining (because
the determinant |T S| is unaffected by the combining rows so we’ll then have that
|T̂ S| = |T S| and hence that d(T̂ ) = d(T )). This check runs: after adding k times
row i of T S to row j of T S, the j, p entry is (kti,1 + tj,1 )s1,p + · · · + (kti,r + tj,r )sr,p ,
which is the j, p entry of T̂ S.
ρi ↔ρj
(c) For the second property, we need only check that swapping T −→ T̂ and then
multiplying to get T̂ S gives the same result as multiplying T by S first and then
swapping (because, as the determinant |T S| changes sign on the row swap, we’ll
then have |T̂ S| = −|T S|, and so d(T̂ ) = −d(T )). This check runs just like the one
for the first property.
kρi
For the third property, we need only show that performing T −→ T̂ and then
computing T̂ S gives the same result as first computing T S and then performing the
scalar multiplication (as the determinant |T S| is rescaled by k, we’ll have |T̂ S| = k|T S|
and so d(T̂ ) = k d(T )). Here too, the argument runs just as above.
The fourth property, that if T is I then the result is 1, is obvious.
(d) Determinant functions are unique, so |T S|/|S| = d(T ) = |T |, and so |T S| = |T ||S|.
Answers to Exercises 279
Four.II.1.24 Any permutation matrix has the property that the transpose of the matrix
is its inverse.
For the implication, we know that |AT | = |A|. Then 1 = |A · A−1 | = |A · AT | =
|A| · |AT | = |A|2 .
The converse does not hold; here is an example.
!
3 1
2 1
Four.II.1.25 Where the sides of the box are c times longer, the box has c3 times as
many cubic units of volume.
Four.II.1.26 If H = P−1 GP then |H| = |P−1 ||G||P| = |P−1 ||P||G| = |P−1 P||G| = |G|.
e3
~
e2
~ ~
β
e1
~ ~
β 2
1
~
β 3
Four.II.1.28 We will compare det(~s1 , . . . ,~sn ) with det(t(~s1 ), . . . , t(~sn )) to show that the
second differs from the first by a factor of |T |. We represent the ~s ’s with respect to
280 Linear Algebra, by Hefferon
Swap the columns in det(~tφ(1) , . . . , ~tφ(n) ) to get the matrix T back, which changes
the sign by a factor of sgn φ, and then factor out the determinant of T .
X X
= sφ(1),1 . . . sφ(n),n det(~t1 , . . . , ~tn ) · sgn φ = det(T ) sφ(1),1 . . . sφ(n),n · sgn φ.
φ φ
Answers to Exercises 281
As in the proof that the determinant of a matrix equals the determinant of its transpose,
we commute the s’s to list them by ascending row number instead of by ascending
column number (and we substitute sgn(φ−1 ) for sgn(φ)).
X
= det(T ) s1,φ−1 (1) . . . sn,φ−1 (n) · sgn φ−1 = det(T ) det(~s1 ,~s2 , . . . ,~sn )
φ
Four.II.1.29 (a) An algebraic check is easy.
0 = xy2 + x2 y3 + x3 y − x3 y2 − xy3 − x2 y = x · (y2 − y3 ) + y · (x3 − x2 ) + x2 y3 − x3 y2
simplifies to the familiar form
y = x · (x3 − x2 )/(y3 − y2 ) + (x2 y3 − x3 y2 )/(y3 − y2 )
(the y3 − y2 = 0 case is easily handled).
For geometric insight, this picture shows that the box formed by the three
vectors. Note that all three vectors end in the z = 1 plane. Below the two vectors
on the right is the line through (x2 , y2 ) and (x3 , y3 ).
x2
y
2
x 1
y
1 x3
y
3
1
The box will have a nonzero volume unless the triangle formed by the ends of the
three is degenerate. That only happens (assuming that (x2 , y3 ) 6= (x3 , y3 )) if (x, y)
lies on the line through the other two.
(b) This is how the answer was given in the cited source. We find the altitude
through (x1 , y1 ) of a triangle with vertices (x1 , y1 ) (x2 , y2 ) and (x3 , y3 ) in the usual
way from the normal form of the above:
x x2 x3
1
1
y1 y2 y3 .
p
(x2 − x3 )2 + (y2 − y3 )2
1 1 1
Another step shows the area of the triangle to be
x x2 x3
1
1
y1 y2 y3 .
2
1 1 1
This exposition reveals the modus operandi more clearly than the usual proof of
showing a collection of terms to be identical with the determinant.
(c) This is how the answer was given in the cited source. Let
x
1 x2 x3
D = y1 y2 y3
1 1 1
282 Linear Algebra, by Hefferon
then the area of the triangle is (1/2)|D|. Now if the coordinates are all integers,
then D is an integer.
Laplace’s Formula
expanded on the first row gives a · (+1)|d| + b · (−1)|c| = ad − bc (note the two 1×1
minors).
Four.III.1.18 The determinant of
a b c
d e f
g h i
is this.
e f d f d e
a· −b· +c· = a(ei − fh) − b(di − fg) + c(dh − eg)
h i g i g h
! !
t2,2 − t1,2
T1,1 T2,1 t 2,2 −t 1,2
Four.III.1.19 (a) = =
T1,2 T2,2 − t2,1 t1,1
−t2,1 t1,1
!
t2,2 −t1,2
(b) (1/t1,1 t2,2 − t1,2 t2,1 ) ·
−t2,1 t1,1
Four.III.1.20 No. Here is a determinant
whose value
1 0 0
0 1 0 = 1
0 0 1
doesn’t equal the result of expanding down the diagonal.
1 0 1 0 1 0
1 · (+1) + 1 · (+1) + 1 · (+1) =3
0 1 0 1 0 1
Four.III.1.21 Consider this diagonal matrix.
d1 0 0 ...
0 d 0
2
0 0 d
D= 3
..
.
dn
If i 6= j then the i, j minor is an (n − 1)×(n − 1) matrix with only n − 2 nonzero
entries, because we have deleted both di and dj . Thus, at least one row or column
of the minor is all zeroes, and so the cofactor Di,j is zero. If i = j then the minor
Answers to Exercises 285
By the way, Theorem 1.9 provides a slicker way to derive this conclusion.
Four.III.1.22 Just note that if S = T T then the cofactor Sj,i equals the cofactor Ti,j
because (−1)j+i = (−1)i+j and because the minors are the transposes of each other
(and the determinant of a transpose equals the determinant of the matrix).
(b) x = 2, y = 2
2 z=1
3 Determinants are unchanged by combinations, including column combinations,
so det(Bi ) = det(~ ~ 1 + · · · + xi a
a1 , . . . , x1 a ~ i + · · · + xn a ~ n, . . . , a
~ n ). Use the op-
eration of taking −x1 times the first column and adding it to the i-th column,
etc., to see this is equal to det(~ ~ i, . . . , a
a1 , . . . , x i a ~ n ). In turn, that is equal to
xi · det(~ ~ n ) = xi · det(A), as required.
~ i, . . . , a
a1 , . . . , a
4 (a) Here is the case of a 2×2 system with i = 2.
! ! !
a1,1 x1 + a1,2 x2 = b1 a1,1 a1,2 1 x1 a1,1 b1
⇐⇒ =
a2,1 x1 + a2,2 x2 = b2 a2,1 a2,2 0 x2 a2,1 b2
(b) The determinant function is multiplicative det(Bi ) = det(AXi ) =
det(A) · det(Xi ). The Laplace expansion shows that det(Xi ) = xi , and solving for
xi gives Cramer’s Rule.
5 Because the determinant of A is nonzero, Cramer’s Rule applies and shows that
xi = |Bi |/1. Since Bi is a matrix of integers, its determinant is an integer.
6 The solution of
ax +by = e
cx +dy = f
is
ed − fb af − ec
x= y=
ad − bc ad − bc
provided of course that the denominators are not zero.
7 Of course, singular systems have |A| equal to zero, but we can characterize the
infinitely many solutions case is by the fact that all of the |Bi | are zero as well.
8 We can consider the two nonsingular cases together with this system
x1 + 2x2 = 6
x1 + 2x2 = c
where c = 6 of course yields infinitely many solutions, and any other value for c yields
no solutions. The corresponding vector equation
! ! !
1 2 6
x1 · + x2 · =
1 2 c
gives a picture of two overlapping vectors. Both lie on the line y = x. In the c = 6
case the vector on the right side also lies on the line y = x but in any other case it
does not.
Answers to Exercises 289
1 (a) Under Octave, rank(rand(5)) finds the rank of a 5×5 matrix whose entries are
(uniformly distributed) in the interval [0..1). This loop which runs the test 5000
times
octave:1> for i=1:5000
> if rank(rand(5))<5 printf("That's one."); endif
> endfor
0.2
0.15
0.1
0.05
0
20 40 60 80 100
2 The number of operations depends on exactly how we do the operations.
(a) The determinant is −11. To row reduce takes a single row combination with two
multiplications (−5/2 times 2 plus 5, and −5/2 times 1 plus −3) and the product
down the diagonal takes one more multiplication. The permutation expansion takes
two multiplications (2 times −3 and 5 times 1).
(b) The determinant is −39. Counting the operations is routine.
(c) The determinant is 4.
3 One way to get started is to compare these under Octave: det(rand(10));, versus
det(hilb(10));, versus det(eye(10));, versus det(zeroes(10));. You can time
them as in tic(); det(rand(10)); toc().
4 Yes, because the J is in the innermost loop.
where p + 1 6= i and q + 1 6= j.
3 Sarrus’s formula uses 12 multiplications and 5 additions (including the subtractions
in with the additions). Chiò’s formula uses two multiplications and an addition (which
is actually a subtraction) for each of the four 2×2 determinants, and another two
multiplications and an addition for the 2×2 Chió’s determinant, as well as a final
division by a1,1 . That’s eleven multiplication/divisions and five addition/subtractions.
So Chiò is the winner.
4 Consider an n×n matrix.
a1,1 a1,2 ··· a1,n−1 a1,n
a
2,1 a2,2 ··· a2,n−1 a2,n
..
A= .
· · · an−1,n−1
an−1,1 an−1,2 an−1,n
an,1 an,2 ··· an,n−1 an,n
Rescale every row but the first by a1,1 .
a1,1 a1,2 ··· a1,n−1 a1,n
a a
2,1 1,1 a 2,2 a1,1 ··· a2,n−1 a1,1 a2,n a1,1
a1,1 ρ2 ..
−→
a1,1 ρ3 .
..
an−1,1 a1,1 an−1,2 a1,1 · · · an−1,n−1 a1,1
an−1,n a1,1
.
a1,1 ρn an,1 a1,1 an,2 a1,1 ··· an,n−1 a1,1 an,n a1,1
That rescales the determinant by a factor of an−1
1,1 .
Next perform the row operation −ai,1 ρ1 + ρi on each row i > 1. These row
operations don’t change the determinant.
−a2,1 ρ1 +ρ2
−→
−a3,1 ρ1 +ρ3
..
.
−an,1 ρ1 +ρn
The result is a matrix whose first row is unchanged, whose first column is all zero’s
(except for the 1, 1 entry of a1,1 ), and whose remaining entries are these.
a1,2 ··· a1,n−1 a1,n
a2,2 a1,1 − a2,1 a1,2 · · · a2,n−1 an,n − a2,n−1 a1,n−1 a2,n an,n − a2,n a1,n
..
.
an,2 a1,1 − an,1 a1,2 · · · an,n−1 a1,1 − an,1 a1,n−1 an,n a1,1 − an,1 a1,n
The determinant of this matrix is an−11,1 times the determinant of A.
Denote by C the 1, 1 minor of the matrix, that is, the submatrix consisting of the
first n − 1 rows and columns. The Laplace expansion down the final column of the
above matrix gives that its determinant is (−1)1+1 a1,1 det(C).
If a1,1 6= 0 then setting the two equal and canceling gives det(A) = det(C)/an−2
n,n .
292 Linear Algebra, by Hefferon
The solar eclipse picture also shows the converse. If we picture the projection as
going from left to right through the pinhole then the ellipse I projects through P to a
circle S.
6 A spot on the unit sphere
p1
p 2
p3
6 0. In that case it corresponds to this point on
is non-equatorial if and only if p3 =
the z = 1 plane
p1 /p3
p2 /p3
1
since that is intersection of the line containing the vector and the plane.
7 (a) Other pictures are possible, but this is one.
T0
U0
V0
V2 T2
U2
V1
U1
T1
u0 ∈ R). Similarly, U2 is on T1 V0
0 1 d
RepB (~u2 ) = c 1 + d 1 = c + d
0 1 d
and so has this homogeneous coordinate vector.
1
u 2
1
Also similarly, U1 is incident on T2 V0
0 1 f
RepB (~u1 ) = e 0 + f 1 = f
1 1 e+f
and has this homogeneous coordinate vector.
1
1
u1
(d) Because V1 is T0 U2 ∩ U0 T2 we have this.
1 1 u0 0 g + h = iu0
g 0 + h u2 = i 1 + j 0 =⇒ hu2 = i
0 1 1 1 h=i+j
Substituting hu2 for i in the first equation
hu0 u2
hu2
h
shows that V1 has this two-parameter homogeneous coordinate vector.
u0 u2
u2
1
(e) Since V2 is the intersection T0 U1 ∩ T1 U0
1 1 0 u0 k + l = nu0
k 0 + l 1 = m 1 + n 1 =⇒ l=m+n
0 u1 0 1 lu1 = n
and substituting lu1 for n in the first equation
lu0 u1
l
lu1
Answers to Exercises 295
Similarity
Five.II.1.7 (a) Because we describe t with the members of B, finding the matrix
representation is easy:
0 1 0
2
RepB (t(x )) = 1 RepB (t(x)) = 0 RepB (t(1)) = 0
1 −1 3
B B B
gives this.
0 1 0
RepB,B (t) 1 0 0
1 −1 3
(b) We will find t(1), t(1 + x), and t(1 + x + x2 , to find how each is represented
with respect to D. We are given that t(1) = 3, and the other two are easy to see:
t(1 + x) = x2 + 2 and t(1 + x + x2 ) = x2 + x + 3. By eye, we get the representation
of each vector
3 2 2
RepD (t(1)) = 0 RepD (t(1 + x)) = −1 RepD (t(1 + x + x2 )) = 0
0 1 1
D D D
and thus the representation of the map.
3 2 2
RepD,D (t) = 0 −1 0
0 1 1
(c) The diagram
t
Vwrt B −−−−→ Vwrt B
T
idyP idyP
t
Vwrt D −−−−→ Vwrt D
T̂
shows that these are P = RepB,D (id) and P−1 = RepD,B (id).
0 −1 1 0 0 1
P = −1 1 0 P−1 = 0 1 1
1 0 0 1 1 1
Five.II.1.8 One possible choice of the bases is
! ! ! !
1 −1 1 0
B=h , i D = E2 = h , i
2 1 0 1
(this B comes from the map description). To find the matrix T̂ = RepB,B (t), solve
the relations ! ! ! ! ! !
1 −1 3 1 −1 −1
c1 + c2 = ĉ1 + ĉ2 =
2 1 0 2 1 2
Answers to Exercises 299
f
R2wrt E2 −−−x−→ R2wrt E2
S
Answers to Exercises 301
Now the conclusion follows from the transitivity part of Exercise 12.
We can also finish without relying on that exercise. Write RepB,B (fx ) = QT Q−1 =
QRepE2 ,E2 (fx )Q−1 and RepD,D (fy ) = RSR−1 = RRepE2 ,E2 (fy )R−1 . By the equation
in the first paragraph, the first of these two is RepB,B (fx ) = QPRepE2 ,E2 (fy )P−1 Q−1 .
Rewriting the second of these two as R−1 · RepD,D (fy ) · R = RepE2 ,E2 (fy ) and
substituting gives the desired relationship
Five.II.1.17 The k-th powers are similar because, where each matrix represents the map
t, the k-th powers represent tk , the composition of k-many t’s. (For instance, if
T = reptB, B then T 2 = RepB,B (t ◦ t).)
Restated more computationally, if T = PSP−1 then T 2 = (PSP−1 )(PSP−1 ) =
PS2 P−1 . Induction extends that to all powers.
For the k 6 0 case, suppose that S is invertible and that T = PSP−1 . Note that T
is invertible: T −1 = (PSP−1 )−1 = PS−1 P−1 , and that same equation shows that T −1
is similar to S−1 . Other negative powers are now given by the first paragraph.
Five.II.1.18 In conceptual terms, both represent p(t) for some transformation t. In
computational terms, we have this.
p(T ) = cn (PSP−1 )n + · · · + c1 (PSP−1 ) + c0 I
= cn PSn P−1 + · · · + c1 PSP−1 + c0 I
= Pcn Sn P−1 + · · · + Pc1 SP−1 + Pc0 P−1
= P(cn Sn + · · · + c1 S + c0 )P−1
Five.II.1.19 There are two equivalence classes, (i) the class of rank zero matrices, of
which there is one: C1 = { (0) }, and (2) the class of rank one matrices, of which there
are infinitely many: C2 = {(k) | k 6= 0 }.
Each 1×1 matrix is alone in its similarity class. That’s because any transformation
of a one-dimensional space is multiplication by a scalar tk : V → V given by ~v 7→ k · ~v.
~ the matrix representing a transformation tk with respect
Thus, for any basis B = hβi,
~
to B, B is (RepB (tk (β))) = (k).
So, contained in the matrix equivalence class C1 is (obviously) the single similarity
class consisting of the matrix (0). And, contained in the matrix equivalence class
C2 are the infinitely many, one-member-each, similarity classes consisting of (k) for
k 6= 0.
Five.II.1.20 No. Here is an example that has two pairs, each of two similar matrices:
! ! ! !
1 −1 1 0 2/3 1/3 5/3 −2/3
=
1 2 0 3 −1/3 1/3 −4/3 7/3
and ! ! ! !
1 −2 −1 0 −1 −2 −5 −4
=
−1 1 0 −3 −1 −1 2 1
(this example is not entirely arbitrary because the center matrices on the two left
sides add to the zero matrix). Note that the sums of these similar matrices are not
similar ! ! ! ! ! !
1 0 −1 0 0 0 5/3 −2/3 −5 −4 0 0
+ = + 6=
0 3 0 −3 0 0 −4/3 7/3 2 1 0 0
since the zero matrix is similar only to itself.
Answers to Exercises 303
Five.II.2: Diagonalizability
Five.II.2.6 Because we chose the basis vectors arbitrarily, many different answers are
possible. However, here is one way to go; to diagonalize
!
4 −2
T=
1 1
take it as the representation of a transformation with respect to the standard basis
T = RepE2 ,E2 (t) and look for B = hβ ~ 1, β
~ 2 i such that
!
λ1 0
RepB,B (t) =
0 λ2
~ ~
that is, such that t(β1 ) =!λ1 and t(β2 ) = λ2 . !
4 −2 ~ ~ 4 −2 ~ ~2
β 1 = λ1 · β 1 β 2 = λ2 · β
1 1 1 1
We are looking for scalars x such that ! this ! equation !
4 −2 b1 b1
=x·
1 1 b2 b2
has solutions b1 and b2 , which are not both zero. Rewrite that as a linear system
(4 − x) · b1 + −2 · b2 = 0
1 · b1 + (1 − x) · b2 = 0
If x = 4 then the first equation gives that b2 = 0, and then the second equation gives
that b1 = 0. We have disallowed the case where both b’s are zero so we can assume
that x 6= 4.
(−1/(4−x))ρ1 +ρ2 (4 − x) · b1 + −2 · b2 = 0
−→
((x2 − 5x + 6)/(4 − x)) · b2 = 0
Consider the bottom equation. If b2 = 0 then the first equation gives b1 = 0 or x = 4.
The b1 = b2 = 0 case is not allowed. The other possibility for the bottom equation is
that the numerator of the fraction x2 − 5x + 6 = (x − 2)(x − 3) is zero. The x = 2 case
gives a first equation of 2b1 − 2b2 = 0, and so associated with x = 2 we have vectors
whose first and second!components are equal: ! ! !
~1 = 1 4 −2 1 1
β (so =2· , and λ1 = 2).
1 1 1 1 1
304 Linear Algebra, by Hefferon
If x = 3 then the first equation is b1 − 2b2 = 0 and so the associated vectors are those
whose first component is twice their second:
! ! ! !
~ 2 4 −2 2 2
β2 = (so =3· , and so λ2 = 3).
1 1 1 1 1
This picture
t
R2wrt E2 −−−−→ R2wrt E2
T
idy idy
t
R2wrt B −−−−→ R2wrt B
D
shows how to get the diagonalization.
! !−1 ! !
2 0 1 2 4 −2 1 2
=
0 3 1 1 1 1 1 1
Comment. This equation matches the T = PSP−1 definition under this renaming.
! !−1 ! !
2 0 1 2 1 2 4 −2
T= P= P−1 = S=
0 3 1 1 1 1 1 1
Five.II.2.7 (a) Setting up
! ! !
−2 1 b1 b1 (−2 − x) · b1 + b2 = 0
=x· =⇒
0 2 b2 b2 (2 − x) · b2 = 0
gives the two possibilities that b2 = 0 and x = 2. Following the b2 = 0 possibility
leads to the first equation (−2 − x)b1 = 0 with the two cases that b1 = 0 and that
x = −2. Thus, under this first possibility, we find x = −2 and the associated vectors
whose second component is zero, and whose first component is free.
! ! ! !
−2 1 b1 b1 ~1 = 1
= −2 · β
0 2 0 0 0
Following the other possibility leads to a first equation of −4b1 + b2 = 0 and so the
vectors associated with this solution have a second component that is four times
their first component.
! ! ! !
−2 1 b1 b1 ~2 = 1
=2· β
0 2 4b1 4b1 4
The diagonalization is this.
! ! !−1 !
1 1 −2 1 1 1 −2 0
=
0 4 0 2 0 4 0 2
(b) The calculations are like those in the prior part.
! ! !
5 4 b1 b1 (5 − x) · b1 + 4 · b2 = 0
=x· =⇒
0 1 b2 b2 (1 − x) · b2 = 0
Answers to Exercises 305
The bottom equation gives the two possibilities that b2 = 0 and x = 1. Following
the b2 = 0 possibility, and discarding the case where both b2 and b1 are zero, gives
that x = 5, associated with vectors whose second component is zero and whose first
component is free. !
~1 = 1
β
0
The x = 1 possibility gives a first equation of 4b1 + 4b2 = 0 and so the associated
vectors have a second component that is the negative
! of their first component.
~1 = 1
β
−1
We thus have this diagonalization.
! ! !−1 !
1 1 5 4 1 1 5 0
=
0 −1 0 1 0 −1 0 1
Five.II.2.8 For any integer p, we have this.
p p
d1 0 d1 0
0 ... = 0 ...
dn dp
n
Five.II.2.9 These two are not similar ! !
0 0 1 0
0 0 0 1
because each is alone in its similarity class.
For the second half, these ! !
2 0 3 0
0 3 0 2
are similar via the matrix that changes bases from hβ ~ 1, β
~ 2 i to hβ
~ 2, β
~ 1 i. (Question.
Are two diagonal matrices similar if and only if their diagonal entries are permutations
of each others?)
Five.II.2.10 Contrast these two. ! !
2 0 2 0
0 1 0 0
The first is nonsingular, the second is singular.
Five.II.2.11 To check that the inverse of a diagonal matrix is the diagonal matrix of the
inverses, justmultiply.
a1,1 0 1/a1,1 0
0 a2,2 0 1/a2,2
.. ..
. .
an,n 1/an,n
306 Linear Algebra, by Hefferon
Five.II.2.17 If ! !
1 c a 0
P P−1 =
0 1 0 b
then ! !
1 c a 0
P = P
0 1 0 b
so ! ! ! !
p q 1 c a 0 p q
=
r s 0 1 0 b r s
! !
p cp + q ap aq
=
r cr + s br bs
The 1, 1 entries show that a = 1 and the 1, 2 entries then show that pc = 0. Since
c 6= 0 this means that p = 0. The 2, 1 entries show that b = 1 and the 2, 2 entries
then show that rc = 0. Since c 6= 0 this means that r = 0. But if both p and r are 0
then P is not invertible.
Five.II.2.18 (a) Using the formula for the inverse of a 2×2 matrix gives this.
! ! !
a b 1 2 1 d −b
· ·
c d 2 1 ad − bc −c a
!
1 ad + 2bd − 2ac − bc −ab − 2b2 + 2a2 + ab
=
ad − bc cd + 2d2 − 2c2 − cd −bc − 2bd + 2ac + ad
Now pick scalars a, . . . , d so that ad − bc 6= 0 and 2d2 − 2c2 = 0 and 2a2 − 2b2 = 0.
For example, these will do.
! ! ! !
1 1 1 2 1 −1 −1 1 −6 0
· · =
1 −1 2 1 −2 −1 1 −2 0 2
(b) As above,
! ! !
a b x y 1 d −b
· ·
c d y z ad − bc −c a
!
1 adx + bdy − acy − bcz −abx − b2 y + a2 y + abz
=
ad − bc cdx + d2 y − c2 y − cdz −bcx − bdy + acy + adz
we are looking for scalars a, . . . , d so that ad−bc 6= 0 and −abx−b2 y+a2 y+abz = 0
and cdx + d2 y − c2 y − cdz = 0, no matter what values x, y, and z have.
For starters, we assume that y 6= 0, else the given matrix is already diagonal.
We shall use that assumption because if we (arbitrarily) let a = 1 then we get
−bx − b2 y + y + bz = 0
(−y)b2 + (z − x)b + y = 0
308 Linear Algebra, by Hefferon
The eigenspace is the set of vectors whose second component is twice the first
component.
! ! ! !
b2 /2 3 0 b2 /2 b2 /2
{ | b2 ∈ C } =3·
b2 8 −1 b2 b2
(Here, the parameter is b2 only because that is the variable that is free in the above
system.) Hence, this is an eigenvector associated with the eigenvalue 3.
!
1
2
Finding an eigenvector associated with λ2 = −1 is similar. This system
4 · b1 + 0 · b2 = 0
8 · b1 + 0 · b2 = 0
leads to the set of vectors whose first component is zero.
! ! ! !
0 3 0 0 0
{ | b2 ∈ C } = −1 ·
b2 8 −1 b2 b2
And so this is an eigenvector associated with λ2 .
!
0
1
(b) The characteristic equation is
3 − x 2
0= = x2 − 3x + 2 = (x − 2)(x − 1)
−1 −x
and so the eigenvalues are λ1 = 2 and λ2 = 1. To find eigenvectors, consider this
system.
(3 − x) · b1 + 2 · b2 = 0
−1 · b1 − x · b2 = 0
For λ1 = 2 we get
1 · b1 + 2 · b2 = 0
−1 · b1 − 2 · b2 = 0
leading to this eigenspace and eigenvector.
! !
−2b2 −2
{ | b2 ∈ C }
b2 1
For λ2 = 1 the system is
2 · b1 + 2 · b2 = 0
−1 · b1 − 1 · b2 = 0
leading to this. ! !
−b2 −1
{ | b2 ∈ C }
b2 1
310 Linear Algebra, by Hefferon
Five.II.3.31 Any two representations of that transformation are similar, and similar
matrices have the same characteristic polynomial.
Five.II.3.32 It is not true. All of the eigenvalues
! of this matrix are 0.
0 1
0 0
Five.II.3.33 (a) Use λ = 1 and the identity map.
(b) Yes, use the transformation that multiplies all vectors by the scalar λ.
Five.II.3.34 If t(~v) = λ · ~v then ~v 7→ ~0 under the map t − λ · id.
Five.II.3.35 The characteristic equation
a − x b
0= = (a − x)(d − x) − bc
c d − x
simplifies to x2 + (−a − d) · x + (ad − bc). Checking that the values x = a + b and
x = a − c satisfy the equation (under the a + b = c + d condition) is routine.
Five.II.3.36 Consider an eigenspace Vλ . Any w ~ ∈ Vλ is the image w ~ = λ · ~v of some
~v ∈ Vλ (namely, ~v = (1/λ) · w ~ ). Thus, on Vλ (which is a nontrivial subspace) the
action of t−1 is t−1 (~ ~ , and so 1/λ is an eigenvalue of t−1 .
w) = ~v = (1/λ) · w
Five.II.3.37 (a) We have (cT + dI)~v = cT~v + dI~v = cλ~v + d~v = (cλ + d) · ~v.
(b) Suppose that S = PT P−1 is diagonal. Then P(cT + dI)P−1 = P(cT )P−1 +
P(dI)P−1 = cPT P−1 + dI = cS + dI is also diagonal.
Five.II.3.38 The scalar λ is an eigenvalue if and only if the transformation t − λ id is
singular. A transformation is singular if and only if it is not an isomorphism (that is,
a transformation is an isomorphism if and only if it is nonsingular).
Five.II.3.39 (a) Where the eigenvalue λ is associated with the eigenvector ~x then
Ak~x = A · · · A~x = Ak−1 λ~x = λAk−1~x = · · · = λk~x. (The full details require
induction on k.)
(b) The eigenvector associated with λ might not be an eigenvector associated with µ.
Five.II.3.40 No. These are two same-sized, equal rank, matrices with different eigenval-
ues. ! !
1 0 1 0
0 1 0 2
Five.II.3.41 The characteristic polynomial has an odd power and so has at least one real
root.
Five.II.3.42 The characteristic polynomial x3 + 5x2 + 6x has distinct roots λ1 = 0,
λ2 = −2, and λ3 = −3. Thus the matrix
can be
diagonalized into this form.
0 0 0
0 −2 0
0 0 −3
316 Linear Algebra, by Hefferon
Five.II.3.43 We must show that it is one-to-one and onto, and that it respects the
operations of matrix addition and scalar multiplication.
To show that it is one-to-one, suppose that tP (T ) = tP (S), that is, suppose that
PT P−1 = PSP−1 , and note that multiplying both sides on the left by P−1 and on the
right by P gives that T = S. To show that it is onto, consider S ∈ Mn×n and observe
that S = tP (P−1 SP).
The map tP preserves matrix addition since tP (T + S) = P(T + S)P−1 = (PT +
PS)P−1 = PT P−1 +PSP−1 = tP (T +S) follows from properties of matrix multiplication
and addition that we have seen. Scalar multiplication is similar: tP (cT ) = P(c·T )P−1 =
c · (PT P−1 ) = c · tP (T ).
Five.II.3.44 This is how the answer was given in the cited source. If the argument
of the characteristic function of A is set equal to c, adding the first (n − 1) rows
(columns) to the nth row (column) yields a determinant whose nth row (column) is
zero. Thus c is a characteristic root of A.
Nilpotence
Five.III.1: Self-Composition
Five.III.1.9 For the zero transformation, no matter what the space, the chain of range
spaces is V ⊃ {~0 } = {~0 } = · · · and the chain of null spaces is {~0 } ⊂ V = V = · · · . For
the identity transformation the chains are V = V = V = · · · and {~0 } = {~0 } = · · · .
Five.III.1.10 (a) Iterating t0 twice a + bx + cx2 7→ b + cx2 7→ cx2 gives
t2
a + bx + cx2 7−→
0
cx2
and any higher power is the same map. Thus, while R(t0 ) is the space of quadratic
polynomials with no linear term {p + rx2 | p, r ∈ C}, and R(t20 ) is the space of purely-
quadratic polynomials {rx2 | r ∈ C }, this is where the chain stabilizes R∞ (t0 ) =
{ rx2 | n ∈ C }. As for null spaces, N (t0 ) is the space of purely-linear quadratic
polynomials {qx | q ∈ C }, and N (t20 ) is the space of quadratic polynomials with
no x2 term { p + qx | p, q ∈ C}, and this is the end N∞ (t0 ) = N (t20 ).
Answers to Exercises 317
Five.III.1.13 It is both ‘if’ and ‘only if’. A linear map is nonsingular if and only if it
preserves dimension, that is, if the dimension of its range equals the dimension of its
domain. With a transformation t : V → V that means that the map is nonsingular if
and only if it is onto: R(t) = V (and thus R(t2 ) = V, etc).
Five.III.1.14 The null spaces form chains because because if ~v ∈ N (tj ) then tj (~v) = ~0
and tj+1 (~v) = t( tj (~v) ) = t(~0) = ~0 and so ~v ∈ N (tj+1 ).
Now, the “further” property for null spaces follows from that fact that it holds
for range spaces, along with the prior exercise. Because the dimension of R(tj ) plus
the dimension of N (tj ) equals the dimension n of the starting space V, when the
dimensions of the range spaces stop decreasing, so do the dimensions of the null spaces.
The prior exercise shows that from this point k on, the containments in the chain are
not proper — the null spaces are equal.
Five.III.1.15 (Many examples are correct but here is one.) An example is the shift
operator on triples of reals (x, y, z) 7→ (0, x, y). The null space is all triples that start
with two zeros. The map stabilizes after three iterations.
Five.III.1.16 The differentiation operator d/dx : P1 → P1 has the same range space as
null space. For an example of where they are disjoint — except for the zero vector —
consider an identity map, or any nonsingular map.
Five.III.2: Strings
Five.III.2.19 Three. It is at least three because `2 ( (1, 1, 1) ) = (0, 0, 1) 6= ~0. It is at most
three because (x, y, z) 7→ (0, x, y) 7→ (0, 0, x) 7→ (0, 0, 0).
Five.III.2.20 (a) The domain has dimension four. The map’s action is that any vector
in the space c1 · β~ 1 + c2 · β
~ 2 + c3 · β
~ 3 + c4 · β
~ 4 goes to c1 · β
~ 2 + c2 · ~0 + c3 · β
~ 4 + c4 · ~0 =
~ 3 + c3 · β
c1 · β ~ 4 . The first application of the map sends two basis vectors β ~ 2 and
~
β4 to zero, and therefore the null space has dimension two and the range space has
dimension two. With a second application, all four basis vectors go to zero and so
the null space of the second power has dimension four while the range space of the
second power has dimension zero. Thus the index of nilpotency is two. This is the
canonical form.
0 0 0 0
1 0 0 0
0 0 0 0
0 0 1 0
Answers to Exercises 319
(b) The dimension of the domain of this map is six. For the first power the dimension
of the null space is four and the dimension of the range space is two. For the second
power the dimension of the null space is five and the dimension of the range space
is one. Then the third iteration results in a null space of dimension six and a range
space of dimension zero. The index of nilpotency is three, and this is the canonical
form.
0 0 0 0 0 0
1 0 0 0 0 0
0 1 0 0 0 0
0 0 0 0 0 0
0 0 0 0 0 0
0 0 0 0 0 0
(c) The dimension of the domain is three, and the index of nilpotency is three. The
first power’s null space has dimension one and its range space has dimension two.
The second power’s null space has dimension two and its range space has dimension
one. Finally, the third power’s null space has dimension three and its range space
has dimension zero. Here is the canonical
form
matrix.
0 0 0
1 0 0
0 1 0
Five.III.2.21 By Lemma 1.4 the nullity has grown as large as possible by the n-th
iteration where n is the dimension of the domain. Thus, for the 2×2 matrices, we
need only check whether the square is the zero matrix. For the 3×3 matrices, we need
only check the cube.
(a) Yes, this matrix is nilpotent because its square is the zero matrix.
(b) No, the square is not the zero matrix.
!2 !
3 1 10 6
=
1 3 6 10
(c) Yes, the cube is the zero matrix. In fact, the square is zero.
(d) No, the third power is not the zero matrix.
3
1 1 4 206 86 304
3 0 −1 = 26 8 26
5 2 7 438 180 634
(e) Yes, the cube of this matrix is the zero matrix.
Another way to see that the second and fourth matrices are not nilpotent is to note
that they are nonsingular.
Five.III.2.22 The table of calculations
320 Linear Algebra, by Hefferon
p Np N (Np )
0 1 1 0 1 r
0 0 1 1 1 u
−u − v | r, u, v ∈ C }
{
1 0
0 0 0 0
0 0 0 0 0 u
0 0 0 0 0 v
0 0 1 1 1 r
0 0 0 0 0 s
{
−u − v | r, s, u, v ∈ C}
2 0
0 0 0 0
0 0 0 0 0 u
0 0 0 0 0 v
2 –zero matrix– C5
gives these requirements of the string basis: three basis vectors map directly to zero,
one more basis vector maps to zero by a second application, and the final basis vector
maps to zero by a third application. Thus, the string basis has this form.
~ 1 7→ β
β ~ 2 7→ β~ 3 7→ ~0
~ 4 7→ ~0
β
~ 5 7→ ~0
β
From that the canonical form is immediate.
0 0 0 0 0
1 0 0 0 0
0 1 0 0 0
0 0 0 0 0
0 0 0 0 0
Five.III.2.23 (a) The canonical form has a 3×3 block and a 2×2 block
0 0 0 0 0
1 0 0 0 0
0 1 0 0 0
0 0 0 0 0
0 0 0 1 0
corresponding to the length three string and the length two string in the basis.
(b) Assume that N is the representation of the underlying map with respect to the
standard basis. Let B be the basis to which we will change. By the similarity
diagram
n
C2wrt E2 −−−−→ C2wrt E2
N
idyP idyP
n
C2wrt B −−−−→ C2wrt B
Answers to Exercises 321
p Np N (Np )
−1 1 −1 u
1 1
0 1
{
0 | u ∈ C}
1 −1 1 −u
1 0 1 u
2 0
0 0
{
v | u, v ∈ C}
−1 0 −1 −u
3 –zero matrix– C3
shows that any map represented by this basis must act on a string basis in this
way.
~ 1 7→ β
β ~ 2 7→ β
~ 3 7→ ~0
and apply t.
~0 = c1,−1 β
~ 1 + c1,0 t(β ~~ 1 ) + c1,h1~0
~ 1 ) + · · · + c1,h1 −1 th1 (β
+ c2,−1 β ~ 2 + · · · + ci,hi −1 thi (β~i ) + ci,hi~0
324 Linear Algebra, by Hefferon
Conclude that the coefficients c1,−1 , . . . , c1,hi −1 , c2,−1 , . . . , ci,hi −1 are all zero as B∪ Ĉ
is a basis. Substitute back into the first displayed equation to conclude that the
remaining coefficients are zero also.
Five.III.2.32 For any basis B, a transformation n is nilpotent if and only if N =
RepB,B (n) is a nilpotent matrix. This is because only the zero matrix represents the
zero map and so nj is the zero map if and only if Nj is the zero matrix.
Five.III.2.33 It can be of any size greater than or equal to one. To have a transformation
that is nilpotent of index four, whose cube has range space of dimension k, take a
vector space, a basis for that space, and a transformation that acts on that basis in
this way.
~1
β 7→ ~ 2 7→ β
β ~ 3 7→ β ~ 4 7→ ~0
~
β5 7→ ~ ~
β6 7→ β7 7→ β ~ 8 7→ ~0
..
.
~
β4k−3 7→ β ~ 4k−2 7→ β
~ 4k−1 7→ β~ 4k 7→ ~0
..
.
–possibly other, shorter, strings–
So the dimension of the range space of T 3 can be as large as desired. The smallest
that it can be is one — there must be at least one string or else the map’s index of
nilpotency would not be four.
Five.III.2.34 These two have only zero for eigenvalues
! !
0 0 0 0
0 0 1 0
but are not similar (they have different canonical representatives, namely, themselves).
Five.III.2.35 It is onto by Lemma 1.4. It need not be the identity: consider this map
t : R2 → R2 . ! !
x t y
7−→
y x
For that map R∞ (t) = R2 , and t is not the identity.
Five.III.2.36 A simple reordering of the string basis will do. For instance, a map that is
associated with this string basis
~ 1 7→ β
β ~ 2 7→ ~0
~ 1, β
is represented with respect to B = hβ ~ 2 i by this matrix
!
0 0
1 0
Answers to Exercises 325
~ 2, β
but is represented with respect to B = hβ ~ 1 i in this way.
!
0 1
0 0
Five.III.2.37 Let t : V → V be the transformation. If rank(t) = nullity(t) then the
equation rank(t) + nullity(t) = dim(V) shows that dim(V) is even.
Five.III.2.38 For the matrices to be nilpotent they must be square. For them to commute
they must be the same size. Thus their product and sum are defined.
Call the matrices A and B. To see that AB is nilpotent, multiply (AB)2 = ABAB =
AABB = A2 B2 , and (AB)3 = A3 B3 , etc., and, as A is nilpotent, that product is
eventually zero.
The sum is similar; use the Binomial Theorem.
Five.III.2.39 Some experimentation gives the idea for the proof. Expansion of the second
power
t2S (T ) = S(ST − T S) − (ST − T S)S = S2 − 2ST S + T S2
the third power
t3S (T ) = S(S2 − 2ST S + T S2 ) − (S2 − 2ST S + T S2 )S
= S3 T − 3S2 T S + 3ST S2 − T S3
and the fourth power
t4S (T ) = S(S3 T − 3S2 T S + 3ST S2 − T S3 ) − (S3 T − 3S2 T S + 3ST S2 − T S3 )S
= S4 T − 4S3 T S + 6S2 T S2 − 4ST S3 + T S4
suggest that the expansions follow the Binomial Theorem. Verifying this by induction
on the power of tS is routine. This answers the question because, where the index of
nilpotency of S is k, in the expansion of t2k
S
X
2k i 2k
tS (T ) = (−1) Si T S2k−i
i
06i62k
for any i at least one of the S and S2k−i has a power higher than k, and so the term
i
Jordan Form
Five.IV.1.13 For each, the minimal polynomial must have a leading coefficient of 1 and
Theorem 1.8, the Cayley-Hamilton Theorem, says that the minimal polynomial must
contain the same linear factors as the characteristic polynomial, although possibly of
lower degree but not of zero degree.
(a) The possibilities are m1 (x) = x − 3, m2 (x) = (x − 3)2 , m3 (x) = (x − 3)3 , and
m4 (x) = (x − 3)4 . Note that the 8 has been dropped because a minimal polynomial
must have a leading coefficient of one. The first is a degree one polynomial, the
second is degree two, the third is degree three, and the fourth is degree four.
(b) The possibilities are m1 (x) = (x + 1)(x − 4), m2 (x) = (x + 1)2 (x − 4), and
m3 (x) = (x + 1)3 (x − 4). The first is a quadratic polynomial, that is, it has degree
two. The second has degree three, and the third has degree four.
(c) We have m1 (x) = (x − 2)(x − 5), m2 (x) = (x − 2)2 (x − 5), m3 (x) = (x − 2)(x − 5)2 ,
and m4 (x) = (x − 2)2 (x − 5)2 . They are polynomials of degree two, three, three,
and four.
(d) The possibilities are m1 (x) = (x + 3)(x − 1)(x − 2), m2 (x) = (x + 3)2 (x − 1)(x − 2),
m3 (x) = (x + 3)(x − 1)(x − 2)2 , and m4 (x) = (x + 3)2 (x − 1)(x − 2)2 . The degree
of m1 is three, the degree of m2 is four, the degree of m3 is four, and the degree of
m4 is five.
transformations.
0 0 0 ... 0
1 0 0 0
0 1 0
..
.
0 0 0 1 0
This is an (n + 1)×(n + 1) matrix with an easy characteristic polynomial, c(x) = xn+1 .
(Remark: this matrix is RepB,B (d/dx) where B = hxn , nxn−1 , n(n−1)xn−2 , . . . , n!i.)
To find the minimal polynomial as in Example 1.12 we consider the powers of T −0I = T .
But, of course, the first power of T that is the zero matrix is the power n + 1. So the
minimal polynomial is also xn+1 .
Five.IV.1.17 Call the matrix T and suppose that it is n×n. Because T is triangular,
and so T − xI is triangular, the characteristic polynomial is c(x) = (x − λ)n . To see
that the minimal polynomial is the same, consider T − λI.
0 0 0 ... 0
1 0 0 . . . 0
0 1 0
..
.
0 0 ... 1 0
Recognize it as the canonical form for a transformation that is nilpotent of degree n;
the power (T − λI)j is zero first when j is n.
m2 (x) = (x − 1)2 ,
0 0 2 6
0 0 0 6
(T − 1I)2 =
0 0 0 0
0 0 0 0
m3 (x) = (x − 1)3 ,
0 0 0 6
0 0 0 0
(T − 1I)3 =
0 0 0 0
0 0 0 0
and m4 (x) = (x − 1)4 . Because m1 , m2 , and m3 are not right, m4 must be right, as
is easily verified.
In the case of a general n, the representation is an upper triangular matrix with
ones on the diagonal. Thus the characteristic polynomial is c(x) = (x − 1)n+1 . One
way to verify that the minimal polynomial equals the characteristic polynomial is
argue something like this: say that an upper triangular matrix is 0-upper triangular if
there are nonzero entries on the diagonal, that it is 1-upper triangular if the diagonal
contains only zeroes and there are nonzero entries just above the diagonal, etc. As the
above example illustrates, an induction argument will show that, where T has only
nonnegative entries, T j is j-upper triangular.
Five.IV.1.19 The map twice is the same as the map once: π ◦ π = π, that is, π2 = π and
so the minimal polynomial is of degree at most two since m(x) = x2 − x will do. The
fact that no linear polynomial will do follows from applying the maps on the left and
right side of c1 · π + c0 · id = z (where z is the zero map) to these two vectors.
0 1
0 0
1 0
Thus the minimal polynomial is m.
Five.IV.1.20 This is one answer.
0 0 0
1 0 0
0 0 0
Five.IV.1.21 The x must be a scalar, not a matrix.
Five.IV.1.22 The characteristic polynomial of
!
a b
T=
c d
332 Linear Algebra, by Hefferon
A2 + 2A + 3I = (PBP−1 )2 + 2 · PBP−1 + 3 · I
= (PBP−1 )(PBP−1 ) + P(2B)P−1 + 3 · PP−1 = P(B2 + 2B + 3I)P−1
shows that f(A) is similar to f(B).
Answers to Exercises 335
(a) Taking f to be a linear polynomial we have that A−xI is similar to B−xI. Similar
matrices have equal determinants (since |A| = |PBP−1 | = |P|·|B|·|P−1 | = 1·|B|·1 = |B|).
Thus the characteristic polynomials are equal.
(b) As P and P−1 are invertible, f(A) is the zero matrix when, and only when, f(B)
is the zero matrix.
(c) They cannot be similar since they don’t have the same characteristic polynomial.
The characteristic polynomial of the first one is x2 − 4x − 3 while the characteristic
polynomial of the second is x2 − 5x + 5.
Five.IV.1.32 Suppose that m(x) = xn + mn−1 xn−1 + · · · + m1 x + m0 is minimal for
T.
(a) For the ‘if’ argument, because T n + · · · + m1 T + m0 I is the zero matrix we have
that I = (T n + · · · + m1 T )/(−m0 ) = T · (T n−1 + · · · + m1 I)/(−m0 ) and so the matrix
(−1/m0 )·(T n−1 +· · ·+m1 I) is the inverse of T . For ‘only if’, suppose that m0 = 0 (we
put the n = 1 case aside but it is easy) so that T n +· · ·+m1 T = (T n−1 +· · ·+m1 I)T
is the zero matrix. Note that T n−1 + · · · + m1 I is not the zero matrix because
the degree of the minimal polynomial is n. If T −1 exists then multiplying both
(T n−1 + · · · + m1 I)T and the zero matrix from the right by T −1 gives a contradiction.
(b) If T is not invertible then the constant term in its minimal polynomial is zero.
Thus,
T n + · · · + m1 T = (T n−1 + · · · + m1 I)T = T (T n−1 + · · · + m1 I)
is the zero matrix.
Five.IV.1.33 (a) For the inductive step, assume that Lemma 1.7 is true for polynomials
of degree i, . . . , k−1 and consider a polynomial f(x) of degree k. Factor f(x) = k(x−
λ1 )q1 · · · (x − λz )qz and let k(x − λ1 )q1 −1 · · · (x − λz )qz be cn−1 xn−1 + · · · + c1 x + c0 .
Substitute:
k(t − λ1 )q1 ◦ · · · ◦ (t − λz )qz (~v) = (t − λ1 ) ◦ (t − λ1 )q1 ◦ · · · ◦ (t − λz )qz (~v)
= (t − λ1 ) (cn−1 tn−1 (~v) + · · · + c0~v)
= f(t)(~v)
(the second equality follows from the inductive hypothesis and the third from the
linearity of t).
(b) One example is to consider the squaring map s : R → R given by s(x) = x2 . It
is nonlinear. The action defined by the polynomial f(t) = t2 − 1 changes s to
f(s) = s2 − 1, which is this map.
s2 −1
x 7−→ s ◦ s(x) − 1 = x4 − 1
Observe that this map differs from the map (s − 1) ◦ (s + 1); for instance, the first
map takes x = 5 to 624 while the second one takes x = 5 to 675.
336 Linear Algebra, by Hefferon
Five.IV.1.34 Yes. Expand down the last column to check that xn + mn−1 xn−1 + · · · +
m1 x + m0 is plus or minus the determinant of this.
−x 0 0 m0
0 1−x 0 m1
0 0 1−x m2
..
.
1 − x mn−1
Five.IV.2.19 (a) The characteristic polynomial is c(x) = (x − 3)2 and the minimal
polynomial is the same.
(b) The characteristic polynomial is c(x) = (x + 1)2 . The minimal polynomial is
m(x) = x + 1.
(c) The characteristic polynomial is c(x) = (x + (1/2))(x − 2)2 and the minimal
polynomial is the same.
(d) The characteristic polynomial is c(x) = (x − 3)3 The minimal polynomial is the
same.
(e) The characteristic polynomial is c(x) = (x − 3)4 . The minimal polynomial is
m(x) = (x − 3)2 .
(f) The characteristic polynomial is c(x) = (x + 4)2 (x − 4)2 and the minimal polyno-
mial is the same.
(g) The characteristic polynomial is c(x) = (x − 2)2 (x − 3)(x − 5) and the minimal
polynomial is m(x) = (x − 2)(x − 3)(x − 5).
(h) The characteristic polynomial is c(x) = (x − 2)2 (x − 3)(x − 5) and the minimal
polynomial is the same.
t
C3wrt B −−−−→ C3wrt B
J
The matrix to move from the lower left to the upper left
is this.
1 −2 0
−1
P−1 = RepE3 ,B (id) = RepB,E3 (id) = 1 0 1
−2 0 0
The matrix P to move from the upper right to the lower right is the inverse of P−1 .
(b) We want this matrix and its inverse.
1 0 3
P−1 = 0 1 4
0 −2 0
338 Linear Algebra, by Hefferon
(c) The concatenation of these bases for the generalized null spaces will do for the
basis for the entire space.
−1 −1 1 0 −1
0 0 1 0 −1
B−1 = h 0 , −1i B3 = h−1 , 0 , 1 i
1 0 0 −2 2
0 1 0 2 0
The change of basis matrices are this one and its inverse.
−1 −1 1 0 −1
0 0 1 0 −1
P−1 = 0 −1 −1 0 1
1 0 0 −2 2
0 1 0 2 0
Five.IV.2.22 The general procedure is to factor the characteristic polynomial c(x) =
(x − λ1 )p1 (x − λ2 )p2 · · · to get the eigenvalues λ1 , λ2 , etc. Then, for each λi we find a
string basis for the action of the transformation t − λi when restricted to N∞ (t − λi ),
by computing the powers of the matrix T − λi I and finding the associated null spaces,
until these null spaces settle down (do not change), at which point we have the
generalized null space. The dimensions of those null spaces (the nullities) tell us the
action of t − λi on a string basis for the generalized null space, and so we can write
the pattern of subdiagonal ones to have Nλi . From this matrix, the Jordan block Jλi
associated with λi is immediate Jλi = Nλi + λi I. Finally, after we have done this for
each eigenvalue, we put them together into the canonical form.
(a) The characteristic polynomial of this matrix is c(x) = (−10−x)(10−x)+100 = x2 ,
so it has only the single eigenvalue λ = 0.
power p (T + 0 · I)p N ((t − 0)p ) nullity
−10 4 2y/5
1 { | y ∈ C} 1
−25 10 y
0 0
2 C2 2
0 0
(Thus, this transformation is nilpotent: N∞ (t − 0) is the entire space). From the
~ 1 7→ β
nullities we know that t’s action on a string basis is β ~ 2 7→ ~0. This is the
canonical form matrix for the action of t − 0 on N∞ (t − 0) = C2
!
0 0
N0 =
1 0
Answers to Exercises 339
(The null space of (t − 4)3 is the same, as it must be because the power of the term
associated with λ2 = 4 in the characteristic polynomial is two, and so the restriction
of t − 2 to the generalized null space N∞ (t − 2) is nilpotent of index at most two —
it takes at most two applications of t − 2 for the null space to settle down.) The
pattern of how the nullities rise tells us that the action of t − 4 on an associated
string basis for N∞ (t − 4) is β
~ 2 7→ β
~ 3 7→ ~0.
Putting the information for the two eigenvalues together gives the Jordan form
of the transformation t.
1 0 0
0 4 0
0 1 4
We can take elements of the null spaces to get an appropriate basis.
0 1 0
_
B = B1 B4 = h1 , 0 , 5i
0 1 5
(d) The characteristic polynomial is c(x) = (−2 − x)(4 − x)2 = −1 · (x + 2)(x − 4)2 .
For the eigenvalue λ−2 , calculation of the powers of T + 2I yields this.
z
N (t + 2) = { z | z ∈ C}
z
The null space of (t + 2)2 is the same, and so this is the generalized null space
N∞ (t + 2). Thus the action of the restriction of t + 2 to N∞ (t + 2) on an associated
~ 1 7→ ~0.
string basis is β
For λ2 = 4, computing the powers of T − 4I yields
z x
N (t − 4) = { −z | z ∈ C} N ((t − 4)2 ) = { −z | x, z ∈ C }
z z
and so the action of t − 4 on a string basis for N∞ (t − 4) is β
~ 2 7→ β
~ 3 7→ ~0.
Therefore the Jordan form is
−2 0 0
0 4 0
0 1 4
and a suitable basis is this.
1 0 −1
_
B = B−2 B4 = h1 , −1 , 1 i
1 1 −1
Answers to Exercises 341
and
N ((t − 6)3 ) = C4
~ 1 7→ β
and the nullities show that the action of t−6 on a string basis is β ~ 2 7→ β
~ 3 7→ ~0
~ ~
and β4 7→ 0. The Jordan form is
6 0 0 0
1 6 0 0
0 1 6 0
0 0 0 6
and finding a suitable string basis is routine.
0 2 3 −1
0 −1 3 −1
B = h , , , i
0 −1 −6 1
1 2 3 0
Five.IV.2.23 There are two eigenvalues, λ1 = −2 and λ2 = 1. The restriction of t + 2 to
N∞ (t + 2) could have either of these actions on an associated string basis.
~ 1 7→ β
β ~ 2 7→ ~0 ~1 →
β 7 ~0
~2 →
β 7 ~0
The restriction of t − 1 to N∞ (t − 1) could have either of these actions on an associated
string basis.
~ 3 7→ β
β ~ 4 7→ ~0 ~ 3 7→ ~0
β
~ 4 7→ ~0
β
In combination, that makes four possible Jordan forms, the two first actions, the
second and first, the first and second, and the two second actions.
−2 0 0 0 −2 0 0 0 −2 0 0 0 −2 0 0 0
1 −2 0 0 0 −2 0 0 1 −2 0 0 0
−2 0 0
0 0 1 0 0 0 1 0 0 0 1 0 0 0 1 0
0 0 1 1 0 0 1 1 0 0 0 1 0 0 0 1
Five.IV.2.24 The restriction of t + 2 to N∞ (t + 2) can have only the action β~ 1 7→ ~0. The
restriction of t − 1 to N∞ (t − 1) could have any of these three actions on an associated
string basis.
~ 2 7→ β
β ~ 3 7→ β
~ 4 7→ ~0 ~ 2 7→ β
β ~ 3 7→ ~0 ~2 →
β 7 ~0
~
β4 7→ 0~ β3 7→ ~0
~
~ 4 7→ ~0
β
Taken together there are three possible Jordan forms, the one arising from the first
action by t − 1 (along with the only action from t + 2), the one arising from the second
Answers to Exercises 343
Five.IV.2.35 False; these two 4×4 matrices each have c(x) = (x−3)4 and m(x) = (x−3)2 .
3 0 0 0 3 0 0 0
1 3 0 0 1 3 0 0
0 0 3 0 0 0 3 0
0 0 1 3 0 0 0 3
Five.IV.2.36 (a) The characteristic polynomial is this.
a − x b
= (a−x)(d−x)−bc = ad−(a+d)x+x2 −bc = x2 −(a+d)x+(ad−bc)
c d − x
Five.IV.2.39 One such ordering is the dictionary ordering. Order by the real component
first, then by the coefficient of i. For instance, 3 + 2i < 4 + 1i but 4 + 1i < 4 + 2i.
Five.IV.2.40 The first half is easy — the derivative of any real polynomial is a real
polynomial of lower degree. The answer to the second half is ‘no’; any complement of
Pj (R) must include a polynomial of degree j + 1, and the derivative of that polynomial
is in Pj (R).
Five.IV.2.41 For the first half, show that each is a subspace and then observe that any
polynomial can be uniquely written as the sum of even-powered and odd-powered
terms (the zero polynomial is both). The answer to the second half is ‘no’: x2 is even
while 2x is odd.
Five.IV.2.42 Yes. If RepB,B (t) has the given block form, take BM to be the first j vectors
of B, where J is the j×j upper left submatrix. Take BN to be the remaining k vectors
in B. Let M and N be the spans of BM and BN . Clearly M and N are complementary.
To see M is invariant (N works the same way), represent any m ~ ∈ M with respect
to B, note the last k components are zeroes, and multiply by the given block matrix.
The final k components of the result are zeroes, so that result is again in M.
Five.IV.2.43 Put the matrix in Jordan form. By non-singularity, there are no zero
eigenvalues on the diagonal. Ape this example:
2
9 0 0 3 0 0
1 9 0 = 1/6 3 0
0 0 4 0 0 2
to construct a square root. Show that it holds up under similarity: if S2 = T then
(PSP−1 )(PSP−1 ) = PT P−1 .
348 Linear Algebra, by Hefferon
1 (a) By eye, we see that the largest eigenvalue is 4. Sage gives this.
sage: def eigen(M,v,num_loops=10):
....: for p in range(num_loops):
....: v_normalized = (1/v.norm())*v
....: v = M*v
....: return v
....:
sage: M = matrix(RDF, [[1,5], [0,4]])
sage: v = vector(RDF, [1, 2])
sage: v = eigen(M,v)
sage: (M*v).dot_product(v)/v.dot_product(v)
4.00000147259
(b) A simple calculation shows that the largest eigenvalue is 2. Sage gives this.
sage: M = matrix(RDF, [[3,2], [-1,0]])
sage: v = vector(RDF, [1, 2])
sage: v = eigen(M,v)
sage: (M*v).dot_product(v)/v.dot_product(v)
2.00097741083
(b) Sage takes a few more iterations on this one. This makes use of the procedure
defined in the prior item.
sage: M = matrix(RDF, [[3,2], [-1,0]])
sage: v = vector(RDF, [1, 2])
sage: v,v_prior,dex = eigen_by_iter(M,v)
sage: (M*v).norm()/v.norm()
2.01585174302
sage: dex
6
Sage does not return (use <Ctrl>-c to interrupt the computation). Adding some error
checking code to the routine
def eigen_by_iter(M, v, toler=0.01):
dex = 0
diff = 10
while abs(diff)>toler:
dex = dex+1
if dex>1000:
print "oops! probably in some loop: \nv=",v,"\nv_next=",v_next
v_next = M*v
if (v.norm()==0):
print "oops! v is zero"
return None
if (v_next.norm()==0):
print "oops! v_next is zero"
return None
v_normalized = (1/v.norm())*v
v_next_normalized = (1/v_next.norm())*v_next
diff = (v_next_normalized-v_normalized).norm()
v_prior = v_normalized
v = v_next_normalized
return v, v_prior, dex
gives this.
oops! probably in some loop:
v= (0.707106781187, -1.48029736617e-16, -0.707106781187)
v_next= (2.12132034356, -4.4408920985e-16, -2.12132034356)
oops! probably in some loop:
v= (-0.707106781187, 1.48029736617e-16, 0.707106781187)
v_next= (-2.12132034356, 4.4408920985e-16, 2.12132034356)
oops! probably in some loop:
v= (0.707106781187, -1.48029736617e-16, -0.707106781187)
v_next= (2.12132034356, -4.4408920985e-16, -2.12132034356)
So it is circling.
5 In theory, this method would produce λ2 . In practice, however, rounding errors in
the computation introduce components in the direction of ~v1 , and so the method will
still produce λ1 , although it may take somewhat longer than it would have taken with
a more fortunate choice of initial vector.
6 Instead of using ~vk = T~vk−1 , use T −1~vk = ~vk−1 .
1 The equation
! ! !
0.89 0 .90 .01 −.01 −.01
0.89I − T = − =
0 0.89 .10 .99 −.10 −.10
leads to this system. ! ! !
−.01 −.01 p 0
=
−.10 −.10 r 0
So the eigenvectors have p = −r.
2 Sage finds this.
sage: M = matrix(RDF, [[0.90,0.01], [0.10,0.99]])
sage: v = vector(RDF, [10000, 100000])
sage: for y in range(10):
....: v[1] = v[1]*(1+.01)^y
....: print "pop vector year",y," is",v
....: v = M*v
....:
pop vector year 0 is (10000.0, 100000.0)
pop vector year 1 is (10000.0, 101000.0)
pop vector year 2 is (10010.0, 103019.899)
pop vector year 3 is (10039.19899, 106111.421211)
pop vector year 4 is (10096.3933031, 110360.453787)
pop vector year 5 is (10190.3585107, 115891.187687)
pop vector year 6 is (10330.2345365, 122872.349786)
pop vector year 7 is (10525.9345807, 131525.973067)
pop vector year 8 is (10788.6008533, 142139.351965)
pop vector year 9 is (11131.1342876, 155081.09214)
So inside the park the population grows by about eleven percent while outside the
park the population grows by about fifty five percent.
3 The matrix equation
! ! !
pn+1 0.95 0.01 pn
=
rn+1 0.05 0.99 rn
means that to find eigenvalues we want to solve this.
λ − 0.95 0.01
0= = λ2 − 1.94λ − 0.9404
0.10 λ − 0.99
Sage gives this.
sage: a,b,c = var('a,b,c')
sage: qe = (x^2 - 1.94*x -.9404 == 0)
sage: print solve(qe, x)
[
x == -1/100*sqrt(18813) + 97/100,
x == 1/100*sqrt(18813) + 97/100
]
sage: n(-1/100*sqrt(18813) + 97/100)
-0.401604899378826
sage: n(1/100*sqrt(18813) + 97/100)
2.34160489937883
3 We have this.
0 0 1 1/2
1/3 0 0 0
H=
1/3 1/2 0 1/2
1/3 1/2 0 0
(c) Page p3 is important, but it passes its importance on to only one page, p1 . So
that page receives a large boost.
So the value of the expression is dominated by the first term. Solving 1000 =
√ √
(1/ 5) · ((1 + 5)/2)n gives this.
sage: a = ln(1000*5^(0.5))/ln(phi)
sage: a
16.0271918385296
sage: psi^(17)
-0.000280033582072583
So by the seventeenth power, the second term does not contribute enough to change
the roundoff. For the ten thousand and million calculations the situation is even more
extreme.
354 Linear Algebra, by Hefferon
sage: b = ln(10000*5^(0.5))/ln(phi)
sage: b
20.8121638053112
sage: c = ln(1000000*5^(0.5))/ln(phi)
sage: c
30.3821077388746
has roots of 2 and 3. Any function of the form f(n) = c1 2n + c2 3n satisfies the
recurrence.
(b) The matrix expression of the relation is
! ! !
f(n) 0 4 f(n − 1)
=
f(n − 1) 1 0 f(n − 2)
and the characteristic equation
2
λ − 2 = (λ − 2)(λ + 2)
has the two roots 2 and −2. Any function of the form f(n) = c1 2n + c2 (−2)n
satisfies this recurrence.
(c) In matrix form the relation
f(n) 5 −2 −8 f(n − 1)
f(n − 1) = 1 0 0 f(n − 2)
f(n − 2) 0 1 0 f(n − 3)
has a characteristic equation with roots −1, 2, and 4. Any combination of the form
c1 (−1)n + c2 2n + c3 4n solves the recurrence.
3 (a) The solution of the homogeneous recurrence is f(n) = c1 2n + c2 3n . Substituting
f(0) = 1 and f(1) = 1 gives this linear system.
c1 + c2 = 1
2c1 + 3c2 = 1
By eye we see that c1 = 2 and c2 = −1.
(b) The solution of the homogeneous recurrence is c1 2n + c2 (−2)n . The initial
conditions give this linear system.
c1 + c2 = 0
2c1 − 2c2 = 1
The solution is c1 = 1/4, c2 = −1/4.
Answers to Exercises 355
(The matrix is square so the sign in front of −λ is −1even ). Application of the inductive
hypothesis gives the desired result.
= (−1)k−1 an−k · 1
− λ · (−1)k−2 (−λk−1 + an−1 λk−2 + an−2 λk−3 + · · · + an−k+1 λ0 )
6 This is a straightforward induction on n.
7 Sage says that we are safe.
sage: T64 = 18446744073709551615
sage: T64_days = T64/(60*60*24)
sage: T64_days
1229782938247303441/5760
sage: T64_years = T64_days/365.25
sage: T64_years
5.84542046090626e11
sage: age_of_universe = 13.8e9
sage: T64_years/age_of_universe
42.3581192819294