Text (Vector Space)
Text (Vector Space)
1 Vector Spaces
Recall properties of Rn .
– (addition) u + v ∈ Rn for u, v ∈ Rn .
– (scalar multiplication) cu ∈ Rn for u ∈ Rn and c ∈ R.
1. u + v = v + u;
2. (u + v) + w = u + (v + w);
3. u + 0 = u, where 0 is the zero vector in Rn ;
4. u + (−u) = 0;
5. c(u + v) = cu + cv;
6. (c + d)u = cu + du;
7. c(du) = (cd)u;
8. 1u = u.
Basically, other properties of Rn can be proven by these properties. This means that these
properties are essential for Rn . If operations of a set satisfy these essential properties, then
the set and Rn should have same properties. We call the set a vector space:
– (addition) u + v ∈ V for u, v ∈ V .
– (scalar multiplication) cu ∈ V for u ∈ V and c ∈ R.
• The two operations satisfy the following properties (axioms) hold for any u, v, w ∈ V
and c, d ∈ R:
1. u + v = v + u;
2. (u + v) + w = u + (v + w);
3. There exists x ∈ V such that for any u ∈ V , u + x = u. We denote this vector by
0, which is called the zero vector of V ;
1
4. For any u ∈ V , there exists u′ ∈ V such that u + u′ = 0. We denote this vector
by −u, which is called the additive inverse of u;
5. c(u + v) = cu + cv;
6. (c + d)u = cu + du;
7. c(du) = (cd)u;
8. 1u = u.
Hence Rn is a vector space under the usual addition and scalar multiplication.
Example 1.3. Let Mn be the set of n × n matrices. We consider the usual matrix addition
and matrix scalar multiplication. For example, for M2 ,
a11 a12 b11 b12 a11 + b11 a12 + b12
• (addition) + = .
a21 a22 b21 b22 a21 + b21 a22 + b22
a11 a12 ca11 ca12
• (scalar multiplication) c = .
a21 a22 ca21 ca22
Then the axioms (1), (2), (5), (6), (7), (8) follow by definition. The zero matrix
0 ··· 0
0 = ... . . . ...
0 ··· 0
satisfies the axiom (3). Moreovre, for A ∈ Mn , (−1)A satisfies the axiom (4). Hence Mn is
a vector space. In general, the set of all n × m matrices is a vector space under the usual
matrix addition and matrix scalar multiplication.
Example 1.4. Let Pn be the set of all polynomials of degree at most n. For f, g ∈ Pn with
f (t) = a0 + a1 t + · · · an tn and g(t) = b0 + b1 t + · · · + bn tn , we define the addition f + g on Pn
by
(f + g)(t) = (a0 + b0 ) + (a1 + b1 )t + · · · + (an + bn )tn ∈ Pn .
Moreover, for f ∈ Pn and c ∈ R, we define the scalar multiplication cf on Pn by
Then it is easy to show that the axioms (1), (2), (5), (6), (7), (8) hold. The zero polynomial
0 defined by
0(t) = 0 + 0t + · · · + 0tn
satisfies the axiom (3). Moreover, for f ∈ Pn , (−1)f satisfies the axiom (4). Hence Pn is a
vector space under these operations.
2
Example 1.5. For R2 , we define the scalar multiplication c ⊙ (a, b) = (ca, b). Then R2 is not
a vector space under the usual addition and this scalar multiplication. Indeed, the axiom (6)
does not always hold. For example,
(r + s) ⊙ (a, b) = ((r + s)a, b),
but
r ⊙ (a, b) + s ⊙ (a, b) = (ra, b) + (sa, b) = ((r + s)a, 2b).
Proposition 1.6. Let V be a vector space. Then
(a) (The uniqueness of 0) If there exists 0′ ∈ V such that u + 0′ = u for any u ∈ V , then
0′ = 0.
(b) (The uniqueness of −u) Let u ∈ V . If there exist u′ , u′′ ∈ V such that u+u′ = u+u′′ =
0, then u′ = u′′ .
Proof. (a) By applying u + 0′ = u for u = 0, one has 0 + 0′ = 0. On the other hand, from
the axiom (3) we obtain 0′ + 0 = 0′ . Moreover, from the axiom (1) one has 0 + 0′ = 0′ + 0.
Hence 0 = 0 + 0′ = 0′ + 0 = 0′ .
(b) By adding u′′ to the both side of u + u′ = 0, one has
(u + u′ ) + u′′ = 0 + u′′
(u′ + u) + u′′ = 0 + u′′ (axiom (1))
u′ + (u + u′′ ) = 0 + u′′ (axiom (2))
u′ + 0 = 0 + u′′ (assumption)
u′ = u′′ (axiom (3))
Proposition 1.7. Let V be a vector space. For any u ∈ V and c ∈ R, one has
(a) 0u = 0,
(b) c0 = 0,
(c) −u = (−1)u.
Proof. (a)
0u = (0 + 0)u
= 0u + 0u (axiom (6)).
By adding −0u to the both sides, one has
0u + (−0u) = (0u + 0u) + (−0u)
0 = (0u + 0u) + (−0u) (axiom (4))
= 0u + (0u + (−0u)) (axiom (1))
= 0u + 0 (axiom (4))
= 0u (axiom (3)).
3
(b)
(c) We want to show that u + (−1)u = 0. This implies that −u = (−1)u from axiom (4).
2 Subspaces
Definition 2.1. Let V be a vector space. A subspace of V is a subset W of V that satisfies
the following three properties:
(a) The zero vector of V is in W ;
4
(c) For any u ∈ W and c ∈ R, cu ∈ W .
Proposition 2.2. Let V be a vector space. Then any subspace of V is a vector space under
the operations on V .
Proof. Properties (b) and (c)define two operations on W . Moreover, the axioms (1), (2),
(5)–(8) automatically hold in W . Property (a) guarantees axiom (3) in W . If u ∈ W , then
(−1)u = −u ∈ W from property (c). This implies axiom (4) in W . Hence W is a vector
space.
Example 2.3. For any nonzero vector space V has at least two subspaces. One is itself.
One is the zero space. Indeed, let W = {0}. Then
(a) 0 ∈ W .
(b) 0 + 0 = 0 ∈ W (axiom(3)).
In Example 2.4, since B and C are subspaces of A, B and C are vector spaces from
Proposition 2.2. On the other hand, C is a subset of B. In this case, we can C as a subspace
of B. In fact, we can show the following easily.
5
Definition 2.6. Let V be a vector space and u1 , . . . , up ∈ V . Then a linear combina-
tion of u1 , . . . , up if a vector in the form c1 u1 + · · · + cp up with c1 , . . . , cp ∈ R. we denote
span{u1 , . . . , up } the set of all linear combinations of u1 , . . . , up . Namely,
span{u1 , . . . , up } := {c1 u1 + · · · + cp up : c1 , . . . , cp ∈ R} ⊂ V.
(a1 u1 + · · · + ap up ) + (b1 u1 + · · · + bp up )
= (a1 + b1 )u1 + · · · + (ap + bp )up ∈ span{u1 , . . . , up }.
3 Linear transformations
Let V and W be vector spaces. Now, we consider a map T from V to T , denoted by
T : V → W , namely, we assign each vector u in V to a vector T (u) in W . We say that T (u)
is the image of u under T . Denote Im(T ) the set of all images T (u) with u ∈ V . We say
that Im(T ) is the image of V under T . We remark that Im(T ) is a subset of W .
Definition 3.1. Let V and W be vector spaces and T : V → W . We say that T is a linear
transformation if
1. T (u + v) = T (u) + T (v) for any u, v ∈ V .
6
Example 3.2. Let A be an m × n matrix and we define a map T : Rn → Rm by x → Ax.
Then T is a linear transformation. In this case, Im(T ) = ColA.
a b
Example 3.3. We define a map det : M2 → R defined by A = → ad − bc. Then for
c d
1 0
I2 = , one has det(I2 ) = 1 and det(2I2 ) = 4. Hence 4 = det(2I2 ) ∕= 2 det(I2 ) = 2. This
0 1
implies that det is not a linear transformation.
a b
Example 3.4. We define a map T : M2 → R defined by → a + d. Then
c d
a b e f a+e b+f
1. For A = ,B = ∈ M2 , A + B = . Hence T (A + B) =
c d g h c+g d+h
(a + e) + (d + h) = (a + d) + (e + h) = T (A) + T (B).
a b xa xb
2. For A = and x ∈ R, xA = . Hence T (xA) = xa + xd = x(a + d) =
c d xc xd
xT (A).
Therefore,
T is a linear transformation. In general, a map Trn : Mn → R defined by
aij 1≤i,j≤n → a11 + · · · + ann is a linear transformation. Trn is called the trace of Mn . Then
Im(Trn ) = R. In particular, Im(Trn ) is a subspace of R.
Proposition 3.5. Let V and W be vector spaces and T : V → W a linear transformation.
Then Im(T ) is a subspace of W .
Proof. (a) T (0V ) = T (00V ) = 0T (0V ) = 0W ∈ Im(T ).
(b) For T (x), T (y) ∈ Im(T ) with x, y ∈ V , one has T (x) + T (y) = T (x + y) ∈ Im(T ).
(c) For T (x) ∈ Im(T ) with x ∈ V and c ∈ R, one has cT (x) = T (cx) ∈ Im(T ).
Therefore, Im(T ) is a subspace of W .
Definition 3.6. Let V and W be vector spaces and T : V → W a linear transformation.
The kernel Ker(T ) of T is the set of all vectors x ∈ V such that T (x) = 0W , where 0W is
the zero vector of W .
Example 3.7. Let A be an m×n matrix and we define a linear transformation T : Rn → Rm
by x → Ax. Then Ker(T ) = NulA.
a b
Example 3.8. We consider the trace Tr2 of M2 . Then A = ∈ Ker(Tr2 ) if and only
c d
if Tr2 (A) = a + d = 0. Hence
a b
Ker(Tr2 ) = : a, b, c ∈ R .
c −a
7
Proof. (a) In the proof of Proposition 3.5, we have already seen that T (0V ) = 0W . Hence
0V ∈ Ker(T ).
• {f1 } is linearly independent. In fact, for any vector space V and a vector x ∈ V , {x}
is linearly independent if and only if x ∕= 0.
• {f1 , f2 } is linearly independent. In fact, for any vector space V and x, y ∈ V with
x, y ∕= 0, {x, y} is linearly dependent if and only if there exists c ∈ R such that y = cx.
T (c1 x1 + · · · + cp xp ) = T (0V )
c1 T (x1 ) + · · · + cp T (xp ) = 0W
8
Example 4.5. We consider the trace Tr2 of M2 . Let
1 0 0 0
A= ,B = .
0 0 0 −1
Then {A, B} is linearly independent. However, since Tr2 (A) + Tr2 (B) = 0, {Tr2 (A), Tr2 (B)}
is not linearly independent.
Definition 4.6. Let V and W be linear spaces. A linear transformation T : V → W is said
to be one-to-one (injective) if T (x) = T (y) with x, y ∈ V implies that x = y.
Proposition 4.7. Let V and W be linear spaces and T : V → W a linear transformation.
Then
(a) T is one-to-one if and only if Ker(T ) = {0V }.
(b) Assume that T is one-to-one. If {x1 , . . . , xp } is linearly independent, then {T (x1 ), . . . , T (xp )}
is linearly independent.
Proof. (a) (⇒) Take x ∈ Ker(T ). Then T (x) = 0W = T (0V ). Since T is one-to-one, x = 0V .
Hence Ker(T ) = {0V }.
(⇐) Take x, y ∈ V such that T (x) = T (y). Then
T (x) = T (y)
T (x) − T (y) = 0W
T (x − y) = 0W .
5 Basis
Definition 5.1. Let W be a subspace of a vector space V . A set of vectors B = {b1 , . . . , bp }
in V is a basis for W if
1. B is linearly independent.
2. W = spanB.
Example 5.2. Set S = {1, t, t2 , . . . , tn } ⊂ Pn . Then S is linearly independent. Suppose that
c0 · 1 + c1 t + · · · + cn tn = 0(t) with c0 , . . . , cn ∈ R. Then c0 = · · · = cn = 0. For example,
when n = 2, c0 = 0 (t = 0), c0 + c1 + c2 = 0 (t = 1) and c0 + 2c1 + 4c2 = 0 (t = 2) imply
c0 = c1 = c2 = 0. Hence S is linearly independent. Since clearly, Pn = spanS, S is a basis
for Pn . In this case, S is called the standard basis for Pn .
9
Example 5.3. Let f1 (t) = 2t − t2 , f2 (t) = 2 + 2t, f3 (t) = 6 + 16t − 5t2 and W =
span{f1 , f2 , f3 } ⊂ P2 . Then since f3 = 5f1 + 3f2 , f3 ∈ span{f1 , f2 }. In particular,
W = span{f1 , f2 }. Since {f1 , f2 } is linearly independent, {f1 , f2 } is a basis for H.
Theorem 5.4 (The Spanning Set Theorem). Let V be a vector space and S = {v1 , . . . , vp } ⊂
V and W = spanS. Then
(a) If one of the vectors in S, say vk , is a linear combination of the remaining vectors in
S, then S \ {vk } still spans W .
Theorem 5.5 (The unique representation theorem). Let V be a vector space and B =
{b1 , . . . , bn } a basis for V . Then for each x ∈ V , there uniquely exist c1 , . . . , cn ∈ V such
that
x = c 1 b1 + · · · + c n bn . (1)
Proof. Since spanB = V , there exist c1 , . . . , cn ∈ V satisfying (1). Suppose that x also has
the representation x = d1 b1 + · · · + dn bn with d1 , . . . , dn ∈ R. Then
n
Example 5.7. We consider the standard basis S = {1, t, . . . , t } for Pn . Then for f (t) =
a0
..
a0 + a1 t + · · · + an t ∈ Pn , f S = . ∈ Rn+1 .
n
an
10
Definition 5.8. Let V and W be vector spaces. We say that a linear transformation T :
V → W is onto (surjective) if for any w ∈ W , there exists v ∈ V such that T (v) = w.
Namely, Im(V ) = W .
Definition 5.9. Let V and W be vector spaces. We say that a linear transformation T :
V → W is an isomorphism from V onto W if it is one-to-one and onto. In this case, we
say that V and W are isomorphic, denoted by V ∼
= W.
Theorem 5.10. Let V be a vector space and B = {b1 , . . . , bn } a basis for V . Then the
coordinate mapping of B is an isomorphism from V onto Rn .
Proof. It is an exercise to show that the coordinate mapping is a linear transformation. We
show that the map is one-to-one and onto.
Let T be the coordinate mapping of B.
Assume that for some x ∈ V , x B = 0Rn . Namely, x ∈ ker(T ). This implies that
x = 0b1 + · · · 0bn = 0V . Hence ker(T ) = {0V }. It then follows from Proposition 4.7 (a) that
T is one-to-one.
a1 a1
..
Take any . ∈ Rn . Set x = a1 b1 + · · · + an bn ∈ V . Then T (x) = x B = ... .
an an
Hence, T is onto.
From Propositions 4.4 and 4.7 (b) we have the following.
Corollary 5.11. Let V be a vector space and B = {b1, . . . , bn } a basis for V . Then
{x1 , . . . , xp } ⊂ V is linearly independent if and only if { x1 B , . . . , xp B } is linearly in-
dependent.
Example 5.12. Let f1 (t) = 1, f2 (t) = t, f3 (t) = 4 − t ∈ P1. We recall that S ={1, t} is the
1 0 4
standard basis for P1 . Then f1 S = =: v1 , f2 S = =: v2 , f3 S = =: v3 . We
0 1 −1
recall that {f1 } and {f1 , f2 } are linearly independent but {f1 , f2 , f3 } is linearly dependent.
Hence it follows that {v1 } and {v1 , v2 } are linearly independent but {v1 , v2 , v3 } is linearly
dependent.
6 Dimension
Theorem 6.1. Let V be a vector space and B = {b1 , . . . , bn } a basis for V . Then any set
in V containing more than n vectors must be linearly dependent.
Proof. Let {u1 , . . . , up } be a set in V with p > n. Then { u1 B , . . . , u1 B } ⊂ Rn is linearly
dependent. It then follows from Corollary 5.11 that {u1 , . . . , up } is also linearly dependent.
Theorem 6.2. Let V be a vector space and B = {b1 , . . . , bn } a basis for V . Then any basis
for V must consist of exactly n vectors.
Proof. Let A = {a1 , . . . , am } be an arbitrary basis for V . Since B is a basis for V and A is
linearly independent, one has m ≤ n from Theorem 6.1. Since A is a basis for V and B is
linearly independent, one has n ≤ m. Hence we obtain n = m.
11
Definition 6.3. Let V be a vector space. We say that V is finite-dimensional. if V
is spanned by a finite set. In this case, it follows from Theorem 5.4 that V has a basis.
The dimension of V , denoted by dim V , is the number of vectors in a basis for V . For
convenience, we define dim{0} = 0. If V is not spanned by a finite set, then we say that V
is infinite-dimensional.
is a basis for M2 . Hence one has dim M2 = 4. In general, it follows that dim Mn = n2 .
Example 6.5. We consider the vector space P2 . Then {1, t, t2 } is a basis for P2 . Hence one
has dim P2 = 3. In general, it follows that dim Pn = n + 1.
Example 6.6. Let P is the set of all polynomials. Then we can easily show that P is a
vector space. Moreover, P is infinite-dimensional. Indeed, if P is spanned by a finite set A.
Then we let n be the maximal degree of the polynomials in A. Hence one has tn+1 ∈ / spanA.
n+1
This contradicts that t ∈ P.
Theorem 6.9 (The Basis Theorem). Let V be a p-dimensional vector space with p ≥ 1 and
S = {v1 , . . . , vp } ⊂ V . Then the following arguments are equivalent:
(c) V = spanS.
Proof. (a) ⇒ (b) and (a) ⇒ (c) follow from the definition of a basis.
((b) ⇒ (a)) By Theorem 6.8, S can be expanded to a basis for V . But that basis must
contain exactly p vectors. Hence S must be a basis for V .
((c) ⇒ (a)) Since V is a nonzero space, by Theorem 5.4, a subset S ′ of S is a basis for
V . Since dim V = p, S ′ must contain p vectors. This implies that S = S ′ .
12
7 Inner product spaces
Recall that for u, v ∈ Rn , the standard inner product of u and v is
u ◦ v = ut v = u 1 v 1 + · · · + u n v n .
Proposition 7.1. Let u, v, w ∈ Rn and c ∈ R. Then
1. u ◦ v = v ◦ u,
2. (u + v) ◦ w = u ◦ w + v ◦ w,
3. (cu) ◦ v = c(u ◦ v),
4. u ◦ u ≥ 0, and u ◦ u = 0 ⇐⇒ u = 0.
Definition 7.2. Let V be a vector space. An inner product on V is a function that, to
each pair u, v ∈ V , associates a real number 〈u, v〉 and satisfies the following four axioms
for u, v, w ∈ V and c ∈ R.
1. 〈u, v〉 = 〈v, u〉,
2. 〈u + v, w〉 = 〈u, w〉 + 〈v, w〉,
3. 〈cu, v〉 = c〈u, v〉,
4. 〈u, u〉 ≥ 0, and 〈u, u〉 = 0 ⇐⇒ u = 0.
A vector space with an inner product is called an inner product space.
Example 7.3. For u, v ∈ R2 , define
〈u, v〉 = 4u1 v1 + 5u2 v2 .
Then for any u, v, w ∈ R and c ∈ R
1. 〈u, v〉 = 4u1 v1 + 5u2 v2 = 4v1 u1 + 5v2 u2 = 〈v, u〉.
2. 〈u + v, w〉 = 4(u1 + v1 )w1 + 5(u2 + v2 )w2 = (4u1 w1 + 5u2 w2 ) + (4v1 w1 + 5v2 w2 ) =
〈u, w〉 + 〈v, w〉.
3. 〈cu, v〉 = 4(cu1 )v1 + 5(cu2 )v2 = c(4u1 v1 + 5u2 v2 ) = c〈u, v〉.
4. 〈u, u〉 = 4u21 + 5u22 ≥ 0 and 〈u, u〉 = 0 ⇐⇒ u = 0.
Therefore, 〈u, v〉 defines an inner product on R2 .
Example 7.4. For f, g ∈ Pn , define
1
〈f, g〉 = f (t)g(t)dt.
−1
1 that 〈f,
It is easy to show g〉 satisfies axioms (1), (2) and (3).
4. 〈f, f 〉 = −1 (f (t)) dt. This is equal to the area of the region between (f (t))2 and the
2
13
Proposition 7.5. Let V be an inner product space. Then for any vector u ∈ V , one has
〈0, u〉 = 0.
Proposition 7.7. Let V be an inner product space. Then the zero vector 0 is orthogonal to
any vector in V .
Proof. (1) follows by the same proof for the usual inner product on Rn .
(2) If u = 0, then from Proposition 7.5 the inequality holds. Assume that u ∕= 0. Let
f (t) = tu + v2 with a scalar t. Then one has f (t) = t2 u2 + 2t〈u, v〉 + v2 . Since
u2 > 0, f (t) ≥ 0 if and only if the discriminant of f (t) is less than or equal to 0. Hence
〈u, v〉2 − u2 · v2 ≤ 0. Namely, 〈u, v〉2 ≤ u2 · v2 . This implies |〈u, v〉| ≤ u · v.
(3) One has
14
8 Complex vector spaces
We regard “scalars” as real numbers and studied real vector spaces. From now we regard
“scalars” as complex numbers. Namely, we consider complex vector spaces.
Remark 8.1. More general, we can consider a vector space over a field. The basic intuition
is that a field is a set where one can define addition, subtraction, multiplication, and division
with the usual properties. Standard examples are the real numbers R, and the complex
numbers C, the rational numbers Q, and F2 = {0, 1} and so on.
Example 8.2. Cn is a complex vector space, which is called the n-dimensional complex
space.
As is the case in real vector spaces, we can define subspaces, linear transformations,
linear independence, basis and dimensions and we can several propositions about such
properties.
Definition 8.3. Let V be a complex vector space. An Hermitian inner product on V is
a function that, to each pair u, v ∈ V , associates a complex number 〈u, v〉 and satisfies the
following four axioms for u, v, w ∈ V and c ∈ C.
1. 〈u, v〉 = 〈v, u〉,
〈x, y〉 = x1 y1 + · · · + xn yn .
〈x, y〉 = x1 y1 + · · · + xn yn .
Then the axioms (1), (2) and (3) are clearly satisfied.
√ √ √ √
(4) Since (a + b −1) · (a + b −1) = (a + b −1) · (a − b −1) = a2 + b2 ≥ 0, 〈x, x〉 ≥ 0.
Moreover, 〈x, x〉 = 0 ⇐⇒ x = 0. Therefore, 〈x, y〉 is an Hermitian inner product on V .
For an Hermitian space, we can define norm, orthogonality, orthogonal set, an or-
thogonal basis, an orthonormal set, orthonormal basis, and we can prove the Gram–
Schmidts process
15