The Dual Space
The Dual Space
Contents
1. The dual of a vector space 1
1.1. Linear functionals 1
1.2. The dual space 2
1.3. Dual basis 3
1.4. Linear functionals as covectors; change of basis 4
1.5. Functionals and hyperplanes 5
1.6. Application: Spline interpolation of sampled data (Based on
Prof. Gerlach’s notes) 5
1.7. The bidual 8
1.8. Orthogonality 8
1.9. The transpose 9
1.10. Eigenvalues of the transpose 11
xn
1. The most often encountered linear functionals are integrals. For exam-
ple, on the linear space of continuous functions on [a, b] let I : C[a, b] → R
Rb
defined as (I, f ) = a f (t) dt.
More generally, given some w(t) > 0 on [a, b], the integral with respect to
Rb
the weight w, (Iw , f ) = a f (t) w(t)dt is a linear functional on C[a, b].
2. Evaluation at a point: if t1 ∈ [a, b] define, for V = C[a, b]
(3) Et1 : V → R, (Et1 , f ) = f (t1 )
It is easy to see that Et1 is linear.
One could think of Et1 as an integral with the weight the Dirac’s delta
function δ(t − t1 ). Except, δ is not a function! It is a distribution.
Linear combinations of evaluation functions: if t1 , . . . , tn ∈ [a, b] and
c1 , . . . , cn ∈ F then c1 Et1 + . . . + cn Etn defined by
(c1 Et1 + . . . + cn Etn , f ) = c1 f (t1 ) + . . . + cn f (tn )
is also a linear functional on C[a, b].
3. On the linear space C ∞ (a, b) of infinitely many times differentiable
functions on an interval (a, b), fix some t0 ∈ [a, b]; then (φ, f ) = f (k) (t0 ) is
a linear functional, and so is (φ, f ) = nk=0 ck f (k) (t0 ).
P
1.3. Dual basis. Given B = {v1 , . . . , vn } a basis in V , its dual is one special
associated basis for V 0 : consider the linear functionals v10 , . . . , vn0 defined by
(vj0 , vi ) = δi,j for all i, j = 1, . . . , n, where δi,j is the Kronecker symbol:
0 if i 6= j
δi,j =
1 if i = j
Note that the row vector representations of v10 , v20 , . . . , vn0 in the basis
{v1 , v2 , . . . , vn } are
[1, 0, . . . , 0], [0, 1, . . . , 0], . . . , [0, 0, . . . , 1]
In particular, it is clear that v10 , v20 , . . . , vn0 are linearly independent.
Consider any functional φ ∈ V 0 and its row vector representation (1) in
the basis B. Then
X n
(4) φ= φj vj0
j=1
representation in B 0 then
x1
(φ, x) = φ1 x1 + . . . + φn xn = [φ1 , . . . , φn ] ...
xn
xn x̃n
1.4.1. Change of the dual basis upon a change of coordinates. To find the
transition matrix from the basis dual to BV to the one dual to B̃V note that
X X X
−1 −1 0 −1 −1
(ṽj0 , vk ) = (ṽj0 , Slk ṽl ) = Slk (ṽj , ṽl ) = Slk δjl = Sjk
l l l
X X X
−1 −1 0 −1 0
= Sji δik = Sji (vi , vk ) = ( Sji vi , vk )
i i i
0
P −1 0
for all k, hence ṽj = i Sji vi therefore the transition matrix between the
dual bases is (S −1 )T .
III. THE DUAL SPACE 5
Exercise. How does this compare with the matrix change of variables
S −1 M S? Why do we get something different?
1.5. Functionals and hyperplanes. Two-dimensional subspaces in R3 are
planes going through the origin; they can be specified as the set of all vectors
orthogonal to a given vector n ∈ R3 as {x ∈ R3 | n · x = (nT , x) = 0}. The
row vector nT is the transpose of n.
The plane in R3 passing through a point x0 and orthogonal to n ∈ R3
is {x ∈ R3 | (nT , x − x0 ) = 0}. Equivalently, this is {x ∈ R3 | (nT , x) =
k} where k = (nT , x0 ). Also, equivalently, this is a translation of a two-
dimensional subspace: x0 + {x ∈ R3 | (nT , x) = 0}.
This situation is generalized in arbitrary dimensions as follows.
Hyperspaces are subspaces on Rn of dimension n−1 (they are said to have
codimension 1). Hyperspaces H of a vector space V are maximal subspaces:
if U is a subspace so that H ⊂ U ⊂ V then U = H or U = V .
Hyperplanes are translations of hyperspaces.
Theorem 6. Let V be a vector space.
a) Every hyperspace in V is the null space of a linear functional.
b) Every hyperplane in V is the level set of a linear functional. In other
words, a hyperplane is the set Hφ,k = {x ∈ V | (φ, x) = k} where φ ∈ V 0 is
a nonzero linear functional and k ∈ F .
We may prove these facts a bit later.
1.6. Application: Spline interpolation of sampled data (Based on
Prof. Gerlach’s notes). Suppose we measure some data: at n + 1 data
points s0 , s1 , . . . , sn , assumed distinct, we measure the quantities y0 , y1 , . . . , yn .
The question we want to address is to fit them on a curve, namely to find a
function ψ(t) so that
(6) ψ(xj ) = yj for all j = 0, 1, . . . , n
Of course, there are infinitely many such functions ψ, and to have a
definite (unique) answer, we need to restrict our search to special classes of
functions. The choice of the type of functions used for interpolation is done
taking into account expected error, calculation time etc.
1.6.1. Polynomial interpolation. Polynomials in Pn , of degree at most n,
depend on n + 1 parameters, so it is reasonable to expect one can find a
unique polynomial ψ ∈ Pn satisfying (6). A polynomial of degree n passing
through the given points is called spline polynomial of degree n. Before
computers, elastic strips were used to produce the interpolation. Look up
on Wikipedia “flat spline”.
To find the spline polynomial, we will seek a general formula for a poly-
nomial taking the value pj at x = xj .
A general formula is derived as follows. Consider the evaluation linear
functionals (3) on V = Pn : Es0 , . . . , Esn which satisfy (Esj , p) = p(sj ) for
all j = 0, . . . , , n and all p ∈ Pn .
6 BASED ON NOTES BY RODICA D. COSTIN
which gives any polynomial p ∈ Pn based on its sample values p(s0 ), . . . , p(sn ).
To find the basis Q0 , . . . , Qn , note that the polynomial Qk must satisfy
(Esj , Qk ) = Qk (sj ) = 0 for all j 6= k, therefore Qk = ck (t − s0 )(t − s1 ) . . . (t −
sj−1 )(t − sj+1 ) . . . (t − sn ) with ck determined so that Qk (sk ) = 1, therefore
(t − s0 )(t − s1 ) . . . (t − sj−1 )(t − sj+1 ) . . . (t − sn )
Qk =
(sk − s0 )(sk − s1 ) . . . (sk − sj−1 )(sk − sj+1 ) . . . (sk − sn )
In conclusion, there is a unique polynomial ψ ∈ Pn satisfying (6), namely
n
X
(7) ψ(t) = yk Qk (t)
k=0
Note: the fact that there exists a unique interpolating function in a certain
linear space is based on the fact that the evaluation functionals form a basis
for that linear space.
1.7. The bidual. The dual V 0 is a vector space, therefore it has a dual
V 00 , called the bidual of V . It turns out that the bidual can be naturally
identified with V by the pairing (φ, x), which can be interpreted either as φ
acting on x, or as x acting on φ. First we show the following result, which
is sometimes phrased as “linear functionals separate points”.
Lemma 7. If x ∈ V is so that (φ, x) = 0 for all φ ∈ V 0 , then x = 0.
P
To prove the Lemma, let v1 , . . . , vn of a basis of V ; then x = j xj vj
and we have in particular, (vk0 , x) = xk = 0 for all k, therefore x = 0.
Theorem 8. The bidual V 00 of a finite dimensional space V is isomorphic
to V .
Define
N (M T ) = {y ∈ F m | M T y = 0} the left null space of M
Note that M T y = 0 is equivalent to yT M = 0, justifying the name of left
null space.
Recall that dim R(M ) + dim N (M ) = n and using this for M T , we obtain
dim R(M T ) + dim N (M T ) = m.
Recall that dim R(M ) = rank(M ) and the rank of a matrix is the order
of the largest nonzero minor; then rank(M ) = rank(M T ).
Summarizing:
Theorem 12. The fundamental theorem of linear algebra
Let M be an m × n matrix. Ler r = rank(M ). Then
dim R(M ) = r
dim R(M T ) = r
dim N (M ) = n − r
dim N (M T ) = m − r
Also
N (M ) = R(M T )⊥
N (M T ) = R(M )⊥
where ⊥ signifies the annihilator of the set.
Only the second to last line needs a proof. Here it is: consider first x ∈
N (M ) and show that x belongs to the annihilator of R(M T ), in other words,
that xT (M T y) = 0 for all y ∈ F m , which is obvious since xT (M T y) =
(M x)T y. Conversely, taking x ∈ R(M T )⊥ , this means that 0 = xT (M T y) =
(M x)T y for all y ∈ F m , which means that M x = 0, hence x ∈ N (M ). 2
III. THE DUAL SPACE 11