Mathematics Notes Chapter 2
Mathematics Notes Chapter 2
Syllabus Inner product space: Definition and properties of inner product space, orthog-
onality, Cauchy Schwarz inequality, Norm and Orthogonal Basis and Gramm-Schmidt
orthonormalisation. Schur’s Theorem, Linear functional, Riesz representation theorem,
orthogonal complement or dual subspace, Singular value and singular vectors, singular
value decomposition.
Definition 2.1 Dot product For x, y ∈ Rn , the dot product of x and y, denoted x · y,
is defined by x · y = x1 y1 + x2 y2 + · · · + xn yn , where x = (x1 , x2 , · · · , xn ) and y =
(y1 , y2 , · · · , yn ).
Note The dot product of two vectors in Rn is a number, not a vector. Obviously
x · x = ∥x∥2 for all x ∈ Rn . The dot product on Rn has the follow- ing properties:
(c) For y ∈ Rn fixed, the map from Rn → R that sends x ∈ Rn to x · y is linear, and
An inner product is a generalization of the dot product. At this point you may be
tempted to guess that an inner product is defined by abstracting the properties of the dot
product discussed above. For real vector spaces, that guess is correct. However, wee need
to examine the complex vector space, before making the general definition so that it will
be useful for both real and complex vector spaces.
Recall that if z = a + ib, where a, b ∈ R, then
(i) the absolute value of z denoted |z| is defined by |z| = a2 + b2 ,
1
2 BSCAIML-301 Notes by Subodh C Bhunia
(d) Homogeneity in first slot: ⟨au, v⟩ = a⟨u, v⟩ for all a ∈ F and all u, v ∈ V ,
Every real number equals its complex conjugate. Thus if we are dealing with a real
vector space, then in the last condition simply states that ⟨u, v⟩ = ⟨v, u⟩ for all u, v ∈ V .
Definition 2.3 Inner product space An inner product space is a vector space V along
with an inner product defined on V .
The most important example of an inner product space is F n with the Euclidean inner
product defined in (a) of Illus. 1.1.
When F n is referred to as an inner product space, you should assume that the inner
product is the Euclidean inner product unless explicitly told otherwise. So that we do
not have to keep repeating the hypothesis that V is an inner product space, for the rest
of this chapter we make the following assumption:
Notation V For the rest of discussion, V denotes an inner product space over F . When
we say that a vector space V is an inner product space, we also want to mean that an
inner product on V is the Euclidean inner product if the vector space is F n .
CHAPTER 2. INNER PRODUCT SPACE 3
Proof: (a) follows from the conditions of additivity in the first slot and homogeneity in
the first slot in the definition of an inner product.
(b) follows from part (a) and the result that every linear map takes 0 to 0.
Part (c) follows from part (b) and the conjugate symmetry property in the definition of
an inner product.
(d) Suppose u, v, w ∈ V . Then
⟨u, v + w⟩ = ⟨v + w, u⟩
= ⟨v, u⟩ + ⟨w, u⟩
= ⟨v, u⟩ + ⟨w, u⟩
= ⟨u, v⟩ + ⟨u, w⟩
(e) Suppose a ∈ F and u, v ∈ V . Then
⟨u, av⟩ = ⟨av, u⟩ = a ⟨v, u⟩
= a⟨u, v⟩.
Norms Our motivation for defining inner products came initially from the norms of
vectors on R2 and R3 . Now we see that each inner product determines a norm.
Illus. 2.2 (a) If (z1 , z2 ,p · · · , zn ) ∈ F n (with the Euclidean inner product), then
∥(z1 , z2 , · · · , zn )∥ = |z1 |2 + |z2 |2 + · · · + |zn |2 .
(b) In the vector Zspace of continuous real-valued functions on [−1, 1] with inner product
1
defined as ⟨f · g⟩ = f (x)g(x)dx, we have
sZ −1
1
∥f ∥ = |f (x)|2 dx.
−1
Illus. 2.3 If u, v are nonzero vectors in R2 , then ⟨u, v⟩ = ∥u∥ ∥v∥ cos θ
where θ is the angle between u and v. Thus two vectors in R2 are orthogonal (with
respect to the usual Euclidean inner product) if and only if the cosine of the angle between
them is 0, which happens if and only if the vectors are perpendicular in the usual sense of
plane geometry.
u-cv
u
cv
For example, the standard basis {(1, 0, 0), (0, 1, 0), (0, 0, 1)} is an orthonormal basis of R3 .
Theorem 2.1 An orthonormal list of the right length (equal to the dim(V ) is an or-
thonormal basis of V .
Proof Since it an orthonormal list, the list must be linearly independent. Since, the
number of vectors in the orthonormal list is equal to the dimension of V , it must be a
basis of V .
Illus.
2.6 Consider the following list
R4 :
of four vectorsin
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
, , , , , ,− ,− , ,− ,− , , − , ,− , .
2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2
Let us verify that the list forms an orthonormal basis in R4 .
s
2 2 2 2
1 1 1 1 1 1 1 1
We have , , , = + + + =1
2 2 2 2 2 2 2 2
Similarly, the other three vectors in the list above also have norm 1.
We
have
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
, , , , , ,− ,− = · + · − · − · =0
2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2
Similarly, the inner product of any two distinct vectors in the list above also equals 0.
Thus the list above is orthonormal. Because we have an orthonormal list of length four
in the four-dimensional vector space R4 , this list is an orthonormal basis of R4 .
Theorem 2.3 Let {u1 , u2 , · · · , um } be an orthogonal basis for a vector space V . For each
y ∈ V , the weights in the linear combination y = c1 u1 + c2 u2 + · · · + cm um are given by
⟨y, uj ⟩
cj = j = 1, 2, · · · , m.
⟨uj , uj ⟩
Proof y = c1 u1 + c2 u2 + · · · + cm um .
The orthogonality of {u1 , u2 , · · · , um } shows that
⟨y, uj ⟩ = ⟨(c1 u1 + c2 u2 + · · · + cm um ), uj ⟩ = cj ⟨uj , uj ⟩
⟨y, uj ⟩
Since ⟨uj , uj ⟩ =
̸ 0, the equation above has a solution
⟨uj , uj ⟩
for cj for each j = 1, 2, · · · , m.
Proof Recall that T has an upper-triangular matrix with respect to some basis of V .
Now apply the previous theorem.
Note : We shall use the name Hilbert’s space inour discussion, so we shall have some
idea about it.
Hilbert’s space : Firstly a metric space is a set X together with a notion of distance
d between its elements, usually called points. The distance is measured by a function
called a metric or distance function. For x, y ∈ X, the distance between them is written
as d(x, y) and the metric space is written as (X, d).
A Cauchy sequence is a sequence { xn } in a metric space (X, d) such that for every
ϵ > 0, there exists an integer N such that d(xn , xm ) < ϵ for all n, m ≥ N . If every Cauchy
sequences defined on a inner-product space converges to a limit point that is also within
the space, such a space is then called complete.
Further,
p in an inner product space, if the norm is defined using the inner product i.e.,
∥x∥ = ⟨x, x⟩, then the inner product space is called a Hilbert space, if it is complete.
Proof First we show there exists a vector u ∈ V such that ϕ(v) = ⟨v, u⟩ for every v → V .
Let e1 , · · · , en be an orthonormal basis of V . Then
ϕ(v) = ϕ(⟨v, e1 ⟩ e1 + · · · + ⟨v, en ⟩ en )
= ⟨v, e1 ⟩ ϕ(e1 ) + · · · + ⟨v, en ⟩ ϕ(en )
= ⟨v, ϕ(e1 ) e1 + · · · + ϕ(en ) en ⟩ for every v ∈ V .
Thus setting u = ϕ(e1 ) e1 + · · · + ϕ(en ) en , we have
ϕ(v) = ⟨v, u⟩ for every v ∈ V , as desired.
Now we prove that only one vector v ∈ V has the desired behavior
CHAPTER 2. INNER PRODUCT SPACE 11
Illus. 2.8 If U is a line in R3 containing the origin, then U ⊥ is the plane containing the
origin that is perpendicular to U . If U is a plane in R3 containing the origin, then U ⊥ is
the line containing the origin that is perpendicular to U .
Thus U ∩ U ⊥ ⊂ {0}.
(e) Suppose U and W are subsets of V and U ⊂ W . Suppose v ∈ W ⊥ .
Then ⟨v, u⟩ = 0 for every u ∈ W , which implies that ⟨v, u⟩ = 0 for every u ∈ U .
Hence v ∈ U ⊥ . Thus W ⊥ ⊂ U ⊥ .
Theorem 2.8 (The Hilbert projection theorem). For a Hilbert space V and a closed
convex subset U , the distance top described above is attained by a unique element of U .
Proof is omitted.
Definition 2.11 (Orthogonal projections) For a vector v and a closed convex subset
U (most often a closed subspace) of a Hilbert space, we use vU to denote this distance-
minimizing element of U , called the orthogonal projection of v into U .
This is the set of all vectors w ∈ Rn that are orthogonal to all of the vectors in V .
v1T
v2T
..
.
T
Then the matrix A = vm
T
vm+1
.
..
vkT
has more columns than rows, so its null space is nonzero.
Let x be a nonzero vector in Nul (A).
T
v1 · x
v1 x
v2T x v2 · x
.. ..
. .
T
Then 0 = Ax = vm x = vm · x
T
vm+1 x vm+1 · x
.. ..
. .
vkT x vk · x
Since v1 · x = v2 · x = · · · = vm · x = vm+1 · x = · · · = vk · x = 0,
it follows that x is in V ⊥ and similarly, x is in (V ⊥ )⊥ .
This again implies x is orthogonal to itself, which contradicts our assumption that x
is nonzero. Therefore, k = n, as desired.
(2) Finally, we prove the second assertion.
Clearly V is contained in (V ⊥ )⊥ .
Let m = dim (V ). By the Property (3), proved above, we have dim (V ⊥ ) = n − m,
so dim ((V ⊥ )⊥ ) = n − (n − m) = m.
The only m-dimensional subspace of (V ⊥ )⊥ is (V ⊥ )⊥ itself, so (V ⊥ )⊥ = V .
Explain in details orthogonal complement or dual subspace, Singular value and singu-
lar vectors, singular value decomposition
Orthogonal Complement or Dual Subspace:
Given a vector space V over a field (usually the real numbers or complex numbers),
the orthogonal complement of a subspace U of V , denoted as U ⊥ , is the set of all vectors
in V that are orthogonal to every vector in U . In other words, for any vector ∈ U and
any vector v ∈ U ⊥ , their inner product (dot product) is zero: ⟨u, v⟩ = 0.
Mathematically, the orthogonal complement can be defined as:
U ⊥ = {v ∈ V |⟨u, v⟩ = 0 for all u ∈ U ⊥ }
The orthogonal complement is itself a subspace of V , and it is always true that U ∩
U ⊥ = {0}, meaning the only vector common to both U and its orthogonal complement is
the zero vector.
Singular Value and Singular Vectors: Consider a matrix A with dimensions m × n (m
rows and n columns).
1. Compute the eigenvalues and eigenvectors of AT A to obtain the right singular vec-
tors (V T ).
CHAPTER 2. INNER PRODUCT SPACE 17
2. Normalize the right singular vectors to obtain the unit right singular vectors.
3. Compute the square root of the eigenvalues of AT A to obtain the singular values
(the diagonal elements of diagonal matrix Σ).
4. Compute the left singular vectors (U ) by dividing the corresponding right singular
vectors by the singular values.
SVD is widely used in various fields including linear algebra, signal processing, data anal-
ysis, image compression, and machine learning. It’s a fundamental tool for understanding
and manipulating matrices.
Example 2.1 For an integer n > 0, let Pn denote the vector space of polynomials with
real coefficients of degree 2 or less.
Define the map T : P2 → P4 by T (f )(x) = f (x2 ).
Determine if T is a linear transformation.
If it is, find the matrix representation for T relative to the basis B = {1, x, x2 } of P2
and C = {1, x, x2 , x3 , x4 } of P4 .
Soln. To prove that T is a linear transformation, we must show that it satisfies both
axioms for linear transformations.
For f, g ∈ P2 , we have
T (f + g)(x) = (f + g)(x2 ) = f (x2 ) + g(x2 ) = T (f )(x) + T (g)(x)
Also, for a scalar c ∈ R, we have
T (cf )(x) = (cf )(x2 ) = cf (x2 ) = cT (f )(x).
We see that T is a linear transformation.
To find its matrix representation, we must calculate T (f ) for each f ∈ B and find its
coordinate vector relative to the basis C.
We calculate T (1) = 1, T (x) = x2 , T (x2 ) = x4 .
Each of these is an element of C. Their coordinate vectors relative to C are thus
1 0 0
0 0 0
[T (1)]B = 0 , [T (x)]B = 1 , and [T (x2 )]B = 0
0 0 0
0 0 1
The matrix representation of T is found by combining these columns, in order, into
one matrix.
18 BSCAIML-301 Notes by Subodh C Bhunia
1 0 0
0 0 0
Thus, [T ]C→B =
0 1 0
0 0 0
0 0 1
3 0
Example 2.2 Find the matrices U, Σ, V for A = .
4 5
Soln. The rank is r = 2. With rank 2, this A has positive singular values σ1 and σ2 .
We will see that σ1 is larger than λmax = 5, and σ2 is smaller than λmin = 3.
Begin with AT A and T
AA :
T 3 4 3 0 25 20
A A= =
0 5 4 5 20 25
T 3 0 3 4 9 12
and AA = =
4 5 0 5 12 41
2 2
These have the same trace √ √ eigenvalues σ1 = 45 and σ2 = 5.
(50) and the same
The square roots are
√σ1 = 45and σ2 = 5.
45 √0 √ 3 0
Thus, we get Σ = = 5 .
0 5 0 1
A key step is to find the eigenvectors of AT A (with eigenvalues 45 and 5) :
25 20 1 1
= 45
20 25 1 1
25 20 1 1
=5
20 25 −1 −1
Then v1 and v2 are orthogonal eigenvectors of AT A rescaled to length 1.
So, right singular
vectors are:
1 1 1 1
v1 = √ and v2 = √
2 1 2 −1
Next, we campute the left singular vectors ui√ ’s. √
Compute Av 1 and
Av 2 which
will be
σ 1 u 1 = 45 u
1 and σ 2 u 2 = 5 u2 respectively.
3 0 1 1 1 3 3 1
Av1 = √ =√ =√
4 5 2 1 2 9 2 3
1 3 1 1 1
So, u1 = √ √ =√
45 2 3 10 3
3 0 1 1 1 3 1 3
Similarly Av2 = √ =√ =√
4 5 −1 −1 −1
2 2 2
1 1 3 1 3
So, u2 = √ √ =√
5 2 −1 10 −1
CHAPTER 2. INNER PRODUCT SPACE 19
√
1 1 1 1 1 3 3 0
Thus, V = √ , U=√ and Σ = 5 .
2 −1 1 10 3 −1 0 1
Thus, the singular value decomposition of A is U ΣV T .
Note that U and V contain orthonormal bases for the column space and the row space
respectively.
1 1
Example 2.3 Find the matrices U, Σ, V for A = 1 0 .
0 1