Linear Algebra
Linear Algebra
in Chemical Engineering
(CH61015)
P. A. Deshpande
Department of Chemical Engineering
Indian Institute of Technology Kharagpur
[email protected]
ii
Contents
iii
iv CONTENTS
Chapter 0
A set is a collection of well-defined objects. The objects present in the set are called its
members or elements. We will generally represent a set by a capital letter and its members
by small letters.
A = {a, e, i, o, u} (1)
A is the set and a, e, i, o, u are its elements. To indicate that a is an element of A, the symbol
∈ is used.
a∈A (2)
The number of elements present in a set is called its cardinality or cardinal number. A set
is said to be a finite set if its cardinality is finite. Else, the set is said to be an infinite set.
A set is said to be an empty set or a null set or a void set if it does not contain any element.
Such a set is generally denoted by φ. A few examples of empty sets are given below.
φ = {} (3)
φ = {x : x ∈ R ∧ x2 = −1} (4)
A set containing exactly one element is called a singleton set. Two sets A and B are said to
be equal if they have same elements.
A = B iff x ∈ A =⇒ x ∈ B ∧ x ∈ B =⇒ x ∈ A (5)
1
2 CHAPTER 0. BACKGROUND: A REVIEW OF SET THEORY
Two sets are said to be equivalent if they have the same number of elements in them i.e.
their cardinalities are equal.
A ∼ B iff |A| = |B| (6)
B ⊂ A iff x ∈ B =⇒ x ∈ A ∀x ∈ B (7)
A set of cardinality |B| has 2|B| number of subsets. Two sets A and B are said to be
comparable if either A ⊂ B or B ⊂ A.
Power set of a set A is a set of all possible subsets of A.
P(A) = {S : S ⊆ A} (8)
The union of two sets A and B is a set which has elements which are either in A or in B or
in both.
A ∪ B = {x : x ∈ A ∨ x ∈ B ∨ x ∈ A ∧ B} (9)
The intersection of two sets A and B is a set which has elements which are common to both
A and B.
A ∩ B = {x : x ∈ A ∧ x ∈ B} (10)
Laws of sets
Idempotent laws:
A∪A=A (11)
A∩A=A (12)
Identity laws:
A∪φ=A (13)
A∩φ=φ (14)
Commutative laws:
Associative laws:
A ∪ (B ∪ C) = (A ∪ B) ∪ C (17)
A ∩ (B ∩ C) = (A ∩ B) ∩ C (18)
Distributive laws:
A ∪ (B ∩ C) = (A ∪ B) ∩ (A ∪ C) (19)
A ∩ (B ∪ C) = (A ∩ B) ∪ (A ∩ C) (20)
Complement laws:
A ∩ A0 = φ (21)
A ∪ A0 = U (22)
φ0 = U (23)
U0 = φ (24)
A − B = {x : x ∈ A ∧ x ∈
/ B} (25)
A∆B = (A − B) ∪ (B − A) (26)
Ordered pair: In any set, the order in which the elements are written is immaterial i.e.
Two ordered pairs (a, b) and (c, d) are equal iff a = c and b = d. As an extension to an
ordered collection of n-elements, we can define ordered n-tuple.
4 CHAPTER 0. BACKGROUND: A REVIEW OF SET THEORY
Cartesian product of sets: Let A and B be two sets. The set of all ordered pairs (a, b) such
that a ∈ A and b ∈ B is called the cartesian product of the two sets. It is denoted by A × B.
A × B = {(a, b) : a ∈ A ∧ b ∈ B} (29)
A × B 6= B × A unless A = B (30)
Relations:
A relation R from a set A to B is a subset of A × B.
R⊂A×B (32)
The domain of a relation R from A to B is the set of all first elements of the ordered pairs
which belong to R.
dom(R) = {a : a ∈ A ∧ (a, b) ∈ R} (33)
The range of a relation R from A to B is the set of all second elements of the ordered pairs
which belong to R.
range(R) = {b : b ∈ B ∧ (a, b) ∈ R} (34)
A × B = {(1, 1), (1, 2), (1, 4), (2, 1), (2, 2), (2, 4), (3, 1), (3, 2), (3, 4)} (35)
We choose a subset of A × B such that the sum of the elements of the ordered pairs is greater
than or equal to 4. Then
R = {(1, 4), (2, 2), (2, 4), (3, 1), (3, 2), (3, 4)} (36)
It can be seen from Figure 1 that there are more than one arrows emerging from A. Hence,
in relations, it is allowed to have more than one ordered pairs with the same first entry i.e.
more than one elements in A can have a corresponding image in B. Now we define a special
case of relations called a function.
5
Functions:
The set A is called the domain of f while the set B is called the co-domain of f . Different
cases given in Figure 2 can be used to identify functions.
Figure 2: Different scenarios depicting a relation to be a function. (i) not a function, (ii)
not a function, (iii) function, (iv) function
6 CHAPTER 0. BACKGROUND: A REVIEW OF SET THEORY
Into function: There is atleast one element in B such that there is no pre-image correspond-
ing to it in A. q in Figure 3(i), for example, does not have any pre-image.
One-One or injective function: For every image, there is exactly one pre-image. But all
images need not have a pre-image. See Figure 3(iii) for example.
One-one onto or bijective function: This function is both one-one, i.e. all the images have
exactly one corresponding pre-image, and onto, i.e. every image has a pre-image. See Figure
3(iv) for example.
Figure 3: Different types of functions. (i) into function, (ii) onto or surjective function, (iii)
one-one or injective function, (iv) one-one onto or bijective function
7
Problems
1. Write the following sets using set builder notation:
(i) A = { 12 , 32 , 34 . . . }
(ii) A = { 21 , 23 , 34 }
(iii) A = {0, 0, 7, 26, 63. . . }
(iv) A = {0, 2, 4, 6. . . }
(v) A = a set of all possible integers whose cube is an odd integer
(vi) A = a set of all real numbers which cannot be written as quotients of two integers
3. Prove that
(i) every set is a subset of itself
(ii) φ is a subset of all sets
(iii) number of subsets of a given set equals 2N where N is the cardinality of the set
(iv) number of proper subsets of a given set equals 2N -2
6. If A = {x : x = 4n + 1, n ∈ N, n ≤ 5} and
B = {x : x = 3n, n ∈ N, n ≤ 8} then determine A∆B.
(a) A ∪ (B ∪ C) = (A ∪ B) ∪ C (b) A ∩ (B ∩ C) = (A ∩ B) ∩ C
(v) Distributive laws:
(a) A ∪ (B ∩ C) = (A ∪ B) ∩ (A ∪ C)
(b) A ∩ (B ∪ C) = (A ∩ B) ∪ (A ∩ C)
(vi) De Morgan’s laws:
(a) (A ∪ B)c = Ac ∩ B c (b) (A ∩ B)c = Ac ∪ B c
(vii) Complement laws:
(a) A ∩ Ac = φ (b) A ∪ Ac = U
(c) φc = U (d) Uc = φ
(viii) Involution law:
(Ac )c = A
8. The Cartesian product A × A has 9 elements with two of the elements being (-1,0)
and (0,1). Identify the set A and determine A × A.
11. For two sets A = {ai }, i = 1 · · · n and B = {bi }, i = 1 · · · n, n ∈ Z, verify that < A, B >
⊂ PP(A ∪ B).
12. If a and b are elements of sets A and B, respectively, then verify the followings:
14. State whether each diagram in the following figures defines a mapping from A = {a, b, c}
into B = {x, y, z}.
9
18. Consider the mapping F : R2 → R2 defined by F (x, y) = (3y, 2x). Let S be the unit
circle in R2 , that is, the solution set of x2 + y 2 = 1. (a) Describe F (S). (b) Find
F −1 (S).
10 CHAPTER 0. BACKGROUND: A REVIEW OF SET THEORY
Binary operation:
Let S be any non-empty set. A function f : S × S → S is called a binary operation on the
set S. The binary operation f on the set S associates every ordered pair (a, b) ∈ S × S to a
unique element f (a, b) ∈ S.
From the above, it can be easily seen that addition is a binary operation on N. Similarly,
multiplication is a binary operation on N. But subtration and division are not binary oper-
ations on N.
a ⊗ b = b ⊗ a ∀a, b ∈ S (1.1)
(a ⊗ b) ⊗ c = a ⊗ (b ⊗ c) ∀a, b, c ∈ S (1.2)
Hence, multiplication and addition of numbers and matrices are associative binary opera-
tions.
11
12 CHAPTER 1. GROUPS, RINGS AND FIELDS
Distributivity:
Let two operations ⊗ and be defined on S. The binary operation ⊗ is said to be
(a) left distributive over if
a ⊗ (b c) = (a ⊗ b) (a ⊗ c) ∀a, b, c ∈ S (1.3)
(b c) ⊗ a = (b ⊗ a) (c ⊗ a) ∀a, b, c ∈ S (1.4)
The binary operation ⊗ is said to be distributive over if it is both left as well as right
distributive.
Closure property:
Let ⊗ be a binary operation on S. For T ⊆ S, T is said to be closed under ⊗ if a ⊗ b ∈
T ∀a, b ∈ T .
Identity element:
Let ⊗ be a binary operation on S. An element e1 ∈ S is called a left identity if
e1 ⊗ a = a ∀a ∈ S (1.5)
a ⊗ e2 = a ∀a ∈ S (1.6)
An element e ∈ S is called an identity element if it is both left and right identity i.e.
e ⊗ a = a = a ⊗ e ∀a ∈ S (1.7)
Inverse of an element:
Let ⊗ be a binary operation on S. An element b1 ∈ S is called the left inverse of an element
a ∈ S if
b1 ⊗ a = e (1.8)
13
Groups:
A group is a non-empty set G equipped with a binary operation ⊗ : G × G → G that
associates an element a ⊗ b ∈ G ∀a, b ∈ G and having the following properties.
a ⊗ (b ⊗ c) = (a ⊗ b) ⊗ c (1.11)
a ⊗ e = e ⊗ a = a i.e. ∃e ∈ G (1.12)
a ⊗ b = b ⊗ a ∀a, b ∈ G (1.14)
eI ⊗ a = a ∀a ∈ G (1.15)
a ⊗ eII = a ∀a ∈ G (1.16)
Similarly, for a = eI ,
eI ⊗ eII = eI (1.18)
Every element in a group has a unique inverse element. To prove this, we assume two inverse
elements of a ∈ G, aI , aII ∈ G.
aI ⊗ a = e (1.19)
a ⊗ aII = e (1.20)
aI ⊗ (a ⊗ aII ) = aI ⊗ e = aI (1.22)
Since ⊗ is associative,
=⇒ aII = aI (1.24)
Subgroups:
Given a non-empty set G, a set H ⊆ G is called a subgroup of G iff
h1 ⊗ h2 ∈ H ∀h1 , h2 ∈ H (1.26)
h−1 ∈ H ∀h ∈ H (1.27)
Rings:
A ring is a non-empty set R equipped with two binary operations
+ : R × R → R (addition) (1.28)
∗ : R × R → R (multiplication) (1.29)
(a + b) + c = a + (b + c) (1.30)
a+b=b+a (1.31)
a+0=0+a=a (1.32)
a ∗ (b ∗ c) = (a ∗ b) ∗ c (1.34)
a∗1=1∗a=a (1.35)
(a + b) ∗ c = (a ∗ c) + (b ∗ c) (1.36)
a ∗ (b + c) = (a ∗ b) + (a ∗ c) (1.37)
Fields:
A set K is a field if it is a ring and the following properties hold.
(a) 06=1
(b) K ∗ = K − {0} is a group i.e. every a 6= 0 has an inverse with respect to *
(c) * is commutative
16 CHAPTER 1. GROUPS, RINGS AND FIELDS
Problems
1. Verify whether the set of integers Z is a group under addition? Identify whether
Z∗ = Z − {0} is a group under multiplication?
2. Verify whether the set Q of rational numbers is a group under addition? Identify
whether Q∗ = Q − {0} is a group under multiplication?
3. Verify whether a set of n × n invertible matrices is a group under (a) matrix addition
and (b) matrix multiplication?
4. Identify whether the following operations make a binary operation on the given sets.
(a) addition on N, Z, I+ , I− , Q, R, C
(b) subtration on N, Z, I+ , I− , Q, R, C
(c) multiplication on N, Z, I+ , I− , Q, R, C
(d) division on N, Z, I+ , I− , Q, R, C
5. For each of the above operations or combination of operations, determine whether the
operations are (a) commutative, (b) associative, (c) distributive.
7. Verify whether a universal set U is closed under every possible binary operation ⊗
defined on U ?
8. Verify whether N is a group under (a) addition, (b) multiplication? If yes, then check
if the group is abelian?
10. Verify whether a set of non-singular matrices (n × n) forms a group under (a) matrix
addition, (b) matrix multiplication? Are the identity elements for the two operations
the same? Is the group abelian under both the operations?
11. Verify whether the additive groups Z, Q, R and C are commutative rings?
17
13. Verify whether the group R[X] of polynomials in one variable with real coefficients is
a ring under multiplication of polynomials? Check whether it is commutative?
14. Verify whether a group of n × n matrices is a ring under matrix multiplication? Check
whether it is commutative?
15. Verify whether a set of fractions of polynomials f (x)/g(x), f (x), g(x) ∈ R(x), g(x) 6= 0
is a field?
a b
16. Let G be a group with elements of the form such that a, b, c, d ∈ {0, 1} and
c d
ad − bc 6= 0. Prove that G is a group under matrix multiplication with order 6.
17. If G is a group such that (ab)2 = a2 b2 ∀a, b ∈ G then prove that G is abelian.
18 CHAPTER 1. GROUPS, RINGS AND FIELDS
Chapter 2
a (u ⊗ v) = a u ⊗ a v (2.1)
(a ⊗ b) v = (a v) ⊗ (b u) (2.2)
(a b) u = a (b u) (2.3)
1u=u (2.4)
The elements of V are called vectors and they will be denoted as ui from now onwards. The
elements of F are called scalars.
- The mapping can be identified as scalar multiplication and the binary operation ⊗ can
be identified as vector addition.
- The vector space is called a real vector space if F = R and it is called a complex vector
space if F = C.
- Strictly speaking, the above properties define a linear vector space. But we will be using
the term vector space to signify the same thing.
19
20 CHAPTER 2. LINEAR VECTOR SPACES
Subspaces:
Let V be a vector space over a field F . A non-empty S ⊂ V is said to be a subspace of V if
S itself is a vector space over F under the operations on V restricted to S.
Restriction:
If S ⊂ V and a binary operation ⊗ is defined on S then ⊗ is said to be a restriction of S on
V if a ⊗ b ∈ S ∀a, b ∈ S.
Criteria for a subset to be a subspace:
Let V be a vector space over F . A non-empty set S ⊂ V is a subspace iff ∀u, v ∈ S and
∀α ∈ F
u+v∈S (2.5)
αu ∈ S (2.6)
Hence, S is an additive abelian group under vector addition and S is closed under scalar
multiplication. The above criteria can be alternatively posed as follows.
A non-empty S ⊂ V is a subspace of V iff αu + βv ∈ S, u, v ∈ S, α, β ∈ F .
21
Problems
1. Show that a set of all ordred n-tuples of elements of any field F is a vector space over
F.
2. Show that a set of F m×n matrices over a field F is a vector space over F with respect
to the addition of matrices as the vector addition and multiplication of a matrix by a
scalar as the scalar multiplication.
3. Show that the set F [x] of all polynomials over a field F is a vector space over F .
4. Let V1 and V2 be two vector spaces over the same field F . Show that their Cartesian
product V1 × V2 = {(v1 , v2 ) : v1 ∈ V1 , v2 ∈ V2 } is a vector space over F .
5. Let F be a field and V be a set of all ordered pairs (a1 , a2 ), a1 , a2 ∈ F . The following
operations have been defined.
λ(a1 , a2 ) = (λa1 , a2 ), λ ∈ F
Verify whether the operations given above make V a vector space over F .
6. Show that a set V of all real valued continuous functions defined on a closed interval
[a, b] is a real vector space with vector addition and scalar multiplication defined as
follows.
∀f, g ∈ V, λ ∈ R.
∀f, g ∈ V X , λ ∈ F .
8. Let V = {(x, y) : x, y ∈ R}. Show that V is not a vector space under the vector
addition and scalar multiplication defined below.
∀(ai , bi ), (aj , bj ) ∈ V, k ∈ R.
α ⊗ β = (xi + xj + 1, yi + yj + 1)
c α = (cxi , cyi )
u ⊗ v = (xi + yi + 1, xj + yj + 1, xk + yk + 1)
12. Identify whether R3 is a vector space over R under the following two operations.
∀(xi , xj , xk ), (yi , yj , yk ) ∈ R3 , a ∈ R.
23
u ⊗ v = (x + y, 1)
a u = (ax, 1)
14. Let V be a set of ordered pairs (a, b), a, b ∈ R. The following operations are defined.
(a, b) ⊗ (c, d) = (a + c, b + d)
k(a, b) = (ka, 0)
Show that V is not a vector space over R under these two operations.
15. Let V be the set of ordered pairs (a, b), a, b ∈ R and k ∈ R. Show that V is not a
vector space over R with the vector addition and scalar multiplication defined by:
(i)
(a, b) + (c, d) = (a + d, b + c)
(ii)
(a, b) + (c, d) = (a + d, b + c)
k(a, b) =(a, b)
(iii)
(iv)
18. Let ai , aj , ak ∈ F . Show that the set S of all triads (xi , xj , xk ) of elements of
F : ai xi + aj xj + ak xk = 0 is a subspace of F 3 . What is the geometrical interpretation
of this when F = R.
19. Show that a set S of all n × n symmetric matrices over a field F is a subspace of the
vector space F n×n matrices over F .
20. Let V be the vector space of all real valued continuous functions over R. Show that
the set S of solutions of the differential equation
d2 y dy
2 2
− 9 + 2y = 0
dx dx
is a subspace of V .
21. Let R be a field of real numbers and S be the set of all solutions of the equation
x + y + 2z = 0. Show that S is a subspace of R3 .
22. Let S be the set of all elements of the form (x + 2y, y, −x + 3y) ∈ R3 , x, y ∈ R. Show
that S is a subspace of R3 .
23. Let V be the vector space of all 2 × 2 matrices over R. Show that
(i) the set of all 2 × 2 singular matrices over R is not a subspace of V .
(ii) the set of all 2 × 2 matrices satisfying A × A = A, A ∈ V is not a subspace of A.
24. Let V be a vector space over R3 . Which of the following subsets of V are subspaces of
V?
(i) S1 = {(a, b, c) : a + b = 0}
(ii) S2 = {(a, b, c) : a = 2b + 1}
(iii) S3 = {(a, b, c) : a ≥ 0}
(iv) S4 = {(a, b, c) : a2 = b2 }
25
26. Let V be the set of all 2 × 3 matrices. V is a vector space over R. Determine which of
the following subsets of V are subspaces of V ?
(i)
( )
a b c
S1 = : a, b, c, d ∈ R
d 0 0
(ii)
( )
a b c
S2 = : a + c = e + f, a, b, c, d, e, f ∈ R
d e f
(iii)
( )
a b c
S3 = : a > 0, b = c, a, b, c, d, e, f ∈ R
d e f
26 CHAPTER 2. LINEAR VECTOR SPACES
Chapter 3
Linear combination:
If V is a vector space over a field F , v1 , v2 , . . . , vn are n vectors in V and λ1 , λ2 , . . . λn are
scalars in F then
n
X
vn+1 = λi vi (3.1)
i=1
Linear span:
Let V be a vector space over a field F and S ⊂ V . A set of all possible finite linear
combinations of vectors in S in called the linear span of S.
( n
)
X
[S] = λi vi : λi ∈ F, n ∈ N, vi ∈ S (3.2)
i=1
Linear independence:
A set of vectors v1 , v2 , . . . , vn in a vector space V over F are said to be linearly independent
if
n
X
λi vi = 0 =⇒ λi = 0∀i, λi ∈ F (3.3)
i=1
27
28 CHAPTER 3. LINEAR INDEPENDENCE, BASIS, DIMENSION AND SPAN
Basis and dimension: A non-empty subset B ⊂ V is said to be a basis for a vector space V
if
(i) B spans V
(ii) all vectors in B are linearly independent.
A vector space is said to be finite dimensional if there exists a finite subset of V that
spans it. The followings are noteworthy.
(i) A null vector cannot be in the basis.
(ii) If the number of vectors in the basis is not finite then the vector space is called infinite
dimensional.
If the number of vectors present in the basis are n then the vector space is said to be n-
dimensional. If V (F ) is a vector space over F then for dimension n,
(i) any set of n + 1 or more vectors in V are linearly dependent.
(ii) no set of n − 1 or lesser vectors can span V .
29
Problems
1. Express the following vectors in R3 as a linear combination of v1 , v2 , v3 .
(i) w1 = [1 −2 5]T ; (ii) w2 = [1 1 1]T ; (iii) w3 = [0 0 0]T
f (x) = x2 + 4x − 3
φ1 (x) = x2 − 2x + 5
φ2 (x) = 2x2 − 3x
φ3 (x) = x + 3
5. Let V = P2 (t) be the vector space of all polynomials of degree lesser than or equal to
2. Express the polynomial f (t) = at2 + bt + c as a linear combination of the following
polynomials.
8. Show that the polynomials 1, 1 + x and (1 + x)2 span the vector space V = P2 (x) of
all polynomials of degree at most 2 over R.
13. Let R2×2 be the vector space of all 2 × 2 matrices. Show that the matrices
1 0 0 1 0 0 0 0
M1 = ; M2 = ; M3 = ; M4 =
0 0 0 0 1 0 0 1
span R2×2 .
v1 = [3 −1 0 − 1]T ; v2 = [2 −1 3 2]T
v3 = [−1 1 1 3]T ; v4 = [1 1 9 − 5]T
15. Find one vector in R3 over R that spans the intersection of subspaces of S and T where
S = {(a, b, 0) : a, b ∈ R} and T = [u1 , u2 ], u1 = [1 1 1]T , u2 = [1 2 3]T .
16. Consider the vector space V = Pn (t) consiting of all polynomials of degree ≤ n. Show
that the set of polynomials 1, t, t2 , t3 , . . . , tn span V .
17. Show that the following vectors are linearly independent in R4 over R.
v1 = [1 1 2 4]T ; v2 = [2 −1 −5 2]T
v3 = [2 1 1 6]T ; v4 = [1 −1 −4 0]T
31
f (x) = 2x3 + x2 + x + 1
g(x) = x3 + 3x2 + x − 2
h(x) = x3 + 2x2 − x + 3
19. Let V be the vector space of functions from R into R. Show that the functions f (t) =
sin(t), g(t) = et , h(t) = t2 are linearly independent in V .
20. Verify if the following matrices in R2×2 over R are linearly independent.
1 1 1 0 1 1
M1 = ; M2 = ; M3 =
1 1 0 1 0 0
23. Which of the following subsets of all continuous function space are linearly indepen-
dent?
(i) S1 = {sin(x), cos(x), sin(x + 1)}
(ii) S2 = {xex , x2 ex , (x2 + x − 1)ex }
(iii) S3 = {sin2 (x), cos(2x), 1}
(iv) S4 = {x, sin(x), cos(x)}
32 CHAPTER 3. LINEAR INDEPENDENCE, BASIS, DIMENSION AND SPAN
24. If the set {v1 , v2 , v3 } is linearly independent in a vector space V (F ) then prove that
the set {v1 + v2 , v2 + v3 , v3 + v1 } is also linearly independent.
26. Find a maximal linearly independent subsystem of a system having the following vec-
tors.
v1 = [2 −2 − 4]T ; v2 = [1 9 3]T
v3 = [−2 −4 1]T ; v4 = [3 7 1]T
27. Determine whether or not each of the following sets forms a basis of R3 (R).
(i) B1 = {[1 1 1]T , [1 0 1]T }
(ii) B2 = {[1 1 1]T , [1 2 3]T , [2 −1 1]T }
(iii) B3 = {[1 1 2]T , [1 2 5]T , [5 3 4]T }
(iv) B4 = {[1 2 3]T , [1 3 5]T , [1 0 1]T , [2 3 0]T }
28. Let v1 = [1 i 0]T , v2 = [2i 1 1]T and v3 = [0 1+i 1 − i]T be three vectors
in C3 (C). Show that the set B = {v1 , v2 , v3 } is a basis of C3 (C).
34. Which of the following subsets Bi form a basis for the given vector space V .
(i) B1 = {[1 0]T , [i 0]T , [0 1]T }; V = C2 [R]
(ii) B2 = {[1 i i + 1]T , [1 i i − 1]T , [i −i 1]T }; V = C3 [C]
(iii) B3 = {1, sin(x), sin2 (x), cos2 (x)}; V = C3 [−π, π]
The above set of equations can be cast as a matrix equation given as follows.
a a ... a1n x b
11 12 1 1
a21 a22 . . . a2n x2 b2
.. .. .. = .. (4.1)
. . . .
an1 an2 . . . ann xn bn
4.1 can be written as A x = b where A is the coefficient matrix, x is the vector of unknowns
to be solved for, and b is the vector consisting of the entries of the RHS.
One of the methods to solve such a system of equations is the Gauss elimination which
involves converting A into its echelon form followed by back substitution. However, here we
35
36CHAPTER 4. ANALYSIS OF SYSTEMS OF SIMULTANEOUS LINEAR EQUATIONS
would be interested in the analysis of the nature of the solutions without actually having to
solve the system. We would be interested in answering the following questions.
(i) Does the solution exist?
(ii) If yes, then is the solution unique?
For this, we will make use of the concept of null space and range space. Let us first consider
the homogeneous case.
The first task is to know whether the null space of A is empty or it has elements in it. Null
space is a space which consists of the vectors which satisfy A x = 0 i.e. it is the space of
solutions of the homogeneous case. If the null space is empty then the only solution is a
trivial solution, x = [0 0 . . . 0]T .
If the null space is non-empty then it is required to determine the dimension and a basis of
the null space. All linear combinations of the basis in the null space satisfy A x = 0.
How to check if the null space is empty? In such a case, the only solution to the homoge-
neous system is a trivial solution. For an n × n system, if det(A)n×n = 0 then the only
solution is the trivial solution. In other words, the homogeneous system will have only a
trivial solution if the rank ρ(A) = n = number of unknowns. If this is not the case then
we need to determine a basis of the null space of A. This can be determined following the
procedure given below.
x1 + x2 + x3 = 0
2x1 + 3x2 + x3 = 0
1 1 1 x1 0
2 3 1 x2 = 0
5 6 4 x3 0
37
R2 → R2 − 2R1
R3 → R3 − 5R1
1 1 1 x 0
1
0 1 −1 x2 = 0
0 1 −1 x3 0
R3 → R3 − R2
1 1 1 x 0
1
0 1 −1 x2 = 0
0 0 0 x3 0
No further operation will further reduce R2 to a zero row. Therefore, we stop here and recast
the matrix equation back to a system of simultaneous equations.
x1 + x2 + x 3 = 0
x2 − x 3 = 0
We have two equations in three unknowns. Hence we have a degree of freedom of one. From
the second equation, we assume x2 = α. Hence x2 = x3 = α and x1 = −2α. The solution to
the homogeneous equations can, hence, be written as
x1 −2
x2 = α 1
x3 1
Hence, the dimension of the null space of A is 1 since there is only one vector in its basis
and a possible basis is [−2 1 1]T . Every vector which is obtained by multiplying α to this
vector will be a solution to the homogeneous equation. It can be seen that this vector or any
multiple of it satisfy the original system of homogeneous equations. It can be noted that we
had three unknowns (x1 , x2 , x3 ) and two equations after row reduction. Hence the degree of
freedom of the system was one. This resulted in the appearance of one parameter α. The
dimension of the null space is also one. Hence, the dimension of the null space will be equal
to the number of parameters or the degree of freedom of the row reduced system.
A system of homogeneous equations always has a solution since a trivial solution always
exists for this case. Hence, for a homogeneous case, we have the following answers.
38CHAPTER 4. ANALYSIS OF SYSTEMS OF SIMULTANEOUS LINEAR EQUATIONS
A x = b; b 6= 0
If A x = 0 has only a trivial solution then A x = b has a unique solution. To prove this, let
us assume that A x = b has two distinct solutions v1 and v2 .
=⇒ A v1 = b
A v2 = b
=⇒ A (v1 − v2 ) = 0
Hence, v1 − v2 must be the solution to the homogeneous case. But the only solution to the
homogeneous case A x = 0 has been given as the trivial solution. Hence v1 − v2 = 0 or
v1 = v2 which means that the solution to A x = b must be unique.
From the foregoing discussion, it can be concluded that to know the nature of solutions of
A x=b, we must first solve for A x=0. If the only solution to A x=0 is a trivial solution then
A x=b must have a unique solution. If the solution to A x=0 is not unique then A x=b may
or may not be solvable. If it is solvable then it will have infinitely many solutions.
Now we focus on the question of determining whether A x=b has a solution or not.
A x=b will have a solution if ρ(A) = ρ(A|b). We make use of this fact to develop the range
space of A. Range space of A contains all the elements which satisfy A x=b. Consider the
previous case with b = [1 2 1]T . The corresponding matrix equation is given below.
1 1 1 x1 1
3 1 x2 = 2
2
5 6 4 x3 1
1 1 1 b1
A|b = 2 3 1 b2
5 6 4 b3
39
R2 → R2 − 2R1
R3 → R3 − 5R1
1 1 1 b1
A|b = 0 1 −1 b2 − 2b1
0 1 −1 b3 − 5b1
R3 → R3 − R2
1 1 1 b1
A|b = 0 1 −1 b2 − 2b1
0 1 −1 b3 − b2 − 3b1
It can be seen from the above rearrangements that ρ(A) = 2. The condition of solvability
dictates that b3 − b2 − 3b1 = 0. Note that despite the entries known in b, we continue with
bi ’s rather than the actual values given in b. We now do the degree of freedom analysis.
Number of unknowns = 3 (b1 , b2 , b3 )
Number of equations = 1 (b3 − b2 − 3b1 = 0)
Hence, we have two arbitarary parameters. Let b1 = α and b2 = β =⇒ b3 = 3α + β.
b 1 0
1
b2 = α 0 + β 1
b3 3 1
Therefore, the dimension of the range space of A = 2 and a possible basis is v1 = [1 0 3]T
and v2 = [0 1 1]T . All possible linear combinations of v1 and v2 satisfy A x=b. Any b
proposed must be expressed as a linear combination of v1 and v2 for the solution to exist.
Therefore, as given originally in the problem,
1 1 0
b = 2 = α 0 + β 1
1 3 1
=⇒ α = 1
β=2
3α + β = 1
40CHAPTER 4. ANALYSIS OF SYSTEMS OF SIMULTANEOUS LINEAR EQUATIONS
No values of α and β satisfy the above set of equations. Hence, the solution does not exist
for the system with b = [1 2 1]T .
41
Problems
1. Determine whether the following set of simultaneous equations have a solution. If yes,
then comment upon the nature of solutions.
(i)
2u + 3v = 0
4u + 5v + w = 0
2u − 5v − 3w = 0
(ii)
x+y =0
x + 2y + z = 0
y + 2z + t = 0
z + 2t = 5
(iii)
2u − v = 0
−u + 2v − w = 0
−v + 2w − z = 0
−w + 2z = 5
(iv)
x1 + x2 + x3 = 6
x1 + 2x2 + 2x3 = 11
(v)
x1 + x2 + x3 = −2
3x1 + 3x2 − x3 = 6
x1 − x2 + x3 = −1
42CHAPTER 4. ANALYSIS OF SYSTEMS OF SIMULTANEOUS LINEAR EQUATIONS
x1 + x2 + 2x3 = 2
2x1 + 3x2 − x3 = 5
3x1 + 4x2 + x3 = c
(i) Determine the dimension and basis for the null space of the coefficient matrix.
(ii) Determine the dimension and basis for the range space.
(c) Find the conditions on b1 , b2 , b3 to have a solution.
4. For each of the following system of equations, find the conditions on bi ’s so as to have
a solution for the system.
1 2 b
1
2 4 x1 b
= 2
2 5 x2 b3
3 9 b4
6. Determine the corresponding ranges space for the following systems of equations.
(i)
1 2 3 b
x 1
1
2 4 6 b2
x =
2
2 5 7 b3
x3
3 9 12 b4
(ii)
x1
1 3 1 2
b
x 1
2
2 6 4 8 = b2
x3
0 0 2 4 b3
x4
(iii)
x1
1 3 3 2 b1
x
2
2
6 9 7 = b
2
x3
−1 −3 3 4 b3
x4
(iv)
x
1
1 2 0 3 x2 b
= 1
2 4 0 7 x3 b2
x4
(v)
x + 2y − 2z = b1
2x + 5y − 4z = b2
4x + 9y − 8z = b3
(vi)
x3 + 2x4 + 3x5 = b3
44CHAPTER 4. ANALYSIS OF SYSTEMS OF SIMULTANEOUS LINEAR EQUATIONS
x1 + 2x2 = b1
2x1 + 4x2 = b2
x 1 = b3
9. For the following set of equations, determine the dimension and a basis for the null
space and the range space.
11. For the following systems of equations, determine the dimension and bases for the null
and range spaces. Determine the solutions to the systems on the basis of your analysis.
(i)
x1 − x2 + 3x3 + 2x4 = b1
3x1 + x2 − x3 + x4 = b2
(ii)
x1 + 2x2 − x3 + x5 = b1
3x1 + 2x2 + x4 = b2
(iii)
x 1 + x 2 − x 3 = b1
−2x1 − x2 + x3 = b2
x1 + 2x2 − 2x3 = b3
(vi)
−x1 + x2 − 2x3 + x4 = 0
2x1 + 3x2 + x3 − x4 = 2
Linear transformations
Till now we considered how to analyze vectors in a given vector space. It is now desired
to study the relationships among different vector spaces over the same field. Let V1 and
V2 be two vector spaces over the same field F . A mapping t : V1 → V2 is called a linear
transformation or linear map or homomorphism if
(i) t(u + v) = t(u) + t(v)
(ii) t(αu) = αt(u)
∀u, v ∈ V1 , α ∈ F . A linear transformation from a vector space to itself is called a linear
operator.
The above two conditions can be merged together to get a single condition as follows.
t : V1 → V2 is called a linear transformation if t(αu + βv) = αu + βv ∀u, v ∈ V1 , α, β ∈ F .
If V1 and V2 are two vector spaces over the same field F and B = {b1 , b2 , . . . , bn } is a basis
in V1 then there exists a unique transformation t : V1 → V2 such that t(bi ) = b0i where b0i are
the vectors in V2 .
47
48 CHAPTER 5. LINEAR TRANSFORMATIONS
To prove the above, the following steps are adopted. Since B forms the basis in V1 ,
n
X
v= λi bi , v ∈ V1 (5.1)
i=1
n
!
X
t(v) = t λi vi
i=1
n
X
= (λi t(bi ))
i=1
n
X
= λi b0i
i=1
Pn
Since λi ’s need to be unique scalars, i=1 λi b0i shall also be unique. If u, v ∈ V1 and
α, β, λi , µi ∈ F ,
n
X
u= λi bi
i=1
n
X
v= µi b i
i=1
n
! n
!!
X X
t(αu + βv) = t α λi bi +β µi bi
i=1 i=1
n
!
X
=t (αλi + βµi )bi
i=1
n
X
= (αλi + βµi )b0i
i=1
n
X
= (α(λi b0i ) + β(µi b0i ))
i=1
n
! n
!
X X
=α λi b0i +β µi b0i
i=1 i=1
= αt(u) + βt(v)
= t(v)
=⇒ t0 (v) = t(v)
Hence, t : V1 → V2 is unique.
If A is an m × n matrix and a mapping tA is defined such that tA : F n×1 → F m×1
tA (X + Y ) = A(X + Y )
=A X +A Y
= tA (X) + tA (Y )
=⇒ tA (X + Y ) = tA (X) + tA (Y )
It is easy to recognize that ker(t) is nothing but the null space of t. Image of t, Im(t), is
defined as
Im(t) = {t(v) : v ∈ V1 } (5.3)
Sylvester’s law:
Problems
1. Let F be a field. Show that the mapping t : F 2 → F 3 given by
is a linear transformation.
3. Let V = R[x] be the vector space of all polynomials over field R and let D : V → V be
d
the mapping associating each polynomial f (x) to its derivative dx
(f (x)). Show that D
is a linear transformation.
4. Let V be a real vector space of all continuous functions from R into itself. Show that
the mapping T : V → V given by
Z x
T [f (x)] = f (t)dt ∀f (x) ∈ V, x ∈ R
0
5. Let C be the vector space of all complex numbers over the field of complex numbers
and let t : C → C be a mapping given by
t(x + iy) = x ∀x + iy ∈ C
t(x, y) = (x + y, x) ∀x, y ∈ R
is a linear transformation.
52 CHAPTER 5. LINEAR TRANSFORMATIONS
(ii) t : R3 → R2 defined by
t(x, y, z) = (x + y + z, 2x − 3y + 4z)
(iii) t : R3 → R2 defined by
(ii) t : R2 → R2 defined by
t(x, y) = (x2 , y 2 )
(iii) t : R3 → R2 defined by
t(x, y, z) = (x + 1, y + z)
(iv) t : R2 → R2 defined by
t(x, y) = (xy, y)
9. Let V = Rn×n be a vector space of n × n matrices and let M be a fixed non-null matrix
in V . Which of the followings is a linear transformation?
(i) t(A) = M A
(ii) t(A) = M A + A M
(ii) t(A) = A M − M A
(iv) t(A) = M + A
where t : V → V .
53
12. Write the nth order ordinary differential equation given below in operator form and
identify whether the operator is linear.
dn u dn−1 u
an + a n−1 + · · · + a0 = f (x)
dxx dxn−1
13. Let t : R2 → R2 be a linear transformation for which t(1, 2) = (2, 3) and t(0, 1) = (1, 4).
Find the formula for t and find t(x, y).
14. Let B = {(−1, 0, 1), (0, 1, −1), (1, −1, 1) be a basis for R3 (R) and t : R3 → R3 be a
linear transformation such that
t(−1, 0, 1) = (1, 0, 0)
15. Let t : R3 → R3 be such that t(1, 2, 3) = (1, 0, 0), t(1, 2, 0) = (0, 1, 0) and t(1, −1, 0) =
(0, 0, 1). Find t(a, b, c) ∀(a, b, c) ∈ R3 .
t(x, y, z) = (x + 2y − z, y + z, x + y − 2z)
(ii) t : R4 → R3
(iii) t : R3 → R2
t(x, y, z) = (x + y + z, 2x + 2y + 2z)
Determine
(i) basis for range of tA
(ii) basis for kernel of tA
(iii) rank and nullity of tA
22. Let C be the field of complex numbers and let t : C3 → C3 be a mapping given by
26. The set B = {e3t , te3t , t2 e3t } is an ordered basis of the vector space V of all functions
f : R → R. Let D be the differential operator defined as d/dt. Find the matrix
representation of D relative to the basis B.
Let V be a real vector space. An inner product on V is a function <, >: V × V → R which
assigns each ordered pair (u, v) ∈ V × V to a real number < u, v > in such a way that the
following axioms are satisfied.
(i) Linearity:
∀u1 , u2 , v ∈ V, a, b ∈ R (6.2)
(ii) Symmetry:
< u, v >=< v, u > (6.3)
A vector space equipped with an inner product is called an inner product space. The above
conditions have been laid down for real vector spaces. As an extension, the most generalized
conditions can be written as given below.
(i) Linearity:
59
60 CHAPTER 6. INNER PRODUCT SPACES
(ii) Symmetry:
< u, v >= < v, u > (6.8)
where overbar represents the complex conjugate. From linearity and symmetry, the following
property follows.
< u, αv >= α < u, v > (6.11)
There can be several functions which can satisfy the conditions given above for an inner
product. A standard inner product is defined as follows.
n
X
< u, v > = ui v i (6.12)
i=1
Z b
< f, g > = f (x)g(x)dx (6.13)
a
Two vectors u and v are said to be orthogonal if their inner product is zero.
Problems
1. Consider the vector space Rn . Prove that Rn is an inner product space with the inner
product defined by
< u, v >= u1 v1 + u2 v2 + · · · + un vn
2. Consider the vector space C[a, b] of all continuous functions defined on the interval
[a, b]. Prove that the following operation makes it an inner product space.
Z b
< f, g >= f (t)g(t)dt, f (t), g(t) ∈ C[a, b]
a
u = (a1 , a2 ), v = (b1 , b2 ) ∈ R2
4. Let V be a real vector space. Show that the sum of two inner products on V is an
inner product on V . Is the same true for the difference of two inner product? What
about positive multiples of an inner product?
u = (a1 , a2 ), v = (b1 , b2 ) ∈ R2
6. Let <, > be the standard inner product on R2 . If u = [1 2]T and v = [−1 1]T ∈ R2
then find w ∈ R2 satisfying
7. For each of the followings, determine whether the operation <, > makes the space an
inner product space.
(i) < u, v >= u1 v1 − 2u1 v2 − 2u2 v1 + 5u2 v2
62 CHAPTER 6. INNER PRODUCT SPACES
9. Let <, > be the standard inner product on R2 . If u = [1 3]T and v = [2 1]T ∈ R2
such that
< w, u >= 3 < w, v >= −1
then determine w.
10. For any u = [u1 u2 ]T and v = [v1 v2 ]T ∈ R2 , the following operation is defined.
h i 1 2 u
< u, v >= u1 v1 2
3 8 v2
Determine whether R2 is an inner product space with the inner product defined by the
above operation.
11. If f (x) = x and g(x) = e−ix then show that < f (x), g(x) >= < g(x), f (x) >.
12. Let A = [aij ] be a 2×2 matrix with real entries. For x, y ∈ R2×1 , let fA (x, y) = yT A x.
Show that fA is an inner product space on R2×1 iff AT = A, a11 > 0, a22 > 0 and |A| >
0.
13. Let C[−π, π] be the inner product space of all continuous functions defined on [−π, π]
with the inner product as the standard inner product. Verify whether sin(t) and cos(t)
are orthogonal in such a space.
17. Let V be the vector space of all polynomials over R of degree ≤ 2 with the inner
R1
product defined as < f, g >= 0 f (t)g(t)dt. Find a basis of the subspace orthogonal
to the polynomial φ(t) = 2t + 1.
18. Consider the inner product space R4 with standard inner product defined. Let four
vectors in R4 be as given below.
v1 = [1 1 0 − 1]T , v2 = [1 2 1 3]T
19. Let f (x) belong to a space of continuous functions with the standard inner product
defined. If f (x) is expressed as an infinite series as follows.
∞
X
f (x) = an sin(nπx)
n=1
φn (x) = exp(2πinx), n ∈ I, 0 ≤ x ≤ 1
21. Consider a piecewise continuous function f (x) defined on the interval [−c, c] with a
period 2c. The function is to be represented as
∞
a0 X nπx nπx
f (x) = + an cos + bn sin
2 n=1
c c
23. Apply Gram-Schmidt orthogonalization to the basis B = {[1 0 1]T , [1 0 −1]T , [0 3 4]T }
of an inner product space R3 to obtain an orthogonal and an orthonormal basis.
24. Let B be a set of vectors in R3 with standard inner product defined. For the follow-
ings, identify whether B is a basis. Is the basis orthonormal? If not then obtain an
orthonormal basis.
(i) B = {[2 0 1]T , [3 −1 5]T , [0 4 2]T }
(ii) B = {[1 0 0]T , [1 1 0]T , [1 1 1]T }
√
d= < v, v > (7.1)
The above inequality is quadratic in α. We want to know the value of α for which the above
equation attains a minima. From the derivative test,
− < u, v >
α= (7.5)
< v, v >
< u, v >2
=⇒ < u, u > − ≥0 (7.6)
< v, v >
=⇒ < u, u >< v, v >≥< u, v >2 (7.7)
65
66 CHAPTER 7. NORM AND METRIC SPACES
Now consider
The inequality given by (7.11) is called the triangle inequality. This inequality can easily be
identified as a case of the definition of “distance” between vectors. Norm of a vector can be
identified as the distance of that vector from the zero vector. We can now define a space,
called the “metric space” if the followings are satisfied.
(i) Positivity
d(u, v) ≥ 0 (7.12)
(ii) Symmetry
d(u, v) = d(v, u) (7.13)
L2 space:
Consider a space X[a, b] of functions defined in the interval [a, b] such that f : [a, b] → C and
Z b
|f (x)|2 dx < ∞ (7.17)
a
α, β ∈ C
Therefore, if f, g ∈ L2 (a, b) then all linear combinations in the space also belong to the space.
68 CHAPTER 7. NORM AND METRIC SPACES
Problems
1. Prove the following (in)equalities.
(i) ||u + v||2 = ||u||2 + ||v||2
(ii) | ||u|| − ||v|| | ≤ ||u − v||
(iii) ||u + v||2 + ||u − v||2 = 2(||u||2 + ||v||2 )
(iv) ||w − u||2 + ||w − v||2 = 12 ||u − v||2 + 2||w||2 − 12 ||u + v||2
(v) ||u − v|| ≤ ||u − w|| + ||w − v||
2. For each of the followings, determine whether they belong to L2 and calculate the norm
if it is defined.
1
1, 0 ≤ x ≤
(i) f (x) = 2
1
0, ≤x≤1
2
1
(ii) f (x) = √x , 0 ≤ x ≤ 1
(iii) f (x) = √1 , 0 ≤ x ≤ 1
x3
1
(iv) f (x) = x
,1 ≤ x ≤ ∞
3. Show that the infinite set of functions {1, cos(x), cos(2x) . . . sin(x), sin(2x) . . . } is or-
thogonal in real inner product space L2 (−π, π).
{einx : n ∈ Z}
5. Verify the CBS inequality for the following functions on [0, 1].
f (x) = 1
g(x) = x
6. Determine which of the following functions belong to L2 (0, ∞) and calculate the norm
in cases it is defined.
69
7. Determine the real values of α for which xα lies in (a) L2 (0, 1) and (b) L2 (1, ∞).
8. Define a function f (x) ∈ L2 (−1, 1) such that < f (x), x2 + 1 >= 0 and ||f (x)|| = 2.
70 CHAPTER 7. NORM AND METRIC SPACES
Chapter 8
Adjoint operators
a11 a12
A=
a21 a22
u1
u=
u2
v1
v=
v2
a11 u1 + a12 u2
Au=
a21 u1 + a22 u2
71
72 CHAPTER 8. ADJOINT OPERATORS
then
1
d2 g
Z
< Df, g > = f + αg dx (8.2)
0 dx2
< Df, g > =< f, D∗ g > (8.3)
Hence, in this case, we have the differential operator given by D which is self-adjoint with
the specified boundary conditions.
We showed the above protocol for obtaining the adjoint of an operator when the operator
was a matrix or a differential operator. The adjoint of a matrix operator can be used to
check the existance and uniqueness of solutions of A x = b.
Problems
1. Determine the adjoint of the following matrix operators. Check whether they are self-
adjoint.
i 0
(i) A =
i 1
1 −2i
(ii) A =
3 i
3 2+i 4
(iii) A = 2 − i −i
2
2 i 1
−4 + 5i 2 + 2i 1 − 2i
(iv) A = 2 + 2i −1 + 8i −2 − 2i
4 + 4i −2 − 2i −4 + 5i
d
2. The momentum operator in Quantum Mechanics is given as p̂ = −i dx . Check if this
operator is self-adjoint for a region [a, b].
−h2 d2 ψ
+ V̂ ψ = Eψ
8π 2 m dx2
As a background study, find out the meaning and significance of all the quantities that
appear in the above equation. Check whether the operator in the above equation is
(a) linear, (b) self-adjoint.
4. Consider the following operator with homogeneous boundary conditions as shown be-
low.
d2
L̂ = + 1, x ∈ [0, π]
dx2
u(0) = u(π) = 0
d2 d
L̂ = p2 2
+ p1 + p0
dx dx
6. Identify the solvability condition and determine the range space using the alternative
theorem for the following set of equations.
x 1 + x 2 + x 3 = b1
2x1 − x2 + x3 = b2
x1 − 2x2 = b3
7. Using the alternative theorem, determine the solvability condition for the following set
of equations.
x1 − x2 + 2x3 = 3
2x1 + x2 + 6x3 = 2
x1 + 2x2 + 4x3 = −1
x1 − x2 + 3x3 + 2x4 = 2
3x1 + x2 − x3 + x4 = −3
10. Solve all relevant problems of chapter 4 using Fredholm’s alternative theorem.
13. Show that the product of two self-adjoint operators is also a self-adjoint operator iff
the two operators commute.
Eigenvalue problems
Consider a family of two first order ordinary differential equations shown below.
dx
= ax + by (9.1)
dt
dy
= cx + dy (9.2)
dt
The family of equations shown above can be written as a single matrix equation as shown
below.
d x a b x
= (9.3)
dt y c d y
For the above equation, the zero vector [0 0]T is always a solution, which is a trivial
solution. In order to know whether non-trivial solutions exist for this system, one needs to
solve for the homogeneous system.
a b x 0
= (9.4)
c d y 0
77
78 CHAPTER 9. EIGENVALUE PROBLEMS
d λt
u0 (t) = (e v)
dt
= λeλt v
= eλt (λv)
= eλt (A v)
= A(eλt v)
n
X
u(t) = ci eλi t vi (9.5)
i=1
The system of equations considered previously was a homogeneous system of the form
u0 = A u. Now we consider the solutions of non-homogeneous initial value problems of
the form u0 = A u + b(t) with an initial condition u(0) = u0 . For solution of such systems,
we make use of similarity transformation.
Similar matrices:
If P is any non-singular matrix such that P −1 A P = B then A and B are said to be similar
matrices. For similar matrices A and B consider the following operations.
B = P −1 A P
=⇒ B P −1 = P −1 A (P P −1 )
=⇒ B P −1 = P −1 A
=⇒ B P −1 u = P −1 A u
79
If u is an eigenvector of A then
B P −1 u = P −1 (λu)
=⇒ B (P −1 u) = λ(P −1 u)
=⇒ B v = λv
where v = P −1 u. The above equation means that if λ and u are the eigenvalue and eigen-
vector of A, respectively, then λ will also be the eigenvalue of the similar matrix B with the
corresponding eigenvector as P −1 u.
Diagonalization of matrices:
For similarity transformation, B = P −1 A P . Consider a matrix P whose columns are made
up of eigenvectors of A.
A P = A[v1 | v2 | . . . | vn ]
= [A v1 | A v2 | . . . | A vn ]
= [λ v1 | λ v2 | . . . | λ vn ]
=P Λ
where
λ1 0 0 . . . 0
0 λ2 0 . . . 0
Λ=. (9.6)
.
0 0 0 . . . λn
Power of matrices: nth power of a matrix can be easily determined using similarity transfor-
80 CHAPTER 9. EIGENVALUE PROBLEMS
mation. As seen before, when the matrix P is made of the eigenvectors of A then
P −1 A P = Λ
=⇒ A = P Λ P −1
=⇒ An = (P Λ P −1 )n
=⇒ An = (P Λ P −1 )(P Λ P −1 ) . . . (P Λ P −1 )
=⇒ An = P Λn P −1
Inverse of a matrix:
Following the above method, it is easy to determine the inverse of a matrix as shown below.
A = P Λ P −1
=⇒ A−1 = (P Λ P −1 )−1
=⇒ A−1 = P Λ−1 P −1
If an n × n matrix does not have n linearly independent eigenvectors then there exists a
non-singular matrix P such that
P −1 A P = J (9.7)
where J is called the Jordan matrix and has the following form.
J 0 0 ... 0
1
0 J2 0 ... 0
J = . (9.8)
.
0 0 0 . . . Jn
where J i ’s are called Jordan blocks in which eigenvalues appear on the diagonal, 1’s are on
the first superdiagonal and the rest of the elements are all zero. As an example, a 3 × 3
Jordan block can be written as follows.
λ 1 0
J = 0 λ 1
0 0 λ
81
Hence, the number of Jordan blocks equals the number of linearly independent eigenvectors
of the matrix.
Now we look into the procedure of determining the generalized eigenvectors. We take the
case of a 3 × 3 matrix. Let us first consider the case of a system with only one eigenvector
available. The following holds true.
P = [v | q1 | q2 ]
A P = [A v | A q1 | A q2 ]
=⇒ A P = [λ v | A q1 | A q2 ]
82 CHAPTER 9. EIGENVALUE PROBLEMS
λ 1 0
P J = [v | q1 | q2 ] 0 λ 1
0 0 λ
=⇒ P J = [λ v | v + λ q1 | q1 + λ q2 ]
For J = P −1 A P , we equate A P = P J.
[λ v | A q1 | A q2 ] = [λ v | v + λ q1 | q1 + λ q2 ]
Hence, q1 and q2 can be obtained as the solutions of the above non-homogeneous problem.
Let us now consider the case where two eigenvectors are available.
A P = [A v1 | A v2 | A q]
=⇒ A P = [λ v1 | λ v2 | A q]
λ 0 0
P J = [v1 | v2 | q] 0 λ 1
0 0 λ
=⇒ P J = [λ v1 | λv2 | v2 + λ q]
=⇒ [λ v1 | λ v2 | A q] = [λ v1 | λv2 | v2 + λ q]
=⇒ (A − λI)q = v2
The generalized vector can be obtained by solving the above equation. By exchanging v1
and v2 , another generalized vector can be obtained by solving the following equation.
(A − λI)q = v1
Now we use all the concepts developed till now for solving non-homogeneous initial value
problems. Given
d
u(t) = A u(t) + b(t)
dt
d −1
=⇒ (P u) = P −1 A u + P −1 b
dt
d −1
=⇒ (P u) = (P −1 A P ) (P −1 u) + P −1 b
dt
The above equation can be solved using usual method except that we do not know the
exponential of a matrix. This can be determined as follows.
(Λt)2 (Λt)3
e−Λ t = I + Λt + + ... (9.12)
2! 3!
eλ1 t 0 0 ... 0
0 e λ2 t 0 ... 0
−Λ t
=⇒ e = . (9.13)
.
λn t
0 0 0 ... e
84 CHAPTER 9. EIGENVALUE PROBLEMS
Problems
1. Find thegeneral
solutions for each of the following problems.
1 2
(i) u0 = u
0 3
1 2
(ii) u0 = u
3 6
1 2
(ii) u0 = u
1 0
1 2
(iv) u0 = u
3 −3
where bc > 0
d2 x dx
2
+ b + kx = 0
dt dt
determine all the values of b and k for which the system has real and distinct eigen-
values. Find the general solution and the solution which satisfies the initial condition
x(0) = 1.
a 1
4. Consider A = . Determine the value of the parameter a for which A has
0 1
repeated real eigenvalues.
d2 x dx
+ +x=0
dt2 dt
d2 x dx
+ 2 +x=0
dt2 dt
a b
7. Consider u0 = u, where a + d 6= 0 and ad − bc 6= 0. Determine the general
c d
solution of the system.
where b ≥ 0 and k ≥ 0. For which values of k and b does the system have complex
eigenvalues? real and distinct eigenvalues? repeated eigenvalues?
10. Consider
the following matrices.
7 −16 −8
A = −16 7
8
−8 8 −5
−2 0 −2i
A= 0 1 0
2i 0 −2
(i) Diagonalize the above matrices.
(ii) Determine the square of the above matrice using similarity transformation. Verify
using multiplication.
86 CHAPTER 9. EIGENVALUE PROBLEMS
du
=Au
dt
u(0) = [1 1 1]T
5 −3 −2
A = 8 −5 4
−4 3 3
d
u(t) = A u(t) + b(t)
dt
d
u(t) = A u(t) + b(t)
dt
where
−i i 0
A = i −i 0
0 0 −i
√
2t
√
b = 2t
−t
e
√1
2
u(0) = √1
2
1
87
d2 u du
2
+ 5 + 6u = e−t
dt dt
with
du
= (u)0 = 1
dt 0
15. Convert the following initial value problem to a matrix equation and solve using simi-
larity transformation and otherwise.
d2 u du du
+ 3 + 2u = 0; u(0) = 1; (0) = 3
dt2 dt dt
16. Verify that {[e2t e2t e2t ]T , [e−t 0 e−t ]T , [−e−t e−t 0]T } is a solution set to the
system
0 1 1
dx
= 1 0 1 x
dt
1 1 0
Sturm-Liouville theory
Numerous classes of Sturm-Liouville problems are identified and solutions derived in the
following sections
89
90 CHAPTER 10. STURM-LIOUVILLE THEORY
Bessel equation
The Bessel equation is given as follows.
d2 y dy
x2 2
+ x + (x2 − n2 )y = 0 (10.5)
dx dx
where n is a non-negative parameter. The solutions to the Bessel equation are called Bessel
functions. It can be seen that the equation is an ordinary differential equation with variable
coefficients. Such equations can be solved using a technique called series solution due to
Frobenius. We first recast the equation to the following form.
d2 y 1 dy n2
+ + (1 − )y = 0 (10.6)
dx2 x dx x2
∞
X
k+r−2 k+r
=⇒ (k + r + n)(k + r − n)x +x =0 (10.10)
r=0
The above equation is an important result and a major landmark in obtaining solutions of
the type discussed here. We compare the coefficients of the above polynomial equation.
r = 0, a0 (k + n)(k − n) = 0 =⇒ k = ±n
−ar
=⇒ ar+2 = (10.11)
(k + r + 2 + n)(k + r + 2 − n)
The above equation is one of the most important results of this solution procedure and is
called the recurrence relation. This relation relates different coefficients appearing in the
solution. From a1 = 0 and the above recurrence relation, we can write,
a1 = a3 = a5 = · · · = 0
For k = +n,
−ar
ar+2 =
(2n + r + 2)(r + 2)
−a0
=⇒ a2 =
(2)(2n + 2)
−a0
=⇒ a2 = 2
2 (n + 1)
−a2 −a2
=⇒ a4 = = 2
(4)(2n + 4) 2 · 2 (n + 2)
a0
=⇒ a4 = 4
2! · 2 (n + 2)(n + 1)
−a4 −a4
=⇒ a6 = =
(6)(2n + 6) 3 · 22 (n + 3)
−a0
=⇒ a6 = 6
3! · 2 (n + 3)(n + 2)(n + 1)
From the above expressions of the coefficients, we can write the solution to the Bessel equa-
tion as follows.
x2 x4 x6
k
y = a0 x 1 − + − − ...
22 (n + 1) 2!24 (n + 2)(n + 1) 3! · 26 (n + 3)(n + 2)(n + 1)
Therefore, for k = +n, the solution to the Bessel equation can be written as follows.
∞
n
X (−1)r x2r
y = a0 x (10.12)
r=0
r!22r (n + 1)(n + 2) . . . (n + r)
We have determined the solution of the Bessel equation. The only unknown is a0 . For a
specific choice of a0 , we get a clean solution of the equation, as given below.
1 1
a0 = = (10.13)
2n Γ(n + 1) 2n n!
∞
1 n
X (−1)r x2r
=⇒ y = x (10.14)
2n Γ(n + 1) r=0 r!22r (n + 1)(n + 2) . . . (n + r)
92 CHAPTER 10. STURM-LIOUVILLE THEORY
∞
X (−1)r x2r+n
=⇒ y = (10.15)
r=0
22r+n r!(r + n)!
This series is called Bessel’s function of the first kind of order n, and is conventionally denoted
s Jn (x). With k = +n and − n, the Bessel’s functions can be written as follows.
∞ 2r+1
r x 1
X
Jn (x) = (−1) (10.16)
r=0
2 r!(r + n)!
∞ 2r+1
r x 1
X
J−n (x) = (−1) (10.17)
r=0
2 r!(r − n)!
The following plots show the first two Bessel functions J0 (x) and J1 (x).
Figure 10.1: Bessel functions of the first kind J0 (x) and J1 (x)
In the previous discussion, we used the concept of gamma function. Let us remind ourselves
of the gamma function which is defined as follows.
Z ∞
Γ(x) = e−t tx−1 dt (10.18)
0
Γ(x + 1) = x! (10.19)
We discussion only about the Bessel functions of the first kind. Bessel functions of the second
kind can be derived from Bessel functions of the first kind. This is left as a background study
and an exercise.
93
Legendre equation
The Legendre equation is given as
d2 y
2 dy
(1 − x ) 2 − 2x + λy = 0; − 1 < x < 1 (10.20)
dx dx
Following the Frobenius method, the series solution to the above equation can be represented
as an infinite series given below.
∞
X
y= ar xk+r (10.21)
r=0
∞
dy X
=⇒ = (k + r)ar xk+r−1 (10.22)
dx r=0
2 ∞
dy X
=⇒ = (k + r − 1)(k + r)ar xk+r−2 (10.23)
dx2 r=0
∞
X
k+r−2 k+r
=⇒ ar (k + r − 1)(k + r)x + λ − (k + r)(k + r + 1) x =0 (10.24)
r=0
The above equation puts us in a position to write the recurrence relation as given below.
(k + r)(k + r + 1) − λ
ar+2 = ar (10.25)
(k + r + 1)(k + r + 2)
Therefore, a knowledge of a0 and a1 will yield the complete series through the recurrence
relation. When k = 0 and λ = n(n + 1), n ∈ I+ ∪ 0, For such a case
Therefore, the infinite series solution reduces to a polynomial for the above choice of param-
eters. In such a case,
r(r + 1) − n(n + 1)
ar+2 = ar
(r + 1)(r + 2)
(r − n)(r + n + 1)
ar+2 = ar (10.26)
(r + 1)(r + 2)
94 CHAPTER 10. STURM-LIOUVILLE THEORY
Hence, the solution to the Legendre equation takes the following form.
n(n + 1) 2 (n − 2)n(n + 1)(n + 3) 4
y = a0 1 − x + x + ...
2! 4!
(n − 1)(n + 2) 3 (n − 3)(n − 1)(n + 2)(n + 4) 5
+ a1 x − x + x + ...
3! 5!
If u0 (x) and u1 (x) are the power series with even only and odd only powers, respectively,
then the above solution can be written as follows.
It can be seen from the above analysis that the solution depends upon n. For each n, a pair
of linearly independent solutions is obtained.
n = 0, u0 (x) = 1
1 1
u1 (x) = x + x3 + x5 + . . .
3 5
1
n = 1, u0 (x) = 1 − x2 − x4 + . . .
3
u1 (x) = x
n = 2, u0 (x) = 1 − 3x2
2 1
u1 (x) = x − x3 − x5 + . . .
3 5
n = 3, u0 (x) = 1 − 6x2 + 3x4 + . . .
5
u1 (x) = x − x3
3
.
95
It can be observed that for every n, one of the solutions is a polynomial solution while the
other is an infinite series. One is generally interested in the polynomial solutions. Such
solutions can be written in the following form.
(2n)! 1 · 3 · 5 · · · (2n − 1)
an = n 2
= (10.29)
2 (n!) n!
The resulting polynomial is called Legendre polynomial of degree n, denoted as Pn (x). The
general expression for Legendre polynomial of degree n can be written as follows.
[n/2]
1 X k (2n − 2k)!
Pn (x) = n (−1) xn−2k (10.30)
2 k=0 k!(n − k)!(n − 2k)!
where
n/2, if n is even
[n/2] =
(n − 1)/2, if n is odd
P0 (x) = 1 P1 (x) = x
1 1
P2 (x) = (3x2 − 1) P3 (x) = (5x3 − 3x)
2 2
1 1
P4 (x) = (35x4 − 30x2 + 3) P5 (x) = (63x5 − 70x3 + 15x)
8 8
Laguerre equation
Laguerre equation is given as follows.
d2 y dy
x 2 + (1 − x) + λy = 0 (10.31)
dx dx
Using the series solution technique, we can write the followings.
∞
X
y= ar xk+r (10.32)
r=0
∞
dy X
=⇒ = (k + r)ar xk+r−1 (10.33)
dx r=0
∞
d2 y X
=⇒ = (k + r − 1)(k + r)ar xk+r−2 (10.34)
dx2 r=0
(k + r)2 ar = (k + r − λ − 1)ar−1 = 0
This polynomial solution is called the Laguerre polynomial. From the previous recurrence
relation, we can write
r2
ar−1 = ar
r−λ−1
We can substitute r = λ, λ−1, λ−2 . . . in the above recurrence relation to obtain the general
coefficient of the polynomial solution. The expression for aλ−n can be written as follows.
(λ!)2
n
aλ−n = (−1) aλ (10.41)
((λ − n)!)2 (n!)
Hence, the solution of the equation as the Laguerre polynomial can be written as
λ
(λ!)2
X
n
Lλ (x) = (−1) aλ xλ−n (10.42)
n=0
((λ − n)!)2 (n!)
L0 (x) = 1 L1 (x) = 1 − x
Hermite equation
The following differential equation is called the Hermite equation.
d2 y dy
2
− 2x + 2λy = 0 (10.43)
dx dx
Using the series solution with k = 0, we get the followings.
∞
X
y= ar x r (10.44)
r=0
∞
dy X
=⇒ = rar xr−1 (10.45)
dx r=0
∞
d2 y X
=⇒ = (r − 1)rar xr−2 (10.46)
dx2 r=0
∞
X ∞
X ∞
X
=⇒ (r − 1)rar xr−2 − 2x rar xr−1 + 2λ ar x r = 0 (10.47)
r=0 r=0 r=0
Hence, the solution to the Hermite equation can be written as the summation of the even
and odd series.
y = a0 yeven + a1 yodd (10.49)
Like before, we can see that the infinite series truncates to polynomials when λ is an integer
(say N ) in which case, the recurrence relation can be written as follows.
2(r − N − 2)
ar = ar−2
r(r − 1)
It can be seen that the series will terminate when r reaches a values of N + 2. Hence,
we get polynomial solutions. Following the previous analysis, we list out first few Hermite
polynomials below and leave it for you to plot and see their nature.
H0 (x) = 1
H1 (x) = 2x
H2 (x) = 4x2 − 2
Problems
1. Show that the Sturm-Liouville operator is:
(i) a linear operator
(ii) self-adjoint
Compare the result with the Legendre polynomials Pi (x), i = 1 − 5. Show that there
is a linear relationship between the two set of polynomials.
101
x = cosθ
y(cosθ) = u(θ)
10. Verify whether the following is a correct generating function for Hermite polynomials.
2 dn −x2
Hn (x) = (−1)n ex (e )
dxn
13. Verify whether the following is a correct generating function for Laguerre polynomials.
n
1 n k
X
k
Ln (x) = (−1) x
k=0
k! k
where
n
n!
=
k k!(n − k)!
14. Verify whether the following is a correct generating function for Laguerre polynomials.
dn n −x
Ln (x) = ex (x e )
dxn
15. Show that Bessel functions Jn and J−n are not linearly independent if n ∈ N0 . Check
the same for the case of n ∈ I.