Review of Linear Algebra
Review of Linear Algebra
Associate Professor
Department of EE
NIT Rourkela
Definition
Field: A field F is an object consisting of a set of elements, and two
binary operations
addition(+)
multiplication(.)
such that the following axioms are obeyed. Addition is:
1 associative (α + β) + γ = α + (β + γ) ∀ α, β, γ ∈ F.
2 commutative α + β = β + α.
3 ∃ identity element 0 : α + 0 = α.
4 ∃ inverse element : ∀ α ∃ − α s.t. α + (−α) = 0.
Definition (contd..)
Similarly, Multiplication is:
5 associative (α.β).γ = α.(β.γ)∀α, β, γ ∈ F.
6 commutative α.β = β.α.
7 ∃ identity element 1 : α.1 = α.
8 ∃ inverse element : ∀ α 6= 0 ∃ α−1 s.t. α.α−1 = 1.
and finally, (.) distributes over (+)
9 α.(β + γ) = α.β + α.γ.
10 (β + γ).α = β.α + γ.α.
Examples
First of all, it’s easy to see that the rational numbers satisfy all the
field axioms, so Q is a field. In fact, verifying the field axioms for Q
simply boils down to the basic arithmetic properties of the integers:
associativity, commutativity and distributivity and the existence of 0
and 1.
Similarly, the set of real numbers R and complex numbers C are the
fields.
The integers Z are not a field, since field axiom 8 isn’t satisfied by Z.
Indeed, the only integers which have multiplicative inverses are ±1.
Definition
Vector Spaces (or linear spaces)
A vector space (V, F) is a set of vectors V, and a field of scalars F, and
two binary operations:
vector addition (+)
scalar multiplication (.) (multiplication of vectors by scalars)
such that:
Addition + : V × V : (x, y ) → x + y ∀ x, y ∈ V
1 associative (x + y ) + z = x + (y + z).
2 commutative x + y = y + x.
3 ∃ identity 0 (zero vector) : x + 0 = 0 + x = x.
4 ∃ inverse, ∀x ∈ V, ∃(−x) s.t. x + (−x) = 0.
Definition (contd..)
Scalar multiplication . : F × V : (α, x) → αx ∀α ∈ F, x ∈ V
5 (αβ)x = α(βx).
6 1.x = x.
7 0.x = 0.
a and, distributive laws:
8 (α + β)x = αx + βx ∀x ∈ V, ∀ α, β ∈ F .
9 α(x + y ) = αx + αy αx, y ∈ V, ∀α ∈ F.
a
Here 1 and 0 are the multiplicative and additive identities of the field.
Examples
The first example is the obvious one: if n ≥ 1, then Rn with the usual
component-wise addition and scalar multiplication is a real vector
space, that is a vector space over the field R. We usually call Rn real
n-space.
More generally, for any field F and n ≥ 1, the set F n of all n-tuples
(x1 , x2 , · · · , xn )> of elements of F can be made into a vector space
over F in exactly the same way. That is,
Examples(contd.)
The set Pn of all polynomials
ÿ + aẏ + by = 0 (1)
Examples(contd.)
Let R2×2 denotes the set of 2 × 2 matrices. If addition and scalar
multiplication are defined to be the standard matrix operations, then
R2×2 is a vector space with the following operations: Consider
a11 a12
A= A ∈ R2×2
a21 a22
b11 b12
B= B ∈ R2×2 (2)
b21 b22
Examples(contd.)
Closure under addition:
a11 + b11 a12 + b12
A+B =
a21 + b21 a22 + b22
In general Rm×n , the set of m × n matrices are also a vector space with
standard matrix addition and scalar multiplication.
Examples(contd.)
a set that is not a vector space:
Let
V = {A ∈ R2×2 : |A| =
6 0}
Reason
This is not a vector space, it is not closed under matrix addition. Both I
and −I are in V . Yet I + (−I ) ∈ / V.
Definition
Let (V, F) be a linear space and W a subset of V. Then (W, F) is called
a subspace of (V, F) if (W, F) is itself a vector space.
How to check if something is a subspace:
Verify that W is a subset of V (thus W inherits the vector space
axioms of V).
Verify closure under vector addition and scalar multiplication, i.e.,
∀w1 , w2 ∈ W, ∀α1 , α2 ∈ F then α1 w1 + α2 w2 ∈ W.
Notice that every subspace of a vector space contains the zero vector
0 (why?). In fact, 0 is itself a subspace, called the trivial subspace.
Example
Consider the vector space R2 of all points (x, y ), where x, y ∈ R.
Let L be a line through the origin 0.
Suppose that L is represented by the equation αx + βy = 0, in other
words
L = {(x y )> ∈ R2 : αx + βy = 0}
Is L is a vector space?
YES! The L forms a vector space over R. In fact every line in R2
through the origin is vector space over R.
Example
A plane passing through origin is a subspace of R3 . In other words,
W ⊂ R3 such that
W = {(x y z)> ∈ R3 : ax + by + cz = 0}
Test
0
Suppose 0 ∈ W . Now, a.0 + b.0 + c.0 = 0. So W is nonempty.
0
x1 x2
Closure under addition: Let y1 and y2 ∈ W .
z1 z2
Example
Then,
Now,
Example
Let F be a field and n be an integer with n > 2, and B ∈ Fn×n , show that
W = {A ∈n×n : AB = BA}
is a subspace.
Solution
To show null vector included in subspace W
0∈W
Solution
To show subspace W is closure under addition and scalar multiplication.
Give A1 , A2 ∈ W , c ∈ F. We have
(A1 + cA2 )B
= A1 B + cA1 B
= BA1 + CBA2
= B(A1 + cA2 )
Definition
Suppose (V, F) is a linear space. Let S = {v1 , v2 , · · · vr } be a finite set of
vectors in V. Any vector of the form
α1 v1 + α2 v2 + · · · + αr vr (3)
Example
1 1
Let V = R3 . S = {u1 , u2 }, where u1 = −1 and u2 = 0 .
2 1
1
Then u = −1 is a linear combination of vector space in S.
3
It is easy to verify
u = 2u1 + (−1)u2 .
Example
Let V = R2×3 . S = {A1 , A2 }, where
1 −1 1 1 1 1
A1 = ; A2 =
0 1 1 0 1 0
Then
1 −3 1
A=
0 1 2
A = 2A1 + (−1)A2 .
Definition
Suppose (V, F) is a linear space. Let S = {v1 , v2 , · · · vn } be a subset of V.
We say S is spanning set of V or that S spans V if for every vector v ∈ V
can be written as a linear combination of vectors in S.
Example
1 0
The set S = {e1 , e2 }, where e1 = and e2 = is a spanning set
0 1
2 x
of R because every vector v = in R2 could be written as a linear
y
combination of e1 and e2 as follows:
v = xe1 + ye2 .
Example
Prove that the matrices
−2 0 3 0
,
0 1 0 −1
which gives
−2c + 3d = a
c −d =b
Dr. Manas Kumar Bera (Associate ProfessorDepartment
Review
of EENIT
of Linear
Rourkela)
Algebra February 11, 2025 25 / 82
Spanning Set
Solution
Solving this system we see that there will always be a solution for any a
and b. Thus any diagonal matrix can be written as a linear combination of
the matrices above so theses two matrices span the set of all diagonal
2 × 2 matrices.
α1 v1 + α2 v2 + · · · + αp vp
=⇒ α1 = α2 = αp = 0
Definition
Suppose (V, F) is a linear space. The set of vectors {v1 , v2 , · · · vp }, vi ∈ V
is said to be linearly independent iff
α1 v1 + α2 v2 + · · · + αp vp = 0 (4)
=⇒ α1 = α2 = αp = 0 (5)
where αi ∈ F.
The set of vectors is said to be linearly dependent iff there ∃ scalars
α1 , α2 , · · · , αp ∈ F at least one of which is not zero such that
α1 v1 + α2 v2 + · · · + αp vp = 0. (6)
Example
Let R3 is the underlying vector space. Is the set S = {v1 , v2 , v3 }, where
1 1 0
v1 = 1 , v2 = 0 , v3 = 1 (7)
0 −1 −1
linearly independent?
Answer
For the linear combination
α1 v1 + α2 v2 + α3 v3 = 0. (8)
find α’s.
solution
1 1 0 0
α1 1 + α2 0 + α3 1 = 0
0 −1 −1 0
1 1 0 α1 0
⇒ 1 0 −1
α2 = 0
0 −1 −1 α3 0
⇒ α1 = α2 = α3 = 0.
Example
Let R3 is the underlying vector space. Is the set S = {v1 , v2 , v3 }, where
1 1 3
v1 = 1 , v2 = 0 , v3 = 2 (9)
0 −1 −1
linearly independent?
Answer
For the linear combination
α1 v1 + α2 v2 + α3 v3 = 0. (10)
find α’s.
solution
1 1 3 0
α1 1 + α2 0 + α3 2 = 0
0 −1 −1 0
1 1 3 α1 0
⇒ 1 0
2 α2 = 0
0 −1 −1 α3 0
⇒ α1 = 2, α2 = 1, α3 = −1.
Definition
Suppose (V, F) is a linear space. Then a set of vectors
B = {b1 , b2 , · · · , bn } is called a basis of V if
1 B spans V.
2 B is a linearly independent set.
Example
1
We have already seen that the set S = {e1 , e2 }, where e1 = and
0
0
e2 = is a spanning set of R2 . It is also linear independent set.
1
Therefore S is basis for R2 .
Example
1
We have already seen that the set B = {e1 , e2 , e3 }, where e1 = 0 ,
0
0 0
e2 = 1 and e3 = 0 is a spanning set of R3 . It is also linear
0 1
independent set. Therefore B is basis for R3 .
Coordinate Representation
Coordinate Representation: Any vector x ∈ V may be written as a linear
combination of the basis vectors:
e.g.,
x =ξ1 b1 + ξ2 b2 + · · · + ξn bn (11)
Xn
= ξi bi
i=1
ξ1
..
and here, ξ = . ∈ F n is called the coordinate vector of x with respect
ξn
to the {bi }.
The ξi ’s are called the coordinates of x with respect to {bi }.
Facts
Fact: The ξi ’s are uniquely defined in terms of x and the {bi }.
Proof: Suppose not. Then ∃ξ, ξ 0 such that
x = ξ1 b1 + ξ2 b2 + · · · + ξn bn (12)
x= ξ10 b1 + ξ20 b2 + ··· + ξn0 bn (13)
which for ξi 6= ξi0 implies that the {bi } are linearly dependent contradicting
the assumption that {bi } is a basis.
Notes:
A basis 3
of
avector
space is not unique,
i.e.,
in R
1
1 0 0 1 0 0
0 1 0 , 1 1 0 are both bases.
0 0 1 0 1 1
2 A vector space (V, F) is said to be finite dimension n or to be n
dimensional, written dimV = n, if there exists linearly independent
vectors {b1 , b2 , · · · , bn } which span V . The set {b1 , b2 , · · · , bn } is
then called a basis for V .
Example
Consider the vectors α1 = (1, 0, −1), α2 = (0, −3, 4), α3 = (1, 2, −1) in R3
Solution
a
a) To show that B is a basis, let γ = b ∈ R3 . If
c
c1 α1 + c2 α2 + c3 α3 = γ has a unique solution, then {α1 , α2 , α3 } is a basis.
Augmented matrix for systems
c1 a 1 0 1 a
A c2 = b → 0 −3 2 b
c3 c −1 4 −1 c
1 0 1 a 1 0 1 a
≈ 0 −3 2 b ≈ 0 −3 2 b
0 4 0 a+c 0 0 8/3 a + c + 4/3b
Solution
b)
Definition
Let V and W be vector spaces over the same field F . A linear
operator is a mapping M : V → W such that
I (a) M(v1 + v2 ) = M(v1 ) + M(v2 ) (additivity)
I (b)M(αv1 ) = αM(v1 ) (homogeneity)
for all v1 , v2 ∈ V and all α ∈ F
Example: The following operator is linear:
M : Rn → Rm : v → Av where A ∈ R m×n
Linear operators on finite dimensional vector spaces are matrices with
the action of matrix-vector multiplication.
Example
Find the bases for the null space and the range space where
1 −1 1 −1
A = 4 −4 5 −2
2 −2 −1 −8
Solution:
rref(A)
1 −1 1 −1 1 1 1 −1 1 −1 0 −3
0 0 1 2 → 0 0 1 2 → 0 0 1 2
0 0 −3 −6 0 0 0 0 0 0 0 0
Example
Now
x1 − x2 − 3x4 = 0
x3 + 2x4 = 0
x1 = t + 3s
x2 = t
x3 = −2s
x4 = s
Example
Now, above can be written as
x1 1 3
x2 1
0
=t +s
x3 0 −2
x4 0 1
1 3
( )
1 0
Basis for N (A) =
0 , −2
;
0 1
( 1 1 )
FACT
The dimension of the row space and of the column space are called
the row rank of A and the column rank A respectively.
The rank of A is defined as the number of linearly independent
columns. It also equals the number of linearly independent rows.
Because of this fact, if A is m × n, then
rank(A) 6 min(m, n)
Example
−3 6 −1 1 −7
Example: 1 −2 2 3 −1 ∼
2 −4 5 8 −4 3×5
1 −2 0− 1 3
0 0 1 2 −2 = rref(A)
0 0 0 0 0 3×5
Row Space:
u1 = 1 −2 0 −1 3 ∈ R5 ; u2 = 0 ∈ R5 .
0 1 2 −2
−3 −1
Column space: V1 = 1 ∈ R3 ; V2 = 2 ∈ R3 .
2 5
R(A) = span{V1 , V2 }. rank(A) = 2.
Ax = y
where x ∈ Rn , y ∈ Rm and A = : Rn → Rm .
a1 · · · an
The matrices A and y are given and x is the unknown to be solved.
Thus the set actually consists of m equations and n unknowns.
The number of equations can be larger than, equal to, or smaller than
the number of unknowns.
What is the existence condition and general form of solution of
Ax = y ?
rank(A) = rank([A y ])
General Solution
Given an m × n matrix A and an m × 1 vector y , let xp be a solution
of Ax = y .
let k := n − rank(A) be the nullity of A. If A has rank n (full column
rank) or k = 0, then the solution xp is unique.
If k > 0, then for every real αi , i = 1, 2, · · · k, the vector
x = xp + α1 n1 + · · · + αk nk
Example
Find the solution of the following systems
4x + 2y − 2z + 3w = 1
2x + y − 2z + 5w = 1
Example
The augmented matrix for this system is
4 2 −2 3 1
2 1 −2 5 1
Example
The solution
x −1/2s + t + 1
y s
z = 7/2t + 3/2
w t
which also can be written as:
1 −1/2 1
0 1 0
3/2 + s 0 + t 7/2
0 0 1
FACT
Let us consider a unforced system
Norm of Vectors
A norm is a function which assigns to every vector x in a vector space
a real positive number which is represented as kxk.
For example, the length of a vector u ∈ R2 is obtained from the
Pythagorean theorem by computing the length of the hypotenuse of a
right triangle as shown in figure.
p
This measured length kuk = x 2 + y 2 is called the Euclidean norm
in R2 and there is an obvious extension to the higher dimension.
Definition
Let the field F be R or C. A linear space (V, F) is said to be a normed
linear space if ∃ a map : k.k : V → R+ satisfying the following axioms.
1 kv1 + v2 k ≤ kv1 k + kv2 k ∀v1 , v2 ∈ V.
2 kαv k = |α|kv k ∀α ∈ F, v ∈ V.
3 kv k = 0 ⇔ v = 0.
>
Examples: (F n , F) with x = x1 x2 · · ·
xn
kxk1 = ni=1 |xi | → `1 -norm.
P
1
Pn 2
kxk2 = 2
i=1 |xi | → `2 -norm.
1
Pn p
kxkp = i=1 |xi |p → `p -norm.
Example
1
1
Let x =
..
.
1
Pn×1
n
kxk1 = i=1 |xi |
=n
1
√
Pn 2
kxk2 = 2
i=1 |xi | = n
Example
3
Let x = 4 − 3i
1 3×1
Pn
kxk1 = i=1 |xi | = 3 + 5 + 1 = 9
1
√ √
Pn 2
kxk2 = 2
i=1 |xi | = 9 + 25 + 1 = 35
Definition
A vector x is said to be normalized if its Euclidean norm is 1 or
x T x = 1.
Note that x T x is scalar and xx T is n × n.
Two vectors x1 and x2 are said to be orthogonal if x1T x2 = x2T x1 = 0.
A set of vectors xi , i = 1, 2, · · · m, is said to be orthonormal if
(
T 0, if i =6 j
xi xj =
1, ifi = j.
Concepts
Notice that in ẋ = Ax , the matrix A operates on the vector x to give
the vector ẋ. This is basic function of any matrix-mapping one vector
into another vector. Generally the derived vector is different from the
source vector, both in magnitude and direction.
But there may be some special directions in the state space such that
if the vector x is in that direction, the resultant vector ẋ also lies
along the same direction. It only gets stretched or squeezed.
Any vector along these special directions are called eigenvectors and
the factor by which any eigenvector expands or contracts when it is
operated on by the matrix A, is called the eigenvalue.
Figure shows this definition of eigenvectors. If Ax is not collinear with
x after the transformation, as in Figure (a), x is not an eigenvector. If
Ax is collinear with x after the transformation, as in Figure (b), x is
an eigenvector.
Dr. Manas Kumar Bera (Associate ProfessorDepartment
Review
of EENIT
of Linear
Rourkela)
Algebra February 11, 2025 65 / 82
Eigenvalues and eigenvectors
Concepts
Definitions
Eigenvectors: The eigenvectors of the matrix A are all vectors, xi 6= 0,
which under the transformation A become multiples of themselves;
that is,
Axi = λi xi (16)
adj(λi I − A)
xi = (λi I − A)−1 0 = 0 (18)
det(λi I − A)
det(λi I − A) = 0 (19)
Eigenvalues:
λ 0 −3 1
det(λI − A) = −
0 λ 1 −3
λ + 3 −1
=
−1 λ + 3
= λ2 + 6λ + 8 = (λ + 2)(λ + 4) (20)
Dr. Manas Kumar Bera (Associate ProfessorDepartment
Review
of EENIT
of Linear
Rourkela)
Algebra February 11, 2025 69 / 82
Eigenvalues and eigenvectors
How to find Eigenvalues and eigenvectors
The eigenvalues are λ = −2, −4. Now we find an eigenvector for each
of these by finding a vector in N (λI − A).
With eigenvalue λ = −2, following can be written as (16),
−3 1 x1 x1
= −2
1 −3 x2 x2
or
−3x1 + x2 = −2x1
x1 − 3x2 = −2x2
from which x1 = x2 .
c
Thus, x = .
c
Dr. Manas Kumar Bera (Associate ProfessorDepartment
Review
of EENIT
of Linear
Rourkela)
Algebra February 11, 2025 70 / 82
Eigenvalues and eigenvectors
Example
Finding a basis for the null space is given by
1
−1
1
Example
The geometric multiplicity λ2 = −2 is q2 = 3 − 1 = 2.
Fact: GM of λi ≤ AM of λi .
x(t) = c1 e λ1 t v1 + c2 e λ2 t v2 + · · · + cn e λn t vn (21)
which implies
n
X
x(0) = ci vi (22)
i=1
−5 3
Consider a system ẋ = Ax; x ∈ Rn . Here A =
−2 2
The characteristics polynomial: λ2 + 3λ − 4 = (λ + 4)(λ − 1).
The eigenvalues are −4 and 1.
3 1
The eigenvectors are and
1 2
3 1
The general form of solution is x(t) = c1 e −4t + c2 et .
1 2
The C-H theorem states that the matrix A satisfies its own
characteristics polynomial, i.e.,
An + p1 An−1 + · · · + pn−1 A + pn I = 0
p(s) s 4 + 3s 3 + 2s 2 + s + 1
=
4(s) s 2 − 5s + 5
146s − 184
= s 2 + 8s + 27 + 2
s − 5s + 5
2
p(s) = (s + 8s + 27)4(s) + 146s − 184
Dr. Manas Kumar Bera (Associate ProfessorDepartment
Review
of EENIT
of Linear
Rourkela)
Algebra February 11, 2025 81 / 82
Cayley- Hamilton Theorem
Application
Let r (s) = 146s − 184 and from C-H theorem 4(A) = 0 which
implies,
p(A) = r (A)
⇒ p(A) = A4 + 3A3 + 2A2 + A + I
= 146A − 184