Basic Opertaion Matrices
Basic Opertaion Matrices
Observe that the tranpose of a row vector is a column vector. Similarly, the transpose of a column
vector is a row vector.
THEOREM 2.3: Let A and B be matrices and let k be a scalar. Then, whenever the sum and product are
defined,
We emphasize that, by (iv), the transpose of a product is the product of the transposes, but in the
reverse order.
The trace of A, written trðAÞ, is the sum of the diagonal elements. Namely,
trðAÞ ¼ a11 þ a22 þ a33 þ þ ann
The following theorem applies.
THEOREM 2.4: Suppose A ¼ ½aij and B ¼ ½bij are n-square matrices and k is a scalar. Then
(i) trðA þ BÞ ¼ trðAÞ þ trðBÞ, (iii) trðAT Þ ¼ trðAÞ,
(ii) trðkAÞ ¼ k trðAÞ, (iv) trðABÞ ¼ trðBAÞ.
EXAMPLE 2.7 Let A and B be the matrices A and B in Example 2.6. Then
diagonal of A ¼ f1; 4; 7g and trðAÞ ¼ 1 4 þ 7 ¼ 4
diagonal of B ¼ f2; 3; 4g and trðBÞ ¼ 2 þ 3 4 ¼ 1
Moreover,
trðA þ BÞ ¼ 3 1 þ 3 ¼ 5; trð2AÞ ¼ 2 8 þ 14 ¼ 8; trðAT Þ ¼ 1 4 þ 7 ¼ 4
trðABÞ ¼ 5 þ 0 35 ¼ 30; trðBAÞ ¼ 27 24 33 ¼ 30
As expected from Theorem 2.4,
trðA þ BÞ ¼ trðAÞ þ trðBÞ; trðAT Þ ¼ trðAÞ; trð2AÞ ¼ 2 trðAÞ
Furthermore, although AB 6¼ BA, the traces are equal.
EXAMPLE 2.8 The following are the identity matrices of orders 3 and 4 and the corresponding scalar
matrices for k ¼ 5:
2 3 2 3
2 3 1 2 3 5
1 0 0 6 7 5 0 0 6 7
4 0 1 0 5; 6 1 7; 4 0 5 0 5; 6 5 7
4 1 5 4 5 5
0 0 1 0 0 5
1 5
Remark 1: It is common practice to omit blocks or patterns of 0’s when there is no ambiguity, as
in the above second and fourth matrices.
Inverse of a 2 2 Matrix
a b
Let A be an arbitrary 2 2 matrix, say A ¼ . We want to derive a formula for A1 , the inverse
c d
of A. Specifically, we seek 22 ¼ 4 scalars, say x1 , y1 , x2 , y2 , such that
a b x1 x2 1 0 ax1 þ by1 ax2 þ by2 1 0
¼ or ¼
c d y1 y2 0 1 cx1 þ dy1 cx2 þ dy2 0 1
Setting the four entries equal to the corresponding entries in the identity matrix yields four equations,
which can be partitioned into two 2 2 systems as follows:
ax1 þ by1 ¼ 1; ax2 þ by2 ¼ 0
cx1 þ dy1 ¼ 0; cx2 þ dy2 ¼ 1
Suppose we let jAj ¼ ab bc (called the determinant of A). Assuming jAj 6¼ 0, we can solve uniquely for
the above unknowns x1 , y1 , x2 , y2 , obtaining
d c b a
x1 ¼ ; y1 ¼ ; x2 ¼ ; y2 ¼
jAj jAj jAj jAj
Accordingly,
1
1 a b d=jAj b=jAj 1 d b
A ¼ ¼ ¼
c d c=jAj a=jAj jAj c a
In other words, when jAj 6¼ 0, the inverse of a 2 2 matrix A may be obtained from A as follows:
(1) Interchange the two elements on the diagonal.
(2) Take the negatives of the other two elements.
(3) Multiply the resulting matrix by 1=jAj or, equivalently, divide each element by jAj.
In case jAj ¼ 0, the matrix A is not invertible.
2 3 1 3
EXAMPLE 2.11 Find the inverse of A ¼ and B ¼ .
4 5 2 6
First evaluate jAj ¼ 2ð5Þ 3ð4Þ ¼ 10 12 ¼ 2. Because jAj 6¼ 0, the matrix A is invertible and
5
1 1 5 3 2 3
A ¼ ¼ 2
2 4 2 2 1
Now evaluate jBj ¼ 1ð6Þ 3ð2Þ ¼ 6 6 ¼ 0. Because jBj ¼ 0, the matrix B has no inverse.
Remark: The above property that a matrix is invertible if and only if A has a nonzero determinant
is true for square matrices of any order. (See Chapter 8.)
Inverse of an n n Matrix
Suppose A is an arbitrary n-square matrix. Finding its inverse A1 reduces, as above, to finding the
solution of a collection of n n systems of linear equations. The solution of such systems and an efficient
way of solving such a collection of systems is treated in Chapter 3.
THEOREM 2.5: Suppose A ¼ ½aij and B ¼ ½bij are n n (upper) triangular matrices. Then
(i) A þ B, kA, AB are triangular with respective diagonals:
ða11 þ b11 ; . . . ; ann þ bnn Þ; ðka11 ; . . . ; kann Þ; ða11 b11 ; . . . ; ann bnn Þ
(ii) For any polynomial f ðxÞ, the matrix f ðAÞ is triangular with diagonal
ð f ða11 Þ; f ða22 Þ; . . . ; f ðann ÞÞ
(iii) A is invertible if and only if each diagonal element aii 6¼ 0, and when A1 exists
it is also triangular.
A lower triangular matrix is a square matrix whose entries above the diagonal are all zero. We note
that Theorem 2.5 is true if we replace ‘‘triangular’’ by either ‘‘lower triangular’’ or ‘‘diagonal.’’
Generally speaking, vectors u1 , u2 ; . . . ; um in Rn are said to form an orthonormal set of vectors if the
vectors are unit vectors and are orthogonal to each other; that is,
0 if i 6¼ j
ui uj ¼
1 if i ¼ j
In other words, ui uj ¼ dij where dij is the Kronecker delta function:
We have shown that the condition AAT ¼ I implies that the rows of A form an orthonormal set of
vectors. The condition AT A ¼ I similarly implies that the columns of A also form an orthonormal set
of vectors. Furthermore, because each step is reversible, the converse is true.
The above results for 3 3 matrices are true in general. That is, the following theorem holds.
THEOREM 2.6: Let A be a real matrix. Then the following are equivalent:
(a) A is orthogonal.
(b) The rows of A form an orthonormal set.
(c) The columns of A form an orthonormal set.
THEOREM 2.7: Let A be a real 2 2 orthogonal matrix. Then, for some real number y,
cos y sin y cos y sin y
A¼ or A¼
sin y cos y sin y cos y
Special Complex Matrices: Hermitian, Unitary, Normal [Optional until Chapter 12]
Consider a complex matrix A. The relationship between A and its conjugate transpose AH yields
important kinds of complex matrices (which are analogous to the kinds of real matrices described above).
(Thus, A must be a square matrix.) This definition reduces to that for real matrices when A is real.
(a) By inspection, the diagonal elements of A are real, and the symmetric elements 1 2i and 1 þ 2i are
conjugate, 4 þ 7i and 4 7i are conjugate, and 2i and 2i are conjugate. Thus, A is Hermitian.
(b) Multiplying B by BH yields I; that is, BBH ¼ I. This implies BH B ¼ I, as well. Thus, BH ¼ B1 ,
which means B is unitary.
(c) To show C is normal, we evaluate CC H and C H C:
2 þ 3i 1 2 3i i 14 4 4i
CC ¼H
¼
i 1 þ 2i 1 1 2i 4 þ 4i 6
14 4 4i
and similarly C H C ¼ . Because CC H ¼ C H C, the complex matrix C is normal.
4 þ 4i 6
We note that when a matrix A is real, Hermitian is the same as symmetric, and unitary is the same as
orthogonal.
The case of matrix multiplication is less obvious, but still true. That is, suppose that U ¼ ½Uik and
V ¼ ½Vkj are block matrices such that the number of columns of each block Uik is equal to the number of
rows of each block Vkj . (Thus, each product Uik Vkj is defined.) Then
2 3
W11 W12 . . . W1n
6 W21 W22 . . . W2n 7
UV ¼ 6
4 ...
7; where Wij ¼ Ui1 V1j þ Ui2 V2j þ þ Uip Vpj
... ... ... 5
Wm1 Wm2 . . . Wmn
The proof of the above formula for UV is straightforward but detailed and lengthy. It is left as an exercise
(Problem 2.85).
The block matrix A is not a square block matrix, because the second and third diagonal blocks are not
square. On the other hand, the block matrix B is a square block matrix.
M 1 ¼ diagðA1 1 1
11 ; A22 ; . . . ; Arr Þ
Analogously, a square block matrix is called a block upper triangular matrix if the blocks below the
diagonal are zero matrices and a block lower triangular matrix if the blocks above the diagonal are zero
matrices.
CHAPTER 2 Algebra of Matrices 41
EXAMPLE 2.17 Determine which of the following square block matrices are upper diagonal, lower
diagonal, or diagonal:
2 3
2 3 1 0 0 0 2 3 2 3
1 2 0 62 3 4 07 1 0 0 1 2 0
A ¼ 4 3 4 5 5; B¼6 4 5 0 6 0 5;
7 C ¼ 4 0 2 3 5; D ¼ 43 4 55
0 0 6 0 4 5 0 6 7
0 7 8 9
(a) A is upper triangular because the block below the diagonal is a zero block.
(b) B is lower triangular because all blocks above the diagonal are zero blocks.
(c) C is diagonal because the blocks above and below the diagonal are zero blocks.
(d) D is neither upper triangular nor lower triangular. Also, no other partitioning of D will make it into
either a block upper triangular matrix or a block lower triangular matrix.
SOLVED PROBLEMS
(b) First perform the scalar multiplication and then a matrix addition:
2 4 6 9 0 6 7 4 0
2A 3B ¼ þ ¼
8 10 12 21 3 24 29 7 36
(Note that we multiply B by 3 and then add, rather than multiplying B by 3 and subtracting. This usually
prevents errors.)
x y x 6 4 xþy
2.2. Find x; y; z; t where 3 ¼ þ :
z t 1 2t zþt 3
Write each side as a single equation:
3x 3y xþ4 xþyþ6
¼
3z 3t zþt1 2t þ 3
Set corresponding entries equal to each other to obtain the following system of four equations:
3x ¼ x þ 4; 3y ¼ x þ y þ 6; 3z ¼ z þ t 1; 3t ¼ 2t þ 3
or 2x ¼ 4; 2y ¼ 6 þ x; 2z ¼ t 1; t¼3
The solution is x ¼ 2, y ¼ 4, z ¼ 1, t ¼ 3.
2.3. Prove Theorem 2.1 (i) and (v): (i) ðA þ BÞ þ C ¼ A þ ðB þ CÞ, (v) kðA þ BÞ ¼ kA þ kB.
Suppose A ¼ ½aij , B ¼ ½bij , C ¼ ½cij . The proof reduces to showing that corresponding ij-entries
in each side of each matrix equation are equal. [We prove only (i) and (v), because the other parts
of Theorem 2.1 are proved similarly.]
42 CHAPTER 2 Algebra of Matrices
(i) The ij-entry of A þ B is aij þ bij ; hence, the ij-entry of ðA þ BÞ þ C is ðaij þ bij Þ þ cij . On the other hand,
the ij-entry of B þ C is bij þ cij ; hence, the ij-entry of A þ ðB þ CÞ is aij þ ðbij þ cij Þ. However, for
scalars in K,
ðaij þ bij Þ þ cij ¼ aij þ ðbij þ cij Þ
Thus, ðA þ BÞ þ C and A þ ðB þ CÞ have identical ij-entries. Therefore, ðA þ BÞ þ C ¼ A þ ðB þ CÞ.
(v) The ij-entry of A þ B is aij þ bij ; hence, kðaij þ bij Þ is the ij-entry of kðA þ BÞ. On the other hand, the ij-
entries of kA and kB are kaij and kbij , respectively. Thus, kaij þ kbij is the ij-entry of kA þ kB. However,
for scalars in K,
kðaij þ bij Þ ¼ kaij þ kbij
Thus, kðA þ BÞ and kA þ kB have identical ij-entries. Therefore, kðA þ BÞ ¼ kA þ kB.
Matrix Multiplication
2 3
3 2 4 2 3
3 6 9 7 5
2.4. Calculate: (a) ½8; 4; 54 2 5, (b) ½6; 1; 7; 56 7
4 3 5, (c) ½3; 8; 2; 44 1 5
1 6
2
(a) Multiply the corresponding entries and add:
2 3
3
½8; 4; 54 2 5 ¼ 8ð3Þ þ ð4Þð2Þ þ 5ð1Þ ¼ 24 8 5 ¼ 11
1
(c) The product is not defined when the row matrix and the column matrix have different numbers of elements.
2.5. Let ðr sÞ denote an r s matrix. Find the sizes of those matrix products that are defined:
(a) ð2 3Þð3 4Þ; (c) ð1 2Þð3 1Þ; (e) ð4 4Þð3 3Þ
(b) ð4 1Þð1 2Þ, (d) ð5 2Þð2 3Þ, (f) ð2 2Þð2 4Þ
In each case, the product is defined if the inner numbers are equal, and then the product will have the size of
the outer numbers in the given order.
(a) 2 4, (c) not defined, (e) not defined
(b) 4 2, (d) 5 3, (f) 2 4
1 3 2 0 4
2.6. Let A ¼ and B ¼ . Find: (a) AB, (b) BA.
2 1 3 2 6
(a) Because A is a 2 2 matrix and B a 2 3 matrix, the product AB is defined and is a 2 3 matrix. To
obtain the entries in the first row of AB, multiply the first row ½1; 3 of A by the columns
2 0 4
; ; of B, respectively, as follows:
3 2 6
1 3 2 0 4 2 þ 9 0 6 4 þ 18 11 6 14
AB ¼ ¼ ¼
2 1 3 2 6
CHAPTER 2 Algebra of Matrices 43
To obtain the entries in the second row of AB, multiply the second row ½2; 1 of A by the columns of B:
1 3 2 0 4 11 6 14
AB ¼ ¼
2 1 3 2 6 4 3 0 þ 2 8 6
Thus,
11 6 14
AB ¼ :
1 2 14
(b) The size of B is 2 3 and that of A is 2 2. The inner numbers 3 and 2 are not equal; hence, the product
BA is not defined.
2 3
2 1 0 6
2 3 1
2.7. Find AB, where A ¼ and B ¼ 4 1 3 5 1 5.
4 2 5
4 1 2 2
Because A is a 2 3 matrix and B a 3 4 matrix, the product AB is defined and is a 2 4 matrix. Multiply
the rows of A by the columns of B to obtain
4 þ 3 4 2 þ 9 1 0 15 þ 2 12 þ 3 2 3 6 13 13
AB ¼ ¼ :
8 2 þ 20 4 6 þ 5 0 þ 10 10 24 2 þ 10 26 5 0 32
1 6 2 2 1 6 1 6
2.8. Find: (a) , (b) , (c) ½2; 7 .
3 5 7 7 3 5 3 5
(a) The first factor is 2 2 and the second is 2 1, so the product is defined as a 2 1 matrix:
1 6 2 2 42 40
¼ ¼
3 5 7 6 35 41
(b) The product is not defined, because the first factor is 2 1 and the second factor is 2 2.
(c) The first factor is 1 2 and the second factor is 2 2, so the product is defined as a 1 2 (row) matrix:
1 6
½2; 7 ¼ ½2 þ 21; 12 35 ¼ ½23; 23
3 5
2.9. Clearly, 0A ¼ 0 and A0 ¼ 0, where the 0’s are zero matrices (with possibly different sizes). Find
matrices A and B with no zero entries such that AB ¼ 0.
1 2 6 2 0 0
Let A ¼ and B ¼ . Then AB ¼ .
2 4 3 1 0 0
The above sums are equal; that is, corresponding elements in ðABÞC and AðBCÞ are equal. Thus,
ðABÞC ¼ AðBCÞ.