An Introduction To Lie Algebra
An Introduction To Lie Algebra
CSUSB ScholarWorks
12-2017
Recommended Citation
Talley, Amanda Renee, "An Introduction to Lie Algebra" (2017). Electronic Theses, Projects, and
Dissertations. 591.
https://scholarworks.lib.csusb.edu/etd/591
This Thesis is brought to you for free and open access by the Office of Graduate Studies at CSUSB ScholarWorks. It
has been accepted for inclusion in Electronic Theses, Projects, and Dissertations by an authorized administrator of
CSUSB ScholarWorks. For more information, please contact [email protected].
An Introduction to Lie Algebra
A Thesis
Presented to the
Faculty of
San Bernardino
In Partial Fulfillment
Master of Arts
in
Mathematics
by
December 2017
An Introduction to Lie Algebra
A Thesis
Presented to the
Faculty of
San Bernardino
by
December 2017
Approved by:
Abstract
Acknowledgements
Table of Contents
Abstract iii
Acknowledgements iv
List of Figures vi
1 Introduction 1
5 Conclusion 53
Bibliography 54
vi
List of Figures
Chapter 1
Introduction
Abstract Lie algebras are algebraic structures used in the study of Lie groups.
They are vector space endomorphisms of linear transformations that have a new operation
that is neither commutative nor associative, but referred to as the bracket operation, or
commutator. Recall that an associative algebra preserves bilinear multiplication between
elements in a vector space V over a field F , where V ×V → V is defined by (x, y) → xy for
x, y ∈ V . The bracket operation yields a new bilinear map, over the same vector space,
that turns any associative algebra into a Lie Algebra, where V × V → V is now defined
by [x, y] → (xy − yx). Central to the study of Lie algebras are the classical algebras,
which will be explored throughout this paper to derive isomorphisms, identify simple
algebras, and determine solvability. After examining some natural Lie algebra examples,
we will delve into the derived algebra, which is analogous to the commutator subgroup of
a group, and make use of it to define a sequence of ideals called the derived series, which
is solvable. We will then use a different sequence of ideals called the descending central
series to classify nilpotent algebras. All of this will arm us with the necessary tools to
prove Engel’s Theorem, which states that if all elements of L are ad-nilpotent, then L is
nilpotent. We must first recall some basic group axioms, ring axioms, and vector space
properties as a precursor to the study of Lie algebra.
2
• Inverse: For each a ∈ G there exists an element a−1 ∈ G such that a ∗ a−1 =
a−1 ∗ a = e. Such an element a−1 ∈ G is called an inverse of a in G.
Example 1.0.1. The general linear group, denoted gl(n), consists of the set of invert-
ible n × n matrices. By the above, given that multiplication of invertible n × n matrices
is associative, and each invertible matrix has an inverse and an identity, gl(n) forms a
group under multiplication.
3
Definition 1.5. The number of elements in a group G is called the order of G and is
denoted |G|. G is a finite group if |G| is finite. If a ∈ G, the order |a| of a in G is
the least positive integer n such that an = e. If there exists no such integer n, then |a| is
infinite.
0
Definition 1.6. A map ϕ : G → G from a group G to a group G0 is called a homo-
morphism if ϕ(ab) = ϕ(a)ϕ(b) for all a, b ∈ G (in ϕ(ab) the product is being taken in
G, while in ϕ(a)ϕ(b) the product is being taken in G0 ).
Example 1.0.2. For any group G, the identity map is always a homomorphism, since
if ϕ : G → G0 is the identity ϕ(x) = x, then ϕ(xy) = xy = ϕ(x)ϕ(y)
Definition 1.8. For the homomorhpism ϕ, the kernel of ϕ is the set {x ∈ G|ϕ(x) = e0 },
denoted Ker ϕ.
Definition 1.10. The group consisting of the cosets of H in G under the operation
(aH)(bH) = (ab)H is called the quotient group of G by H, written G/H.
Definition 1.11. If G is any group, then the center of G, denoted Z(G), consists of the
elements of G that commute with every element of G. In other words,
Note that ey = y = ye for all y ∈ G, so e ∈ Z(G), and the center is a nonempty subset of
G. Let a ∈ G. Then the centralizer of a in G, denoted CG (a), is the set of all elements
of G that commute with a. In other words
Note that for any a ∈ G we have Z(G) ⊆ CG (a). In other words, the center is contained
in the centralizer of any element [Aig02].
The following material for group and ring theory was derived from Joseph A.
Gallian’s Contemporary Abstract Algebra [Gal04].
Definition 1.14. Let G be a group, and let a ∈ G. The function ϕa defined by ϕa (x) =
axa−1 for all x ∈ G is called the inner automorphism of G induced by a.
Definition 1.15. A set R equipped with two operations, written as addition and multi-
plication, is said to be a ring if the following four ring axioms are satisfied, for any
elements, a, b, and c in R:
• Closure: ab ∈ R
5
Theorem 1.17. Ideals are Kernels Every ideal of a ring R is the kernel of a ring
homomorphism of R. In particular, an ideal A is the kernel of the mapping r → r + A
from R to R/A.
Definition 1.18. A unity (or identity) in a ring is a nonzero element that is an identity
under multiplication. A nonzero element of a commutative ring with unity that has a
multiplicative inverse is called a unit of the ring.
Theorem 1.19. First Isomorphism Theorem for Rings Let ϕ be a ring homomor-
phism from R → S. Then the mapping from R/kerϕ to ϕ(r), given by r + kerϕ → ϕ(r)
is an isomorphism. In symbols, R/kerϕ ∼
= ϕ(r)
Theorem 1.21. Third Isomorphism Theorem for Rings Let A and B be ideals of
a ring R with B ⊆ A. A/B is an ideal of R/B and (R/B)(A/B) is isomorphic to R/A.
Definition 1.24. A field is a commutative ring with unity in which every nonzero ele-
ment is a unit.
6
Unless otherwise stated, the material for our vector space review was derived
from Stephen H. Friedberg et al’s, Linear Alebra [FIS03].
Definition 1.25. Let F be a field. A set V equipped with two operations, written as
addition and multiplication, is said to be a vector space over F if
• a(v + w) = av + aw
• a(bv) = (ab)v
• 1v = v
Example 1.0.3. If T : V → W is a linear map, the kernel of T and the image (range)
of T defined by
ker(T ) = {x ∈ W : T (x) = 0}
im(T ) = {w ∈ W : T (x), x ∈ W }
are subspaces of V .
Definition 1.27. Let V be a vector space over a field F. A vector space homomorphism
that maps V to itself is called an endomorphism of V
1. hx + z, yi = hx, yi + hz, yi
2. hcx, yi = c hx, yi
3. hx, yi = hy, xi
4. hx, xi > 0 if x 6= 0
Definition 1.30. Let V be a vector space over a field F . A function f from the set V ×V
of ordered pairs of vectors to F is called a bilinear form on V if f is linear in each
variable when the other variable is held fixed; that is, f is a bilinear form on V if
Definition 1.31. An alternating space is defined as the set of all vector spaces V and
alternate bilinear forms such that f : V × V → F where:
2. f (x + y, z) = f (x, z) + f (y, z)
3. Im A = A = AIm
Definition 1.34. Let A, B and C be matrices where C = cij , A = ajk , and B = bkj .
Then by matrix multiplication
n
X
C = AB where cij = aik bkj
k=1
Where the scalar cij = (−1)i+j det(Ãij ) is called the cofactor of the entry of A if row i,
column j.
Definition 1.36. The trace of a matrix is the sum of its diagonal entries.
Definition 1.38. If the transpose of a matrix is equal to the negative of itself, the matrix
is said to be skew symmetric. This means that for a matrix to be skew symmetric,
At = −A.
9
Definition 1.39. Let F be a field and V, W vector spaces over the field F . A function
T : V → W is said to be a linear transformation from V to W if for all c, d ∈ F , and
all x, y ∈ V we have
• T (0) = 0
• T (x − y) = T (x) − T (y)
Xn n
X
T( ai xi ) = ai T (xi )
i=1 i=1
Therefore A(adjA) is a scalar matrix with diagonal elements all equal to |A|.
⇒ A(AdjA) = I(n)
⇒ (AdjA)A = |A|I(n)
AdjA
A−1 =
|A|
[Eve66]
Theorem 1.47. In the characteristic function f (λ) of matrix A, pn = |A| and p1 = trA.
[Eve66]
11
Chapter 2
Definition 2.1. A vector space L over a field F , is a Lie algebra if there exists a bilinear
multiplication L × L → L, with an operation, denoted (x, y) 7→ [xy], such that:
Example 2.1.1. Given an n dimensional vector space End (V ), the set of all all linear
maps V 7→ V with associative multiplication (x, y) 7→ xy for all x, y, where xy denotes
functional composition, observe, End (V ) is an associative algebra over F . Let us define
a new operation on End (V ) by (x, y) 7→ xy − yx. If we denote xy − yx by [x, y], then
End (V ) together with the map [·, ·] satisfies Definition 2.1, and is thus a Lie algebra.
Proof. The first two bracket axioms are satisfied immediately. The only thing left to
prove is the Jacobi identity. Given x, y, z ∈ End (V ), we have by use of the bracket
operation:
Definition 2.2. The Lie algebra End (V ) with bracket [x, y] = xy − yx, is denoted as
gl(V ), the general linear algebra.
Example 2.1.2. We can show that real vector space R3 is a Lie algebra. Recall the
following cross product properties when a, b and c represent arbitrary vectors and α, β
and γ represent arbitrary scalars:
1. a × b = −(b × a),
([MT03])
Note, a × a = −(a × a), by property (1), letting b = a, therefore, a × a = 0. By the above
properties, the cross product is both skew symmetric (property 1) and bilinear (property
2). By vector triple product expansion: x × (y × z) = y(x · z) − z(x · y). To show that the
cross product satisfies the Jacobi identity, we have:
(Al ) The set of all endomorphisms of V having trace zero is denoted by sl(V ), the
special linear algebra. Letting x, y ∈ sl(V ), we can show that sl(V ) is a subalgebra of
gl(V ).
Proof.
Therefore, both Lie algebra axioms are satisfied, hence by Definition 2.3, sl(V ) is a
subalgebra of gl(V ).
14
Therefore [x, y] ∈ sp(V ), hence sp(V ) is a subalgebra of gl(V ) when sp(v) is of even
dimension. We show that sp(V ) requires even dimensionality by considering the
determinant of x when n is odd. From sx = −xt s,
1 0 0
matrix s =
0 0 , corresponds to the orthogonal algebra, denoted as o(V ), or
Il
0 Il 0
o(2l + 1, F ). Recall a matrix A ∈ Mn×n (F ) is called orthogonal if AAt = I. We can
a b1 b2
partition x as x = c1 m n . The orthogonal algebra is a subalgebra of gl(V ),
c2 p q
where f (x(v), w) = −f (v, x(w)) (the same conditions as for (Cl ).
X = (x1 , x2 , ..., xn ) and Y = (y1 , y2 , ..., yn ). The corresponding Lie algebra is o(n, F )
0 Il
(Dl ) Let Dl be an an orthogonal algebra with s = . Let m = 2l, then Dl
Il 0
consists of all the matrices with dimension 2l2 − l, satisfying f (x(v), w) = −f (v, x(w)).
Definition 2.4. The following is a list of several other subalgebras consisting of the
upper, lower, and triangular matrices of gl(n, F ):
• Let t(n, F ) be the set of upper triangular matrices (aij ), aij = 0 if i > j.
Example 2.2.2. Given algebras t(n, F ), o(n, F ), n(n, F ), we can compute the dimension
of each algebra, by exhibiting a basis.
0 a12 ... a1n
0 0 ... a2n
• For n(n, F ) = .
.. .. ..
. .
0 0 ··· 0
bh
Using the formula for the area of a triangle: A = 2 , the dimension for n(n, F ) is
n(n−1)
therefore 2
16
a11 0 ... 0
0 a22 ...
0
• For o(n, F ) =
.. .. ..
. . .
0 0 · · · ann
The dimension of o(n, F ) is clearly n.
a11 + a12 ... a1n
0 a 22 ... a 2n
• For t(n, F )= .
.. .. . .
. .
0 0 ... ann
Again using the formula for the area of a triangle, the dimension of t(n, F ) is
n(n+1)
2
Definition 2.5. The (external) direct sum of two Lie algebras L, L0 , written L ⊕ L0 is
the vector space direct sum, with [·, ·] defined “componentwise”: [(x1 , y1 ), (x2 , y2 )] where
{(x, 0)} , {(y, 0)} are ideals and thus ”nullify” each other. That is, [L, L0 ] = 0. [Sam90]
Example 2.2.3. We can prove that t(n, F ) is the direct sum of o(n, F ) and n(n, F ). By
the above definition, we must show [t(n, F ), t(n, F )] = 0 where
t(n, F ) = o(n, F ) + n(n, F ).
Where Aik = 0 in the first sum since i ≥ k for A strictly upper triangular. For the
second sum k ≥ i > j − 1, thus Bkj = 0. Therefore A ∩ B = {0}, meaning
17
Conversely:
The derivative between two functions f and g is a linear operator that satisfies the
Leibniz rule:
1. (f g)0 = f 0 g + f g 0
Given an algebra A over a field F , the derivation of A is the linear map δ such that
δ(f, g) = f δ(g) + δ(f )g for all f, g ∈ A. The set of all derivations of A is denoted by
Der(A). Given δ ∈ Der(A), f, g ∈ A, and α ∈ F . By property 2.,
Example 2.3.2. By example 2.3.1, the collection of derivations, Der(V ), satisfies skew
symmetry. By definition 2.6, Der(V ) satisfies the Jacobi identity. Therefore Der(V )
defines a Lie algebra.
Definition 2.7. The map L → DerL sending x to adx is called the adjoint
representation of L.
That is, if and only if the Jacobi identity is satisfied, where [[x, y], z] = −[z, [x, y]] and
−[y, [x, z]] = [y, [z, x]] by skew symmetry.
Let L be a finite dimensional vector space over F . Any vector space can be made into a
Lie algebra simply by setting [x, y] = 0 for all vectors x and y in L. The resulting Lie
algebra is called abelian. Recall that a group G under multiplication is said to be
Abelian if G obeys the commutative law or, that is, if for every a, b ∈ G we have
ab = ba [Aig02]. If an associative algebra is abelian and therefore commutative, then
ab = ba ⇔ [a, b] = 0, where [a, b] is the associated commutator.
Definition 2.9. If L is any Lie algebra with basis x1 , ..., xn , then the multiplication
table for L is determined by the structural constants, the set of scalars {bijk } such
that [xi , xj ] = nk=1 bijk xk .
P
Example 2.4.2. By use of the bracket axioms, the following set of structural constants
define an abstract Lie algebra:
Proof.
20
But:
[xi xj ] = xi xj − xj xi
[xj xi ] = xj xi − xi xj
⇔ [xi xj ] + [xj xi ] = xi xj − xj xi + xj xi − xi xj = 0
Definition 2.10. The assorted systems of structural constants of a given Lie algebra
relative to its basis form an orbit (the set of all transformations of one element) under
this action [Sam90].
Example 2.4.3. The orbit of the system cijk = 0 for all i, j, k forms a natural
isomorphism to the abelian algebra [X, Y ] = 0 of dimension n.
Example 2.4.4. Consider a Lie algebra L with basis x1 , ...xn . The structural constants
are enumerated as followed: e1 = (1, 0, 0), e2 = (0, 1, 0), and e3 = (0, 0, 1). By use of the
Lie bracket axioms, We have:
[e1 , e2 ] = e3
[e1 , e3 ] = −e2
[e2 , e3 ] = e1
21
Example 2.4.5. We can use structural constants to show that a three dimensional
vector space with basis (x, y, z) such that [xy] = z, [xz] = y, [yz] = 0 defines a Lie
algebra.
Where bijk = bijk − bjik = 0 for all structural constants above (by skew-symmetry). By
the Jacobi identity, we have:
e1 = (1, 0, ..., 0)
e2 = (0, 1, ..., 0)
..
.
en = (0, 0, ..., 1)
Theorem 2.12. Let V and W be finite-dimensional vector spaces (over the same field).
Then V is isomorphic to W if and only if they are of equal dimension.
22
Example 2.4.6. By Corollary 2.13, there exists an isomorphism between gl(n) and F n .
23
Chapter 3
3.1 Ideals
Lie algebra ideals are intrinsic to much that follow. In group theory, ideals are used to
define the kernel of a group, normal subgroups, and the structure of nilpotent and
solvable groups. In a ring, every kernal is an ideal of a homomorphism, and every
subgroup is normal. By definition of the normal subgroup, if H is a subgroup of G, for
all g ∈ G we have gH = Hg ⇔ [g, H] = 0. Unless otherwise stated, the material for this
chapter is derived from [Hum72].
By skew-symmetry, all Lie algebra ideals are automatically two sided. That is, if
[x, y] ∈ I, then [y, x] ∈ I. The kernel of a Lie algebra L and L itself are trivial ideals
contained in every Lie algebra.
Example 3.1.1. The set of all inner derivations adx , x ∈ L, is an ideal of Der(L). Let
δ ∈ Der(L). By definition of inner derivations, for all y ∈ L:
The center respresents the collection of elements in L for which the adjoint action ad(z)
yeilds the zero derivation. It can therefore be thought of as the kernel of the adjoint
map from L to Der(L). Clearly, the center is an ideal, therefore Z(L) satisfies skew
symmetry. Given y ∈ L, [z, [x, y]] = −[x, [y, z]] − [y, [z, x]], both [y, z] and [z, x] belong
to Z(L) by definition, therefore, Z(L) satisfies the Jacobi identity. If Z(L) = L, then L
is said to be albelian.
CL (X) = {x ∈ L|[x, X] = 0}
Definition 3.4. Consider the Lie algebra L, having two ideals I and J, the following
are ideals of L:
I + J = {x + y|x ∈ I, y ∈ J}
nX o
[I, J] = xi yi |xi ∈ I, yi ∈ J
Proposition 3.5. Let L be a Lie algebra and let I, J be subsets of L. The subspace
spanned by elements of the form x + y (x ∈ I, y ∈ J), is denoted I + J, and the subspace
spanned by elements of the form [x, y], (x ∈ I, y ∈ J), is denoted [I, J]. If I, I1 , I2 , J, K
are subspaces of L, then
2. [I, J] = [J, I]
[Wan75]
If I is an ideal of L, then the quotient space L/I (whose elements are the linear cosets
L + I) induces the well defined bracket operation, where [x + I, y + I] = [x, y] + I for
x, y ∈ L [Sam90].
25
Definition 3.6. By use of the bracket operation [·, ·], the quotient space L/I becomes a
quotient Lie algebra [Sam90].
Definition 3.7. If the Lie algebra L has no ideals except itself and 0, and if moreover
[LL] 6= 0, we call L simple.
Example 3.1.2. Let L be a Lie algebra, and L = sl(n,F ) where char F 6= 2. It can be
shown that L is simple if we take as a standard basis for L the three matrices
0 1 0 0 1 0
x= ,y = ,h =
0 0 1 0 0 −1
• adx :
• ady:
0 0 0 1 0 1 0 0
ady (x) = [y, x] = yx − xy = −
1 0 0 0 0 0 1 0
0 0 1 0 −1 0
= − = = −h
0 1 0 0 0 1
0 0 2
• ad h:
1 0
0 1 0 1 1 0
adh (x) = [h, x] = hx − xh = −
0 −1 0 0 0 0 0 −1
0 1 0 −1 0 2
= − = = 2x = 2 · x + 0 · h + 0 · y
0 0 0 0 0 0
Example 3.1.3. We can show that sl(3, F ) is simple, unless char F = 3, using the
standard basis h1 , h2 , eij (i 6= j). If I 6= 0 is an ideal, then I is the direct sum of
eigenspaces for ad h1 or ad h2 . We will determine the eigenvalues of by afixing h1 and
h2 to eij , i.e, take the adjoint using diagonal basis matrices. By definition of basis
matrices, for sl(3, F ), h1 = e11 − e22 and h2 = e22 − e33 .
Continuing this way, we construct the table below to denote the eigenvalues ad h1 , ad h2
acting on eij
Therefore the eigenvalues are 2, 1, −1, −2 and 0, and since char F 6= 3, the eigenvalues
are distinct. If I is an ideal, then by definition,
Example 3.1.4. We can show that t(n, F ) and o(n, F ) are self-normalizing subalgebras
of gl(n, F ), whereas n(n, F ) has a normalizer t(n, F ). Note: for all upper triangular
matrices, the diagonal entries of the matrix are precisely it’s eigenvalues.
n
X n
X
[b, ekk ] = bijk xk ekk − ekk bijk xk
k=1 k=1
Xn n
X
= bij eij ekk − bij ekk eij
i,j=1 i,j=1
X n Xn
= bij δjk eik − bij δki ekj ∵ eij ekl = δjk eil
i,j=1 i,j=1
X n n
X
= bik eik − bkj ekj ∈ t(n, F )
i=1 j=1
Pn Pn
Where i=1 bik eik implies that bik = 0, for i < k and j=1 bkj ekj implies that
bjk = 0 for j < k. Therefore bkl = 0 for k < l as required, and thus b ∈ t(n, f ).
Homomorphisms as they appear in ring theory articulate nicely in Lie algebra. Vector
space homomorphisms are linear maps where the group action is abelian under addition
30
and preserves scalar multiplication. Given a Lie algebra L, the next few definitions
allow us to visualize Lie algebra homomorphisms, and generalize the well known
isomorphism theorems to vector spaces modulo I.
The adjoint representation of a Lie algebra L is a homomorphism using the map ad:
L → gl(V ). By definition, ker ad=Z(L). If L is a simple Lie algebra, then Z(L) = 0,
thus by Definition 3.9, adL → gl(L) is a monomorphism, and therefore an invertible
homomorphism, which proves that any simple Lie algebra is isomorphic to a linear Lie
algebra.
Definition 3.10. The special linear group SL(n, F) denotes the kernel of the
homomorphism
det : GL(n, F ) → F x = {x ∈ F |x 6= 0}
where F is a field.
We will now forge a proof of Lie algebra isomorphism theorems analogous to the ring
theory isomorphisms described in the introduction.
L/I
31
2. Let I and J be ideals of L such that I ⊂ J, then J/I is an ideal of L/I and
(L/I)/(J/I) is naturally isomorphic to L/J.
Proof.
Example 3.2.2. Since n(n, F ) is in the kernal of t(n, F ), by Proposition 3.11, we can
show t(n, F )/n(n, F ) ∼
= o(n, F ). Given x, z ∈ t(n, F ), y ∈ n(n, F ), we have
Where [x, z] ∈ t(n, F ). Any upper triangular matrix can be diagonalized, therefore
[x, z] ∈ o(n, F ) and thus [x, z] + y ∈ o(n, F ).
(x + y) + (z + y) = (x + z) + y
Since the sum of any two upper triangular matrices is also upper triangular,
(x + z) ∈ t(n, F ), and therefore (x + z) + y ∈ o(n, F ).
Example 3.2.3. For small values of l, isomorphisms occur among certain classical
algebras. We can show that A1 , B1 , C1 are all isomorphic. Additionally, we will show
that B2 is isomorphic to C2 , and D3 is isomorphic to A3 . We can form our desired
isomorphisms given the following root system decomposition for classical algebras
An − D n :
Al :
eij (i 6= j)
hi = eii − ei+1,i+1 (1 ≤ i ≤ l)
Bl :
eii − el+i,l+i (2 ≤ i ≤ l + 1)
e1,l+i+1 − ei+1,1 (1 ≤ i ≤ l)
e1,i+1 − el+i+1,1 (1 ≤ i ≤ l)
ei+1,j+1 − el+j+1,l+i+1 (1 ≤ i 6= j ≤ l)
ei+1,l+j+1 − ej+1,l+i+1 (1 ≤ i < j ≤ l)
ei+l+1,j+1 − ej+l+1,i+1 (1 6= j < i ≤ l)
Cl :
eii − el+i,l+i (1 ≤ i ≤ l)
eij − el+j,l+i (1 ≤ i 6= j ≤ l)
ei,l+i (1 ≤ i ≤ l)
ei,l+j + ej,l+i (1 ≤ i < j ≤ l)
33
Dl
ei,j − el+j,l+i (1 ≤ i, j ≤ l)
ei,l+j − ej,l+i (1 ≤ i < j ≤ l)
el+i,j − el+j,i (1 ≤ i < j ≤ l)
[Hum72], [Wan75]
• To prove A1 ∼
= B1 , we must show sl(2, F ) ∼
= o(3, F ). The diagonal basis matrix for
A1 is e11 − e22 :
[e11 − e22 , e12 ] = (e11 − e22 )e12 − e12 (e11 − e22 ) = e12 + e12 = 2e12
[e11 − e22 , e21 ] = (e11 − e22 )e21 − e12 (e11 − e21 ) − 2e21
The following map enumerates the root system from A1 to B1 where sl(2, F ) is a
2 × 2 matrix and o(3, F ) is a 3 × 3 matrix:
Similarly, the following map signifies the root system from B1 to C1 where
sp(2, F ) is a 2 × 2 matrix:
Notice that the basis matrices for A1 and C1 are identical, therefore by the
transitive property, the isomorphism between A1 and C1 is immediate. and
A1 ∼
= B1 ∼ = C1
• For B2 ∼
= C2 , we must show o(5, F ) ∼
= sp(4, F ) by adjoining the diagonal matrices
from B2 , C2 to corresponding eigenvectors to determine the eigenvalues involved in
their respective basis matrices. The diagonal matrices for B2 are as followed:
h1 = e22 − e44
h2 = e33 − e55
34
Therefore, the eigenvalue for eigenvector e21 − e14 is (1, 0) = α, and the eigenvalue
for eigenvector e12 − e41 is (−1, 0) = −α. The remaining eigenvalues are derived
similarly:
Denote λ01 as the eigenvalue of [h01 , x] and λ02 as the eigenvalue of [h02 , x] where x is
an eigenvector. Then eigenvalue λ0 = (λ01 , λ02 ).
35
[h01 , e51 − e13 ] = 0 [h02 , e51 − e13 ] = −1 −(α0 + β 0 ) = (−1, 0) + (1, −1) = (0, −1)
[h01 , e24 ] = 0 [h02 , e24 ] = 2 2α0 + β 0 = 2(−1, 1) + (2, 0) = (0, 2)
[h01 , e42 ] = 0 [h02 , e42 ] = −2 −(2α0 + β 0 ) = −[2(−1, 1) + (2, 0)] = (0, −2)
1 1 1 1
H10 = − h01 + h02 H20 = h1 + h2
2 2 2 2
Therefore, B2 ∼
= C2 , where:
• For A3 ∼
= D3 , we must show sl(4, F ) ∼
= o(6, F ). The diagonal matrices for A3 are:
Denote λ1 = [h1 , x], λ2 = [h2 , x], and λ3 = [h3 , x]. To find the eigenvalue that
corresponds to eigenvector e13 , we have:
For D3 , denote λ01 = [h01 , x0 ], λ02 = [h02 , x0 ], λ03 = [h03 , x0 ] where x0 is an eigenvector.
Let λ0 = (λ01 , λ02 , λ03 ). The eigenvalues for A3 and D3 are listed on the following
table:
x λ = (λ1 , λ2 , λ3 ) x0 λ0 = (λ01 , λ02 , λ03 )
e13 α(1, 1, −1) e26 − e25 α0 = (0, 1, 1)
e31 α = (−1, 1, 1) e62 − e53 α = (0, −1, −1)
e24 β(−1, 1, 1) e23 − e65 β 0 (0, 1, −1)
e42 0
−β = (1, −1, −1) e32 − e56 −β 0 (0, −1, 1)
e41 γ = (−1, 0, −1) e12 − e54 γ 0 = (1, −1, 0)
e14 γ = (1, 0, 1) e21 − e45 −γ 0 (−1, 1, 0)
e43 (α + γ) = (0, 1, −2) e16 − e34 (α + γ 0 ) = (1, 0, 1)
0
Therefore A3 ∼
= D3 , where:
3.3 Automorphisms
ϕa : A → A
ϕa (x) = a−1 xa
Given x, y ∈ A:
ϕa (xy) = a−1 (xy)a = a−1 xaa−1 ya = (a−1 xa)(a−1 ya) = ϕa (x)ϕa (y)
Example 3.3.2. Let L be a Lie algebra such that L = sl(n, F ), g ∈ GL(n, F ). The map
ϕ : L → L defined by x → −gxt g −1 (xt = the transpose of x) belongs to Aut L. When
n = 2, g = the identity matrix, we can prove that Aut L is inner.
Therefore:
Example 3.3.3. An automorphism of the form exp(adx ), with adx nilpotent, i.e.,
(adx )k = 0 for some k > 1, is called inner.
To derive the power series expansion for exp(adx ), we begin by constructing the Taylor
series expansion using the differential operator for matrix A. Given an invertible matrix
A, the solution to x0 = A~x might be x0 (t) = λveλt . Multiplying by A, we have:
Ax(t) = Aveλt
where Ax(t) = x0 (t) ⇔ λv = Av
39
n−1
! n−1
X δ i (x) X δ j (y)
exp(δ(x))exp(δ(y)) =
i! j!
i=0 j=0
2k−2
X δ n (xy)
=
n!
n=0
k−1
X δ n (xy)
=
n!
n=0
P∞ adk(x)
Where the first equality holds because exp(adx ) = k=0 k! , and the second equality
satisfies Leibniz’s product rule. Since δ is a nilpotent derivation of L, the last equality
holds because δ k = 0. Therefore, the composition of two inner automorphisms is again
an inner automorphism, thus the derivation of an inner automorphisms exp(ad(x)) is a
homomorphism:
Definition 3.13. When adx is nilpotent, the inner automorphism constructed is called
Int L. For φ ∈ Aut l, x ∈ L,
Example 3.3.4. Let σ be the automorphism of sl(2, F ) give by the following: let
L ∈ sl(2, F ), with standard basis (x, y, h). Define σ =exp ad x· exp ad (−y)· exp ad x.
We can show that σ(x) = −y, σ(y) = −x, σ(h) = −h. From example 3.1.2, we have
[x, y] = h, [h, y] = −2y, [h, x] = −2x
Chapter 4
4.1 Solvability
Definition 4.1. The derived series of a Lie algebra L is a sequence of ideals of L where
[L, I 1 ] = [L, [I, I]] ⊆ [I, [I, L]] + [I, [L, I]] ⊆ [I, I] + [I, I] = I 1 [Wan75]
Therefore I n+1 is an ideal, making each member of the derived series of I an ideal of
L.
G = G0 ≥ G1 ≥ G2 ≥ ... ≥ Gn = e (4.1)
Proposition 4.4.
1. If L is a solvable Lie algebra, then so are all of the subalgebras and homomorphic
images of L.
2. If I is a solvable ideal of L such that the quotient L/I is solvable, then L itself is
solvable.
Proof.
L
2. Let (L/I)(n) = 0. By construction of the canonical homomorphism π : L → I, we
have that π(L(n) ) = 0 by part 1. or L(n) ⊂ I = kerπ. If I (m) = 0, then
(L(i) )(i) = L(i+j) ⇒ L(n+m) = 0 (since L(n) ⊂ I).
3. By Proposition 3.11 (part 3.): if I, J are ideals of L, then from the sequence
I+J
0→I →I +J → J → 0, there exists a natural isomorphism between the third
term (I + J)/J and I/(I ∩ J). Now (I + J)/J is solvable by the Lemma 4.3, since
the homomorphic image of I is solvable. Therefore I/(I ∩ J) is solvable by part 2.
• Li+1 is an ideal of Li .
Li
• Each quotient Li+1 is abelian.
Proof.
⇒
abelian. By definition of the derived series, L(i+1) = [L(i) , L(i) ]. Let [x, y] ∈ L(i+1) ,
with x, y ∈ L(i) :
L(i)
Therefore L(i+1)
is abelian.
L(k−1)
⇐ Given L(k)
is abelian and L = L0 ⊃ L1 ⊃ ... ⊃ Lk = 0, then the derived series for L
L(k−1)
terminates in the zero subalgebra. Therefore, by defintion, Lk is solvable, making L(k)
Lk−2
solvable, and thus Lk−1 is solvable by Proposition 4.4. Similarly, Lk−1 is solvable, and
L0
thus Lk−2 is solvable. Continuing this way, L1 is abelian and thus L1 is solvable.
Therefore L0 = L is solvable.
Example 4.1.3. We can show that L is solvable if and only if adL is solvable.
Proof.
⇒ Let L(n) = 0. By induction, when n = 0, L(0) = L, and adL(0) = adL = (adL )(0) .
Therefore adL is solvable for n = 0. Assume adL(n) = (adL )(n) ,
4.2 Nilpotency
In this section, we shall explore the descending central series, defined as a recursive
sequence of subalgebras that ends in the zero subspace. We shall rewrite this series
using the quotient space to determine the effect of having a nilpotent algebra as it
relates to it’s subalgebras and their homomorphic images.
46
Thus Li is spanned by long brackets [X1 [X2 [...XI+1 ]...] (which we abbreviate to
[X1 X2 ...Xi+1 ]) [Sam90]. Notice that Ln = 0 makes any abelian algebra nilpotent.
Therefore nilpotent implies solvable since L(i) ⊂ Li for all i. However solvable does not
imply nilpotent.
Example 4.2.1. The algebra n(n, F ) of strictly upper triangular matrices is nilpotent.
k = 0, implies
A matrix is strictly upper triangular if and only if i > j − k. So Given Bij
B k = 0 for k ≥ m if B is an m × m matrix. Using induction, we have B 1 = 0 for
k = 0 for i, j − (k − 1):
i.j − 1. Assume Bij
i
X m
X
k
Bij = BBjk−1 = k−1
Bir Brj + k−1
Bir Brj
r=1 r=i+1
i
X m
X
k−1
= 0 · Brj + Bir · 0 = 0
r=1 r=i+1
Note that the first term goes to zero, since r ≤ i where Bir is strictly triangular. The
second term goes to zero since r ≥ i + 1 > (j − k) + 1 = j − (k − 1). Therefore, by the
k = 0.
induction step, Bij
We can show a similar result for lower triangular matrices using the following indices:
Proposition 4.6.
1. If L is a nilpotent Lie algebra, then all the subalgebras and homomorphic images
of L are nilpotent.
Proof.
47
Example 4.2.2. Let I be an ideal of L. Then each member of the descending central
series of I is also an ideal of L.
Therefore I n+1 is an ideal, then each member of the descending central series of I is an
ideal of L.
Example 4.2.3. We can show that L is nilpotent if and only if adL is nilpotent.
Proof. Using induction, let n = 0, then L0 = L. Therefore adL0 = (adL )0 = adL . Let
adLn = (adL )n . Then
⇒ Let Ln = 0 for some n. By the above induction, adLn = (adL )n = 0 and thus adL is
nilpotent.
⇐ Given adL is nilpotent, (adL )n = 0. Since (adL )n = adLn = 0, Ln = 0, therefore L is
nilpotent.
Example 4.2.4. The sum of two nilpotent ideals of a Lie algebra L is again a nilpotent
ideal. Therefore, L possesses a unique maximal nilpotent ideal.
Example 4.2.5. Let charF = 2. We can prove that L = sl(2, F ) is nilpotent. Recall the
standard basis for L:
1 0 0 1 0 0
x= ,y = ,z =
0 −1 0 0 1 0
[x, z] = [(e11 − e22 ), e21 ] = (e11 − e22 )(e21 ) − (e21 )(e11 − e22 ) = −e21 − e21 = −2e21 = −2z
[x, y] = [(e11 − e22 ), e12 ] = (e11 − e22 )(e12 ) − (e12 )(e11 − e22 ) = e12 − (−e12 ) = 2e12 = 2y
[y, z] = [e12 , e21 ] = e12 e21 − e21 e12 = e11 − e22 = x
49
0 x 0
2
L = [sl(2, F ), [sl(2, F ), sl(2, F )]] = [sl(2, F ), F x] = 0.
The purpose of Engel’s Theorem is to connect the property of Lie algebra nilpotence to
operators on a vector space. A Lie algebra is nilpotent if for a sufficiently long sequence
{xi } of elements of L, the nested adjoint ad(xn )[...ad(x2 )[ad(x1 )[y]]] is zero for all y ∈ L.
That is, applying the adjoint sufficiently many times will kill any element belonging to
L, rending each x ∈ L ad-nilpotent. Engel’s theorem states the converse as true. In
preparation for Engel’s theorem, we must first observe the following lemma, then prove
Theorem 4.8.
Proof. Using induction on dimL. If dim L = 0, then clearly L.v = 0 and the theorem is
true. If dim L = 1, then recall that a single nilpotent linear transformation contains at
least one eigenvector corresponding to eigenvalue 0, thus the theorem is true and
L.v = 0.
50
Given L is a nilpotent Lie algebra, L(n) = 0 for all x1 , ..., xn , x ∈ L. Let adx ∈ gl(L)
with x ∈ L ad-nilpotent, then:
[GG78]
Suppose K 6= L is any nontrivial subalgebra of L. Consider the vector spaces of K and
L. Since K is a subalgebra of a nilpotent endomorphism, adk (l + K) = adk (l) + K is a
well defined map where, for k 0 ∈ K:
Thus adk induces a nilpotent linear transformation on the quotient subspace L/K.
Since K is a subalgebra of L, dimK < dimL. From the induction assumption and
Lemma 4.7, it follows that there exists some non-zero vector x + k 6= k in L/K such that
x + k is killed by the image in gl(L/K). By definition, NL (K) = {x ∈ L|[x, k] ⊂ K},
therefore K is properly in the normalizer of K in L since for x ∈ L, y ∈ K, [x, y] ⊂ K
where x ∈
/ K. Let K be a maximal proper subalgebra of L, then every x ∈ L has
[x, k] ∈ K for all k ∈ K, which makes NL (K) an ideal of L. Consider the dimension of
L/K. If dimL/K > 1, then there exists a 1 dimensional subalgebra F v/K ⊆ L/K. By
the correspondance theorem, K ⊆ F v ⊆ L, but L is not one dimensional. Therefore, K
has codimension 1, which makes L/K one dimensional. Since L/K is 1-dimensional,
Span{l ∈ L} = F (z + L)K for some nilpotent endomorphism z ∈ L. For each l ∈ L,
l + K = αz + K, k ∈ K. Therefore αz ∈ K, and thus l = αz + k for z ∈ l − k.
By induction, W = {v ∈ V |k.v = 0} =
6 0. By the above, K is an ideal of L. If
x ∈ L, y ∈ K, w ∈ V , then since yx − xy = −[x, y] in End (V ), we have:
51
Theorem 4.9. Engel I Let V be a vector space; let L be a sub Lie algebra of the
general linear Lie algebra gl(V ), consisting entirely of nilpotent operators. Then L is a
nilpotent Lie algebra.
Proof. Given a Lie algebra of L having only ad-nilpotent elements, since the adoint is a
linear transformation, adL ⊂ gl(L), which satisfies the hypothesis of Theorem 4.8 when
L 6= 0. Thus there exists an x ∈ L such that [L, x] = 0, which implies that the center of
L is nonzero. Therefore L/Z(L) consists of ad-nilpotent elements, where dim
adL/Z(L) < dimL. Using an induction argument on the dimension of L, similar to
Theorem 4.8, it follows that L/Z(L) is nilpotent. Therefore, by Proposition 4.6 (part
2.), if L/Z(L) is nilpotent, then so is L.
52
Chapter 5
Conclusion
Now that we have covered the basics, we can further explore the root system of the
classical Lie algebras. The full set of roots of a Lie algebra L can be generated from the
set of simple roots and their associated Weyl reflections. If we further explore nilpotent
subalgebras of a Lie algebra that are self-normalising, we invariably must investigate
cartan subalgebras and the killing form.
Differentiating a Lie group action yeids a Lie algebra action. A Lie algebra is a
linearization of a Lie group action, and allows one a doorway to study Lie groups. The
foundation of Lie theory is the exponential map relating Lie algebras to Lie groups,
known as the Lie group-Lie algebra correspondence. We briefly touched on this map
earlier in section 3.3, dealing with Lie algebra automorphisms.
The question arises: what real world applications do Lie algebras possess? Classical
Lie algebra representations have applications in physics. They can be used to study
fluid dynamics. Symmetries in physics take on the structure of mathematical groups,
where continuous groups represent Lie groups. Collisions of vector bosons in quantum
field theory is tied to Lie algebra, specifically gauge theory, and as such, becomes a
mechanism to interpolate physical interactions. The algebra can be promoted to a
group and interpreted as a symmetry (gauge symmetry). Lie algebras are thus integral
to describing theories of nature.
54
Bibliography
[Aig02] Papantonopoulou Aigli. Algebra Pure and Applied. Prentice Hall, New Jersey,
2002.
[Eve66] Howard Eves. Elementary Matrix Theory. Dover Publications, Inc, New York,
1966.
[GG78] Morikuni Goto and Frank D. Grosshans. Semisimple Lie Algebras. Marcel
Dekker, Inc, New York, NY, 1978.
[MT03] Jerrold E. Marsden and Anthony J. Tromba. Vector Calculus. Wiley, New
York, NY, 2003.
[Sam90] Hans Samelson. Notes on Lie Algebras. Springer-Verlag, New York, 1990.