0% found this document useful (0 votes)
28 views22 pages

Ward

Uploaded by

林Arthur
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
28 views22 pages

Ward

Uploaded by

林Arthur
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 22

SEMI-SIMPLE LIE ALGEBRAS AND THEIR

REPRESENTATIONS

ABIGAIL WARD

Abstract. This paper presents an overview of the representations of Lie al-


gebras, particularly semi-simple Lie algebras, with a view towards theoretical
physics. We proceed from the relationship between Lie algebras and Lie groups
to more abstract characterizations of Lie groups, give basic definitions of differ-
ent properties that may characterize Lie groups, and then prove results about
root systems, proceeding towards a proof of the Theorem of Highest Weight.
After proving this theorem, we show how it may be applied to the the Lorentz
group SO(1,3) through representations of su(2).

Contents
1. An introduction to Lie algebras 1
1.1. Lie groups and Lie algebras: some beginning motivation 1
1.2. Definitions and examples 3
1.3. Characterizing properties of Lie algebras 4
1.4. A useful example: sl(2, C) 5
2. Nilpotent sub-algebras and diagonalization results 5
3. Weight Spaces and Root spaces 8
3.1. Weight Spaces 8
3.2. Root spaces 10
4. Structure of Root Systems 12
5. The Universal Enveloping Algebra 15
6. Relating roots and weights 16
7. Applying these results: the Lorentz group and its Lie algebra. 20
Acknowledgments 22
References 22

1. An introduction to Lie algebras


1.1. Lie groups and Lie algebras: some beginning motivation. When study-
ing physics, understanding the symmetries of a system is often key to understanding
the system itself. The symmetries of a system will always form a group (with the
natural group action being the composition of two symmetries); furthermore, since
these symmetries often vary continuously, they will form a Lie group: a topological
group which is also a manifold, such that the group action and the action of taking
inverses are smooth functions.

Date: September 3, 2013.


1
2 ABIGAIL WARD

When studying manifolds, we often pass to studying the tangent space, which
is simpler since it is linear; since Lie groups are manifolds, we can apply this
technique as well. The question then arises: if the tangent space is an infinitesimal
approximation to the Lie group at a point, how can we infinitesimally approximate
the group action at this point? The study of Lie algebras gives us an answer to this
question.
The study of matrix groups is well-understood, so a first step in understanding
Lie groups is often representing the group as a group of matrices; for an example
consider the following:
Example 1.1. Recall the circle group, which may be written {eiθ |θ ∈ [0, 2π)}—we
know we may also represent this group as the set of matrices
  
cos θ − sin θ
|θ ∈ [0, 2π) .
sin θ cos θ
All the Lie groups that we give as examples in this paper will be finite-dimensional
matrix groups. Thus when motivating the study of Lie algebras, let us assume our
group G is a matrix group. We know we may write the tangent space at the identity
to this group as

g = {c0 (0)|c : R → G is a curve with c(0) = 1G


that is smooth as a function into matrices}.
Note that the maps t 7→ c(kt) and t 7→ c(t)b(t) shows that g is closed under addition
and multiplication by scalars. Now, we wish to consider how the group action in
G affects this space. Consider the map t 7→ gc(t)g −1 for g ∈ G; differentiating this
operation shows that this space is closed under the operation [X, Y ] = XY − Y X.
This bracket operation provides the structure for the Lie algebra, which leads to an
understanding of the group action; the bracket form [X, Y ] is the differential form
of the operation x−1 yx.
The above provides a way to obtain a Lie algebra from a Lie group; it is also
sometimes possible to reconstruct the Lie group given the Lie algebra. It can be
shown that for such matrix groups, the map X 7→ etX (where this is just the
standard matrix exponential) is a map from g to G, and for certain groups, this is
map is a bijection. Hence understanding the Lie algebras of these matrix groups
can give useful information about the group itself. In this paper, we do not discuss
how this exponential map forms such a correspond; the curious reader is encouraged
to consult [2].
In this paper, we first study general Lie algebras and then move towards studying
the representations of semi-simple Lie algebras, which are Lie algebras with no
Abelian sub-algebras (i.e. sub-algebras consisting of elements X such that [X, Y ] =
0 for all elements Y in the Lie algebra), with the motivation being that these semi-
simple Lie algebras are the most common Lie algebras in areas of theoretical physics.
We conclude by demonstrating how these Lie algebras arise in physics by briefly
sketching how the study of semi-simple Lie algebras relates to the Lorentz group
and its corresponding Lie algebra.
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 3

1.2. Definitions and examples. A Lie algebra is g is a vector space over a field
K, equipped with an operation [·, ·] : L × L → L satisfying the following properties:
for all X, Y, Z ∈ g, and a, b ∈ K.
• linearity: [aX + bY, Z] = a[X, Z] + b[Y, Z]
• anti-symmetry: [X, Y ] = −[Y, X]
• the Jacobi identity:[X, [Y, Z]] = [[X, Y ], Z] + [Y, [X, Z]]
Example 1.2. If g is an algebra, then g becomes a Lie algebra under the bracket
operation defined by the commutator:
[X, Y ] = XY − Y X.
In particular for a vector space K, EndK (V ) becomes a Lie algebra under this
operation.
As we might expect, a Lie algebra homomorphism is a map between Lie
algebras that respects the vector space and bracket structure: if ϕ : g → g0 is a
Lie algebra homomorphism, then ϕ([X, aY + Z]) = [ϕ(X), aϕ(Y ) + ϕ(Z)] for all
X, Y, Z ∈ g and all a in the underlying field. If a, b ⊂ g are subsets of a Lie algebra
g, then we take [a, b] = span{{[a, b]|{a ∈ a, b ∈ b}. A subalgebra h ⊆ g is a vector
subspace of g such that [h, h] ⊆ h; if in addition [g, h] ⊆ h, then h is an ideal. g
is said to be Abelian if [g, g] = 0; this definition makes sense since the bracket
operation often arises as the commutator of operators1. We may also define in the
usual way a direct sum of Lie algebras.
If g is a vector space over a field K, and V is a vector space over the same
field, and π : g → EndK V is a homomorphism, then π is a representation of g
on V . If π : g → EndK V and π 0 : g → EndK V 0 are two representations, these
representations are equivalent if there exists an isomorphism E : V → V 0 such that
Eπ(X) = π 0 (X)E for all X ∈ g. An invariant subspace for a representation is a
subspace U such that π(X)U ⊆ U for all X ∈ g.
We can consider the representation of g on itself denoted ad and defined by
ad(X)(Y ) = [X, Y ]; that this is a linear map that respects the bracket structure
follows from the linearity of the bracket operation and the Jacobi identity. Fur-
thermore, for any subalgebra h ⊆ g, the adjoint representation restricted to h is a
representation of h on g.
Example 1.3. The Heisenberg Lie algebra is a Lie algebra H over C generated
by elements {P1 , . . . , Pn , Q1 , . . . , Qn , C} satisfying
(1.4) [Pi , Pj ] = [Qi , Qj ] = [Pi , C] = [Qj , C] = [C, C] = 0
(1.5) [Pi , Qj ] = δij C.
(We use C since this denotes the center of the Lie algebra). If we take C = i~1,
we may note that this models the quantum mechanical position and momentum
operators, with Pi and Qi being the momentum and position operators of the i-th
coordinate, respectively.
In the case where n = 1, one representation of this is given by
 
0 a c
π(aP + bQ + cC) = 0 0 b  .
0 0 0
1Indeed, when studying the universal enveloping algebra of a Lie group, we will see that every
bracket operation of two elements can be regarded as a commutator of the two elements
4 ABIGAIL WARD

1.3. Characterizing properties of Lie algebras.


Definition 1.6. Define recursively:
g0 = g, g1 = [g, g], gj+1 = [gj , gj ].
This is the commutator series for g. Note that gj is an ideal for all j. g is
solvable if gj = 0 for some j.
Definition 1.7. Similarly, define recursively
g0 = g, g1 = [g, g], gj+1 = [g, gj ].
This is the lower central series for g. Again note that gj is an ideal for all j. g
is nilpotent if gj = 0 for some j.
Proposition 1.8. A Lie algebra is solvable if it is nilpotent.
Proof. Note that gj ⊆ gj for all j. 
Proposition 1.9. Any sub-algebra or quotient algebra of a solvable Lie algebra
is solvable. Any sub-algebra or quotient Lie algebra of a nilpotent Lie algebra is
nilpotent.
Proof. If h is a subalgebra of g, then hk ⊆ gk , so if g is solvable, h is solvable. If
π : g → g/h is the projection to the quotient algebra, then π(gk ) = hk , so g solvable
imples hk solvable. We may replace ‘solvable’ with ‘nilpotent’ in the above without
changing anything else to obtain the same results for nilpotent Lie algebras. 
Definition 1.10. A finite-dimensional Lie algebra g is simple if g has no nonzero
proper ideals.
Definition 1.11. A finite-dimensional Lie algebra g is semi-simple if g has no
nonzero proper solvable ideals.
This paper focuses on finite-dimensional semi-simple Lie algebras, as many Lie
algebras encountered in physics are of this type. The above definition does not give
as much insight into the structure of a semi-simple Lie algebra as other characteri-
zations do; we list those alternate characterizations below, after first introducing a
bilinear form on Lie algebras, the Killing form:
Let g be a a finite-dimensional Lie algebra over R. For X, Y ∈ g, note that
ad X ad Y ∈ EndV g, and we can define
B(X, Y ) = Tr(ad X, ad Y ).
Then B is a symmetric bilinear form which we call the Killing form. It is also
associative, in the sense that B([X, Y ], Z) = B(X, [Y, Z]) for all X, Y, Z ∈ g, which
is clear from the fact that Tr(AB) = Tr(BA).
Having introduced the Killing form, we may now list useful alternate character-
izations of semi-simplicity:
Theorem 1.12. Let g be a finite-dimensional Lie algebra. The following are equiv-
alent:
(a) g is semi-simple.
(b) its Killing form is non-degenerate ( Cartan’s criterion for semi-simplicity).
(c) it is the direct sum of simple Lie algebras.
(d) if it has no Abelian sub-algebras.
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 5

We omit the proof of these statements. In many cases it is obvious that a Lie
algebra is semi-simple from (b) or (c), and some texts (especially physics texts)
often take (c) as the definition.
1.4. A useful example: sl(2, C). One of the most commonly encountered Lie
algebraz is
sl(n, C) = X ∈ gl(n, C)| Tr(X) = 0}
That this is a Lie algebra follows from the fact that Tr(AB) = Tr(BA) for all
matrices A, B, so Tr([A, B]) = 0 for any A, B ∈ gl(n, C)| Tr(X) = 0.
Remark 1.13. sl(2, C) has a basis given by
     
1 0 0 1 0 0
h= , e= , f= .
0 −1 0 0 1 0
The study of the representations of (2, C) has implications in classifying the
structure of semi-simple Lie algebras. While we do not prove the major theorems
about these representations found in e.g. [1], we do present a result that has
important implications in our study. If we let ϕ be a representation of a Lie algebra
g on a vector space V , then ϕ is completely reducible if there exist invariant
subspaces U1 , . . . , Un such that V = U1 ⊕ . . . ⊕ Un and the restriction of ϕ to Ui is
irreducible.
We have the following (which is unproven here).
Theorem 1.14. If ϕ is a representation of a semi-simple Lie algebra on a finite
dimensional vector space, then ϕ is completely reducible; in particular, if ϕ is a
representation of sl(2, C), then ϕ is completely reducible.
We also have the following, which is specific to sl(2, C):
Theorem 1.15. With h the basis vector for sl(2, C) above, ϕ(h) is diagonalizable
with all eigenvalues integers and the multiplicity of each eigenvalue k equal to the
multiplicity of −k.
For a proof of these results, see [1].

2. Nilpotent sub-algebras and diagonalization results


Our study of semi-simple Lie algebras begins with the following results, which
allows us to understand how solvable Lie algebras can be understood to act jointly
on a space:
Theorem 2.1 (Lie’s theorem). Let g be solvable, let V 6= 0 be a finite-dimensional
vector space over C, and let π : g → EndC V be a representation. Then there is a
simultaneous eigenvector v 6= 0 for all members ofπ(g). More generally, for V a
vector space over a subfield C ⊂ C, and if all the eigenvalues of π(X), X ∈ g lie in
K, then there exists such an eigenvector.
Proof. Proceed by induction on the dimension n of g. If n = 1, then π(g) is one
dimensional and the result is immediate.
Now, assume the result for Lie algebras of dimension k < n. Consider the com-
mutator ideal h = [g, g]; this must be a proper ideal since otherwise the commutator
series would never terminate. As [g, g] is a proper subspace of g, it must have di-
mension less than n; we may therefore find a subspace h ⊂ g of codimension 1
6 ABIGAIL WARD

such that [g, g] ⊆ h. Now, note that [g, h] ⊆ [g, g] ⊆ h, so h is an ideal. As any
subalgebra of a solvable Lie algebra is solvable, we have that h is solvable, and since
it has dimension n − 1, we may apply the induction hypothesis to obtain a vector
v ∈ V such that v is a simultaneous eigenvector for all H ∈ h. We then have for
each H ∈ h, π(H)v = λ(H)v, where λ : h → K is a functional.
Now, choose nonzero X ∈ g not in h, and define recursively:
e−1 = 0, e0 = v, ep = π(X)ep−1
and let E = span{e0 , e1 , . . .}. Note that π(X)E ⊆ E, so we may regard π(X) as
an operator on E; then, since all the eigenvalues of π(X) lie in K, we may chose
an eigenvector w ∈ E of π(X).
Denote Ep = span{e0 , e1 , . . . , ep }. We first show that π(H)ep ≡ λ(H)ep mod Ep−1 .
We may proceed by induction on p. We chose e0 = v such that this condition holds.
Now assume the result for p and let H ∈ h. We have
π(H)ep+1 = π(H)π(X)ep
= π([H, X])ep + π(X)π(H)ep .

Now, note that π([H, X]) ∈ h, so by induction we have π([H, X])ep ≡ λ([H, X])ep
mod Ep−1 . We also have that π(H)ep ≡ λ(H)ep mod Ep−1 , so
π(X)π(H)ep ≡ π(X)λ(H)ep mod span{Ep−1 , π(X)Ep−1 }.
But
span{Ep−1 , π(X)Ep−1 } = span{e0 , . . . , ep−1 , π(X)e0 , . . . , π(X)ep−1 } = Ep ,
so we have that
π(H)ep+1 ≡ λ(H)ep + λ(H)π(X)ep mod Ep
≡ λ([H, X])ep + λ(H)π(X)ep mod Ep
≡ 0 + λπ(H)π(X)ep mod Ep
≡ λπ(H)ep+1 mod Ep
and so the result follows by induction.
Now, we may chose n such that {e1 , . . . , en } form a basis for E. Then as
π(H)ep ≡ λ(H)ep mod Ep−1 , relative to this basis, π(H) when considered as an
operator on E has the form
 
λ(H) ?

 λ(H) 

 .. .. 
 . . 
0 λ(H)
(i.e. is upper triangular with all diagonal entries λ(H).) We see then that Tr(π(H)) =
λ(H) dim E. Applying this result to [H, X], we obtain that
λ[H, X] dim E = Tr π([H, X]) = Tr[π(H), π(X)] = 0
(where here we have used the fact that Tr[A, B] = 0 for any matrices A and B).
Since dim E 6= 0, we have that λ[H, X] = 0. We may use this to refine our pre-
vious result. Before, we had π(H)ep ≡ λ(H)ep mod Ep−1 ; we wish to refine this
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 7

to π(H)ep = λ(H)ep . Again, we proceed by induction, with the result true by


definition for p = 0. Then using the inductive hypothesis, we have that
π(H)ep+1 = π(H)π(X)ep
= π([H, X])ep + π(X)π(H)ep
= λ([H, X])ep + λ(H)π(X)ep
= λ(H)π(X)ep = λ(H)ep+1

and we may conclude that π(H)ep = λ(H)ep for all p. Then π(H)e = λ(H)e for
all e ∈ E and in particular π(H)w = λ(H)w. Since then w is an eigenvector for
all H ∈ h, and we chose w to be an eigenvector of X, we conclude that w is a
simultaneous eigenvector for all G ∈ g.

Corollary 2.2. Under the assumptions on g, V, π, and K as above, there exists a
sequence of subspaces
V = V0 ⊇ V1 ⊇ . . . ⊇ Vm = 0
such that each Vi is stable under π(g) and dim Vi /Vi+1 = 1. This means that there
exists a basis with respect to which all the matrices of π(g) are upper triangular. ??
Proof. Proceed by induction on the dimension of V . If m = 1, the result is obvious.
For m > 1, the above theorem tells us that we may find a simultaneous eigenvector
v of π(X) for all X ∈ g. Let U be the subspace spanned by this vector. Then
since π(g)U = U , π acts a representation π̃ on the quotient space V /U , which has
dimension strictly less than V . By the inductive hypothesis we may thus find a
sequence of subspaces
W = W 0 ⊇ V 1 ⊇ . . . ⊇ Wm = 0
such that each Wi is stable under π̃(g) and dim Wi /Wi+1 = 1. Now, consider the
corresponding sequence of subspaces in V , where each Vi has image Wi under the
projection to the quotient space; we obtain the sequence
V = V0 ⊇ V1 ⊇ . . . ⊇ Vm = 0
which satisfies the desired properties. Hence by induction, such a sequence may
always be found.
Now, chose a vector vi ∈ Vi−1 for 1 ≤ i ≤ m such that Kvi + Vi = Vi−1 ; with
respect to the basis {v1 , . . . , vm }, all matrices of π(g) are upper triangular. 
The following result allows us to relate the property of nilpotency of a Lie algebra
and nilpotentcy of its adjoint representation:
Theorem 2.3. A Lie algebra h is nilpotent if and only if the Lie algebra ad H is
nilpotent.
Proof. Let H ∈ h. We have
[[...[Hk+1 , Hk ], Hk−1 , . . . , H1 ] = ad[...[Hk+1 , Hk ], ..., H2 ](H1 )
so using the fact that
ad[...[Hk+1 , Hk ], ..., H2 ] = [. . . [ad Hk+1 , ad Hk ], . . . , ad H2 ]
8 ABIGAIL WARD

shows us that if h is nilpotent, then the left side is 0 for high enough k, and
conversely, if ad h is nilpotent, then the right side is 0 for high enough k and hence
h is nilpotent. 
In the future, these theorems will allow us to put matrices in our Lie algebras
into useful forms.

3. Weight Spaces and Root spaces


3.1. Weight Spaces. Let h be a finite-dimensional Lie algebra over C, and let π be
a representation of h onto a complex vector space V . For such π and V , whenever
α is in the dual h∗ , let Vα be defined as
{v ∈ V |(π(H) − α(H)1)n v = 0 for all H ∈ h and some n = n(H, v).}
These look like generalized eigenspaces. Clearly, we may assume that n ≤ dim V .
Definition 3.1. If Vα 6= 0, Vα is called a generalized weight space and α is a
weight. Members of Vα are generalized weight vectors.
Proposition 3.2. Suppose that h is a nilpotent Lie algebra over C and that π is
a representation of h on a finite-dimensional complex vector space V . Then there
are only finitely many generalized weights, each generalized weight space is stable
under h, and V is the direct sum of all the generalized weight spaces.
Proof. First , we will show each Vα is stable under h. For H ∈ h, let
Vα,H = {v ∈ V |(π(H) − α(H)1)n v = 0 for some n = n(H, v).}
We wish to show that Vα,H is invariant under π(H); to do this, we use the nilpotency
of h. Since h is nilpotent, ad H is nilpotent as a linear map by Theorem 2.3. We
may define
h(m) = {Y ∈ h|(ad H)m Y = 0},
and the nilpotency guarantees that h = ∪dm=0 h(m) (where d is the dimension of h).
We wish to show that π(Y )Vα,H ⊆ Vα,H for Y ∈ h(m) by induction on m.
For m = 0, we have that h(0) = 0 and the result follows trivially. Now, assume
the result is true for m − 1. Let Y ∈ h(m) , and note then that [H, Y ] ∈ h(m−1) . We
have that
(π(H) − α(H)1)π(Y ) = π(H)π(Y ) − α(H)π(Y )
= π([H, Y ]) + π(Y )π(H) − α(H)π(Y )
= π([H, Y ]) + π(Y )(π(H) − α(H)1).

Some computation shows that


(π(H) − α(H)1)2 π(Y )
= (π(H) − α(H)1)π(Y )(π(H) − α(H)1) + (π(H) − α(H)1)π([H, Y ])
= π(Y )(π(H) − α(H)1)2 + (π(H) − α(H)1)π([H, Y ]) + π([H, Y ])(π(H) − α(H)1),
and we see that for a given n,
(π(H) − α(H)1)n π(Y ) = π(Y )(π(H) − α(H)1)n
n−1
(π(H) − α(H)1)n−s−1 π([H, Y ])(π(H) − α(H)1)s .
X
+
s=0
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 9

Now, recall that for m ≥ dim V , for v ∈ Vα,H , (π(H) − α(H)1)m v = 0. In the
above expression, take n = 2 dim V .Then for v ∈ Vα,H , we have that
(π(H) − α(H)1)n π(Y )v = π(Y )(π(H) − α(H)1)n v
n−1
(π(H) − α(H)1)n−s−1 π([H, Y ])(π(H) − α(H)1)s v
X
+
s=0
n/2
(π(H) − α(H)1)n−s−1 π([H, Y ])(π(H) − α(H)1)s v
X
=
s=0

(all other terms vanish). Now, consider each remaining term of the form
(π(H) − α(H)1)n−s−1 π([H, Y ])(π(H) − α(H)1)s v.
Note that (π(H) − α(H)1)s v ∈ Vα,H , and since [H, Y ] ∈ h(m−1) , we have by the
inductive hypothesis that [H, Y ] leaves Vα,H stable, so
π([H, Y ])(π(H) − α(H)1)s v ∈ Vα,H .
But then since n − s − 1 ≥ dim V , (π(H) − α(H)1)n−s−1 acts as 0 on Vα,H , so we
obtain that
n/2
(π(H) − α(H)1)n−s−1 π([H, Y ])(π(H) − α(H)1)s v = 0.
X

s=0

The sum above then gives us


(π(H) − α(H)1)n π(Y )v = 0.
Thus π(Y )v ∈ Vα,H , and we conclude that Vα,H is stable under π(Y ); then by
induction, π(Y )Vα,H ⊆ Vα,H for all Y ∈ h.
Now we are in a position to prove that V may be written as the direct sum of
the weight spaces. First, chose a basis H1 , . . . , Hr for h. Since we work over C, we
know we may write V as the direct sum of the generalized eigenspaces of π(H1 ):
M
V = Vλ,H1
λ

Now, note that we may regard each λ as µ(H1 ) for some µ ∈ h∗ .


We may then write
M
V = Vµ(H1 ),H1
values of µ(H1 )

(this direct sum decomposition makes since there are only finitely many µ(H1 ) such
that Vµ(H1 ),H1 6= 0). Now, note that for each µ ∈ h∗ , Vµ(H1 ),H1 is invariant under
π(h), so in particular is invariant under H2 ; Thus we may further decompose V in
the same way:
M M
V = Vµ(H1 ),H1 ∩ Vµ(H1 ),H2 ,
values of µ(H1 ) values of µ(H2 )

and so on, iterating to obtain


 
M r
\
V =  Vµ(Hj ),Hj  .
values of µ(H1 ) ... j=1
10 ABIGAIL WARD

Each of these spaces is invariant under π(h). Now, by Corollary ?? (which is


applicable since h is nilpotent and hence solvable), there is some basis in which all
π(Hi ) are upper triangular matrices. P Then in this basis, for each Hi , π(Hi ) must
have diagonal entries
P µ(H i ). Thus π( ci Hi ) must be a triangular matrix
Pr with all
diagonal entries ci µ(Hi ).T If we define a linear functional L
α by α( i=1 ci Hi ) =
r
ci µ(Hi ), we conclude that j=1 Vµ(Hj ),Hj = Vα . Hence V = α Vα . 
3.2. Root spaces. Now, let g be a Lie algebra, h a subalgebra, and recall that
the adjoint representation defined by ad H(X) = [H, X] is a representation of h on
g. The weights of ad h relative to this representation are called roots. There are
many statements we can make about the roots.
Proposition 3.3. If g is any finite-dimensional Lie algebra over C and h is a
nilpotent Lie subalgebra, then the generalized weight spaces of g relative to adg h
satisfy
L
(a) g = gα
(b) h ⊆ g0
(c) [gα , gβ ] ⊆ gα+β . If α+β is not a root, then this statement says that [gα , gβ ] = 0.
Proof. (a) This is by the above proposition.
(b) Since h is nilpotent, for all H ∈ h, there is an n such that (ad H)n = 0 on H,
which immediately gives that h ⊆ g0 .
(c) Let A ∈ gα and B ∈ gβ . Let H ∈ h. Note that we have
(ad(H) − α(H) − β(H))[X, Y ]
= [H, [X, Y ]] − α(H)[X, Y ] − β(H)[X, Y ]
= [[H, X], Y ] + [X, [H, Y ]] − α(H)[X, Y ] − β(H)[X, Y ]
= [(ad H − α(H))X, Y ] + [X, (ad H − β(H))Y ].
Induction gives that
(ad(H) − α(H) − β(H))n [X, Y ]
n  h
X k k n−k
i
= (ad H − α(H)) X, (ad H − β(H)) Y
n
k=1
(this is just the Binomial Theorem and the calculations above). Then chosing
n > 2 dim V guarantees that either k > dim V or n − k > dim V , so [(ad H −
α(H))k X, (ad H − β(H))n−k Y ] = 0. Thus [A, B] ∈ gα+β .

Note that (c) above implies that g0 is a subalgebra. If g0 = h, we call h a Cartan
subalgebra. It can be shown ([1, p. 90]) that any finite-dimensional Lie algebra
over C has a Cartan subalgebra.
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 11

The above definition of a Cartan subalgebra is hard to work with; the following
characterization is more useful:
Proposition 3.4. If g is a finite-dimensional Lie algebra and h is a nilpotent
subalgebra, then h is a Cartan subalgebra if and only if
h = Ng (h) = {X ∈ g|[X, h] ⊆ h}.
Proof. First note that if h is any nilpotent subalgebra, we have that h ⊆ Ng (h) ⊆ g0 :
the first of these inclusions holds because h is a subalgebra, and the second holds
because ad H n X = ad H n−1 [H, X], and ad H n−1 is 0 on h for large enough n, since
h is nilpotent.
Now, assume that h is not a Cartan subalgebra, so g0 6= h. Then g0 /h is nonzero,
and as the quotient of a solvable Lie algebra, is solvable. By Lie’s theorem, there
exists a nonzero X̃ + h ∈ g0 /h that is a simultaneous eigenvector of ad h, and its
eigenvalue has to be zero (as otherwise X would not be in g0 ). But then for H ∈ h,
[H, X] ∈ h, so X is not in h but is in Ng (h). Thus h 6= Ng (h).
Conversely, note that if h is a Cartan subalgebra, g0 = h, so Ng (h) = h. 
In particular, semi-simple Lie algebras have easily understood Cartan sub-algebras.
Proposition 3.5. If g is a complex finite-dimensional semi-simple Lie algebra and
h is a Cartan subalgebra then h is Abelian.
Proof. Since h is nilpotent, ad h is solvable and we may apply Lie’s theorem to
obtain a basis for g in which ad h is simultaneously triangular. Now, note that for
triangular matrices, Tr(ABC) = Tr(BAC), so for any H1 , H2 , H ∈ h, we have
B(ad[H1 , H2 ], ad H) = 0
Now, let H ∈ h and X ∈ gα where α is any nonzero weight. By the above
proposition, may find a basis
{Gα1 ,1 , . . . , Gα1 ,n1 , . . . , Gαm ,1 , . . . , Gαm ,nm }
where each set {Gαi ,1 , . . . , Gαi ,ni } is a basis for gαi , and the set of all such Gα,i
forms a basis for g. Now, recall that for any B ∈ gβ ,
ad H(ad XB) = ad H([X, B]) ∈ gα+β .
In this basis, we see that ad H ad X has all zeros on the diagonal, so Tr(ad H ad X) =
0. In particular, for any H1 , H2 ∈ h, Tr(ad[H1 , H2 ] ad X) = 0.
We thus have proven for any X ∈ gα or X ∈ h, for any H1 , H2 ∈ H,
Tr(ad[H1 , H2 ] ad X) = 0. Since any X ∈ g may be written as a linear combination
of such elements and because trace is linear, we obtain that for any H1 , H2 ∈ h,
any X ∈ g,
B([[H1 , H2 ], X]) = 0.
Since g is semi-simple, the Killing form is non-degenerate; hence [H1 , H2 ] = 0 for
any H1 , H2 ∈ h. 
12 ABIGAIL WARD

We also have a partial converse:


Proposition 3.6. If g is a complex semi-simple Lie algebra and h is a maximal
Abelian sub-algebra such that adg h is diagonalizable, then h is a Cartan subalgebra.

Proof. h is Abelian and hence nilpotent, so we may decompose g as


M
g = g0 ⊕ gα .
α6=0

Then since
L adg h is simultaneously diagonalizable, we know we may write g0 as
g0 = h τ , where [h, τ ] = 0. Now, note that if τ is nonzero, then there exists
X ∈ τ , and [h + CX, h + CX] = [h, h] + [h, CX] + [X, X] = 0. Hence h + CX is an
Abelian subalgebra properly containing h, a contradiction. Hence τ = 0 and h is a
Cartan subalgebra. 

We will show later that all maximal Abelian sub-algebras are diagonalizable, so
we may leave out that assumption. Many texts take this as the definition of a
Cartan subalgebra.
Furthermore, we have the following: (for a proof, see [1], pp. 92-93).
Proposition 3.7. Any two Cartan sub-algebras of a finite-dimensional complex
Lie algebra g are conjugate by an inner automorphism of g.
Corollary 3.8. All Cartan sub-algebras h of g have the same dimension; this is
the rank of g.

4. Structure of Root Systems


Again, let g be a complex semi-simple Lie algebra, B its Killing form, and h a
Cartan subalgebra of g. Recall that if we define for α ∈ h∗
{X ∈ g|(ad H − α(H)1)n X = 0for some n ∈ N, for all H ∈ h}
and denote ∆ to be the set of nonzero roots (i.e. nonzero α ∈ h∗ such that gα 6= 0),
then we may write
M
g=h⊕ gα .
α∈∆
We will soon show that we may take n = 1, which simplifies this expression
significantly.
There are many statements we can make about these roots:
Proposition 4.1. (a) If α and β are in ∆∪{0} and α+β 6= 0, then B(gα , gβ ) = 0.
(b) If α ∈ ∆ ∪ {0}, then B(gα , g−α ) 6= 0.
(c) If α ∈ ∆, then −α ∈ ∆
(d) B|h,h is nondegenerate. Hence for each α ∈ ∆, there is a corresponding Hα ∈ h
with α(H) = B(H, Hα ) for all H ∈ h.
(e) ∆ spans h∗ , and {Hα } spans h
Proof. (a) We know that ad gα ad gβ (gγ ) ⊆ gα+β+γ , so we see than in a basis
compatible with the root space decomposition of g, for any A ∈ gα and any
B ∈ gβ , ad A ad B must have all zeros on the diagonal. Hence Tr(ad A ad B) =
B(A, B) = 0, and we conclude that B(gα , gβ ) = 0.
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 13

(b) Recall L that since g is semi-simple, B is nondegenerate. Since we may write


g = h α∈∆ gα and we know B(gα , gβ ) = 0 for all β with β 6= −α, but
B(gα , g) 6= 0, we must have B(gα , g−α ) 6= 0.
(c) This is immediate from the above, since if there were no root space g−α , we
would have B(gα , g) = 0, which is impossible since B is nondegenerate.
(d) By the above, we see that B|h×h is non-degenerate: if H ∈ h is such that
B(H, X) = 0 for all X ∈ H, then the fact that B(H, gβ ) = 0 for all β 6= 0
gives that H is such that B(H, Y ) = 0 for all Y ∈ g. Thus H 7→ B(·, H) is an
isomorphism from h to h∗ , and thus there exists Hα such that α(H) = B(H, Hα )
for some Hα ∈ g.
(e) Note that if ∆ spans h∗ , then for all nonzero H ∈ h, there exists α ∈ h∗ such
that α(H) 6= 0. Thus suppose H ∈ h has α(H) = 0 for all α ∈ ∆. Then since
we may decompose g into root space and on each root space H is nilpotent,
ad H is nilpotent. Then since ad h is Abelian, (ad H 0 ad H)n = ad H 0n ad H n
for any H 0 ∈ h, so ad H 0 ad H is also nilpotent. Hence Tr(h, H) = B(h, H) = 0;
but since B is non-degenerate, this implies H = 0. Hence ∆ spans h∗ . Since h
is finite-dimensional it is isomorphic to h∗ , and we conclude that {Hα } spans
h.


Now, for each α ∈ ∆ ∪ {0} and consider the action of h on gα . Since h fixes
gα and ad h is nilpotent and hence solvable, by Lie’s theorem that we may find a
simultaneous eigenvector for ad h on gα . Let Eα be such an eigenvector; then for
all H ∈ h, [H, Eα ] = α(H)Eα .
We have the following further information about the roots of h:
Proposition 4.2. (a) If α is a root, X is in g−α , then [Eα , X] = B(Eα , X)Hα .
(b) If α, β ∈ ∆, then β(Hα ) = qα(Hα ) for some q ∈ Q.
(c) If α ∈ ∆, then α(Hα ) 6= 0.
Proof. (a) We have that
B([Eα , X], H) = B(X, [H, Eα ])
= α(H)B(X, Eα )
= B(Hα , H)B(Eα , X)
= B(B(Eα , X)Hα , H)
and since B is nondegenerate, we conclude that B(Eα , X)Hα = [Eα , X].
0
(b) By Proposition 4.1b we can chose X−α ∈ g such that B(Eα , X−α ) 6= 0; further-
more, any such X−α ∈ g−α . After normalizing, we have that B(Eα , X−α ) = 1.
Then by a, we have that [Eα , X−α ] =LHα .
Now, consider the subspace g0 = n∈Z gβ+nα , which by Proposition 4.1b
is invariant under ad Hα . Now, consider the trace of ad Hα considered as an
operator on this subspace in two ways. First, note that for any n, ad Hα acts
on gβ+nα with generalized eigenvalue β(Hα ) + nα(Hα ), so we have
X
Tr(ad Hα ) = (β(Hα ) + nα(Hα )) dim gβ+nα .
n∈Z
0
We also know that g is invariant under Eα and X−α , so we regard these
as operators on this subspace. Since we have that [Eα , X−α ] = Hα , we then
14 ABIGAIL WARD

have Tr(ad Hα ) = Tr([Eα , X−α ]) = 0 Tr(ad[Eα , X−α ]) = 0. Thus the equation


above shows that β(Hα ) is a rational multiple of α(Hα ).
(c) If α(Hα ) = 0, then the above shows that β(Hα ) = 0 for all β ∈ g. But since we
know that ∆ forms a basis for h∗ , this implies that Hα = 0, which is impossible
by how we chose Hα . Hence α(Hα ) 6= 0.

Proposition 4.3. For all α ∈ ∆, dim gα = 1, and nα ∈
/ ∆ for any n ≥ 2.
Proof. Again choose X −α ∈ g−α with B(Eα , X−α ) = 1 to obtain Hα = [Eα , X−α ].
Let g0 = CEα ⊕CHα ⊕ n<0 gnα . Then this subspace is invariant under ad Hα , ad Eα ,
L
and ad X−α . We know that Hα = [Eα , X−α ] and Eα , Xα ∈ g0 , so when consid-
ered as an operator on g0 , Tr(Hα ) = Tr([Eα , Xα ]) = 0. However, we also know
that ad Hα (Eα ) = α(Hα )Eα , that ad Hα (CHα ) = 0, and that ad Hα acts on each
summand with generalized eigenvalue nα(Hα ). We thus see that
X
Tr(ad Hα ) = α(Hα ) + nα(Hα ) dim gnα .
n<0
Setting this equal to zero, we see that
X
α(Hα ) = − nα(Hα ) dim gnα
n<0

and since the above tells us that α(Hα ) 6= 0, we have that


X
n dim gnα = −1
n<0
or reversing the sign of the indices,
X
n dim g−nα = 1.
n>0
Thus we see that dim−nα = 1 for n = 1, and 0 otherwise. 
This result also allows for some very useful results: We had
gα = {X ∈ g|(ad H − α(H)1)n X = 0 for some n ∈ N, for all H ∈ h}.
We can now simplify this:
Corollary 4.4. For any α ∈ ∆, gα = {X ∈ g| ad HX = α(H)X}; in other words
we can always take n = 1 above.
Proof. By the above, we know that dimα = 1 for all α ∈ ∆. 
Corollary 4.5. The action of ad h on g is simultaneously diagonalizable.
L
Proof. This follows from the fact that g = h ⊕ α∈∆ gα and that each gα is one
dimensional. 
This gives the promised generalization of Proposition 3.6.
Corollary 4.6. For a semi-simple complex Lie algebra g, a subalgebra h ⊂ g is a
Cartan subalgebra if and only if h is a maximal Abelian subalgebra.
Corollary 4.7. The pair of vectors Eα , E−α may be normalized so B(Eα , E−α ) =
1: that is, B(Eα , E−α ) 6= 0.
Proof. gα is one dimensional for all α, and B(gα , g−α ) 6= 0. 
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 15

With this normalization and proposition 4.2 above, we have the relations
[Hα , Eα ] = α(Hα )Eα ,
[Hα , E−α ] = −α(Hα )E−α ,
[Eα , E−α ] = Hα
After re-normalization, we obtain that
[Hα0 , Eα0 ] = 2Eα ,
[Hα0 , E−α0 ] = −2E−α ,
[Eα0 , E−α
0
] = Hα0 .
Then some checking shows that
     
0 1 0 0 0 1 0 1
Hα 7→ , Eα 7→ , E−α0 7→
0 −1 0 0 0 0
0
defines an isomorphism between the vector space spanned by Hα0 , Eα0 , and E−α and
the vector space spanned by these three matrices. This the vector space
  
a b
|a, b, c ∈ C = sl(2, C).
c −a
So, we conclude that every semi-simple Lie algebra contains within it embedded
copies of sl(2, C).

5. The Universal Enveloping Algebra


With all this in mind, we are almost in a position to state our final results, but
we need to define one more concept: that of the universal enveloping algebra.
Definition 5.1. For a complex Lie algebra g, the universal enveloping algebra
U (g) is a complex associative algebra with identity having the following property:
there exists a Lie algebra homomorphism i : g → U (g) such for any complex
associative algebra A with identity and Lie algebra homomorphism ϕ : g → A (so
that for all X, Y ∈ g, ϕ(A)ϕ(B) − ϕ(B)ϕ(A) = ϕ([X, Y ])), there exists a unique
extension ϕ̃ : U (g) → A such that the diagram
ϕ
g A

i ϕ̃

U (g)

commutes.
A universal enveloping algebra g can be explicitly constructed for any complex
Lie algebra; for the details of this construction, refer to [1, pp. 164–180]. Further-
more, any two universal enveloping algebras are the same up to unique isomorphism.
Proposition 5.2. Representations of g on complex vector spaces are in one-to-one
correspondence with left U (g) modules, with the correspondence being π 7→ π̃.
Proof. If π : g → EndC (V ) is a representation of g on V , then by the definition
of U (g), there exists an extension π̃ : U (g) → EndC (V ), and V becomes a left
U (g) module with the left action being defined as uv = π̃(u)v for all u ∈ U (g).
16 ABIGAIL WARD

Conversely, if V is a left V module, then we may define for all X ∈ g π(X)v = i(X)v,
and since i is a Lie algebra homomorphism, this defines a representation. Since
π̃ ◦ i = π these two operations are inverses, and thus this correspondence is one-to-
one. 
To proceed, we need an important theorem about the universal enveloping al-
gebra which describes a basis for U (g). We want to describe this basis in terms of
the basis for g, and we may do this as follows:
Theorem 5.3. (Poincaré-Birkhoff-Witt). Let g be a complex Lie algebra, and let
{Xi }i∈A be a totally ordered basis for g (note that here this basis can be uncountable,
although we will always work with A finite). Then the set of all finite monomials
i(Xi1 )j1 · · · i(Xik )jk
with all jk ≥ 0 forms a basis for U (g).
Corollary 5.4. The canonical map i : g → U (g) is one-to-one.
The proof of the Poincaré-Birkhoff-Witt theorem, while not difficult, is rela-
tively long, and is omitted here; for reference, see [[1] pp. 168-171]. This theorem
gives us many useful results, among of which is the following, which follows almost
immediately:
Corollary 5.5. If h is a Lie subalgebra of g, then the associative subalgebra of U (g)
generated by 1 and h is canonically isomorphic to U (h).

6. Relating roots and weights


Finally, we are in a position to tie together all of this information. Before stating
the final theorem, we must define an ordering on the roots α; one way to do this is
pick a basis and order the roots lexographically relative to this basis, so α1 > α2 if
α1 (Hi ) = α2 (Hi ) for all i < j, and α1 (Hj ) > α2 (Hj ).
We also make two more definitions that describe linear functionals on the Cartan
subalgebra h:
Definition 6.1. Let λ ∈ h∗ ; we call λ algebraically integral if 2hλ, αi/kαk2 is
an integer for all roots. We call λ dominant if hλ, αi ≥ 0 for all positive roots.
Proposition 6.2. Let g be a complex semi-simple Lie algebra, let h be a Cartan
subalgebra, letP∆ be the nonzero roots of the adjoint representation of h over g,
and let h0 = α∈∆ RHα . If ϕ is a representation of g on the finite-dimensional
complex vector space V , then:
(a) ϕ(h) acts diagonally on V , so that every generalized weight vector is a weight
vector and V is the direct sum of all the weight spaces.
(b) Every weight is real-valued on h0 .
(c) Every weight is algebraically integral.
(d) Roots and weights are related by ϕ(gα )Vλ ⊆ Vλ+α .
Proof. (a, b) We saw above that for any nonzero root α and corresponding
root vectors Eα , E−α spansa subalgebra
 slα isomorphic to sl(2, C) with
1 0
2|α|−2 Hα corresponding to (where the 2|α|−2 is just a normal-
0 −1
ization constant.)Thus π(slα ) is a finite-dimensional representation of slα ,
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 17

and thus Example 2 shows that ϕ(2|α|−2 Hα ) is diagonalizable with inte-


ger eigenvalues. Since the collection of Hα span h and since diagonalizable
commuting operators are simultaneously diagonalizable, this proves that
ϕ(h) acts diagonally on V and is real valued, which proves (a) and (b).
(c) Let λ be a weight and choose a nonzero V ∈ Vλ . Then the above shows
that
ϕ(2|α|−2 Hα −) = 2|α|−2 ϕ(Hα )v = 2|α|−2 λ(Hα )v = 2|α|−2 hλ, αi
is an integer multiple of v, so 2|α|−2 hλ, αi is an integer and λ is algebraically
integral.
(d) Let Eα ∈ gα , let v ∈ Vλ , and let H ∈ h. Then
ϕ(H)ϕ(Eα )v = ϕ(Eα )ϕ(H)v + ϕ([H, Eα ])v
= λ(H)ϕ(Eα )v + α(H)ϕ(Eα )v
= (λ + α)HEα v.
so we conclude that ϕ(Eα )v ∈ Vλ+α .

Definition 6.3. A representation ϕ of a Lie algebra g on a vector space V is
irreducible if ϕ(g) has no irreducible subspaces besides for 0 and V .
With all of this built up, we can finally prove the following:
Theorem 6.4. (Highest Weight Theorem) Let g be a complex semi-simple Lie
algebra, let h be a Cartan subalgebra, let ∆ be the set of roots of the adjoint repre-
sentation, and introduce an ordering on ∆. Let h0 be as above. Let ∆+ denote the
positive roots.
If two irreducible representations ϕ1 and ϕ2 of g have the same highest weight
λ, then ϕ1 and ϕ2 are equivalent and may be labeled ϕλ . In addition:
(a) λ is independent of the ordering on h0 .
(b) Vα , the weight space of λ, is one dimensional.
(c) For all α ∈ ∆+ , the root vector Eα annihilates Vλ , and the members of Vλ are
the only vectors with this property.
P`
(d) every weight of ϕλ is of the form λ − i=1 ni αi with each ni ≥ 0 and the
αi ∈ ∆+ .
We also have the following, which we will not prove, but is useful in the following
section where we apply these results:
Theorem 6.5. For any dominant algebraically integral linear functional λ ∈ h∗ ,
there exists such an irreducible representation ϕλ .
Proof. (of Theorem 6.4). First, we prove that such a correspondence exists. Let
ϕ be an irreducible representation of g on a vector space V . Let λ be the highest
weight of this representation. By Proposition 6.2, we have that λ is algebraically
integral.
If α ∈ ∆+ , then λ + α cannot be a weight, since we chose λ to be the highest
weight. Then by Proposition X, ϕ(Eα )Vλ ⊆ Vλ+α = 0, and the first conclusion in
c follows.
Now, extend ϕ to be defined on all of U (g) with ϕ(1) = 1, as discussed above
in the definition of U (g). First, we may show that ϕ(U (g))v = V for all v ∈ V .
18 ABIGAIL WARD

Proceed as follows: Let w ∈ V . Define recursively W0 = Cv, Wn = ϕ(U (g))Wn−1 .


Then since ϕ is irreducible, we have that W0 ⊂ W1 ⊂ . . . Wn = V , with each
inclusion strict since ϕ is irreducible; this process must thus terminate at n ≤ dim V .
But then we have that there exists X1 , . . . , Xn ∈ U (g) such that ϕ(w1 ) . . . ϕ(wn )v =
w, and this gives that ϕ(X1 ) . . . ϕ(Xn )v = ϕ(X1 . . . Xn )v = w, so we see that we
must have U (g)v = V .
Now, if β1 , . . . , βk is an ordering of ∆+ , H1 , . . . , Hl be a basis for H, and note
that the set {E−β1 , . . . , E−βk , H1 , . . . , Hk , Eβ1 , . . . , Eβk } forms a basis for g. By the
Poincaré-Birkhoff-Witt theorem, the monomials of the form
p1 pk
E−β 1
. . . E−β k
H1m1 . . . Hlml Eβq11 . . . Eβqkk
form a basis for U (g).
Let v ∈ Vλ , and consider
 
p1 pk m1 ml q1 qk
ϕ E−β 1
. . . E −βk 1 H . . . H l Eβ1 . . . Eβk v
   
p1
= ϕ E−β 1
. . . ϕ Eβpkk v

We know from the above that Eβk v = 0, and that E−βk v ∈ Vλ−β , and H` v = λv
from Proposition X. We conclude that if the mutliplication by the representation
of the monomial gives a result in Vλ then q1 , . . . qk , p1 , . . . , pk = 0, and hence we
obtain a multiple of v. Since we know that U (g)v = V , we conclude that Vλ consists
precisely of the span of v, proving (b).
Furthermore, we know that if the monomial acts nontrivially on v, then the
result lands in the weight space with weight
k
X
λ− ∂i βi .
i=1

Again, since U (g)v = V , we obtain that these are all the weight spaces there are;
Pk
in other words, every weight is of the form λ − i=1 ∂i βi , and d is proved. Since λ
is the only weight in ∆ that can have this property, and this is independent of the
ordering chosen, a is proven.
We can now prove that the second half of c: that if Eα v = 0 for all α ∈ ∆+ , then
v ∈ Vλ . Assume not, so there exists some such v ∈ / Vλ . Without loss of generality,
we may assume that v has no component in Vλ . Let λ0 be the largest weight such
that v has a nonzero component in Vλ0 and let v 0 be this component. Then since
Eα v = 0 for all α, and in particular the component of Eα v in Vα+λ0 is 0, we must
have Eα v 0 = 0 for all α (as Eα (v − v 0 ) will always have 0 component in Vα+λ0 ). We
also have that ϕ(h)v 0 ⊆ spanv 0 . Then applying a general monomial to v and using
the result that U (g)v = V , we conclude that
X p  
pk

1
V = E−β 1
. . . ϕ Eβk Cv = V
   
p1
. . . ϕ Eβpkk Cv ∈ Vλ0 − i βi pi , so we would
P
but this is impossible since E−β 1

obtain that V is contained within only weight spaces with weight less than or equal
to λ0 , a contradiction. Hence if Eα v = 0 for all α ∈ ∆+ , then v ∈ Vλ , and this
property characterizes Vλ .
Now, we may prove that λ is dominant, i.e. that hλ, αi ≥ 0 for all α ∈ ∆+ . For
α ∈ ∆+ , form Hα , Eα , and E−α , normalized to obey equations X. As we have seen,
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 19

these vectors span an isomorphic copy of sl(2, C) that we label slα . For v 6= 0 in
Vλ , the subspace of V spanned by all monomials

ϕ(E−α )p ϕ(Hα )q ϕ(Eα )r v


is stable under slα , and the fact that Eα annihilates the members of Vλ shows that
this is equal to the span of all ϕ(E−α )p . On these vectors ϕ(Hα0 ) has eigenvalue
2hλ, αi
(λ − pα)(Hα0 ) = − 2p
|α|2
and thus the largest eigenvalue is thus 2hλ,αi
|α|2 . Since we know from Theorem 1.15
that for any representation of sl(2, C) on a finite-dimensional vector space, π(h)
has positive eigenvalues, this shows that hλ, αi ≥ 0, so λ is therefore dominant.
Now, we can prove that this correspondence is one-to-one. Let ϕ and ϕ0 be
irreducible finite-dimensional representations on V and V 0 with the same highest
weight, λ. We know we may regard ϕ and ϕ0 as module representations of U (g).
Let v0 ∈ Vλ and v00 ∈ Vλ0 be nonzero, and consider the representation ϕ ⊕ ϕ0 on
V ⊕ V 0 . Consider the subspace
M
S = ϕ ⊕ ϕ0 (U (g))(v0 v00 )
of V ⊕V 0 . Certainly S is invariant under ϕ⊕ϕ0 (U (g)); we claim that it is irreducible.
Let T ⊆ S be an irreducible invariant subspace and let v ⊕ v 0 be a nonzero
highest weight vector of T . Then for α ∈ ∆+ , we have that
0 = (ϕ ⊕ ϕ0 )Eα v ⊕ v 0 = ϕ(Eα )v ⊕ ϕ0 (Eα0 )v 0
so ϕ(Eα )v = ϕ0 (Eα0 )v 0 = 0. Since the c shows that the only vectors that are
annihilated by all weight vectors are the vectors in Vλ , we must have that v = cv0
and v 0 = cv00 . By assumption cv0 ⊕ c0 v00 ∈ ϕ ⊕ ϕ0 (U (g))v0 ⊕ v00 , since v0 and v00 are
weight vectors.
Now, note that when we apply one of the monomials in equation X, we obtain
that the Eβ annihilate v ⊕ v 0 and the E−β lower the weights, and the H ∈ h act by
M
ϕ ϕ0 (H)(v0 ⊕ v00 ) = ϕ(H)v0 ⊕ ϕ0 (H)v00 = λ(H)v0 ⊕ v00 .
Since these monomials act as a basis, we conclude that c = c0 and thus
S = (ϕ ⊕ ϕ0 )Eα v0 ⊕ v00 ⊆ T
so T = S, and S is irreducible.
Now, note that the projection π : S → V commutes with the representations
and is not identically 0, and we may apply the following Lemma:
Lemma 6.6. (Schurr’s Lemma) Let ϕ and ϕ0 be irreducible representations of a
Lie algebra g on finite-dimensional vector spaces V and V 0 . If L : V → V 0 is
a nontrivial linear map such that ϕ0 (X)L = Lϕ(X) for all X ∈ g, then L is a
bijection.
Proof. Note that ker L is invariant under ϕ, so if ker L 6= L, then ker L = 0;
similarly, im L is invariant under ϕ0 , so since im L 6= 0, im L = L. 
This gives that π is a bijection between S and V . Similarly, π 0 : S → V 0 is a
bijection from S to V 0 . Hence ϕ and ϕ0 are equivalent. 
20 ABIGAIL WARD

7. Applying these results: the Lorentz group and its Lie algebra.
In this section we apply the results proven above to sketch out how the study
of the Lie algebra of the Lorentz group gives information in theoretical physics.
For the sake of brevity, we assume the reader has some familiarity with theoretical
physics and in particular the Lorentz group and Einstein summation convention;
we also leave out many computational details that are easy to check but lengthy
to write out. For much more detail on the applications of Lie algebras to particle
physics, the reader is referred to [2].
Briefly, the Lorentz group is the group of symmetries that acts on events in
Minkowski space that leaves the relativistic interval between two events invariant.
The the characterizing condition states that for two events x and y, for all Λ in the
Lorentz group
gαν (Λx)α (Λy)ν = gαβ xα y β .
(where gαν is the Minkowski metric). If we use the matrix representation where
 
1 0 0 0
0 −1 0 0
G= ,
0 0 −1 0
0 0 0 1

and we consider only transformations that preserve orientation and the direction
of time, we obtain the group SO(1, 3), which can be represented as matrices X
satisfying the equation
X T G + GX = 0.
This is a quadratic equation in the entries of X that defines the manifold SO(1, 3).
In identifying the Lie algebra with the tangent space so(1, 3) at the identity by
differentiating, we obtain that for X ∈ so(1, 3), we must have X + X T = 0, so
so(1, 3) is the space of skew symmetric matrices (which is clearly a Lie algebra, as
the commutator of two skew symmetric matrices is skew symmetric). Note that
so(1, 3) has dimension 6, since any element is characterized by its entries above the
diagonal.
Our study of this Lie algebra takes us through the study of a related Lie algebra,
su(2) (denoted as such since it is the Lie algebra of the group SU (2)). Consider
the real Lie algebra of dimension three with the following basis:
     
0 −i 0 −1 −i 0
τ1 = 1/2 , τ2 = 1/2 , τ3 = 1/3 .
−i 0 1 0 0 i

These matrices satisfy vj vk + vs vr = δjk 1.


Now, consider the adjoint representation of su(2) on itself. For example, com-
putations will show that

ad(τ1 )τ1 = 0τ1 + 0τ2 + 0τ3


ad(τ1 )τ2 = 0τ1 + 0τ2 + 1τ3
ad(τ1 )τ3 = 0τ1 − 1τ2 + 0τ3
SEMI-SIMPLE LIE ALGEBRAS AND THEIR REPRESENTATIONS 21

and so on; we obtain that with respect to the basis {τ1 , τ2 , τ3 }, the adjoint repre-
sentation has matrices
     
0 0 0 0 0 0 0 −1 0
ad(τ1 ) = 0 0 −1 , ad(τ2 ) = 0 0 −1 , ad(τ2 ) = 1 0 0 .
0 1 0 0 1 0 0 0 0
Consider now the complexification of the aforementioned Lie algebras–i.e., the
set of matrices when considered as a complex vector space. Ticciati identifies the
complexification of so(1, 3) with C ⊗ so(1, 3) and the complexification of su(2) with
C ⊗ su(2); we follow Knapp in denoting these instead as so(1, 3)C and su(2)C . This
is useful because it allows us to apply our above results with roots and weights,
which all depended on having a complex Lie algebra. Note that so(1, 3)C is the Lie
algebra of anti-Hermitian matrices, and su(2)C is exactly sl(2, C).
Consider the matrices
     
0 0 0 0 0 0 0 0 0 0 0 0
 , X3 = 0 0 −1 0 ,
0 0 0 0  0 0 0 1  
X1 =  0 0 0 −1
, X 2 = 
0 0 0 0 0 1 0 0
0 0 1 0 0 −1 0 0 0 0 0 0
     
0 1 0 0 0 0 1 0 0 0 0 1
1 0 0 0 0 0 0 0 0 0 0 0
B1 =  0 0 0 0 ,
 B2 = 1 0 0 0 , B3 = 0 0 0 0 .
   

0 0 0 0 0 0 0 0 1 0 0 0
and the basis of so(1, 3)C given by Ti = 1/2(Xi + iB1 ), Ti = 1/2(Xi − iB1 ).
Computations will show that the maps Ti 7→ τi , Ti 7→ τi are isomorphisms, so we
conclude that so(1, 3)C ∼= su(2)C ⊕ su(2)C . Thus studying the representations of
su(2)C will give us information on the representations of su(2)C .
To study the representations of su(2)C , we first choose a Cartan subalgebra.
Since su(2)C is semi-simple and we cannot chose two linearly independent elements
of su(2)C that commute with each-other, the subspace generated by any element
is a maximal Abelian subalgebra and hence a Cartan subalgebra. For notational
convenience’s sake, denote ρj = iτj for j = 1, 2, 3; then, following the usual conven-
tion intheoretical
 physics (where our z axis is always the axis of choice), we chose
1 0
ρ3 = as our basis for this algebra. We wish to find then operators X
0 −1
such that
[ρ3 , X] = λX
where such X will be in our root spaces. Computation will show that for operators
of the form R = iρ1 + iρ2 , λ = 1, and for operators of the form L = ρ1 − iρ2 ,
λ = −1. Dimension considerations and Proposition 3.2 then show that we may
write X = Cρ3 ⊕ CL ⊕ Ŗ; this is one root space decomposition of su(2)C .
We may now apply the Theorem of Highest Weight to note that for any dominant
linear functional λ on Cρ3 , it makes sense to talk about the irreducible represen-
tation Dλ where λ is the highest weight. The linear functionals of Cρ3 are just
determined by the value they send ρ3 to; the condition that these are dominant
amounts to the condition that this value must be greater than zero, and the fact
that they are algebraically integral just amounts to the fact that this value must
be n/2 for some integer n ≥ 0. Furthermore, given that our roots here are just α
and −α, we see that the weights of this representation are of the form λ − n for
22 ABIGAIL WARD

0 ≤ n ≤ 2λ, and the operator L acts as a lowering operator. The reader familiar
with quantum mechanics will recognize this as precisely the conditions describing
the spin of a particular particle.
We can therefore see how the choice of a particular representation of this Lie
algebra can be useful in describing the actual physical system that the Lie algebra
acts upon; furthermore, we may chose the highest weight as appropriate to the
physical system at hand. The above shows briefly how the representations of su(2)C
can be useful in describing physical systems; in more complicated systems described
by SO(1, 3), similar methods apply.

Acknowledgments
It is a pleasure to thank my mentor, Jonathan Gleason, for his guidance and
prompt and thorough answers to all of my questions. I would also like to thank
Wouter van Limbeek for his idea for this project and suggestions of reading material.

References
[1] Anthony W. Knapp Lie Groups: Beyond an Introduction Birkhauser. 1996.
[2] R. Ticciati Quantum Field Theory for Mathematicians. Cambridge University Press. 1999.

You might also like