Representation of Linear Groups - Berndt
Representation of Linear Groups - Berndt
Representations
of Linear Groups
Rolf Berndt
Representations
of Linear Groups
An Introduction Based
on Examples from Physics
and Number Theory
Bibliografische information published by Die Deutsche Nationalbibliothek
Die Deutsche Nationalbibliothek lists this publication in the Deutschen Nationalbibliografie;
detailed bibliographic data is available in the Internet at <http://dnb.d-nb.de>.
ISBN 978-3-8348-0319-1
Preface
There are already many good books on representation theory for all kinds of groups.
Two of the best (in this author’s opinion) are the one by A.W. Knapp: “Representation
Theory for Semisimple Groups. An Overview based on Examples” [Kn1] and by G.W.
Mackey: “Induced Representations in Physics, Probability and Number Theory” [Ma1].
The title of this text is a mixture of both these titles, and our text is meant as a very
elementary introduction to these and, moreover, to the whole topic of group represen-
tations, even infinite-dimensional ones. As is evident from the work of Bargmann [Ba],
Weyl [Wey] and Wigner [Wi], group representations are fundamental for the theory of
atomic spectra and elementary physics. But representation theory has proven to be an
inavoidable ingredient in other fields as well, particularly in number theory, as in the
theory of theta functions, automorphic forms, Galois representations and, finally, the
Langlands program. Hence, we present an approach as elementary as possible, having in
particular these applications in mind.
This book is written as a summary of several courses given in Hamburg for students of
Mathematics and Physics from the fifth semester on. Thus, some knowledge of linear
and multilinear algebra, calculus and analysis in several variables is taken for granted.
Assuming these prerequisites, several groups of particular interest for the applications in
physics and number theory are presented and discussed, including the symmetric group
Sn as the leading example for a finite group, the groups SO(2), SO(3), SU(2), and SU(3)
as examples of compact groups, the Heisenberg groups and SL(2, R), SL(2, C), resp. the
Lorentz group SO(3, 1) as examples for noncompact groups, and the Euclidean groups
E(n) = SO(n) Rn and the Poincaré group P = SO(3, 1)+ R4 as examples for semidi-
rect products.
This text would not have been possible without the assistance of my students and colleagues; it
is a pleasure for me to thank them all. In particular, D. Bahns, S. Böcherer, O. v. Grudzinski,
M. Hohmann, H. Knorr, J. Michaliček, H. Müller, B. Richter, R. Schmidt, and Chr. Schweigert
helped in many ways, from giving valuable hints to indicating several mistakes. Part of the
material was treated in a joint seminar with Peter Slodowy. I hope that a little bit of his way
of thinking is still felt in this text and that it is apt to participate in keeping alive his memory.
Finally, I am grateful to U. Schmickler-Hirzebruch and S. Jahnel from the Vieweg Verlag for
encouragement and good advice.
Contents
Introduction ix
3 Continuous Representations 31
3.1 Topological and Linear Groups . . . . . . . . . . . . . . . . . . . . . . . . 31
3.2 The Continuity Condition . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
3.3 Invariant Measures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
3.4 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Bibliography 261
Index 266
Introduction
In this book, the groups enumerated in the Preface are introduced and treated as matrix
groups to avoid as much as possible the machinery of manifolds, Lie groups and bundles
(though some of it soon creeps in through the backdoor as the theory is further devel-
oped). Parallel to information about the structure of our groups we shall introduce and
develop elements of the representation theory necessary to classify the unitary represen-
tations and to construct concrete models for these representations. As the main tool
for the classification we use the infinitesimal method linearizing the representations of a
group by studying those of the Lie algebra of the group. And as the main tools for the
construction of models for the representations we use
– tensor products of the natural representation,
– representations given by smooth functions (in particular polynomials) living on a
space provided with an action of the group,
and
– the machinery of induced representations.
Moreover, because of the growing importance in physics and the success in deriving
branching relations, the procedure of geometric quantization and the orbit method, de-
veloped and propagated by Kirillov, Kostant, Duflo and many others shall be explained
via its application to some of the examples above.
Besides the sources already mentioned, the author was largely influenced by the now
classical book of Kirillov: “Elements of the Theory of Representations” [Ki] and the
more recent “Introduction to the Orbit Method” [Ki1]. Other sources were the books
by Barut and Raczka: “Theory of Group Representations and Applications” [BR], S.
Lang: “SL(2, R)” [La], and, certainly, Serre: “Linear Representations of Finite Groups”
[Se]. There is also the book by Hein: “Einführung in die Struktur- und Darstellungs-
theorie der klassischen Gruppen” [Hei], which follows the same principle as our text,
namely to do as much as possible for matrix groups, but does not go into the infinite-
dimensional representations necessary for important applications. Whoever is further
interested in the history of the introduction of representation theory into the theory of
automorphic forms and its development is referred to the classical book by Gelfand, Graev
and Pyatetskii-Shapiro: “Representation Theory and Automorphic Forms” [GGP], Gel-
bart’s: “Automorphic Forms on Adèle Groups” [Ge], and Bump’s: “Automorphic Forms
and Representations” [Bu]. More references will be given at the appropriate places in our
text; as already said, we shall start using material only from linear algebra and analysis.
But as we proceed more and more elements from topology, functional analysis, complex
function theory, differential and symplectic geometry will be needed. We will try to in-
troduce these as gently as possible but often will have to be very rudimentary and will
have to cite the hard facts without the proofs, which the reader can find in the more
refined sources.
To sum up, this text is prima facie about real and complex matrices and the nice and
sometimes advanced things one can do with them by elementary means starting from
a certain point of view: Representation theory associates to each matrix from a given
group G another matrix or, in the infinite-dimensional case, an operator acting on a
Hilbert space. One may want to ask, why study these representations by generally more
complicated matrices or operators if the group is already given by possibly rather simple
matrices? An answer to this question is a bit like the one for the pudding: the proof
is in the eating. And we hope our text will give an answer. To the more impatient
reader who wants an answer right away in order to decide whether to read on or not, we
offer the following rough explanation. Certain groups G appear in nature as symmetry
groups leaving invariant a physical or dynamical system. For example, the orthogonal
group O(3) is the symmetry group for the description of the motion of a particle in a
central symmetric force field, and the Poincaré group P is the symmetry group for the
motion of a free particle in Minkowski space. Then the irreducible unitary representa-
tions of G classify indivisible intrinsic descriptions of the system and, boldly spoken, can
be viewed as “elementary particles” for the given situation. Following Wigner and his
contemporaries, the parameters classifying the representations are interpreted as quan-
tum numbers of these elementary particles. . .
The importance of representations for number theory is even more difficult to put into a
nutshell. In the Galois theory of algebraic number fields (of finite degree) Galois groups
appear as symmetry groups G. Important invariants of the fields are introduced via cer-
tain zeta- or L-functions, which are constructed using finite-dimensional representations
of these Galois groups. Another aspect comes from considering smooth (holomorphic
resp. meromorphic) functions in several variables which are periodic or have a more gen-
eral covariant transformation property under the action of a given discrete subgroup of
a continuous group G, like for instance G = SL(2, R) or G a Heisenberg or a symplectic
group. Then these functions with preassigned types, e.g., theta functions or modular
forms, generate representation spaces for (infinite-dimensional) representations of the re-
spective group G.
Finally, we will give an overview over the contents of our text: In a prologue we will fix
some notation concerning the groups and their actions that we later use as our first ex-
amples, namely, the general and special linear groups over the real and complex numbers
and the orthogonal and unitary groups. Moreover, we present the symmetric group Sn
of permutations of n elements and some facts about its structure. We stay on the level of
very elementary algebra and stick to the principle to introduce more general notions and
details from group theory only when needed in our development of the representation
theory. We follow this principle in the first chapter where we introduce the concept of
linear representations using only tools from linear algebra. We define and discuss the
fundamental notions of equivalence, irreducibility, unitarity, direct sums, tensor product,
characters, and give some first examples.
The theory developed thus far is applied in the second chapter to the representations of
finite groups, closely following Serre’s exposition [Se]. We find out that all irreducible
representations may be unitarized and are contained in the regular representation.
In the next step we move on to compact groups. To do this we have to leave the purely
algebraic ground and take in topological considerations. Hence, in the third chapter, we
define the notion of a topological and of a (real or complex) linear group, the central
notion for our text. Following this, we refine the definition of a group representation by
adding the usual continuity condition. Then we adequately modify the general concepts
of the first chapter. We try to take over as much as possible from finite to compact
groups. This requires the introduction of invariant measures on spaces with a (from now
on) continuous group action, and a concept of integration with respect to these measures.
In the forth chapter we concentrate on compact groups and prove that the irreducible
representations are again unitarizable, finite-dimensional, fixed by their characters and
contained in the regular representation. But their number is in general not finite, in
contrast to the situation for finite groups. We state, but do not prove, the Peter-Weyl
Theorem. But to get a (we hope) convincing picture, we illustrate it by reproducing
Wigner’s discussion of the representaions of SU(2) and SO(3). We use and prove that
SU(2) is a double cover of SO(3). Angular momentum, magnetic and spin quantum
numbers make an appearance, but for further application to the theory of atomic spectra
we refer to [Wi] and the physics literature.
In a very short fifth chapter, we assemble some material about the representations of
locally compact abelian groups. We easily get the result that every unitary irreducible
representation is one-dimensional. But as can be seen from the example G = R, their
number need not be denumerable. More functional analysis than we can offer at this
stage is needed to decompose a given reducible representation into a direct integral of
irreducibles, a notion we not consider here.
Chapters 7 and 8 are the core of our book. In the seventh chapter we introduce the
concept of induced representations, which allows for the construction of (sometimes
infinite-dimensional) representations of a given group G starting from a (possibly one-
dimensional) representation of a subgroup H of G. To make this work we need again a
bit more Hilbert space theory and have to introduce quasi-invariant measures on spaces
with group action. We illustrate this by considering the examples of the Heisenberg
group and G = SU(2), where we rediscover the representations which we already know.
Then we use the induction process to construct models for the unitary representations of
SL(2, R) and SL(2, C). In particular, we show how holomorphic induction arises in the
discussion of the discrete series of SL(2, R) (here we touch complex function theory). We
insert a brief discussion of the Lorentz group GL = SO(3, 1)0 and prove that SL(2, C)
is a double cover of GL . To get a framework for the discussion of the representations of
the Poincaré group GP , which is a semidirect product of the Lorentz group with R4 , we
define semidirect products and treat Mackey’s theory in a rudimentary form. We outline
a recipe to classify and construct irreducible representations of semidirect products if
one factor is abelian. We do not prove the general validity of this procedure as Mackey’s
Imprimitivity Theorem is beyond the scope of our book, but we apply it to determine
the unitary irreducible representations of the Euclidean and the Poincaré group, which
are fundamental for the classification of elementary particles.
Under the heading of Geometric Quantization, in the eighth chapter we take an alterna-
tive approach to some material from chapter 7 by constructing representations via the
orbit method. Here we have to recall (or introduce) more concepts from higher analysis:
manifolds and bundles, vector fields, differential forms, and in particular the notion of
a symplectic form. We can again use the information and knowledge we already have
of our examples G = SL(2, R), SU(2) and the Heisenberg group to get a feeling what
should be done here. We identify certain spheres and hyperboloids as coadjoint orbits of
the respective groups, and we construct line bundles on these orbits and representation
spaces consisting of polarized sections of the bundles.
Finally, in the nineth and last chapter, we give a brief outlook on some examples where
representations show up in number theory. We present the notion of an automorphic
representation (in a rudimentary form) and explain its relation with theta functions and
automorphic forms. We have a glimpse upon Hecke’s and Artin’s L-functions and men-
tion the Artin conjecture.
We hope that some of the exercises and/or omitted proofs may give a starting point for
a bachelor thesis, and also that this text motivates further studies in a master program
in theoretical physics, algebra or number theory.
¡ Libro, afán
de estar en todas partes,
en soledad!
J. R. Jiménez
Chapter 0
This text is mainly on groups which some way or another come from physics and/or
number theory and which can be described in form of a real or complex matrix group.
O(n) := {A ∈ Mn (R); t AA = En },
resp. for n = p + q
where Dp,q is the diagonal matrix having p times 1 and q times −1 in its diagonal, and
the unitary group
U(n) := {A ∈ Mn (C); t AĀ = En },
2 0. Prologue: Some Groups and their Actions
resp. for n = p + q
U(p, q) := {A ∈ Mn (R); t ADp,q Ā = Dp,q }.
Again, addition of the letter S to the symbol of the group indicates that we take only
matrices with determinant 1, e.g.
SO(n) := {A ∈ O(n); det A = 1 }.
These groups together with some other families of groups, in particular the symplectic
groups showing up later, are known as classical groups.
Later on, we will often use subgroups consisting of certain types of block matrices, e.g.
the group of diagonal matrices
An := {D(a1 , . . . , an ); a1 , . . . , an ∈ K∗ },
where D(a1 , . . . , an ) denotes the diagonal matrix with the elements a1 , . . . , an in the di-
agonal, or the group of upper triangular matrices (or standard Borel group) Bn consisting
of matrices with zeros below the diagonal and the standard unipotent group Nn , the sub-
group of Bn where all diagonal elements are 1.
In view of the importance for applications, moreover, we distinguish several types of
Heisenberg groups: Thus the group N3 we just defined, is mostly written as
⎛ ⎞
1 x z
Heis (K) := { g = ⎝ 1 y ⎠ ; x, y, z ∈ K }.
1
In the later application to theta functions it will become clear that, though it may seem
more complicated, we shall better use the following description for the Heisenberg group.
⎛ ⎞
1 0 0 µ
⎜ λ 1 µ κ ⎟
Heis(K) := { g = (λ, µ, κ) := ⎜ ⎟
⎝ 0 0 1 −λ ⎠ ; µ, λ, κ ∈ K },
0 0 0 1
and the “higher dimensional” groups, which for typographical reasons we here do not
write as matrix groups
Heis(Kn ) := { g = (x, y, z); x, y ∈ Kn , z ∈ K }
with the multiplication law given by
gg = (x + x , y + y , z + z + t xy − t yx ).
We suggest to write elements of Kn as columns.
Exercise 0.1: Write Heis(Kn ) in matrix form. Show that Heis(Kn ) for n = 1 and the
other two Heisenberg groups above are isomorphic.
G × X −→ X , (g, x) −→ g · x
g · (g · x) = (gg ) · x, e · x = x
Remark 0.1: If Aut X denotes the group of all bijections of X onto itself, the definition
says that we have a group homomorphism G −→ Aut X associating to every g ∈ G the
transformation x −→ g · x.
The group action is called effective iff no element except the neutral element e acts as
the identity, i.e. the homomorphism G −→ Aut X is faithful.
The group action is called transitive iff for every pair x, x ∈ X there is a g ∈ G with
x = g · x.
G · x0 := {g · x0 ; g ∈ G }
Gx0 := {g ∈ G; g · x0 = x0 }
Example 0.1: G = GL(n, K) and its subgroups act on X = Kn from the left by matrix
multiplication
(A, x) −→ Ax
Example 0.2: A group acts on itself from the left in three ways:
a) by left translation (g, g0 ) −→ g · g0 = gg0 =: λg g0 ,
b) by (the inverse of) right translation (g, g0 ) −→ g · g0 = g0 g −1 =: ρg−1 g0 ,
c) by conjugation (g, g0 ) −→ g · g0 = gg0 g −1 =: κg g0 .
Left- and right translations are obviously transitive actions. For a given group, the de-
termination of its conjugacy classes, i.e. the classification of the orbits under conjugation
is a highly interesting question, as we shall see later.
Exercise 0.4: Determine a family of matrices containing exactly one representative for
each conjugacy class of G1 = GL(2, C) and G2 = GL(2, R).
Definition 0.2: A set is called a homogeneous space iff there is a group G acting tran-
sitively on X .
Remark 0.2: In this case one has X = G · x for each x ∈ X and any two stabilizing
groups Gx and Gx are conjugate. (Prove this as Exercise 0.5.)
Definition 0.3: For any G-set we shall denote by X /G or XG the set of G-orbits in
X and by X G the set of fixed points for G, i.e. those points x ∈ X for which one has
g · x = x for all g ∈ G.
If the set X has some structure, for instance, if X is a vector space, we will (tacidly)
additionally require that a group action preserves this structure, i.e. in this case that
x −→ g · x is a linear map for each g ∈ G (or later on, is continuous if X is a topological
space). It is a fundamental question whether the orbit space X /G inherits the same
properties as the space X may have. For instance, if X is a manifold, is this true also
for X /G? We will come back to this question later several times. Here let us only look
at the following situation: Let X be a homogeneous G-space, x0 ∈ X , and H = Gx0 the
isotropy group. Then we denote by G/H the set of left cosets gH := {gh; h ∈ H}. These
are also to be seen as H-orbits H · g where H acts on G by right translation. G/H is a
G-set, G acting by (g, g0 H) −→ gg0 H. Then we shall identify G/H and X via the map
G/H −→ X , gH −→ g · x0 .
This map is an example for the following general notion.
Definition 0.4: Let X and X be G-sets and f : X −→ X be a map. The map f is
called G-equivariant or a G-morphism iff one has for every g ∈ G and x ∈ X
g · f (x) = f (g · x).
Now we come back to the question raised above.
Exercise 0.6: Prove that G/H has the structure of a group iff H is not only a subgroup
but a normal subgroup, i.e. one has ghg −1 ∈ H for all g ∈ G, h ∈ H.
We mention here another useful fact: If H is a subgroup of G, one defines the normalizer
of H in G as
NG (H) := {g ∈ G; gHg −1 = H}.
It is clear that NG (H) is the maximal subgroup in G that has H as a normal sub-
group. Then the group Aut (G/H) of G-equivariant bijections of G/H is isomorphic to
NG (H)/H. (Prove this as Exercise 0.7.)
0.3 The Symmetric Group 5
As well as left actions one defines and studies right actions of a group G on a set or a
space X :
G × X −→ X , (g, x) −→ x · g,
(x · g) · g = x · (gg ), x·e=x
Remark 0.3: Obviously it is easy to switch from right action to left action and vice
versa using the antiautomorphism g −→ g −1 of G. Namely, on any right G-space X one
can naturally define a left action by g · x := x · g −1 .
Right actions often show up in number theory and they are written there as xg := x · g
(for instance in the action of the Galois group on a number field).
Definition 0.6: The symmetric group Sn is the group of permutations, i.e. bijections,
of a set of n elements.
Remark 0.6: The map σ −→ (σ) is a group homomorphism of Sn to the subgroup
{1, −1} of the multiplicative group R∗ = R \ {0}. Thus, the number of transpositions in
a representation of σ as a product of transpositions is either even or odd.
The kernel of is usually written as An . It is a normal subgroup and called the alternating
group.
Permutations are often written as products of cycles, i.e. permutations (i1 , . . . , ir ) with
ij −→ ij+1 for j < r and ir −→ i1 if r > 1 and the identity for r = 1. More precisely,
one has
Remark 0.7: Each permutation is (up to order) a unique product of disjoint cycles.
Theorem 0.1: The number of conjugacy classes of Sn is equal to the number p(n) of
partitions of n.
– {id}
Exercise 0.9: Write down the matrices A(σ) and verify that the map given by associat-
ing σ −→ A(σ) is in fact an isomorphism from S3 to the subgroup of GL(3, R) consisting
of the permutation matrices.
Chapter 1
We first collect the basic definitions and concepts concerning representations of groups,
using only algebraic methods. As is common, we introduce the concept of a linear repre-
sentation, which is completely adequate and sufficient for the study of representations of
finite groups. In a later chapter we will enrich the discussion with elements from topology
to refine the notion of a representation to that of a continuous representation.
We start by stating the essential definitions and only afterwards illustrate these by some
examples and constructions.
with
π(gg ) = π(g)π(g ) for all g, g ∈ G.
Aut V is equivalently denoted by GL(V ) and stands for the group of all (linear) isomor-
phisms of V .
In the case of a finite-dimensional vector space V , say of dimension n, one says that π is
of degree n or π is a n-dimensional representation.
8 1. Basic Algebraic Concepts
As every group homorphism transforms the neutral elements of the groups into one
another, it is clear that this prescription associates the unit matrix En to the neutral
element e of the group G, and that we have π(e) = idV in the situation of the general
definition above.
If G is a matrix group, G ⊂ GL(n, C), as in Section 0.1, obviously, we have the natural
representation π0 given by
π0 (A) = A
for each A ∈ G. Another even more ubiquous representation is the trivial representation
π = id, associating the identity idV to each g ∈ G.
Using the notion of group action introduced in Section 0.2, we can also say that a lin-
ear representation of G in V is the same thing as a left action of G on the vector space V .
Next, we restrict our objects still more, in particular in view to the applications in physics.
We assume now that V is a unitary complex vector space, i.e. V is equipped with a scalar
product
< ., . >: V × V −→ C, (v, v ) −→ < v, v >
which is
– linear in the second variable and antilinear in the first,
– hermitian: for every v, v ∈ V one has < v, v >= < v , v >, and
– positive definite: for every v ∈ V one has < v, v > ≥ 0 and = 0 exactly for v = 0.
1.2 Equivalent Representations 9
n
< x, y >:= x̄i yi for all x, y ∈ Cn .
i=1
F π(g) = π (g)F,
V
F / V
π(g) π (g)
V
F / V .
F : V −→ V
Remark 1.1: The space of intertwining operators between π and π is again a vec-
tor space. It is denoted by HomG (V, V ) or C(V, V ). Moreover, we use the notation
C(V ) := C(V, V ) and c(π, π ) = c(V, V ) = dim C(V, V ). c(π, π ) is also called the multi-
plicity of π in π and denoted by mult(π, π ).
Determine the unitary (linear) dual Ĝ of a given group G, i.e. the set of equivalence
classes of unitary irreducible representations of G.
10 1. Basic Algebraic Concepts
Example 1.1: In Section 1.1 we already mentioned that every matrix group G has its
natural representation π0 , i.e. for each real or complex matrix group G ⊂ GL(n, K) one
has the representation in V = Cn associating to every A ∈ G the matrix A itself. These
natural representations are obviously unitary for G = SO(n) or SU(n) but in general
neither unitary nor irreducible, as the following example shows:
Example 1.2: Let g be an element of the group G = SO(2). Then we usually write
α β
g = r(ϑ) := ( ), α = cos ϑ, β = sin ϑ, ϑ ∈ R.
−β α
α − iβ 0 e−iϑ 0
U r(ϑ)U −1 = ( )=( ).
0 α + iβ 0 eiϑ
This can be interpreted as follows. The matrix U is the matrix for an intertwining
operator F intertwining the natural representation π0 with the representation π1 on
V = C2 given by
e−iϑ 0
π1 (r(ϑ)) = ( ).
0 eiϑ
This representation is reducible and we see that the natural representation is equivalent
to a reducible one. Or we realize that the vectors
√ 1
√ i
e1 := U e1 = (1/ 2) , e2 := U e2 = (1/ 2)
i 1
we see that each element g ∈ S3 can be written as a product of (in general) several
factors σ and τ . We indicate this here (and later on in a similar fashion in other cases)
by the notation
S3 =< σ, τ > .
We easily find one-dimensional representations in V = C, namely the trivial representa-
tion
π1 (g) = 1 for all g ∈ S3
and the signum representation
with the canonical basis vectors e1 = t (1, 0, 0), . . . , e3 = t (0, 0, 1) and z1 , z2 , z3 ∈ C. Then
π0 is given by
< e1 + e2 + e3 , w >= zi ,
V3 := {w, zi = 0} is the orthogonal complement to V1 and as i zi = i zg−1 (i) holds
for all g ∈ S3 , V3 is an invariant subspace.
Remark 1.2: By the general methods to be developed later, it will be clear that all ir-
reducible representations of S3 are equivalent to π1 , π2 , or π3 , i.e. the equivalence classes
of πi (i=1,2,3) constitute the unitary dual of S3 .
12 1. Basic Algebraic Concepts
Exercise 1.3: Prove this in a direct elementary way (see [FH] §1.3).
We take G ⊂ GL(2, C) and as vector space V the space Vm = C [u, v]m of homogeneous
polynomials P of degree m in two variables u, v. By counting, we see dim V = m + 1.
There are two essentially equivalent ways to define a representation of G on Vm :
a b a b
For P ∈ Vm and g = ( ) ∈ G resp. g −1 = ( ) we put
c d c d
or
(ρm (g)P )(u, v) := P (au + cv, bu + dv).
Obviously λ(g)P and ρ(g)P are both again homogeneous polynomials of degree m. The
functional equations λ(gg ) = λ(g)λ(g ) and ρ(gg ) = ρ(g)ρ(g ) can be verified directly.
It becomes also clear from the general discussion in the following Example 1.5.
We remark that λm and ρm are irreducible unitary representations of G = SU(2), if Vm
is provided with a suitable scalar product. This will come out later in section 4.2.
Example 1.5: Let X be a G-set with a left G-action x −→ g · x and V = F(X ) a vector
space of complex functions f : X −→ C such that for f ∈ V we also have fg ∈ V where
fg is defined by
fg (x) = f (g −1 x).
(λ(g)f )(x) := f (g −1 x)
defines a representation λ of G on V .
−1 −1 −1
λ(gg )f (x) = f ((gg )−1 · x) = f (g g · x) = f (g · g −1 · x)
and
−1
λ(g)λ(g )f (x) = λ(g)fg (x) = fg (g −1 · x) = f (g · g −1 · x).
Analogously, one treats a set X with a right G-action x −→ x · g and a vector space
V = F(X ) of complex functions f on X closed under f −→ f g where f g is defined by
f g = f (x · g).
1.3 First Examples 13
(ρ(g)f )(x) := f (x · g)
defines a representation ρ of G on V .
and
ρ(g)ρ(g )f (x) = ρ(g)f g (x) = f (x · g · g ).
For X = Cn and V again a space of functions in n variables, this time viewed as a row,
and G ⊂ GL(n, C), we have again the action x −→ x · g = xg by matrix multiplication.
For n = 2 this leads naturally to the second version of Example 1.4.
Another (as, later to be seen, crucial) application of this procedure is given by making G
act on itself, i.e. X = G, by left or right translation and V = L2 (G), the space of square
integrable functions with respect to a left resp. right invariant Haar measure (these no-
tions will be explained later). We then get the left (right) regular representation of G.
Example 1.6: Let G = Heis(R) be the Heisenberg group from Section 0.1, in the form
of the group of triples g = (λ, µ, κ), λ, µ, κ ∈ R with the multiplication law
gg = (λ + λ , µ + µ , κ + κ + λµ − λ µ),
Exercise 1.4: Prove that πm is a unitary representation. (The proof of the irreducibility
will become easier later when we have prepared some more tools.)
Example 1.7: Prove (as Exercise 1.5) that the Weil group of R, WR from Exercise
0.2 in Section 0.1, has the irreducible representations π = π0 and π = πm , m ∈ Z \ {0}
given by
0 1 0 1 z 0 z 0
π0 ( ) := det = 1, π0 ( ) := det = z z̄,
−1 0 −1 0 0 z̄ 0 z̄
and for m = 0
m
z 0 z 0 0 1 0 1
πm ( ) := , π ( ) := .
0 z̄ 0 z̄ m m
−1 0 (−1)m 0
14 1. Basic Algebraic Concepts
A(g) 0
(π ⊕ π )(g) = ( ) ∈ GL(n + m, K).
0 A (g)
For V = Kn , V = Km and π(g) = A(g) ∈ GL(n, K), π (g) = A (g) ∈ GL(m, K), the
tensor product is given by the Kronecker product of the matrices A(g) and A (g)
⎛ ⎞
a1,1 A (g) . . . a1,n A (g)
(π ⊗ π )(g) = ⎝ ... ⎠ ∈ GL(nm, K) .
an,1 A (g) . . . an,n A (g)
We will not go into the definition of the tensor product (as to be found in the afore
mentioned sources) but simply recall the following. If V has a basis (ei )i∈I and V a
basis (fj )j∈J , then V ⊗ V has a basis (ei ⊗ fj )(i,j)∈I×J .
In V ⊗ V we have the rules
(v1 + v2 ) ⊗ v = v1 ⊗ v + v2 ⊗ v ,
v⊗ (v1
+ v2 )
= v ⊗ v1 + v ⊗ v2 ,
λ(v ⊗ v ) = (λv) ⊗ v = v ⊗ (λv ) for all v, v1 , v2 ∈ V, v , v1 , v2 ∈ V , λ ∈ K.
It is clear that one can also define tensor products of more than two factors, and in par-
ticular tensor powers like ⊗3 V = V ⊗ V ⊗ V (It is not difficult to see that these products
are associative.)
1.4 Basic Construction Principles 15
Moreover, there are subspaces of tensor powers with a preassigned symmetry property.
In particular we are interested in p-fold symmetric tensors S p V and p-fold antisymmetric
tensors ∧p V .
For instance, if V is three-dimensional with basis (e1 , e2 , e3 ),
(e1 ⊗ e1 , e1 ⊗ e2 , e1 ⊗ e3 , e2 ⊗ e1 , . . . , e3 ⊗ e3 ),
e1 ⊗ e1 , e1 ⊗ e2 + e2 ⊗ e1 , e1 ⊗ e3 , e2 ⊗ e2 , e2 ⊗ e3 + e3 ⊗ e2 , e3 ⊗ e3 .
e1 ∧ e2 := e1 ⊗ e2 − e2 ⊗ e1 , e1 ∧ e3 := e1 ⊗ e3 − e3 ⊗ e1 , e2 ∧ e3 := e2 ⊗ e3 − e3 ⊗ e2 .
More generally, one has symmetric and antisymmetric subspaces in ⊗p V with bases
resp.
For instance, if V is three-dimensional, S p V can be identified with the space K [u, v, w]p
of homogeneous polynomials of degree p in three variables. And if π is a representation
of G on V, the prescription ei −→ π(g)ei defines by linear continuation representations
S p π and ∧p π on S p V resp. ∧p V .
It is one of the most important construction principles for finite-dimensional representa-
tions to take the natural representation π0 and to get (other) irreducible representations
by taking tensor products and reduce these to sums of irreducibles. This will be analysed
later.
V ∗ = Hom(V, K) = {ϕ : V −→ K, ϕ K − linear }.
and, if dim V = n with basis (e1 , . . . , en ), we denote (e∗1 , . . . , e∗n ) for the dual basis for
V ∗ , i.e. with < e∗i , ej > = δi,j (= 1 for i = j and = 0 for i = j).
16 1. Basic Algebraic Concepts
The proof that we really get a representation is already contained in the example 1.5 in
section 1.3.
It is clear from linear algebra that in matrix form we have A∗ (g) = t A(g −1 ).
There is still another essential construction principle, namely the concept of induced
representations. It goes back to Frobenius and is also helpful in our present algebraic
context. But we leave it aside here and postpone it for a more general treatment in
Chapter 7.
1.5 Decompositions
We now have at hand procedures to construct representations and we have the notion of
irreducible representations which we want to look at as the building blocks for general
representations. To be able to do this, we still have to clarify some notions.
We defined in Section 1.1 that (π, V ) is irreducible, iff there is no nontrivial subrepre-
sentation. Now we add:
Theorem 1.1: Let (π, V ) be the representation of a finite group G and (π1 , V1 ) a sub-
representation. Then there exists an invariant complementary subspace V2 .
1.5 Decompositions 17
Proof: Let <, > be any scalar product for V (such a thing exists as is obvious at least
for finite-dimensional V Kn ) . We can find a G-invariant scalar product <, > , namely
the one defined by
Remark 1.5: If G is any group and (π, V ) a unitary representation of G, again every
subrepresentation (π1 , V1 ) has an invariant complement (π2 , V2 ) given as in the proof
above by V2 := V1⊥ = {v ∈ V, < v, v1 > = 0 for all v1 ∈ V1 } because V is already
provided with a π-invariant scalar product.
Remark 1.6: If G is compact, the theorem holds as well, as the sum in the definition
of the invariant scalar product can be replaced by an integral. This will be done later in
Chapter 4.
V = V0 ⊃ V1 ⊃ · · · ⊃ Vn = {0}
such that that the representations πi appearing on Vi /Vi+1 are irreducible. One has the
Jordan - Hölder Theorem: The length of such a chain of subspaces and the equivalence
classes of the representations πi (up to order) are uniquely defined by the equivalence
class of π.
Theorem 1.2: (Schur’s Lemma): If two linear representations (π, V ) and (π , V ) are
irreducible, then every intertwining operator F ∈ C(π, π ) is either zero or invertible. If
V = V , π = π and dim V = n, then F is a homothety, i.e. F = λ id, λ ∈ C.
Proof: i) As defined in Section 1.2 an intertwining operator for π and π is a linear map
F : V −→ V with
Remark 1.9: The first part of the Theorem is also expressed as follows: Two irreducible
representations of the same group are either equivalent or disjoint.
The second part can be written as C(π) = C.
There are several more refined versions of Schur’s Lemma which come up if one sharpens
the notion of a linear representation by adding a continuity requirement as we shall do
later on. For the moment, though it is somewhat redundant, we add another version
which we take over directly from Wigner ([Wi] p. 75) because it contains an irreducibil-
ity criterium used in many physics papers and its proof is a very instructive example for
matrix calculus.
1.5 Decompositions 19
Theorem 1.3 (the finite-dimensional unitary Schur): Let (π, Cn ) be a unitary matrix
representation of a group G, i.e. with π(g) = A(g) ∈ U(n). Let M ∈ GL(n, Cn ) be a
matrix commuting with all A(g), i.e.
Proof: i) First, we show that we may assume M is a Hermitian matrix: We take the
adjoint (with A −→ A∗ := tĀ) of the commutation relation (1.2) and get
A(g)∗ M ∗ = M ∗ A(g)∗ ,
multiply this from both sides by A(g) and, using the unitarity A(g)A(g)∗ = En , get
M ∗ A(g) = A(g)M ∗ .
Hence, if M commutes with all A(g), not only M ∗ does but also the matrices
H1 := M + M ∗ , H2 := i(M − M ∗ ),
which are Hermitian. It is therefore sufficient to show that every Hermitian matrix,
which commutes with all A(g), is a scalar matrix, since if H1 and H2 are multiples of
En , so must be 2M = H1 − iH2 .
ii) If M is Hermitian, one of the fundamental theorems from linear algebra (see for
instance [Ko], p.256 or [Fi] Kapitel 4) tells that M can be diagonalized, i.e. conjugated
into a diagonal matrix D by a matrix U ∈ U(n)
U M U −1 = D.
Then we put Ã(g) := U A(g)U −1 . Ã(g) is unitary as can be easily verified. From the
commutation relation (1.2), we have DÃ(g) = Ã(g)D. With D = (di,j ), di,j = 0 for
i = j, Ã(g) = (ãi,j ), this leads to
Hence for di,i = dj,j we have ãi,j = ãj,i = 0, that is, all Ã(g) must have zeros at all
intersections of rows and columns, where the diagonal elements are different. This would
mean that g −→ Ã(g) resp. g −→ A(g) are reducible representations. Hence, as this
is not the case, we have that all diagonal elements are equal and D is a scalar matrix
λEn , λ ∈ C.
The proof shows that we can sharpen the statement from the Theorem to the irreducibiliy
criterium:
Corollary: If there exists a nonscalar matrix which commutes with all matrices of a
finite-dimensional unitary representation π, then the representation is reducible. If there
exists none, it is irreducible.
This Corollary will help to prove the irreducibility in many cases. Here we show the use-
fulness of Schur’s Lemma by proving another fundamental fact (later we shall generalize
this to statements valid also in the infinite-dimensional cases):
20 1. Basic Algebraic Concepts
F : V −→ V, v −→ π(g0 )v
The second part of Schur’s Lemma says that F is a homothety, i.e. we have a λ ∈ C with
To finish our statements on decompositions in this section, we state here without proofs
(for these see [Ki] p.121ff) some standard facts about the kind of uniqueness one can
expect for a decomposition of a completely reducible representation. By the way, this
theorem is a reflex of a central theorem from Algebra about the structure of modules
over a ring (a representation (π, V ) will be seen also as module over the group ring C[G]).
Reminding the notions of disjoint representations from Section 1.2 and of finite repre-
sentations from above, we add the following notion.
As an example one may think here of a representation of the form π k , k ∈ N, the sum
of k representations (π, V ) given on the k-th sum V k = V ⊕ · · · ⊕ V .
The central fact can be stated like this:
Among other things, the proof uses the notion of projection operators, known from linear
algebra. As we have to use these here later on, we recall the following concept.
1.6 Characters 21
As usual Tr denotes the trace. If π is given by the matrices A(g) = (aij (g)) ∈ GL(n, C),
one has
n
χπ (g) = aii .
i=1
Therefore the trace of a representation is well defined independently of the choice of the
matrix representation. And for conjugate g, g ∈ G we have
χπ (g) = χπ (g ).
This is also expressed by saying the character is a class function.
Obviously one has for the neutral element e of G
χπ (e) = n = dim V.
We assume moreover that π is unitary (we shall see soon that this is not really a restric-
tion if the group is finite or compact). Then π(g) is in matrix form A(g) conjugate to a
diagonal matrix D = D(λ1 , . . . , λn ), where λ1 , . . . , λn are the eigenvalues of A(g). Hence:
n
Remark 1.10: χπ (g) = i=1 λi .
22 1. Basic Algebraic Concepts
If G is finite, each element g has finite order and so has π(g). Thus the eigenvalues λi
must have | λi |= 1 and we have λ−1
i = λi .
Proof: The first equation is simply the expression of the fact that the trace of a block
A
matrix ( ) is the sum of the traces of the two matrices A and A . The second
A
relation can be seen by inspection of the trace of the Kronecker product in 1.4.2.
Representations of Finite
Groups
As we are mainly interested in continuous groups like SO(3) etc., this topic plays no cen-
tral rôle in our text. But here we can learn a lot about how representation theory works.
Thus we present the basic facts. We will not do it by using the abstract notions offered
by modern algebra (as for instance in [FH]), where eventually the whole theory can be
put into the nutshell of one page. Instead we follow very closely the classic presentation
from the first pages of Serre’s book [Se].
KG := {u : G −→ K}.
KG is a K-vector space as we have an addition u + u defined by
(u + u )(g) := u(g) + u (g) for all g ∈ G
and scalar multiplication λu by (λu)(g) := λu(g) for u ∈ KG and λ ∈ K. But we also
have a multiplication uu defined by
Remark 2.1: One can show that KG is an associative K-algebra (for this notion see
Section 6.1).
There is an alternative way to describe KG, namely as a set of formal sums g∈G u(g)g
built from the values u(g) of u in g ∈ G, i.e.
As we restrict our interest in this text to complex representations, in the sequel we write
H = CG and H0 = Ccl G for the algebra of class functions, i.e. with u(g) = u(g ) for
g ∼ g . If (π, V ) is a representation of G, by Remark 1.8 in 1.5 we may assume that it is
unitary. Conjugate elements of G have the same character. Hence the character u = χπ
is not only an element in H but also in H0 .
Our first aim is to prove that, for a complete system π1 , . . . , πh of irreducible represen-
tations, the characters χπ1 , . . . , χπh are an orthonormal system (=: ON-system) in H.
Here the scalar product <, > is given by
If u = χπ is a character, by the Remark 1.11 in 1.6 we have u(t) = u(t−1 ) and hence
(χπ , v) = < χπ , v > for v ∈ H.
< χ, χ > = 0, if π ∼ π ,
< χ, χ > = 1.
F 0 := (1/m) π (g −1 )F π(g).
g∈G
1) F 0 = 0, if π ∼ π (“case 1”),
2) F 0 = λ id, if V = V , π = π , where λ = (1/ dim V )Tr F (“case 2”).
2.1 Characters as Orthonormal Systems 25
Proposition 2.2: For π and π as in Proposition 2.1 given in matrix form by π(g) = A(g)
and π (g) = B(g), we have
(1/m) Aij (t−1 )Akl (t) = (1/n)δil δjk for all i, j, k, l in case 2 .
t∈G
states
and this requires that all the coefficients of all Fjk are zero. Hence, we have the first
relation. In the second case we have
(1/m) Aij (t−1 )Fjk Akl (t) = λδil = (1/n) Fjk δjk δil .
t∈G jk jk
Comparing the coefficients of Fjk on both sides leads to the second relation.
Proof of the Theorem: We have χ (t) = Bii (t) and χ(t) = Aii (t) and hence
< χ, χ > = (1/m) χ(t−1 )χ(t) = (1/m) Aii (t−1 )Ajj (t)
t∈G t∈G ij
< χ, χ >= (1/m) χ (t−1 )χ(t) = (1/m) Bii (t−1 )Ajj (t)
t∈G t∈G ij
V = V1 ⊕ · · · ⊕ Vk .
From 1.2 we know that this number is also recognized as mult(π , π).
Proof: By Theorem 1.6 in 1.6 we have χ = i χi . Theorem 2.1 in 2.1 says that the
summands in
Corollary 3: If two representations π and π have the same character χπ = χπ , they
are equivalent.
Proof: By Corollary 1 the multiplicities are equal for all irreducible components of π
and π and so both representations are equivalent.
V = m1 V1 ⊕ · · · ⊕ mh Vh ,
one has
h
< χ, χ > = m2i .
i=1
Proof: This follows immediately from Theorem 2.1 in 2.1, as χ = mi χi and the χi
are an ON-system.
If g = e, we have gt = t for all t ∈ G, which shows that the diagonal elements of λ(g) are
zero. In particular we have Tr λ(g) = 0 and Tr λ(e) = Tr Em = m. So we have proven:
Proposition 2.3: The character χλ of the regular representation λ of G is given by
χλ (e) = #G = m,
χλ (g) = 0 for all g = e.
By Proposition 2.3 and the general fact that the character of the neutral element equals
the dimension, we finally get
Proof: By Proposition 2.4 we have χλ (t) = Σni χi (t) for all t ∈ G. Taking t = e we
obtain a) and for t = e we obtain b).
This result is very useful for the determination of the irreducible representations of a finite
group: suppose one has constructed some mutually nonequivalent irreducible representa-
tions of dimensions n1 , . . . , nh . In order that they be all irreducible representations (up
to equivalence) it is necessary and sufficient that one has n21 + · · · + n2h = m.
28 2. Representations of Finite Groups
One can prove more facts in this context, for instance the dimensions ni all divide the
order m of G (see [Se] p.53]).
πf := f (t)π(t).
t∈G
By the second part of Schur’s Lemma, πf = λid is a scalar multiple of the identity. We
have Tr λ id = nλ and
Tr πf = Σf (t)Tr π(t) = Σf (t)χ(t).
Hence
λ = (1/n)Σf (t)χ(t) = (m/n) < χ̄, f > .
As the characters are class functions, they are elements of H0 .
2.3 Characters as Orthonormal Bases 29
Proof: Theorem 2.1 from Section 2.1 says that the (χi ) form an ON-system in H0 . We
have to prove that they generate H0 . That is, we have to show that every f ∈ H0 ,
orthogonal to all χ̄i , is zero. Let f be such an element. For each representation π of G,
we put πf = Σf (t)π(t) as above. Proposition 2.6 shows that πf is zero if π is irreducible.
But as each π may be decomposed into irreducible representations, πf is always zero. We
apply this to the regular representation , i.e. to π = λ, and compute for the first basis
vector of the representation space for λ
πf e1 = f (t)λ(t)e1 = f (t)et .
t t
Since πf is zero, we have πf e1 = 0 and this requires f (t) = 0 for all t ∈ G. Hence f is zero.
h
fg = λi χi with λi = < χi , fg > = (c(g)/m)χi (g).
i=1
h
fg (t) = (c(g)/m) χi (g)χi (t).
i=1
Remark 2.2: We recall the statement from 0.3 that the number of conjugacy classes of
the symmetric group Sn equals the number p(n) of partitions of n.
It is a nice excercise to redo the example of G = S3 in the light of the general facts now
at hand: We have p(3) = 3 and thus know of three irreducible representations π1 , π2 , π3
with respective characters χ1 , χ2 , χ3 . Moreover we know that G = S3 can be generated
by the transposition σ = (1, 2) and the cycle τ = (1, 2, 3). The three conjugacy classes
are represented by e = id, σ, and τ . We have
σ 2 = e, τ 3 = e, στ = τ 2 σ.
30 2. Representations of Finite Groups
Hence, for each representation π the equation π(σ)2 = id holds. So there are two one-
dimensional representations, namely π1 = 1 and π2 = sgn with
χ1 (e) = 1, χ1 (σ) = 1, χ1 (τ ) = 1,
χ2 (e) = 1, χ2 (σ) = −1, χ2 (τ ) = 1.
χ1 + χ2 + 2χ3 = χλ .
These formulae should also be read off from the results of Exercise 2.1.
There is a lot more to be said about the representations of finite groups in general
and of the symmetric group in particular. Here, we only mention that the irreducible
representations of Sn are classified by the Young tableaus realizing a permutation of n.
And the following theorem about the canonical decomposition of a representation of a
finite group, which is a special case of our Theorem 1.5 in 1.5 and can be proved rather
easily (see [Se] p. 21):
Let π1 , . . . , πh be as before irreducible representations representing each exactly one
equivalence class and let (π, V ) be any representation. Let V = U1 ⊕ · · · ⊕ Uk be a
direct sum decomposition of V into spaces belonging to irreducible representations. For
i = 1, . . . , h denote by Wi the direct sum of those of the U1 , . . . , Uk which are isomorphic
to Vi (the Wi belong to the primary representations introduced at the end of 1.5). Then
we have
V = W 1 ⊕ · · · ⊕ Wh ,
which is called the canonical decomposition. Its properties are as follows:
Theorem 2.3: i) The decomposition V = W1 ⊕· · ·⊕Wh does not depend on the initially
chosen decomposition of V into irreducible subspaces.
ii) The projection Pi of V onto Wi associated to the decomposition is given by
We stop here our treatment of finite groups and propose to the reader to treat the fol-
lowing examples.
Exercise 2.2: Determine the irreducible representations of the cyclic group Cn and of
the dihedral group Dn (This is the group of rotations and reflections of the plane which
preserve a regular polygon with n vertices, it is generated by a rotation r and a reflection
s fulfilling the relations rn = e, s2 = e, srs = r−1 ).
Continuous Representations
Among the topological groups, compact groups are the most easy ones to handle. Since
we would like to treat their representations next, it is quite natural that the appearance
of topology leads to an additional requirement for an appropriate definition, namely a
continuity requirement. We will describe this first and indicate the necessary changes
for the general concepts from Sections 1.1 to 1.5 (to be used in the whole text later). In
the next chapter we specialize to compact groups. We will find that their representation
theory has a lot in common with that of the finite groups, in particular the fact that all
irreducible representations are finite-dimensional and contained in the regular represen-
tation (the famous Theorem of Peter and Weyl). But there is an important difference:
the number of equivalence classes of irreducible representations may be infinite.
We do not prove all the modifications in the general theorems when linear representa-
tions are specialized to continuos representations, but concentrate in the next chapters
on an explicit description of the representation theory for SU(2) resp. SO(3) and the
other examples mentioned in the introduction.
The map
G × G −→ G, (a, b) −→ ab−1
is continuous.
Remark 3.1: This requirement is equivalent to the following three conditions which are
more convenient for checking
Every abstract group can be regarded as a topological group if one gives it the discrete
topology where every subset is open.
All the matrix groups presented in Section 0.1 and later on are topological groups: As
2 2
they are subsets of Mn (R) Rn , they inherit the (standard) topology from Rn . Open
balls in Mn (R) (and analogously in Mn (C)) are
where
(3.1) A := < A, A >, < A, B >:= Re Tr tĀB.
This topology is consistent with the group structure: multiplication of two matrices gives
a matrix where the coefficients of the product matrix are polynomials in the coefficients of
the factors and thus the two continuity requirements 1.) and 3.) in the remark above are
fulfilled. Similarly with condition 2.), as the inverse of a matrix has coefficients which
are polynomials in the coefficients of the matrix divided by the determinant, another
polynomial here not equal to zero, i.e. we have again a continuous function.
Now we have the possibility to define the class of groups we will restrict to in this text:
Remark 3.3: More generally, one calls also linear those groups, which are isomorphic
to a closed subgroup of GL(n, H), H the quaternion skewfield, which in this text will
only appear in a later application.
Remark 3.4: In another kind of generality one treats Lie groups, which are topological
groups and at the same time differentiable manifolds such that
are differentiable maps. As we try to work without the (important) notion of a manifold
as long as possible, it is quite adequate to stay with the concept of linear groups as
defined above (which comprises all the groups we are interested in here).
3.2 The Continuity Condition 33
While treating topological notions, let us remind that a group G is called (path-)connected
if for any two a, b ∈ G there is a continuous map of a real interval mapping one endpoint
of the interval to a and the other to b.
Exercise 3.1: Show that SU(2) and SL(2, R) are connected and GL(2, R) is not.
Before we come to the definition of continuous representations, let us adapt the notion
of a G-set from Section 0.2 to the case that G is a topological group.
is continuous.
If the space X is Hausdorff (as all our examples will be), then the stabilizing subgroup
Gx ⊂ G of each point x ∈ X is a closed subgroup of G. Conversely, if H is a closed
subgroup of a Hausdorff group G, then the space G/H of left cosets gH, g ∈ G, is a
homogeneous Hausdorff G-space when given the usual quotient space topology. If X is
any other topological G-space with the same stabilizing group Gx = H for x ∈ X , then
the natural map
Φ : G/H −→ X , gH −→ gx
is 1-1 and continuos. If G and X are locally compact, then Φ is a homeomorphism and
allows an identification of both spaces.
• An operator T is said to be bounded iff there exists a constant C such that one has
T v ≤ C v for all v ∈ D(T ).
• The norm T of a bounded operator T is defined as the smallest C such that
Tv ≤ C v for all v ≤ 1
.
Remark 3.5: A bounded linear operator is uniformly continuous. Conversely, if a linear
operator T is continuous at a point v0 (e.g. v0 = 0), then T is bounded.
• Let A be a linear operator in H with dense domain D(A) ⊂ H. Then one can prove
that there are v, v ∈ H such that < Au, v > = < u, v > holds for all u ∈ H. One
sets v =: A∗ v, verifies that A∗ is a linear operator, and calls A∗ the adjoint of A.
• For operators A and B in H one writes B ⊃ A iff one has D(B) ⊃ D(A) and
Bu = Au for all u ∈ D(A). Then B is called an extension of A.
• A linear operator A with dense domain D(A) is called symmetric iff one has A∗ ⊃ A
and
< Au, v > = < u, Av > for all u, v ∈ D(A).
If not mentioned otherwise, L(H) is provided with the strong operator topology ([BR]),
i.e. we have as basis of open neighbourhoods
Remark 3.6: It is quite obvious that one could also ask for the following continuity
requirements:
a) The map G × H −→ H, (g, v) −→ π(g)v is continuous.
b) The map G −→ GL(H), g −→ π(g) is continuous.
and
c) For each v ∈ V, the map G −→ H, g −→ π(g)v is continuous in g = e and
there is a uniform bound for π(g) in some neighbourhood of e ([Kn] p.10).
In all examples in this section and in most cases of the following sections all these con-
ditions are fulfilled. We will not discuss the hypotheses necessary so that all these
conditions are equivalent and refer to this to [Ki] p.111, [Kn] p.10 and [BR] p.134. But
we will show for some examples how the continuity of a representation comes out. Later
we will not examine continuity and leave this to the reader’s diligence. (To check this
and the question whether a representation from the examples to be treated is bounded
or not may even be (part of) the subject of a bachelor thesis.)
Later we will modify the notions of reducibility and irreducibility for continuous rep-
resentations. But here we can already prepare the way how to do this and prove the
following:
Remark 3.7: Each irreducible continuous representation π of G = SO(2) is equivalent
to some χk , k ∈ Z.
Proof: We look at (π, V ) only as a linear representation and apply Schur’s Lemma from
1.5. As G is abelian, we can conclude by Theorem 1.4 following from Schur’s Lemma in
1.5 that V = C. Hence we can assume π : SO(2) −→ C∗ . As SO(2) is homeomorphic to
R/Z, we have a continuous map r : R −→ R/Z. Now we use the assumption that π is
continuous and we have a sequence of continuous maps
R −→ SO(2) −→ C∗ , t −→ r(t) −→ π(r(t)),
and in consequence a continuous map
R
t −→ π(r(t)) =: χ(t) ∈ C∗ .
As π is a linear representation, χ has the property
χ(t1 + t2 ) = χ(t1 )χ(t2 ), χ(t) = 1 for all t ∈ 2πZ.
Here we have to recall a fact from standard calculus (see for instance [Fo] p.75): Every
continuous function χ : R −→ C∗ obeying the functional equation χ(t1 +t2 ) = χ(t1 )χ(t2 )
is an exponential function, i.e. χ(t) = c exp(tζ) for c ∈ C∗ , ζ ∈ C. The condition
χ(2πn) = 1 for all n ∈ Z fixes c = 1 and ζ = ki for k ∈ Z.
By the way, already here, we can can anticipate a generalization of Proposition 2.4 in
2.2 from finite to compact groups: L2 (SO(2)) can be identified with a space of periodic
functions. And from the theory of Fourier series one knows that this space has as a
Hilbert space basis the trigonometric functions ej , j ∈ Z, with
ej (t) := e2πijt , t ∈ R.
36 3. Continuous Representations
This can be translated into the statement that each irreducible continuous representation
of SO(2) is contained (with multiplicity one) in the left- (or right-)regular representation
λ resp. ρ of SO(2) and λ and ρ are direct sums of the χj , j ∈ Z.
Example 3.2: Let G be SO(3) and, for fixed m ∈ N, let πm be the representation of G
given on the space Vm of all homogeneous polynomials of degree m in 3 variables, i.e.
by
π(g)P (x) := P (g −1 x), x = t (x1 , x2 , x3 ) (a column).
Moreover, let us announce here that we shall have to discuss the relationship of the two
examples G = SO(3) and G = SU(2) very thoroughly later. The representations πj of
SU(2) will come out as irreducible, but here we see already that for instance for m = 2
+ x23 )C since each element g ∈ SO(3)
(πm , Vm ) has an invariant subspace V0 := (x21 + x22
acts on R3 leaving invariant the spheres given by x2i = ρ2 .
The addition of the continuity condition to the definition of the representation has nat-
ural consequences for the notions related to the definition of the linear representation in
Sections 1.1 and 1.2: Here only closed subspaces are relevant. Hence in the future, we
use definitions modified as follows.
• The representation π2 on the factor (or quotient) space V /V1 is called a (topological)
factor representation.
• unitary iff V is a Hilbert space H and every π(g) respects the scalar product in H.
Theorem 3.1: If π and π are unitary irreducible representations in the Hilbert spaces
H resp. H and F : H −→ H is an intertwining operator, then F is either an isomor-
phism of Hilbert spaces or F = 0.
The proofs of these theorems (see e.g. [BR] p. 143/4 or [Kn] p.12) use some tools from
functional analysis (the spectral theorem) which we do not touch at this stage.
We indicate another very useful notion (in particular concerning unitary representations).
Theorem 3.3: Every unitary representation (π, H) of G is the direct sum of cyclic rep-
resentations.
As the proof is rather elementary and gives some indications how to work in the infinite-
dimensional case and how the continuity condition works, we give it here (following [BR]
p.146):
Let Hv1 be the closure of the linear span of all π(g)v1 , g ∈ G for any 0 = v1 ∈ H.
Then Hv1 is π-invariant: Indeed, let Hv 1 be the linear span of all π(g)v1 , then for each
u ∈ Hv1 we have a sequence (un ), un ∈ Hv 1 which converges to u. Obviously, one has
π(g)un ∈ Hv 1 . The continuity of each π(g) implies π(g)un −→ π(g)u and hence we have
π(g)u ∈ Hv1 and Hv1 is invariant. Thus the subrepresentation π1 = π |Hv1 is cyclic with
cyclic vector v1 .
If Hv1 = H, choose 0 = v2 ∈ Hv⊥1 = H \ Hv1 , consider the closed linear span Hv2 which
is π-invariant and orthogonal to Hv1 , and continue like this if H = Hv1 ⊕ Hv2 .
Let ξ denote the family of all collections {Hvi }, each composed of a sequence of mutu-
ally orthogonal, invariant and cyclic subspaces. We order the family by means of the
38 3. Continuous Representations
inclusion relation. Then ξ is an ordered set to which Zorn’s Lemma applies: It assures
the existence of a maximal collection {Hvi }max . By the separability of H, there can
be at most a countable number of subspaces in {Hvi }max and their direct sum, by the
maximality of {Hvi }max must coincide with H.
To close these general considerations let us recall the following. If H and H are Hilbert
spaces with scalar products <, > resp. <, > , then we can define a scalar product in
H ⊗ H by the formula
< v ⊗ v , w ⊗ w >:=< v, w >< v , w > .
If either H or H is finite-dimensional, then the space H ⊗ H equipped with this scalar
product is complete. If both H and H are infinite-dimensional, we complete H ⊗ H by
the norm defined by the scalar product above and denote this by H⊗H ˆ .
If X is a topological space with the family V = {Oi }i∈I of open sets, it is also a Borel
space with a family B of Borel sets B, where these sets are obtained from open and closed
sets by the operation of countable unions, countable intersections and complementation.
A little more generally, a Borel space is a set with a family B of subsets B such that B
is a σ-algebra, i.e. one has
X \ B ∈ B for B ∈ B,
i∈I Bi ∈ B for Bi ∈ B, i ∈ I N,
i∈I Bi ∈ B for Bi ∈ B, i ∈ I N.
where µ is σ-additive,
i.e.
i) µ(∪Bi ) = µ(Bi ) if the Bi are pairwise disjoint (i ∈ I N), and
ii) there is a covering of X by sets Bi of finite measure, i.e. we have Bi ∈ B, i ∈ I N
with µ(Bi ) < ∞ and ∪Bi = X .
Then the measure µ is called right-invariant iff one has µg = µ for all g ∈ G. Similarly
µ is called left-invariant iff we have a left action and gµ = µ for all g ∈ G where
gµ(B) = µ(gB) for B ∈ B.
Theorem 3.4 (Haar): On every locally compact group with countable basis for the
topology, there exists a nonzero left-invariant (and similarly a right-invariant) measure.
It is defined uniquely up to a numerical factor.
We will not go into the proof of the Theorem (see for instance [HR] Ch.IV, 15) but give
later on several concrete examples for its validity. The most agreable case is the following:
To make these up to now only abstract concepts work in our context, we have to recall
a more practical aspect due to the general integration theory:
Let (X , B, µ) be a measure space and f : X −→ C a measurable function, i.e. f is the
sum f = f1 − f2 + i(f3 − f4 ) of four real nonnegative functions fj (j = 1, .., 4) having the
property
{x ∈ X ; fj (x) > a} ∈ B for all a > 0.
40 3. Continuous Representations
For measurable nonnegative functions f general integration theory provides a linear map
f −→ µ(f ) = f (x)dµ(x) ∈ C ∪ {∞}.
X
We call a measurable function f integrable iff |f (x)| dµ(x) < ∞.
If X = G as above, one usually writes
Proposition 3.1: Let G be a locally compact group. Then there is a continuous homo-
morphism ∆G , called the modular function of the group G, into the multiplicative group
of positive real numbers for which the following equalities hold
Proposition 3.2: If G is compact, then the Haar measure is finite and biinvariant.
Proof: The image of a compact group under the continuous homomorphism ∆G is again
a compact group. The multiplicative group R>0 has only {1} as a compact subgroup.
So
by Proposition 3.1, we have a biinvariant measure which can be normalized to G dg = 1.
3.4 Examples
Most of the noncompact groups to be treated later on are also unimodular, e.g. the group
SL(2, R) or the Heisenberg group Heis(R). We will show here how this can be proved.
gg = (λ + λ , µ + µ , κ + κ + λµ − λ µ).
with det Jg0 (g) = 1. Hence we see that dg is left-invariant. Similarly we prove the right-
invariance.
Remark 3.8: Without going into the general theory of integration on manifolds, we
indicate that, for calculations like this, it is very convenient to associate the infinitesimal
measure element with an appropriate exterior or alternating differential form and then
check if this form stays invariant. In the example above we take
and calculate using the the general rules for alternating differential forms (we shall say
a bit more about this later in section 8.1), in particular
dµ ∧ dλ = − dλ ∧ dµ, dµ ∧ dµ = 0,
as follows
cos ϑ sin ϑ
SO(2)
r(ϑ) = ( ).
− sin ϑ cos ϑ
This group will be our main example for the representation theory of compact groups.
As standard parametrization we use the same as Wigner does in [Wi] p.158: The general
form of a two–dimensional unitary matrix
a b
g=( ).
c d
of determinant one is
a b
g=( ), a, b ∈ C with | a2 | + | b2 | = 1.
−b̄ ā
42 3. Continuous Representations
With
eiα cos ϑ sin ϑ
s(α) := ( ) and r(ϑ) = ( )
e−iα − sin ϑ cos ϑ
one has a decomposition
a b
SU(2)
g = ( ) = s(−α/2)r(−β/2)s(−γ/2), α, γ ∈ [0, 2π], β ∈ [0, π].
−b̄ ā
In particular, one can see that SU(2) is generated by the matrices s(α), r(β).
Later on we shall see later that the angles α, β, γ correspond to the Euler angles for
the three-dimensional rotation introduced by Wigner in [Wi] p.152. In Hein [He] one
finds all these formulas, but unfortunately in another normalization, namely one has
zHein = a, uHein = −b, αHein = β, βHein = −α, γHein = −γ. Our parametrization
is one-to-one for α, γ ∈ [0, 2π], β ∈ [0, π] up to a set of measure zero. An invariant
normalized measure is given by
Example 3.6: An example for a group G which is not unimodular is the following
a b
G = {g = ( ) ; a ∈ R∗ , b ∈ R}
1
aa ab + b
We have gg = ( ) =: g , hence
1
a a a b + b
so that dr g = a−1 dadb and, from g g = ( ) =: g ∗
1
1 x y 1/2
G = {g = ( )( ); y > 0, x ∈ R}.
1 y −1/2
Chapter 4
Representations of Compact
Groups
As already mentioned, the representation theory of compact groups generalizes the the-
ory of finite groups. We shall closely follow the presentation in [BR], and cite from [BR]
p.166: “The representation theory of compact groups forms a bridge between the rela-
tively simple representation theory of finite groups and that of noncompact groups. Most
of the theorems for the representations of finite groups have direct analogues for compact
groups and these results in turn serve as the starting point for the representation theory
of noncompact groups.”
Proposition 4.1: Let (π, H) be a representation of G in the Hilbert space H with scalar
product <, > . Then there exists a new scalar product <, > defining a norm equivalent to
the initial one, relative to which the map g −→ π(g) defines a unitary representation of G.
Proof: i) As to be expected from the case of finite groups, we define <, > by
< u, v > := < π(g)u, π(g)v > dg.
G
Sesquilinearity and hermiticity of <, > are obvious. Let us check that we can conclude
u = 0 from
< u, u > = < π(g)u, π(g)u > dg = 0 :
G
44 4. Representations of Compact Groups
From G < π(g)u, π(g)u > dg = 0 one deduces that < π(g)u, π(g)u > = 0 almost every-
where. If g ∈ G is such that π(g)u = 0, then π(g)−1 π(g)u = u = 0.
because dg is left-invariant.
iii) The norms . and . induced by <, > resp. <, > are equivalent:
We have
u 2 = < π(g)u, π(g)u > dg
G
< u, u > dg = N 2 u
2
≤ sup π(g) 2
g∈G G
it follows that
u < u, u > dg
2
=
G
≤ N2 < π(g)u, π(g)u > dg = N 2 < u, u >= N 2 u 2 .
G
iv) Equivalent norms define the same families of open balls and thus equivalent topolo-
gies, i.e. if g −→ π(g)v, v ∈ H is continuous for the topology belonging to . , it is also
continuous for the topology deduced from . .
We now turn to the already announced result that every unitary irreducible represen-
tation is finite-dimensional. In its proof we use an interesting tool, namely the Weyl
operator Ku defined for elements u, v from a Hilbert space H with π-invariant scalar
product by
Ku v := < π(g)u, v > π(g)udg.
G
(4.1) < Ku v, v > = < v, π(g)u >< π(g)u, v > dg
G
= |< π(g)u, v >|2 dg = λ(u) v 2 .
G
By interchanging the rôles of u and v and using the equality (which is equivalent to the
unimodularity of G)
f (g −1 )dg = f (g)dg,
we get
λ(v) u 2 = |< π(g)v, u >|2 dg =
|< u, π(g)v >|2 dg
−1
= |< π(g )v, u >| dg =
2
|< π(g)u, v >|2 dg
= λ(u) v 2 .
As the nonnegative continuous function g −→ |< π(g)u, u >| assumes the value u = 1
at g = e, we must have c > 0.
Let {ei }i=1,...,n be a set of ON-vectors in H. In (4.1) we now put u = ek and v = e1 to
obtain
|< π(g)ek , e1 >|2 dg = λ(ek ) e1 2 = c.
and
n
n
nc = |< π(g)ek , e1 >|2 dg = |< π(g)ek , e1 >|2 dg
k=1 G G k=1
≤ e1 2 dg = 1.
G
46 4. Representations of Compact Groups
The proof uses some more tools from functional analysis and so we skip it here (see [BR]
p.169/170). But we mention the appearance of the important notion of a Hilbert–Schmidt
operator, which is an operator A in H, s.t. for an arbitrary basis {ei }i∈I of H we have
Aei 2 < ∞.
i∈I
Again, as in the theory of finite groups, we have ON-relations for the matrix elements of
an irreducible representation.
Theorem 4.3: Let π and π be two irreducible unitary representations of G and A(g)
resp. A (g) their matrices with respect to a basis {ei } of H and {ek } of H . Then one
has relations
Aij (g)Akl (g)dg = 0 if π ∼ π
= (1/n)δik δjl if π ∼ π and n := dim H.
For π = π , Schur’s Lemma requires Fij = λij id. Hence for (r, t) = (i, j), the orthogo-
nality relations just obtained are still satisfied. For (r, t) = (i, j), we have
(Fij )ij = Aii (g)Ajj (g)dg = λij = 0 if i = j,
= | Aii (g) |2 dg = λii if i = j.
4.1 Basic Facts 47
Exercise 4.1: Show the same fact for the left regular representation.
Remark 4.2: With the same reasoning as in the case of finite groups, the characters χπ
of finite-dimensional irreducible
unitary representations π, defined by χπ (g) := Tr π(g)
resp. χπ (g) := i Aii (g) = i < A(g)ei , ei > for a matrix form with respect to a basis
e1 , . . . , en , have the following properties.
1) The characters are class functions, i.e. χ(g0 gg0−1 ) = χ(g) for all g, g0 ∈ G.
2) We have χ(g −1 ) = χ(g).
3) π ∼ π , then χπ = χπ .
If
4) χπ (g)χπ (g)dg = 0 if π ∼ π and = 1 if π ∼ π .
h
χπ = mi χπi .
i=1
and
h
m2i = χπ (g)χπ (g)dg.
i=1 G
The big difference to the theory of finite groups is that for compact groups the number of
equivalence classes of irreducible representations no longer necessarily is a finite number.
We already know this from the example G = SO(2) in section 3.2. The central fact is
now the famous Peter–Weyl Theorem:
Theorem 4.5: Let Ĝ = {πi }i∈I be the unitary dual of G, i.e. a complete set of represen-
tatives of the equivalence classes of irreducible unitary representations of G. For every
i ∈ I let Ai (g) = (Aijk (g))j,k=1,...,ni be a matrix form of πi and
i √
Yjk (g) := ni Aijk (g).
Then we have
1) The functions Yjki
, i ∈ I, j, k = 1, . . . , ni form a complete ON-system in L2 (G).
2) Every irreducible unitary representation π of G occurs in the decomposition of
the right regular representation with a multiplicity equal to the dimension of π, i.e.
mult (πi , ρ) = ni .
3) Every C-valued continuous function f on G can be uniformly approximated by a
i
linear combination of the (Yjk ).
4) The characters (χπi )i∈I generate a dense subspace in the space of continuous class
functions on G.
The proof of these facts use standard techniques from higher analysis. We refer to [BR]
p. 173 - 176 or [BtD] p.134.
We put
a b
G = SU(2)
g = ( ), | a | 2 + | b | 2 = 1
−b̄ ā
and for j a half integral nonnegative number, we let V (j) be the space C[x, y]2j of
homogeneous polynomials of degree 2j in two variables. Thus we have dim V (j) = 2j + 1.
We choose as a basis of V (j) monomials fp , which are conveniently normalized by
xj+p y j−p
fp (x, y) :=
, p = −j, −j + 1, . . . , j.
(j + p)! (j − p)!
We define the representation πj by the action
x −1 x
(g, ) −→ g ,
y y
4.2 The Example G = SU(2) 49
i.e. we put
πj (g)fp (x, y) := fp (āx − by, b̄x + ay),
and get by a straightforward computation
πj (g)fp (x, y) = Σp fp (x, y)Ajp p (g)
with
j+p
(j + p)!(j − p)!(j + p )!(j − p )!
Ajp p (g) = (−1) k
aj−p −k āj+p−k bk b̄k+p −p
(j − p − k)!(j + p − k)!k!(k + p − p)!
k=0
in particular
(2j)!
(4.2) Ajjp (g) = .
(j + p)!(j − p)!
Proof: i) Unitarity
Wigner’s proof relies on the fact that the polynomials fp are normalized so that we have
j
1 (| x2 | + | y 2 |)2j
f¯p fp = | x2 |j+p | y 2 |j−p = .
p=−j p=−j
(j + p)!(j − p)! (2j)!
j
j
| āx − by |2(j+p) | b̄x − ay |2(j−p)
| πj (g)fp (x, y) |2 =
p=−j p=−j
(j + p)!(j − p)!
(| āx − by |2 + | b̄x − ay |2 )2j (| x |2 + | y |2 )2j
= = ,
(2j)! (2j)!
where the last equality follows from the unitarity of g. So we have the invariance
Σp | π(g)fp |2 = Σp | fp |2 .
If we know that the (2j + 1)2 functions f¯p fp are linearly independent, we see that we
have
Σp Ājp p (g)Ajp p (g) = Σp Ājp p (g)tAjpp (g) = δp p ,
i.e. the unitarity of Aj (g). Hence we still have to show the linear independence of the
functions f¯p fp : We look at a relation
cp p x̄j+p ȳ j−p xj+p y j−p = 0, cp p ∈ C.
p ,p
50 4. Representations of Compact Groups
But this condition implies that cp ,q−2j−p = 0 (and thus the linear independence of the
f¯p fp ), because we can write y/ȳ = eit , t ∈ R, and a relation
Σcp ,q−2j−p eitp = 0 for all t ∈ R
requires that all the coefficients c must vanish.
ii) Irreducibility
One can also find a proof by application of the criterium given in the Corollary at the end
of Section 4.1 using the character of πj . We will see this as a byproduct of the following
third part.
iii) Completeness
eiα
s(α) := ( ), α ∈ [0, π), resp. s(α/2), α ∈ [0, 2π).
e−iα
its trace is
j
χπj (s(−α/2)) = eipα =: ξj (α).
p=−j
So we have
ξ0 (α) = 1,
ξ1/2 (α) = e−iα/2 + eiα/2 = 2 cos(α/2),
ξ1 (α) = e−iα + 1 + eiα = 2 cos α + 1,
ξ3/2 (α) − ξ1/2 (α) = 2 cos(3α/2), etc.
It is now evident that SU(2) can have no other representations than the πj , j ∈ (1/2)N0 :
For the character of such a representation must, after multiplication by a weighting
function, be orthogonal to all ξj , and therefore to ξ0 , ξ1/2 , ξ1 − ξ0 , ξ3/2 − ξ1/2 , . . . . But
a function, which is orthogonal to 1, 2 cos(α/2), 2 cos α, 2 cos(3α/2), . . . in [0, 2π) must
vanish according to Fourier’s Theorem.
Exercise 4.3: Verify the irreducibilty of πj using the character criterium (Corollary to
Remark 4.2 in Section 4.1).
i.e.
π1 ⊗ π1 = π0 + π1 + π2 .
j+j
πj ⊗ π =j π .
=|j−j |
52 4. Representations of Compact Groups
The elements g of SO(3) are real 3 × 3-matrices consisting of three ON-rows (or columns)
(a1 , a2 , a3 ), i.e. we have six conditions < aj , ak >= δjk for the nine entries in g. Hence
we expect three free parameters. It is a classical fact that the elements of SO(3) are
parametrized by three Euler angles α, β, γ. Again we follow Wigner’s notation ([Wi]
p.152):
Proposition 4.2: For every g ∈ SO(3) there exist angles α, β, γ ∈ R, such that
Here S3 (α) and S3 (γ) are rotations about the z-axis (or e3 ∈ R3 ) through α resp. γ and
S2 (β) is a rotation about the y-axis (or e2 ∈ R3 ), i.e. we put
⎛ ⎞
cos α − sin α
S3 (α) := ⎝ sin α cos α ⎠,
1
⎛ ⎞
cos α − sin α
S2 (α) := ⎝ 1 ⎠,
sin α cos α
⎛ ⎞
1
S1 (α) := ⎝ cos α − sin α ⎠ .
sin α cos α
Exercise 4.5: Prove this.
Remark 4.3: i) There are several notational conventions for the Euler angles in com-
mon use. In Hein [He] p.56 we find a statement (with proof) analogous to our Proposi-
tion 4.2 with S2 (β) replaced by S1 (β). For a parametrization of SO(3) one needs only
α, γ ∈ [0, 2π) and β ∈ [0, π]. This parametrization is not everywhere injective: for β = 0
α and γ are only fixed up to their sum. Goldstein [Go] p.145-148 gives an overview of
other descriptions.
ii) SO(3) is generated by rotations S3 (t) and S2 (t) or by rotations of type S3 (t) and S1 (t).
The key fact for our discussion here is given by the following theorem (we shall later
see that this is a special case of an analogous statement about SL(2, C) covering the
(restricted) Lorentz group SO(3, 1)+ ):
4.3 The Example G = SO(3) 53
ϕ : SU(2) −→ SO(3)
1 i −1
sx := ( ), sy := ( ), sz := ( ).
1 −i 1
V := {H ∈ M2 (C); t H = H̄, Tr H = 0}
−z x + iy
= {H = ( ) = xsx + ysy + zsz ; x, y, z ∈ R}.
x − iy z
a b
which for H as above and g = ( ) gives by a straightforward computation
−b̄ ā
−z x + iy
H = ( )
x − iy z
with
E.g. from the fact that the map det is multiplicative, we deduce
det H = x2 + y 2 + z 2 = det H = x + y + z
2 2 2
x = A(g)x,
we see that multiplication with A(g) respects the euclidean quadratic form, i.e. we have
A(g) ∈ O(3). Moreover, by a continuity argument, we have A(g) ∈ SO(3): as every
g ∈ SU(2) can be deformed contiuously to g = E2 and det is a continuous function, we
cannot have det A(g) = −1.
54 4. Representations of Compact Groups
If a representation π : SU(2) −→ GL(V ) factorizes through SO(3), i.e. one has a homo-
morphism π : SO(3) −→ GL(V ) with π = π · ϕ resp. a commutative diagram
π
SU(2) - GL(V )
@
ϕ@ π
@
@
R
SO(3)
then, this way, π induces the representation π of SO(3). Obviously this is the case for a
representation π with π(−E2 ) = id, where we get a unique prescription by putting
π (A) := π(g) if A = ϕ(g).
For π(−E2 ) = −id one gets a double valued representation by putting
π (A) := ±π(g) if A = ϕ(g) = ϕ(−g).
(This is a special case of a projective representation, which will be discussed below.)
In consequence, we know the unitary dual of SO(3): The representations πj of SU(2) with
j ∈ N0 are exactly those with π(−E2 ) = π(E2 ) = id and thus induce (up to equivalence)
all irreducible unitary representations π of SO(3): If there were any π not in this family,
it would composed with ϕ produce a representation of SU(2), which by our completeness
result for SU(2) would be equivalent to a πj .
For g ∈ SU(2) and ϕ(g) = g , Wigner ([Wi] p.167) conjugates the π (g ) := πj (g) = Aj (g)
from 4.2 by the diagonal matrix M = ((i)−2k δkl ) and calls the result Dj ({α, β, γ}) with
{α, β, γ} representing the Euler angles of g ∈ SO(3) in the parametrization chosen above.
Hence we have from the explicit form of the matrices Aj (g) obtained in 4.2 the explicit
form
(j + p)!(j − p)!(j + p )!(j − p )!
j k
D ({α, β, γ})p p = (−1)
(j + p − k)!(j + p − k)!k!(k + p − p)!
k
ip α
× e (cos(β/2))2j+p+p −2k (sin(β/2))2k+p −p .
4.3 The Example G = SO(3) 55
j
χπj (S3 (α)) = eipα = 1 + 2 cos α + · · · + 2 cos(jα), j integral
p=−j
= 2 cos(α/2) + 2 cos(3α/2) + · · · + 2 cos(jα), j half − integral.
π̃ : G −→ GL(V ),
where
c : G × G −→ K∗
with
c(g, g )c(gg , g ) = c(g, g g )c(g , g ) for all g, g , g ∈ G.
π̃(g(g g )) = π̃((gg )g ).
In the second part of his book [Wi], Wigner describes the application of these results
to the theory of spin and atomic spectra. Though we are tempted to do more, from
this we only take over that, roughly, an irreducible unitary (projective) representation πj
of SO(3) corresponds to a particle having three-dimensional rotational symmetry. The
number j specifying the representation then is its angular momentum quantum number
or its spin and the indices numbering a basis of the representation space are interpreted
as magnetic quantum numbers. We strongly recommend to the reader interested in this
topic to look into Wigner’s book or any text from physics.
56 4. Representations of Compact Groups
To finish this chapter, we briefly indicate another approach to the determination of the
representations of G = SO(3), which is a first example for the large theory of eigenspace
representations promoted in particular by Helgason.
From analysis it is well known that the three-dimensional euclidean Laplace operator
x = r sin ϑ cos ϕ,
y = r sin ϑ sin ϕ,
z = r cos ϑ.
in the variables (ϑ, ϕ) results. The (2 + 1) linearly independent solutions Y = Ym (with
m = −, − + 1, . . . , , ) are known as spherical harmonics of the −th degree. They have
the form
Ym (ϑ, ϕ) = Φm (ϕ)Θm (ϑ)
where
1
Φm (ϕ) := √ eimϕ ,
2π
2 + 1 ( − m)! 1/2 dm
Θm (ϑ) := (−1)m ( ) (sin ϑ)m Pl (cos ϑ),
2 ( + m)! d(cos ϑ)m
2 + 1 ( − m)! 1/2 dm
Θ,−m (ϑ) := ( ) (sin ϑ)m Pl (cos ϑ),
2 ( + m)! d(cos ϑ)m
with m ≥ 0 in the last two equations. The P are the Legendre polynomials defined by
1 d 2
P (x) := (x − 1) .
2 ! dx
4.3 The Example G = SO(3) 57
π (g)r Ym (ϑ, ϕ) =
Ym (ϑ, ϕ)Dm m (g).
m =−
imα
Dm m (S3 (α)) = e δm m .
By applying Schur’s commutation criterium, one can prove here, like we did already
above, that D is irreducible for all = 0, 1, 2, . . . .
The conjugacy classes of SO(3) are parametrized by the angle α, α ∈ [0, π). For the
characters we obtain the formula
χ() (S3 (α)) = eimα = 1 + 2 cos α + · · · + 2 cos(α).
m=−
Again we can see that we got all irreducible representations as {cos nϕ}n∈N0 is a complete
system in L2 ([0, π)) by Fourier’s Theorem.
Chapter 5
Representations of Abelian
Groups
Remark 5.1: We already defined the unitary dual Ĝ of any group G as the set of
equivalence classes of unitary irreducible representations of G. For G abelian, Ĝ consists
just of all characters of G. It is also called the Pontrjagin dual of G.
Remark 5.2: For G abelian, Ĝ is also a group: We define the composition of two
characters χ and χ by
(χχ )(g) := χ(g)χ (g) for all g ∈ G.
If G is a topological group, Ĝ is also. More precisely, if G is locally compact (resp. com-
ˆ
pact, resp. discrete), Ĝ is locally compact (resp. discrete, resp. compact). One has G Ĝ.
60 5. Representations of Abelian Groups
i.e. we have Ĝ = G.
Example 5.2: For G = S 1 = {ζ ∈ C; | ζ | = 1} every character has the form
χ(ζ) = ζ k , k ∈ Z,
f (t) = ikt
c(k)e , c(k) := (1/(2π)) e−ikt f (t)dt
k∈Z 0
for G = SO(2) S 1 to other compact groups. Now, the analogon of Fourier series for
G = R is the Fourier transform associating to a sufficiently nice function f its Fourier
transform
√ ∞
(5.1) ˆ
Ff (k) = f (k) := (1/ 2π) χk (t)f (t)dt, χk (t) := eikt .
−∞
This may be interpreted in the following way: In the case of compact groups we had a
(denumerable) direct sum decomposition of the regular representation ρ into irreducible
subrepresentations πj
ρ= πj .
j∈J
operators into a direct integral. Following [BR] p.649 resp. [Ki] p.57, we present them
briefly as they are also essential for subsequent chapters.
Spectral Measure
Let [a, b] ⊂ R be a finite or infinite interval and E be a function on [a, b] with values E(λ)
for λ ∈ [a, b], which are operators in a Hilbert space H. E is called a spectral function if
it satisfies the following conditions:
i) E(λ) is self-adjoint for each λ,
ii) E(λ)E(µ) = E(min(λ, µ)),
iii) The operator function E is strongly right continuous, i.e. one has
lim E(λ + )u = E(λ)u for all u ∈ H,
→0+
Conditions i) and ii) mean that E(λ), λ ∈ [a, b], are (refining Definition 1.7) orthogonal
projection operators. For an interval ∆ = [λ1 , λ2 ] ⊂ [a, b] one denotes the difference
E(λ2 ) − E(λ1 ) by E(∆).
For ∆1 , ∆2 ⊂ [a, b] one has
E(∆1 )E(∆2 ) = E(∆1 ∩ ∆2 ),
in particular
E(∆1 )E(∆2 ) = 0 if ∆1 ∩ ∆2 = ∅,
i.e. the subspaces H1 = E(∆1 )H and H2 = E(∆2 )H are orthogonal.
is a spectral function.
The properties of the spectral function imply that, for any u ∈ H, the function
σu (λ) := < E(λ)u, u >
is a right continuous, non-decreasing function of bounded variation with
σu (−∞) = 0, σu (∞) = u 2 .
Moreover it is denumerable additive: For a pairwise disjoint decomposition {∆n } with
∆ = ∪∞
n=1 ∆n , one has
∞
∞
The function σu is called the spectral measure. We now can state the fundamental theo-
rem of spectral decomposition theory as follows:
62 5. Representations of Abelian Groups
The operation of direct sum of Hilbert spaces admits a generalization: Let there be given
a set X with measure µ and a family {Hx }x∈X of Hilbert spaces. It is natural to try to
define a new space
H= Hx dµ(x),
X
the elements of which are to be functions f on X, assuming values in Hx for all x ∈ X,
and the scalar product being defined by the formula
< f1 , f2 > := < f1 (x), f2 (x) >Hx dµ(x).
X
The difficulty consists in the fact that the integrand in this expression may be non-
measurable. In the case where all of the spaces Hx are separable, one can introduce
an appropriate definition of measurable vector-functions f , which will guarantee this
measurability of the numerical function x −→< f1 (x), f2 (x) >: In the special example
that all Hx coincide with a fixed separable H0 with basis {ej }, the numerical functions
x −→< f (x), ej > are measurable (for more general cases see [Ki] p.57). Once an
appropriate concept of measurable vector-functions
has been adopted one defines the
continuous sum (or direct integral ) H = X Hx dµ(x) as the set of equivalence classes of
measurable vector-functions f with summable square of norms.
An operator A in H = X Hx dµ(x) is called decomposable iff there exists a family {A(x)}
of operators in the spaces Hx such that
By a suitable generalization of the notion of spectral measure to the dual Ĝ (which for
an abelian locally compact group G can be given the structure of a topological space),
one has the general statement:
A is called
xy = yx for all x, y ∈ A,
64 6. The Infinitesimal Method
– a Lie algebra iff the multiplication xy, usually written as xy =: [x, y] (the Lie
bracket), is anti-symmetric, i.e.
[x, [y, z]] + [y, [z, x]] + [z, [x, y]] = 0 for all x, y, z ∈ A.
A × A −→ A, (x, y) −→ xy
n
∂f ∂g ∂f ∂g
{f, g} := ( − ) for all f, g ∈ C ∞ (R2n ),
i=1
∂q i ∂p i ∂p i ∂q i
Exercise 6.2: Verify that each Poisson algebra is also a Lie algebra.
gl(n, K) := Mn (K).
Example 6.3: If V is any K-vector space, the space End V of linear maps F from V to
V is a Lie algebra with
Example 6.4: If g is a Lie algebra and g0 is a subspace with [X, Y ] ∈ g0 for all X, Y ∈ g0 ,
g0 is again a Lie algebra. One easily verifies this and that
1 1
F := ( ), G := ( ), H := ( )
1 −1
−i −1 i
X1 := (1/2)( ), X2 := (1/2)( ), X3 := (1/2)( )
−i 1 −i
Exercise 6.6: Show that g = R3 provided with the usual vector product as composition
is a Lie algebra isomorphic to so(3).
and relations
[R, P ] = [R, Q] = 0, [P, Q] = 2R.
Remark 6.1: It will soon be clear that gl(n, R), sl(n, R), so(n), su(n), and heis(R) are
associated to the groups GL(n, R), SL(n, R), SO(n), SU(n), resp. Heis(R).
In analogy with the concept of the linear representation of a group G in a vector space
V , one introduces the following notion.
Each Lie algebra has a trivial representation π̂ = π̂ 0 with π̂ 0 (X) = 0 for all X ∈ g and
an adjoint representation π̂ = ad given by
In analogy with the respective notions for group representations, we have the following
concepts.
– (π̂0 , V0 ) is a subrepresentation of (π̂, V ) iff V0 ⊂ V is a π̂-invariant subspace, i.e. π̂(X)v0 ∈
V0 for all v0 ∈ V0 , and π̂0 = π̂ |V .
– π̂ is irreducible iff π̂ has no nontrivial subrepresentation.
If π̂1 and π̂2 are representations of g we have
– the direct sum π̂1 ⊕ π̂2 as the representation on V1 ⊕ V2 given by
and
– the tensor product π̂1 ⊗ π̂2 as the representation on V1 ⊗ V2 given by
There is a lot of material about the structure and representations of Lie algebras, much
of it going back to Elie and Henri Cartan. As already indicated, a good reference for this
is the book by Humphreys [Hu]. We will later need some of this and develop parts of it.
Here we only mention that one comes up with facts like this: If g is semisimple (i.e. has
no nontrivial ideals), then each representation π̂ of g is completely reducible.
Before we discuss more examples, we establish the relation between linear groups and
their Lie algebras.
Everyone interested in concrete examples could do the following Exercise 6.12 preparing
the ground for the next definition: Verify
1 t 1
exp(tX) = , for X = ,
1
cos t sin t 1
exp(tX) = , for X = ,
− sin t cos t −1
t
e 1
exp(tX) = , for X = ,
e−t −1
⎛ ⎞ ⎛ ⎞
cos t sin t 1
exp(tX) = ⎝ − sin t cos t ⎠ , for X = ⎝ −1 ⎠.
1 0
Exercise 6.13: Verify that one has the following relations
{exp(tX), X ∈ so(3); t ∈ R} = SO(3),
{exp(tX), X ∈ su(2); t ∈ R} = SU(2),
{exp(tX), X ∈ sl(2, R); t ∈ R} = SL(2, R).
This exercise shows that there are groups, which are examples for groups of exponential
type, i.e. images under the expontial map of their Lie algebras. For G = SL(2, R) one
has
{exp(tX), X ∈ sl(2, R), t ∈ R} = SL(2, R),
but the matrices exp(tX), X ∈ sl(2, R), t ∈ R, generate the identity component of
G = SL(2, R).
We take this as motivation to associate a Lie algebra to each linear group by a kind of
inversion of the exponential map:
A hurried or less curious reader may simply believe this and waive the proof. (See for
instance [He] p.114: it is, in principle, not too difficult but lengthy. It shows why (in our
approach) we had to restrict ourselves to closed matrix groups.) We emphasize that it is
essentially the heart of the whole theory that not only a linear group but also every Lie
group has an associated Lie algebra. This is based on another important notion, which
we now introduce because it gives us a tool to determine explicitly the Lie algebras for
those groups we are particularly interested in (even if we we have skipped the proof of
the central theorem above).
This suggests the following procedure to determine the Lie algebra g of a given matrix
group G (which is also the background of our Exercise 6.13 above): Look for “sufficiently
many independent” one-parameter subgroups γ1 , . . . , γd and determine their generators
X1 , . . . , Xd . Then g will come out as the R-vector space generated by these matrices Xi .
Obviously, one needs some explanation:
– Two 1-PUGs γ and γ are independent iff their infinitesimal generators are linearly
independent over R as matrices in Mn (C).
– The number d of the sufficiently many independent (γi ) is just the dimension of our
group G as a topological manifold: Up to now we could work without this notion, but
later on we will even have to consider differentiable manifolds. Therefore, let us recall (or
introduce) that the notion of a real manifold M of dimension d includes the condition
that each point m ∈ M has a neighbourhood which is homeomorphic to an open set in
Rd . This amounts to the practical recipe that the dimension of a group G is the number
of real parameters needed to describe the elements of G.
For instance for G = SU(2) and SO(3) we have as parameters the three Euler angles
α, β, γ, which we introduced in 4.3 and 4.2. Hence, here we have d = 3, as also in the
next example.
we take
⎛ ⎞ ⎛ ⎞ ⎛ ⎞
1 t 1 1 t
γ1 (t) := ⎝ 1 ⎠ , γ2 (t) := ⎝ 1 t ⎠ , γ3 (t) := ⎝ 1 ⎠,
1 1 1
and get ⎛ ⎞ ⎛ ⎞ ⎛ ⎞
1 1
X1 = ⎝ ⎠ , X2 = ⎝ 1 ⎠ , X3 = ⎝ ⎠.
G
g −→ π(g)v ∈ H
is differentiable.
This definition makes sense if we accept that a linear group is a differentiable manifold
and extend the notion of differentiability to vector valued functions. In our examples
things are fairly easy as we have functions depending on the parameters of the group
where differentiability is not difficult to check.
Definition 6.6: Let (π, H) be a continuous representation of a linear group G. Then its
associated derived representation is the representation dπ of g given on the space H∞
of smooth vectors by the prescription
d
dπ(X)v := π(exp(tX))v |t=0 for all v ∈ H∞ .
dt
6.3 Derived Representations 71
for
f ∈ V (1) = < f−1 , f0 , f1 >
with √ √
f−1 (x, y) = (1/ 2)y 2 , f0 (x, y) = xy, f1 (x, y) = (1/ 2)x2 .
As we did in Exercise 6.4 in 6.1, we take as basis of su(2)
−i −1 i
X1 := (1/2)( ), X2 := (1/2)( ), X3 := (1/2)( ).
−i 1 −i
It is clear that V (1) consists of smooth vectors and hence we can compute
and
dt π(exp tX1 )f0 |t=0
d
X̂1 f0 =
= d
dt (cos(t/2)x + i sin(t/2)y)(i sin(t/2)x + cos(t/2)y) |t=0
√
= (i/2)(x2 + y 2 ) = (i/ 2)(f−1 + f1 ),
dt π(exp tX1 )f1 |t=0
d
X̂1 f1 =
√
= d
+ i sin(t/2)y)2 |t=0
dt (1/ 2)(cos(t/2)x
√ √
= (i/ 2)xy = (i/ 2)f0 .
Similarly we get
√ √ √
X̂2 f−1 = −(1/ 2)f0 , X̂2 f0 = (1/ 2)(f−1 − f1 ), X̂2 f1 = (1/ 2)f0 ,
X̂3 f−1 = if−1 , X̂3 f0 = 0, X̂3 f1 = −if−1 .
72 6. The Infinitesimal Method
one has
X̂0 fp = pfp , X̂± fp = fp±1 , p = −1, 0, 1 (f2 = f−2 = 0).
ii) Do the same for π = πj , j > 1.
Example 6.9: We take the Heisenberg group G = Heis(R) ⊂ GL(4, R) and the
Schrödinger representation π = πm , m ∈ R \ {0}, given by
with
g = (λ, µ, κ) ∈ G, x ∈ R, f ∈ H = L2 (R), em (u) := e2πimu .
Here H∞ is the (dense) subspace of differentiable functions in L2 (R). We realize the Lie
algebra g = heis(R) by the space spanned by the four-by-four matrices
⎛ ⎞ ⎛ ⎞ ⎛ ⎞
1
⎜ 1 ⎟ ⎜ 1 ⎟ ⎜ 1 ⎟
P =⎜⎝
⎟, Q = ⎜
⎠ ⎝
⎟, R = ⎜
⎠ ⎝
⎟.
⎠
−1
i.e. we have
P̂ = ∂x , Q̂ = 4πimx, R̂ = 2πim.
Exercise 6.17: i) Verify that these operators really fulfill the same commutation rela-
tions as P, Q, R.
2
ii) Take f0 (x) := e−πmx and
and determine
fp := Ŷ+p f0 , Ŷ− fp and Ŷ0 fp for p = 0, 1, 2, . . . .
These Examples show how to associate a Lie algebra representation to a given group
representation. Now the following question arises naturally: Is there always a nontrivial
subspace H∞ of smooth vectors in a given representation space H? We would even want
6.4 Unitarily Integrable Representations of sl(2, R) 73
that it is a dense subspace. The answer to this question is positive. This is easy in our
examples but rather delicate in general and one has to rely on work by Garding, Nelson
and Harish Chandra. It is not very surprising that equivalent group representations are
infinitesimally equivalent, i.e. their derived representations are equivalent algebra repre-
sentations. The converse is not true in general. But as stated in [Kn] p.209, at least for
semisimple groups one has that infinitesimally equivalent irreducible representations are
unitarily equivalent.
Taking these facts for granted, a discussion of the unitary irreducible representations of
a given linear group G can proceed as follows: We try to classify all irreducible repre-
sentations of the Lie algebra g = Lie G of G and determine those, which can be realized
by derivation from a unitary representation of the group. In general, this demands for
some more information about the structure of the Lie algebra at hand. Before going a
bit into this structure theory, we discuss in extenso three fundamental examples whose
understanding prepares the ground for the more general cases.
with
1 1
F =( ), G = ( ), H = ( )
1 −1
and the relations
[H, F ] = 2F, [H, G] = −2G, [F, G] = H.
The exercises at the end of the examples in the previous section should give a feeling
that it is convenient to complexify the Lie algebra, i.e. here we go over to
where
1 ±i −i
(6.2) X± := (1/2)(H ± i(F + G)) = (1/2)( ), Z := −i(F − G) = ( )
±i −1 i
with
[Z, X± ] = ±2X± , [X+ , X− ] = Z.
This normalization will soon prove to be adequate and the operators X± will get a
meaning as ladder operators: We consider a representation π̂ of gc = < X± , Z > on a
C-vector space V =< vi >i∈I where we abbreviate
We are looking for representations (π̂, V ) of gc , which may be integrated, i.e. which are
complexifications of derived representations dπ for a representation π of SL(2, R).
74 6. The Infinitesimal Method
k±2:
Zv = ZX± v = (X± Z ± 2X± )v = X± (k ± 2)v = (k ± 2)v .
Here we used the Lie algebra relation [Z, X± ] = ±2X± , which for the operators acting
on V is transformed into the relation
ZX± − X± Z = ±2X± .
This remark has immediate consequences:
s s
s s
s s
s s
s s
s s s
s s s
s case 3 s s
X+ 6 X−
s ? s
s s s
s s s
The following statement offers the decisive point for the reasoning that we really get
all representations by our procedure. It is a special case of a more general theorem by
Godement on spherical functions. It also is a first example for the famous and important
multiplicity-one statements.
For a proof we refer to Lang [La1] p.24 or the article by van Dijk [vD].
and put
k
vn−2k := X− vn .
76 6. The Infinitesimal Method
As above, we see that we have m = −n with negative n in configuration 1 and that n has
to be positive in configuration 3. For instance, the following are possible configurations.
6 s
4 s
n=2 s m=2 s
s
X+ 6 X−
s ? case 2 case 3
s
m = −2 s n = −2 s
−4 s
case 1 −6 s
s
6.4 Unitarily Integrable Representations of sl(2, R) 77
X± vn = a± ±
n vn±2 with an = (1/2)(s + 1 ± n),
where s ∈ C, such that a±n = 0 for all n ∈ 2Z (resp. 2Z + 1), i.e. s ∈ 2Z + 1 for n ∈ 2Z
and s ∈ 2Z for n ∈ 2Z + 1.
We chose a symmetric procedure for our configuration 4. But as well we could have
j j
chosen any nonzero vn ∈ Vn , and then vn+2j := X+ vn , vn−2j := X− vn for j ∈ N would
constitute an equivalent representation.
Exercise 6.20: Check this.
– π̂s,+ or π̂s,− for the even resp. odd case in configuration 4 with s = 2Z + 1 for π̂s,+
and s = 2Z for π̂s,− , s ∈ C,
– π̂n+ for configuration 3 with lowest weight n, n ∈ N0 ,
– π̂n− for configuration 2 with highest weight −n, n ∈ N0 ,
– σ̂n for configuration 1 with highest weight n and lowest weight −n, n ∈ N0 .
Remark 6.8: If in π̂s,+ and π̂s,− we do not exclude the values of s as above, we get all
representations as sub - resp. quotient representations of these π̂s,± .
Proof: i) Let V = < vn > and V = < vn > be representation spaces for π̂s,+ resp. π̂s ,+
and F : V −→ V an intertwining operator. Because of Zvn = nvn , we have for each n
ZF vn = F Zvn = nF vn .
78 6. The Infinitesimal Method
X± F vn = F X± vn
we deduce
(s + 1 ± n)bn vn±2
= bn±2 (s + 1 ± n)vn±2 ,
i.e.
bn /bn−2 = (s + 1 − n)/(s + 1 − n)
and
bn+2 /bn = (s + 1 + n)/(s + 1 + n)
resp.
bn /bn−2 = (s + n − 1)/(s + n − 1).
Now we see that we have
Therefore we have
< v, i dπ(X)w > = − < v, i dπ(X)∗ w >,
i.e. X̂ = dπ(X) is skew-Hermitian.
Now, we apply this to g = sl(2, R). The Proposition leads to the condition:
Proof: Using the notation from the beginning of this section we have
1 ±i
X± = (1/2)(H ± i(F + G)) = (1/2)( ) =: (X1 ± iX2 )
±i −1
Now, if we ask for unitarity, we have to check whether we find a scalar product <, >
defined on the representation space V such that we get
(6.3) < X̂+ vj , vj+2 > = − < vj , X̂− vj+2 > for all j ∈ J.
X̂− vj+2 = vj
i.e.
ak vj+2 2 = − vj 2 .
An equation like this is only possible for ak < 0, which is true for the ak in configuration
2 but not for configuration 1, where we have seen that all ak are positive.
Similarly Remark 6.6 shows that in configuration 3 the condition (6.3) is no obstacle.
X̂± vj = a± ±
j vj±2 with aj = (1/2)(s + 1 ± j),
At first we treat the even case V =< vj >j∈2Z . Condition (6.4) requires for j = 0
(s̄ + 1) v2 2 = −(s − 1) v0 2 .
80 6. The Infinitesimal Method
As the norm has to be positive, a relation like this is only possible if one has
s2 = σ 2 − τ 2 + 2στ i < 1,
or
τ =0 and σ 2 < 1.
In the odd case V = < vj >j∈2Z+1 we get for j = −1 from (j )
s̄ v1 2 = −s v−1 2 .
σ=0 and τ ∈ R.
In the next chapter we will prove that, in fact, all these representations π̂ can be inte-
grated to unitary representations π of G. They get the following names
– π = πk∓ discrete series representation of highest (lowest) weight ∓k,
– π = πis,± even (odd ) principal series representation,
– π = πs complementary (or supplementary) series representation.
Because of the equivalences stated in Proposition 6.1, we have for the unitary dual of
G = SL(2, R)
Ĝ = {π0 ; πk± , k ∈ N; πis,± , s ∈ R>0 ; πs , 0 < s < 1}.
We will study this more closely in the next chapter but already indicate a kind of visu-
alization.
6.4 Unitarily Integrable Representations of sl(2, R) 81
iR
πis,+
πk− πk+
πs
s s s s s s s s s s s s s s s
−3 −2 −1 1 2 3
πis,−
We can observe a very nice fact, which will become important later and is the starting
point for a general procedure in the construction of representations:
Ω := X+ X− + X− X+ + (1/2)Z 2
acts as a scalar for the three types of representations in Theorem 6.2 and determine these
scalars.
π : SL(2, R) −→ U(n).
Since
a 1 b a −1 1 a2 b
( −1 )( )( −1 ) =( )
a 1 a 1
1
b
all n(b) = ( ) are conjugate. So all π(n(b)) with b > 0 are conjugate. It is a not too
1
difficult topological statement that in a compact group all conjugacy classes are closed.
As U(n) is compact, we can deduce
and therefore π(n(b)) = En for all b > 0. But we can repeat this conclusion for b < 0
1
and as well for n̄(b) = ( ) and get
b 1
−i −i −1
(6.5) H1 = ( ), H2 = ( ), H3 = ( ),
i −i 1
with
[Hj , Hk ] = 2Hl , (j, k, l) = (1, 2, 3), (2, 3, 1), (3, 1, 2).
su(2) and sl(2, R) have the same complexifications. Hence we can express the matrices
Z, X± from (6.2) in the previous section as
Z = iH3 , X± = (1/2)(iH1 ∓ H2 )
and repeat the discussion of the representation space V = < vj >j∈J to get the configu-
rations 1 to 4. Let us look at configuration 1: We take a vector vn ∈ V of highest weight
n ∈ N, put
k
vn−2k = X̂− vn , X̂+ vn−2k =: ak vn−2k+2 ,
get ak = kn − k(k − 1) and m = −n for the lowest weight. But in this case the skew-
hermiticity of the Hj changes things and Remark 6.9 in 6.4 is to be replaced by:
dπ(X± )∗ = dπ(X∓ ).
Hence, the discussion in 6.4 shows that for unitarity in this case the ak have to be
positive as, in fact, they are for n ≥ k ≥ 0. We see that our discussion rediscovers the
representation space V = V (j) from section 4.2 with n = 2j. By the way, we get here for
free another proof that the representation πj from 4.2 is irreducible.
Configurations 2 to 4 are not unitarizable.
with derived representation dψ(κ) = 2πim. Having this in mind (and the discussion of
Example 6.9 in 6.3), we propose the following construction of a representation (π̂, V ),
where we abbreviate again π̂(Y )v =: Y v.
Y0 v0 = µv0 , Y− v0 = 0, µ = 2πm.
We check whether this representation can be unitarized: We look at the necessary con-
dition in Proposition 6.2 in 6.4. The skew-Hermiticity of P, Q, R leads to
Hence, as in the sl-discussion, we get the necessary condition for a scalar product <, >
on V
< Ŷ± v, w > = − < v, Ŷ± w >,
in particular, for v = vj−1 , w = vj , we have
vj 2 = jµ vj−1 2 .
We see that unitarity is possible iff µ = 2πm > 0. A model, i.e. a realization of this
representation π̂ by integration, is given by the Schrödinger representation πm , which we
discussed already in several occasions (see Example 6.9 and Exercise 6.17 in 6.3).
Remark 6.12: This discussion above is already a good part of the way to a proof of the
famous Stone-von Neumann Theorem stating that up to equivalence there is no other
irreducible unitary representation π of Heis(R) with π |C = ψm for m = 0 (for a complete
proof see e.g. [LV] p.19ff).
Remark 6.13: What we did here is essentially the same as that appears in the physics
2
literature under the heading of harmonic oscillator. Our v0 = f0 with f0 (x) = e−2πmx
describes the vacuum, and Y+ and Y− are creation resp. annihilation operators producing
higher resp. lower excitations.
We have to introduce some more notions to get the tools for an appropriate structure
theory. The reader interested mainly in special examples and eager to see the repre-
sentation theory of the other groups mentioned in the introduction may want to skip
this section. Otherwise she or he is strongly recommended to a parallel study of a more
detailed source like [Hu], [HN] or [Ja]. Here we follow [Ki] p.92f, [KT], and [Kn1], which
don’t give proofs neither (for these see [Kn] p.113ff, and/or [Kn2] Ch.I - VI).
The reports [Kn1] and [KT] take as their central objects real or complex linear connected
reductive groups. The definition of a reductive group is not the same with all authors
but the following definition ([Kn] p.3, [KT] p.25) is very comprehensive and practical for
our purposes.
Definition 6.7: A linear connected reductive group is a closed connected group of real
or complex matrices that is stable under conjugate transpose, i.e. under the Cartan
involution
Θ : Mn (C) −→ Mn (C), X −→ (t X̄)−1 .
A linear connected semisimple group is a linear connected reductive group with finite
center.
More standard is the following definition: A group is called simple iff it is non-trivial,
and has no normal subgroups other than {e} and G itself.
For the moment we follow another approach (as in [Ki] p.17). Non-commutative groups
can be classified according to the degree of their non-commutativity: Given two subsets
A and B of the group G, let [A, B] denote the set of all elements of the form aba−1 b−1
as a runs through A and b runs through B. We define two sequences of subgroups:
Let G0 := G and for n ∈ N let Gn be the subgroup generated by the set [Gn−1 , Gn−1 ].
Gn is called the n-th derived group of G.
86 6. The Infinitesimal Method
Let G0 := G and for n ∈ N assume Gn to be the subgroup generated by the set [G, Gn−1 ].
We obviously obtain the following inclusions
G = G0 ⊃ G 1 ⊃ · · · ⊃ G n ⊃ . . . ,
G = G0 ⊃ G 1 ⊃ · · · ⊃ Gn ⊃ . . . .
For a commutative group these sequences are trivial, we have Gn = Gn = {e} for all
n ∈ N.
Now we pass to the corresponding concepts for Lie algebras (following [Ki] p.89).
In every Lie algebra g we can define two sequences of subspaces, where here the expression
[a, b] denotes the linear hull of all [X, Y ], X ∈ a, Y ∈ b:
g1 := [g, g], g2 := [g, g1 ], . . . , gn+1 := [g, gn ],
g1 := [g, g], g2 := [g1 , g1 ], . . . , gn+1 := [gn , gn ].
It is clear that one has the following inclusions
gn ⊃ gn+1 , gn ⊃ gn+1 , gn ⊃ gn , n ∈ N.
Exercise 6.27: Prove that all gn and gn are ideals in g and that gn /gn+1 and gn /gn+1
are commutative.
For dim g < ∞ the sequences (gn ) and (gn ) must stabilize, beginning with a certain n
we have gn = gn+1 = · · · = g∞ and gn = gn+1 = · · · = g∞ .
Example 6.11: heis(R) is nilpotent, and hence solvable as we have more generally
Now we pass to the other side of Lie algebra structure theory. As in [Kn1] p.1 ff we
restrict ourselves to the treatment of finite dimensional real or complex algebras g.
In this case one has [g, g] = g, which shows that we are as far from solvability as possible.
There are other (equivalent) definitions, for instance g is said to be semisimple iff one
has rad g = 0. In this definition the radical rad g is the sum of all solvable ideals of g.
Theorem 6.5: g is semisimple iff g is the sum of simple ideals. In this case there are
no other simple ideals, the direct sum decomposition is unique up to order of summands
and every ideal is the sum of the simple ideals. Also in this case [g, g] = g.
Definition 6.13: A reductive Lie algebra is a Lie algebra that is the direct sum of two
ideals, one equal to a semisimple algebra and the other to an abelian Lie algebra.
Theorem 6.6: If g is a real Lie algebra of real or complex (even quaternion) matrices
closed under conjugate transposition, then g is reductive. If moreover the center of g is
trivial, i.e. Zg := {X ∈ g; [X, Y ] = 0 for all Y ∈ g} = 0, then g is semisimple.
As to be expected, to some extent the notions just defined for groups and algebras fit
together and one has statements like the following:
Proposition 6.4 ([KT] p.29): If G is a linear connected semisimple group, then g = Lie G
is semisimple. More generally, if G is linear connected reductive, then g is reductive with
g = Zg ⊕ [g, g] as a direct sum of ideals. Here Zg denotes the center of g, and the com-
mutator ideal is semisimple.
The main tool for the structure theory of semisimple Lie algebras is the following bilinear
form that was first defined by Killing and then extensively used by E. Cartan.
Definition 6.14: The Killing form B is the symmetric bilinear form on g defined by
The starting point for structure theory of semisimple Lie algebras is Cartan’s Criterium
for Semisimplicity.
Theorem 6.7: g is semisimple if and only if the Killing form is nondegenerate, that is
B(X, Y ) = 0 for all Y ∈ g implies X = 0.
The trace form is invariant in the same sense as the Killing form. Both forms are related
(see e.g. [Fog] IV.4):
A variant of the trace form already appeared in our definition of topology in matrix
spaces. We introduced for X, Y ∈ Mn (C)
As infinitesimal object corresponding to the Cartan involution for our matrix groups G
Hence we have
< X, Y > := −Re B0 (X, θY )
as a scalar product on g as a real vector space. This will become important later on.
6.6 Some Structure Theory 89
Definition 6.15: Given a complex Lie algebra g, a real algebra g0 is called a real form
of g iff (g0 )c := g0 ⊗ C = g.
Example 6.13: g = sl(2, C) has just two real forms g0 = sl(2, R) and su(2).
More about this is to be found in [Kn1] p.15.
The main ingredients for a classification scheme are the root and Dynkin diagrams. We
give a sketch starting by the case of complex semisimple algebras where the theory is
most accessible, and then go to compact real and finally briefly to noncompact real al-
gebras. The central tool to define roots and their diagrams is a certain distinguished
abelian subalgebra.
There are other equivalent ways to characterize a Cartan subalgebra h, for instance (see
[Kn1] p.2): h is a nilpotent subalgebra whose normalizer Ng (h) satisfies
Ng (h) := {X ∈ g, [X, H] ∈ h for all H ∈ h} = h.
Each semisimple complex algebra has a Cartan subalgebra. Any two are conjugate via
Int g ([Kn1] p.24), where Int g is a certain analytic subgroup of GL(g) with Lie algebra
ad g. We refer to [Kn2] p.69/70 for the notion of an analytic subgroup, which comes up
quite naturally when one analyzes the relation between Lie algebras and Lie groups.
90 6. The Infinitesimal Method
The theory of Cartan subalgebras for the complex semisimple case extends to a complex
reductive Lie algebra g by just saying that the center of g is to be adjoined to a Cartan
subalgebra of the semisimple part of g.
as follows: For all H ∈ h the maps ad H are diagonizable, and, as the elements of h
all commute, they are simultaneously diagonizable (by the finite-dimensional spectral
theorem from Linear Algebra). Let V1 , . . . , V be the eigenspaces in g for the different
systems of eigentupels. If h := < H1 , . . . , Hr > and ad Hi acts as λij · id on Vj , define
a linear functional λj on h by λj (Hi ) = λij . If H := Σci Hi , then ad H acts on Vj by
multiplication with
for the eigenspace corresponding to λ ∈ h∗ . The nonzero such λ are called roots and the
corresponding gλ a root space. The (finite) set of all roots is denoted by ∆.
The following are some elementary properties of root space decompositions (for proofs
see for instance [Kn2] II.4):
Some deeper properties of root space decompositions are assembled in the following state-
ment (see again [Kn2] II.4).
6.6 Some Structure Theory 91
a) dim gα = 1 for α ∈ ∆.
b) nα ∈ ∆ for α ∈ ∆ and any integer n ≥ 2.
c) [gα , gβ ] = gα+β for α + β = 0.
d) The real subspace h0 of h on which all roots are real is a real form of h and B |h0 ×h0
is an inner product.
We transfer B| h0 ×h0 to the real span h∗0 of the roots obtaining a scalar product <, >
and a norm . . It is not too difficult to see that for α ∈ ∆ there is an orthogonal
transformation sα given on h∗0 by
2 < ϕ, α >
sα (ϕ) := ϕ − α for ϕ ∈ h∗0 .
α 2
The analysis of the root space shows that it has very nice geometrical and combinatorical
properties. The abstraction of these is the following:
Definition 6.17: An abstract root system is a finite set ∆ of nonzero elements in a real
inner product space V such that
a) ∆ spans V ,
b) all reflections sα for α ∈ ∆ carry ∆ to itself,
c) 2 < β, α > / α 2 ∈ Z for all α, β ∈ ∆.
The abstract root system is called reduced iff α ∈ ∆ implies 2α ∈ ∆.
And it is called reducible iff ∆ = ∆ ∪ ∆ with ∆ ⊥ ∆ , otherwise it is irreducible.
The root system of a complex semisimple Lie algebra g with respect to a Cartan subal-
gebra h forms a reduced abstract root system in h∗0 . And a semisimple Lie algebra g is
simple if the corresponding root system is irreducible.
Definition 6.18: The dimension of the underlying space V of an abstract root system
∆ is called its rank, and if ∆ is the root system of a semisimple Lie algebra g, we also
refer to r = dim h as the rank of g.
6
J
]
I 6
@ Q
kJ
]J
3
-
6
J
- @ - Q -
J @ Q
J^ ?@
R
JJ
+
^Q
s
?
?
92 6. The Infinitesimal Method
A root α is called simple if α > 0 and α does not decompose as α = β1 + β2 with β1 and
β2 positive roots. And a root α is called reduced if (1/2)α is not a root.
Relative to a given simple system α1 , . . . , αr , the Cartan matrix C is the r × r-matrix
with entries
cij = 2 < αi , αj > / αi 2 .
The main facts are that a Cartan matrix and equivalently the CDW-diagram determine
the Lie algebra uniquely (up to isomorphism). In particular this correspondence does
not depend on the choice of a basis of the root system.
Example 6.15: For g = sl(r + 1, C), r ≥ 1 and its standard Cartan subalgebra
h := {D(d1 , . . . , dr ); d1 , . . . , dr ∈ C, Σdi = 0}
Ar r r r ... r r r r ≥ 1.
α1 α2 αr
6.6 Some Structure Theory 93
The diagrams for the other classical algebras are the following
Br r r r ... r r -
-r r ≥ 2,
α1 α2 αr−1 αr
Cr r r r ... r r
r r ≥ 3,
α1 α2 αr−1 αr
r αr−1
Dr r r r ... r r r r ≥ 4.
α1 α2 αr−2 αr
The remaining exceptional graphs of indecomposable reduced root systems are the fol-
lowing
E6 r r r r r
E7 r r r r r r
E8 r r r r r r r
r r -
-r r r -r
-
F4 G2 -
Theorem 6.10: Each complex semisimple Lie algebra has a compact real form.
From [Kn2] Proposition 4.30, Theorem 4.34 and 4.36 we take over:
Theorem 6.12: If G is compact connected and T a maximal torus, then each element
of G is conjugate to a member of T .
Example 6.16:
1. For G = SU(n) one has as a maximal torus, its Lie algebra and its complexified Lie
algebra
T = {D(eiϑ1 , . . . , eiϑn ); Σϑj = 0},
t0 = {D(iϑ1 , . . . , iϑn ); Σϑj = 0},
t = {D(z1 , . . . , zn ); Σzj = 0}.
2. For G = SO(2n) and SO(2n + 1) one has as a maximal torus T the diagonal block
matrices with 2 × 2-blocks r(ϑj ), j = 1, . . . , n, in the diagonal, resp. additionally 1 in the
second case.
We go back to the general case and use the notation we have introduced. In this setting,
we can form a root-space decomposition
g=t⊕ gα .
α∈∆
Another central concept is the one of the Weyl group. We keep the notation used above:
G is compact connected, g0 = Lie G, T a maximal torus, t0 = Lie T , t = t0 ⊗ C, ∆(g, t) is
the set of roots, and B is the negative of a G invariant scalar product on g0 . We define
tR = it0 . As roots are real on tR , they are in t∗R . The form B, when extended to be
complex bilinear, is positive definite on tR , yielding a scalar product <, > on t∗R . Now, the
Weyl group W = W (∆(g, t)) is in this context defined as the group generated by the root
reflections sα , α ∈ ∆(g, t), given (as already fixed above) on t∗R by sα (λ) = λ − 2<λ,α>
|α|2 α.
This is a finite group, which also can be characterized like this:
One defines W (G, T ) as the quotient of the normalizer by the centralizer
W (G, T ) := NG (T )/ZG (T ).
By Corollary 4.52 in [Kn2] p.260 one has ZG (T ) = T and hence also the formula
W (G, T ) = NG (T )/T.
Theorem 6.13: The group W (G, T ), when considered as acting on t∗R , coincides with
W (∆(g, t)).
Bθ (X, Y ) := −B(X, θY )
g0 = k0 ⊕ p0 , k0 := {X ∈ g0 ; θX = X}, p0 := {X ∈ g0 ; θX = −X}
The Cartan decomposition of Lie algebras has a global counterpart (see for instance
[Kn1] Theorem 4.3). Its most rudimentary form is the following:
96 6. The Infinitesimal Method
Proposition 6.6: For every A ∈ GL(n, C) resp. GL(n, R), there are S1 , S2 ∈ O(n)
resp. U(n) and a diagonal matrix D with real positive elements in the diagonal such that
A = S1 DS2 . This decomposition is not unique.
Proof as Exercise 6.29 (use the fact that to each positive definite matrix B there is a
unique positive definite matrix P with B = P 2 and/or see [He] p.73).
Restricted Roots
We want to find a way to describe Cartan subalgebras in our general situation: Let
G be semisimple linear group with Lie G = g0 , g = g0 ⊗ C, θ a Cartan involution of
g0 , and g0 = k0 ⊕ p0 the corresponding Cartan decomposition. Let B be the Killing
form (or more generally any nondegenerate symmetric invariant bilinear form on g0 with
B(θX, θY ) = B(X, Y ) such that Bθ (X, Y ) = −B(X, θY ) is positive definite).
Definition 6.20: Let a0 be a maximal abelian subspace of p0 . Restricted roots are the
nonzero λ ∈ a∗0 such that
Let Σ be the set of restricted roots and m0 := Zk0 . We fix a basis of a0 and an associated
lexicographic ordering in a∗0 and define Σ+ as the set of positive resticted roots. Then
n0 = (g0 )λ
λ∈Σ+
Theorem 6.14: The semisimple Lie algebra g0 is a vector space direct sum
g0 = k0 ⊕ a0 ⊕ n0 .
For a proof see for instance [Kn2] Proposition 6.43. There is also a global version ([Kn2]
Theorem 6.46) stating (roughly) that one has a diffeomorphism of the corresponding
groups K × A × N −→ G given by (k, a, n) −→ kan. We already know all this in our
standard example G = SL(2, R): If we take θX = −t X, we have
g0 = sl(2, R) = k0 ⊕ a0 ⊕ n0 ,
with
1 1 1
k0 =< ( ) >, a0 =< ( ) >, n0 =< ( )>
−1 −1
and as in Example 6.6 in 6.2
As to be expected, roots and restricted roots are related to each other. If t0 is a maximal
abelian subspace of m0 := Zk0 (a0 ), then h0 := a0 ⊕ t0 is a Cartan subalgebra of g0 in
the sense we defined at the beginning ([Kn2] Proposition 6.47). Roots are real valued on
a0 and imaginary valued on t0 . The nonzero restrictions to a0 of the roots turn out to
be restricted roots. Roots and restricted roots can be ordered compatibly by taking a0
before it0 . Cartan subalgebras in this setting are not always unique up to conjugacy.
1 1
Exercise 6.30: Show that h0 :=< ( ) > and h0 :=< ( ) > are Cartan
1 −1
subagebras of g0 = sl(2, R) and determine the corresponding Iwasawa decomposition.
Every Cartan subalgebra of g0 is conjugate (via Intg0 ) to this h0 or the h0 above.
is nonzero and called the weight space corresponding to λ. One introduces a lexicographic
ordering among the weights and hence has the notion of highest and lowest weights. The
set of weights belonging to a representation π̂ is denoted by Γ(π̂). For dim h = 1 we
have h h∗ C and so the weights are simply the complex numbers we met in our
examples in 6.4 and 6.5, for instance for g = sl(2, C) we found representations σ̂N with
dim VN = N + 1 and weights −N, −N + 2, . . . , N − 2, N . In continuation of this, we
treat a more general example.
Example 6.17: Let G = SU(n) and, hence, g = su(n)c = sl(n, C) and the Cartan
subalgebra the diagonal subalgebra h :=< H; Tr H = 0 >. We choose as generators of h
the matrices Hj := Ejj − Ej+1,j+1 and (slightly misusing) we write also Hj = ej − ej+1
where ej denote the canonical basis vectors of Cn−1 h. We have three natural types
of representations :
At first, let V be the space of homogeneous polynomials P of degree N in z1 , . . . , zn and
their conjugates and take the action of g ∈ SL(n, C) given by
n
(dπ(H)P )(z, z̄) = (−iϑj zj )∂zj P (z, z̄) + (−iϑj z̄j )∂z̄j P (z, z̄).
j=1 j=1
If P is a monomial
n
P (z, z̄) = z1k1 · · · · · znkn · z̄1l1 · · · · · z̄nln with (kj + lj ) = N,
j=1
then we get
n
dπ(H)P = (lj − kj )(iϑj )P.
j=1
Exercise 6.31: Describe the weights for this representation with respect to the lexico-
graphic ordering for the basis elements of h0 =< iH1 , . . . , iHn−1 >R . Do this also for the
two other “natural” representations, namely on the subspaces V1 of holomorphic and V2
of antiholomorphic polynomials in V (i.e. polyomials only in z1 , . . . , zn resp. z̄1 , . . . , z̄n ).
We will come back to this for the case n = 3 in the next section.
In our examples g0 = sl(2, R) and g0 = su(2) in 6.4 resp. 6.5 we saw that the finite-
dimensionality of the representation space V was equivalent to an integrality condition
for the weight. Now we want to look at the general case where the weights are l-tupels
if one fixes a basis of a Cartan subalgebra h resp. its dual h∗ and the real form h∗0 .
Then as a generalization of what we saw in 6.4 and 6.5, we have the elementary properties
of the weights for a finite-dimensional representation π̂ on a vector space V :
Proposition 6.7: a) π̂(h) acts diagonally on V , so that every generalized weight vector
is a weight vector and V is the direct sum of all weight spaces.
b) Every weight is real valued on h0 and algebraically integral.
c) Roots and weights are related by π̂(gα )Vλ ⊂ Vλ+α .
l
d) Every weight of π̂λ is of the form λ − i=1 ni αi with ni ∈ N0 and αi ∈ Π.
e) Each weight space Vµ for π̂λ has dimVwµ = dim Vµ for all w in the Weyl group W (∆),
and each weight µ has | λ |≥| µ | with equality only if µ is in the orbit W (∆)λ.
We already introduced in 3.2 the concept of complete reducibility. Here we can state the
following fact ([Kn2] Th. 5.29).
The proofs of these two theorems use three tools, which are useful also in other contexts:
– Universal enveloping algebras,
– Casimir elements,
– Verma modules.
Again following [Kn1] p.10f, we briefly present these as they will help us to a better
understanding in our later chapters.
This is a general and far reaching concept applicable for any complex Lie algebra g:
T (g) := C ⊕ g ⊕ (g ⊗ g) ⊕ . . . .
X ⊗ Y − Y ⊗ X − [X, Y ], X, Y ∈ T 1 (g).
Then the universal enveloping algebra is the associative algebra with identity given by
U (g) := T (g)/a.
This formal definition has the practical consequence that U (g) consists of sums of mono-
mials usually written (slightly misusing) as a(j) X1j1 . . . Xnjn , a(j) ∈ C, X1 , . . . , Xn ∈ g
with the usual addition and a multiplication coming up from the Lie algebra relations
by Xi Xj = Xj Xi + [Xi , Xj ]. More carefully and formally correct, one has the following:
Let ι : g −→ U (g) be the composition of natural maps
so that
ι([X, Y ]) = ι(X)ι(Y ) − ι(Y )ι(X).
100 6. The Infinitesimal Method
Theorem 6.17: Let {Xi }i∈I be a basis of g, and suppose that a simple ordering has
been imposed on the index set I. Then the set of all monomials
with i1 < · · · < in and with all jk ≥ 0, is a basis of U (g). In particular the canonical
map ι : g −→ U (g) is one-to-one.
The name universal enveloping algebra stems from the following universal mapping prop-
erty.
Theorem 6.18: Whenever A is a complex associative algebra with identity and we have
a linear mapping ϕ : g −→ A such that
then there exists a unique algebra homomorphism ϕ̃ : U (g) −→ A such that ϕ̃(1) = 1
and ϕ = ϕ̃ ◦ ι.
The map ϕ̃ from the theorem may be thought of as an extension of ϕ from g to all of
U (g). This leads to the following useful statement.
We now come back to the case that g is semisimple and to the notation introduced above.
We enumerate the positive roots as α1 , . . . , αm and we let H1 , . . . , Hl be a basis of h.
Then for the construction of highest weight representations it is appropriate to use the
ordered basis
p1 pm
E−α 1
. . . E−α m
H1k1 . . . Hlkl Eαq11 . . . Eαqm
m
is a basis of U (g). If one applies members of this basis to a nonzero highest weight vector
v0 of V, one gets control of a general member of U (g)v0 : Eαq11 . . . Eαqmm
will act as 0 if
q1 + · · · + qm > 0, and H1 . . . Hl will act as a scalar. Thus one has only to sort out
k1 kl
p1 pm
the effect of E−α 1
. . . E−α m
and most of the conclusions of the Theorem of the Highest
Weight follow readily. If one looks at our examples in 6.4 and 6.5, one can get an idea
how this works even in the case of representations, which are not finite-dimensional, and
how the integrality of the weight leads to finite-dimensionality.
6.6 Some Structure Theory 101
For a complex semisimple Lie algebra g with Killing form B, the Casimir element Ω is
the element
of U (g), where (Xi ) is a basis of g and (X̃i ) is the dual basis relative to B. One can show
that Ω0 is defined independently of the basis (Xi ) and is an element of the center Z(g)
of U (g).
Exercise 6.32: Check this for the case g = sl(2, C) and determine the Casimir element.
Theorem 6.20: Let Ω0 be the Casimir element, (Hi )i=1,.,l an orthogonal basis of h0
relative to B, and choose root vectors Eα so that B(Eα , E−α ) = 1 for all roots α. Then
l
a) Ω0 = i=1 Hi2 + α∈∆ Eα E−α .
b) Ω0 operates by the scalar |λ|2 + 2 < λ, δ > = |λ + δ|2 − |δ|2 in an irreducible finite-
dimensional representation of g of highest weight λ, where δ is half the sum of the positive
roots.
c) The scalar by which Ω0 operates in an irreducible finite-dimensional representation of
g is nonzero if the representation is not trivial.
Remark 6.16: The center of U (g) is important also in the context of the determination
of infinite-dimensional representations. We observed in Exercise 6.23 in 6.4 that Ω, a
multiple of the Casimir element Ω0 , acts as a scalar for the unitary irreducible represen-
tations of sl(2, R). For more general information we recommend [Kn] p.214 where as a
special case of Corollary 8.14 one finds the statement: If π is unitary, then each mem-
ber of the center Z(gc ) of U (gc ) acts as a scalar operator on the K−finite vectors of π.
We will come back to this in 7.2 while explicitly constructing representations of SL(2, R).
where δ is again half the sum of the positive roots and this term is introduced to simplify
calculations with the Weyl group.
102 6. The Infinitesimal Method
Verma modules are essential for the construction of representations. They have the fol-
lowing elementary properties:
a) V (µ) = 0.
b) V (µ) is a universal highest weight module for highest weight modules of U (g) with
highest weight µ − δ.
c) Each weight space of V (µ) is finite-dimensional.
d) V (µ) has a unique irreducible quotient L(µ).
In our treatment of finite and compact groups we already saw the effectiveness of the
theory of characters of representations. As for the moment we look at finite-dimensional
representations we can use characters here too. To allow for more generalization, we
treat them for now as formal exponential sums (again following [Kn1] p.12/3):
Let again g be a semisimple Lie algebra, h a Cartan subalgebra, ∆ a set of roots provided
with a lexicographic ordering, α1 , . . . , αl the simple roots, and W (∆) the Weyl group.
∗
We regard the set Zh of functions f from h∗ to Z as an abelian group under pointwise
addition. We write
f= f (λ)eλ .
λ∈h∗
∗
The support of f is defined to be the set of λ ∈ h∗ for which f (λ) = 0. Within Zh ,
let Z[h∗ ] be the subgroup of all f of finite support. The subgroup Z[h∗ ] has a natural
commutative ring structure, which is determined by eλ eµ = eλ+µ . Moreover, we introduce
a larger ring Z < h∗ > with
∗
Z[h∗ ] ⊆ Z < h∗ > ⊆ Zh
∗
consisting of all f ∈ Zh whose support is contained in the union of finitely many sets
νi − Q+ , νi ∈ h∗ and
l
Q+ := { ni αi ; ni ∈ N0 }.
i=1
( cλ eλ )( cµ eµ ) := ( cλ cµ )eν .
λ∈h∗ µ∈h∗ λ∈h∗ λ+µ=ν
∗
as an element of Zh . This definition is meaningful if V is finite-dimensional or if V is a
Verma module.
6.6 Some Structure Theory 103
d := eδ Πα∈∆+ (1 − e−α ).
Now we leave the treatment of the complex semisimple case and go over to the follow-
ing situation: G is compact connected, g0 := Lie G, g the complexification of g0 , T a
maximal torus, t0 := Lie T , ∆(g, t) the set of roots, and B the negative of a G invariant
inner product on g0 , and tR := it0 . As we know, roots are real on tR , hence are in t∗R .
The form B, when extended to be complex bilinear, is positive definite on tR , yielding an
inner product <, > on t∗R . W (∆(g, t)) is the Weyl group generated by the root reflections
sα for α ∈ ∆(g, t). Besides the notion of algebraic integrality already exploited in the
complex semisimple case above, we have here still another notion of integrality:
We say that λ ∈ t∗ is analytically integral if the following equivalent conditions hold:
1) Whenever H ∈ t0 satisfies exp H = 1, then λ(H) is in 2πiZ.
2) There is a multiplicative character ψλ of T with ψλ (exp H) = eλ(H) for all H ∈ t0 .
In [Kn1] p.18 one finds a list of properties of these notions. We cite part of it:
a) Weights of finite-dimensional representations of G are analytically integral. In partic-
ular every root is analytically integral.
b) Analytically integral implies algebraically integral.
c) If G is simply connected and semisimple, then algebraically integral implies analyti-
cally integral.
For instance, the half sum δ of positve roots is algebraically integral but not analytically
if G = SO(3).
In our situation the Theorem 6.15 (Theorem of the Highest Weight) comes in the following
form:
Theorem 6.22: Let G be a compact connected Lie group with complexified Lie alge-
bra g, let T be a maximal torus with complexified Lie algebra t, and let ∆+ (g, t) be a
positive system for the roots. Apart from equivalence the irreducible finite-dimensional
representations π of G stand in one-one correspondence with the dominant analytically
integral linear functionals λ on t, the correspondence being that λ is the highest weight
of π.
104 6. The Infinitesimal Method
In the next section we shall illustrate this by treating an example, which was a milestone
in elementary particle physics.
This basis is also a basis for the complexification su(3)c = sl(3, C). One may be tempted
to use the elementary matrices Eij with entries (Eij )kl = δik δjl and take as a basis
One has the commutation relations [Eij , Ekl ] = δjk Eil − δil Ekj . For a diagonal matrix
H := D(h1 , h2 , h3 ) we get
and hence
[H1 , E12 ] = 2E12 , [H2 , E12 ] = −E12 ,
[H1 , E13 ] = E13 , [H2 , E13 ] = E13 ,
[H1 , E23 ] = −E23 , [H2 , E23 ] = 2E23 ,
and
[E12 , E21 ] = H1 , [E23 , E32 ] = H2 , [E13 , E31 ] = E11 − E33 =: H3 .
6.7 The Example su(3) 105
gα1 := < E12 >, gα2 := < E13 >, gα3 := < E23 >
are root spaces, i.e. nontrivial eigenspaces for the roots α1 , α2 , α3 ∈ h∗ given by
(In the physics literature these tuples α̃j themselves sometimes are called roots.)
We choose these roots αj as positive roots and then get in this coordinization a slightly
unsymmetric picture:
α̃3 r2
AK
A
A r α̃2
A
A 1 2
r r AH r r
-2 -1 HH
HH
r -1 j α̃1
H
Hence we better use Gell-Mann’s matrices and change to Xj := −iAj . Then we get
H1 = X3 , H2 := X8 . Moreover, for X± := (1/2)(X1 ± iX2 ), Y± := (1/2)(X6 ± iX7 ), and
Z± := (1/2)(X4 ±iX5 ), we have X+ = E12 , X− = E21 etc. and the commutation relations
Now the Cartan subalgebra is h =< H1 , H2 > and the root spaces are
α1 and α2 are simple roots, we have α1 + α2 = α3 and we get a picture with hexagonal
symmetry:
r2
ᾱ2 ᾱ3
AK
A
A r
A
A ᾱ1
r r A r -r
-2 -1 1 2
r -1
106 6. The Infinitesimal Method
In the literature (see for instance [Co] p.518) we find still another normalization, which
is motivated like this: The Killing form B for su(3) is given by diagonal matrices
One uses B to introduce a nondegenerate symmetric bilinear form <, > on h∗ as follows:
We define a map
h∗ ⊃ ∆
α −→ Hα ∈ h by B(Hα , H) = α(H) for all H ∈ h,
and then
< α, β >:= B(Hα , Hβ ) for all α, β ∈ ∆.
This leads to inner products on hR = < Hα1 , . . . , Hαl >R (α1 , . . . , αl simple positive
roots) and its dual space. For su(3) we have l = 2 and with α1 , α2 from above
It is convenient to introduce an orthonormal basis Ȟ1 , . . . , Ȟl of h, i.e. with B(Ȟp , Ȟq ) =
δpq . Then we have the simple rule
and √ √ √ √
α̌1 = ( 3/6)ᾱ1 = (1/ 3, 0), α̌2 = ( 3/6)ᾱ2 = (−1/(2 3), 1/2).
We see that, for the Cartan matrix C = (Cpq ) = (2 < αp , αq > / < αp , αp >) from 6.6.2,
we have in this case
2 −1
C=( ).
−1 2
The Weyl group W (see the end of 6.6.3) is generated by the reflections
at the planes in h∗R Rl orthogonal to the α̌p with l = 2 and p, q = 1, 2 in our case. For
g0 = su(3) we have 6 elements in W .
Now we proceed to the discussion of the representations (π̂, V ) of su(3). The general
theory tells us that we can assume that π̂(H) is a diagonal matrix for each H ∈ h. The
diagonal elements π̂(H)jj =: λj (H) fix the weights λj , i.e. linear functionals on h. We
write
λ = (λ(Ȟ1 ), . . . , λ(Ȟl ))
for an ON-basis of h. By the Theorem of the Highest Weight (Theorem 6.15 and Theorem
6.22), an irreducible representation is determined by its highest weight Λ and this highest
6.7 The Example su(3) 107
weight is simple, i.e. has multiplicity one. Moreover the general theory says ([Co] p.568)
that these highest weights are linear combinations with non negative integer coefficients
of fundamental weights Λ1 , . . . , Λl defined by
l
Λj (H) = (C −1 )kj αk (H)
k=1
< Λ + δ, αj >
d = Πlj=1 .
< δ, αj >
We write π̂(n1 , n2 ) for the representation with highest weight Λ = n1 Λ1 + n2 Λ2 and get
by Weyl’s formula for its dimension
In the physics literature one often denotes the representation by its dimension. Then one
has
a) π̌(0, 0) = {1}, the trivial representation, which has only one weight, namely the high-
est weight Λ = 0,
b) π̌(1, 0) = {3} with highest weight Λ = Λ1 and dimension 3,
c) π̌(0, 1) = {3∗ } with highest weight Λ = Λ2 and dimension 3,
d) π̌(2, 0) = {6} with highest weight Λ = 2Λ1 and dimension 6,
e) π̌(0, 2) = {6∗ } with highest weight Λ = 2Λ2 and dimension 6,
f) π̌(1, 1) = {8} with highest weight Λ = Λ1 + Λ2 and dimension 8,
g) π̌(3, 0) = {10} with highest weight Λ = 3Λ1 and dimension 10,
and so on. The weights appearing in the representations can be determined by the fact
that they are of the form
λ = Λ − m1 α1 − m2 α2 , m1 , m2 ∈ N0 ,
and that the Weyl group transforms weights into weights (preserving the multiplicity).
We sketch some of the weight diagrams and later discuss an elementary method to get
these diagrams, which generalizes our discussion in 6.4 and 6.5.
108 6. The Infinitesimal Method
t
sα2 sα1 Λ = (1/6)(0, 2) = −(1/3)α1 − (2/3)α2
∗
π̌(0, 1) = {3 }
λ(Ȟ2 )
6
Λ = (1/6)(0, 2) = (1/3)α1 + (2/3)α2
t
λ(Ȟ1 )
-
√
sα2 Λ = (1/6)( 3, −1) = −(1/3α1 − (1/3)α2
t t
√
sα1 sα2 Λ = (1/6)(− 3, −1) = − (2/3)α1 − (1/3)α2
-
sα1 sα2 (Λ − α1 ) = −(2/3)α1 − (1/3)α2 λ(Ȟ1 )
t t
√
sα2 (Λ − α1 ) = (1/6)( 3, −1) = (1/3)α1 − (1/3)α2
λ(Ȟ2 )
6
π̌(1, 1) = {8}
u u
√
sα1 Λ = α2 Λ = (1/6)( 3, 3) = α1 + α2
λ(Ȟ1 )
u j
u u -
sα1 sα2 Λ = −α1 0 = (0, 0) sα2 Λ = α1
√
sα1 sα2 sα1 Λ = −α1 − α2 sα2 sα1 Λ = (1/6)( 3, −3) = −α2
u u
π̌(3, 0) = {10} 6
λ(Ȟ2 )
u u u u
√
−α1 + α2 α2 α1 + α2 Λ = (1/6)(3 3, 3) = 2α1 + α2
λ(Ȟ1 )
u u u -
−α1 0 = (0, 0) α1
u u
−α1 − α2 −α2
u −α1 − 2α2
In {8} we find a first example of a weight, which has not multiplicity one, namely the
weight (0, 0).
110 6. The Infinitesimal Method
Similar to our discussion for G = SU(2) and g = su(2) in 4.2, one has here the problem
of the explicit decomposition of representations into irreducible components. This goes
under the heading of Clebsch Gordon series and can be found for instance in [Co] p.611ff.
As examples we cite
{3} ⊗ {3∗ } {8} ⊕ {1} and {3} ⊗ {3} ⊗ {3} {10} ⊕ 2{8} ⊕ {1}.
The proofs of these formulae need some skill or at least patience. But it is very tempt-
ing to imagine the bigger weight diagrams in our examples as composition of the tri-
angles {3} and {3∗ }. This leads to an important interpretation in the theory of ele-
mentary particles: We already remarked that the discussion of the irreducible repre-
sentations of SU(2) resp. SO(3) is useful in the description of particles subjected to
SO(3)-symmetry (in particular an electron in an hydrogen atom), namely a representa-
tion πj , j = 0, 1/2, 1, 3/2, . . . describes a multiplet of states of the particle with angular
momentum resp. spin 2j + 1 and the members of the multiplet are distinguished by a
magnetic quantum number m ∈ Z with m = −(2j + 1), −(2j − 1), . . . , (2j + 1). As one
tried to describe and classify the heavier elementary particles, one observed patterns,
which could be related to the representation theory of certain compact groups, in partic-
ular G = SU(3): In Gell-Mann’s eightfold way we take SU(3) as an internal symmetry
group to get the ingredients of an atom, the proton p and the neutron n as members of
a multiplet of eight states, which correspond to our representation {8}. To be little bit
more precise, one looks at a set of hadrons, i.e. particles characterized by an experimen-
tally fixed behaviour (having strong interaction), and describes these by certain quantum
numbers, namely electric charge Q, hypercharge Y, baryon number B, strangeness S,
isospin I with third component I3 . The baryon number is B = 1 for baryons like proton
or the neutron, -1 for antibaryons, and zero for other particles like the π-mesons. In this
context, one has the following relations
Y = B + S, Q = I3 + (1/2)Y
quark B I I3 Y S Q
u 1/3 1/2 −1/2 1/3 0 2/3
d 1/3 1/2 1/2 1/3 0 −1/3
s 1/3 0 0 −2/3 −1 −1/3
and for more information refer to p. 50 f of [FS]. In this picture the weight diagrams look
like this:
6.7 The Example su(3) 111
u d
u 1/3 r u 1/3 r
r r - r r - I3
−1/2 1/2 I3 −1/2 1/2
−1/3 r u −1/3 r u
ū d¯
−2/3 u s
Up to now the experimental evidence for the existence of the quarks (as far as we know) is
still indirect: they are thought of as building blocks of the more observable particles like
the proton and the neutron. The following diagram shows the quantum numbers of the
baryon octet {8}. By a simple addition process (comparing the (I3 , Y )-coordinates) one
can guess the quark content of the neutron as n = (udd) and of the proton as p = (uud).
One can do as well for the other particles showing up in the diagram.
π̌(1, 1) = {8}
Y
6
u r 1 u
n p
−1/2 ∆0 1/2 I3
u r j
u r u -
Σ− Σ0 Σ+
Ξ− Ξ+
u r −1 u
112 6. The Infinitesimal Method
π̌(3, 0) = {10} Y
6
u u r 1 u u
∆− ∆0 ∆+ ∆++
u r −1 u
Ξ− Ξ0
Ω− u −2
6.7 The Example su(3) 113
To finish this section, let us look what happens when we apply to our example the
procedure, which was successful in our constructions of the representations of su(2) and
sl(2, R) in 6.4 and 6.5. We stay in the unnormalized picture with
r s t
V = vrst C, vrst := X+ Y+ Z+ v.
From the PBW-Theorem we recall that we have to be careful with the order in which we
apply the creation operators. For instance, one has
v110 := Y+ X+ v = (X+ Y+ + [Y+ , X+ ])v = v110 − v001
Before we interpret this and get into danger to lose track in the next shells, let us see
how the natural representation of su(3) looks in this picture:
Here we have
3
V = ej C, e1 = t (1, 0, 0), e2 = t (0, 1, 0), e3 = t (0, 0, 1),
j=1
and hence
⎛ ⎞ ⎞⎛
1 1
H1 e1 = ⎝ −1 ⎠ ⎝ 0 ⎠ = e1 ,
0 0
⎛ ⎞⎛ ⎞
√ 1 1 √
H2 e1 = (1/ 3) ⎝ 1 ⎠ ⎝ 0 ⎠ = (1/ 3)e1 .
−2 0
√
H1 e2 = −e2 , H2 e2 = (1/ √3) e2
H1 e3 = 0, H2 e3 = −(2/ 3) e3 .
Recalling X+ = E12 , X− = E21 and the analogous relations for the other matrices, one
has
X+ e1 = 0, Y+ e1 = 0, Z+ e1 = 0,
X− e1 = e2 , Y− e1 = 0, Z− e1 = e3 ,
X+ e2 = e1 , Y+ e2 = 0, Z+ e2 = 0,
X− e2 = 0, Y− e2 = e3 , Z− e2 = 0,
X+ e3 = 0, Y+ e3 = e2 , Z+ e3 = e1 ,
X− e3 = 0, Y− e3 = 0, Z− e3 = 0.
Moreover, the equations translate to the fact that the creation operators applied to the
highest weight vector v3 and to v2 give zero with the exception X+ v2 = v3 . The only non-
trivial actions of the annihilation operators are Y− v2 = v1 , X− v3 = v2 and Z− v3 = v1 .
We compare this to the general form for the action obtained at the beginning (now with
v100 = 0, v010 = v2 , v001 = v3 ) and find as neccessary condition √for a three-dimensional
representation in our general scheme the condition k = 0, l = −2/ 3, just as it should be.
6.7 The Example su(3) 115
To show how to come to higher dimensional representations, one has to go to higher shells:
2
We have v200 = X+ v, v110 = X+ Y+ v, v101 = X+ Z+ v and so on. Here we remember
to pay attention to the order if the creation operators do not commute: As already
remarked, we have v110 := Y+ X+ v = v110 − v001 since one has [X+ , Y+ ] = Z+ . Though it
is a bit lengthy we give the list of the action of the annihilation operators (to be obtained
parallel to the case X− v100 above and using the actions of the minus operators on the
first shell)
X− v110 = −(k − 1)v √010 ,
Y− v110 = −(1/2)(l√3 − k)v100 ,
Z− v110 = −(1/2)(l 3 − k)v,
X− v101 = −(k + 1)v001 + v110 ,
Y− v101 = −v200 , √
Z− v101 = −(1/2)(l 3 + k + 2)v100 ,
X− v011 = v200 , √
Y− v011 = −(1/2)(l√3 − k)v001 − v110 ,
Z− v011 = −(1/2)(l 3 + k + 2)v010 ,
X− v200 = −2(k + 1)v100 ,
Y− v200 = 0,
Z− v200 = 0,
X− v020 = 0, √
Y− v020 = −(l 3 − k + 2)v010 ,
Z− v020 = 0,
X− v002 = 2v001 ,
Y− v002 = −2v√ 101 ,
Z− v002 = −(l 3 − k + 2)v001 .
Just as well one can determine the action of the minus operators in the next shell and so
on. As an example, we show how the eight-dimensional representation {8} (with highest
weight Λ = α1 + α2 ) comes up in this picture: As above we start with a lowest weight
vector v1 := v, which has in our coordinates the weight Λ0 = (k, l). In the first shell we
have
v2 := X+ v = v100 , v3 := Y+ v = v010 , v4 = Z+ v = v001
of respective weights Λ0 +α1 , Λ0 +α2 , Λ0 +α1 +α2 . For the second shell we put (secretely
looking at the weight diagram we already constructed using the results from the general
theory, otherwise the reasoning would take a bit more time)
2
X+ v = v200 = 0, Y+ X+ v = v110 =: v5 , Z+ X+ v =: v6 ,
and using the relation v110 = v110 − v001 of weight Λ0 + α1 + α2 obtained at the beginning
If we look at the relations coming from the application of the minus operators to the
elements of the second shell listed above, we find in particular the equations
√
X− v200 = −2(k + 1)v100 , Y− v020 = (l 3 − k + 2)v010 .
116 6. The Infinitesimal Method
√
Since we have fixed v200 = v020 = 0, these equations lead to k = −1 and l = 2/ 3.
One verifies easily that all the other relations are fulfilled and the weight diagram for the
eightfold way from above in this context leads to the following picture where only part
of the relations between the operators and generators are made explicit:
π̌(1, 1) = {8} Y
6
u r 2/√3 u
v7 = Y+ v4 v8 = Z+ v4
v5 = Y+ v2 1
u r j
u r u - I3
v3 = Y+ v1 −1 v4 = Z+ v1 v6 = Z+ v2
v1 v2 = X+ v1
u r−2/√3 u
Exercise 6.35: Do the same to construct the representations {6} and {10}.
We hope that these examples help to get a feeling how the discreteness of the weights is
related to the finite-dimensionality of the representations. And, as this procedure gets
more and more complicated, one gets perhaps some motivation to look into the proofs
of the general theorems in the previous section about the integrality of the weights, the
existence of fundamental weights, the action of the Weyl group etc.
Chapter 7
Induced Representations
given by right translation π(g0 )φ(g) = φ(gg0 ). Before we make all this more precise, in
the hope to add some motivation, we treat as an example a representation, which we
already treated as one of our first examples in 1.3.
118 7. Induced Representations
Example 7.1: Let G be the Heisenberg group Heis(R) = {g := (λ, µ, κ); λ, µ, κ ∈ R},
H the subgroup consisting of the elements h := (0, µ, κ) and π0 the one-dimensional
representation of H, which for a fixed m ∈ R∗ is given by
As we know from 3.3 that the Heisenberg group and the subgroup H are unimodular,
one has δ ≡ 1 (we shall see this below in 7.1.2) and we have to deal with the functional
equation (7.1)
By this functional equation the complex function φ(g) = φ(λ, µ, κ) is forced to have the
form
Comparing with (7.2) above, we rediscover the Schrödinger representation from Section
1.3: If we choose H such that in the decomposition of the elements φ above we have
f ∈ L2 (R), then
is the prescription for the unitary representation from Example 1.6 in 1.3. The repre-
sentation given by right translation on the space of functions φ on G of the form (7.2) is
called Heisenberg representation.
H × K −→ G, (h, k) −→ hk
is a topological isomorphism,
• δ = ∆H be the modular function of H, and
• π0 a representation of H in H0 .
7.1 The Principle of Induction 119
Then the homogeneous space of right H-cosets X = H\G can be identified with K
(provided with an action of K from the right). Let Hπ be the space of functions
φ : G −→ H0
satisfying the functional condition (7.1)
φ(hg) = δ(h)1/2 π0 (h)φ(g) for all h ∈ H, g ∈ G
and the finiteness condition that the restriction φ |K =: f is in L2 (K). We define a norm
for Hπ by
(7.3) φ 2Hπ := f 2L2 (K) = |f (k)|2H0 dk,
K
Proof: We have to show that φg0 given by φg0 (g) := φ(gg0 ) is in Hπ . One has
φg0 (hg) = φ(hgg0 ) = δ(h)1/2 π0 (h)φ(gg0 ) = δ(h)1/2 π0 (h)φg0 (g),
i.e. φg0 satifies the functional equation. Now to the finiteness condition: We write g = hk
and kg0 =: hk k with h, hk ∈ H, k, k ∈ K. Then we have (with δ = ∆H )
φg0 (k) = φ(kg0 ) = φ(hk k ) = δ(hk )1/2 π0 (hk )φ(k ),
and, since π0 is bounded, φg0 2Hπ is bounded by a constant times
∆H (hk ) | φ(k ) |2 dk,
K
and equality holds if π0 is unitary. The proof is done if we can show that for f ∈ Cc (K)
one has
f (k)dk = ∆H (hk )f (k )dk.
K K
We take φ ∈ Cc (G) and use Fubini’s Theorem. Since G is unimodular, we get
φ(hk)dkdh = G φ(g)dg = G φ(gg0 )dg
H K
= H K φ(hkg0 )dkdh
= K H φ(hhk k )dhdk
= K H φ(hk )∆H (hk )dhdk
= H K φ(hk )∆H (hk )dkdh.
Here we take φ such that
φ(hk) = ϕ(h)f (k) with f ∈ C(K), ϕ ∈ Cc (H), ϕ(h)dh = 1,
H
and get the desired result.
120 7. Induced Representations
Exercise 7.1: Verify that the case of the Heisenberg group and its Schrödinger repre-
sentation is covered by this theorem.
Remark 7.1: Using the notions from our section 6.6 where we introduced a bit more
structure theory, one can also state a useful variant of Theorem 7.1 (as in Flicker [Fl]
p.39):
If G has an Iwasawa decomposition G = N AK, K a maximal compact subgroup, A
the maximal torus in H = N A, N the unipotent radical, and χ a character of H, the
representation space consists of smooth functions φ : G −→ C with
π := indG
H π0
Before we elaborate on the fact that we really get a unitary representation, this condition
introducing the norm . Hπ needs some explanation:
7.1 The Principle of Induction 121
and for g ∈ G we have a density function ρg , which may be understood as the quotient
of dµg and dµ (in the general theory this goes under the name of a Radon-Nikodym
derivative). The construction of quasi-invariant measures can be done as follows.
We take over (from [BR] p.70:) [Ma1] part I, Lemma 1.1:
g = hs, h ∈ H, s ∈ B.
p : G −→ X , g −→ Hg =: x,
this leads to the existence of a Borel section s : X −→ G, i.e. a map preserving Borel
sets with p ◦ s = idX . Hence for g ∈ G, we have a unique Mackey decomposition
(7.5) G
g = h(x)s(x), h(x) ∈ H, x ∈ X .
∆G (h)
dr g =: dr hdµs (x)
∆H (h)
(by a reasoning, which is a refinement of the last part of the proof of Theorem 7.1, or by
solving the problems in [Ki] p.132) and we have the quasi-measure relation
Heis(R)
g = (λ, µ, κ) = h(x)s(x)
with
h(x) = (0, µ, κ + λµ), s(x) = (λ, 0, 0), x = λ.
122 7. Induced Representations
And, since the groups G and H are both unimodular, we simply have
In this special case, one easily verifies directly that dx is not only quasi-invariant but
invariant. And the general finite norm condition above reduces to the condition that f
is an L2 -function on X R, i.e. the condition we found above while rediscovering the
Schrödinger representation.
In the general case we will need the explicit form of the action of G on X in the framework
of the Mackey decomposition: We apply the Mackey decomposition (7.5) to s(x)g0 and
get a consistent picture by chosing xg0 such that, for h∗ := h(xg0 ) ∈ H (which, for
reasons to be seen later, we also write as h(g0 , x)), we have the Master Equation (the
name is from [Ki1] p.372)
Now, let us verify that the space Hπ , as defined above, is in fact the space for our induced
representation with π(g0 )φ = φg0 :
If φ satisfies the functional equation (7.1), by the same reasoning as in the proof of
Theorem 7.1, so does φg0 for every g0 ∈ G.
The main point is the norm condition. The more experienced reader will see that, in
principle, we do the same as in the proof of Theorem 7.1: Using (7.7) above, the functional
equation (7.1) and the unitarity of π0 , one has from (7.4)
φg0 2Hπ = X φg0 (s(x)) 2H0 dµs (x)
= X φ(s(x)g0 ) 2H0 dµs (x)
= X φ(h∗ s(xg0 )) 2H0 dµs (x)
= X φ(s(xg0 )) 2H0 δ(h(g0 , x))dµs (x).
Replacing x −→ xg0−1 , we see that translation does not change the norm
φg0 2Hπ = X φ(s(x)) 2H0 δ(h(g0 , xg0−1 ))dµs (xg0−1 )
= X φ(s(x)) 2H0 dµs (x) = φ 2Hπ ,
since from the Master Equation (7.7) (as expression of the associativity of the group
operation) one deduces the cocycle condition
Via the chain rule the quasi-measure relation (7.6) dµs (xg0 )/dµs (x) = ρg0 (x) = δ(h(x, g0 ))
provides for
δ(xg0−1 , g0 )dµs (xg0−1 ) = dµs (x).
It is now evident that π is unitary if we put
< φ, ψ >:= < φ(s(x)), ψ(s(x)) >H0 dµs (x).
X
7.1 The Principle of Induction 123
Thus we have the essential tools to extend the validity of Theorem 7.1 to the more gen-
eral hypothesis stated above (we leave out to show that changing the section ϕ leads to
equivalent representations). This realization of the induced representation by functions
living on the group is usually called the first realization or induced picture.
Before we go over to another realization, let us point out that the representation by
right translation has an equivalent version where G acts by left translation, i.e. we have
π (g0 )φ(g) = φ(g0−1 g) where here we have the space H consisting of functions φ fulfilling
π
We stay with the same hypotheses as above and turn to functions living on the homoge-
neous space X = H\G. We denote by
Hπ := L2 (X , dµs , H0 )
the space of H0 -valued functions f on X , which are quadratic integrable with respect
to the quasi-invariant measure dµs . In this context we have as fundamental fact the
following
ϕ : Hπ −→ Hπ , f −→ φf
Using this, we can carry over the induced representation from Hπ to Hπ and from our
Theorem 7.2 come to an equivalent picture as follows.
124 7. Induced Representations
H π0 is realized on Hπ by π̃ with
Theorem 7.4 (Second Realization): π = indG
with
A(g0 , x) := δ(h(g0 , x))1/2 π0 (h(g0 , x))
for s(x)g0 = h(g0 , x)s(xg0 ), x ∈ X , g0 ∈ G.
π(g0 )
Hπ - Hπ
6
ϕ ψ
π̃(g0 ) ?
Hπ - Hπ
By the Master Equation s(x)g0 = h(g0 , x)s(xg0 ) and the definition of φ this can be
written as
Exercise 7.4: Repeat this discussion to get a representation space consisting of functions
f living on the space of left cosets X = G/H by starting from the induced representation
π given by left translation on the space H . Deduce the representation prescription
π
Remark 7.2: In the special case that π0 is the trivial one-dimensional representation
with π0 (h) = 1 for all h ∈ H, we fall back to the situation of our Example 1.5 in Section
1.3 where we constructed a representation on a space of functions on a homogeneous
space X = H\G. This may be seen as starting point of the construction of induced
representations.
7.1 The Principle of Induction 125
Remark 7.3: From the cocycle condition (7.8) for h = h(g, x), one can deduce that the
multiplier system A in Theorem 7.4 as well fulfills this relation, i.e. we have
A(g1 g2 , x) = A(g1 , x)A(g2 , xg1 ) for all g1 , g2 ∈ G, x ∈ X .
And, vice versa, one can take such a system as tool to construct a representation on a
homogeneous space X with a right G-action (see for instance [Ki1] p.388).
For instance in [BR] p.478 a proof can be found that essentially all this does not depend
on the choice of the section s. In most practical cases one has a more or less natural
decomposition of the given group into subgroups as at the beginning of this section, so
one can avoid to talk about different sections, since one has a natural decomposition.
We already discussed the Heisenberg group as an example for this. Here the Schrödinger
representation came out as the Second Realization with π̃ = πm acting on Hπ = L2 (R).
The First Realization by functions living on the group is called Heisenberg representation .
M AN (σ ⊗ exp ν ⊗ 1)
π = indG
and in the induced picture given by left translation π(g0 )φ(g) = φ(g0−1 g) on the represen-
tation space, which is given by the dense subspace of continuous functions φ : G −→ V σ
with functional equation
φ(gman) = e−(ν+ρ) log a σ(m)−1 φ(g)
and the norm
φ2 := |φ(k)|2 dk.
K
By restriction to the compact group K one gets the compact picture. A dense subspace
is given by continuous functions f : K −→ V σ with
f (km) = σ(m)−1 f (k)
and the same norm as above. Here the representation π̃ is given by
−1
π̃(g0 )f (k) := e−(ν+ρ)H(g0 k)
σ(µ(g0−1 k))−1 f (κ(g0−1 k))
where g decomposes under G = KM AN as g = κ(g)µ(g)eH(g) n.
126 7. Induced Representations
In [Kn] p.169 there is a third picture, the noncompact picture, given by restriction to
N̄ = ΘN . Here the representation π̃ is given by
−1
π̃(g0 )f (n̄) := e−(ν+ρ) log a(g0 n̄)
σ(m(g0−1 n̄))−1 f (n̄(g0−1 n̄))
if one takes account of the fact that N̄ M AN exhausts G except for a lower dimensional
set and that in general one has a decomposition g = n̄(g)m(g)a(g)n. The representation
space is here L2 (N̄ , e2Re νH(n̄) dn̄).
Here we can not give satisfying answers to these questions (but hope that our examples
later on will be sufficiently illuminating). We only mention some remarks.
is a dense set in Hπ .
Ad ii) We have seen that the Schrödinger representation of the Heisenberg group is an
irreducible induced representation. In general, this is not the case and in our examples
below we will have to discuss this. We shall find conditions to be added to the definition
of our space if one wants to come to an irreducible representation (e.g. holomorphic in-
duction). In [BR] p.499, there are general criteria to get irreducibility (using the concept
of a canonical system of imprimitivity).
Ad iii) This concept of canonical system of imprimitivity is also used to give a criterium
of inducibility stating under which conditions a unitary representation is equivalent to
an induced representation. One finds versions of this in [BR] p.495, [Ki] p.191, and in
[Ki1] p.389 (Mackey Inducibility Criterion). We do not reproduce these because the
statements need more new notions.
7.1 The Principle of Induction 127
Ad iv) Induction and restriction are adjoint functors. We do not try to make this more
precise. At least for finite groups, one finds statements of the Frobenius Reciprocity as in
[Se] p.57 or [Ki] p.185, which, using our notation (see section 1.2), for a representation
π of G can be written as
c(π, indG
H π0 ) = c(π|H , π0 ).
Ad v) One has the concept of induction in stages, which symbolically can be written as
H indK indK
indG H G
Ad vi) If π0 and π0 are representations of the subgroup H of G the operations of taking
direct sum and induction are interchangeable
G
H (π0 ⊕ π0 ) indH π0 ⊕ indH π0
indG G
([BR] p.488). And if π1 and π2 are unitary representations of the closed subgroups H1
and H2 of the separable, locally compact groups G1 resp. G2 , then one has ([BR] p.491)
1 ×G2
H1 ×H2 (π1 ⊗ π2 ) indH1 π1 ⊗ indH2 π2 .
indG G1 G2
Up to now, we treated only the example of the Heisenberg group. Before we do more se-
rious work with G = SL(2, R), we discuss another example we already looked at, namely
G = SU(2), and the two-dimensional form of the Poincaré group, which later on will
be discussed more thoroughly in the four-dimensional form in the context of Mackey’s
method for semidirect products.
A n a
G = {g = ( ) ∈ SL(3, R); A ∈ SO(1, 1), n = ( ) ∈ R2 }
0 1 t
In analogy to the case of SO(2), we write the elements of SO(1, 1) in the form
cosh ϑ sinh ϑ
A=( ) =: rh(ϑ), ϑ ∈ R.
sinh ϑ cosh ϑ
From the addition theorem we have AA = rh(ϑ)rh(ϑ ) = rh(ϑ + ϑ ), and dϑ is a bi-
invariant measure on SO(1, 1). We take the subgroup
E2 n
H := {h = ( ); n ∈ R2 }
0 1
t α
π0 (h) := eπi(aα+tβ) = eπi nn̂
=: χn̂ (n), n̂ := ( ) ∈ R2 .
β
128 7. Induced Representations
The induced representation has its Second Realization on the space Hπ = L2 (X , dµ(x)) =
L2 (R, dϑ). We have the isomorphism given by
Hπ
φ −→ f ∈ Hπ with f (ϑ) := φ((rh(ϑ), 0)),
and the prescription for the representation from Theorem 7.4, namely
π̃(g0 )f (x) = A(g0 , x)f (xg0 ),
in this case leads to
t
π̃((rh(ϑ0 ), n0 ))f (ϑ) = χn̂ (rh(ϑ)n0 )f (ϑ + ϑ0 ) = eπi n0 rh(ϑ)n̂
f (ϑ + ϑ0 ).
The representation is irreducible. This is a byproduct of Mackey’s general theory for
representations of semidirect products, upon which we will report later in section 7.3.
The reader is invited to find a direct proof using the infinitesimal method analogous to
the example in 6.5 (Exercise 7.5). In the next example the induced representation will
come out to be reducible.
Using the Infinitesimal Method, in Theorem 6.2 in section 6.4, we obtained a list fixing the
possible equivalence classes of irreducible unitary representations of G = SL(2, R). Here
we will show that these representations really exist by presenting concrete realizations
or, as one also says, models. In most occasions it is important to have different models,
depending on the occasion where the representation is showing up. The method we follow
here is just the application of the induction procedure that we developed in the preceding
section. Thus, we have to start by choosing a subgroup H of our G. Already in several
occasions in this text, we specified subgroups of G = SL(2, R). We recall (for instance
from 6.6.4) the notation
cos ϑ sin ϑ
K := SO(2) = {r(ϑ) = ( ); ϑ ∈ R},
− sin ϑ cos ϑ
y 1/2
A := {t(y) = ( ); y ∈ R, y > 0},
y −1/2
1 x
N := {n(x) = ( ); x ∈ R},
1
1
N̄ := {n̄(x) = ( ); x ∈ R},
x 1
M := {±E2 },
a b
B := M AN = {h = ( ); a, b ∈ R, a = 0},
a−1
B̄ := M AN̄ .
a b d−1 b 1
g=( ) = ( )( ) = h(x)s(x) for d = 0,
c d d c/d 1
b −a 1
= ( )( ) = h(x)s(x) for d = 0.
b−1 −1
7.2 Unitary Representations of SL(2, R) 131
1 1
Hence X is in bijection to the set S := {( ); x ∈ R} ∪ {( )}. As in our
x 1 −1
1
context a point has measure zero, we can disregard the element ( ) and work
−1
with
G −→ X = H\G ⊃ X R, g −→ x := c/d, d = 0
1 a b0
with section s : X −→ G, x −→ s(x) = ( ). For g0 = ( 0 ) our Master
x 1 c0 d0
Equation (7.7)
s(x)g0 = h(g0 , x)s(xg0 )
is realized by
(b0 x + d0 )−1 b0 1 a0 x + c0
h(x, g0 ) = ( ), s(xg0 ) = ( ), xg0 = .
0 b0 x + d 0 xg0 1 b0 x + d0
and
φ :=2
| φ(s(x)) |2 dx < ∞.
R
Remark 7.4: This exercise provides as byproduct a proof that this induced representa-
tion is irreducible. There are more direct ways to see this (see for instance [Kn] p.36).
Exercise 7.8: We propose to verify in a special situation that the construction es-
sentially does not depend on the choice of the quasi-invariant measure: As in [BR]
p.483, repeat the former discussion for the quasi-invariant measure dµ(x) = ϕ(x) dx
with ϕ(x) = (1 + x2 )−2 to get a representation space H̃π linked to our Hπ by the inter-
twining operator f˜ −→ f, f (x) = f˜(x)/(1 + x2 )1/2 .
132 7. Induced Representations
To get more familiarity with these constructions and because of the general importance of
the result, we also discuss the realization of our induced representation by left translation,
s.t., for functions φ living on G, we have a representation given by π (g0 )φ(g) = φ(g0−1 g).
a b
We keep H and π0 as above and now put X = G/H. For g = ( ) with a = 0, we
c d
have a Mackey decomposition
1 a b
g = s(y)h(y), s(y) = ( ), y = c/a, h(y) = ( ),
y 1 a−1
with
1 a0 y − c0
s(g0−1 y) = ( ), g0−1 y = ,
g0−1 y 1 −b0 y + d0
d0 − b 0 y −b0
h(y, g0 ) = ( ) =: h∗ .
(d0 − b0 y)−1
put φg0 (g) := φ(g0−1 g), and calculate using the Master Equation g0−1 s(y) = s(g0−1 y)h∗
and the unitarity of π0
φg0 2 = R | φg0 (s(y)) |2 dµ(y)
= R | φ(s(g0−1 y)h∗ ) |2 dµ(y)
= R | φ(s(g0−1 y)) |2 δ(h∗ )−1 dµ(y)
= R | φ(s(ỹ)) |2 dµ(ỹ),
since one has by the quasi-measure relation
dµ(ỹ) dµ(g0−1 y) 1
= = = δ(h∗ )−1 .
dµ(y) dµ(y) (−b0 y + d0 )2
So we have as the First Realization the induced representation defined by functions φ on
G with the functional equation
1
From here, by the map φ −→ f, f (y) := φ(( )), we come to the Second Realization
y 1
given on the space Hπ = L2 (R, dy) by the prescription
In [Kn] p.168, one finds a proof that this norm and the one we used above are equal. We
propose to search here for a direct proof (Exercise 7.9).
Moreover, in [Kn] p.37, there is a proof of the irreducibility of these induced representa-
tions πis,± except for π0,− , which is based on euclidean Fourier Transform.
We will discuss several models, which realize the discrete series representations.
a b
For g = ( ) ∈ SL(2, R) and τ ∈ H, we put
c d
aτ + b
g(τ ) :=
cτ + d
and call this a linear fractional transformation.
Exercise 7.10: Prove that this is well defined and provides a transitive left action of
G = SL(2, R) (even of GL(2, R)) on H.
Verify that H = SO(2) is the stabilizing group of τ = i.
Show that the automorphic factor
fulfills the cocycle relation (7.8) (in the version for the action from the left)
x − t(x2 + y 2 )
x = ,
t2 y 2 + (1 − tx)2
y
y = 2 2 ,
t y + (1 − tx)2
1 − t(x − iy)
eiϑ = eiϑ
.
t2 y 2 + (1 − tx)2
RF = −∂x ,
(7.14) RG = (x2 − y 2 )∂x + 2xy∂y + y∂ϑ ,
RH = −2x∂x − 2y∂y ,
which combine to
RZ = i((1 + x2 − y 2 )∂x + 2xy∂y + y∂ϑ ),
(7.15)
RX± = ±(i/2)(((x ± i)2 − y 2 )∂x + 2(x ± i)y∂y + y∂ϑ ).
LZ = −i∂ϑ ,
(7.16)
LX± = ±(i/2)e±2iϑ (2y(∂x ∓ i∂y ) − ∂ϑ ).
Ω := X+ X− + X− X+ + (1/2)Z 2 ,
After these general preparations, we finally come to the discrete series representations.
There are a lot of different models in the literature. We present the most customary one
and then discuss how it fits into the scheme of the induction procedure.
For instance in [La1] p.181, we find the representation πm with the representation space
Hm := L2hol (H, µm ), m ≥ 2. This space consists of holomorphic (i.e. complex differen-
tiable) functions f on the upper half plane H, which have a finite norm f m < ∞,
i.e. with
dxdy
f 2m := | f (x, y) |2 y m 2 < ∞
H y
and one has an scalar product defined by
dxdy
< f1 , f2 >:= f1 (x, y)f2 (x, y)y m 2 .
H y
As will be apparent soon (even for those readers, which are not so familiar with complex
function theory), the main point is here reduced to the fact that holomorphic functions
f are essentially smooth complex valued functions characterized by the Cauchy Riemann
differential equation
(∂x + i∂y )f = 0.
a b
We write f (τ ) := f (x, y) and put g −1 =: ( ) (this is some trick to get nicer formulae
c d
but, if not remembered carefully, sometimes leads to errors). Then the representation
πm is in the half plane model given by
jm (g, τ ) := j(g, τ )m
Remark 7.5: In [Kn] p.35, one finds a representation given on the same space by the
prescription
a b aτ − c
π̌m ( ))f (τ ) := (−bτ + d)−m f ( ).
c d −bτ + d
And in [GGP] p.53, the authors define a representation Tm given by
a b aτ + c
Tm ( ))f (τ ) := (bτ + d)−m f ( ).
c d bτ + d
The reader is invited to compare this to our definition. We stick to it because we want
to take over directly some proofs from Lang’s presentation in [La1] p.181ff. Namely, one
has to verify several things:
– Hm is complete and non-trivial.
– π fulfills the algebraic relation π(gg ) = π(g)π(g ).
– π is continuous and unitary.
7.2 Unitary Representations of SL(2, R) 137
The algebraic relation is an immediate consequence of the facts stated as Exercise 7.10.
We verify the unitarity using Exercise 7.11:
∞∞
π(g)f 22 = 0 −∞ | f (g −1 τ ) |2 | cτ + d |−2m y m−2 dxdy
∞∞
= 0 −∞ | f (τ ) |2 y m−2 dxdy
= f 22 .
For the other statements we refer to Lang. The idea is to transfer the situation from the
upper half plane to the unit disc where the function theory is more lucid. Since this is
an important technique, we report on some facts:
c̃
G = SL(2, R) - G = SU(1, 1)
p p
? c ?
H = SL(2, R)/SO(2) - D = SU(1, 1)/U (1)
with
1 + iτ̄ −1 − iτ
c̃(n(x)t(y)) = (1/2)y −1/2
−1 + iτ̄ 1 − iτ
τ −i
and c(τ ) = ζ = τ +i .
Now, it is a rather easily acceptable fact that the space of holomorphic functions on D
is spanned by the monomials ζ n , n ∈ N0 . And some calculation shows that
dudv dudv
dν(ζ) := , dνm (ζ) := 41−m (1− | ζ |2 )m ,
1− | ζ |2 1− | ζ |2
given by
ζ + 1 −2i m
f˜(ζ) = f (−i )( ) .
ζ −1 ζ −1
Using a convenient extension of the group action given by (7.18), this can also be written
as
f˜(ζ) = f (C −1 (ζ))jm (C −1 , ζ).
The map Cm intertwines the representation πm from the half plane model given by (7.18)
with the equivalent representation π̃m on the space of functions f˜ ∈ L2hol (D, dνm ) given
by the prescription
π̃m (g)f˜(ζ) := f˜(c̃(g)−1 (ζ))jm (c̃(g)−1 , ζ).
We call this the unit disc model. By the way, if one replaces c̃(g) by g ∈ G , one also
has a representation of G = SU(1, 1).
Already above, we stated the (plausible) fact that L2hol (D, dνm ) is spanned by functions
f˜n with f˜n (ζ) = ζ n , n ∈ N0 . As to be seen easily, via Cm , these functions correspond to
functions fn living on H with
τ −i n
fn (τ ) := ( ) (τ + i)−m .
τ +i
7.2 Unitary Representations of SL(2, R) 139
Intertwined with these two equivalent models, there is a third one, which we will call the
group model because its representation space consists of functions living on the group
G = SL(2, R). We go back to the map c̃ : G −→ G̃. In the formulae (7.19) one has the
elements of G̃ as functions of g ∈ G. We introduce the parameters x, y, ϑ describing g in
the Iwasawa decomposition as in (7.12) and get by a small calculation
Here we use the fact that one has jm (g, i) = (cτ + d)−m = y m/2 eimϑ if g is fixed by
our Iwasawa parametrization (7.12). With the help of the cocycle condition for the
automorphic factor (or multiplier) jm , we deduce for these lifted functions
This shows that the lift ϕm intertwines (algebraically) the representation πm on the space
L2hol (H, dµm ) with the left regular representation λ on the space of functions φ living on
G and coming out by the lift ϕm . In [La1] p.180 one finds a proof that the functions φn
are square integrable with respect to the (biinvariant) measure dg on G given (again in
our Iwasawa parameters) by
dg = y −2 dxdydϑ.
We call this representation, given by left translation on the space spanned by the func-
tions φn living on the group, the group model.
Already earlier, while introducing the other two models, we could have identified the
−
class of our representation with the one of the representation πm of highest weight −m
from the list of (possible) irreducible unitary representations obtained by infinitesimal
considerations in 6.4. We will do this now. The reader is invited to watch very closely
because Lang in [La1] p.184 (erroneously if our reasoning is correct) identifies it as a
lowest weight representation.
There are several ways to proceed: For the isomorphism c̃ : G = SL(2, R) −→ SU(1, 1)
one has c̃(r(ϑ)) = s(ϑ), c̃(r(ϑ)−1 g)) = s(−ϑ)c̃(g). Hence we deduce
and
φn (r(ϑ)−1 g) = e−(2n+m)iϑ φn ,
i.e. our representation has the weights −m, −(m + 2), −(m + 4), . . . . We compare this
with our results from 6.4 and conclude that we have a representation equivalent to the
140 7. Induced Representations
−
highest weight representation πm . But we can do more: In (7.15) we determined the op-
erators RF , RG , RH and RZ , RX± . These operators are just the operators producing the
derived representation of the left regular representation λ. Thus, by a small calculation
(Exercise 7.14), one can verify the formulae
Rm
Z = π̂m (Z) = i((1 + τ 2 )∂τ + mτ ),
Rm
X± = π̂m (X± ) = ±(i/2)((i ± τ )2 ∂τ + m(τ ± i)),
and
Rm
Z fn = −(m + 2n)fn ,
Rm
X+ fn = −nfn−1 ,
Rm
X− fn = (n + m)fn+1 .
+
Similarly the representations πm of lowest weight can be realized by holomorphic func-
tions on the lower half plane H̄ = {τ = x+iy ∈ C, y < 0} or by antiholomorphic functions
f , i.e. with (∂x − i∂y )f = 0.
Exercise 7.16: Determine the class of the representation π̌m taken over from Knapp in
the Remark 7.5 above.
Finally, we want to show how these realizations of the discrete series representations fit
into the discussion of the induction procedure. Namely one can obtain them as subrep-
resentations of suitable induced representations:
From the functional equation we learn that the functions φ in the representation space
are all of the type
φ(g) = e−ikϑ F (x, y)
with a function F depending only on x, y if g is given in our Iwasawa parametrization.
Comparing this to the functions establishing the group model above, we see that we have
to find a condition, which further restricts these functions F . Here we remember from
the infinitesimal considerations in 6.4 that functions spanning the space of a discrete
series representation πk± have to be eigenfunctions of the Casimir element
Ω = X+ X− + X− X+ + (1/2)Z 2 .
We get
i.e. we have
∆φ = (k/2)((k/2) + 1)φ
if
∆k f := 4y 2 (∂x2 + ∂y2 ) + 4yik(∂x + i∂y )f = 0,
and this is obviously the case if f fulfills the equation
i.e. if f is holomorphic. Hence for k = −m, we have recovered the lift of the representation
πm as a subrepresentation of our induced representation πkind .
There is a shorter way to single out in Hχk a subspace carrying πm : To the two conditions
characterizing the space of the induced representation, namely the functional equation
and the norm condition, we add simply the condition
LX− φ = 0.
For φ(g) = e−ikϑ y −k/2 f (x, y) this again leads to the holomorphicity condition ∂τ̄ f = 0.
Remark 7.7: Here we began by stating the prescription for the representation πm in a
space of functions living on the homogeneous space H = G/SO(2). It is a nice exercise to
142 7. Induced Representations
rediscover this formula by our general procedure to change from the First to the Second
Realization of an induced representation:
a b
We start with Mackey’s decomposition as above, g = s(τ )r(ϑ), and for g0−1 = ( )
c d
the Master Equation, using s(τ ) = n(x)t(y), states
with
iϑ∗ cτ + d
e = .
cτ̄ + d
Moreover one has the standard formulae
dµ(g0−1 (τ ) 1 y
= , Im (g0−1 (τ )) = ,
dµ(τ ) (cτ + d)2 | cτ + d |2
such that the left regular representation π(g0 )φ(g) = φ(g0−1 g) in the Second Realization
with F (τ ) = φ(s(τ )) translates to the prescription
dµ(g0−1 (τ )) 1/2 −1
π(g0 )F (τ ) = ( ) π0 (r(ϑ∗ ))F (g0−1 (τ )).
dµ(τ )
The representations π1± are called mock discrete series representations or limits of discrete
±
series. They are realized by the same prescriptions as the πm with m ≥ 0 but here the
−
space of holomorphic functions f on H for π1 gets the new norm
f 2 = sup | f (x + iy) |2 dx,
y>0 R
The courageous reader may be tempted to apply this principle of holomorphic induction
to reconstruct as Exercise 7.17 the irreducible unitary representations of G = SU(2).
For the sake of completeness, we indicate here also a model, but only briefly, as done
in most sources. (The best idea is to go back to Bargmann’s original paper [Ba].) We
take up again the prescriptions as in the construction of the principal series but to get
a model for πs , 0 < s < 1 we change the norm on the space of complex valued functions
on R and introduce
f (x)f (y)
f s :=
2
dxdy.
R R x−y |
| 1−s
7.3 Unitary Representations of SL(2,C) and of the Lorentz Group 143
a b
For g −1 = ( ) the prescription is here
c d
Some more models can be found in [vD] p.465 and [GGP] p.36.
We close this section by pointing out to the Subrepresentation Theorem, which goes
back to Harish–Chandra and roughly states (for a precise formulation and proof see [Kn]
p.238) that a certain class of representations, the admissible representations, may be
realized as subrepresentations of parabolically induced representations starting eventually
by non-unitary representations of the subgroup. Admissible representations are very
important, in particular in number theory. An admissible representation (π, V ) of G
is (in first approximation) characterized as follows ([Ge] p.55): Let K be a maximal
compact subgroup of G. For an irreducible representation χ of K, denote by V (χ) the
subspace of vectors in V, which transform according some multiple of χ. Then one has a
decomposition
V = ⊕χ V (χ)
where each V (χ) is finite dimensional.
The interested reader can exemplify the statement of the subrepresentation theorem in
the case of G = SL(2, R) by constructing as Exercise 7.18 another model of the discrete
series representation by starting the induction procedure from
Thus, in contrast to the smaller group SL(2, R), we have no discrete series representa-
tions. We reproduce models for these representations from [Kn] p. 33/4 and discuss their
construction via the induction procedure afterwards.
a b −bz + d −k az − c
(7.20) πk,iv (( ))f (z) :=| −bz + d |−2−iv ( ) f( ).
c d | −bz + d | −bz + d
Here, the notion of the fractional linear transformation used in the last section as a map
SL(2, R) × H −→ H, (g, τ ) −→ g(τ ), is naturally extended to a map
SL(2, C) × {C ∪ ∞} −→ {C ∪ ∞}
with
a b
( )(z) := az+b
cz+d for z ∈ C, cz + d = 0,
c d
:= ∞ for z ∈ C, cz + d = 0,
:= a
c for z = ∞.
In a similar way, for k ∈ Z, w = u + iv ∈ C, there is a non–unitary principal series πk,w
given by the prescription
a b −bz + d −k az − c
πk,w (( )f (z) :=| −bz + d |−2−w ( ) f( )
c d | −bz + d | −bz + d
for functions f in the space L2 (C, dµw ) with dµw (z) = (1+ | z |2 )u dzdz̄. However, in the
special case k = 0, w = u ∈ R, 0 < u < 2, the representation πu := π0,u is unitary if the
representation space is provided with the inner product
f (z)f˜(ζ)
< f, f˜ >:= dzdζ.
C C | z − ζ |2−u
As to the construction, these representations are induced representations where one takes
the subgroup
α β
H := {h = ( ); α, β ∈ C, α = 0}
α−1
and the representation π0 with π0 (h) :=| α |w ( |α|
α k
) . We have the Mackey decomposition
a b
g=( ) = s(τ )h
c d
where (as in our discussion of the principal series for SL(2, R) in the previous section) it
suffices to treat the case a = 0, such that one has
1 a b
s(τ ) = ( ), z := b/a and h = ( ),
z 1 a−1
7.3 Unitary Representations of SL(2,C) and of the Lorentz Group 145
az − c −bz + d −b
z∗ = , h∗ = ( ).
−bz + d (−bz + d)−1
Moreover, we have
dz ∗ dz̄ ∗ = | −bz + d |−4 dzdz̄.
Hence, since the induced representation in left regular version in its Second Realization
on the space H = L2 (C) has the general prescription
dµ(z ∗ )
π(g)f (z) = π0 (h∗ )f (z ∗ ), for all f ∈ H,
dµ(z)
In [Kn] p.33/4 Knapp proves the irreducibility of πk,iv using Schur’s Lemma in the uni-
tary version cited in 3.2. One could also do this (and verify that really we only have
the above mentioned irreducible unitary representations) by infinitesimal considerations
as in our discussion of SL(2, R). But since sl(2, C) has real dimension 6, this is leads
to considerable non-trivial computations (to be found for instance in Naimark [Na] or
Cornwell [Co] p.668ff in the framework of the discusssion of the Lorentz group).
The interest in the representations of G = SL(2, C) is enlarged by the fact that they
immediately lead to projective representations of the proper orthochronous Lorentz group,
resp. those, which are trivial on −E2 , to genuine representations, the same way as in our
discussion of SU(2) and SO(3) in 4.3. Before we describe the relation between SL(2, C)
and the Lorentz group we have to fix some notation.
We write
⎛ ⎞
1
⎜ 1 ⎟
GL := {A ∈ GL(4, R); t AD3,1 A = D3,1 = ⎜
⎝
⎟}
⎠
1
−1
for the complete homogeneous Lorentz group, i.e. the group of linear transformations,
which leave invariant the Minkowski metric from the theory of special relativity
One may be tempted to replace (x1 , x2 , x3 , x4 ) by (x, y, z, ct), (c the velocity of light)
and put
ds2 = c2 dt2 − dx2 − dy 2 − dz 2 ,
as often done in the physics literature, but we will not do this here. In our former notation
from 0.1 we also have GL = O(3, 1) and GL + = SO(3, 1) for the proper homogeneous
Lorentz group, i.e. the invariant subgroup consisting of matrices A with det A = 1. We
write the matrices in block form
B x
A=( t ), B ∈ GL(3, R), x, y ∈ R3 , a ∈ R.
y a
146 7. Induced Representations
GL
+ has two connected components distinguished by a > 0 and a < 0 since one has
t
B y E B x t
BB − y t y t
Bx − ya
( )( 3 )( )=( t ),
t
x a −1 ty a xB − at y t
xx − a2
E3
and the condition that this should be equal to ( ) asks for
−1
(7.21) t
BB − y t y = E3 , t Bx − ya = 0, t xx − a2 = −1,
B
SO(3)
B −→ B̃ := ( ).
1
These elements are called Lorentz rotations. Moreover, we introduce the Lorentz boosts,
which are of the type
⎛ ⎞
cosh t sinh t
⎜ 1 ⎟
B(t) = ⎜⎝
⎟ , t ∈ R.
⎠
1
sinh t cosh t
Proposition 7.2: For every A ∈ G0 there are R, S ∈ SO(3) and t ∈ R such that
A = R̃B(t)S̃,
A proof can be found in any source treating the Lorentz group, for instance in [He] p.75.
Our main fact in this context is the next theorem.
α z
V := {X ∈ M2 (C); X = t X̄ =: X ∗ } = {( ); α, β ∈ R, z ∈ C}
z̄ β
−a + d b + ic
ϕ : R4 −→ V, v = t (a, b, c, d) −→ ( ).
b − ic a + d
7.4 Unitary Representations of Semidirect Products 147
For X = ϕ(v), Y = ϕ(w) this map transforms the Minkowski (quasi-)inner product
1
and deduce ker ρ = {±E2 } by evaluating the defining condition for X = E2 , ( ),
−1
1
and ( ). The main task is to prove the surjectivity of ρ. We go back to the proof
1
of Theorem 4.7 in 4.3 and use the fact that G0 is generated in GL(V ) by elements
S̃2 (α), S̃3 (α) and B(t). By some computation, one verifies that these are images under
ρ of respectively r(α/2), s(α/2) and s(−it/2).
This theorem justifies the statement made above that the representations π of SL(2, C)
with π(−E2 ) = id create the representations of SO(3, 1)0 . E.g., in [Co] p.671ff one finds
a discussion of the finite-dimensional (non-unitary) representations of the Lorentz group.
with the additional condition that every ρ(a) is an automorphim of N , i.e. ρ(a) ∈ Aut N
for all a ∈ G0 . Then we define the associated semidirect product G := G0 N as the set
of pairs g = (a, n), a ∈ G0 , n ∈ N provided with the composition law
G comes out as a group with neutral element e consisting of the pair of neutral elements
in G0 and in N . The inverse to g = (a, n) is g −1 := (a−1 , ρ(a−1 )n−1 ). (Verification of
the associative law as Exercise 7.19.)
148 7. Induced Representations
Examples:
A a
GE (n) = {g = (A, a) = ( ); A ∈ SO(n), a ∈ Rn }.
1
A a A a AA a + Aa
gg = ( )( )=( )
1 1 1
reflects exactly the composition law (7.22) leading in this case to gg = (AA , a + Aa ).
This law comes out very naturally if one asks for the composition of the affine transfor-
mations x −→ Ax + a, which preserve the euclidean norm x2 = x21 + · · · + x2n .
Example 7.5: The Poincaré group GP (n) in n dimensions: Paralell to the case above
we take G0 = SO(n − 1, 1) and N = Rn and get
A a
GP (n) = {g = (A, a) = ( ), A ∈ SO(n − 1, 1), a ∈ Rn }
1
with the same composition law. This is the invariance group of the affine transformations
preserving the Minkowski (pseudo-)norm x2 = x21 + · · · + x2n−1 − x2n .
a b
Example 7.6: Interprete the group B := {g = ( ); a, b ∈ R, a > 0} as a semidi-
a−1
rect product (Exercise 7.20).
Example 7.8: The Jacobi group GJ is the semidirect product of G0 = SL(2, R) acting
on the Heisenberg group N = Heis(R) via
a b
(( ), (λ, µ, κ)) −→ (dλ − cµ, −bλ + aµ, κ).
c d
This group can be presented in several ways as a matrix group (see our Section 8.5 or
for instance [EZ] or [BeS] p.1).
With exception of the last one, in all these examples the group N is abelian. In the sequel
we will restrict our treatment to this case as things become easier under this hypothe-
sis and use additive notation in the second item of the pair g = (a, n) (as we already
7.4 Unitary Representations of Semidirect Products 149
did in the examples). Whoever is interested in more general cases is refered to Mackey’s
original papers or the treatment of the representation theory of the Jacobi group in [BeS].
G0 −→ G = G0 N, a −→ (a, 0) =: a
and
N −→ G = G0 N, n −→ (e, n) =: n.
Here one has to be a bit careful because we have n · a = (e, n)(a, 0) = (a, n) = g but
a · n = (a, 0)(e, n) = (a, ρ(a)n) = (a, n) = g in general. There is the useful relation
and
N := {n = (0, y, z); y, z ∈ R}. |=
Using the multiplication law gg = (x + x , y + y , z + z + xy ), one calculates
and this is consistent with the usual semidirect product composition law (7.22) given by
gg = (a, n)(a , n ) = (aa , nρ(a)n ).
a b
B := {g = ( ) =: (a, b); a, b ∈ R, a > 0}
a−1
one may be tempted to try G0 := {a = (a, 0); a > 0} and N := {n = (1, b); b ∈ R}.
−1
From matrix multiplication we have gg = (aa , ab + ba ) and (using the embedding
introduced above) ana−1 = (1, a2 b) = ρ(a)n. But then the composition law would give
gg = (a, b)(a , b ) = (aa , nρ(a)n ) = (aa , b + a2 b ) which is not the multiplication law in
this group. To make things consistent (and thus solve the problem from Exercise 7.20),
one constructs the semidirect product G = R>0 R = {[a, x], a, x ∈ R, x > 0} with
ρ(a)x = a2 x and defines an isomorphism
The agreable thing about these semidirect products G = G0 N, N abelian, is the fact
that Mackey’s Theory provides an explicit way to construct their irreducible unitary
representations as induced representations. We describe this method without going ev-
erywhere into the intricacies of the proofs, which rely on some serious functional analysis:
150 7. Induced Representations
π((a, n)) = π((e, n)(a, 0)) = π((e, n))π((a, e)) = π (n)π (a).
we deduce
Ôn̂ := Gn̂.
Mackey’s construction works under the assumption that the semidirect product is regu-
lar. All the groups we treat as examples are of this type. Hence, since we will not give
complete proofs anyway, a not so ambitious reader at first sight may skip the following
rather technical condition (from [BR] p.506).
One also has a slightly different looking condition for regularity used by Mackey ([Ma]
p.77):
There is a subset J ⊂ N̂ , which meets each G−orbit Ô exactly once and is an analytic
subset.
Now as our central prescription for the construction of representations in this section, we
take over from [BR] Theorem 4 on p.508:
It is quite evident that this recipe produces unitary representations. But it is far from
evident that these are irreducible and that we really get the unitary dual of the semidi-
rect product and, hence, have a classification scheme at hand. As already said, we can
not go into the proofs here (which essentially use Mackey’s imprimitivity theorem) and
refer to the Chapters 16 and 17 of [BR] or §13 of [Ki] and Section 10 of [Ma] where
slightly different approaches are adopted. Now, we discuss some examples to show how
the method works. In these examples one can verify directly the validity of the procedure
encapsulated in Theorem 7.7 for instance with the help of the infinitesimal method from
our Chapter 6.
Examples:
• Type 1: For m = 0 and n̂0 = (0, m) one has the line Ôn0 = {(−xm, m), x ∈ R},
• Type 2: For m = 0 and n̂0 = (l, 0) there is the point Ôn0 = {(l, 0)}.
r r r r r r r r r r-
r l
Ô(l,0)
One can see that G is a regular semidirect product: The set {(l, m) ∈ R2 ; lm = 0} is an
analytic set, which meets each orbit exactly once.
– Step 3: The stabilizing group of n̂0 = (0, m), m = 0 is H0 = G0(0,m) = {0} and that
of n̂0 = (l, 0), l ∈ R is H0 = G0(l,0) = G0 R.
– Step 4: In the first case the unitary dual of H0 simply consists of the trivial identity
representation and in the second case we have Ĥ0 R with representatives given by
π1 (a) = eijx =: χj (x), j ∈ R.
– Step 5 and 6: In the second case, the type 2 orbits, the product of the stabilizing
group H0 = G0 with N already exhausts the whole group G. Hence there is no induction
necessary and one has for (j, l) ∈ R2 a one-dimensional representation π of G = G0 N
given by
π((x, y, z)) = eijx eily .
In the first case, belonging to the type 1 orbits, one has H = H0 N N R2 with a
representation π0 given by π0 (h) = eimz for all h = (0, y, z). Here the induction makes
sense and we have a representation π = indG H π0 (for instance) given by right translation
on the completion H of the space of smooth functions φ : G −→ C with functional
equation
φ(hg) = eimz φ(g) for all h ∈ H, g ∈ G,
and φ2 = R | φ((x, 0, 0) |2 dx < ∞. As the composition law says
an element φ in H is of the form φ(g) = eimz f (x) with f ∈ L2 (R) and one has
This is the Schrödinger representation πm from our Example 1.6 in 1.3 translated to
Heis (R) by the isomorphism
from Exercise 0.1. Hence the unitary dual of the Heisenberg group G is
Ĝ = R2 (R \ {0}).
7.4 Unitary Representations of Semidirect Products 153
Ôn0 = {y ∈ R2 , y2 = r2 }.
We can verify the regularity of the semidirect product: One has a countable family of
Borel subsets of N̂ consisting of the following sets: Z0,0 := the point Ôn0 = {(0, 0)} and
for any two positive rational numbers r1 < r2 Zr1 ,r2 := the union of all orbits of type 2
with r1 < r < r2 . Then each orbit is the intersection of the members of the subfamily,
which contain the orbit.
– Step 3: For the type 1 orbits the stabilizing group is G0(r,0) = {E2 } and for the type
2 one has G0(0,0) = G0 .
– Step 4: The unitary dual of G0 = SO(2) is Z.
– Step 5 and 6: In the second case, the type 2 orbit, the product of N and the stabilizing
group is G. Hence for every k ∈ Z, we have a one-dimensional representation π of G given
by
π((r(ϑ), x)) = eikϑ .
In the first case, the type 1 orbit, one has for r > 0 the genuine subgroup H N and
its one-dimensional representation π0 given by
Ĝ = R>0 Z.
Example 7.13: As Exercise 7.22 apply our construction recipe to construct represen-
tations of
a b
G = B := {g = ( ) =: (a, b); a, b ∈ R, a > 0}
a−1
and
a b
G = B := {g = ( ) =: (a, b); a, b ∈ R, a = 0}
a−1
and determine the unitary dual.
Exercise 7.23: Do the same for the Euclidean group in three dimensions GE (3).
154 7. Induced Representations
E3
| x |2 := x21 + x22 + x23 − x24 = t x( )x, for all x = t (x1 , x2 , x3 , x4 ) ∈ R4 .
−1
We restrict our treatment to a discussion of the simplest case (it should not be too
difficult to extend our results to the more general cases). It should also be not difficult
to translate our procedure to the situation apparently prefered by the physics literature
([Co] p.677ff, [BR] p.513ff]), namely that the coordinates are x0 , x1 , x2 , x3 and the norm
is given by | x |2 = x20 − x21 − x22 − x23 . Moreover, there are reasons from physics (see the
hints in 4.3 while discussing the relation of the representations of SO(3) and SU(2)) and
mathematics to define and treat the semidirect product
G̃ := SL(2, C) R4 .
For the definition of this product, one goes back to the proof of Theorem 7.6 in 7.3
where we fixed the surjection ρ : SL(2, C) −→ SO(3, 1): There we had an identification
ϕ : R4 −→ V = Herm2 (C) and defined an action of SL(2, C) on V (and hence on R4 )
by
ρ(Ã)X = ÃX Ã∗ for all X ∈ V, Ã ∈ SL(2, C).
We start with the discussion of G = G0 N, G0 = SO(3, 1)0 , N = R4 (say, for aesthetical
reasons) but soon shall be lead to take in also G̃. We follow the construction procedure
from Theorem 7.7 in the last section.
(In the physics literature often minus signs appear in this identfication of N and its dual,
but we want to avoid the discussion of co- and contravariant vectors at this stage.)
i.e. A transforms p into pA−1 . It is elementary calculus that the orbits for this action
are subsets in the three types of quadrics
More precisely, one verifies using the explicit description of SO(3, 1) in 7.3 that R4 de-
composes exactly into the following orbits:
• Type 1+ : +
Ôm := Ôn̂0 = G0 n̂0 for n̂0 = (0, 0, 0, m),
− −
• Type 1 : Ôm := Ôn̂0 = G0 n̂0 for n̂0 = (0, 0, 0, −m),
• Type 2 : Ôim := Ôn̂0 = G0 n̂0 for n̂0 = (m, 0, 0, 0),
• Type 3 :
+
Ô0+ = Ôn̂0 = G0 n̂0 for n̂0 = (−1, 0, 0, 1),
−
• Type 3 : Ô0− := Ôn̂0 = G0 n̂0 for n̂0 = (−1, 0, 0, −1),
• Type 3 :
0
Ô00 := Ôn̂0 = G0 n̂0 = {(0, 0, 0, 0)} for n̂0 = (0, 0, 0, 0),
−
The type 1+ orbit Ôm+
is called the positive mass shell, Ôm the negative mass shell and
±
Ô0 the forward resp. backward light cone. The situation is sketched in the picture:
p4
+
Ôm 6
@
@
@
@ Ô0+ (0, m)
@
@ (m, 0)
- p1
@
Ôim @
Ô0− @
@
@
− @
Ôm
The regularity of the semidirect product can be verified analogously as in our example 2
in the last section (see [BR] p.517/8).
Now, for each orbit Ô, we have to determine its little group, i.e. the stabilizing group
H0 := G0 n̂0 .
B x
A=( t ), B ∈ SO(3), x, y ∈ R3 , a ∈ R.
y a
ii) The other assertions can be proved similarly by direct computations. But as an-
nounced above, it is more agreable to use the fact from 7.3 that one has SL(2, C) =: G̃0
via the surjection ρ : G̃0 −→ G0 as a double cover of the Lorentz group. In the proof
of Theorem 7.6 in 7.3 we identified R4 via a map ϕ with the space V of hermitian
2 × 2−matrices
−a + d b + ic
t
(a, b, c, d) −→ ( ),
b − ic a + d
and defined the map
i.e. Ã ∈ SU(1, 1). It is not difficult to see (Exercise 7.24) that the homomorphism
ρ : SL(2, C) −→ SO(3, 1)0 restricts to a surjection SU(1, 1) −→ SO(2, 1)0 .
By a small calculation, one verifies that the stabilizing group {Ã; ÃX0 Ã = X0 } here
i
is SL(2, R): For σ = ( ) one has σ = σ −1 and t A−1 = σAσ −1 . Hence from
−i
AσA∗ = σ we deduce σ −1 Aσ = A∗ −1 = t A−1 , i.e. A = Ā.
2
iii) One has ϕ(−1, 0, 0, 1) = ( ) =: X0 , and ÃX0 Ã = X0 calls for
a b
à = ( ), with a, b ∈ C, | a |= 1.
a−1
We put a =: eiϑ/2 , b =: e−iϑ/2 z, ϑ ∈ [0, 4π], z ∈ C, and get the multiplication law
a b
ÃÃ = ( −1 ), with a = ei(ϑ+ϑ )/2 , b = e−i(ϑ+ϑ )/2 (z + eiϑ z ).
a
7.5 Unitary Representations of the Poincaré Group 157
This shows that in this case the stabilizing group H̃0 ⊂ G̃ is isomorphic to the semidirect
product of a rotation group {eiϑ } and a translation group R2 C with the composition
law
(eiϑ/2 , z)(eiϑ /2 , z ) = (ei(ϑ+ϑ )/2 , z + eiϑ z ).
The homomorphism (eiϑ/2 −→ (r(ϑ)) makes S̃1 := {eiϑ/2 ; 0 < ϑ < 4π} a double cover
of SO(2) and this extends to a surjection of H̃0 to the Euclidean group GE (2).
Step 4: We already determined the unitary dual of covering groups H̃0 of all the little
groups coming in here:
i) In the first case (type 1± ), for H̃0 = SU(2), the elements of the unitary dual are rep-
resented by the representations πj , j ∈ (1/2)N0 discussed in 4.2.
ii) In the second case (type 2), one has H̃0 = SU(1, 1) and from 6.2 we know that
SU(1, 1) is isomorphic to SL(2, R) and that the representations appear in three series
πk± , k ∈ N; πis,± , s ∈ R; πs , 0 < s < 1.
iii) For the third case (type 3± ), as Example 7.12 to our construction procedure in 7.4 we
elaborated that for the two-dimensional Euclidean group one has one-dimensional repre-
sentations χk , k ∈ Z (hence χj , j ∈ (1/2)Z for the double cover) and infinite-dimensional
representations πr , r > 0.
iv) For the type 30 , one has the representations of the Lorentz group resp. its covering
SL(2, C) discussed in 7.3, which come in two series πk,iv , k ∈ Z, v ∈ R; π0,u , 0 < u < 2.
Step 5 and 6: Finally, we have to extend the representations of the little group to its
product with N = R4 and, if this product is not the whole group, proceed to an induced
representation. This way we get the following representations (as usual, we list those of
the covering group G̃ of the Poincaré group):
i) In the first case, one has H̃ = SU(2) R4 . For the type 1+ with m > 0, we have for
this group the representation π0 given by
indG̃ π =: πm,+;j .
H̃ 0
indG̃ π
H̃ 0
=: πim;k,± for π = πk± , k ∈ N; m > 0,
=: πim;is,± for π = πis,± , s ∈ R; m > 0,
=: πim;s for π = πs , s ∈ R, 0 < s < 1; m > 0.
indG̃ π
H̃ 0
=: π0,±;j for π = χj , j ∈ (1/2)Z,
=: π0,±;r for π = πr , r ∈ R, r > 0.
iv) In the forth case (for type 30 ) the little group is the whole group and we get the
representations of G̃ by the trivial extension of the representations π of SL(2, C), i.e. one
has π((Ã, x)) = π(Ã) for (Ã, x) ∈ G̃0 . For the corresponding representation we write
π =: π0;j,iv,± for π = πj,iv,± , v ∈ R
=: π0;0,u for π = π0,u , u ∈ R, 0 < u < 2.
Among these, the representations πm,+;j , m > 0, j ∈ (1/2)N0 , coming from the type 1+
orbit and π0,±;j , j ∈ (1/2)Z, coming from the type 3± orbit have a meaning for physics
since they are interpreted as describing elementary particles of spin j and mass m resp.
of zero mass.
Via ϕ the action of G̃0 on V given by (Ã, X) −→ ρ(Ã)X = ÃX Ã∗ is translated into the
action of G̃0 on R4 given by
By a slight abuse of notation, as we already did above, we also take ρ(Ã) to be a matrix
from G0 = SO(3, 1)0 and understand ρ(Ã)x as multiplication of the column x by the
t
matrix ρ(Ã). We identify the unitary dual N̂ with R4 by writing n̂(n) = ei px , i.e. n̂ is
coordinatized by (the row) t p, p ∈ R4 . Then the action of G̃0 on N̂ is simply given by
the multiplication of the row t p by the matrix ρ(Ã−1 ), i.e.
t
p −→ t pρ(Ã−1 ) =: Ã · p.
To get some more practice and since it is important for application in physics, we work
out the “ Second Realization” of the induced representation indG̃ π , i.e. with a repre-
H̃ 0
sentation space consisting of functions u living on the homogeneous space
(At this moment, one usually does not bother with any finiteness or integrability condi-
tion and simply takes smooth functions. We shall have occasion to do better in the next
chapter.)
The general theory of the Iwasawa decomposition or a direct calculation say that we have
a decomposition
λ z
G̃0 = SL(2, C)
à = Ap Rp , Rp ∈ SU(2), Ap = ( ), λ ∈ R∗ , z ∈ C.
λ−1
G̃
g = s(p)h, s(p) = (Ap , 0), h = (Rp , x) ∈ H̃
with
with
t
π0 (h∗ −1 ) = ei px0
πj (R∗ −1 )
t
as one can verify by a small calculation starting from π0 ((R, x)) = ei p0 x πj (R): Namely,
one has
h∗ −1 = (R∗ −1 , x̃) with x̃ = −ρ(R∗ −1 )x∗
and hence
t
p0 x̃ = −t p0 ρ(R∗ −1 )x∗ = t p0 ρ(R∗ −1 (Ag−1 (p) )−1 A0 )x0
0
i.e.
t
p0 x̃ = t p0 ρ(A−1 t
p )x0 = px0
−1
since we have R∗ = (Ag−1 (p) ) A−1
0 Ap .
0
elements of the representation πj from our discussion in 4.2 and then come to
t
∗ −1
(7.24) (π((Ã0 , x0 ))u)n (p) = ei px0 j
Dnn (R )un (Ã−1
0 · p).
n
From here one derives the following interpretation that we have a free particle of spin j
and mass m :
In the rest system p = p0 and under rotations (R, 0), R ∈ SU(2) (7.24) restricts to
j
(π̃((R, 0))un )(p0 ) = Dnn (R)un (p0 ),
n
Pµ un (p) = pµ un (p).
Hence we have M un = mun , i.e. un is an eigenfunction with eigenvalue m for the mass
operator
M := P42 − P32 − P22 − P12 .
7.6 Induced Representations and Vector Bundles 161
(In this example everything is unimodular, hence it does not matter whether we treat un-
or normalized induction, but in general several authors stick to unnormalized induction
though one can also treat normalized induction as we shall see below.) For V = C we
introduce as our first example of a line bundle
G acts on Bk by
g0 [g, t] := [g0 g, t] for all g0 ∈ G, [g, t] ∈ B k .
Obviously the projection and the action are well defined and one has the fiber over x
This leads to the picture that Bk is a bundle standing over the base space X and consisting
of the fibers (B k )x over all base points x ∈ X .
Our main object is the space Γ(B k ) of sections of Bk , i.e. maps
ϕ : X −→ B k with pr ◦ ϕ = idX .
These sections are of the form ϕ(x) = [g, φ(g)] for x = gH with a function φ : G −→ V .
Since the coset gH may be represented also by gh, h ∈ H, the welldefinedness of the
section asks for
[g, φ(g)] = [gh, φ(gh)] = [gh, π0 (h−1 )φ(g)]
where the last equation is expression of the equivalence introduced in the definition of
the bundle Bk . This leads to the fact that the function φ has to fulfill the functional
equation
φ(gh) = π0 (h−1 )φ(g) for all h ∈ H, g ∈ G,
which we know from the construction of the representation space for the induced repre-
sentation. We define an action of G on the space Γ(Bk ) of sections by
Hence at least formally, we recover the prescription for the definition of the induced
representation π. This can be completed to a precise statement if we provide our bundle
with an appropriate topology. And this appears to be largely independent of the special
example G = SU(2), which we proposed at the beginning. In the following, we will try
to show how this works by giving some rudiments of the theory of bundles. We start
by following the treatment of Warner [Wa] p.377, which is adapted to the application in
representation theory, and proceed to another more general approach afterwards.
One calls X the base space, B the bundle space, and p the bundle projection. By abuse
of notation the pair (B, p) is often abbreviated by the letter B to indicate the bundle.
For x ∈ X , the set p−1 (x) is called the fiber over x and often denoted by Bx . A cross
section or simply section of the bundle is a map s : X −→ B with p ◦ s = idX .
A Hilbert bundle B over X is a bundle (B, p) over X together with operations and a norm
making each fiber Bx , x ∈ X into a complex Hilbert space and satisfying the following
conditions:
Example 7.15: Let E be any Hilbert space. The trivial bundle with constant fiber E
is given by B := X × E equipped with the product topology and the projection p given
by p(x, a) = x for x ∈ X , a ∈ E. The cross sections are given by the functions s : X −→ E.
Let B = (B, p) be a Hilbert bundle over X and Γ(B) the set of continuous cross sections
of B . Then it is clear that Γ(B) is a complex vector space under pointwise addition
and scalar multiplication. In the application to representation theory this space shall
provide the representation space of a representation. So we need a topology on it. And
here things start to be a bit technical. Following [Wa], we assume that X is locally
compact and for Γ(B) define the uniform on compacta topology. For our purposes it is
sufficient that in particular this leads to the consequence that each sequence (sn ) in Γ(B)
converges to a continuous cross section ϕ iff for each compact subset C ⊂ X one has
supx∈C sn (x) − s(x) −→ 0 for n −→ ∞. (In [Wa] the sequence is replaced by the more
general notion of net but we will not go into this here.)
The integration theory in this context is still more delicate. So for details, we also refer
to [Wa]. For our purpose it will suffice to fixa measure µ on X and think of Lq (B, µ)
as the space of sections s of B with sq := ( X s(x)q dµ(x))1/q < ∞. Then the inner
product on the space of sections is given by
< s, t > := < s(x), t(x) > dµ(x) for all s, t ∈ L2 (B, µ).
X
Let G be a linear group, H a closed subgroup, δ the quotient of the modular functions
of G and H treated in 7.1 (normalized by the condition δ(1) = 1) and π0 a unitary
representation of H on the Hilbert space V .
The orbit under H of a pair (g, a) ∈ G × E will be denoted by [g, a] := (g, a)H. The
space of all such orbits equipped with the quotient topology will be called E(G, H, π0 )
or simply E. This space is Hausdorff and the natural projection map
Obviously, there are some verifications to be done. Whoever does not want to do these
on his own can look into [Wa] p.380/1. The bundle is called the Hilbert bundle induced
by π0 (and δ) and denoted by E(π0 ) or again E. G acts continuously from the left on E
by
g0 [g, a] := [g0 g, a].
Finally, we rediscover the normalized induced representation π = indG H π0 of G as a
representation by unitary operators on the space of square integrable sections of E: To
this end, one first observes that there exists a one-to-one correspondence between cross
sections ϕ for E = E(π0 ) and functions φ : G −→ V fulfilling the functional equation
If φ and ϕ are connected in this way, one says that φ is the unitary form of ϕ and ϕ is the
bundle form of φ. Evidently a cross section ϕ is continuous on X = G/H iff its unitary
form φ is continuous on G. It is a bit less evident but true (see [Wa] p.381) that E(π0 )
admits enough cross sections. Therefore it makes sense to consider the Hilbert space
Hπ0 := L2 (E, µ)
Some routine verifications parallel to those we did in 7.1 (see [Wa] p.382) show that we
have got a unitary representation equivalent to the one constructed in 7.1. And one
7.6 Induced Representations and Vector Bundles 165
should get the feeling that all this is realized by the example of the (complex!) bundle
E = B k over X = P1 (C) from the beginning of this section. We will say more when we
come back to this at the end of the section.
Here we need the notion of a differentiable manifold, which already was alluded to in 6.2
and is to be found in most text books on advanced calculus. We recommend in particular
the Appendix II in [Ki1] (or the Appendix A in [Be]).
ϕi,j := ϕi ◦ ϕ−1
j |ϕj (Ui ∩Uj )
– the pairs (Ui , ϕi ) (and often also the sets Ui ) are called charts;
It has to be clarified which atlases are really different: We say that two atlases (Ui )i∈I
and (U j )j∈J are called equivalent if the transition functions from any chart of the first
atlas to any chart of the second one are smooth (resp. analytic). The structure of a
smooth (resp. analytic) manifold on M is an equivalence class of atlases.
When dealing with a smooth manifold, we use only one atlas keeping in mind that we
can always replace it by an equivalent one.
In parallel to the just given notions for real manifolds, one defines a smooth (or holo-
morphic) complex p-manifold where Rp is replaced by Cp and the transition functions
are required to be holomorphic, i.e. complex differentiable.
166 7. Induced Representations
Examples:
Example 7.18: M = P1 (C) (as in Example 7.3 in 7.1.5 and the beginning of this
section) is a smooth one-dimensional complex manifold:
We cover P1 (C) by two open subsets U := {(u : v); u, v ∈ C, v = 0} and U := {(u :
v); u, v ∈ C, u = 0}. We have the homeomorphisms
ϕ : U −→ C, (u : v) −→ u/v =: w ∈ C,
(7.25) ϕ : U −→ C, (u : v) −→ v/u =: w ∈ C.
And the map ϕ ◦ ϕ |{w=0} (w) = 1/w = w is complex differentiable. So here one has an
atlas consisting of two charts U and U .
Certainly, at the same time, P1 (C) is an example for a two-dimensional real manifold.
with (u1 , . . . , up+1 ) ∼ (u 1 , . . . , u p+1 ) iff there exists a λ ∈ R∗ with u i = λui for all i.
A manifold is called oriented iff it has an oriented atlas, i.e. an atlas where for any two
charts with coordinates x and y the determinant of the functional matrix ( ∂x
∂y ) is positive
whenever it is defined.
As already said, there is much more to be discussed in this context (and we refer again to
e.g. [Ki1]) but, we hope, this rudimentary presentation suffices to grab an understanding
of our second approach to bundles:
7.6 Induced Representations and Vector Bundles 167
Definition 7.4: E is called a (fiber) bundle over the base X with fiber F iff E, F and X
are smooth manifolds and there is given a smooth surjective map p : E −→ X such that
locally E is a direct product of a part of the base and the fiber F .
More precisely, we require that any point x ∈ X has a neighbourhood Ux such that
EUx := p−1 (Ux ) can be identified with Ux ×F via a smooth map hx so that, for m ∈ EUx ,
one has hx (m) = (x , v) if x = p(m), i.e. the following diagram (where p1 denotes the
projection to the first factor) is commutative
h
p−1 (Ux ) −→
x
Ux × F
↓p ↓ p1
∼
Ux −→ Ux .
From the definition it follows that all sets Ex := p−1 (x), x ∈ X , called fibers, are smooth
submanifolds diffeomorphic to F .
F
One often writes E −→ X to denote a bundle with fiber F and base X .
F
Let E −→ X and E −→ X be
F
There is the natural notion of a map between bundles:
two bundles. Then a bundle map is given by the pair of smooth maps
Φ : E −→ E , f : X −→ X
E
Φ
E −→
p↓ ↓ p
f
X −→ X .
Two bundles over the same base X are called equivalent if there is a bundle map with
f = idX and Φ a diffeomorphism.
F F
A bundle E −→ X is called a trivial bundle if it is equivalent to the bundle X × F −→ X .
We are mainly interested in vector or line bundles where the fiber is F := V a vector
space (with dim V = n = 1 for the case of a line bundle) and the maps hx are linear in
each fiber p−1 (x ), x ∈ Ux .
Primarily, we think of real vector spaces and real manifolds but we also will treat complex
cases.
In any case, the dimension of the fiber as a vector space is called the rank of the vector
bundle.
168 7. Induced Representations
To make this more explicit and show a way to construct bundles, we take an atlas
of the real manifold X given by open sets Ui (i ∈ I), and V = Rn . Then we have
homeomorphisms
h := hi : EUi = p−1 (Ui ) −→ Ui × Rn ,
which a) respect the fibers, i.e. with
h |Ex : Ex −→ Rn .
We denote by
hi,j := hi ◦ h−1
j |(Ui ∩Uj )×Rn : (Ui ∩ Uj ) × R −→ (Ui ∩ Uj ) × R
n n
the bundle transition maps and introduce (differentiable) functions, which are called
bundle transition functions
by
hi,j (x, v) = (x, ψi,j (x)v) for all (x, v) ∈ (Ui ∩ Uj ) × Rn .
It is not difficult to see that these ψi,j are uniquely determined and for i, j, k with
Ui ∩ Uj ∩ Uk = ∅
Here, our main point is that we can construct bundles with fiber Rn over a manifold
with atlas (Ui )i∈I by pasting together columns Ui × Rn via such a given cocycle system
ψi,j . We have the following construction principle, whose validity is not too difficult to
prove.
Proposition 7.4: Let X be a manifold with atlas (Ui )i∈I and ψi,j : Ui ∩Uj −→ GL(n, R)
be a system of smooth functions with
ψi,i = 1 in Ui ,
ψi,j ◦ ψj,i = 1 in Ui ∩ Uj ,
ψi,j ◦ ψj,k ◦ ψk,i = 1 in Ui ∩ Uj ∩ Uk .
Let Ẽ be the union of the manifolds Ui × Rn . One says that points (xi , vi ) ∈ Ui × Rn
and (xj , vj ) ∈ Uj × Rn are equivalent if xi = xj and vi = ψi,j (xj )vj . Then the factor
space E of Ẽ defined by this equivalence relation is a vector bundle with fiber Rn over X
with projection p induced by the natural projections (xi , vi ) −→ xi . And each such fiber
bundle is equivalent to a bundle of this type.
7.6 Induced Representations and Vector Bundles 169
Proposition 7.5: Let E and E be vector bundles of the same rank n over X defined
by the systems of bundle transition functions ψi,j and ψi,j (without loss of generality,
by a refinement argument, one can assume that both bundles are defined by the same
atlas (Ui )i∈I ). Then both bundles are equivalent iff there is a family of matrix functions
fi : Ui −→ GL(n, R) such that one has
ψi,j = fi−1 ψi,j fj for all i, j ∈ I.
As in our first approach, for open sets U ⊂ X we introduce (vector) spaces Γ(U, E) of
sections s over U , i.e. smooth maps s : U −→ E with p ◦ s = idU . For U = X we talk of
global sections.
Parallel to our treatment of real bundles, one can look at the complex case by replacing
appropriately R by C.
Examples:
Example 7.20: The most prominent example, which led to the introduction of the
notion of bundles, is the tangent bundle T M , which comes about if for an n-dimensional
manifold M with charts (Ui , ϕi ), i ∈ I, as matrix of bundle transition functions one takes
the Jacobian J of the manifold transition functions ϕi,j
n
∂yi
bi = (y)aj
j=1
∂x j
when m appears in the chart (U, ϕ) with coordinates x = (x1 , . . . , xn ) and simultane-
ously in the chart (U , ϕ ) with coordinates y = (y1 , . . . , yn ). This has a geometrical
background: in first approximation, we take the manifold M and to each point m ∈ M
associate as a fiber Em the tangent space Tm M to M at this point m. From higher
calculus one knows that there are different ways to describe the elements Xm ∈ Tm M
i.e. the tangent vectors to a manifold. We follow the one already prepared in 6.2 while
discussing the definition of the Lie algebra associated to a linear group and think of the
tangent space of M at m as the space of all tangent vectors of all smooth curves γ inM
passing through m. More precisely, a tangent vector Xm to M at the point m is defined
as the equivalence class of smooth parametrized curves γ = γ(t) passing through m. Two
curves γ and γ̃ in M with γ(0) = γ̃(0) = m are equivalent iff their tangent vectors in m
coincide in the coordinates of a chart (U, ϕ) with m ∈ U , i.e. one has γ ∼ γ̃ if there is a
chart (U, ϕ) with m ∈ U and coordinates ϕ(m) = x such that
d d
ϕ(γ(t)) |t=0 = ϕ(γ̃(t)) |t=0 .
dt dt
With respect to this chart, a tangent vector can be symbolized by the n-tupel of real
d
numbers ai wherenai is just the i-th component of dt ϕ(γ(t))|t=0 or by the differential
operators Xm = i=1 ai ∂xi acting on smooth functions f in the coordinates x. Addition
and scalar multiplication of these operators resp. tuples define a vector space structure,
which can and will be carried over to the first given description of the space Tm M .
170 7. Induced Representations
If (U , ϕ ) with coordinates y is another chart for m and (b1 , . . . , bn ) are the components
of the tangent vector to γ expressed in this coordinate system, by the chain rule we have
with ϕ (ϕ−1 (x)) = y(x)
d d
∂yi
bi = ( ϕ (γ(t)) |t=0 )i = ( (ϕ (ϕ−1 (ϕ(γ(t)))) |t=0 )i = aj ,
dt dt j
∂xj
i.e. the contravariant transformation property from our general construction above.
Let (U, ϕ) be a chart with coordinates x = (x1 , . . . , xn ). With a (now usual) slight misuse
of notation we write
n
X |U =: ai (x)∂xi
i=1
where the coefficients ai are smooth functions defined in U . For a chart (U , ϕ ) with
coordinates y = (y1 , . . . , yn ), one has as well
n
X| = U bj (y)∂yj
j=1
∗
Example 7.21: The cotangent bundle to M has as its fibers the cotangent spaces Tm M,
i.e. the duals to the tangent spaces Tm M . It turns up if, in the same situation as above
in Example 7.20, as matrix of bundle transition functions we take the transpose of the
Jacobian of the map ϕj ◦ ϕ−1i =: ϕj,i , i.e.
Example 7.22: The first two examples concerned real bundles but similarly one can
treat the complex case. Here we take the one-dimensional complex manifold M = P1 (C)
covered by the two charts (U, ϕ) and (U , ϕ ) fixed above in (8.10) while discussing M
as a manifold. For k ∈ Z we get a line bundle Lk by choosing ψ(w) = w−k as bundle
transition function between these two charts. Let us work out the construction of this
bundle along the line of the general construction principle given above: We have two
“columns”, one over U and one over U , namely
w = 1/w, t = w−k t.
Remark 7.9: It can be shown that up to (holomorphic) equivalence there are no other
holomorphic line bundles over P1 (C) than these Lk .
s : (u : v) −→ ∈ Lk ,
where h() = (w, t) ∈ L for w = u/v, v = 0, and h () = (w , t ) ∈ L for w = v/u, u = 0.
The section is holomorphic iff t is given by a holomorphic function in w and t by a
holomorphic function in w and we have the compatibility condition t = w−k t. As
examples and prototypes for such a holomorphic function we take the monomials t = wp
p
and t = w with p, p ∈ N0 . The compatibilty condition says for w = 0
p
t (w ) = w = wk t(w) = w−k+p
and, since one has w = 1/w, we get w−p = wp−k and see that everything is consistent
for p = 0, 1, . . . , k if k is not negative. And the holomorphicity of the relation breaks
down for negative k. Hence we have k + 1 essentially different holomorphic sections for
non-negative k and none for negative k. One needs a bit more complex function theory
to be sure that there are no more (see for instance [Ki] p.80).
After this short digression into the world of bundles, which will be useful later, we return
to the problem of the construction of group representations. As we know from our first
approach to bundles, the space of sections is meant to provide a representation space
and, hence, should get the structure of a Hilbert space. In general this needs some
work, which we shall attack in the next chapter. But if the space of sections is a finite-
dimensional vector space, this is no problem. To get an idea we return to our example
G = SU(2) ⊃ H U(1) and the unitary representation πj = indG H π0 induced from the
representation π0 = χk , k = 2j ≥ 0, of H U(1). In our first approach we constructed
172 7. Induced Representations
the bundle space Bk = {[g, t]; g ∈ G, t ∈ C} with projection pr onto X = G/H = P1 (C)
given by
a b
pr([g, t] := gH = (u : v) = (b : ā) for g = .
−b̄ ā
We define a map of Bk to the line bundle Lk constructed by pasting together the sets
L = {(w, t); w, t ∈ C} and L = {(w .t ); w , t ∈ C} via the bundle transition function
ψ(w) = wk as follows: For g with a = 0 we put
B k
[g, t] −→ (w = b/ā; b̄k t) ∈ L
and
π0 (h) = ζ k .
Hence application of the mapping prescription to the element [gh−1 , π0 (h)t], which as
well represents [g, t], leads to the same images. And the images in L and L are consistent
with the pasting condition (w , t ) = (1/w, w−k t) for w = 0, since with w = b/ā one has
w−k (bk t) = āk t. Thus both maps induce a map of Bk to Lk , which obviously is a bijection.
Here we have a complex line bundle over the one-dimensional complex manifold P1 (C)
whose space of holomorphic sections is a complex vector space and the representation
space for the representation πj , 2j = k of G = SU(2). In principle, the base spaces for
the bundles in our first approach are real manifolds and the section spaces real vector
spaces. This example shows that we may get the wanted irreducible representation by
some more restrictions, namely we put more structure on the real construction to get the
more restrictive complex case (as in holomorphic induction). We shall see more of this
under an angle provided from the necessities of physics in the next chapter.
Chapter 8
The orbit method is an object of pure mathematics, but it got a lot of impetus from
physics under the heading of geometric quantization. Historically it was proposed by
Kirillov in [Ki2] for the description of the unitary dual of nilpotent Lie groups. By fur-
ther work of Kirillov and many others, in particular B. Kostant, M. Duflo, M. Vergne and
D. Vogan, it grew into an important tool for explicit constructions of representations of
various types of groups. The construction mechanism uses elements that also appear in
theoretical physics when describing the transition from classical mechanics to quantum
mechanics. We therefore start by recalling some of this background, even though we
cannot give all the necessary definitions.
dqi ∂H dpi ∂H
(8.1) q̇i := = , ṗi := =−
dt ∂pi dt ∂qi
where the function H = H(q, p, t) is the Hamiltonian of the system, i.e. encodes all the
relevant information about the system. For instance, if our system is simply a mass one
(point) particle in R3 moving
under the influence of the force with a radial symmetric
potential V (q) = 1/r, r = q12 + q22 + q32 , one has H = p21 + p22 + p23 + V (q).
174 8. Geometric Quantization and the Orbit Method
X = a(q, p)∂q + b(q, p)∂p := a1 (q, p) ∂q1 + . . . an (q, p) ∂qn + b1 (q, p) ∂p1 + · · · + bn (q, p) ∂pn
iff at each point of the curve its tangent vector is a vector from the vector field. We write
this as
γ̇(t) = X(γ(t)).
The statement that γ is an integral curve for the Hamiltonian vector field
∂H ∂H
X = XH with ai = , bi = − ,
∂pi ∂qi
To state this, we have to provide some more general notions about differential forms,
which we will need anyway in the following sections. Hence we enlarge our considera-
tions about the tangent and cotangent bundle and their sections from the last section:
We look again at a (real) smooth n-manifold M .
– A vector field X is defined as a section of the tangent bundle T M and in local coordi-
nates x = (x1 , . . . , xn ) written as
As one knows from calculus, there also are higher order differential forms of different
types. We shall need the symmetric and the antisymmetric or exterior forms.
– A symmetric r-form β is written in the local coordinates from above as
where the coefficients b are (smooth) functions of the coordinates x and the differentials
dxi are commuting symbols, i.e. with dxi · dxj = dxj · dxi for all i, j, the sum is meant
over all r-tuples i1 , . . . , ir ∈ {1, . . . , n} and the coefficients b are invariant under all
permutations of the indices. For instance, if n = 2, one has the two-form
β = b11 dx21 + b12 dx1 dx2 + b21 dx2 dx1 + b22 dx22 , b12 = b21 .
Certainly, there is the possibility to introduce a reduced form where one restricts the
index-tuples by taking exactly one representative from each orbit of the symmetric group
Sr acting on the index-tuples. Hence, as reduced form one has in the example above
If one wants to be more sophisticated (and precise), one says that a symmetric r−form
is a (smooth) section of the r−th symmetric power of the cotangent bundle. But for our
treatment here and for the antisymmetric case below the naive description above should
be sufficient.
α= ai (x)dxi1 ∧ · · · ∧ dxir
where here the differentials are anticommuting, i.e. with dxi ∧ dxj = −dxj ∧ dxi for all
i, j. The summation is again over all r−tuples i = (i1 , . . . , ir ) and the coefficients have
the special property of antisymmetry
hence, in particular one has ai = 0 if two of the indices coincide. If we take in our
example n = r = 2, we get
It is very convenient to introduce here the reduced form where the sum is taken only over
all r−tuples (i) = (i1 , . . . , ir ) with 1 ≤ i1 < · · · < ir ≤ n. Hence the reduced form of our
example is
α = ã12 dx1 ∧ dx2 , ã12 = 2a12 .
After changing coordinates the differential forms are transformed according to an ap-
propriate generalization of the transformation property (7.27) in Example 7.21 in 7.6
and the complete definition is again given by characterizing exterior r−forms as smooth
sections of the r−th exterior power of the cotangent bundle.
n
∂a
da := dxi .
i=1
∂xi
To handle the exterior differential, one usually reorders it to its reduced form.
176 8. Geometric Quantization and the Orbit Method
A simple but most important example appears if we look at the coordinates given by
n
x = (q1 , . . . , qn , p1 , . . . , pn ) and take the one-form α = i=1 qi dpi . We get
n
(8.2) ω0 := dα = dqi ∧ dpi
i=1
A form α is called closed iff one has dα = 0 and it is called exact iff there is a form β
such that α = dβ.
Moreover, we point to the fact that exterior differentiation commutes with substitution
of the variables and, if Φ : M −→ M is a smooth map of manifolds and α a form on M ,
then one can draw back α to a form Φ∗ α on M . Without being more pecise here, for
∂yi of the local coordinates x −→ y = y(x)
practical purposes this comes out by substitution
and using the well known formula dyi = j ∂x j
dxj . We do not have enough space to go
into this more deeply but later on shall see in examples how this works.
For α = a(i) (x)dxi1 ∧ · · · ∧ dxir and the vector field X = bj (x)∂j , we have an inner
product i(X)α, which comes out (this certainly is not the most elegant definition, but
avoids more generalities) by applying successively the prescriptions
i(X)dxi = bi (x)
and
i(X)(α ∧ β) = (i(X)α) ∧ β + (−1)r α ∧ (i(X)β) for α ∈ Ωr , β ∈ Ωq .
To show how this works, we give an example, which is the most interesting n for us: As
above,
nwe take coordinates x = (q1 , . . . , qn , p 1 , . . . , pn ) and α = ω = i=1 dq i ∧ dpi and
X = i=1 (ai (q, p) ∂qi + bi (q, p) ∂pi ). Then we have
n
i(X)ω = (ai dpi − bi dqi ).
i=1
Finally, we use this for the symplectic formulation of the Hamilton formalism: We go
from the configuration space Q with coordinates q to its cotangent bundle M = T ∗ Q,
calledthe phase space, with coordinates (q, p). We provide this space with the two-form
n
ω = i=1 dqi ∧ dpi (and hence get the protype of a symplectic manifold, which will be
analysed later more thoroughly). Now we can say that a curve γ provides a solution to
Hamilton’s equations (8.1) if it is an integral curve of the vector field X, which fulfills
the equation
(8.3) dH = i(X)ω.
In 6.1 we introduced the Poisson bracket for two smooth functions f and g with coordi-
nates (q, p)
n
∂f ∂g ∂f ∂g
{f, g} := − .
i=1
∂q i ∂p i ∂p i ∂qi
U (t) =: eitĤ
(we use units such that Planck’s constant is one) and then has the Schrödinger equation
dψ
= iĤψ.
dt
In our standard example of a system consisting of a particle with mass m moving in
R3 in a central force field with potential V, one translates the position variables qj from
the classical world into operators Bqj acting by multiplication with iqj , the momentum
variables pj go to the differential operators Bpj = ∂qj acting on (a subspace of smooth
functions in) the Hilbert space H = L2 (R3 ) and one has the Schrödinger equation
dψ
3
= iĤψ = i((1/2m) ∂q2j − V (q))ψ.
dt j=1
178 8. Geometric Quantization and the Orbit Method
In general, it can be a serious problem to find a Hilbert space H and operators corre-
sponding to the symplectic space and the observables from the classical description. For
some discussion of this we refer to Chapter 5 in [Be]. For our purposes here, we retain
the following observation even though it is somewhat nebulous:
In important cases the behaviour of a physical system is guided by its symmetry group G,
i.e. a group of transformations, which do not change the physical content of the system.
If we determine unitary representations of G we will be provided with a Hilbert space
and a lot of unitary operators.
In the next sections we shall see that important groups, like the Heisenberg group, the
Euclidean and the Poincaré group, SU(2) and SL(2, R), give rise to symplectic manifolds
and line bundles living on these such that the spaces of sections of these bundles are
representation spaces of unitary representation of G.
8.2.1 Prequantization
In Section 8.1 we already established the cotangent bundle T ∗ Q to a manifold Q and in
particular the space Rn with coordinates (q1 , . . . , qn , p1 , . . . , pn ) and provided with the
standard two-form (8.2)
n
ω0 := dqj ∧ dpj
j=1
– i) dω = 0,
and
– ii) on each tangent space Tm M, m ∈ M, the condition holds: if for X ∈ Tm M one has
then X = 0.
8.2 Coadjoint Orbits and Representations 179
Slightly paraphrasing the condition ii), we remark that M neccessarily has to have even
dimension p = 2n and that for ω in m ∈ M with local coordinates (q1 , . . . , qn , p1 , . . . , pn )
given by (8.2)
n
ω0 = dqj ∧ dpj
j=1
n
X= (aj (q, p) ∂qj + aj (q, p) ∂pj ), Y = (bj (q, p) ∂qj + bj (q, p) ∂pj ),
j=1 j=1
from the general duality formula dxi (∂xj ) = δij and the antisymmetry of the forms one
has the relation
n
ω0 (X, Y ) = (aj (q, p) bj (q, p) − aj (q, p) bj (q, p)).
j=1
At first sight, this form ω0 may look very special. But there is a beautiful theorem stating
that this really a general situation:
Theorem 8.1 (Darboux): For every point m of a symplectic manifold (M, ω) of dimen-
sion 2n, there is an open neighbourhood U of m and a smooth map
The examples of symplectic manifolds, which are of interest for us, are the coadjoint
orbits belonging to the action of a group G on the dual g∗ of its Lie algebra g. For
instance the projective space P1 (C) will come out as such an orbit. Before we go into
this, we elaborate a bit on the way to construct bundles on symplectic manifolds. As we
see it, this construction essentially goes back to Weil’s book Variétés Kählériennes [We].
The integrality of ω can be defined in several equivalent ways. Perhaps the most elemen-
tary one is to require (as in [Wo] p.159]:
(I.) The integral of ω over any closed oriented surface in M is an integral multiple of 2π.
To understand (II.), one has to have at hand the notions of the second cohomology group
H 2 (M, R) and its integral subgroup H 2 (M, Z). Though we are not too far from the
possibility to define these, we renounce to do it here but discuss the construction of the
prequantum line bundle where we can see how this condition works (later we shall be led
to a more practical criterion to decide the quantizability of the manifolds showing up as
coadjoint orbits):
180 8. Geometric Quantization and the Orbit Method
Recalling our definitions from 7.6, this bundle can be constructed as follows. We take a
contractible open cover U = (Uj ) of M . Then Poincaré’s Lemma assures that on each
Uj , there exists a real one–form βj such that
ω = dβj .
The same way, one knows that on each contractible Uj ∩ Uk = ∅ a smooth real function
fjk = −fkj exists such that
dfjk = βj − βk
It is the decisive point that the integrality of ω says that the ajk may be assumed to be
integers (this is really the essence of the definition above). If we put
we have
dcjk = icjk (βj − βk )
Hence Proposition 7.4 in 7.6 shows that the cjk are transition functions for a line bundle
B over M . This bundle has a compatible Hermitian structure <, > (see [Wo] p.264),
which we define below in 8.2.3. Since in Poincaré’s Lemma a form β with dβ = α is
not unique (it can changed by an exact form), this construction is not unique. The gen-
eral statement is that these constructions of prequantum bundles B are parametrized by
H 1 (M, R)/H 1 (M, Z) ([Wo] p.286).
We look at P1 (C) as a real two-dimensional manifold and take the covering U = (Uj )j=1,...,6
with
U1 = {(z : 1); z = x + iy ∈ C; |z| < 1},
U2 = {(1 : w); w = u + iv ∈ C; |w| < 1}, w = 1/z = (x − iy)/(x2 + y 2 ),
U3,4 = {(z : 1); z ∈ C; Re z = x ≷ 0},
U5,6 = {(z : 1); z ∈ C; Im z = y ≷ 0}.
This covering is chosen such that all intersections of the covering neighborhoods are
simply connected (this is not the case for the covering by U and U , which we used in
Example 7.18 in 7.6 as one has U ∩ U C∗ ). With c = /π ∈ R we take as a symplectic
form on M
dx ∧ dy i dz ∧ dz
ω := c =c for (z : 1) ∈ Uj , j = 2,
(1 + x2 + y 2 )2 2 (1 + |z|2 )2
du ∧ dv i dw ∧ dw
:= c =c for (1 : w) ∈ U2 .
(1 + u2 + v 2 )2 2 (1 + |w|2 )2
Exercise 8.1: Verify that this two–form ω in the appropriate coordinates has potential
forms θz and θw with dθz = dθw = ω given by
c xdy − ydx i zdz − zdz
θz = = −c for (z : 1) ∈ Uj , j = 2,
2 1 + x2 + y 2 4 1 + |z|2
c udv − vdu i wdw − wdw
θw = = −c for (1 : w) ∈ U2 .
2 1 + u2 + v 2 4 1 + |w|2
On Uj ∩ U2 (j = 2) we have
i zdz − zdz 1 dz dz
θz − θw = −c − −
4 1 + |z|2 1 + |z|2 z z
i zdz − zdz
= −c
4 |z|2
c xdy − ydx
=
2 x2 + y 2
c ˜
= df2j
2
with a potential function f˜2j . If we fix the tangens as a function on (−π/2, π/2) and
f = arctan as its inverse, we can take
−f˜52 (x, y) = −f (x/y) + π/2 = − arctan(x/y) + π/2 and − f˜42 (x, y) = π + arctan(y/x).
182 8. Geometric Quantization and the Orbit Method
Then, on U2 ∩ U4 ∩ U5 , we have
telling that the volume is an integral multiple of 4π = vol(S 2 ) exactly for c = 4n, n ∈ Z.
π M = P1 (C)
B −→
in the standard way as
(m , v , j ) ∼ (m, v, j)
exactly for
m = m, v = ψj j (m)v.
The bundle transition functions ψjk are fixed (in the charts with coordinates x, y) by
By the way, the transition function z comes out if one does not take real potential forms
as above but complex forms like θ z = (1 + |z|2 )−1 z̄dz.
Following Weil ([We] p. 90), the same construction can be repeated by looking at P1 (C)
as a one-dimensional holomorphic manifold and searching for holomorphic transition
functions
Fjk : Uj ∩ Uk −→ C∗ .
8.2 Coadjoint Orbits and Representations 183
Here we take the same covering U = (Uj )j=1,...,6 as above and the symplectic form
dz ∧ dz
ω := = ddΦj forj = 2,
2πi (1 + |z|2 )2 2πi
with
Φj (z) := log(1 + |z|2 )
resp. for j = 2
dw ∧ dw
ω := = ddΦ2
2πi (1 + |w|2 )2 2πi
with
Φ2 (w) := log(1 + |w|2 ).
On Uj ∩ U2 , j = 2, we have
Φj (z) − Φ2 (1/z) = log(|z|2 ).
Fixing appropriate branches of the complex log-function, we take for j = 2
f2j := log z.
2πi
Then we have
Φj (z) − Φ2 (1/z) = −2πi(f2j − f 2j )
and
z dz
ηj = − dΦj = d log(1 + |z|2 ) =
2πi 2πi 2πi 1 + |z|2
with
dz ∧ dz
ω= .
2πi (1 + |z|2 )2
Using Weil’s notation, for the construction of a holomorphic bundle E = E we have
transition functions
Fjk (z) = exp (2πi fjk (z)) = z
exactly if ∈ Z. We see immediately that for ∈ N0 the holomorphic sections are given
in the z-variable by
s(z) = z λ , λ = 0, 1, . . . ,
(as exactly in these cases we have (1/w)λ w holomorphic in w), i.e. we find
Γ (E ) ∼
= C+1 .
As described in a general procedure below, this space consists of the polarized sections
of the prequantum bundle B = B , where the polarization is given by assigning to the
point m with coordinate z the complex subspace spanned by ∂z in the complexification
Tm P1 (C) ⊗ C C ∂z ⊕ C ∂z
of the tangent space Tm P1 (C) for each m ∈ P1 (C), i.e. Γ(E ) can be identified with the
space of those sections of B , which are constant “along the direction of the z̄−coordinate”.
We try to make this a bit more comprehensive. But to do it, again we need more general
notions from complex and algebraic geometry. As seen above, these geometric notions
can and shall be made later rather tangible by replacing them by algebraic ones in our
concrete examples (and, hence, may be skipped at first try).
184 8. Geometric Quantization and the Orbit Method
8.2.3 Quantization
The prequantization Hilbert space H built from C ∞ –sections of the prequantum bundle
B over M - as we just have seen - often turns out to be too large to be useful: This is to be
seen again in the case of the coadjoint orbits, as the representations of the group G on H
as a representation space are in general far from being irreducible. From representation
theory procedures are known to construct appropriate representation spaces, for instance
by parabolic or holomorphic induction. But in this approach here, it is “natural” to use
a notion from the Kähler geometry to single out a subspace of H, namely the notion of
the Kähler polarization. We take over from [Wo] p.92 ff:
g(X, Y ) = 2ω(X, JY ), X, Y ∈ V (M )
Perhaps the reader wants some more explanations. We follow [Wo] p.269:
such that for any function f and sections s, s ∈ CB∞ (M ) one has
Vice versa, a given Kähler manifold (M, ω, J), i.e. a manifold which is simultaneously in
a compatible manner a symplectic and a complex manifold, carries two Kähler polariza-
tions: the holomorphic polarization P spanned at each point by the vectors (∂za ) and its
complex conjugate, the anti–holomorphic polarization P spanned by the (∂za )’s. Locally
it is possible to find a real smooth function f such that
ω = i∂∂f
and
Θ = −i∂f resp. Θ = i∂f
is a symplectic potential adapted to P resp. P .
– A Hermitian structure on a complex vector bundle B is a Hermitian inner product
(·, ·) on the fibers, which is smooth in the sense that B −→ C, v −→ (v, v) is a smooth
function. It is compatible with the connection if for all sections s, s and all real vector
fields X
i(X)d(s, s ) = (X s, s ) + (s, X s ), X s := i(X) s.
Now, let be given a Kähler polarization P on the quantizable (M, ω) with prequantum
line bundle B over M with connection , Hermitian structure <, > and Hilbert space H
of square integrable sections of B. Then we define a space of polarized sections
∞ ∞
CB (M, P ) := {s ∈ CB (M ); X s = 0 for all X ∈ VC (M, P )} .
comes out as our “new” Hilbert space. As remarked above, it consists of all integrable
holomorphic sections of B.
Alternatively, it can be said that the prequantum bundle B (via P ) is given a structure
of a complex line bundle. Anyway, in important cases, this HP is the representation
space of a discrete series representation.
186 8. Geometric Quantization and the Orbit Method
We leave aside the problem, which of the classical observables have a selfadjoint coun-
terpart acting in HP , discussed to some length in [Wo] and the modification by the
introduction of half–densities and go directly to the application in the construction of
unitary representations via line bundles on the coadjoint orbits.
Let G be a (real) linear group with Lie algebra g and its dual space g∗ . As usual, we
realize all this by matrices and then have the adjoint representation Ad of G on g given
by
Ad(g)X := gXg −1 .
(This is the derived map of the inner automorphism κg : G −→ G, g0 −→ gg0 g −1 .)
Following our preparation in section 1.4.3, we can construct the contragredient represen-
tation Ad∗ defined on the dual g∗ to g by
The asterisk on the right-hand side indicates a dual operator: If < η, X > denotes the
value η(X) of the linear functional η applied to the vector X and A is an operator on g
the dual A∗ is defined by
This representation Ad∗ is called the coadjoint representation. Often we will abbreviate
g · η := Ad∗ (g)η.
In 6.6.1 we already worked with the trace form for A, B ∈ Mn (C) given by
g∗
η −→ Xη ∈ g, defined by η(Y ) = < Xη , Y > for all Y ∈ g.
Then the coadjoint representation is fixed by the simple prescription Xη −→ gXη g −1 .
Oη := G · η.
Oη = {gXη g −1 , g ∈ G}.
8.2 Coadjoint Orbits and Representations 187
It is an outstanding and most wonderful fact that our coadjoint orbits are symplectic
manifolds. We cite from [Vo] p.187:
Tη M g/gη .
where ξX̂ denotes the vector field associated to X̂ in the form of a differential operator
acting on smooth functions f on M by the definition
d
(8.6) (ξX̂ f )(η) := (f (exp tX̂ · η))|t=0 .
dt
For a proof we refer to [Ki] 15.2 or [Kos] Theorem 5.4.1. It is obvious that ωη is skew
symmetric. Also it is, in principle, not deep that a form given for the tangent space of
one point can be transported to the whole space if the space is homogeneous. But it may
be an intricate task to give an explicit description of ω in local coordinates. We shall
give an example soon how (8.5) can be used. The fact that ω comes out as a closed form
can be proved by direct calculation or by more sophisticated means as in [Ki1] p.5-10.
The coadjoint orbits still have more power, they are Hamiltonian G-spaces:
µ̃ : g −→ C ∞ (M )
Then the highlight of this theory is the following result to be found also in the above cited
sources. We shall not have occasion to use it here but state it to give a more complete
picture.
Theorem 8.3: The homogeneous Hamiltonian G-spaces are the covering spaces of coad-
joint orbits. More precisely, suppose M is such a space, with map µ̃ : g −→ C ∞ (M ) and
moment map µ. Then µ is a G-equivariant local diffeomorphism onto a coadjoint orbit,
and the Hamiltonian G-space structure on M is pulled back from that on the orbit by
the map µ.
If G is reductive and we have identified g with its dual g∗ as n × n−matrices, one has the
following classification of the coadjoint orbits, which later will be reflected in a mecha-
nism to construct representations of G.
We take
G = SL(2, R) ⊃ K = SO(2)
and
∗ x y+z
g g = Lie G = U (x, y, z) := ; x, y, z ∈ R = < X, Y, Z >R
y−z −x
with
1 1 1
X= , Y = , Z=
−1 1 −1
One can guess the form of the orbits generated by these matrices, but to be on the safe
side we calculate:
a b
gU (x, y, z)g −1 =: U (x̃, ỹ, z̃) for g = ( )
c d
is given by
Hence, for U = αX, α = 0, as a hyperbolic orbit OαX we have the one-sheeted hyper-
boloid in R3 given by
x2 + y 2 − z 2 = α2 .
For U = αZ, α ≷ 0, as elliptic orbits OαZ we have the upper resp. lower half of the
two-sheeted hyperboloid given by
x2 + y 2 − z 2 = −α2 , z ≷ 0.
For U = F and U = G, as nilpotent orbits OF and OG we have the upper resp. lower
half of the cone given by
x2 + y 2 − z 2 = 0, z ≷ 0.
Finally, for U = 0, there is the nilpotent orbit O0 consisting only of the origin in R3 .
Hence we have a disjoint decomposition of R3 into the orbits in their realization as
subsets of R3
R3 = OαX OαZ (OF ∪{0} ∪ OG ).
α>0 α≷0
GαZ = SO(2),
±1 b
GF = ; b∈R ,
±1
±1 0
GG = ; c∈R .
c ±1
From the general theory we have the prediction that the coadjoint orbits are symplectic
manifolds. We can verify this here directly:
190 8. Geometric Quantization and the Orbit Method
We want to give an explict expression for the Kirillov-Kostant form ω in this case. Here
one has to be rather careful. We used the trace form for matrices < A, B >= Tr AB to
identify g∗ with g via η −→ Xη with < Xη , U >= η(U ) for all U ∈ g. Now, we write
g∗ =< X∗ , Y∗ , Z∗ >R with
1 1 1
X∗ = , Y∗ = , Z∗ = ,
−1 1 −1
We put
g∗
η = x∗ X∗ + y∗ Y∗ + z∗ Z∗
and
X̂ = xX + yY + zZ, Ŷ = x X + y Y + z Z ∈ g.
Hence, by the commutation relations for X, Y, Z, we have
M : = G · kZ∗
= {(x∗ , y∗ , z∗ ) ∈ R3 ; x2∗ + y∗2 − z∗2 = −k 2 }.
It is perhaps not the shortest one but rather instructive to proceed like this: One knows
that
dẋ ∧ dẏ
ω1 =
ẏ 2
is a G-invariant symplectic form on M := H G/SO(2) where
τ̇ = ẋ + iẏ := g(i)
with
bd + ac 1 a b
ẋ = 2 , ẏ = 2 for g = .
c + d2 c + d2 c d
8.2 Coadjoint Orbits and Representations 191
The coadjoint G-action on M and the usual action of G on M = H are equivariant (Exer-
cise 8.2: Verify this). From (8.7) we have the description of the elements (x∗ , y∗ , z∗ ) ∈ M
as
x∗ = −(ac + bd)k,
y∗ = (1/2)(a2 + b2 − c2 − d2 )k,
z∗ = (1/2)(a2 + b2 + c2 + d2 )k,
given by
x∗ k
ẋ = , ẏ = .
y∗ − z∗ z∗ − y∗
We use ψ
to pull back ω1 to a symplectic form ω0 , which in local coordinates (x∗ , y∗ )
with z∗ = x2∗ + y∗2 + k 2 , i.e. z∗ dz∗ = x∗ dx∗ + y∗ dy∗ comes out as
dx∗ ∧ dy∗
ω0 = ψ ∗ ω1 = .
kz∗
Now we have to compare this with the Kirillov-Kostant prescription above and determine
the right scalar factor to get the Kirillov-Kostant symplectic form ω. We shall see that
one has
dx∗ ∧ dy∗
ω=− .
z∗
Perhaps the more experienced reader can directly pick this off the Kirillov-Kostant pre-
scription. As we feel this is a nice occasion to get some more practice in handling the
notions we introduced here, we propose to evaluate the general formula (8.5)
in this situation. For the vector fields ξX̂ acting on functions f on M we have the general
formula
d
(ξX̂ f )(η) = f ((exp tX̂) · η)|t=0 .
dt
Putting gX̂ (t) := exp tX̂ one has
et
gX (t) = ,
e−t
1 t
gF (t) = ,
1
1
gG (t) = .
t 1
192 8. Geometric Quantization and the Orbit Method
And, putting A(g) for the matrix of the coadjoint action η −→ g · η, one deduces from
(8.7)
⎛ ⎞
1
A(gX (t)) = ⎝ (1/2)(e2t + e−2t ) (1/2)(e2t − e−2t ) ⎠ ,
(1/2)(e2t − e−2t ) (1/2)(e2t + e−2t )
⎛ ⎞
1 t −t
A(gF (t)) = ⎝−t (1/2)(2 − t2 ) (1/2)t2 ⎠ ,
−t (1/2)t 2
(1/2)(2 + t2 )
⎛ ⎞
1 −t −t
A(gG (t)) = ⎝ t (1/2)(2 − t2 ) −(1/2)t2 ⎠ .
−t (1/2)t 2
(1/2)(2 + t2 )
When we treat the functions f as functions in the two local coordinates (x∗ , y∗ ) with
z∗2 = x2∗ + y∗2 + k 2 , these operators reduce to
ξX = 2z∗ ∂y∗ ,
ξY = −2z∗ ∂x∗ ,
ξZ = 2(y∗ ∂x∗ − x∗ ∂y∗ ).
and
[H1 , H2 ] = 2H3 , [H2 , H3 ] = 2H1 , [H3 , H1 ] = 2H2 .
su(2) is isomorphic to so(3) (see the exercises 6.4, 6.5 and 6.6 in Section 6.1) and, as
usual, we identify X ∈ su(2) and a = (a1 , a2 , a3 ) ∈ R3 with
⎛ ⎞
0 −a3 a2
A = ⎝ a3 0 −a1 ⎠ =: a1 A1 + a2 A2 + a3 A3 ∈ so(3)
−a2 a1 0
Using this identification the Lie bracket corresponds to the vector product in R3
[A, B] = a ∧ b
and the trace form <, > in M3 (R) to the scalar product (·, ·) in R3
Since <, > is proportional to the euclidean scalar product on R3 , we can use the trace
form to identify so(3)∗ with so(3) via the prescription
(8.8) so(3)∗
η −→ Aη such that η(B) =< Aη , B > for all B ∈ so(3) :
We choose a basis (A∗1 , A∗2 , A∗3 ) with < A∗i , Aj >= δij and for an element η ∈ so(3)∗ we
write η ≡ Aη = a∗1 A∗1 + a∗2 A∗2 + a∗3 A∗3 .
As is well known (see the proof of the fact that SU(2) is a twofold covering of SO(3)
in section 4.3), using these identifications the coadjoint operations of SU(2) on su(2)∗
resp. so(3)∗ are simply the rotations of R3 . The orbits are the spheres of radius s,
Ms := G · ηs for ηs := sA∗3 with s > 0, centered at the origin and the origin itself. One
has a symplectic structure on the sphere Ms of radius s given by the volume form, which
is given by restriction to Ms of the 2–form
As local coordinates we use (a∗1 , a∗2 ) such that Ms is given by the parametrization
and
vol2 (Ms ) = 2 ψ ∗ ω0 = 4πs2 .
J
resp. (8.5)
ω(ξA , ξB ) =< η , [A, B] > .
We have
A = a1 A1 + a2 A2 + a3 A3 , B = b1 A1 + b2 A2 + b3 A3 ,
i.e.
[A, B] = (a1 b2 − a2 b1 )A3 + (a2 b3 − a3 b2 )A1 + (a3 b1 − a1 b3 )A2 ,
and with η = a∗1 A∗1 + a∗2 A∗2 + a∗3 A∗3
As in the preceding example we determine the associated vector fields ξAj . Namely using
here gj (t) = exp tAj , j = 1, 2, 3,
d
ξAj f (η) = f (gj (t) · η)|t=0
dt
leads to
= α∂a∗1 + β∂a∗2 ,
with
α = (a2 a∗3 − a3 a∗2 ), β = (a3 a∗1 − a1 a∗3 )
and the corresponding expressions α , β for ξB replacing the aj by bj . We get
Now that we have examples for coadjoint orbits and bundles living on them, we continue
in the program to use the sections of these bundles as representation spaces and we have
to discuss how the general polarization business works in practical cases.
196 8. Geometric Quantization and the Orbit Method
Step 1. For each orbit Oη ⊂ g∗ we look for real subalgebras n ⊂ g which are subordinate
to η. This means that the condition
holds or, equivalently (this looks quite different but is not too hard to prove), the map
is satisfied. The notion of a complex algebraic polarization is defined in the same way:
we extend η to gc = g ⊗ C by complex linearity and consider complex subalgebras n ⊂ gc
that satisfy the corresponding conditions as in the real case.
An algebraic polarization is called admissible if it is invariant under Ad Gη . In the sequel,
we only look for admissible algebraic polarizations.
The relation of these “algebraic” polarizations to the “geometric” ones defined earlier is
explained in [Ki1] p.28/9 and will be made explicit in a simple example later.
Step 3. Moreover, we suppose (as in [Ki] p.238) that in the complex case mc = n + n is
a subalgebra of gc such that we have closed subgroups Q and M of G with Lie algebras
q and m fulfilling the relations
L := Gη ⊂ Q = Gη Q◦ ⊂ M, n ∩ n = qc , n + n = mc .
Step 4. Finally, and this is our practical criterium to guarantee the integrality or
quantizability of the orbit, we suppose that η can be integrated to a unitary character
χQ of Q with dχQ = η . Then we get a unitary representation π of our group G on (the
completion of) the space F+ := C ∞ (G, n, Q, , χQ ) of smooth C–valued functions φ on
G with
φ(gl) = Q (l)−1/2 χQ (l−1 )φ(g) for all l ∈ L, g ∈ G
and
LY̌ φ = 0 for all Y̌ ∈ u, n =: gη + u,
where Q is the usual modular function and LY̌ φ(g) = dt
d
φ(g exptY̌ )|t=0 .
(Here we follow [Vo] p.194 and p.199. For an essentially equivalent approach see [Ki]
p.199ff).
Though this looks (and is) rather intricate, we shall see that the scheme can easily be
realized in our simple examples, in particular for the Heisenberg group in the next but
one section. As already said, there are a lot of subtle questions hidden here. For instance
(see [Ki] p.236), as another condition Pukanszky’s condition, i.e. the additional property
η + n⊥ ⊂ Oη comes in where n⊥ denotes the annihilator of n in g∗ . And, as discussed in
[Ki] p.239, the fundamental group π1 (O) of the orbit plays an important role in the inte-
gration of the representation η of the Lie algebra q to a representation of the group Q.
This leads to the notion of rigged orbits, i.e. orbits in g∗rigg , the set of pairs (η, χ) where
η ∈ g∗ and χ is here a one-dimensional representation of Gη such that dχ = 2πiη|gη (see
[Ki1] p.123 (and Vogan’s criticism of Kirillov’s book on his homepage)). Here we only
want to give an impression, and (from [Ki] p.241) we cite a theorem by Kostant and
Auslander:
Theorem 8.4: Assume G to be a connected and simply connected solvable Lie group.
Then representations that correspond to different orbits or different characters of the
fundamental group of the orbit are necessarily inequivalent.
Theorem 8.5: All irreducible representations of a compact connected, and simply con-
nected Lie group G correspond to integral G-orbits of maximal dimension in g∗ .
In more detail, in our special examples the result of this procedure is as follows.
We begin by taking up again the here already often treated example G = SU(2). Using
(6.5) in Section 6.5, we take
g = {X = Σaj Hj ; a1 , a2 , a3 ∈ R}
and, similar to (8.8) in 8.2.4, we choose (Hj∗ ) such that < Hj∗ , Hk > = δjk and
From (8.9) for the elliptic orbit Oη SU(2)/U(1) passing through to η = sH1∗ we have
the Kirillov-Kostant prescription
ωη (X, Y ) = s(a2 b3 − a3 b2 ).
H0 = iH1 , H± = H2 ± iH3 .
As one easily verifies, both algebras are subordinate to η in the sense defined above in
Subsection 8.2.5 and, for s = k ∈ N0 , the character of gη
gη
Y −→ 2πi < η, Y >
LH± φ = 0
since one has n+ = gη + < H+ > or n− = gη + < H− >. And this space can be under-
stood exactly as the subspace of holomorphic resp. antiholomorphic smooth sections of
the real bundle B over P1 (C) from 8.2 and the end of 7.6: To show this, we determine
the appropriate differential operators acting on the smooth sections f of B resp. the func-
tions φ on G with φ(gh) = χ2s (h−1 )φ(g) where we denote φ(g) =: φ(a, b) =: φ(α, α , β, β )
decomposing our standard coordinates for g into real coordinates
a =: α + iα , b =: β + iβ .
and hence, using the Wirtinger relations ∂a = (1/2)(∂α − i∂α ), ∂ā = (1/2)(∂α + i∂α ),
we get the operators
acting on the functions φ here written in the coordinates a, ā, b, b̄. If φ comes from a
section f of the bundle B over P1 (C), for b = 0 it is of the form (see (8.10) in 8.2)
Hence we get
LH2 +iH3 φ = (2i/b2 )fz̄ , LH2 −iH3 φ = (2i/b̄2 )fz ,
and we see that the polarization by n± provides that the section f has to be holomorphic
resp. antiholomorphic.
Summary: For G = SU(2) (and SO(3)) all irreducible unitary representations can be
constructed by the orbit method.
Each integral orbit Oη carries a line bundle and has two complex polarizations.
The correponding polarized sections provide equivalent representations.
2. G = SL(2, R)
We continue by completing the results obtained for the orbits from the SL(2)-theory in
the last section. The orbits are all realized as subsets of
with convenient β ∈ R.
and
Q1 = M N A and Q2 = M N̄ A,
Then by inducing these characters from Qj to G one gets irreducible unitary repre-
sentations. This coincides quantitatively with the fact that one has the principal series
representations P ±,is := πis,± , which we constructed in 7.2 and which have models
consisting of (smooth) functions φ = φ (g) on G with
1/2 (is+1) ε ε for −
φ (mn(x)t(y)g) = γ(m) (y ) φ(g) , γ = ,
ε 1 for +
φs,2k , k ∈ Z for + .
and
φs,2k+1 , k ∈ Z for −
with
The unitary equivalence P ±,is P ±,−is can be translated into the fact that both polar-
izations produce equivalent representations (see the discussion in [GS1] p. 299 f). Thus,
by the orbit method, the construction above yields only one half of the representations.
This arises from the fact that the orbit Mα2 is not simply connected but has here a
fundamental group π1 (Mα2 ) Z , so that we have one representation (the even one)
belonging to id ∈ π1 (Mα2 ) and the other one to a generator of π1 (Mα2 ) (see the dis-
cussion in [Ki3] p.463). This may serve as motivation to introduce the notion of rigged
orbits already mentioned at the end of our Subsection 8.2.5.
−i 1 ±i
Z1 = −iZ = , X± = (1/2) (X ± iY ) = (1/2)
i ±i 1
with
[Z1 , X± ] = ±2X± , [ X+ , X− ] = Z1 .
and have
gc = Z1 , X+ , X− .
Bη ([ X̌, Y̌ ]) = iα(a+ b− − a− b+ )
for
X̌ := a0 Z1 + a+ X+ + a− X− , Y̌ := b0 Z1 + b+ X+ + b− X− .
8.3 The Examples SU(2) and SL(2, R) 201
n := n± = Z1 , X± C
of gc and n + n̄ = gc , n ∩ n̄ = kc = Z1
n± ∩ g1 = Y − ZR = k .
and this integrates to a representation χk of SO(2) (with χk (r(θ)) = eikθ ) exactly for
2α =: k ∈ Z \ 0 . Thus the elliptic orbits are quantizable for 2α = k ∈ Z \ {0} . Then we
use n+ and n− to define a complex structure on M = Mα±2 G1 /SO(2) H± . And
by the usual procedure we get the discrete series representations πk± , k ∈ N , realized
±
on the L2 -spaces of holomorphic sections of the holomorphic line bundles L on M−α 2 .
With similar calculations as we did for SU(2), these can also be interpretated as spaces
±
of polarized sections φ of the bundle belonging to χk on M−α 2 , i.e. with
In the nilpotent case Oη = G · (Y∗ + Z∗ ) = M0+ = G/N̄ we have one real polarization.
The form ωY∗ +Z∗ is zero on the Ad N̄ - invariant subalgebra of g given by
n = < X, Y − Z > .
a b
The group B = M N A = {±( ); a > 0, b ∈ R} has Lie B = n. We can (at
0 a−1
least formally) proceed as follows. If χ = id denotes the trivial representation of B, one
has here the representation indGB id given by right translation on the space of functions
φ : G −→ C with
φ(bg) = y 1/2 φ(g) for b ∈ B, g ∈ G,
(spanned by φ(g) = y 1/2 eiθ , ∈ 2Z), i.e. the representation P +,0 . As in the previous
case, here we have a fundamental group π1 Z, so there is a second representation
P −,0 consisting of odd functions and spanned by φ(g) = y 1/2 eiθ , ∈ 1 + 2Z), which
decomposes into irreducible halves
(see for instance [Kn] p.36). The lower half of the cone M0− = G1 · Y∗ generates an
equivalent situation.
202 8. Geometric Quantization and the Orbit Method
All this is a small item in the big and important theme of how to attach representations
to nilpotent orbits (see in particular the discussion in Example 12.3 of Vogan [Vo1] p.386).
Summary
Let us finally point to a remark by Kirillov in [Ki3] 8.4 and [Ki1] 6.4 that he sees the
possibility to extend the correspondence framework such that even the complementary
series fits in. This should be further elucidated. Up to now one has here an example of an
imperfect matching between the unitary dual and the set of coadjoint orbits. But there
is a conjecture that by the orbit method one gets all representations appearing in the
decomposition of the regular representation of a group ([Ki1] p 204): “Indeed, according
to the ideology of the orbit method, the partition of g∗ into coadjoint orbits corresponds
to the decomposition of the regular decomposition into irreducible components”.
As the formulae are a bit more smooth, we work with the realization of the Heisenberg
group by three-by-three matrices, i.e. we take
⎛ ⎞
1 a c
G := Heis (R) = {g = ⎝ 1 b ⎠ ; a, b, c ∈ R}
1
and
⎛ ⎞
0 x z
g := {X̌ = ⎝ 0 y ⎠ = xX + yY + zZ; x, y, z ∈ R}.
0
8.4 The Example Heis(R) 203
The coadjoint representation Ad∗ is the contragredient of Ad (see 1.4.3) and given on g∗
by Ad∗ g = (Ad g −1 )∗ , i.e. if we write
We have ⎛ ⎞
0 x z − ay + bx
(Ad g −1 )X̌ = ⎝ 0 y ⎠.
0
And if we take X∗ , Y∗ , Z∗ as a basis for g∗ , dual to X, Y, Z with respect to <, >, and
write
g∗
η = x∗ X∗ + y∗ Y∗ + z∗ Z∗ ,
we have
Ad∗ g η = x̃∗ X∗ + ỹ∗ Y∗ + z̃∗ Z∗
with
There is a more elegant approach to this result: In [Ki1] p.61 Kirillov uses the trace form
on M3 (R) to identify g∗ with the space of lower-triangular matrices of the form
⎛ ⎞
∗ ∗ ∗
η = ⎝x ∗ ∗⎠ .
z y ∗
Here the stars remind us that one actually considers the quotient space of M3 (R) by the
subspace g⊥ of upper-triangular matrices (including the diagonal). This way, one gets
⎛ ⎞ ⎛ ⎞
∗ ∗ ∗ ∗ ∗ ∗
Ad∗ g η = g ⎝x ∗ ∗⎠ g −1 = ⎝x + bz ∗ ∗⎠ .
z y ∗ z y − az ∗
204 8. Geometric Quantization and the Orbit Method
Exercise 8.5: Do the same for the realization of the Heisenberg group and its Lie algebra
by four-by-four matrices as in 0.1 and 6.3. Verify that Ad∗ g is given by
(p∗ , q∗ , r∗ ) −→ (p∗ + 2µr∗ , q∗ − 2λr∗ , r∗ )
for g = (λ, µ, κ) and η = p∗ P∗ + q∗ Q∗ + r∗ R∗ , where P∗ , Q∗ , R∗ is a basis of g∗ dual to
P, Q, R.
From (8.12) it is easy to see that one has just two types of coadjoint orbits Oη = G · η
G/Gη .
i) For η = mZ∗ , m = 0 we have a two-dimensional orbit
Om := OmZ∗ = {(x∗ , y∗ , m); x∗ , y∗ ∈ R} R2 .
ii) For η = rX∗ + sY∗ , r, s ∈ R, as orbit we have the point
Ors := OrX∗ +sY∗ = {(r, s, 0)}.
∗
Obviously g R is the disjoint union of all these orbits and the set of orbits is in
3
one-by-one correspondence with the unitary dual Ĝ described above. Now, we construct
representations following the general procedure outlined in 8.2:
and, by (8.12),
and, hence,
ξX = −z∗ ∂y∗ , ξY = z∗ ∂x∗ , ξZ = 0,
that is
with
α = yz∗ , β = −xz∗ .
Thus we have
ω0 (ξX̂ , ξŶ ) = (αβ − α β) = z∗2 (xy − x y).
If we compare this with the Kirillov-Kostant prescription above, we see that one has the
symplectic form ω on M = Om given by
ω = (1/m)dx∗ ∧ dy∗ .
The algebraic approach to real polarizations asks for subalgebras n of g = X, Y, Z sub-
ordinate to η = mZ∗ . Obviously one has the (abelian) subalgebras n1 = X, Z and
n2 = Y, Z and more generally, for α, β ∈ R with αβ = 0, n = αX + βY, Z. In the
geometric picture these correspond to a polarization given by a splitting of Om R2
into the union of parallel lines αx∗ + βy∗ = const., which are the orbits of Ad∗ Q where
Q is the subgroup of G with Lie Q = n, i.e. Q = {(αt, βt, c); t, c ∈ R}.
Construction of Representations
We want to follow the scheme outlined at the end of 8.2. As up to now our representations
of the Heisenberg group were realized by right translations and the recipe in 8.2.5 is given
for left translations, we have to make the appropriate changes. For the zero-dimensional
orbits Or,s , r, s ∈ R there is not much to be done: One has η = rX∗ + sY∗ and Gη = G
and hence the one-dimensional representation π given by π(a, b, c) = exp(2πi(ra + sb)).
For the two-dimensional orbits Om , m = 0 we have η = mZ∗ , Gη = {(0, 0, c), c ∈ R} = L
and for each polarization given by n the representation Y̌ −→ 2πi < η, Y̌ >= 2πimz
integrates to a character χm of the group Q belonging to n.
206 8. Geometric Quantization and the Orbit Method
Hence each orbit Om is integral and we get a representation space consisting of sections
of the real line bundle over M = Om (constructed using the character χm of Gη with
χm ((0, 0, c)) = exp (2πimc) = e(mc)) which are constant along the directions fixed by the
polarization n. This leads to consider the following functions: We look at φ : G −→ C
with
φ(lg) = χm (l)φ(g),
i.e. φ is of the form φ(a, b, c) = e(mc)F (a, b) with a smooth function F : R2 −→ C
restricted by the polarization condition, which in our case for n = < U, Z > is given by
RU φ = 0
d
RU φ(g) = φ(gU (t)−1 g)|t=0 , gU (t) := exp tU.
dt
We have
gX (t) = (t, 0, 0), and gX (t)−1 g = (a − t, b, c − tb),
gY (t) = (0, t, 0), gY (t)−1 g = (a, b − t, c),
gZ (t) = (0, 0, t), gZ (t)−1 g = (a, b, c − t).
and hence
From the Stone-von Neumann Theorem one knows that this representation (and any
other constructed this way via another polarization) is equivalent to the Schrödinger
representation. In our context we can see this directly using the Fourier transform as an
intertwining operator:
8.4 The Example Heis(R) 207
We get
πm (g0 )fˆ(b) = e(m(c0 − a0 (b + b0 ))) f (t)e(mt(b + b0 ))dt
R
and see that this coincides with the Fourier transform of the application of the Schrödinger
representation πm (g0 )f , namely with a + a0 =: t we get
(πm (g0 )f )(b) = e(m(c0 + ab0 )f (a + a0 )e(mab)da
= e(mc0 )e(−m(a0 (b + b0 ))) e(mtb0 )f (t)e(mbt)dt.
To adopt our presentation to the more general one given in Igusa [Ig] p.31ff, we introduce
complex coordinates for our Heisenberg group
z := −ia + b, z̄ = ia + b,
i.e.
a = (z̄ − z)/(2i), b = (z + z̄)/2
and
∂z = (1/2)(i∂a + ∂b ), ∂z̄ = (1/2)(−i∂a + ∂b ).
We use the complexification gc =< Y± , Z0 > with Y± := ±iX + Y, Z0 = 2iZ and the
complex polarization given by n =< Y− , Z0 >. Then, in order to construct a representa-
tion space following our general procedure, we come to look at smooth complex functions
φ on G, which are of the form
RY− φ = 0.
and, hence, that one has 0 = RY− φ = (−2Fz̄ − 2πmbF )e(mc), i.e.
with a holomorphic function f˜. As we soon will see, to get a unitary representation on
the space Hm introduced above, we have to choose here f˜(z) = exp(πmz 2 /4)f (z) with
f ∈ Hm , i.e. our φ is of the form
2
/2−z 2 /2)
φ(a, b, c) = e(mc)e−(πm/2)(zz̄+z̄ f (z).
Perhaps it is not a complete waste of time and energy to show here how this representation
comes up also (as it must) by the induction procedure we outlined in 7.1: We have the
standard projection
with the Mackey section s given by s(x) = (a, b, 0), i.e. g = (0, 0, c)(a, b, 0), and the
master equation (7.7)
with
s(xg0 ) = (a + a0 , b + b0 , 0), h∗ = (0, 0, c0 + ab0 ) = (0, 0, c0 + (zz0 − z z̄0 + z̄0 z0 − z̄ z̄0 )/(4i)).
dµ(xg0 ) 2
= e−mπ(zz̄0 +z̄z0 +|z0 | ) = δ(h∗ ).
dµ(x)
Hence, injecting these data into the formula from Theorem 7.4 in 7.1.2 for the represen-
tation in the second realization
provides the formula (8.15) for the Fock representation obtained above.
provides an isometry I between the spaces L2 (R) and Hm for m = 1 and intertwines the
corresponding Schrödinger and Fock representations.
Summary
For the Heisenberg group (as for each solvable group), one has a perfect matching be-
tween coadjoint orbits and irreducible unitary representations.
Everything goes through as well for the higher-dimensional Heisenberg groups if one
changes the notation appropriately.
GE (3) = SO(3) R3 .
The reader will find material for this in [GS] p.124ff. As the author of this text is
particularly fond of the Jacobi group GJ , i.e. – in its simplest version – a semidirect
product of SL(2, R) with Heis(R), as another exercise we propose here to treat some
questions in this direction concerning the Jacobi group. Most anwers to these can be
found in [BeS] and [Be1] where we collected some elements of the representation theory
of GJ and the application of the orbit method to GJ . But at a closer look there are
enough rather easily accessible open problems to do some original new work:
The Jacobi group is in general the semidirect product of the symplectic group with an
appropriate Heisenberg group. Here we look at
In this case we fix the multiplication law by the embedding into the symplectic group
Sp(2, R) given by
⎛ ⎞
1 µ
⎜ λ 1 µ κ ⎟
Heis(R)
(λ, µ, κ) −→ ⎜
⎝
⎟ ,
1 −λ ⎠
1
⎛ ⎞
a b
a b ⎜ 1 ⎟
SL(2, R)
M = −→ ⎜
⎝
⎟ .
⎠
c d c d
1
We write
g = (p, q, κ)M or g = M (λ, µ, κ) ∈ GJ (R).
As in [BeS] or [Ya], we can describe the Lie algebra gJ as a subalgebra of g2 = sp(2, R)
by ⎛ ⎞
x 0 y q
⎜p 0 q r ⎟
G(x, y, z, p, q, r) := ⎜
⎝ z 0 −x
⎟
−p⎠
0 0 0 0
and denote
X := G(1, 0, . . . , 0), . . . , R := G(0, . . . , 0, 1).
We get the commutators
all others are zero. Hence, we have the complexified Lie algebra given by
Exercise 8.9 : Verify all this and compute left invariant differential operators
LZ0 = i∂κ
LY± = (1/2)y −1/2 e±iθ (∂p − (x ± iy)∂q − (p(x + iy) + q)∂κ )
LX± = ±(i/2)e±2iθ (2y(∂x ∓ i∂y ) − ∂θ )
LZ1 = −i∂θ
8.5 Some Hints Concerning the Jacobi Group 211
g = (p, q, κ)n(x)t(y)r(θ)
where
1 x y 1/2 α β
n(x) = , t(y) = , r(θ) =
1 y −1/2 −β α
with
α = cos θ, β = sin θ,
and
g(i, 0) = (τ, z) = (x + iy, pτ + q).
As usual, we put
N := {n(x); x ∈ R}, A := {t(a); a ∈ R>0 },
K := SO(2) = {r(θ); θ ∈ R}, M := {±E}.
Elements of the representation theory of GJ (R) can be found in the work of Pyatetskii–
Shapiro, Satake, Kirillov, Howe, Guillemin, Sternberg, Igusa, Mumford and many others.
To a large part, they are summed up in [BeS], the essential outcome being that the
representations π of GJ (R) with nontrivial central character are essentially products
of projective representations of SL(2, R) with a fundamental (projective) representation
m
πSW of GJ called the Schrödinger–Weil representation in [BeS], as it is in turn the
product of the Schrödinger representation πS of the Heisenberg group Heis(R) and the
m
(Segal–Shale–)Weil or oscillator representation πW of SL(2, R). This representation πSW
is a representation of lowest weight k = 1/2 and index m. It is characterized by the fact
that all its K J –types are one-dimensional (K J := SO(2) × C(Heis(R)) SO(2) × R),
m
which in turn is characterized by the fact that the lowest weight vector for πSW satisfies
the heat equation. To be a bit more explicit, the irreducible unitary representations π of
GJ (R) with central character ψ m (with ψ m (x) := em (x)) for m = 0 are infinitesimally
equivalent to
+
And, for instance, the discrete series representation π̂k−1/2 of sl(2) is given on
by
+
Vm,k := Vm1/2 ⊗ Wk−1/2 = vj ⊗ wl j∈N, l∈2N0
with
Z0 (vj ⊗ wl ) = µ(vj ⊗ wl ),
Z(vj ⊗ wl ) = (k + j + l)(vj ⊗ wl ),
Y+ (vj ⊗ wl ) = vj+1 ⊗ wl ,
X+ (vj ⊗ wl ) = −(1/(2µ))vj+2 ⊗ wl + vj ⊗ wl+2 .
In particular, for the spaces of Z-weight k + λ, λ ≥ 0,
V (λ) := {v ∈ Vm,k
+
; Zv = (k + λ)v}
we have
dim V (0) = 1, dim V (1) = 1, dim V (2) = 2, dim V (3) = 2, dim V (4) = 3, . . .
we have
+
mult (ρm,k̃ , πm,k ) = 1 for k̃ = k,
= 1+l for k̃ = k + 2l or = k + 2l + 1, l ∈ N0
= 0 for k̃ < k,
and
mult (ρm,k̃ , π) = ∞
in the other cases listed above. Now, as to be seen for instance in [BeS] p.28ff, a lowest
+
weight or vacuum vector φ0 for a discrete series representation πm,k of GJ (R) realized by
right translation in a space Hm,k of smooth functions φ = φ(g) on G (R) is characterized
+ J
by the equations
The Jacobi group is not semisimple (and indeed a prominent example for a non-reductive
group). We describe its coadjoint orbits using our embedding of GJ into Sp(2, R) :
and put
X∗ := M (1, 0, . . . , 0), . . . , R∗ := M (0, . . . , 0, 1).
Then
Xη1 = X∗ , . . . , Xη6 = R∗
Lemma: If g −1 is denoted as
⎛ ⎞
a 0 b aµ − bλ
⎜ λ 1 µ κ ⎟
g −1 =⎜
⎝ c
⎟,
0 d cµ − dλ ⎠
0 0 0 1
with
p̃ = ap + bq + r(aµ − bλ),
q̃ = cp + dq + r(cµ − dλ),
r̃ = r.
214 8. Geometric Quantization and the Orbit Method
Exercise 8.10: Verify this and determine the coadjoint orbits. For instance, show that
for
U∗ = mR∗ + αX∗ , m = 0, α = 0,
one has the 4-dimensional hyperbolic orbit OU∗ contained in
It is natural to try to apply the general procedure outlined in this section to realize
representations of GJ by bundles carried by these orbits and to ask, which ones you get
and, if not, why? As already said above, there are some answers in [Be1] and [Ya], but
far from all.
Chapter 9
The goal of these last sections will be to lead a way to automorphic representations:
While we already encountered the decomposition of the regular representation on the
space L2 (G), we shall here take a discrete subgroup Γ of G and ask for the representa-
tions appearing in the decomposition of L2 (Γ \ G), i.e., we consider representation spaces
consisting of functions with a certain periodicity, invariance or covariance property. All
this is most easily understood for the case of the Heisenberg group and leads to the
notion of the theta functions. So we will start by treating this case, and afterwards will
look at the group SL(2, R) and introduce modular forms as examples for more general
automorphic forms. Finally, we shall consider several kinds of L-functions appearing in
connection with representation theory.
216 9. Epilogue: Outlook to Number Theory
Since this is an epilogue and an outlook and the material has such a great extension and
depth, the reader will excuse (we hope) that we shall not be able to give more than some
indications, and even less proofs, than in the preceding chapters.
Moreover, today, one has variants, which unfortunately are differently normalized by
different authors. For instance in [Ig] p.V, one finds the theta function with characteristics
m = (m , m ) ∈ R2
In our text, we shall follow the notation of [EZ] p.58 where, for 2m ∈ N, µ = 0, 1, ., 2m−1,
one has
2
(9.2) θm,µ (τ, z) := q r /(4m) r
r∈Z, r≡µ mod 2m
Obviously, we have ϑ(z, τ ) = θ1/2,0 (τ, z). Theta functions have a rather evident quasiperi-
odicity property concerning the complex variable z and a deeper modular property con-
cerning the variable τ ∈ H, which makes them most interesting for applications in number
theory and algebraic geometry, and – concerning both variables – ϑ satisfies the “one-
dimensional” heat equation
4πiϑτ = ϑzz .
At first, we treat the dependance on the complex variable where, as we shall see, the
Heisenberg group comes in. But to give already here at least a tiny hint, for instance,
theta functions are useful if one asks for the number of ways a given natural number can
be written as, say, four squares of integers (see [Ma] p.354 or our Example 9.3 in 9.2).
9.1 Theta Functions and the Heisenberg Group 217
As we keep in mind that the group really behind theta functions is the Jacobi group
introduced in 8.5, we take the version of the Heisenberg group, which comes from its
realization by four-by-four matrices though in other sources different coordinizations are
used (like we did in the last chapter). Thus, we take
C
z −→ g(z) := z + λτ + µ, g = (λ, µ, κ) ∈ G.
X := G(R)/C(R) C.
Recalling the cocycle condition (7.8) in 7.1 and the first example of an automorphic
factor in 7.2, we refine this action by introducing for m ∈ R as an automorphic factor
for the Heisenberg group
If we relate this to the theta functions defined above, we are led to the following central
result, which is not too difficult to prove (using the periodicity and the functional equa-
tion (9.3)).
θm,α , α = 0, 1, . . . , 2m − 1
Proposition 9.1: Under the map F −→ φF , the space Θ(m) is isomorphic to the space
H(m) of functions φ : G −→ C satisfying
i) φ(γg) = φ(g) for all γ ∈ Γ,
ii) φ(gκ) = φ(g)em (κ) for all κ := (0, 0, κ) ∈ C(R),
iii) φ is smooth and satisfies LY− φ = 0.
The third statement is result of a discussion parallel to the one of the Fock representation
in 8.4: here we use Y− := P − τ Q to express the holomorphicity of F by a differential
equation for φ: For
and for the left invariant operators LU acting on functions φ living on G defined by
d
LU φ(g) := φ(ggU (t))|t=0 ,
dt
we get
LP = ∂λ − µ∂κ ,
LQ = ∂µ + λ∂κ ,
LR = ∂κ ,
and
LY− = LP −τ Q = ∂λ − τ ∂µ − (λτ + µ)∂κ .
In the prescription for the lifting from C to G we assume z = λτ + µ, z̄ = λτ̄ + µ and
(via Wirtinger calculus) deduce
∂z̄ = (τ ∂µ − ∂λ )/(τ − τ̄ ).
If we have
φ(λ, µ, κ) = F (z, z̄)em (κ + λ2 τ + λµ),
a tiny computation shows that the condition LY− φ = 0 translates into the holomorphicity
condition ∂z̄ F = 0.
9.1 Theta Functions and the Heisenberg Group 219
This leads us directly to the following representation theoretic interpretation. The con-
ditions i) and ii) of the space H(m) show up if one looks at still another standard
representation of the Heisenberg group, namely the lattice representation:
π0 (h) := em (t).
z −→ z + λτ + µ
(As X is compact, we have no problem with the condition of the finite norm integral.)
This induced representation is not irreducible and we proceed as follows. For Γ = Heis(Z)
we denote by L2 (Γ \ G) the Hilbert space of Γ−invariant measurable functions G −→ C
with the scalar product
φ, ψ := φ(g)ψ(g)dg.
Γ\G
L2 (Γ \ G) = ⊕m∈Z L2 (Γ \ G)m
Now for m > 1/2, this induced representation on L2 (Γ \ G)m is not irreducible. We find
here the first example of an important general statement.
Theorem 9.2 (Duality Theorem for the Heisenberg group): For m ∈ N, the multiplicity
of the Schrödinger representation πm in L2 (Γ \ G) is equal to the dimension of the space
Θ(m).
Proof: There are several different approaches. In the our context, we can argue like
this:
The Schrödinger representation πm transforms a function f ∈ L2 (R) into
and is intertwined with the Heisenberg representation (see Example 7.1 at the beginning
of 7.1) given by right translation on functions φ on G via
Obviously, one has φ(κg) = em (κ)φ(g) but as the Heisenberg representation is induced
by the two-dimensional subgroup {(0, µ, κ); µ, κ ∈ R} of G, one also has invariance
concerning the µ−variable. Now, to produce a subspace of the lattice representation,
one has to create in- or covariance concerning the λ−variable: We define for n ∈ Z and
α = 0, 1, . . . , 2m − 1
λn,α := (n + α/(2m), 0, 0)
and get
λn,α g = (λ + n + α/(2m), µ, κ + µ(n + α/(2m)).
Hence for every α, we have a map ϑα : f −→ φf,α , called theta transform, with
The reader could try to find another version of the proof using the bundle approach to
induced representations. One can prove (Theorem of Appell-Humbert) that every line
bundle on the torus X := L \ C, L = Zτ + Z is of the form L(H, β), i.e. the quotient of
the trivial bundle C × C −→ C by the action of L given by
(, (v, z)) −→ (β()eiH(z,)+(1/2)H(,) v, z + ) for all ∈ L, v, z ∈ C.
Here H : C × C −→ C is a hermitian form such that E = Im (H) is integer valued on L
and β : L −→ S 1 a map with
β(1 + 2 ) = eiπE(1 ,2 ) β(1 )β(2 ).
A computation shows that these hermitian forms are exactly the forms Hm , m ∈ Z, given
by
Hm (z, w) = mz̄w/Im(τ ).
Up to now, we only treated the theta functions as quasiperiodic functions in the complex
variable z. There is also a remarkable modular behaviour concerning the variable τ ∈ H.
Before we discuss this, we return to some more general considerations.
9.2 Modular Forms and SL(2,R) 221
We already met several times with the action of G = SL(2, R) on the upper half plane
X = H := {τ = x + iy ∈ C; y > 0} given by
a b aτ + b
(g = , τ ) −→ g(τ ) := .
c d cτ + d
As g and −g produce the same action, it is more convenient to treat the group
But since we worked so far with G and its representation theory can be rather easily
adopted to Ḡ, we stay with this group.
fulfilling the cocycle relation (7.8) j(gg , τ ) = j(g, g (τ ))j(g , τ ). Let f be a smooth func-
tion on H. Then one writes
There are several important types of discrete subgroups Γ of G and the theory becomes
more and more beautiful if one treats more and more of these. For our purpose to give
an introduction, it is sufficient to restrict ourselves to the (full) modular group
Γ = SL(2, Z)
There is a lot to be said about these groups. We mention only two items:
Exercise 9.3: Prove this and realize that T acts on H as a translation τ −→ T (τ ) = τ +1
and S as a reflection at the unit circle τ −→ S(τ ) = −1/τ .
Determine the fixed points for the action of Γ and their stabilizing groups.
Here we use the definion from [Sh] p.15: For any discrete subgroup Γ of G = SL(2, R),
we call F a fundamental domain for Γ \ H (or simply for Γ) if
– i) F is a connected open subset of H,
– ii) no two points of F are equivalent under Γ,
– iii) every point of H is equivalent to some point of the closure of F under Γ.
(Other authors sometimes use different definitions.)
Exercise 9.4: Verify this and show that the set of Γ−orbits on H is in bijection to
If you have trouble, search for help in [Sh] p.16 or any other book on modular forms.
f is called cusp form (forme parabolique in French) if one has moreover c(0) = 0.
We denote by Ak (Γ) the space of all elliptic modular forms of weight k and by Sk (Γ)
the space of cusp forms. Both are C vector spaces and their dimensions are important
characteristic numbers attached to Γ. From i) taken for γ = −E2 , one sees immediately
that for odd k there are no non-trivial forms. For even k = 2 one has
where [x] denotes the integral part of x, i.e. the largest integer n such that n ≤ x. For a
proof see for instance [Se] p.88 or (with even more general information) [Sh] p.46.
Remark 9.3: One also studies meromorphic modular forms where in ii) holomorphic is
replaced by meromorphic and in iii) one has a q−development f (τ ) = n>−∞ c(n)q n .
9.2 Modular Forms and SL(2,R) 223
f (τ + 1) = f (τ ), f (−1/τ ) = τ k f (τ ).
Hence a modular form of weight k is given by a series f (τ ) = n≥0 c(n)q n converging
for |q| < 1 and fulfilling the condition
f (−1/τ ) = τ k f (τ ).
Remark 9.4: The condition iii) concerning the Fourier expansion can also be understood
in several different ways, for instance as a certain boundedness condition for y −→ ∞
(in particular for f ∈ Sk (SL(2, Z)), |f (τ )|y k/2 is bounded for all τ ∈ H), or, if k is even,
as holomorphicity of the differential form α = f (τ )dτ k/2 at ∞: one can compactify the
upper half plane H by adjoining as cusps the rational numbers and the “point” ∞, i.e.
one takes H∗ := H ∪ Q ∪ {∞}. Then one extends the action of SL(2, R) on H to an action
on H∗ and identifies Γ \ H∗ with the Riemann sphere or P1 (C). In this sense modular
forms are understood as holomorphic k/2-forms on P1 (C).
All this is made precise in books on the theory of modular forms. Here, for our purpose,
the essential point is that, similar to the theta functions in the previous section, modular
forms are realizations of dominant weight vectors of discrete series representations of
G = SL(2, R). Before we go into this, let us at least give some examples of modular
forms and the way they can be constructed.
Example 9.1: The most well-known elliptic modular form is the ∆−function given by
its q−expansion
(2π)−12 ∆(q) : = q Π∞
n=1 (1 − q )
n 24
= Σ∞
n=1 τ (n)q = q − 24q + 252q − 1472q + . . . .
n 2 3 4
This is a cusp form of weight 12. The function n −→ τ (n) is called the Ramanujan func-
tion. It has many interesting properties, for instance there is the Ramanujan conjecture
which has been proved by Deligne in 1974 (as consequence of his proof of the much more
general Weil conjecture concerning Zeta functions of algebraic varieties (see Section 9.6)).
But it is still an open question whether all τ (n) are non-zero.
A modular form in a slightly more refined sense is the η−function, also appearing in the
physics literature in several contexts,
η(τ ) := q 1/24 Π∞
n=1 (1 − q ).
n
where the sum is taken over all m, n ∈ Z with exception of (m, n) = (0, 0).
224 9. Epilogue: Outlook to Number Theory
It is not too much trouble to verify that this is really a modular form of weight k and
has a Fourier expansion (see [Sh] p.32)
∞
(2πi)k
ζ(s) := n−s
n=1
and σk (n) denotes the sum of dk for all positive divisors d of n. These functions appear
naturally in the discussion of lattices L = τ Z + Z in C. They can also be introduced by
a far reaching averaging concept similar to the theta transform we used in the previous
section to understand the construction of theta functions:
Every function ϕ on H formally gives rise to an object satisfying the covariance property
i) of a modular form by the averaging
ϕ|k [γ].
γ∈Γ
One has the problem whether this series converges (and has the properties ii) and iii)).
In general this is not to be expected. But one can use the following fact: For g0 = T
our automorphic factor has the property j(g0 , τ ) = 1 , and, hence, one has
ϕ0 |k [T ] = ϕ0 .
Ek (τ ) := (ϕ0 |k [γ])(τ )
γ∈Γ∞ \Γ
induces a bijection
Γ∞ \ Γ {(c, d) ∈ Z2 ; (c, d) = 1}
where here (c, d) denotes the greatest common divisor. Hence, one can write
Ek (τ ) = (cτ + d)−k .
c,d∈Z, (c,d)=1
ζ(k)Ek (τ ) = Gk (τ ).
9.2 Modular Forms and SL(2,R) 225
The same procedure goes through for every Γ∞ -invariant ϕ. For instance, ϕ(τ ) = e(mτ )
leads to the so called Poincaré series, which are used to construct bases for spaces of
cusp forms.
Example 9.3: As already mentioned in 9.1, Jacobi’s theta function ϑ has a modular
behaviour concerning the variable τ . Here we restrict the function to its value at z = 0
2
ϑ(τ ) := ϑ(0, τ ) = eπin τ .
n∈Z
The first one is easy to see and the second one requires some work (see for instance [FB]
p.344/5). These relations entail that this ϑ is a modular form with respect to the theta
group Γϑ if one generalizes the definition of modular forms appropriately. There are
more general theta series associated to quadratic forms resp. lattices leading directly to
modular forms:
ϑ(S; τ ) := eπiS[x]τ .
x∈Zn
In particular, for
n
an (m) := a(En , m) = #{x ∈ Zn ; x2j = m}
j=1
Proofs can be found for instance in [FB] or [Mu] I. The second relation can be verified
using the nice identity
ϑ4 (τ ) = π −2 (4G2 (2τ ) − G2 (τ /2)).
Example 9.4: As a final example we present the j−invariant or modular invariant,
which is a meromorphic modular form: We put
g2 (τ ) := 60G4 (τ ), g3 (τ ) := 140G6 (τ ).
j is a modular function with Fourier expansion (see for instance [Sh] p.33)
∞
j(τ ) = q −1 (1 + cn q n ), n ∈ Z.
n=1
This function is of special importance because it induces a bijection Γ\H∗ P1 (C). And
the coefficients cn hide some deep mysteries (essentially assembled under the heading of
monstrous moonshine (see [CN], [FLM])).
Parallel to the relation between the Second and First Realization in the induction pro-
cedure one defines a lifting of functions f from the homogeneous space X = G/K Gx0
with K = Gx0 to the group using the automorphic factor j(g, τ ) satisfying the cocycle
condition (7.8) by f −→ φf where we put
Our aim is to translate the conditions i) to iii) from the definition of modular forms f to
appropriate conditions characterizing the lifts of these forms among the functions φ on
G. Classic sources giving all details for this are [GGP], [Ge] and [Bu].
g = n(x)t(y)r(ϑ).
and, hence, for a function φ = φf with φf (g) = ck (g)f (x, y), ck (g) := y k/2 eikϑ
LZ φ = kφ,
LX± φ = ck±2 D̃± f
where
D̃+ = i(∂x − i∂y − ik/y), D̃− = −iy 2 (∂x + i∂y ).
Furthermore, using Wirtinger calculus with
we get
LZ φ = kφ,
LX± φ = ck±2 D± f
where
D+ = 2i(∂τ + k/(τ − τ̄ )), D− = (i/2)(τ − τ̄ )2 ∂τ̄ .
Hence in particular, we have found:
Remark 9.7: Let φ = φf be the lift of f . The condition ii) that f is holomorphic on H
is equivalent to
ii’) LX− φ = 0.
228 9. Epilogue: Outlook to Number Theory
The last condition iii) is easily seen to be equivalent to the boundedness condition
With a bit more work, one translates the cuspidality of a modular form, i.e. f ∈ Sk (Γ),
to the following
Lemma: f ∈ Ak (Γ) is in Sk (Γ) iff for every y0 > 0 there is an M > 0 such that
Hence one has for the lifted functions φ = φf the following property.
Theorem 9.3: Via the lifting ϕk , the space Ak (Γ) of modular forms on H is isomorphic
to the space Ak of smooth functions φ on G with
ia) φ(γg) = φ(g) for all γ ∈ Γ = SL(2, Z),
ib) φ(gr(ϑ)) = φ(g)eikϑ for all r(ϑ) ∈ K = SO(2),
ii’) LX− φ = 0,
iii’) |φ(g)y −k/2 | < M for all y >> 0.
Moreover, the space of cusp forms Sk (Γ) is isomorphic to
Remark 9.9: The cusp condition in Remark 9.8 and in the Theorem above is equivalent
to
(cusp) Γ∩N \N
φ(n(x)g)dx = 0 for all g ∈ G.
Theorem 9.4: The space of cusp forms A0k is characterized by the following conditions
ia) φ(γg) = φ(g) for all γ ∈ Γ = SL(2, Z),
ib) φ(gr(ϑ)) = φ(g)eikϑ for all r(ϑ) ∈ K = SO(2),
ii) ∆φ = LΩ φ = k((k/2) − 1)φ,
iii) φ is bounded,
iv) Γ∩N \N
φ(n(x)g)dx = 0 for all g ∈ G.
Here
∆ = LΩ = 2y 2 (∂x2 + ∂y2 ) − 2y∂x ∂ϑ
is the Laplacian (7.17) from 7.2 and, by a small calculation, it is easy to see that the lift
φf of a cusp form f ∈ Sk (Γ) satisfies the condition ii) and hence all the conditions from
the theorem. The proof of the fact that the lifting map ϕk is a surjection onto the space
given by the conditions ia) to iv) reduces to the verification that for φ ∈ A0k the function
f with
is well defined for each g with g(i) = τ (which is quite obvious) and that f is a cusp form.
This assertion is based on a considerably deeper theorem, the Theorem of the discrete
spectrum, which we shall state below.
Decomposition of L2 (Γ \ G)
We need another tool from functional analysis (for background see for instance [La1]
p.389ff):
Remark 9.11: The minimal closed extension of the Laplacian ∆ is a selfadjoint operator
ˆ As ∆ does by constuction, the extension ∆
∆. ˆ also commutes with the right translation
and hence, by Schur’s Lemma (“the unitary Schur” mentioned in 3.2) ∆|H0 is scalar for
each G-invariant subspace H0 of H. From functional analysis we take over that ∆ˆ and ρ
have a discrete spectrum only for the subspace of cuspidal functions
H0 := {φ ∈ L2 (Γ \ G) ; φ(n(x)g)dx = 0 for almost all g ∈ G}.
Γ∩N \N
230 9. Epilogue: Outlook to Number Theory
Theorem 9.5 (Theorem of the discrete spectrum): The right regular representation ρ of
SL(2, R) on H0 is completely reducible
H0 = ⊕j∈Z Hj
As this leads too much into (most fascinating) functional analysis, we can not go into the
proof of this theorem and refer to [La1] and [GGP] p.94ff. A main point is a multiplicity-
one statement going back to Godement, which we already mentioned (without proof)
in 6.4. But we try to give an indication of the power of this theorem: For the proof of
characterization of cusp forms by the Laplacian in the theorem above, we have to show
that the function f from (9.5)
φ= φj ,
j
the components φj have to fulfill the same relation ib) as φ does. And one has
because by the last remark above, ∆ is scalar on Hj , i.e. one has a λj with
∆φj = λj φj .
ˆ is self adjoint,
and, as ∆
As another consequence we come to the statement parallel to the one we proved in 9.1
for theta functions.
Theorem 9.6 (Duality Theorem for the discrete series of SL(2, R)): The multiplicity
of πk− in the right regular representation ρ on the space H0 of cuspidal functions is equal
to the number of linear independant cusp forms of weight k
Proof: There is a nice and comprehensive classic proof in [GGP] p.53-57. As we al-
ready have at hand the necessary tools, we reproduce its highlights changing only the
parametrization of G = SL(2, R) used by these authors to the one used here all the time.
a)At first, we show that mk := mult(πk− , ρ) is equal to the number m̃k of linear indepen-
dent functions Ψ on X = Γ \ G in H0 with
ai) For each irreducible subspace Hj of H0 , which is a model for πk− , there is - up to a
nonzero factor - exactly one highest weight vector where ρ restricted to Hj has highest
weight −k and, hence ∆ has eigenvalue λk .
aii) Every Ψ ∈ H0 satisfying (9.6) is a linear combination of vectors of highest weight
from irreducible subspaces of H0 equivalent to a representation space of πk− .
b) In the second step, we show that m̃k is equal to the number of cusp forms:
The space of functions Ψ as above is in bijection to the space of functions Φ on G with
Hence, for g0−1 with components a0 , b0 , c0 , d0 , the matrix g0−1 g has parameters τ1 , ϑ1 with
τ1 = g0−1 (τ ), ϑ1 = ϑ − arg(c0 τ + d0 ). And gr(ϕ) has the parameters τ, ϑ + ϕ. Thus the
second equation in (9.7) shows that Φ(g) is a function of type f1 (τ )e−ikϑ . Guided by the
proof of our Remark 9.7, in order to also fulfill the other conditions in (9.7) we try the
ansatz
Φ(g) = e−ikϑ y k/2 f (τ ).
−k
Using our previous notation, this can also be written as Φ(g) = j(g, i) f (τ ). We see
that using this ansatz, the third condition in (9.7) is equivalent to the functional equation
for modular forms
For the real coordinates x, y of τ the first equation in (9.7) comes down to
resp.
y(fxx + fyy ) − ik(fx + ify ) = 0
If f is a modular form, ergo holomorphic, one has fτ̄ = 0 and this equation is fulfilled.
On the other hand, if Φ fulfills (9.7), by the Decomposition Theorem, it is a linear
combination of highest weight vectors with higest weight −k, which are annihilated by
LX+ = e2iϑ (2y(∂x − i∂y ) − ∂ϑ ). And this again leads just to the condition fτ̄ = 0 showing
that Φ corresponds to a linear combination of cusp forms.
The reader again is invited to watch very closely whether all signs are correct. It has to
be emphasized that the notions highest and lowest weight depend on our parametrization
of K = SO(2) and interchange if one uses r(−ϑ) in place of our r(ϑ) (as Gelbart does in
[Ge]). [GGP] wisely use the term vector of dominant weight.
232 9. Epilogue: Outlook to Number Theory
Remark 9.12: It is a natural question to ask under which conditions the lift φ = φf
of a cusp form f is a cyclic vector, i.e. generates an irreducible representation. There
is a nice answer to this: if f is an eigenfunction to the Hecke operators T (p). We will
introduce these operators later in the context of associating L−functions to modular
forms and here only mention that the space of cusp forms has a basis consisting of such
eigenfunctions.
The big difference between this duality statement and the one for the Heisenberg group
in 9.1 is that in this case we have no way to an explicit construction of the modular forms
starting from their property of being a dominant weight vector. Apparently, this reflects
the fact that the structure of the group at hand is considerably more complicated than
that of the Heisenberg group. Some progress here should give a deep insight into the
modular group.
The characterization theorem is the starting point for several generalizations leading to
the notion of automorphic forms. It can be extended to other groups, even adelic ones
as to be found for instance in [Ge]. There it is also pointed out that the right habitat
for the modern theory of modular forms is not SL(2) but GL(2). Here we can not be too
ambitious and we stay with G = SL(2, R), K = SO(2) and Γ = SL(2, Z) and only treat
the following generalization.
Besides the cusp forms realizing highest weight vectors of πk− we just treated, as their
counterparts there come in forms realizing the lowest weight vectors of πk+ , corresponding
to antiholomorphic modular forms on H and the famous Maass wave forms: these are
bounded right K-invariant eigenfunctions of the Laplacian ∆φ = λs φ with
λs = −(1 − s2 )/2, s ∈ iR.
In analogy to the duality theorem stated above, one has here the fact that the dimension
of the space Ws (Γ) of these wave forms is equal to the multiplicity of the principal series
representation πis in ρ (for this and in general more precise information see [GGP] in
particular p.50). In [Ge] p.29 we find the statement: one knows that Ws (Γ) is non-trivial
for infinitely many values of s, but almost nothing is known concerning the specific val-
ues of s for which Ws (Γ) is non-trivial. Still now, Maass wave forms are an interesting
subject. For information on more recent research we recommend the paper by Lewis and
Zagier [LZ].
We finish this section by some remarks concerning the continuous spectrum of the Lapla-
cian, resp. the orthocomplement Hc of the subspace Hd of H = L2 (Γ \ G) consisting of
9.2 Modular Forms and SL(2,R) 233
the cuspidal and the constant functions. One has rather good control about this space
as it can be spanned by the so called non-holomorphic Eisenstein series. As sources for
this we go back to [Ku] and [La1] p.239ff.
fµ (τ ) := y µ , µ ∈ C
and for k ∈ N0
1 b
with Γ∞ = {γ = ; b ∈ Z}.
1
This prescription formally has the transformation property of a modular form of weight
k and it can be verified that for µ with Re µ > 2 it defines a (real analytic) function
on H, which is non-cuspidal. Using the standard lift ϕk , this function can be lifted to a
function on G
= φs,k (γg),
γ∈Γ∞ \Γ
come in, which span the space of the principal series representations. We have
and these functions are further examples for the definition of automorphic forms above.
Already in our (brief) chapter on abelian groups, we met with a continuous decomposi-
tion of the regular representation for G = R by throwing in elements of Fourier analysis
without really explaining the background from functional analysis. Now, this gets worse
here: There is the central statement ([Ge] p.33):
Theorem 9.7: The restriction ρc of ρ to Hc is the continuous sum of the principal series
representations πis,+ , i.e.
∞
c
ρ (g) = πis,+ (g)ds.
0
234 9. Epilogue: Outlook to Number Theory
one can say that any slowly increasing function f on H, which is orthogonal to all cusp
forms, may be expressed by a generalized Fourier integral
∞
y (is)/2
f (τ ) = fˆ(s)E(τ, s)ds, fˆ(s) := (f, E(., s)), E(τ, s) := .
0 |cτ + d|2s
c,d∈Z;(c,d)=1
Though we will not try to get to the bottom of the proofs, we indicate the following also
otherwise useful tools:
θϕ(g) := ϕ(γg).
γ∈Γ∞ \Γ
It is part of nice analysis to discuss under which conditions this leads to convergent ob-
jects and it is no wonder that it works at least if ϕ is a Schwartz function.
φ(n(x)g) = φj (g)e(jx).
j
i.e. we have for functions assuring the convergence of our prescriptions the relation
i.e. an element fulfilling the equation (7.11) in 7.2 characterizing the elements of His,+ .
Remark 9.13: For ϕ in the Schwartz space S(N \G) and Re µ > 0, the integral defining
Z(ϕ, g, µ) converges absolutely and Z(ϕ, g, µ) is an entire function in µ if moreover ϕ has
finite support.
With some more work composing the Zeta and the Theta transform, one obtains a
Γ−invariant function:
is absolutely convergent.
By a small computation, we get the relation between this Eisenstein series and the one
we defined above.
one has
E(ϕ, g, µ) = π −(+µ) Γ( + µ)Ek (g, µ).
The functions ϕ,k , ∈ N0 , k ∈ Z are a Hilbert basis of the space L2 (N \ G) with respect
to the measure dν = y −2 dydϑ. For every fixed and µ = is + 1, the images of these ϕ,k
by the Zeta transform are a basis for the representation spaces His,± if one takes all even
resp. odd k ∈ Z.
There is still much to be said here. We shall come back to the Zeta transform in the
context of a discussion of ζ− and L−functions. But before doing this, we shall finish our
outlook to automorphic forms by returning to theta functions.
236 9. Epilogue: Outlook to Number Theory
and its variant (9.2) (following the notation of [EZ] p.58) with 2m ∈ N, µ = 0, 1, . . . , 2m − 1
2
θm,µ (τ, z) := q r /(4m) r
r∈Z, r≡µ mod 2m
One has ϑ(z, τ ) = θ1/2,0 (τ, z). We already explored the quasiperiodicity property con-
cerning the complex variable z
2
ϑ(z + rτ + s, τ ) = e−πi(r τ +2rz)
ϑ(z, τ )
and (in Example 9.3 in 9.2) the modular property of the value ϑ(τ ) := θ(0, τ ) for z = 0
(also called by the German word Thetanullwert)
To be consistent with [EZ] and the usual Jacobi Theory, we have to change (from Jacobi’s
and Mumford’s) notation and put
θ(τ, z) := ϑ(z, τ ).
aτ + b z
(9.8) θ( , ) = (cτ + d)1/2 ζ e((1/2)cz 2 /(cτ + d))θ(τ, z).
cτ + d cτ + d
The unit root ζ can be determined as follows:
9.3 Theta Functions and the Jacobi Group 237
We
may assume c > 0 or c = 0 and d > 0 (otherwise one would multiply the matrix
a b
by −1) and hence have (cτ + d)1/2 in the first quadrant, i.e. with real- and
c d
imaginary part > 0.
We take the Jacobi symbol ( ab ), i.e. the multiplicative extension of the Legendre symbol,
which for a, p ∈ Z, p prime is given by
( ap ) = 0 if p | a,
= 1 p a and there is an x ∈ Z with x2 ≡ a mod p,
= −1 p a and there is no x ∈ Z with x2 ≡ a mod p.
The proof of this theorem is a very nice piece of analysis but too lengthy for this text.
So we refer to the books of Eichler [Ei], in particular p.59-62, Lion-Vergne [LV] p.145ff,
or Mumford [Mu] I.
Another most remarkable fact is the following characterization of θ, which comes about
as an exercise in complex function theory.
By a closer look one sees that the quasi-periodicity and the modular transformation
property of θ can be combined to a transformation property under an action of the
Jacobi group GJ , which is a semi-direct product of SL(2, R) and Heis(R) and in 8.5 was
realized as a subgroup of Sp(2, R) given by four-by-four matrices. As in 8.5, we use two
notations
a b a b
GJ
g = (λ, µ, κ) = (p, q, r) .
c d c d
By a short calculation one can verify that for (τ, z) ∈ H × C
aτ + b z + λτ + µ
(τ, z) −→ g(τ, z) := ( , )
cτ + d cτ + d
And for k, m ∈ Q one has (see the first pages of [EZ]) an automorphic factor
c(z + λτ + µ)2
(9.9) jk,m (g; (τ, z)) := (cτ + d)−k em (− + λ2 τ + 2λz + λµ + κ).
cτ + d
f −→ f |k,m [g] with f |k,m [g](τ, z) := f (g(τ, z))jk,m (g; (τ, z)).
Similar to modular forms and theta functions as automorphic forms for SL(2, R) resp.
Heis(R), for the Jacobi group GJ , its discrete subgroup ΓJ := SL(2, Z) Z3 , and
k, m ∈ N0 one defines (as in [EZ]):
φf (g) := f (g(i, 0))jk,m (g; (i, 0)) = f (x, y, p, q)em (κ+pz)eikϑ y k/2 , g = (p, q, κ)n(x)t(y)r(ϑ).
Similarly to the analysis in the former cases, one can characterize the images of Jacobi
forms.
Proposition 9.2: Jk,m is isomorphic to the space Ak,m of complex functions C ∞ (GJ )
with
– i) φ(γg) = φ(g) for all γ ∈ ΓJ ,
– ii) φ(gr(ϑ)(0, 0, κ)) = φ(g)em (κ)eikϑ ,
– iii) LX− φ = LY− φ = 0,
– iv) φ(g)y −k/2 is bounded in domains of type y > y0 .
cusp
Jk,m is isomorphic to the subspace A0k,m of Ak,m with
– iv’) φ(g) is bounded.
For a proof of this and the following statement see [BeB] or [BeS] p. 76ff.
9.4 Hecke’s Theory of L−Functions Associated to Modular Forms 239
cusp
Theorem 9.10 (Duality Theorem for the Jacobi Group): For m, k ∈ N, dim Jk,m is
J
equal to the multiplicity of the appropriate discrete series representation of G in the
right regular representation 0 on the space Hm0
of cuspidal functions of type m, which
consists of the elements φ ∈ L (Γ \ G ) satisfying the functional equation
2 J J
At a closer look, one can see that the classic ϑ has the transformation property of a Jacobi
form of weight and index 1/2 for a certain subgroup of ΓJ . As already the infinitesi-
mal considerations indicated, one has to multiply with modular forms of half-integral
weight to get Jacobi forms in the sense of the definition above. This is very nicely ex-
ploited in [EZ]. The appearance of the half-integral weights can be explained by the fact
that, somewhat similar to the relation of the representation theory of SO(3) and SU(2)
resp. SO(3, 1) and SL(2, C), here one has coming in a double cover of SL(2, R) called
the metaplectic group Mp(2, R). But this group is no more a linear group and hence,
unfortunately, outside the scope of this book.
In our context, we take a cusp form f ∈ Sk (Γ) and get an associated Dirichlet series
convergent for s with Re s > k as the coefficients c(n) of the Fourier expansion of f fulfill
an appropriate growth condition. The most important Dirichlet series have an Euler
product similar to the prototype of a Dirichlet series, namely the Riemann Zeta series
∞
ζ(s) := n−s ,
n=1
which is convergent for s with Re s > 1 and has the Euler product (taken over all prime
numbers p)
1
ζ(s) = Π .
1 − p−s
Moreover, it is known that ζ fulfills a functional equation under s −→ 1 − s and can be
continued to a meromorphic function on the whole plane C with a simple pole at s = 1.
240 9. Epilogue: Outlook to Number Theory
Roughly said, Hecke’s theory analyses how these and more general properties are related
to properties of modular forms. The main tools are the following operators.
Hecke Operators
Definition 9.4: For a natural number n one introduces the Hecke operator T (n) as an
operator acting on functions f on the upper half plane H by
aτ + b
(9.10) T (n)f (τ ) := nk−1 d−k f ( ).
d
a≥1,ad=n,0≤b<d
The origin of this prescription can be understood in several ways. There is the more
general background asking for the decomposition of double cosets into cosets for Γ: For
instance, for a prime p one has
1 1 b
Γ Γ= Γ
p p
0≤b<p
(as exploited in [Sh] p.51ff). But we follow the presentation in [Se] p.98 based on the
notion of correspondences: Let E be a set and let GE be the free abelian group generated
by E. A correspondence on E with integer coefficients is a homomorphism T of GE to
itself. We can describe T by its values on the elements x of E:
T (n)L = Ln .
[L:Ln ]
The sum on the right side is finite. If n is prime, one sees that one has n + 1 such
sublattices. For the general case one has
and the set of lattices Ln of index n in L. For L with basis (τ1 , τ2 ) this bijection is given
by associating to the matrix M the lattice with basis τ 1 = aτ1 + bτ2 , τ 2 = d τ2 .
Rλ L = λ L.
It makes sense to compose the correspondences T (n) and Rλ , since they are endomor-
phisms of the abelian group GR .
Rλ Rµ = Rλµ , λ, µ ∈ C∗ ,
Rλ T (n) = T (n)Rλ , n ∈ N, λ ∈ C∗ ,
T (m)T (n) = T (mn), if (m, n) = 1,
T (pn )T (p) = T (pn+1 ) + pT (pn−1 )Rp , p prime, n ∈ N.
For the description of the action of these operators on modular forms of weight k, we
associate to f a corresponding function F defined on the set R of lattices by
T (n)f (τ ) = γ(m)q m
m∈Z
with
Using the relations obtained earlier, the proof is straightforward. Again we refer to [Se]
p.101. In particular, we come to
Since the Hecke operators commute, one can hope for simultaneous diagonalizability.
Hecke Eigenforms
Let f be a Hecke eigenform of weight k, i.e. a modular form of weight k, which is not
identically zero and an eigenfunction for all T (n), n ∈ N,
T (n)f (τ ) = λ(n)f (τ ).
Examples are the Eisenstein series Gk and the discriminant ∆. One has
Applying the first statement of the Theorem to the difference of two normalized Hecke
eigenforms, we get
Corollary 1: Two modular forms of weight k, k > 0, which are both normalized Hecke
eigenforms with the same eigenvalues, coincide.
By the second statement of the Theorem the relations (9.11) for the Hecke operators
translate immediately to relations for the Fourier coefficients. We get
n
Now, these relations for the Fourier coefficients of a Hecke eigenform f = −sc(n)q trans-
late to the statement that the associated Dirichlet series Lf (s) = c(n)n (convergent,
as already mentioned, for Re s > k) has an Euler product.
Corollary 3: The Dirichlet series Lf of a Hecke eigenform f of weight k has the Euler
product
1
(9.16) Lf (s) = Πp∈P .
1− c(p)p−s + pk−1−2s
Proof: From Corollary 2 we know that n −→ c(n) is a multiplicative function, so with
t := p−s we can write
∞
c(pn )tn = Lf,p (t) for Lf,p (t) := (1 − c(p)t + pk−1 t2 )−1 .
n=0
We compute
∞
and as coefficient of t get d(1) = c(p) − c(p) = 0 and, for n ≥ 1, as coefficient of tn+1
Remark 9.16: The statement that Lf has an Euler product as in Corollary 3 is equiv-
alent to the fact that the Fourier coefficients of f fulfill the relations in Corollary 2.
The proof is based on the classical Mellin transform: Let f be a normalized Hecke cusp
form of weight k, i.e. we have
∞
0
∞
= c(n) e−2πny y s−1 dy
0
= c(n)(2πn)−s Γ(s)
In the integral on the left hand side, we put u = 1/y and use the transformation formula
for f to get
∞ 0
du
f (iy)y s−1
dy = − f (i/u)u−s
u
0
∞∞
du
= f (iu)ik uk−s
0 u
= ik L∗f (k − s).
Remark 9.18: The whole theory has been extended to other groups than Γ = SL(2, Z),
in particular to Γ0 (N ) (see for instance [Ge]).
There is also a converse to the statement in Remark 9.17: Every Dirichlet series L,
which fulfills a functional equation of the type just treated and has certain regularity
and growth properties, comes from a modular form of f of weight k. This was first re-
marked by Hecke and then refined considerably by Weil in [We1] and is now cited under
the heading of the famous Converse Theorem.
Remark 9.19: The Hecke operators are hermitian with respect to the Petersson scalar
product for forms f1 , f2 of weight k given by
< f1 , f2 > := f1 (τ )f2 (τ )y k−2 dxdy, F as in Remark 9.2,
F
Though it does not directly belong to our topic, we can not resist the temptation to
complete this section by indicating a concept of arithmetically distinguished modular
forms, which has been exploited by Shimura and many others (see [Sh]): We look at
modular forms (for the modular group Γ, but this is open to wide generalizations) with
integral Fourier coefficients
One can show that one can find a Z-basis of Ak (Γ, Z), which is also a C-basis of Ak (Γ).
Prototypes of these forms are, up to factors, forms we already met with, namely
∞
F (τ ) := q (1 − q n )24 ,
n=1
∞
E4 (τ ) := 1 + 240 σ3 (n)q n ,
n=1
∞
E6 (τ ) := 1 − 504 σ5 (n)q n .
n=1
Our analysis of the action of the Hecke operators above shows that Ak (Γ, Z) is stable
under all T (n). Hence the coefficients of the characteristic polynomial of T (n) acting
on Ak (Γ, Z) are integers and one can deduce that the eigenvalues are algebraic integers.
There are explicit formulas for the traces of the T (n), the Eichler-Selberg trace formulas,
starting by [Ei1] and [Sel].
Summary
In this section we looked at L-functions L(s) = Lf (s) coming from modular cusp forms f .
Via a functional equation these functions have analytic continuation to the whole com-
plex plane. This functional equation is a consequence of the covariant transformation
property of the modular form. The modular form is related to an (infinite-dimensional)
automorphic representation of SL(2, R) (or GL(2, R)). We call such L-functions auto-
morphic L-functions. If f is a Hecke eigenform, Lf has an Euler product with factors of
degree 2, i.e. with
Definition 9.5: An algebraic number field is a finite extension K of the field Q of ra-
tional numbers, i.e. a field containing Q, which is a finite-dimensional vector space over Q.
We can embed Q(ζN ) into C in such a way that ζN −→ e(1/N ). But this is not the
only possible embedding of QN into C, since we could also send ζN −→ e(m/N ) where
(m, N ) = 1.
Suppose now that K is an algebraic number field and F a finite extension of K, i.e. another
field containing K, which has a finite dimension as a vector space over K. This dimension
is called the degree of F over K and denoted by degK F, in particular deg F for K = Q.
Definition 9.6: The group of all field automorphisms σ of F, preserving the field struc-
ture and such that σ(x) = x for all x ∈ K, is called the Galois group of F over K and
denoted by Gal(F/K), in particular Gal(F) for K = Q.
Example 9.5: The Galois group Gal(QN ) is naturally identified with the group
with respect to multiplication. The element n̄ ∈ (Z/N Z)× gives rise to the automor-
n m mn
phism of QN sending ζN to ζN , and hence ζN to ζN for all m. If M divides N , then
QM is contained in QN , and the corresponding homomorphism of the Galois groups
Gal(QN ) −→ Gal(QM ) coincides under the above identification with the natural surjec-
tive homomorphism
Definition 9.7: An algebraic field extension F/K is called a Galois extension iff the
order of the Galois group G = Gal(F/K) is equal to the relative degree degK F.
It is a main part of the theory to give different versions of this definition. Then, Galois
theory consists in establishing for a Galois extension a (contravariant) bijection between
subgroups of the Galois group and subfields of F containing K.
Next, one has class field theory, which tries to collect information about a Galois field
extension F/K, in particular the structure of the Galois group, only from the structure
of the basefield K. We shall need some of this and have to recall some more notions and
facts from algebra:
Let R be an integral domain, i.e. a commutative ring with unit e and without zero-
divisors. A subring a in R is an ideal in R iff one has ra ∈ a for all r ∈ R and a ∈ a.
An ideal a is a principal ideal iff it can be generated by one element, i.e. is of the form
a = aR for an element a ∈ R. An ideal p in R is a prime ideal iff from ab ∈ p one can
conclude a or b ∈ p, or, equivalently, iff the factor ring R/p has no zero-divisors.
If for two ideals we have a ⊃ b, we say a divides b. And (a, b) denotes the greatest
common divisor of a and b (corresponding to the usual notions for integers if R = Z and
a = aZ, b = bZ).
From two ideals a and b one can build new ideals, the sum a + b and the product a · b
(the last one consisting of finite sums of products ab, a ∈ a, b ∈ b).
In number theory, the main occasion to use these notions is the maximal order of an
algebraic number field K, i.e. the ring o of integral elements of K: an element a ∈ K
is called integral iff it is the root of a normalized polynomial f ∈ Z[x] (one has to show
that the set of these elements is a ring). The determination of the maximal orders is
a fundamental task of number theory. For instance, for the field K = Q(i) of Gauss
numbers one has the Gauss integers o = Z[i] = {a + bi; a, b ∈ Z} (this is not difficult, try
to prove it as Exercise 9.7). For the cyclotomic field K = Q(ζN ) one has also the nice
result o = Z[ζN ].
Main tools in this context are the maps norm and trace: For x ∈ F one has
λx : F −→ F, a −→ xa for all a ∈ F,
the (additive) trace map
T rF/K : F −→ K, x −→ Tr λx
and the (multiplicative) norm map
NF/K : F −→ K, x −→ det λx .
248 9. Epilogue: Outlook to Number Theory
An ideal in o is also called an ideal of K and this notion is extended to the notion of a
fractional ideal of K: This is an o-submodule a = 0 of K, which has a common denom-
inator, i.e. there is an element 0 = d ∈ o with da ⊆ o.
All this is done with the intention to associate another finite group to an algebraic num-
ber field: It is not difficult to see that the set JK of fractional ideals provided with
multiplication as composition has the structure of an abelian group. It is called the ideal
group of K. The (almost unique) decomposition of an integer in Z into a product of
powers of primes is extended to the statement (see for instance [Ne] p.23):
ClK := JK /PK .
This comes out as a finite abelian group and is called the ideal class group or simply
class group of K.
Class field theory tries to relate the Galois group of a field extension to the ideal class
group of the ground field. An extension F/K with abelian Galois group G is called a
Hilbert class field iff one has
G(F/K) = ClK .
Example 9.6: The cyclotomic field QN is a Hilbert class field over Q. We have
ClQN (Z/N Z)× .
Obviously, for the trivial character χ0 mod 1 with χ0 (n) = 1 for all n, we obtain the
Riemann Zeta series. Hecke proved that these series have an Euler product (with degree
one factors)
1
L(χ, s) = Πp∈P ,
1 − χ(p)p−s
analytic continuation to C (resp. C\{1} for χ0 ), and a functional equation for s −→ 1−s.
(The proof relies again on the transformation formula of a (generalized) theta series.)
9.5 Elements of Algebraic Number Theory and Hecke L-Functions 249
These are examples of Hecke’s L-functions, which can be defined more generally: Let K
be an algebraic number field with maximal order o, m ⊂ o an ideal, and J m the group of
fractional ideals a in K, which are prime to m
J m := {a ⊂ K; (a, m) = 1}.
χ(a)
L(χ, s) := .
a⊂o
N (a)s
Here the sum is taken over all ideals a in o, N (a) denotes the number of elements of the
residue class ring o/a, and again we put χ(a) = 0 for (a, m) = 1. The assumption that χ
is a character translates easily to the existence of an Euler product taken over all prime
ideals p in o
1
L(χ, s) := Πp .
1 − χ(p)N (p)−s
If we take χ = χ0 with χ0 (a) = 1 for all a, we get the usual Dedekind Zeta function of
the number field K
The question asking for the biggest class of characters, for which one can prove a func-
tional equation (and hence has analytic continuation) leads to Hecke’s definition of the
notion “Grössencharakter”:
χf : (o/m)∗ −→ S 1 , χ∞ : R∗ −→ S 1
such that
χ((a)) = χf (a)χ∞ (a) for all a ∈ o with ((a), m) = 1.
We have remarked (in 5.1) that the set of characters of the additive group R is simply
in bijection to R. A sign that multiplicative structure is more intricate than additive
structure is again given by the fact that the set of these characters χ∞ of R∗ is more
complicated (for a complete statement see [Ne] p.497).
As already said, a more elegant way to understand these constructions is based on the
theory of adèles and idèles introduced into number theory by Tate and then pursued by
many others. Here one has the central notion of a Hecke character as a character of the
idel class group. The relation to the notions presented here is also to be found in [Ne]
p.501. There are further refinements using ray class groups and their characters but we
will stop here and change the topic to get again more concrete examples.
250 9. Epilogue: Outlook to Number Theory
Two Examples:
xm + y m = z m .
For m = 2 there are infinitely many solutions, the Pythagorian triples, e.g. (3, 4, 5), and
for m > 2, as Wiles (and Taylor) proved, there are no non-trivial integral solutions. But
if there are no integral solutions to a diophantine problem, one can look for solutions
mod p, p prime, try to count their number and encode these numbers into a function.
(9.18) x2 + 1 = 0.
The primes p for which the congruence x2 + 1 ≡ 0 mod p can be solved are 2, 5, 13, . . .
all of which leave the remainder 1 upon division by 4, whereas primes like 7, 19, 23, . . .
that leave the remainder 3 upon division by 4 never do so. In 9.3 we already presented
the Legendre symbol
a
( ) = ±1 for all a ∈ Z, with (p, a) = 1,
p
χ(p) = ( −1
p ) =1 for p ≡ 1 mod 4,
= −1 for p ≡ 3 mod 4,
= 0 1,
for (p, 4) =
We have
∞
χ(p) −1
χ(n) 1 1 1 1
L(χ, s) = (1 − s
) = s
= 1 − s + s − s + s + ... .
p
p n=1
n 3 5 7 9
Like every Dirichlet series, this series defines a function in a right half plane. It is an
expression of the mysterious symmetry of our problem that one can bring into play a
theta function, which, by its transformation property, produces a function valid for all
s ∈ C.
Now, having prepared some material from algebraic number theory, we can look at the
function L(χ, s) from another point of view: The polynomial f (x) = x2 + 1 defines the
Gauss field
K = Q(i) = {a + bi, a, b ∈ Q}
and this field has the Galois group G = G(K) of automorphisms elementwise fixing Q
consisting
√ just of the identity map and the map ϕ, uniquely fixed by changing the root
θ1 := −1 = i into θ2 := −i. Hence, one has G {±1} and this group has exactly two
one-dimensional representations, the trivial one and a representation π with π(ϕ) = −1.
One can use this to define another type of L-function, the Artin L-function. To explain
this (following Langlands’ article), we go back some steps:
The solvability of the congruence f (x) = x2 + 1 ≡ 0 mod p is equivalent to the possi-
bility to factorize f mod p into a product of two linear factors, e.g.
x2 + 1 ≡ x2 + 2x + 1 = (x + 1)2 mod 2
x + 1 = x − 4 + 5 ≡ (x − 2)(x + 2)
2 2
mod 5.
For the primes p = 7 and 11 and, in general, for all primes, for which the congruence has
no solutions, the polynomial x2 + 1 stays irreducible mod p.
The polynomial is irreducible mod p for p = 7, 13, 19, 29, 43, 47, 59, . . . and factorizes into
linear factors mod p for p = 2063, 2213, 2953, 3631, . . . .
F (α1 , . . . , αn ) = 0.
√ √
For example, the roots of x3 − 1 = 0 are α1 = 1, α2 = (−1 + −3)/2, α3 = (−1 − −3)/2
and two of the valid relations for these roots are
α1 = 1, α2 α3 = α1 .
To the equation f (x) = 0 we can associate the group of all permutations of its roots that
preserve all valid relations. In the last example the sole possibility in addition to the
252 9. Epilogue: Outlook to Number Theory
trivial permutation is the permutation that fixes α1 and interchanges α2 and α3 . The
group G that we get is called the Galois group of the equation f (x) = 0. It is identical
with the Galois group of the algebraic number field K defined by the polynomial f .
We introduce the discriminant ∆ of the equation (resp. the number field) as defined by
the equation
∆ := (αi − αj ).
i=j
where the primes dividing ∆ are omitted. By closer analysis, one can realize that in the
example we get just the Hecke L-function L(χ, s) if we take for ρ the non-trivial character
χ of G {±1}. This is another expression of (a special case of) the main theorem of
abelian class field theory.
the primes dividing the conductor being omitted from the product. This product con-
verges for Re s > 1 and it is a special case of fhe famous Artin Conjecture from 1923 to
show that it can be continued to an analytic function for all s ∈ C. For the equation
(9.19) this was done in 1970 by J. Buhler. The general case is still open.
9.6 Arithmetic L-Functions 253
To start with, we look again at the example K = QN = Q(ζN ): Here we have the abelian
Galois group G = Gal(QN ) (Z/N Z)× . This isomorphism relates every irreducible
representation ρ of G uniquely to a character χ of (Z/N Z)× . To each prime number p,
which does not divide N , we associate the class p̄ := p mod N in (Z/N Z)× and then
p
the Frobenius automorphism Fp ∈ G given by ζN −→ ζN . Now we can define the Artin
L-function
1
L(ρ, s) :=
1 − ρ(Fp )p−s
pN
For the general case, we look at an algebraic field extension F/K with Galois group
G = Gal(F/K) and rings of integral elements O ⊃ o. For a prime ideal P in O and a
prime ideal p in o we say that P lies over p if P contains p and hence P ∩ o = p. We
say that p decomposes into the primes Pi if p generates in O an ideal (p), which has the
decomposition
h
(p) = Pei i
i=1
into the primes Pi lying over p with the ramification indices ei ∈ N. We denote by k(P)
the residue field O/P and similarly k(p) := o/p. Both fields are finite fields. If Pi lies
over p, k(Pi ) is a finite extension of k(p) and we denote fi := degk(p) k(Pi ). These degrees
and the ramification indices are related by the central formula of Hilbert’s ramification
theory
h
ei fi = n.
i=1
From now on, we assume that the extension F/K is a Galois extension. Then the
decomposition simplifies to
h
(p) = ( Pi )e
i=1
and all degrees fi coincide (=: f ). The Galois group G acts on O and moreover even
transitively on the primes Pi lying over a given p ⊂ o. We introduce the decomposition
group
GP := {σ ∈ G; σP = P}.
The index of the decomposition group in G is equal to the ramification index, which is
defined as e = [G : GP ]. One has the two extreme cases:
Since this automorphism preserves the elements of the subfield k(p), we get a homomor-
phism
ϕP : GP −→ Gal(k(P)/k(p)),
which is surjective. We define the inertia group as its kernel, IP := ker ϕP . If p is
unramified, IP consists only of the identity, one has GP Gal(k(P)/k(p)) and, hence, this
group may be seen as a subgroup of G. In this case we can find exacly one automorphism
FP ∈ G such that
FP a ≡ aq mod P for all a ∈ O
with q := N (p). This is the Frobenius automorphism, it generates the cyclic group GP .
If IP is not trivial, we take as Frobenius FP an element of G such that its image in
Gal(k(P)/k(p)) has again the action a mod P −→ aq mod P.
Now we are ready to define Artin’s L-function for this general situation. We take a
representation ρ of G in the finite-dimensional vector space V and denote by V P the
subspace of elements fixed by ρ(IP ). The characteristic polynomial
for the action of ρ(FP ) on V P depends only on the prime p and not on the prime P lying
over p.
Definition 9.9: Let F/K be a Galois extension of algebraic number fields and (ρ, V ) a
representation of its Galois group G. Then Artin’s L-series for ρ is defined by
1
L(ρ, s) := L(F/K, ρ, s) :=
p
det(1 − ρ(FP )N (p)−s ; V IP )
where the product is taken over all prime ideals p in K (i.e. in o).
This series converges in the right half plane Re s > 1 to an analytic function. For K = Q
and the trivial one-dimensional representation ρ0 with ρ0 (σ) = 1 for all σ ∈ G, Artin’s
L-function is just the Dedekind Zeta-function, one has
L(F/Q, ρ0 , s) = ζF (s).
In [Ne] Kapitel VII, we find a lot of results concerning the Artin L-function. To finish
with this topic, we only mention Theorem (10.6) in [Ne], clarifying the relation between
Artin’s and Hecke’s L-function in the abelian case.
Theorem 9.12: Let F/K be an abelian extension, f the conductor of F/K, ρ a nontrivial
character of Gal(F/K) and χ the associated grössencharacter mod f. Then we have the
relation
1
L(F/K, ρ, s) = L(χ, s), S := {p | f, χ(IP ) = 1}.
1 − ρ(FP )N (p)−s
p∈S
Here we should still explain the notion of a conductor, but again we have to refer to [Ne].
9.6 Arithmetic L-Functions 255
Again we follow Langlands’ article [L] and look at another diophantine problem: We ask
for integral or at least rational solutions of the equation
(9.20) y 2 = x3 + Dx,
A curve like this is distinguished among all smooth curves because it can be given the
structure of an abelian group. Elliptic curves now are very popular even outside of
number theory because of their importance for cryptography. On the way to define an
L-function via Euler factors for the primes p, it seems a good idea to count again the
number Np of solutions of (9.20) mod p (or of the associated homogeneous equation, but
we stay here with Langlands in his article). For instance, for D = −1 and p = 5, we get
N5 = 7. We define αp and βp by the conditions
Np = p + αp + βp ; βp = ᾱp ; αp βp = p.
√
(These conditions will define αp and βp only if |Np − p| ≤ 2 p, but it can be proven that
this is true.) Then one should be tempted to define as L-function for the elliptic curve
1 1
L(ED , s) =
p
1 − αp p 1 − βp p−s
−s
For D = (6577)2 , one omits p = 2 and 3 and gets a function for any s and not only for
Re s > 3/2 where the Euler product converges.
The L-functions of elliptic curves are special cases of the Hasse-Weil Zeta functions asso-
ciated to projective varieties over number fields. Starting with work by Hasse, Weil, and
Dwork, and up to work by Deligne, Langlands and many others, there are deep theorems
and conjectures for these relating them again to automorphic L-functions when the vari-
eties are Shimura varieties, i.e. come as homogeneous spaces from a linear group, as for
instance (roughly said) the completion of the modular curve Γ \ H = Γ \ SL(2, R)/SO(2).
There was a famous conjecture, connected mainly with the names of Taniyama, Shimura,
and A. Weil, indicating (again roughly) that each elliptic L-function is an automorphic
L-function belonging to some modular form, which has been proved by Wiles and Tay-
lor in the framework of their proof of the Fermat conjecture already mentioned at the
beginning of this section.
256 9. Epilogue: Outlook to Number Theory
Hence we finish our story by giving a rudimentary survey of what Langlands says in [L]
concerning the archimedian case, in particular as several notions, which we treated in our
chapters 6 and 7, reappear in a language using intuition from elementary particle physics.
If G is not compact, the situation changes considerably as we saw in our discussion of the
example SL(2, R) in 7.2. Langlands proposes that the system being treated is best com-
pared to a Schrödinger equation for which both asymptotically independent and bound
states appear (for instance, in the case of the hydrogen atom, for the electron one has
discrete orbits and a continuous spectrum):
Since every element g ∈ GL(n, R) can be written as a product k1 ak2 where k1 and k2 are
orthogonal matrices and a := D(α1 , . . . , αn ) a diagonal matrix with positive eigenvalues
αj , the simplest representations should have n freely assignable parameters, and they are
analogous to a system of n interacting but asymptotically independent particles to which
arbitrary momenta can be assigned. In addition, the presence of the factors k1 and k2
may entail the presence of discrete quantum numbers. We exemplified this in 7.2.1 while
constructing the principal series representations of SL(2, R) where we had the purely
imaginary “momentum” is and the discrete parameter ∈ {0, 1}. In the GL(n, R) case,
the general induction procedure from 7.1 proposes to look at first at the group B of
9.7 Summary and Final Reflections 257
superdiagonal matrices
⎛ ⎞
α1 ∗ ∗ ... ∗
⎜ α2 ∗ ... ∗ ⎟
⎜ ⎟
⎜ α3 ... ∗ ⎟
(9.21) b = ⎜
⎜
⎟.
⎟
⎜ ... ⎟
⎝ ⎠
αn
χk : α −→ sgn(α)
k |α|isk
Here we have δ = (n + 1)/2 and the supplementary exponent comes from the modular
function and makes things unitary: Similar to our treatment in the special case in 7.2 we
associate to χ the induced representation indGB χ given by right translation on the space
of functions φ on G that satisfy the functional equation
If one replaces the exponents isk by arbitrary complex numbers, one gets representations,
which in general will be neither irreducible nor unitary. But there is a well-determined
process for choosing a specific irreducible factor of the representation that is then taken
as the representation associated to the characters χ1 , . . . , χn . In the intuitive language,
this means that one allows complex momenta.
In our study of SL(2, R) in 7.2 we met with the discrete series representations. Keeping
this in mind, one can deduce that for GL(2, R) there are representations associated to
one continuous parameter (or, if the analogy is pursued, one momentum) and one dis-
crete parameter (or quantum number).
In the treatment of GL(n, R) the case n = 2 plays a special role because there is just
one algebraic extension of the field R, namely C, and this is of degree 2 over R.
For the group GL(n, C) of complex matrices, the representation theory is simpler. One
has no discrete series and therefore no bound states. In Langlands’ language, for K = R
or C the general irreducible representation of GL(n, K) is analogous to r interacting but
asymptotically bound states with n1 , . . . , nr particles, Σnk = n and the nk being subject
to constraints appropriate to the field. To construct the representation, one introduces
the group P of matrices of the form (9.21) where n is replaced by r, each αk is a square
nk –matrix, and the asterisks represent block matrices of the appropriate size. From
here, we can again apply the induction process but we have to replace the character χ by
258 9. Epilogue: Outlook to Number Theory
A similar statement for K = R requires a group that is not commutative but whose
irreducible representations are of degree at most two, in order to accomodate the existence
of two particle bound states. The appropriate group is the Weil group of R, WR . It is
obtained by adjoining to the group C× an element w such that
This is a kind of dihedral group and, in our text, it appeared as an example already in
the Chapters 0 and 1. It has as non-trivial irreducible representations the character π0
given by
z −→ z z̄, w −→ −1
and the two-dimensional representations πm , for 0 = m ∈ Z given by
m
z 0 0 1
z −→ , w −→ .
0 z̄ m (−1)m 0
The Weil group WC of the complex number field C is just C× . One has the nice result
that for K = R and C the irreducible representations of GL(n, K) are parametrized by
n-dimensional representations of the Weil group WK . The Weil group can be attached
to many of the fields appearing in number theory. It can be rather complicated because,
as Langlands assures, it incorporates many of the deepest facts about the theory of equa-
tions known at the present.
The principle that the irreducible representations of the real group G are classified by
the (continuous) homomorphisms of WR into L G is valid, but this has consequences. A
reducible n-dimensional representation of WR is a homomorphism of WR into GL(n, C)
that is isomorphic to C× and not in the center of GL(n, C). Hence the notion of irre-
ducibility has an obvious analogue for the homomorphisms of WR into L G, and it can
be verified that the homomorphisms which are irreducible in this sense correspond to
9.7 Summary and Final Reflections 259
representations whose matrix coefficients are square-integrable over the group. If the
quadratic form chosen is Euclidean, then G is compact and all representations have this
property, so that any homomorphism of WR into L G that is not irreducible in this sense
will have to be excluded from the list of parameters. A similar but less restrictive con-
dition must also be imposed for the groups with other indices.
It also turns out that the classification provided by WR and L G is coarse. Some homo-
morphisms correspond to several irreducible representations of G. This finally leads to
the notion of L-packets which indicates representations with the same L-functions.
For Langlands the introduction of the L-group is an essential step towards his famous
principle of functoriality. A homomorphism from a group H to a group G does not pro-
vide any way to transfer irreducible representations from one of these groups to another,
unless G is abelian, nor are homorphisms between H and G usually related to homomor-
phisms between L H and L G. On the other hand, a homomorphism from L H to L G, by
composition,
WR −→ L H −→ L G
yields a map from the set of parameters for the irreducible representations of H to the
irreducible representations of G, and thus implicitly a way to pass from irreducible rep-
resentations of H to irreducible representations of G. Langlands calls this passage the
principle of functoriality in the L-group. A possibility to realize this principle in general
would have deepest consequences.
D(αis1 , . . . , αisn ), s1 , . . . , sn ∈ R.
Even if we still avoid to go into the representation theory of the non-archimedian field
Qp , it should not be too hard to accept that the analogon of the matrix above in the
p-adic theory is the matrix
In this product perhaps some “special” primes are left out or are represented by special
factors and one has to find an appropriate “factor at ∞” associated to the archimedian
representation. Such a product has the chance to converge for, say, Re s ≥ a + 1, pro-
vided that the eigenvalues of each A(πp ) are less than pa in absolute value. The outcome
is called a standard L-function. Whoever wants to see that this really happens (and
even more...) should be motivated to take a look into the non-archimedian theory, as for
instance in [Ge] or [GeS].
To finish our fantastic and magical story (the German word Märchen appears well in the
title of one of Langlands’ articles ([L1])), we look at an (automorphic) representation π
of a group G (in particular G = GL(n)), take a finite-dimensional representation ρ of
the L-group L G belonging to G. Then we have again a finite Euler polynomial given by
det(E − ρ(A(πp )t) and one can introduce as automorphic L-function
1
L(π, ρ, s) :=
det(E − ρ(A(πp ))p−s )
where again for some primes p we have to put in special factors, which we do not discuss
here. In particular, one has to find a completing factor belonging to the archimedian
theory. The final goal is to generalize what we sketched in the last section by showing
that to given π and ρ there is an automorphic representation Π such that
A(Πp ) = ρ(πp )
L(π, ρ, s) = L(Π, s)
There are many other important topics coming up in this context. We only mention the
topic of the trace formula but can not touch this here. The only theme whose formulation
is within our reach is the following: Given a representation π of a group G = GL(n) and
a representation π of G = GL(n ), by the methods from of our text one can construct
representations π + π of GL(n + n ) and π ⊗ π of GL(nn ). It is natural to ask for a
similar construction for automorphic representations. This is related to functoriality and
up to now there are very few results.
We repeat our hope that the vagueness of the presentation of some of the material in
the last sections would make the reader curious to know more about all this and to be
stimulated to more intense studies of this most fascinating part of mathematics. All the
time, we were driven by J. R. Jiménez’ words:
[BR] Barut, A.O., Raczka, R.: Theory of Group Representations and Applications.
PWN Polish Scientific Publishers Warszawa 1980.
[Ba] Bargmann, V.: Irreducible Unitary Representations of the Lorentz Group. An-
nals of Math. 48 (1947) 568-640.
[Be1] Berndt, R.: The Heat Equation and Representations of the Jacobi Group. Con-
temporary Mathematics 389 (2006) 47 – 68.
[BeS] Berndt, R. Schmidt, R.: Elements of the Representation Theory of the Jacobi
Group. PM 163, Birkhäuser, Basel 1998.
[BR] Brateli, O., Robinson, D.W.: Operator Algebra and Quantum Statistical Me-
chanics. Springer, New York 1979.
[BtD] Bröcker, T., tom Dieck, T.: Representations of Compact Lie Groups. Springer,
New York 1985.
[CF] Cassels, J.W.S., Fröhlich, A.: Algebraic Number Theory. Thompson, Washing-
ton, D.C. 1967.
[CN] Conway, J.H., Norton, S.P.: Monstrous Moonshine. Bull. London Math. Soc.
11 (1979) 308 - 339.
[Co] Cornwell, J. F.: Group Theory in Physics. Academic Press, London 1984.
[Do] Donley, R. W.: Irreducible Representations of SL(2, R). p.51 - 59 in: Repre-
sentation Theory and Automorphic Forms (Bailey, T.N., Knapp, A.W., eds.),
PSPM Vol. 61 AMS 1997.
[Du] Duflo, M.: Théorie de Mackey pour les groupes de Lie algébriques. Acta Math.
149 (1982) 153 - 213.
262 BIBLIOGRAPHY
[Ei] Eichler, M.: Einführung in die Theorie der algebraischen Zahlen und Funktio-
nen. Birkhäuser, Basel 1963.
[Ei1] Eichler, M.: Einige Anwendungen der Spurformel im Bereich der Modularkor-
respondenzen. Math. Ann. 168 (1967) 128 - 137.
[EZ] Eichler, M., Zagier, D.: The Theory of Jacobi Forms. Birkhäuser, Boston 1985.
[Fi] Fischer, G.: Lineare Algebra. Vieweg, Wiesbaden 2005.
[Fl] Flicker. Y. Z.: Automorphic Forms and Shimura Varieties of PGSp(2). World
Scientific, New Jersey 2005.
[Fog] Fogarty,J.: Invariant Theory. Benjamin,New York 1969.
[Fo] Forster, O.: Analysis 1. Vieweg, Wiesbaden 2006.
[FB] Freitag, E., Busam, R.: Funktionentheorie. Springer, Berlin 1993.
[Fr] Frenkel, E.: Lectures on the Langlands Program and Conformal Field Theory.
arXiv:hep-th/0512172v1 15 Dec 2005.
[FLM] Frenkel, E., Lepowski, J., Meurman, A.: Vertex Operator Algebras and the
Monster. Academic Press, Boston 1988.
[FS] Fuchs, J., Schweigert, Chr.: Symmetries, Lie Algebras and Representations.
Cambridge University Press 1997.
[FH] Fulton, W., Harris, J.: Representation Theory. GTM 129, Springer, New York
1991.
[Ge] Gelbart, S.: Automorphic Forms on Adele Groups. Annals of Math. Studies 83,
Princeton University Press 1975.
[GeS] Gelbart, S., Shahidi, F.: Analytic Properties of Automorphic L-Functions. Aca-
demic Press, Boston 1988.
[Gel] Gell-Mann, M.: Phys. Rev. 125 1067-1084, 1962.
[GH] Griffiths, P., Harris, H.: Principles of Algebraic Geometry. Wiley, New York
1978.
[GGP] Gelfand, I., Graev, M., Pyatetskii-Shapiro, I.: Representation Theory and Au-
tomorphic Functions. W.B. Saunders, Philadelphia 1963.
[Go] Goldstein, H.: Classical Mechanics. Addison-Wesley, Reading 1980.
[GS] Guillemin, V., Sternberg, S.: Symplectic Techniques in Physics. Cambridge
University Press 1984.
[GS1] Guillemin, V., Sternberg, S.: Geometric Asymptotics. Math. Surv. and Mono-
graphs 14, AMS 1990.
[Ha] Halmos, P.: Measure Theory. Van Nostrand, New York 1950.
[He] Hein, W.: Struktur- und Darstellungstheorie der klassischen Gruppen. Springer
HT Berlin 1990.
BIBLIOGRAPHY 263
[HN] Hilgert, J., Neeb, K.-H.: Lie-Gruppen und Lie-Algebren. Vieweg, Braun-
schweig/Wiesbaden 1991.
[HR] Hewitt, E., Ross, K.A.: Abstract Harmonic Analysis I. Springer, Berlin 1963.
[Hu] Humphreys, J.E.: Introduction to Lie Algebras and Representation Theory.
GTM 9, Springer New York 1972.
[Ig] Igusa, J.: Theta Functions. Springer, Berlin 1972.
[JL] Jacquet, H., Langlands, R.P.: Automorphic Forms on GL(2). LNM 114,
Springer, New York 1970.
[Ja] Jacobson, N.: Lie Algebras. Interscience, New York 1962.
[Ki] Kirillov, A.A.: Elements of the Theory of Representations. Springer, Berlin
1976.
[Ki1] Kirillov, A.A.: Lectures on the Orbit Method. GSM 64, AMS 2004.
[Ki2] Kirillov, A.A.: Unitary Representations of Nilpotent Lie Groups. Uspekhi Mat.
Nauk 17 (1962), 57 - 110; English transl. in Russian Math. Surveys 17 (1962).
[Ki3] Kirillov, A.A.: Merits and Demerits of the Orbit Method. Bulletin of the AMS
36 (1999) 433 - 488.
[Ko] Koecher, M.: Lineare Algebra. Springer, Berlin 1983.
[Kos] Kostant, B.: Quantization and Unitary Representations. In Lectures in Modern
Analysis III. (ed. Taam, C.T.) LNM 170, Springer, Berlin 1970.
[Kn] Knapp. A.W.: Representation Theory of Semisimple Groups. An Overview
Based on Examples. Princeton University Press 1986.
[Kn1] Knapp. A.W.: Structure Theory of Semisimple Lie Groups. p.1 - 27 in: Repre-
sentation Theory and Automorphic Forms (Bailey, T.N., Knapp, A.W., eds.),
PSPM Vol. 61, AMS 1997.
[Kn2] Knapp. A.W.: Lie Groups Beyond an Introduction. PM 140, Birkhäuser, Boston
1996.
[KT] Knapp, A.W., Trapa, P.E.: Representations of Semisimple Lie Groups. p. 7
- 87 in: Representation Theory of Lie Groups (Adams, J., Vogan, D., eds.),
IAS/Park City Math. Series 8, AMS 2000.
[Ku] Kubota, T.: Elementary Theory of Eisenstein Series. Halsted Press, New York
1973.
[La] Lang, S.: Algebra. Addison-Wesley, Reading, Mass. 1965.
[La1] Lang, S.: SL(2,R). Springer, New York 1985.
[La2] Lang, S.: Algebraic Number Theory. Addison - Wesley, Reading, Mass. 1970.
[L] Langlands, R.P.: Representation Theory: Its Rise and Its Role in Number The-
ory. p. 181-210 in Proceedings of the Gibbs Symposium, Yale University 1989,
AMS 1990.
264 BIBLIOGRAPHY
[LV] Lion, G., Vergne, M.: The Weil representation, Maslov index and Theta series.
Birkhäuser, Boston 1980.
[LZ] Lewis, J., Zagier, D.: Period functions for Maass wave forms I. Ann. Math.
153 (2001) 191-253.
[Mu] Mumford, D.: Tata Lectures on Theta I,II,III. PM 28, 43, 97, Birkhäuser, Boston
1983, 1984, 1991.
[Na] Naimark, M.A.: Linear Representations of the Lorentz Group. Pergamon Press,
London 1964.
[RV] Ramakrishnan, D., Valenza, R.J.: Fourier Analysis on Number Fields. GTM
186, Springer, New York 1999.
[Se] Serre, J.P.: Linear Representations of Finite Groups. Springer, New York 1977.
[Se1] Serre, J.P.: A Course in Arithmetics. GTM 7, Springer, New York 1973.
[Sel] Selberg, A.: Harmonic analysis and discontinuous groups in weakly symmetric
riemannian spaces with application to Dirichlet series. J. Indian Math. Soc. 20
(1956) 47 - 87.
[vD] van Dijk, G.: The irreducible Unitary Representations of SL(2, R). In: Rep-
resentations of Locally Compact Groups with Applications. ed. Koornwinder,
T.H. Mathematisch Centrum Amsterdam 1979.
[Ve] Vergne, M.: Geometric Quantization and Equivariant Cohomology. First Eu-
ropean Congress of Mathematics, Vol I, p.249–298. PM 119, Birkhäuser, Basel
1994.
[Vo] Vogan, D.A.: The Method of Coadjoint Orbits for Real Reductive Groups. p.179
– 238 in: Representation Theory of Lie Groups (Adams, J., Vogan, D., eds.),
IAS/Park City Math. Series 8 AMS 2000.
[Vo1] Vogan, D.A.: Associated Varieties and Unipotent Representations. p.315 – 388
in: Harmonic Analysis on Reductive Lie Groups (Barker, W., Sally, P., eds.),
Birkhäuser, Boston 1991.
[Vo2] Vogan, D.A.: Cohomology and group representations. PSPM Vol 61, 219–243
(1997).
[Wa] Warner, G.: Harmonic Analysis on Semi-Simple Lie Groups. Springer, Berlin
1972.
[We] Weil, A.: Variétés Kählériennes. Hermann, Paris 1957.
[We1] Weil, A.: Über die Bestimmung Dirichletscher Reihen durch ihre Funktionalglei-
chung. Math, Ann. 168 149-156 (1967).
[Wo] Woodhouse, N.: Geometric Quantization. (Second Edition) Oxford University
Press 1991.
[Ya] Yang, Y.-H.: The Method of Orbits for Real Lie Groups. Kyungpook Math. J.
42 (2002) 199–272.
Index
symmetric, 34 continuous, 34
orbit contragredient, 16
admissible, 196 cyclic, 37
coadjoint, 186 decomposable, 16, 36
elliptic, 188 derived, 70
hyperbolic, 188 discrete series, 142
nilpotent, 188 discrete series of SL(2, R), 133
rigged, 197 disjoint, 9
eigenspace, 56
parabolic induction, 130 equivalent, 9
parabolic subgroup, 130 factor representation, 16
partition, 6 Fock representation, 207
Pauli matrices, 53 induced, 119, 164
permutation matrix, 6 irreducible, 8, 36
Peter–Weyl Theorem, 48 lattice, 219
Petersson scalar product, 244 Lie algebra, 66
phase space, 176 linear, 7
Poincaré group, 154 permutation representation , 11
two-dimensional, 127 primary, 20
Poincaré half plane, 133
principal series, 130
Poincaré series, 225
projective representation, 55
Poincaré-Birkhoff-Witt Theorem, 100
regular representation of a compact group,
Poisson algebra, 64
47
Poisson bracket, 64, 177
regular representation of a finite group,
polarization
27
admissible, 196
Schrödinger representation, 13, 118, 152,
algebraic approach, 196
206
geometric approach, 184
Schrödinger-Weil representation, 211
Pontrjagin dual, 59
smooth, 70
prequantum Hilbert space, 180
unitary, 9
prime ideal, 247
representations
principal series of SL(2, C), 143
of SL(2, C), 143
principal series of SL(2, R), 130, 200, 235
principle of functoriality, 259 of SL(2, R), 130
projection operator, 20 of SO(3), 54
projective representation, 55, 145 of SU(2), 48
projective space, 55, 129, 166 of the Poincaré group GP , 157
proton, 110 of the two-dimensional Poincaré group,
127
q-development, 222 unitarily integrable representations of
quantization, 177 sl(2, R), 78
quasi-measure relation, 121 unitarily integrable representations of
su(2), 82
Radon-Nikodym derivative, 121 Riemann Zeta function, 224, 239
Ramanujan function, 223 right translation, 4
representation root, 90
adjoint, 186 positive, 92
admissible, 143 reduced, 92
coadjoint, 186 restricted, 96
completely reducible, 16, 36 simple, 92
270 INDEX
vacuum vector
of Heis(R), 83
vector field, 170, 174
Hamiltonian, 174
vector of lowest weight, 75
Verma module, 101
weight, 74, 97
algebraically integral, 98
Weil group of R, 2, 13, 258