Escolar Documentos
Profissional Documentos
Cultura Documentos
Stefan Weinzierl
February 13, 2012
1
Contents
1 Introduction 4
1.1 Literature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
3 Lie groups 22
3.1 Manifolds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.1.1 Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.1.2 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
3.1.3 Morphisms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
3.2 Lie groups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
3.2.1 Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
3.2.2 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
3.3 Algebras . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
3.3.1 Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
3.3.2 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
3.4 Lie algebras . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
3.4.1 Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
3.4.2 The exponential map . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
3.4.3 Relation between Lie algebras and Lie groups . . . . . . . . . . . . . . . 29
3.4.4 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.4.5 The Fierz identity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
3.5 Gauge symmetries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
4 Representation theory 38
4.1 Group actions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
4.2 Representations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
4.3 Schurs lemmas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
4.4 Representation theory for finite groups . . . . . . . . . . . . . . . . . . . . . . . 44
4.4.1 Characters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
4.4.2 Application: Molecular vibration . . . . . . . . . . . . . . . . . . . . . 49
2
4.4.3 Application: Quantum mechanics . . . . . . . . . . . . . . . . . . . . . 53
4.5 Representation theory for Lie groups . . . . . . . . . . . . . . . . . . . . . . . . 56
4.5.1 Irreducible representation of SU (2) and SO(3) . . . . . . . . . . . . . . 56
4.5.2 The Cartan basis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
4.5.3 Weights . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
4.6 Tensor methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
4.6.1 Clebsch-Gordan series . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
4.6.2 The Wigner-Eckart theorem . . . . . . . . . . . . . . . . . . . . . . . . 70
4.6.3 Young diagrams . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
4.7 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
4.7.1 Selection rules in quantum mechanics . . . . . . . . . . . . . . . . . . . 76
4.7.2 Gauge symmetries and the Standard Model of particle physics . . . . . . 77
4.7.3 The naive parton model . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
3
1 Introduction
1.1 Literature
Introductory texts:
- J.P. Elliot and P.G. Dawber, Symmetry in Physics, Macmillan Press, 1979
- W. Fulton and J. Harris, Representation theory, Springer, 1991
- B. Hall, Lie groups, Lie Algebras and Representations, Springer, 2003
Physics related:
- D.B. Lichtenberg, Unitary symmetry and elementary particles, Academic Press, 1970
- H. Georgi, Lie Algebras in Particle Physics, Benjamin/Cummings Publishing Company,
1982
- M. Schottenloher, Geometrie und Symmetrie in der Physik, Vieweg, 1995
- M. Nakahara, Geometry, Topology and Physics, IOP, 1990
Classics:
Differential geometry:
- S. Helgason, Differential Geometry, Lie Groups and Symmetric Spaces, AMS, 1978
Hopf algebras:
Specialised topics:
1.2 Motivation
Symmetries occur in many physical systems, from molecules, crystals, atoms, nuclei to elemen-
tary particles. Symmetries occur in classical physics as well as in quantum physics. A symmetry
is expressed by a transformation, which leave the physical system invariant. Examples for
such transformations are translations, rotations, inversions, particle interchanges. The symmetry
transformations form a group and we are led to study group theory.
4
A few examples:
mx = 0
mx = const,
x = r cos ,
y = r sin
we obtain
x V y V V V
V = + = y +x ,
x y x y
and therefore
V V d
0 = V = y +x = y (mx) + x (my) = m (yx xy) .
x y dt
5
This shows that the angular momentum is conserved:
qi = fi (~q,t, ), 1 i n,
depending on a real parameter , such that all functions fi are continously differentiable with
respect to and sucht that = 0 corresponds to the identity transformation
qi = fi (~q,t, 0), 1 i n.
+ d (~q,t, ) + O 2 ,
L ~q ,~q ,t = L ~q,~q,t
dt
with
(~q,t, 0) = 0,
is conserved. This is Noethers theorem. It states the every continuously symmetry of the La-
grange function leads to a conserved quantity. Note that it is only required that the transformed
Lagrange function agrees with the original one only up to a gauge transformation.
mx1 = x1 (x1 + x2 ) ,
mx2 = x2 (x1 + x2 ) .
6
This suggests new coordinates
q1 = x1 + x2 ,
q2 = x1 x2 .
mq1 = 3q1 ,
mq2 = q2 .
In terms of the coordinates q1 and q2 the solutions are harmonic oscillations with frequencies
r r
3
1 = , 2 = .
m m
Let us denote the symmetry transformation x1 x2 by . The new coordinates q1 and q2 are
even and odd, respectively, under the symmetry transformation:
q1 = q1 ,
q2 = q2 .
V (~x) = V (~x) ,
it follows that also the Hamilton operator has this property: H(~x) = H(~x). Now, if (~x) is an
eigenfunction of the Hamilton operator with eigenvalue E,
H(~x) = E(~x),
H(~x) = E(~x).
(~x) = c(~x).
(~x) = c2 (~x),
and thus
c = 1.
7
Hence, (~x) is either even or odd. This leads to selection rules in quantum mechanics: The
transition probability for a decay from some initial state i to a final state f is proportional to
the square of hte integral
Z
I = d 3 x f (~x) O (~x) i (~x)
where O (~x) depends on the particular decay process. If O (~x) is an even function of~x, the integral
is non-zero only if i and f are both even or both odd.
Summary: Understanding the symmetry properties of a physical system is useful for the follow-
ing reasons:
Gauge symmetries are the key ingredient for the understanding of the fundamental forces.
8
2 Basics of group theory
2.1 Definition of a group
A non-empty set G together with a composition : G G G is called a group (G, ) if
The first system of axioms clearly implies the second system of axioms. To show that the second
system also implies the first one, we show the following:
a) If e is a left-neutral element, and e is a right-neutral element, then e = e .
Proof:
e = e e e is left-neutral
= e, e is right-neutral
b = be e is right-neutral
= b ab b is right-inverse of a
= (b a) b associativity
= e b b is left-inverse of a
= b e is left-neutral
(a b) (a b) = a (b a) b
= aeb
= ab
9
Therefore a b = e.
d) If e is the left-neutral element, then e is also right-neutral.
Proof:
a = ea
= a1 a a
= a a1 a
= a a1 a
= ae
This completes the proof that the second system of axioms is equivalent to the first system of
axioms. To verify that a given set together with a given composition forms a group it is therefore
sufficient to verify axioms (G2) and (G3) instead of axioms (G2) and (G3).
More definitions:
A group (G, ) is called Abelian if the operation is commutative : a b = b a.
The number of elements in the set G is called the order of the group. If this number is fi-
nite, we speak of a finite group. In the case where the order is infinite, we can further distinguish
the case where the set is countable or not. For Lie groups we are in particular interested in the
latter case. For finite groups we can write down all possible compositions in a composition
table. In such a composition table each element occurs exactly once in each row and column.
Examples
a) The trivial example: Let G = {e} and e e = e. This is a group with one element.
b) Z2 : Let G = {0, 1} and denote the composition by +. The composition is given by the
following composition table:
+ 0 1
0 0 1
1 1 0
Z2 is of order 2 and is Abelian.
c) Zn : We can generalise the above example and take G = {0, 1, 2, ..., n 1}. We define the
addition by
a + b = a + b mod n,
where on the l.h.s. + denotes the composition in Zn , whereas on the r.h.s. + denotes the
usual addition of integer numbers. Zn is a group of order n and is Abelian.
10
d) The symmetric group Sn : Let X be a set with distinct n elements and set
G = {| : X X permutation of X }
As composition law we take the composition of permutations. The symmetric group has order
|Sn | = n!
e) (Z, +): The integer numbers with addition form an Abelian group. The order of the group is
infinite, but countable.
f) (R, +): The real numbers with addition form an Abelian group. The order of the group is
not countable.
g) (R , ): Denote by R = R\{0} the real numbers without zero. The set R with the mul-
tiplication as composition law forms an Abelian group.
together with matrix multiplication as composition. To check this, one has to show that
cos sin cos sin
sin cos sin cos
Using the addition theorems of sin and cos one finds = + . The elements of this group are
not countable, but they form a compact set.
11
2.2 Group morphisms
Let a and b be elements of a group (G, ) with composition and let a and b be elements of a
group (G , ) with composition . We are interested in mappings between groups which preserve
the structure of the compositions.
f (a b) = f (a) f (b).
Automorphism: We call a mapping f : G G from the group G into the group G itself an
automorphism, if it is an isomorphism.
G = (Zn , +) ,
and
n 0 1 (n1)
o
G = e2i n , e2i n , ..., e2i n , .
This is the group of the n-th roots of unity. These two groups are isomorphic. The isomorphism
is given by
f : G G ,
k
k e2i n .
2.3 Subgroups
A non-empty subset H G is said to be a subgroup of a group G, if H is itself a group under
the same law of composition as that of G.
Z6 = {0, 1, 2, 3, 4, 5}.
12
The set
H = {0, 2, 4}
is a subgroup of Z6 , isomorphic to Z3 . The isomorphism is given by
f : H Z3 ,
2n n.
Generating set: Let G be a group and S G. The group
hSi = {U |U subgroup of G with S U}
is called the subgroup generated by S. We say that G is generated by S, if
G = hSi.
We say that G is finitely generated if there is a finite set S = {a1 , ..., an} such that
G = ha1 , ..., ani.
Note that finitely generated does not imply that the group has a finite number of elements. The
group (Z, +) is generated by the element 1, but has an infinite number of elements.
The orders of the elements of a group give us useful information about the group: Let us consider
the following example. Assume that we are considering a group with four elements {e, a, b, c}.
Assume further that the elements a and b are of order two:
a2 = b2 = e.
Based on this knowledge we know the following entries in the composition table:
13
e a b c
e e a b c
a a e ?
b b e
c c
In the place with the question mark we must put c, otherwise c would occur in the last col-
umn twice. All other entries can be figured out with similar considerations and we obtain the
composition table
e a b c
e e a b c
a a e c b
b b c e a
c c b a e
2.4 Cosets
Consider a subgroup H of a group G. Let a G. The set
is called a left coset of H in G. The number of distinct left cosets of H in G is called the index
of H in G and is denoted by
|G : H|
Theorem (Lagrange):
1.
[
G = aH
aG
3. If two of the numbers |G|, |H| and |G : H| are finite, then also the third one is and we have
the relation
|G| = |G : H| |H| .
Proof:
14
On the other hand it is obvious that
[
aH G
aG
a1 h1 = a2 h2
1
a2 a1 = h2 h1
1 H
a1
2 a1 H = H
a1 H = a2 H,
|aH| = |H| .
The theorem of Lagrange has some important consequences for finite groups: Let G be a finite
group.
The order |H| of each subgroup divides the order |G| of the group.
The order of each element a divides the order |G| of the group.
a|G| = e.
As a further consequence we have the following: If G is a finite group, where the order is a prime
number, then G is cyclic.
b = gag1 .
a is conjugated to itself.
15
If a is conjugated to b, then b is conjugated to a.
The set of all elements conjugate to a is called the conjugacy class of a. The set of all elements
of a group can be decomposed into disjoint conjugacy classes.
It is often useful to consider conjugacy classes instead of all individual elements, because group
elements of the same conjugacy class will behave similar.
aNa1 N.
Remark: This means that for all n1 N and for all a G there exists an n2 N such that
an1 a1 = n2 .
For a normal subgroup the left and right cosets are equal
aN = Na.
This is exactly the property we need such that the left cosets have a well-defined composition
law:
or in more detail
(an1 ) (bn2 ) = a (n1 b) n2 = a bn1 n2 = ab n1 n2 = (ab) n3 .
|{z}
n3
G/N = {aN|a G}
16
is a group, called the factor group of G by N.
Remark: We have
|G/N| = |G : N| .
Example: Let us consider the dihedral group Dn for n 3. This is the symmetry group of a
regular polygon. The symmetry operations are rotations through an angle 2/n and reflexions.
In mathematical terms, this group is generated by two elements a and b with
an = b2 = e, ab = ban1 .
The element a corresponds to a rotation through the angle 2/n, hence an = e, the element b
corresponds to a reflexion, hence b2 = e. This group is non-Abelian and one easily convinces
oneself that ab = ban1 (note that a1 = an1 ). The element a generates a cyclic subgroup hai
of order n. We can write Dn as left cosets of hai:
Dn = hai bhai.
This shows that Dn has (2n) elements. The subgroup hai is a normal subgroup:
17
2.7 Direct product
Let G1 , ..., Gn be groups. In the set
we define a composition by
G is called the direct product of the groups G1 , ..., Gn . The neutral element in G is given by
(e1 , ..., en), the inverse element to (a1 , ..., an) is given by (a1 1
1 , ..., an ).
Gi = E ... E} Gi E
| {z | {z
... E}
(i1) times (ni) times
G = N1 N2 ...Nn ,
Ni (N1 ...Ni1Ni+1 ...Nn) = {e} .
Furthe properties of inner direct products: If G = N1 ...Nn is an inner direct product, then we have
ai a j = a j ai .
a = a1 ...an,
with ai Ni .
18
The converse is also true: If Gi (1 i n) are subgroups of G, such that the elements of Gi
commute with the elements of G j (for i 6= j) and each element a G has a unique representation
a = a1 ...an
with ai Gi , then the Gi s are normal subgroups of G and G is an inner direct product of the Gi s.
Consequence: If n and m are two positive numbers, which share no common divisor, then
Znm
= Zn Zm .
Without a proof we state the following theorem, which characterises completely the finitely gen-
erated Abelian groups:
Theorem on finitely generated Abelian groups: Every finitely generated Abelian group G
is a direct product of finitely many cyclic groups, in other words
G
= Z pk1 ... Z pkr r Z ... Z,
1
An example where direct products occur in physics is given by the gauge symmetry of the Stan-
dard Model of particle physics. The gauge group is given by
where U (1) is the gauge group corresponding to the hypercharge, SU (2) is the gauge group
corresponding to the weak isospin and SU (3) is the gauge group corresponding to the colour
charges.
First theorem of Sylow: Let G be a finite group of order n = pr m, where p is a prime num-
ber and gcd(p, m) = 1. Then there exists for each j with 1 j r a subgroup H of G with order
p j.
Corollary: Let G be a finite group and let p be a prime number, which divides the order of
G. Then G contains an element of order p.
19
Definition: Let p be a prime number and let G be a group. We call G a p-group, if the or-
der of every element is a power of p.
Second theorem of Sylow: Let G be a finite group of order n = pr m, where p is a prime number
and gcd(p, m) = 1. Let P be a p-Sylow group of G and let H be a p-subgroup of G. Then there
exists an element a such that
aHa1 P.
Corollary: Assume that for a given prime number p there is only one p-Sylow group P of G.
Then P is a normal subgroup.
Third theorem of Sylow: Let G be a finite group and assume that the prime number p is a
divisor of the order of the group. Then the number of the p-Sylow groups of G is also a divisor
of the order of the group and of the form
1 + kp,
with k 0.
aG = {ag|g G} .
We have
aG = G.
20
An important application of the group rearrangement theorem is the following: Consider a func-
tion
f : GR
f (a) = f (ab)
aG aG
for all b G.
21
3 Lie groups
3.1 Manifolds
3.1.1 Definition
A topological space is a set M together with a family T of subsets of M satisfying the following
properties:
1. 0/ T , M T
2. U1 ,U2 T U1 U2 T
A topological space is called Hausdorff if for any two distinct points p1 , p2 M there exists
open sets U1 ,U2 T with
/
p1 U1 , p2 U2 , U1 U2 = 0.
A map between topological spaces is called continous if the preimage of any open set is again
open.
22
If p U and
(p) = (x1 (p), ..., xn(p)) ,
the set U is called the coordinate neighbourhood of p and the numbers xi (p) are called the
local coordinates of p.
Note that in each coordinate neighbourhood M looks like an open subset of Rn . But note that we
do not require that M be Rn globally.
Consider two manifolds M and N with dimensions m and n. Let xi be coordinates on M and
y j be coordinates on N. A mapping f : M N between two manifolds is called analytic, if for
each point p M there exits a neighbourhood U of p and n power series Pj , j = 1, ..., n such that
y j ( f (q)) = Pj (x1 (q) x1 (p), ..., xm(q) xm (p))
for all q U .
3.1.2 Examples
a) Rn : The space Rn is a manifold. Rn can be covered with a single chart.
b) S1 : The circle
S1 = {~x R2 ||~x|2 = 1}
is a manifold. For an atlas we need at least two charts.
3.1.3 Morphisms
Homeomorphism: A map f : M N between two manifolds M and N is called a homeo-
morphism if it is bijective and both the mapping f : M N and the inverse f 1 : N M are
continous.
23
3.2 Lie groups
3.2.1 Definition
A Lie group G is a group which is also an analytic manifold, such that the mappings
G G G,
(a, b) a b,
and
G G,
a a1
are analytic.
Remark: Instead of the two mappings above, it is sufficient to require that the mapping
G G G,
(a, b) a b1
is analytic.
3.2.2 Examples
The most important examples of Lie groups are matrix groups with matrix multiplication as com-
position. In order to have an inverse, the matrices must be non-singular.
a) GL(n, R), GL(n, C): The group of non-singular n n matrices with real or complex entries.
GL(n, R) has n2 real parameters, GL(n, C) has 2n2 real parameters.
b) SL(n, R), SL(n, C): The group of non-singular n n matrices with real or complex entries
and
det A = 1.
SL(n, R) has n2 1 real parameters, while SL(n, C) has 2(n2 1) real parameters.
RRT = 1.
The group O(n) has n(n 1)/2 real parameters. The group O(n) can also be defined as the
transformation group of a real n-dimensional vector space, which preserves the inner product
n
x2i
i=1
24
d) SO(n): The group of special orthogonal n n matrices defined through
UU = 1.
The group U (n) has n2 real parameters. The group U (n) can also be defined as the transformation
group of a complex n-dimensional vector space, which preserves the inner product
n
zi zi
i=1
UU = 1 and det U = 1.
The group Sp(n, R) has (2n + 1)n real parameters. The group Sp(n, R) can also be defined as the
transformation group of a real 2n-dimensional vector space, which preserves the inner product
n
x j y j+n x j+n y j .
j=1
3.3 Algebras
3.3.1 Definition
Let K be a field and A a vector space over the field K. A is called an algebra, if there is an
additional composition
AA A
(a1 , a2 ) a1 a2
25
Remark: It is not necessary to require that K is a field. It is sufficient to have a commutative ring
R with 1. In this case one replaces the requirement for A to be a vector space by the requirement
that A is an unital R-modul. The difference between a field K and a commutative ring R with 1
lies in the fact that in the ring R the multiplicative inverse might not exist.
a1 a2 = a2 a1
1A a = a.
Note that it is not required that A has a unit element. If there is one, note that difference between
1A A and 1K K: The latter always exists and we have the scalar multiplication with one:
1K a = a.
3.3.2 Examples
a) Consider the set of n n matrices over R with the composition given by matrix multiplication.
This gives an associative, non-commutative algebra with a unit element given by the unit matrix.
[a, b] = ab ba.
[a, a] = 0,
[a, [b, c]] + [b, [c, a]] + [c, [a, b]] = 0.
Remark: Consider again the example above of the set of n n matrices over R where the com-
position is defined by the commutator
[a, b] = ab ba.
26
Clearly this definition satisfies [a, a] = 0. It fullfills the Jacobi identity:
[a, [b, c]] + [b, [c, a]] + [c, [a, b]] =
= abc acb bca + cba + bca bac cab + acb + cab cba abc + bac
= 0.
Matrix algebras with the commutator as composition are therefore Lie algebras.
Let A be a Lie algebra and X1, ..., Xn a basis of A as a vector space. [Xi , X j ] is again in A and
can be expressed as a linear combination of the basis vectors Xk :
n
ci jk Xk .
Xi , X j =
k=1
The coefficents ci jk are called the structure constants of the Lie algebra. For matrix algebras the
Xi s are anti-hermitian matrices.
The notation above is mainly used in the mathematical literature. In physics a slightly differ-
ent convention is often used: Denote by T1 , ..., Tn a basis of A as a (complex) vector space. Then
n
[Ta , Tb ] = i fabc Tc .
c=1
A few properties:
1. We have
exp(0) = 1.
27
2. exp X is invertible and
(exp X )1 = exp (X ) .
3. We have
4. If XY = Y X then
5. If A is invertible then
exp AX A1 = A exp (X ) A1 .
6. We have
d
exp (tX ) = X exp (tX ) = exp (tX )X .
dt
In particular
d
exp (tX ) = X.
dt t=0
Point 1 is obvious. Points 2 and 3 are special cases of 4. To prove point 4 it is essential that X
and Y commute:
Xi Y j n
X i Y ni 1 n n
exp X expY = = = X iY ni
i=0 i! j=0 j! n=0 i=0 i! (n i)! n=0 n! i=0
i
1
= n! (X +Y )n = exp (X +Y ) .
n=0
Proof of point 5:
1 1
n! n! AX nA1 = A exp (X ) A1.
1 n
exp AX A1 = AX A =
n=0 n=0
Proof of point 6:
d d 1
1
exp (tX ) = (tX ) =
n
t n1 X n = X exp (tX ) = exp (tX )X .
dt dt n=0 n! n=0 (n 1)!
Case 1: X is diagonalisable.
28
If X = ADA1 with D = diag (1 , 2 , ...) we have
exp X = exp ADA1 = A exp (D) A1 = A diag e1 , e2 , ... A1 .
Case 2: X is nilpotent.
A matrix X is called nilpotent, if X m = 0 for some positive m. In this case the series termi-
nates:
m1
Xn
exp X =
n=0 n!
Case 3: X is arbitrary.
A general matrix X may be neither diagonalisable nor nilpotent. However, any matrix X can
uniquely be written as
X = S + N,
where S is diagonalisable and N is nilpotent and SN = NS. Then
29
The T a s (and the X a s) are called the generators of the Lie group G.
Theorem: The commutators of the generators T a of a Lie group are linear combinations of
the generators and satisfy a Lie algebra.
h i n
T a , T b = i f abc T c .
c=1
In order to proove this theorem we have to show that the commutator is again a linear combina-
tion of the generators. We start with the definition of a one-parameter subgroup of GL(n, C): A
map g : R GL(n, C) is called a one-parameter sub-group of GL(n, C) if
1. g(t) is continous.
2. g(0) = 1.
If g(t) is a one-parameter sub-group of GL(n, C) then there exists a unique n n matrix X such
that
X is given by
d
X = g(t) .
dt t=0
X = ia T a
30
Therefore AYA1 is a linear combination of the generators. Now we take for A = exp (X ). This
implies that
exp (X )Y exp (X )
is a linear combination of the generators. Since the vector space spanned by the generators is
topologically closed, also the derivative with respect to belongs to this vector space and we
have shown that
d
exp (X )Y exp (X ) = XY Y X = [X ,Y ]
d =0
We have seen that by studying a Lie group G in the neighbourhood of the identity we can obtain
from the Lie group G the corresponding Lie algebra g. We can now ask if the converse is also
true: Given the Lie algebra g, can we reconstruct the Lie group G ? The answer is that this can
almost be done. Note that a Lie group need not be connected. The Lorentz group is an example
of a Lie group which is not connected. Given a Lie algebra we have information about the con-
nected component in which the idenity lies. The exponential map takes us from the Lie algebra
into the group. In the neighbourhood of the identity we have
!
n
g(1 , ..., n) = exp i a T a .
a=1
3.4.4 Examples
As an example for the generators of a group let us study the cases of SU (2) and SU (3), as well
as the groups U (2) and U (3). A common normalisation for the generators is
1 ab
Tr T a T b = .
2
a) The group SU (2) is a three-paramter group. The generators are proportional to the Pauli
matrices:
1 1 0 1 2 1 0 i 3 1 1 0
T = , T = , T = .
2 1 0 2 i 0 2 0 1
31
b) The group SU (3) has eight parameters. The generators can be taken as the Gell-Mann matri-
ces:
0 1 0 0 i 0 1 0 0
1 1 1
T 1 = 1 0 0 , T 2 = i 0 0 , T 3 = 0 1 0 ,
2 2 2
0 0 0 0 0 0 0 0 0
0 0 1 0 0 i 0 0 0
1 1 1
T4 = 0 0 0 , T5 = 0 0 0 , T6 = 0 0 1 ,
2 2 2
1 0 0 i 0 0 0 1 0
0 0 0 1 0 0
1 1
T 7 = 0 0 i , T 8 = 0 1 0 .
2 2 3
0 i 0 0 0 2
for U (3).
Tr T a T b T a T b ,
We first consider the case for SU (N). The Fierz identity reads for SU (N):
1 1
a a
Ti j Tkl = il jk i j kl .
2 N
Proof: T a and the unit matrix form a basis of the N N hermitian matrices, therefore any hermi-
tian matrix A can be written as
A = c0 1 + ca T a .
32
The constants c0 and ca are determined using the normalization condition and the fact that the
T a are traceless. We first take the trace on both sides:
Tr (A) = c0 Tr 1 + ca Tr T a = c0 N,
therefore
1
c0 = Tr (A) .
N
Now we multiply first both sides with T b and take then the trace:
1
Tr AT b = c0 Tr T b + ca Tr T a T b = ca ab ,
2
therefore
ca = 2Tr (T a A) .
In the case of a U (N)-group the identity matrix is part of the generators and the Fierz identity
takes the simpler form
1
Tiaj Tkla = il jk .
2
33
As a consequence we have for a U (N)-group for the traces
1
Tr(T a X ) Tr(T aY ) = Tr (XY ) ,
2
1
Tr(T a X T aY ) = Tr (X ) Tr(Y ) .
2
It is also useful to know, that the structure constants f abc can expressed in terms of traces over
the generators: From
[T a , T b ] = i f abc T c
This yields an expression of the structure constants in terms of the matrices of the fundamental
representation. We can now calculate for the group SU (N) the fundamental and the adjoint
Casimirs:
N2 1
(T a T a )i j = CF i j = i j ,
2N
f abc f dbc = CA ad = Nad .
For the group SU (N) we define the symmetric tensor d abc through
n o 1
T a, T b = d abc T c + ab .
N
Here, {..., ...} denotes the anti-commutator
{A, B} = AB + BA.
From this expression and the fact that the trace is cyclic we see explicitly that d abc is symmetric
in all indices.
F = A (x) A (x)
34
the field strength tensor. We denote further by U (x) an element of a U (1)-Lie group, smoothly
varying with x. We may write
where (x) is a real smooth function. Recall that the group U (1) has one generator, which is
the 1 1 unit matrix. This is simply one and we dont write it explicitly. We then consider the
transformation
i
A (x) = U (x) A (x) + U (x) .
e
Here, e denotes the electric charge. Working out the expression for A (x) in terms of (x) we
find
1
A (x) = A (x) (x).
e
This is nothing else than a gauge transformation A (x) = A (x) (x) with (x) = (x)/e.
The field strength transforms as
F = U FU .
Therefore for a U (1)-transformation the field strength is invariant. As a consequence, also the
Lagrange density
1
L = F F
4
is invariant:
L = L.
Now let us see if this can be generalised to SU (N). In electrodynamics we can interpret the gauge
potential as a quantity, which takes values in the Lie algebra of U (1). There is only one generator
for the U (1)-group. In the group SU (N) there are more generators (N 2 1 to be precise), and
we start from the following generalisation of the gauge potential:
T a Aa (x),
where the T a s are the generators of the Lie algebra of SU (N) and a ranges from 1 to N 2 1. We
consider again a transformation of the form
i
T A (x) = U (x) T A (x) + U (x) ,
a a a a
g
35
where U (x) is now an element of SU (N). (g is a coupling constant replacing the electromagnetic
coupling e). U (x) can be written in terms of the generators as
The group SU (N) is non-Abelian and the generators T a do not commute. As a consequence the
expression
(T a Aa ) T a Aa
A = T a Aa .
We have
i i
A
= U A + U = UAU + U U
g g
i i
= U A U + U AU +UA U + U U + U U
g g
Since UU = 1 we have
0 = UU = U U +U U
and hence
U U = U U .
36
We now define the Lagrange density as
1
L = Tr F F .
2
This Lagrange density is invariant under SU (N)-gauge transformations:
1 1 1
L = Tr F
F = Tr U FU U F U = Tr F F = L .
2 2 2
Going back from our short-notation to the more detailed notation
A = T a Aa ,
F = T a F
a
.
a the Lagrange density reads
In terms of F
1
a a
L = F F .
4
37
4 Representation theory
4.1 Group actions
An action of a group G on a set X is a correspondence that associates to each element g G a
map g : X X in such a way that
g1 g2 = g1 g2 ,
e is the identity map on X,
where e denotes the neutral element of the group. Instead of g (x) one often writes gx.
G is said to act effectively on X , if the homomorphism from G into the group of transforma-
tions of X is injective.
G is said to act transitively on X , if there is only one orbit. A set X where a group G acts
transitively is called a homogeneous space. Every orbit of a (not necessarily transitive) group
action is a homogeneous space.
The stabilizer (or the isotropy subgroup or the little group) Hx of a point x X is the subgroup
of G that leave x fixed, e.g. h Hx if hx = x. When Hx is the trivial subgroup for all x X , we
say that the action of G on X is free.
4.2 Representations
Let V be a finite-dimensional vector space and GL(V ) the group of automorphisms of V . Typi-
cally V = Rn or V = Cn and GL(V ) = GL(n, R) or GL(V ) = GL(n, C).
g (g).
38
This implies
(e) = 1,
g1 = [(g)]1 .
(g) = 1, g.
Remark: In general more than one group element can be mapped on the identity. If the mapping
: G GL(V ) is one-to-one, i.e.
(g) = (g ) g = g
(g)w W g G and w W.
If V1 and V2 are representations of G, the direct sum V1 V2 and the tensor product V1 V2
are again representations:
Two representations 1 and 2 of the same dimension are called equivalent, if there exists a
non-singular matrix S such that
1 (g) = S2(g)S1 , g G.
For finite groups and compact Lie groups it can be shown that any representation is equivalent to
a unitary representation.
For finite groups the proof goes as follows: Suppose we start with an arbitrary (i.e. not neces-
sarily unitary) representation 2 (g). We would like to find a S such that 1 (g) = S2 (g)S1 is a
unitary matrix for all g. We set S to be a hermitian matrix which satisfies
S 2 = 2 g 2 g .
g G
39
Then
2 (g) S2 2 (g) = 2 (g) 2 g 2 g 2 (g) = 2 g g 2 g g
g G
g G
= 2 g 2 g
= S2 .
g G
We therefore have
2 (g) S2 2 (g) = S2 ,
2 (g) S2 = S2 2 (g)1 ,
S1 2 (g) S = S2 (g)1 S1 ,
1
S2 (g) S1 = S2 (g) S1 ,
1 (g) = 1 (g)1,
This proof carries over to the case of compact Lie groups by replacing the sum in the defini-
tion of S by an integration over all group elements.
The goal of representation theory: Classify and study all representations of a group G up to
equivalence. This will be done by decomposing an arbitrary representation into direct sums of
irreducible representations.
M = c1.
Proof: We have
(g)M = M(g) g G.
If (g) is of dimension n, then M must be square of dimension n. Let us assume that (g) is
unitary. Then
M (g) = (g) M .
(g)M = M (g).
40
Therefore also M commutes with all (g), and so do the hermitian matrices
H1 = M + M ,
H2 = i M M .
D = U 1 HU.
If we define now
(g) = U 1 (g)U,
we have
(g)D = D (g).
Let D = diag(1 , ..., n) and consider now the i, j element of this matrix equation:
(g) i j j = i (g) i j ,
i j (g) i j = 0.
Suppose that a certain eigenvalue of D occurs k times and that, by a suitable ordering the first
k positions of D are occupied by . Then
1 = ... = k 6= l , k + 1 l n.
and is thus reducible, contrary to the initial assumption. Thus if and only if all the eigenvalues
of D are the same (g) will be irreducible. In other words, D and hence M must be a multiple
of the unit matrix.
Lemma 2: If 1 (g) and 2 (g) are two irreducible representations of a group G of dimensions
n1 and n2 respectively and if a rectangular matrix M of dimension n1 n2 exists sucht that
1 (g)M = M2 (g), g G
41
then either
(a) M = 0 or
(b) n1 = n2 and detM 6= 0, in which case 1 (g) and 2 (g) are equivalent.
Proof: Let us assume without loss of generality that 1 (g) and 2 (g) are unitary representations.
M 1 (g) = 2 (g)M ,
M 1 (g1 ) = 2 (g1 )M .
M 1 (g1 )M = 2 (g1 )M M.
M M2 (g1 ) = 2 (g1 )M M.
By lemma 1 we conclude
M M = 1.
If 6= 0 then det M 6= 0 and therefore M 1 exists. From 1 (g)M = M2 (g) it follows that
1 (g) = M2 (g)M 1
Mik Mki = 0,
k
|Mki|2 = 0.
k
M = 0.
To complete the proof we consider the case n1 6= n2 . Let us assume n1 < n2 . Construct M from
M by adding n2 n1 rows of zeros:
M
M =
0
42
M = M 0
We have
M M = MM
and thus
det M M = det M M = det M det M = 0.
Application: Orthogonality theorem for finite groups. Let G be a finite group and let 1 and
2 be irreducible representations of dimension n1 and n2 . Then
1 1
=
|G| gG
1 (g)X 2(g1 g ) =
|G| gG
1 (g)X 2(g1 )2 (g ) = M2 (g ).
1(g)i j 2(g1)kl = 0.
gG
Now consider the case where 1 and 2 are identical: 1 = 2 = . Take again
1
M = (g)X (g1).
|G| gG
43
One shows again
(g)M = M(g).
Therefore by Schurs first lemma
1
|G| gG
(g)i j X j k (g1 )k l = cil .
Again take X = j j kk :
1
|G| gG
(g)i j (g1)kl = cil .
Another consequence of Schurs first lemma: All irreducible representation of an Abelian group
are one-dimensional.
Example: Consider the symmetric group S3 , the permutation group of three elements, which
is the simplest non-abelian group. This group has two one-dimensional representations: The
trivial one I and the alternating representation A defined by
gv = sign(g)v.
There is a natural representation, in which S3 acts on C3 by
g (z1, z2 , z3 ) = zg1 (1) , zg1 (2) , zg1 (3)
This representation is reducible: The line spanned by the sum
e1 + e2 + e3
is an invariant sub-space. The complementary sub-space
V = {(z1 , z2 , z3 )|z1 + z2 + z3 = 0}
defines an irreducible representation. This representation is called the standard representation.
It can be shown that any representation of S3 can be decomposed into these three irreducible
representations
W = I n1 An2 V n3 .
44
4.4.1 Characters
Definition:If V is a representation of G, its character V is the complex-valued function on the
group defined by
V (g) = Tr ((g)) .
In particular we have
V hgh1 = V (g).
V W = V + W ,
V W = V W ,
V = (V ) ,
1
2V (g) = V (g)2 V (g2 ) ,
2
1
Sym2V (g) = V (g)2 + V (g2 )
2
Orthogonality theorem for characters: For finite groups we had the orthogonality theorem. If
we consider unitary representations and if we make the agreement that if two representations are
equivalent, we take them to be identical, the orthogonality theorem can be written as
|G|
(g)i j (g)lk =
n1
il k j
gG
(Note that for an unitary representation we have (g1 )kl = (g)lk .) Now we set i = j and sum,
and we set l = k and sum:
(g)(g) = |G|
gG
= n ,
gG classes
where n denotes the number of elements in the class C . Therefore
n(C)(C) = |G|
Character table:
45
n1C1 n2C2 n2C2 ...
1 1 (C1 ) 1 (C2 ) 1 (C3 ) ...
2 2 (C1 ) 2 (C2 ) 2 (C3 ) ...
3 3 (C1 ) 3 (C2 ) 3 (C3 ) ...
... ... ... ... ...
The number of orthogonal vectors corresponds to the number of inequivalent representations.
The dimension of the space is given by the number of classes. Therefore the number of inequiv-
alent representations is smaller or equal to the number of classes. In fact equality holds. To show
this, we consider one specific (reducible) representation, called the regular representation. The
regular representation is defined by
ga gb = Rcb (ga) gc
c
Note that the matrix Rcb (ga ) has in each column exactly one 1 and all other entries in this column
are zero. This defines a representation: We have
Therefore it follows
or in matrix notation
We have already seen that the 1 appears in each column of the matrix R (g) exactly once. The 1
appears always on the diagonal if g = e, otherwise it appears never on the diagonal. This implies
for the character
0 g 6= e,
(g) =
R
|G| g = e.
R (g) = a (g) ,
M
46
where the sum is over all irreducible representations and a gives the number of times the irre-
ducible representation is contained in R . For the characters we have then
R (g) = a (g) .
Using the fact that R (g) = 0 for all g 6= e we have on the other side
1 1
|G|
n R (g) =
|G|
(e) |G| = (e) = n ,
a = n .
|G| = a n = n2.
From the orthogonality theorem for finite groups it follows that we can view (g)i j for fixed
and fixed i, j as a vector in a |G|-dimensional space. The orthogonality theorem tells us, that
there are
n2
orthogonal vectors. On the other hand we have just shown that this number equals |G|, therefore
the vectors (g)i j span the full space. Therefore any vector in this space can be written as a
linear combination of these basis vectors. In particular the component va with respect to the
standard bases is given by
va = c(, i, j) (ga)i j .
,i, j
47
Let us now focus on vectors which are constant on classes. For those vectors we have with
gc = g1
b ga gb
1 |G|
va = vc
|G| b=1
1 |G|
= c(, i, j) gb gagb i j
|G| b=1
1
,i, j
1 |G|
=
|G| b=1
c(, i, j) b ik (ga)kl (gb)l j
g 1
,i, j k,l
1
= c(, i, j) n i j lk (ga )kl
,i, j k,l
c(, i, i)
= n
(ga )
,i
These vectors span a subspace of dimension equal to the number of classes nclass . The above
equations show that the characters of the irreducible representations span this subspace. Hence
there must be exactly nclass such characters, which is the desired result.
(g) = a (g)
M
Then
(g) = a(g).
Conisder now
1 = 1 irreducible
|G| |a|
(g)(g) = 2
g > 1 reducible
n(C)(C) = |G|
If we define
r
n
= (C )
|G|
48
we have
= .
We can view as an entry of a nclass nclass -matrix. In matrix notation we can write
= 1.
= 1.
In other words
= .
|G|
(C) (C ) = .
n
This defines an orthogonality relation between the columns of the character table.
Let us assume that the molecules has n atoms. We will denote the positions of the atoms by
(~x1 , ...,~xn)
(0) (0)
and the equilibrium position by (~x1 , ...,~xn ). It is convenient to introduce 3n coordinates i ,
i = 1, ..., 3n, describing the displacement from the equilibrium position. The Lagrange function
of the system reads
L = T V,
1 3n
T = Ti j i j
2 i=1
1 3n
V = Vi j i j .
2 i=1
49
From classical mechanics we know that by suitable change of coordinates we can achive that the
Lagrange function reads
1 3n 2 1 3n 2 2
L = i 2 i i .
2 i=1 i
The i s are called the frequencies of the normal modes. The change of coordinates involves the
diagonalisation of (3n) (3n) matrices.
We will now discuss how the task of obtaining the normal modes can be simplified using group
theory. As an example we will discuss the water molecule H2 O. Let us agree that the equilibrium
position of the water molecule is in the x z plane, with the O-atom along the z-axis and the two
H-atoms along the x-axis. We assume that the two H-atoms cannot be distinguished. We first
determine the symmetry group, which leaves the equilibrium position invariant. We can rotate
the system by 180 along the z-axis, since the two H-atoms are not distinguished, we obtain the
same configuration again. This defines a Z2 -group. We denote the element, which generates
the group by a. Secondly, we have a reflection in the x z plane: Changing y y will not
affect the equilibrium position. This defines another Z2 -group. We denote the element, which
generates the group by b. In summary we find the symmetry group Z2 Z2 , with the generators
a and b. This group has four elements {e, a, b, ab} and is Abelian. We therefore have four classes
(each element is in a class of its own) and as a consequence four irreducible representations 1 ,
..., 4 . The character table is easily obtained:
e a b ab
1 1 1 1 1
2 1 1 1 1
3 1 1 1 1
4 1 1 1 1
The (3n) coordinates i define a (3n)-dimensional representation of this group. This represen-
tation is in general reducible. We will now discuss how often a given irreducible representation
occurs in the (3n)-dimensional one, i.e. we look for the decomposition
= a .
M
50
or writing this alternatively
n 3
i, j = (g)i,k; j,l k,l .
k=1 l=1
We can think of (g)i,k; j,l as a n n-matrix, whose entries are 3 3-matrices. For the character
of this representation we have
n 3
(g) = (g)i,i; j, j.
i=1 j=1
In particular we observe that for the trace only the 3 3-matrices on the diagonal of the n n-
matrix contribute. In other words: Only the displacements of the atoms which are left unmoved
by the symmetry operation are relevant. Let us now consider an atom i which is not moved by
a symmetry operation. The effect of a rotation by an angle through the z-axis on the three
displacements i,1 , i,2 , i,3 is given by
cos sin 0
sin cos 0
0 0 1
The trace of this matrix is
(R ()) = 2 cos + 1.
For = 180 we find
= 1.
The reflection in the x z plane is described by
1 0 0
0 1 0 ,
0 0 1
its character is
= 1.
Finally, the combined operation of a rotation by = 180 and a reflection is given by
1 0 0
0 1 0 ,
0 0 1
with character
= 1.
We can now obtain the characters of the representation :
51
e a b ab
9 1 3 1
Not all irreducible representations in the decomposition of correspond to true vibrations: The
9 generalised coordinates i contain 6 coordinates desribing the centre-of-mass motion and the
rigid rotation of the molecule. The centre-of-mass motion is described by the vector
1 n
~X = mi~xi,
M i=1
n
with M = mi . This is a three-dimensional representation of the symmetry group with character
i=1
e a b ab
trans 3 1 1 1
The rigid rotation can be described by the three quantities
(0) (0)
n n
xi,1 i,2 xi,2 i,1
1 1
(0) (0) (0)
mi~xi ~i = mi xi,2 i,3 xi,3 i,2 .
M i=1 M i=1 (0) (0)
xi,3 i,1 xi,1 i,3
This transforms as a three-dimensional vector under rotations, but as a pseudo-vector under re-
flections. Indeed, for the symmetry transformation y y we have the transformation matrix
1 0 0
0 1 0
0 0 1
= 1.
52
We can now find the decomposition of trans rot in terms of the irreducible representations.
We make the ansatz
trans rot = a .
M
For the water molecule |G| = 4 and all classes have exactly one element: n = 1. We find
1
a1 = (1 3 + 1 1 + 1 3 + 1 1) = 2,
4
1
a2 = (1 3 + (1) 1 + 1 3 + (1) 1) = 1,
4
1
a3 = (1 3 + 1 1 + (1) 3 + (1) 1) = 0,
4
1
a4 = (1 3 + (1) 1 + (1) 3 + 1 1) = 0.
4
Therefore
trans rot = 21 + 2 .
We therefore find three vibrational modes. Two transform as the trivial representation 1 . The
displacements in these modes are left invariant under the symmetry group Z2 Z2 . One of these
two modes is given by the vibration, where both H-atoms move along the z-axis in the same
direction, while the O-atom moves along the z-axis in the opposite direction. The other mode is
given by an oscillation, where the O-atom is at rest, and the H-atoms move along the x-axis in
opposite directions.
The third mode transforms as the representation 2 . It transforms trivially under reflections,
therefore the motion is in the x z plane. It has however a non-trivial transformation under the
rotation of 180 around the z-axis. This is oscillation is given by a motion, where the O-atom
moves along the x-axis in one direction, the two H-atoms have both a component along the x-
axis in the opposite direction. In addition the two H-atoms have opposite components along the
z-axis.
i~ (~x,t) = H (~x,t) .
t
53
If the Hamilton operator H is time-independent we can make the ansatz
i
(~x,t) = (~x) exp Et
~
H (~x) = E (~x) .
~x = g~x, g G.
An operator transforms as
O = (g)O(g1 ).
We are in particular interested in transformations, which leave the Hamilton operator invariant:
H = (g)H(g1).
Multiplying this equation by (g) from the right we obtain H(g) = (g)H, and we see that this
is equivalent to the statement that (g) commutes with the Hamilton operator
H, (g) = 0.
Remark: Usually the Hamiltonian of a quantum mechanical system is given by the sum of the
kinetic and potential energy operator. As symmetry transformations we will usually consider
translations, rotations and reflections. The kinetic energy operator is invariant under these trans-
formations, therefore the full Hamiltonian is invariant if the potential energy operator is:
V (~x) = V (g~x) .
Let us now consider a quantum mechanical system with a Hamilton operator H, which is invari-
ant under a finite symmetry group G. In this case :
The eigenfunctions for a given eigenvalue E form a representation of the symmetry group
G.
54
Proof: The set of all degenerate eigenfunctions for the eigenvalue E form a vectorspace V . If
and are two eigenfunctions with the eigenvalue E, so is any linear combination of them. This
vector space defines a representation of the symmetry group G. If V and = (g) then
dim V dim V = n .
Example: We consider the quantum mechanical harmonic oscillator in one dimension. The
Hamilton operator is
p2 m2 2
H = + x .
2m 2
We define the characteristic length
r
~
x0 = .
m
Obviously, the Hamilton operator is invariant under the reflection x x. The reflection a
generates a symmetry group Z2 . The character table of this group is
e a
1 1 1
2 1 1
It is well-known that the eigenvalues are given by
1
En = ~ n +
2
and the eigenfunctions are given by
2 !
1 1 x x
n (x) = p exp Hn ,
2n n! x0 2 x0 x0
55
4.5 Representation theory for Lie groups
4.5.1 Irreducible representation of SU (2) and SO(3)
The groups SU (2) and SO(3) have the same Lie algebra:
[Ia , Ib ] = iabc Ic .
For SU (2) we can take the I a s proportional to the Pauli matrices
1 0 1 1 0 i 1 1 0
I1 = , I2 = , I3 = .
2 1 0 2 i 0 2 0 1
This defines a representation of SU (2) which is called the fundamental representation. (It is not
a representation of SO(3).)
Quite generally the structure constants provide a representation known as the adjoint or vector
representation:
(Mb )ac = i fabc .
For SU (2) and SO(3):
0 0 0 0 0 i 0 i 0
M1 = 0 0 i , M2 = 0 0 0 , M3 = i 0 0 .
0 i 0 i 0 0 0 0 0
The dimension of the adjoint representation equals the dimension of the parameter space of the
group and the numbers of generators.
56
Here, {..., ...} denotes the anti-commutator
{A, B} = AB + BA.
We then define C3 by
C3 = d abc T a T b T c .
C3 is called the cubic Casimir operator. Since C2 and C3 are Casimir operators, we have
[C2 , T a ] = 0, [C3 , T a ] = 0.
Continuing in this line, it can be shown that SU (n) has (n 1) independent Casimir operators
C2 , C3 , ..., Cn1 . The group U (n) has n independent Casimir operators C1 , C2 , ..., Cn1 .
Definition: The rank of a Lie algebra is the number of simultaneously diagonalisable gen-
erators.
Example 1: SU (2) has rank one, the convention is to take I3 diagonal.
Example 2: SU (3) has rank two, in the Gell-Mann representation T3 and T8 are diagonal.
Theorem: The number of independent Casimir operators is equal to the rank of the Lie alge-
bra. The proof can be found in many textbooks.
The eigenvalues of the Casimir operators may be used to label the irreducible representations.
The eigenvalues of the diagonal generators can be used to label the basis vectors within a given
irreducible representation.
Example SU (2):
I 2 |, mi = |, mi ,
I3 |, mi = m |, mi .
Consider
I12 + I22 |, mi = I 2 I32 |, mi = m2 |, mi .
Further
m2 0.
57
Define
1
I = (I1 iI2 )
2
[I3 , I ] = I , I 2, I = 0.
Therefore the operators I dont change . From the commutation relation with I3 we obtain
(I3 I I I3 ) |, mi = I |, mi ,
I3 (I |, mi) = (m 1) (I |, mi) .
I+ |, mmax i = 0,
I |, mmini = 0.
Now
Therefore
I 2 |, mmax i = I I+ + I32 + I3 |, mmax i ,
|, mmax i = mmax (mmax + 1) |, mmax i ,
and
= mmax (mmax + 1) .
Similar:
I 2 |, mmin i = I+ I + I32 I3 |, mmin i ,
|, mmin i = mmin (mmin 1) |, mmini ,
and
= mmin (mmin 1) .
58
From
it follows
mmin = mmax .
Since the ladder operators raise or lower m by one unit we must have that mmax and mmin differ
by an integer, therefore
2mmax = integer.
= j ( j + 1)
Normalisation:
I |, mi = A |, m 1i
With I = I we have
|A |2 = h, m| I I |, mi = h, m| I I |, mi = h, m| I 2 I3 (I3 1) |, mi
and therefore
|A |2 = j ( j + 1) m (m 1)
Condon-Shortley convention:
p
A = j ( j + 1) m (m 1).
59
4.5.2 The Cartan basis
Definition: Suppose a Lie algebra A has a sub-algebra B such that the commutator of any element
of A (T a say) with any element of B (T b say) always lies in B, then B is said to be an ideal of A:
h i
a b
T ,T B.
A Lie algebra is called reductive if it is the sum of a semi-simple and an abelian Lie alge-
bra.
A simple Lie algebra is also semi-simple and a semi-simple Lie algebra is also reductive.
are simple.
Semi-simple Lie algebras are sums of simple Lie algebras:
su(n1) su(n2 ).
From Schurs lemma we know that abelian Lie groups have only one-dimensional irreducible
representations. Therefore let us focus on Lie groups corresponding to semi-simple Lie algebras.
A Lie group, which has a semi-simple Lie algebra, is for obvious reasons called semi-simple.
We first would like to have a criterion to decide, whether a Lie algebra is semi-simple or not: If
h i
a b
T ,T = i f abc T c ,
define
A criterion due to Cartan say that a Lie algebra is semi-simple if and only if
det g 6= 0.
60
For SU (n) we find
gab = CA ab .
Let us now define the Cartan standard form of a Lie algebra. To motivate the Cartan standard
form let us as an example suppose that we have
1 2
T , T = 0, T 1 , T 3 6= 0, T 2 , T 3 =6 0.
none of the new commutators vanishes. It is certainly desirable to pick a basis, where a maximum
number of commutators vanish and the non-vanishing ones are rather simple. This will bring us
to the Cartan standard form. Let us assume that
n
A = ca T a ,
a=1
n
X = xa T a ,
a=1
such that
[A, X ] = X .
In general the secular equation will give a n-th order polynomial is . Solving for one obtains
n roots. One root may occur more than once. The degree of degeneracy is called the multiplicity
of the root.
Theorem (Cartan): If A is chosen sucht that the secular equation has the maximum number
of distince roots, then only the root = 0 is degenerate. Further if r is the multiplicity of that
root, there exist r linearly independent generators Hi , which mutually commute
Hi , H j = 0, i, j = 1, ..., r.
61
r is the rank of the Lie algebra.
Notation: Latin indices for 1, ..., r, e.g. Hi and greek indices for the remaining (n r) gener-
ators E ( = 1, ..., n r).
Example SU (2):
h i
I a , I b = iabc I c .
Take A = I 3:
3
I , X = X .
Secular equation:
det i3bc bc = 0,
i 0
i 0 = 0,
0 0
3 + = 0,
2 1 = 0.
Therefore the roots are 0, 1. We have
3
=0 I , X = 0 X = I 3 = H1 ,
3 1
=1 I , X = X X = I 1 + iI 2 = E1 ,
2
3 1
= 1 I , X = X X = I 1 iI 2 = E2 .
2
Theorem: For any compact semi-simple Lie group, non-zero roots occur in pairs of opposite sign
and are denoted E and E ( = 1, ..., (n r)/2).
62
Cartan standard form of SU (2):
1 1 0 1 0 1 1 0 0
H1 = , E1 = , E1 = .
2 0 1 2 0 0 2 1 0
The roots are
Theorem: If ~ is a root vector, so is ~, (since roots always occur in pairs of opposite sign).
2~ ~ 2~ ~
and
||2 ||2
63
are integers. Suppose these integers are p and q. Then
2
~ ~ pq
2 2
= = cos2 1.
|| || 4
Therefore
pq 4.
It follows that
1 1 3
cos2 = 0, , , , 1.
4 2 4
Case = 0 : This is the trivial case ~ = ~.
Case = 30 : We have pq = 3 and p = 1, q = 3 or p = 3, q = 1. Let us first discuss p = 1, q = 3.
This means
2~ ~ 2~ ~
= 1, = 3.
||2 ||2
Therefore
||2
= 3.
||2
The case p = 3, q = 1 is similar and in summary we obtain
||2 1
2
= 3 or .
|| 3
Case = 45 : We have pq = 2 and p = 1, q = 2 or p = 2, q = 1. It follows
||2 1
2
= 2 or .
|| 2
Case = 60 : We have pq = 1 and p = 1, q = 1. It follows
||2
= 1.
||2
Case = 90 : In this case p = 0 and q = 0. This leaves the ratio ||2 /||2 undetermined.
The cases = 120 , = 135 , = 150 and = 180 are analogous to the ones discussed above.
64
60
4.5.3 Weights
Let us first recall some basic facts: The rank of a Lie algebra is the number of simultaneously
diagonalizable generators. In the following we will denote the rank of a Lie algebra by r.
We already mentioned the following theorem : The rank of a Lie algebra is equal to the number
of independent Casimir operators. (A Casimir operator is an operator, which commutes with all
the generators.)
For a Lie algebra of rank r we therefore have r Casimir operators and r simultaneously diag-
onalizable generators Hi .
The eigenvalues of the Casimir operators may be used to label the irreducible representations.
The eigenvalues of the diagonal generators Hi may be used to label the states within a given
irreducible representation.
Let ~ be a shorthand notation for ~ = (1 , ..., r ), a set of eigenvalues of Casimir operators and
let ~m be a shorthand notation for ~m = (m1 , ..., mr ), a set of eigenvalues of the diagonal generators:
Hi |~,~mi = mi |~,~mi
The vector ~m is called the weight vector.
Example SU (3): Let us consider the fundamental representation. The vector space is spanned
by the three vectors
1 0 0
e1 = 0 , e2 = 1 , e3 = 0 .
0 0 1
We have
1 1
(H1 , H2 ) e1 = , e1 ,
6 3 2
1 1
(H1 , H2 ) e2 = , e2 ,
6 3 2
2
(H1 , H2 ) e3 = 0, e3 .
3 2
65
This gives the weight vectors
! !
1 16
6
0
~m1 = 1 , ~m2 = 1 , ~m3 = .
3 2
3 2
32
m1
~m3
Consider now the complex conjugate representation of the fundamental representation: If
= exp (ia T a )
T a = T a .
It follows that the weights of the complex conjugate representation are negatives of those of the
fundamental representation: m2
m1
SI a S1 = I a , SS1 = .
The generators E are generalisations of the raising and lowering operators I of SU (2). Sup-
pose
Hi |~,~mi = mi |~,~mi
66
and
[Hi , E ] = i E .
Then
Therefore E |~,~mi is proportional to |~,~m +~i unless zero. Therefore the weight vectors within
an irreducible representation differ by a linear combination of root vectors with integer coeffi-
cients.
Example: In the SU (2) case the weight vectors were one-dimensional. Within one irreducible
representation all weights could be obtained from mmax by applying the lowering operator I .
The action of I corresponds to a shift in the weight proportional to a root vector.
In the SU (2) case we also found that within an irreducible representations the weights are
bounded, i.e. there is a maximal weight, for which I+ |, mmax i = 0. We now look how this
fact generalises: We start with the definition of the multiplicity of a a weight: The number of
different eigenstates with the same weight is called the multiplicity of the weight. A weight is
said to be simple if the multiplicity is 1.
For Lie algebras with r 2, weights are not necessarily simple.
Ordering of weights: The convention for SU (n) is the following: ~m is said to be higher than ~m
if the rth component of (~m ~m ) is positive (if zero look at the (r 1)th component, if this one
is also zero, look at the (r 2)th component, etc. ).
67
Theorem: For any compact semi-simple Lie algebra there exists for any irreducible representa-
tion a highest weight. Furthermore this highest weight is also simple.
(Recall: In the SU (2) case we first showed that the values of m are bounded, and then obtained
all other states in the irreducible representation by applying the lowering operator to the state
with mmax .)
Theorem: For every simple Lie algebra of rank r there are r dominant weights M ~ (i) , called
~ is a linear combination
fundamental dominant weights, such that any other dominant weight M
~
of the M (i)
r
~ =
M ni M
~ (i)
i=1
Note that there exists r fundamental irreducible representations, which have the r different M ~ (i) s
as their highest weight. We can label the irreducible representations by (n1 , n2 , ..., nr ) instead of
the eigenvalues of the Casimirs.
Example: For SU (2) we can label the irreducible representations either by the eigenvalue
of the Casimir operator I 2 , or by a number n with the relation
n
= j ( j + 1) , j = .
2
Note that there is one fundamental representation, which is two-dimensional. The dominant
weight of the fundamental representation is 1/2. All other dominant weights j are non-negative
integer multiples of this fundamental dominant weight:
1
j = n
2
We now turn to the question how to construct new irreducible representations out of given irre-
ducible ones. If V1 and V2 are irreducible representations, the direct sum is V1 V2 is reducible
and decomposes into the irreducible representations V1 and V2 . Nothing new here. More inter-
esting is the tensor product, which we will study in the following.
68
4.6.1 Clebsch-Gordan series
To motivate the discussion of tensor methods we start again from the SU (2) example and its
relation to the spin of a physical system. Suppose we have to independent spin operators J~1 and
J~2 , describing the spin of particle 1 and 2, respectively.
J1 i , J2 j = 0 i, j
Let us now define the total spin as
J~ = J~1 + J~2 ,
Jz = J1 z + J2 z .
We use the following notation:
| j 1 , m1 i eigenstate of J12 and J1 z
| j 2 , m2 i eigenstate of J22 and J2 z
We define
| j 1 , j 2 , m1 , m2 i = | j 1 , m1 i | j 2 , m2 i .
The set
{| j1 , j2, m1 , m2 i}
are eigenvectors of
2 2
J1 , J2 , J1 z , J2 z
and is referred to as the uncoupled basis. In general these states are not eigenstates of J 2 and the
basis is reducible. This can be seen easily:
2 ~ ~ ~ ~
J = J1 + J2 J1 + J2 = J12 + J22 + 2J~1 J~2 ,
and 2J~1 J~2 fails to commute with J1 z and J2 z . To find a better basis, we look for a set of mutually
commuting operators. The set
2
J , Jz, J12 , J22
is such a set and an eigenbasis for this set is labelled by
{| j, m, j1, j2 i} .
This basis is called the coupled basis and carries and irreducible representation of dimension
2 j + 1. Of course we can express each vector in the coupled basis through a linear combination
of the uncoupled basis:
jm
| j, m, j1, j2 i = C j 1 j 2 m1 m2 | j 1 , j 2 , m1 , m2 i
m1 ,m2 ;m1 +m2 =m
69
jm
The coefficients C j1 j2 m1 m2 are called the Clebsch-Gordan coefficients. The Clebsch-Gordan co-
efficients are tabulated in the particle data group tables.
70
Then
D E
M = O = U U OU U = O = M.
If
O = O ,
or equivalently
O ,U = 0
we say that the operator O transforms as a scalar (or as a singlet). This is the simplest case. We
discuss now more general cases. We first fix the group to be SU (2). Suppose we are given a set
of (2k + 1) operators Tqk , k q k, which transform irreducible under the group SU (2). That
is to say that
Tqk = U TqkU
can be expressed as
(k)
Tqk = Dqq Tqk ,
71
We can now state the Wigner-Eckart theorem:
D E 1 D E
k j m
C jkmq j T k j .
j m Tq jm =
2j +1
The important point is that the double bar matrix element h j ||T k || ji is independent of m, m and
j m
q. The dependence on m, m and q is entirely given by the Clebsch-Gordan coefficients C jkmq .
2 2 = 3 1,
into a direct sum of irreducible representations. We generalise this now to general irreducible
representations of SU (N).
Let us denote the number of boxes in row j by j . Then a Young diagram is a partition of m
defined by the numbers (1 , 2 , ..., n) subject to
1 + 2 + ... + n = m,
1 2 ... n .
(1 , 2 , 3 , 4 ) = (4, 2, 1, 1)
The number of rows is denoted by n. For SU (N) we consider only Young diagrams with n N.
p1 = 1 2 ,
p2 = 2 3 ,
...
pn1 = n1 n2 .
72
The example above has
(p1 , p2 , p3 ) = (2, 1, 0)
Correspondence between Young diagrams and irreducible representations: Recall from the last
lecture that we could label any irreducible representation of a simple Lie algebra of rank r by
either the r eigenvalues of the Casimir operators or by the r numbers (p1 , ..., pr ) appearing when
expressing the dominant weight of the representation in terms of the fundamental dominant
weights:
r
~ =
M pi M
~ (i)
i=1
The group SU (N) has rank N 1 and we associate to an irreducible representation of SU (N)
given through (p1 , ..., pN1) the Young diagram corresponding to (p1 , ..., pN1 ).
As only differences in the number boxes between succesive rows matter, we are allowed to add
any completed column of N boxes from the left. Therefore in SU (4) we have
The trivial (or singlet) representation is alway associated with a column of N boxes. For SU (3):
The complex conjugate representation of a given representation is associated with the conjugate
Young diagram. This diagram is obtained by taking the complement with respect to complete
columns of N boxes and rotate through 180 to obtain a legal Young diagram.
Examples for SU (3):
complement rotation
,
complement rotation
,
complement rotation
.
73
i) Place integers in the boxes, starting with N in the top left box, increase in steps of 1 across
rows, decrease in steps of 1 down columns:
N N +1 N +2
N 1 N
N 2 N 1
iii) The denominator is given by multiplying all hooks of a Young diagram. A hook is the
number of boxes that one passes through on entering the tableau along a row from the right hand
side and leaving down a column.
3 4 : 234
dim = = 8,
2 131
345
3 4 5 : dim = = 10,
123
3 4 5 : 234345
dim = = 10.
2 3 4 123234
Rules for tensor products: We now give rules for tensor products of irreducible representations
represented by Young diagrams. As an example we take in SU (3)
a a.
b
ii) Add the boxes with the as from the lettered diagram to the right-hand ends of the rows of the
unlettered diagram to form all possible legitimate Young diagrams that have no more than one a
per column.
a
a a,
a
74
Note that the diagram
a
a
is not allowed since it has one column with two as.
iii) Repeat the same with the bs, then with the cs, etc.
a a a b a
a a b, a a, , b
b
b a a
Note that the diagram
a
a
b
is not allowed for SU (3), since it has more than 3 rows.
iv) A sequence of letters a, b, c, ... is admissible if at any point in the sequence at least as
many as have occured as bs, at least as many bs have occured as cs, etc. Thus abcd and aabcb
are admissible sequences, while abb and acb are not. From the diagrams in step iii) throw away
all diagrams in which the sequence of letters formed by reading right to left in the first row, then
in the second row, etc., is not admissible. This leaves
a a a
a a, , b
b
b a
Removing complete columns of 3 boxes, we finally obtain
3 8 = 15 6 3.
As a further example let us calculate in SU (3) the tensor product of the fundamental representa-
tion with its complex conjugate representation:
3 3 = 1 8.
75
As a final example let us consider
3 3 3 = 1 8 8 10.
4.7 Applications
4.7.1 Selection rules in quantum mechanics
We consider a time-independent quantum mechanical system
H (~x) = E (~x) ,
and on operators by
O = (g)O(g1 ).
H = (g)H(g1),
we may label the states by the irreducible representations of the group G. Likewise, we focus
now on operators O, which transform as an irreducible representation of the group G. We denote
by () (~x) a set of states transforming like the irreducible representation :
() (~x) = (g)() (~x) ,
Further we denote by O() a set of operators transforming like like the irreducible representation
O() = (g)O(),
Recall that we defined for SU (2) a tensor operator as a set of (2k + 1) operators Tqk (k q k),
which transform irreducible as
(k)
Tqk = Dqq Tqk
76
under SU (2). The transformation law O() = (g)O() is nothing else than the straightforward
generalisation to an arbitrary group.
We are interested in the matrix elements
D E
() O() () .
Suppose the decomposition of the tensor product into irreducible representations reads
= n ,
where the number n indicates, how often the irreducible representation occurs in the decom-
position.
We then have the following theorem: If does not occur in the decomposition n , then
the matrix
D E
()
() ()
O
where SU (3)colour corresponds to the strong interactions, SU (2)weak isospin to the weak interac-
tion and U (1)Y to the hypercharge. The gauge symmetry of SU (2)weak isospin U (1)Y is sponta-
neously broken to a subgroup U (1)Q , where U (1)Q corresponds to the electroc charge. We will
not go here into the details of the mechanism of spontaneous symmetry breaking, but focus on an
unbroken gauge theory. The strong interactions with gauge group SUcolour provide an example.
All particles are classified according to representations of the gauge group. Fermions trans-
form as the fundamental representation of the gauge group. For SU (3)colour the quark fields form
a three-dimensional representation qi (x), i = 1, 2, 3. Let us denote a gauge transformation by
77
the dependence on the space-time coordinates x indicates that the gauge transformation may vary
from one space-time point to another. The hermitian matrices T a are the generators of the gauge
group in the fundamental representation. Under this gauge transformation, the quarks transform
as follows:
qi (x) = U (x)i j q j (x),
or in vector/matrix-notation without indices
q (x) = U (x)q(x).
The anti-fermions transform as the complex conjugate of the fundamental representation. In
physics we usually take the anti-quark fields qi (x), i = 1, 2, 3 as the components of a bra-vector
and write the transformation law as
q (x) = q(x)U (x).
Taking the transpose of this equation we get
T
q (x) = U (x)q(x)T .
We note that the combination q(x) q(x) is gauge-invariant:
q (x) q (x) = q(x)U (x) (U (x)q(x)) = q(x) q(x).
The gauge boson fields are given by the gauge potentials Aa (x), where the index a runs from 1
to the number of generators of the Lie group. For SU (N) this number is given by N 2 1. For
SU (3)colour we have eight gauge boson fields, which are called gluon fields. We have already
seen that the gauge potential transforms as
i
T A (x) = U (x) T A (x) + U (x) .
a a a a
g
This transformation law ensures that the expression
1
a a
L = F a
F , with F = Aa Aa + g f abc Ab Ac
4
is invariant. For space-time independent transformations U (x) = U the transformation law re-
duces to
T a Aa (x) = U T a Aa (x)U .
For infinitessimal transformations we have
T a Aa (x) = 1 ib T b + ... T a Aa (x) (1 + ic T c + ...)
h i
= T A (x) + i T , T b Aa + ... = T a Aa (x) f abc T c b Aa + ...
a a a b
78
We therefore find
Aa (x) = + f b A = exp ib Mac Ac ,
ac abc c b
where
b
Mac = i f abc
qq = q q .
3 3 = 1 8.
79
Under a SU (3)flavour transformation, the quarks and the antiquarks transform as
qi = Ui j q j , qi = Uij q j .
Before discussing the octet representation we first look how the operators H1 , H2 , E1 , E2
and E3 act on the states qq of the nine-dimensional representation. A finite SU (3)flavour trans-
formation acts on such a state as
U q q = (U q) U q .
U = exp (ia T a ) .
In order to find the action of the generators on the tensor representation we expand to first order
in a :
(1 ia T a + ...) q q = q q (ia T a q) q + q (ia T a ) q + ...
The generators H1 and H2 are diagonal and real (H1 = H1 , H2 = H2 ). We can use this formula
to obtain the action of H1 and H2 on the states qq . For example H1 acts on uu as
1 1
H1 (uu) = uu uu = 0.
6 6
Doing this for all other basis vectors qq of the nine-dimensional representation, we find that H1
is given in this representation by
2 1 2 1 1 1
H1 = diag 0, , , , 0, , , , 0 .
6 6 6 6 6 6
80
Similar, H2 is given by
1 1 1 1
H2 = diag 0, 0, , 0, 0, , , , 0 .
2 2 2 2
For the non-diagonal generators we have to be a little bit more carefully. Let us first discuss the
simpler SU (2) case. We can write
i1 I1 + i2 I2 = i+ I+ + i I
with
1 1
= (1 i2 ) , I = (I1 iI2 ) .
2 2
Note that the coefficients are now complex. Therefore complex conjugation gives
(i+ I+ + i I ) = i+ I+
i I
= i I+ i+ I
Therefore
I+ = I , I = I+ .
= (1, 0, 0, 0, 1, 0, 0, 0, 1)T .
81
We have
H1 = 0, H2 = 0,
E = 0,
~ 1 and M
M ~ 2 are the two fundamental dominant weights. The octet representation has the dominant
weight
1
!
~ = M
M ~ 1 +M
~2= 6
1
2
+ . The weights 1/ 6
To which
state does this weight vector correspond ? Let us call the state K
and 1/ 2 are the eigenvalues of H1 and H2 applied to K + . In other words, we must have
1
H1 K + = K + ,
6
1
H2 K + = K + .
2
82
These equations are easily solved and one finds K + = cus, where c is some constant. Requiring
that K + has unit norm leads to
K + = us.
There are two ways to obtain the other states in this representation. The first possibility constructs
first all possible weights of the representation. For the octet representation the occuring weights
are
1
! 1
!
0 6
~m0 = , ~m1 = 1
6 , ~m2 = 1
6 , ~m3 = 3 ,
0 0
2 2
1
!
1
!
6
6
6
~m4 = 1 , ~m5 = 1 , ~m6 = 3 .
0
2 2
For each weight we can then repeat the exercise and solve
(i) (i)
H1 = m1 , H2 = m2 ,
(i) (i) (i) (i)
where m1 and m2 are the components of ~mi = (m1 , m2 ) and is the state which we would
like to solve for. For the weight vectors ~m1 to ~m6 we find
~m1 : K + = us,
~m2 : K 0 = d s,
~m3 : = d u,
~m4 : K = su,
~m5 :
K 0 = sd,
~m6 : + = ud.
H1 = 0, H2 = 0
yields
= c1 uu + c2 d d + c3 ss
83
Both states belong to the octet representation: We know that the octet representation is eight-
dimensional and that the weight spaces corresponding to ~m1 to ~m6 are one-dimensional. There-
fore the weight space corresponding to ~m0 must be two-dimensional.
The second possibility of finding the remaining states in the octet representation starting from
the state K + with the dominant weight is given by repeatidly applying the operators E to the
state K + . We then obtain multiples of the other states. For example:
1 1
E1 K + = d s = K 0 ,
3 3
1 1 1
E2 K + = (ss uu) = 0 ,
3 6 2
1 1
E3 K + = ud = + .
3 3
We can repeat this procedure with the newly found states K 0 , + and the linear combination of
0 and , until we have found all states in the representation.
The classification of the pseudo-scalar meson 0 , , K 0 , K 0 , K , and according to the
representations of SU (3)flavour was very important in early days of particle physics. With four
quark flavours (up, down, strange, charm) the symmetry group can be extended to SU (4)flavour.
Adding a fifth quark (bottom quark) would bring us to SU (5)flavour. In principle one could also
thing about SU (6)flavour by adding the top quark. However, the classification of pseudo-scalar
mesons according to SU (6)flavour is not useful, since the top quark is so heavy and does therefore
not form composite bound states like pseudo-scalar mesons.
It should be added that the naive parton model has some short-comings. The most important
ones are:
SU (3)flavour is only an approximate symmetry. The flavour symmetry is explicitly broken
by mass terms. As the strange quark mass differs the most from the quark masses of
the up- and the down-quark, corrections due to the strange quark mass give the dominant
contribution to SU (3)flavour-breaking terms.
In the modern understanding, a meson does not consist of a quark and an antiquark alone,
but in addition contains an indefinite number of gluons and quark-antiquark pairs.
The physical particles and do not correspond exactly to the pure octet state and the
pure singlet state, but are mixtures of both:
cos sin
= uu + d d 2ss uu + d d + ss .
6 3
sin cos
= uu + d d 2ss + uu + d d + ss .
6 3
84
5 The classification of semi-simple Lie algebras
Recall: For a semi-simple Lie algebra g of dimension n snd r we had the Cartan standard form
Hi , H j = 0,
[Hi , E ] = i E ,
with generators Hi , i = 1, ..., r as well as the generators E and E with = 1, ..., (n r)/2.
The generators Hi generate an Abelian sub-algebra of g. This sub-algebra is called the Car-
tan sub-algebra of g.
The r numbers i , i = 1, ..., r are the components of the root vector ~ = (1 , ..., r ).
2~ ~
~ = ~ 2 ~
Let us now put this a little bit more formally. For any root vector we define a mapping W
from the set of root vectors to the set of root vectors by
2~ ~
W () = ~ 2 ~
W can be described as the reflection by the plane perpendicular to . It is clear that this
mapping is an involution: After two reflections one obtains the original root vector again. The
set of all these mappings W generates a group, which is called the Weyl group.
Since W maps a root vector to another root vector, we have the following theorem:
Theorem: The set of root vectors is invariant under the Weyl group.
Actually, a more general result holds: We have seen that if ~m is a weight and if ~ is a root
vector then
2~ ~m
W (~m) = ~m ~
2
is again a weight vector. Therefore we can state that the following theorem:
Theorem: The set of weights of any representation of g is invariant under the Weyl group.
The previous theorem is a special case of this one, as the root vectors are just the weights of
the adjoint representation.
85
For the weights we defined an ordering. ~m is said to be higher than ~m if the rth component
of (~m ~m ) is positive (if zero look at the (r 1)th component). This applies equally well to
roots.
R = R+ R ,
where R+ denotes the positive roots and R denotes the negative roots.
Definition: The (closed) Weyl chamber relative to a given ordering is the set of points ~x in
the r-dimensional space of root vectors, such that
~x ~
2 2 0 ~ R+ .
Example: The Weyl chamber for SU (3):
2~ ~ 2~ ~
and
||2 ||2
are integers. This restricts the angle between two root vectors to
For = 30 or = 150 the ratio of the length of the two root vectors is
||2 1
= 3 or .
||2 3
For = 45 or = 135 the ratio of the length of the two root vectors is
||2 1
2
= 2 or .
|| 2
86
For = 60 or = 120 the ratio of the length of the two root vectors is
||2
= 1.
||2
Let us summarise: The root system R of a Lie algebra has the following properties:
1. R is a finite set.
2. If ~ R, then also ~ R.
3. For any ~ R the reflection W maps R to itself.
This is the root system of SU (2). For rank 2 we first note that due to property (3) the angle
between two roots must be the same for any pair of adjacent roots. It will turn out that any of the
four angles 90 , 60 , 45 and 30 can occur. Once this angle is specified, the relative lengths of
the roots are fixed except for the case of right angles. Let us start with the case = 90 . Up to
rescaling the root system is
A1 A1 :
This corresponds to SU (2) SU (2). This group is semi-simple, but not simple. In general, the
direct sum of two root systems is again a root system. A root system which is not a direct sum is
called irreducible. An irreducible root system corresponds to a simple group. We would like to
classify the irreducible root systems.
A2
87
This is the root system of SU (3).
B2 :
G2 :
88
The angle between the two simple roots is = 120 .
The Dynkin diagram of the root system is constructed by drawing one node for each sim-
ple root and joining two nodes by a number of lines depending on the angle between the two
roots:
no lines if = 90
Theorem: Two complex semi-simple Lie algebras are isomorphic if and only if they have the
same Dynkin diagram.
Theorem: A complex semi-simple Lie algebra is simple if and only if its Dynkin diagram is
connected.
1 2 3 n1 n
Bn
= SO(2n + 1, C)
1 2 3 n1 n
89
Cn
= Sp(n, C)
1 2 3 n1 n
Dn
= SO(2n, C)
n1
1 2 3 n2 n
E6
1 2 3 5 6
4
E7
1 2 3 5 6 7
4
E8
1 2 3 5 6 7 8
4
F4
1 2 3 4
G2
1 2
An = SU (n + 1)
Bn = SO(2n + 1)
Cn = Sp(n)
Dn = SO(2n)
E6 , E7 , E8 , F4 , G2
90
5.3 Proof of the classification
Recall: The root system R of a Lie algebra has the following properties:
1. R is a finite set.
2. If ~ R, then also ~ R.
With the help of an ordering we can divide the root vectors into a disjoint union of positive and
negative roots:
R = R+ R .
A positive root vector is called simple if it is not the sum of two other positive roots.
The angle between two simple roots is either 90 , 120, 135 or 150
The Dynkin diagram of the root system is constructed by drawing one node for each sim-
ple root and joining two nodes by a number of lines depending on the angle between the two
roots:
no lines if = 90
Theorem: The only possible connected Dynkin diagrams are the ones listed in the previous
section.
To prove this theorem it is sufficient to consider only the angles between the simple roots, the
relative length do not enter the proof.
Such diagrams, without the arrows to indicate the relative lengths, are called Coxeter diagrams.
Define a diagram of n nodes, with each pair connected by 0, 1, 2 or 3 lines, to be admissible if
there are n independent unit vectors ~e1 , ..., ~en in a Euclidean space with the angle between ~ei and
~e j as follows:
91
no lines if = 90
(i) Any sub-diagram of an admissible diagram, obtained by removing some nodes and all lines
to them, will also be admissible.
(ii) There are at most (n 1) pairs of nodes that are connected by lines. The diagram has no
loops.
(iv) In an admissible diagram, any string of nodes connected to each other by one line, with
none but the ends of the string connected to any other nodes, can be collapsed to one node,
and the resulting diagram remains admissible.
Proof of (i): Suppose we have an admissible diagram with n nodes. By definition there are n
vectors ~e j , such that the angle between a pair of vectors is in the set
Removing some of the vectors~e j does not change the angles between the remaining ones. There-
fore any sub-diagram of an admissible diagram is again admissible.
2~ei ~e j 1.
Now
! !
0 < ~ei ~ei = n + 2 ~ei ~e j < n # connected pairs
i j i< j
92
Therefore
Connecting n nodes with (n 1) connections (of either 1, 2 or 3 lines) implies that there are no
loops.
Consider the node ~e1 and let ~ei , i = 2, ..., j bet the nodes connected to ~e1 . We want to show
j
(2~e1 ~ei)2 < 4.
i=2
Since there are no loops, no pair of ~e2 ,...,~e j is connected. Therefore ~e2 , ..., ~e j are perpendicular
unit vectors. Further, by assumption~e1 , ~e2 ,...,~e j are linearly independent vectors. Therefore ~e1 is
not in the span of ~e2 ,...,~e j . It follows
j
1 = (~e1 ~e1 )2 > (~e1 ~ei)2
i=2
and therefore
j
(~e1 ~ei)2 < 1.
i=2
Proof of (iv):
1 2 r
If ~e1 , ..., ~er are the unit vectors corresponding to the string of nodes as indicated above, then
~e ~e = (~e1 + ... +~er )2 = r + 2~e1 ~e2 + +2~e2 ~e3 + ... + +2~er1 ~er
= r (r 1) = 1.
Further~e satisfies the same conditions with respect to the other vectors since~e ~e j is either~e1 ~e j
or ~er ~e j .
93
With the help of these lemmata we can now prove the original theorem:
From (iii) it follows that the only connected diagram with a triple line is G2 .
Further we cannot have a diagram with two double lines, otherwise we would have a sub-
diagram, which we could contract as
...
contradicting again (iii). By the same reasoning we cannot have a diagram with a double line
and a triple node:
...
To finish the case with double lines, we rule out the diagram
1 2 3 4 5
We find
By a similar reasoning one rules out the following (sub-) graphs with single lines:
These sub-diagrams rules out all graphs not in the list of the previous section. To finish the proof
of the theorem it remains to show that all graphs in the list are admissible. This is equivalent
94
to show that for each Dynkin diagram in the list there exists a corresponding Lie algebra. (The
simple root vectors of such a Lie algebra will then have automatically the corresponding angles
of the Coxeter diagram.)
To prove the existence it is sufficient to give for each Dynkin diagram an example of a Lie alge-
bra corresponding to it. For the four families An , Bn , Cn and Dn we have already seen that they
correspond to the Lie algebras of SU (n + 1), SO(2n + 1), Sp(n) and SO(2n) (or SL(n + 1, C),
SO(2n + 1, C), Sp(n, C) and SO(2n, C) in the complex case). In addition one can write down
explicit matrix representations for the Lie algebras corresponding to the five exceptional groups
E6 , E7 , E8 , F4 and G2 .
Finally for the uniqueness let us recall the following theorem: Two complex semi-simple Lie
algebras are isomorphic if and only if they have the same Dynkin diagram.
95