In this chapter we shall obtain the classification of simple Lie algebras over an algebraically closed field of characteristc 0. This was given first by Killing, modulo some errors which were corrected by Cartan. Later simplifications are due to Weyl, van der Waerden, Coxeter, Witt, and Dynkin. In our discussion we shall follow Dynkin’s method which is fairly close to Cartan’s original method. However, we shall formulate everything in terms of “split” semi-simple Lie algebras over an arbitrary base field of characteristic 0. It is easy to see that the assumption of algebraic closure in the classical treatments is used only to ensure the existence of a decomposition of the algebra as =
⊕
α ⊕
β …
δ where
is a Cartan subalgebra and the
α are the root spaces relative to
. This can be achieved by assuming the existence of a “splitting Cartan subalgebra” (cf. § 1). It appears to be clearer and more natural to employ this hypothesis in place of the stronger one of algebraic closure of the base field. For the benefit of a reader who has some familiarity with the associative theory it might be remarked that the split simple Lie algebras which are singled out in the classical theory are the analogues of the simple matrix algebras Φn of the associative theory.
A part of the results of this chapter (the isomorphism and existence theorems) will be derived again in Chapter VII in a more sophisticated way. In Chapter X we shall take up the problem of extending the classification from algebraically closed base fields—or from split Lie algebras—to simple Lie algebras over any field of characteristic 0. It should be noted that the classification we shall give is valid also for characteristic p ≠ 0 under fairly simple hypotheses which are stronger than simplicity. This has been shown by Seligman and, in an improved form, by Mills and Seligman.
We shall call a Cartan subalgebra of a finite-dimensional Lie algebra
a splitting Cartan subalgebra (abbreviated s.c.s.) if the characteristic roots of every ad
h, h ε
, are in the base field. We shall say that
is split if it has a splitting Cartan subalgebra. If the base field Φ is algebraically closed, any finite-dimensional
is split and any Cartan subalgebra is a s.c.s.
Example. Let = ΦnL. and let
denote the subalgebra of diagonal matrices. We have seen (§3.3) that
is a s.c.s. in ΦnL, so ΦnL is split. Next let Φ be the field of real numbers and let A be a matrix whose characteristic roots ξi (in the complex field) are distinct but not every ξi ε Φ. Then the polynomial algebra Φ[A] is a Cartan subalgebra of
. The characteristic polynomial of ad A is
and some of its roots are not in Φ. Hence Φ[A] is not a s.c.s.
In the remainder of this chapter will denote a split finitedimensional semi-simple Lie algebra over a field Φ of characteristic 0,
will be a splitting Cartan subalgebra of
and we shall write (a, b) for f(a, b) = tr ad a ad b, the Killing form on
. We know that (a, b) is non-degenerate. Our assumption on
is that ad
is a split algebra of linear transformations. Hence we know that we can decompose
as
where α, β, …, δ are the non-zero roots. These are linear functions on and
α is the set of elements xα ε
such that xα(ad h — a(h))r = 0 for some r = r(h), h ε
(cf. §2.4 and §3.2). In the same way
=
0, the Fitting null component of
relative to ad
. We have [
α
β]
α + β
if α + β is a root while [
α
β] = 0 if α + β is not a root. Our first task will be to obtain additional information on
, on the roots α and on the corresponding root spaces
α. We shall number these results by Roman numerals.
I. If α and β are any two roots (including 0) and β ≠ — α, then α and
β are orthogonal relative to the Killing form.
Proof: We show first that =
0 ⊥
a if a ≠ 0. Let h ε
eα ε
α, and choose h′ so that α(h′) ≠ 0. Then the restriction of ad h′ to
α is a non-zero scalar plus a nilpotent and so this is nonsingular. It follows that for any k = 1, 2, … we can find an
such that
Since the Killing form Is invariant we have
Since is nilpotent, k can be chosen so that [h′ … [h′h]…] = 0. Then the above relation implies that (eα, h) = 0. Thus
⊥
α for a ≠ 0. Now let β ≠ — α and let eβ ε
β. As before, write eα =
Then (eα, eβ) = (
, eβ) = –
. If a + β is a root it is non-zero and
ε
β+α. Hence
= 0. If a + β is not a root,
= 0 and again
,) = 0. Hence (eα, eβ) = 0 and
α ⊥
β.
II. is a non-isotropic subspace of
(relative to (a, b)). If α is a root, then —α is a root and
α, and
-α are dual spaces relative to (a, b).
Proof. If z ε and z ⊥
, then z ⊥
since, by I, z ⊥
for all α ≠ 0. Then z = 0 by the non-degeneracy of (a, b). If α is a root and —α. is not a root, then
α ⊥
β for every root β. Then
α ⊥
contrary to the non-degeneracy of the Killing form. Also the argument used for α = 0 shows that if z ≠ 0 is in
α, then there exists a w ε
-α such that (z, w) ≠ 0. Similarly, if w ≠ 0 is in
-α then there exists a z ε
α, such that (z, w) ≠ 0. This shows that
α and
-α are dual spaces relative to (a, b).
We recall that the matrices of the restrictions of ad h, h ε , to
α can be taken simultaneously in the form
(§2.6). If dim α = nα and h, k ε
, then this gives the formula
We recall also that α(h′) = 0 for every h′ ε ′. These results imply our next two results.
III. There are l linearly independent roots where l = n0 = dim .
Proof. The roots are linear functions and so belong to the conjugate space * of
. We have dim
*= l. Hence if the assertion is false, then the subspace of
* spanned by the roots has dimensionality l′ < l. This implies that there exists a non-zero vector h ε
such that α(h) = 0 for every root α. Then (3) implies that (h, k) = 0 for every k ε
, contrary to II.
IV. is abelian.
Proof: If h′ ε ′ = [
], then α(h′) = 0 for all α. Then (h′, k) = 0 for all k and h′ = 0. Hence
′ = 0 and
is abelian.
The fact that the restriction of the Killing form to is non-degenerate implies that if ρ(h) is any element of
*. that is, any linear function, then there exists a unique vector hρε
such that
The mapping ρ → hρ of into * is surjective and 1:1. If ρ,σ ε
* we define
Then
and it is immediate that (ρ, σ) is a non-degenerate symmetric bilinear form on *.
V. Let eα be an element of α such that [eαh] = α(h)ea, h ε
, and let e-α, be any element of
-α. Then
proof: We have
Hence (7) follows from the non-degeneracy of (h, k) in .
VI. Every non-zero root α is non-isotropic relative to the bilinear form (ρ, σ) in *.
Proof: We note first that ad eα is nilpotent if α is a non-zero root and eα ε α. For this it suffices to show that if xβε
β, then there exists a positive integer k such that
. Consider the sequence
The vectors in this sequence are either 0 or belong respectively to the roots β + α, β + 2α, β + 3α, … Since there are only a finite number of distinct roots for it follows that
for a suitable k. We can choose eα ≠ 0 in
α so that [eαh] = α(h′)eα (cf. (2)) and e-α ε
_α so that (eα, e-a) = 1. The latter choice can be made since
α,
-α are dual spaces relative to the Killing form. If ρ is any element of
* we define
ρ = 0 if ρ is not a root and
ρ is the root space corresponding to ρ if ρ is a root. Set
It is clear from the rule on the product of root spaces that –ka is a nilpotent subalgebra of
. Also
is a subalgebra containing
–ka as ideal. Hence
is solvable. If x-α ε
-α, then [eαx-α] is a multiple of hα, by (7). Also if k > l, then [eα
-kα]
-(k-1)α and [eαhα] = α(hα) eα = (α, α)eα. This shows that
is a subalgebra of
and that if (α, α) = 0, then
is an ideal in
. Since this ideal is solvable and has one-dimensional difference algebra, (α, α) = 0 implies that
is solvable. Then ad
is a solvable algebra of linear transformations acting in
. Since ad eα is nilpotent, this element is in the radical of the enveloping associative algebra of ad
(Corollary 2 to Theorem 2.8). Hence the same is true of (ad eα) (ad e-α), which implies that (eα,e-α) = tr ad eα ad-α = 0, contrary to (eα, e-α) = 1. Hence (α, α) ≠ 0.
VII. If α is a non-zero root, then nα = dim α = 1. Moreover, the only integral multiples kα of α which are roots are α, 0 and — α.
Proof: Let eα, e-α, be defined as in the proof of VI (cf. (8)). Then
is an invariant subspace of
relative to ad h, h ε
, since [eαh] = α(h)eα, [hαh] = 0 and [
-kαh]
-kα. The restriction of ad h to
-kα has the single characteristic root —ka(h). Hence we have
and, in particular
Since is a subalgebra containing eα and e-α it is invariant under ad eα and ad e-α and since [e-αeα] = hα, [ad
e-α, ad
eœ] = ad
hα. Hence tr (ad
hα) = 0. Since (α, α) ≠ 0 this and (9) imply that 1 — n-α — 2n-2α —…= 0. This occurs only if n-α = 1, n-2α = n-3α = … = 0. Thus —2α., —3α, … are not roots and n-α = 1. Since we can replace α by —α in the argument, we have also that na = 1 and 2α, 3α, … are not roots.
If α is a non-zero root, α = Φeα and [eah] = α(h)eα. Moreover, (7) shows that if we choose eα and e-α so that (eα, e-α) = — 1 then [eαe-α] = hα. Then [eαhα] = (α, α)eα, [eαhα] = — (α, α)eα. If we set
Thus (e′α, e′ -α, h′α) is the type of normalized basis we have considered before for a split three-dimensional simple algebra. This proves
VIII. If α is a non-zero root then Φhα + α +
-αis a split threedimensional simple subalgebra of
.
At this point we have all the information we need on the multiplication in , the products of elements of ·
by elements of the
α and the products [eαe-α], eα ε
α, e-α ε
-α. It remains to investigate products of the form [eαe-α] where α and β are non-zero roots and β ≠ — α. We shall obtain the results required after we have established in the next section a basic result on representations of semi-simple Lie algebras.
We shall need only a part of the following theorem at this point. Later the full result will play an important role in the representation theory.
THEOREM 1. Let be a finite-dimensional split semi-simple Lie algebra over a field of characteristic 0,
a splitting Cartan subalgebra,
=
+ Σ Φeα the decomposition of
into root spaces relative to
. Let hα and (ρ, σ) for ρ, σ in the conjugate space
* be defined as before (cf. (4), (5)) and let eα and e -α be chosen so that [eαe-α] = hα. Let
be a finite-dimensional module for
, R the representation. Then
is a split module for
and if
Λ is the weight module of
corresponding to a weight Λ, then the linear transformation induced by hR in
, is the scalar Λ(h) 1. Let
(α) =
+ Φeα + Φe-α, Then
(α) is a subalgebra and
is a completely reducible
(α)module. Any irreducible
(α)-submodule
of
has a basic (y0, y1, … ym) such that
where M is a weight and 2(M, α)/(α, α) = m. Moreover, if y is any non-zero vector such that yh = M(h)y and yeα = 0, then y generates an irreducible (a)-submodule of
. Let Λ be a weight of
in
and let Σ be the collection of weights of the form Λ + iα, i an integer, α a fixed non-zero root. Then Σ is an arithmetic progression with first term Λ — ra, difference α, and last term Λ + qα and we have
If x is a non-zero vector such that xh = (Λ + qα)(h)x, then x generates an irreducible (α) -submodule and every weight belonging to Σ occurs as weight of
in this submodule. If Λ is a weight, then
is also a weight and
Proof: By III we can find a basis for * consisting of l roots α1, α2 … αl. The corresponding elements hα1 hα2 … hαl form a basis for
. Also
i = Φhαi + Φeαi+ Φe-αi is a split three-dimenional simple Lie algebra and, by (10), e′αi, e′-αi = 2e-αi/(αi, αi), h′αi = 2hαi/(αi,αi) is a canonical basis for
i. If we recall the form of the irreducible modules for such an algebra given in § 3.8 and use complete reducibility of finite-dimensional modules, we see that there exists a basis for
such that h′Rαi hence also
= (1/2)(αi, αi)h′Rαi has a diagonal matrix relative to this basis. This is equivalent to two statements about
: the characteristic roots of
are in Φ and
a semi-simple linear transformation. Since the
commute, a standard argument shows that there exists a basis (u1, u2, … uN) for
such that every
has a diagonal matrix relative to this basis (cf. the proof of Theorem 3.10). Since the hαi form a basis for
it follows that hR has a diagonal matrix relative to (uj), that is, we have: ujh = Λj((h)uj, h ε
, j = 1, 2, ···, N. Thus it is clear that
R is a split abelian Lie algebra of linear transformations and that the Λj = Λj((h) are the weights of
. Moreover, hR is the scalar Λ(h)1 in the weight space
Λ (whose basis is the set of uj such that Λj = Λ). This proves the first statement. Now let α be a non-zero root and consider the subspace
(α) =
+ Φeα + Φe-α Since
is an abelian subalgebra and [eαhh] = α(h)eα, [e-αh] = — α(h)e-α,
(α) is a subalgebra. Let
o be the subspace of
defined by α(h) = 0. Then
=
0 + Φhα and
(α) =
0 ⊕
, where
is the split three-dimensional simple Lie algebra with basis (eα, e-α, hα). We have [
0
] = 0, so
0 is the center of
(α). Now we have seen that hR is semi-simple for every h ε
. Hence the main criterion for complete reducibliity (Theorem 3.10) shows that
is completely reducible as an
(α)-module. Let
be an irreducible
(α) -submodule of
. Then
contains a vector y ≠ 0 such that yh = M(h)y where M is a weight. If we replace y by one of the vectors in the sequence y, yeRα, y(eRα)2, … we may suppose also that
where (e′α, e′-α, h′α) is the basis for
given in (10). Then
and yE = 0. The argument of § 3.8 shows that 2(M, α)/(α, α) is a nonnegative integer m (strictly speaking m. 1) and that (y, yF, …, yFm) is a basis for an irreducible -module. Also we have yFm+1 = 0 and induction on i shows that (yFi)h = (M — iα)(h)yFi Hence ΣΦyFi is an
(a) -submodule of
and since
is irreducible as
-module,
= ΣΦyFi. We now modify the basis for
by replacing yFi by
Then we have yih = (M — iα)(h)yi, yie-α = yi+1 for i = 0,1, ···, m — 1 and yme-α = 0. Also the last formula of (3.36):
becomes
which is the same as the last equation in (12). Thus we have established (12). The argument shows also that any non-zero y in satisfying yh = M(h)y, yeα = 0 generates an irreducible
(α)-submodule of
. This proves our assertions about
(α). Now let Λ be any weight and let Σ be the set of weights of the form Λ + iα, α a fixed non-zero root, l an integer. The weight Λ is a weight for
in one of the irreducible
(α)-modules
into which
can be decomposed and we may suppose that
is generated by y such that yh = M(h)y, yeα = 0. Then Λ = M — kα where 2(M, α)/(α, α) = m and 0
k
m. Let q be the largest integer such that Λ + qα is a weight. Then if x ≠ 0 is chosen so that xh = (Λ + qα)(h)x, we have xeα = 0, since Λ + (q + 1) α is not a weight. Hence x generates an irreducible
(α) -submodule of dimensionality s + 1 where
On the other hand, M = Λ + kα is a weight, so k q, and we have m = 2(M, α)/(α, α) = 2(Λ + kα, α)/(α, α) = 2(Λ, α)/(α,α) + 2k. Then
so k — m q — s. Now the weights in the module
generated by y and those in the module generated by x are, respectively,
Since k q and q — s
k — m,. all those in the first progression are contained in the second and since Λ was arbitrary in Σ, it follows that Σ coincides with the second sequence. Since Λ is contained in this sequence we have q — s
0. If we set r = — (q —s), then the last term becomes Λ — rα where r
0. Also 2(Λ, α)/(α,α) = s — 2q = r — q, which proves (13). We have Λ′ = Λ — [2(Λ, α)/(α, α)]α = Λ + (2q — s)α and —r
2q — s
q by our inequalities. Hence Λ′ ε Σ and so Λ′ is a weight. It remains to prove (15). We observe that the argument just used shows that Λ′ also occurs in the first sequence in (17), that is, Λ′ is a weight in every irreducible
(α)-submodule
which has Λ as weight. We have dim
Λ = 1 = dim
Λ′. Hence if
= Σ ⊕
j, is a direct decomposition of
into irreducible
(α)-submodules, then dim
Λ is the number of
j which have Λ as weight for
. Hence dim
Λ = dim
Λ′.
This result applies in particular to the adjoint representation of . Our result states that if β and α are any two roots, α ≠ 0, then the roots of the form β + iα, i an integer, form an arithmetic progression with difference α. We call this the α-string of roots containing β. If the string is β — rα, β — (r — 1)α, · · ·, β + qα, then
A number of consequences can now be drawn from Theorem 1. We continue the numbering of § 1.
IX. If α, β and, α + β are non-zero roots, then [eαeβ] ≠ 0 for any eα, ≠ 0 in α and any eβ ≠ 0 in
β.
Proof: The «-string containing β is β — rα, …, β + qα and q 1. None of these roots is 0 since no integral multiple of α is a root except 0, ±α. Since the root spaces corresponding to nonzero roots are one-dimensional this holds for our string. Let x be a non-zero element of
β+qα. Then [xh] = (β + qa)(h)x. If we choose e-α so that [eαe-α] = hα, then the theorem shows that x, x ad e-a, x(ade-α)2, ···, x(ad e-α)r+q are non-zero and these span the spaces
β+qa,
β+(q-1)α. In particular, eβ is a non-zero multiple of x(ade-α)q. Formula (12) implies that
which is not zero r 0 since and q > 0. Hence [eβeα] ≠ 0.
We see also that [[eβeα]e-α] = — (q(r + l)/2)(α, α)eβ. This formula is valid also if β + α is not a root since in this case [eβeβ] = 0 and q = 0. Also, if β = — α, then the α-string containing β is α, 0, § α, so r = 0, q = 2. The right-hand side of the formula is —(α, α)e-α and the left-hand side is [[e-αeβ]e-§] = — [hχβ-α] = [e-α h hα] = —(α, α)e-α, so again the formula holds. We have therefore proved
X. Let a and β be non-zero roots and eβ ε β, eβ ε
β, e-α ε
-β satisfy [eβe-β] = hα. Suppose the α-string containing β is β — rα, … β, β + qα. Then
We remark also that it a and —a are interchanged, r and q are interchanged and we have [e-αeα] = — hα = hα, (—α, —α) = (α, α). Hence another form of (20) is
XI. No multiple of a non-zero root a is a root except 0, a, — a.
Proof : Let β = kα be a root. Then 2(α, β)(α, α) = 2k is an integer and we may assume this is odd and positive. Then it is immediate that the α-string containing β contains γ = (1/2) α. This contradicts the fact that α = 2γ cannot be a root.
XII. The a-string containing β (α, β ≠ 0) contains at most four roots. Hence 2(α, β)(α, α) = 0, ±1, ±2, ±3.
Proof. We may assume that β ≠ α, —α since the α-string containing α consists of the three roots α, 0, —α. Assume we have at least five roots. By re-labelling these we may suppose that β — 2α, β — α, β, β + α, β + 2α are roots. Then 2α = (β + 2α) — β and 2(β + α) = (β + 2α) + β are not roots. Hence the β-string containing β + 2α has just the one term β + 2α. Hence (β + 2α, β) = 0. Similarly β — 2α — β and β — 2α + β are not roots, so that (β — 2α, β) = 0. Adding we obtain (β, β) = 0, contradicting the fact that the non-zero roots are not isotropic. We now have 2(α, β)/(α, α) = (r — q) while r + q + 1 4. Hence r
3, q
3 and 2(α, β)/(α, a) = 0, ±1, ±2, ±3.
From now on we shall identify the prime field of Φ with the field Q of rational numbers. As before, let * be the conjugate space of
and now let
* denote the Q-space spanned by the roots. We have seen that the roots span the space
*. We now prove
XIII. dim 0* = l = dim
.
Proof. It suffices to show that if (α1, α2, ···,αl) is a basis for * consisting of roots, then every root β is a linear combination of the αi with rational coefficients. We have β = Σ λiαi, λi ε Φ. Hence (β, αj) = Σ λi(αi, αj), j = 1, 2, …, l and
This system of equations has integer coefficients [2(αiαj)/(αj, α,j)], [2(β, αj)/(αj, αj,)] and the determinant
since (ρ, σ) is non-degenerate and the αi are a basis for *. Hence (21) has a unique solution which is rational. Thus the λi are rational numbers.
If Λ is a weight of for a reβresentation of
, then we have seen that 2(Λ, α)/(α, α) is an integer for every non-zero root α. The argument just given for roots shows that Λ is a rational linear combination of the roots αj, j = 1, 2, ···, l. Thus the weights Λ ε
*0.
XIV. (ρ, σ) is a rational number for ρ, σ ε *0 and (ρ, σ) is a positive definite symmetric bilinear form on
*0.
Proof: We recall the formula (3) for (h, k) h, k ε . If we take into account the fact that the
α are one-dimensional we can rewrite this as
Now let β be a non-zero root. Then (β, β) = Σα(β, α)2. Let the β-string containing α be
Then, by (18), 2(α, β)/(β, β) = rαβ - qαβ and (α, β) = [(rα β — qαβ)/2](β, β). Hence
Since (β, β) ≠ 0, Σα(rαβ — qαβ)2 ≠ 0 and
is a rational number and (α, β) = [(rα β — qαβ)/2](β, β). is rational for any roots roots. Then
and (ρ, ρ) = 0 implies that every (ρ, α) = 0. Then ρ = 0 since the roots α span
*.
We recall that in a space with a non-degenerate bilinear form (ρ, σ), if α is a non-isotropic vector, then the mapping Sα
is a linear transformation which leaves fixed every vector in the hyperplane orthogonal to α and sends α into —α. We call this the reflection determined by α. This belongs to the orthogonal group of the form (ρ, σ).
We have seen (Theorem 1) that if Λ is a weight of in a representation of
, then Λ′ = ΛSα = Λ — [2(Λ, α)/(α, α)](α) is also a weight. The reflections Sα, α a root, generate a group of linear transformations in
0* called the Weyl group W of
(relative to
). This group plays an important role in the representation theory for
. The result we have just noted is that the weights of a particular reβresentation are a set of vectors which is invariant under the Weyl group. In particular, this holds for the roots. If two elements of W produce the same permutation of the roots they are identical since the roots span
*0. Since there are only a finite number of roots it follows that W is a finite group.
We now introduce an ordering in the rational vector space 0*. For this purpose we choose a basis of roots α1 α2…. αl and we call
positive if the first non-zero λi is positive. The set of positive vectors is closed under addition and under multiplication by positive rationals. If σ, ρ ε
*0 we write σ > ρ if σ — ρ > 0. Then
* is totally ordered in this way and if σ > ρ then a + τ > ρ + τ and λσ > λρ or λσ < λρ according as λ > 0 or λ < 0. We shall refer to the ordering of
0* just defined as the lexicographic ordering determined by the ordered set of roots (α1, α2 ··αl) (which form a basis for
0*).
LEMMA 1. Let Suppose the ρi > 0 and (ρi, ρj)
0 if i ≠ j. Then the ρ’s are linearly independent over Q.
Proof: Suppose Since
We have
. Hence
On the other hand, which is a contradiction. Hence the ρ’s are Q-independent.
DEFINITION 1. Relative to the ordering defined in *0 we call a root α simple if α > 0 and α cannot be written in the form β + γ where β and γ are positive roots.
XV. Let π be the collection of simple roots relative to a fixed lexicographic ordering of *0. Then:
(i) If α, β ε π, α ≠ β, then α — β is not a root.
(ii) If α, β ε π and α ≠ β, then (α, β) 0.
(iii) The set π is a basis for 0* over Q. If β is any positive root, then β = Σαεπkαα where the kαare non-negative integers.
(iv) If β is a positive root and β π, then there exists an α ε π such that β — α is a positive root.
Proof: (i) If α, β ε π and α — β is a positive root, then α = β + (α — β) contrary to the definition of π. If α — β is a negative root we again obtain a contradiction on writing β = (β — α) + α. (ii) Let β — rα, β — (r — 1 )α, …, β + qα be the α-tsring containing β. Then 2(α, β)/(α, α) = r — q. Since r = 0 by (i) and (α, α) > 0, (α, β) 0. (iii) The linear independence of the roots contained in π is clear from (ii) and the lemmα. Let β be a positive root and suppose we already know that every root γ such that β > γ > 0 is of the form Σαεπkαα, kα a non-negative integer. We may suppose also that β
π, so that β = βι + β 2, βi > 0. Then β > βi and
non-negative integers; hence β = Σ (k′α + k′′α) which is the required form for β. If β is a negative root then — β is a positive root. Hence β = Σkαα where the kα are integers
0. The first statement of (iii) follows from this and the linear independence of the elements of π. (iv) Let β be a positive root not in π. The lemma and (iii) imply that there is an α ε π such that (β, α) > 0. Then 2(β, α)/(α, α) = r — q > 0 (r, q as before). Hence r > 0 and β — α is a root. If β — α < 0, then α — β > 0 and α = β + (α — β) contrary to: α ε π. Hence β — α > 0 and β = (β — α) + α where α ε π.
We now write π = (αl α2, …, αl) and we call this the simple system of roots for relative to
and the given ordering in
o*. We have seen that every root β = Σkiαi where the ki are integers and either all ki
0 or all ki
0. This property is characteristic of simple systems. Thus let π = (α1, α2, …, αl) where l = dim
and the are roots such that every root β = Σ kiαi where the ki are integers of like sign (rationale would do too). Evidently our hypothesis implies that π is a basis for
0*. We introduce the lexicographic ordering of
0* based on the
Then the positive roots β = Σkiαi are those such that the ki 0 and some ki > 0. It is clear that no αi is a sum of positive roots. Hence the αi are simple. Since any simple system consists of l roots the set π = (α1 α2, …, αl) is the simple system defined by the ordering.
If π = (α1 α2, …, αl) is a simple system of roots, the matrix (Aij), Aij = 2(αi, αj)/(αi, αi) is called a Cartan matrix for (relative to
). The diagonal entries of this matrix are Aii = 2 and the off diagonal entries are Aij = 0, —1, —2 or —3. (XII and XV (ii)). If i ≠ j the αi and αj are linearly independent so that if θij is the angle between αi and αj then
This gives
This implies that either both Aij and Aji are 0 or one is —1 while the other is § 1, —2, or —3. The determinant of the Cartan matrix (Aij) is a non-zero multiple of that of ((αi, aj)). Hence det (Aij) ≠ 0.
We now choose and we now write
These elements have the canonical multiplication table for a split three-dimensional simple Lie algebra: [eihi] = 2ei, [fihi] = —2fi,[eifi] = hi. Also we have [eifj] = 0 if i ≠ j since αi — αj is not a root, [eihj] = 2(αi αj)/(αj, αj)ei = Ajiei and [fihj] = — Ajifi. The last relations include [eihi] = 2ei, [fihi] = — 2fi. Thus we have the following relations for the ei, fi, hi:
i, j = 1,2, ··· l
We wish to show that the 3l elements ei, fi, hi (or the 2l elements ei, fi, since hi = [eifi]) generate . Moreover, we shall show that we can obtain a basis for
whose multiplication table is completely determined by the Aij. This will prove that
is determined by the Cartan matrix.
If β = Σ kiαi is a root then we define the level |β| = Σ |ki|. The level is a positive integer and the positive roots of level one are just the αi ε π.
XVI. The set of roots is determined by the simple system π and the Cartan matrix. In other words, the sequences (k1 k2, …, kl) such that Σ kiαi are roots can be determined from the matrix (Aij).
Proof: It suffices to determine the positive roots. The positive roots of level one are just the αi, ε π. Now suppose we already know the positive roots of level n, n a positive integer. We give a method for determining the positive roots of the next level. By XV these are of the form β = α + αj, α > 0 of level n, αj ε π. Hence the problem is to determine for a given α > 0 of level n the αj ε π such that α + αj is a root. If α = αj, α + αj is not a root. Hence we may assume that α = Σ kiαi and some ki > 0 for i ≠ j. Then the linear forms α — αj, α — 2αj, … which are roots are positive of level less than n. Hence one knows which of these are roots. Thus the number r such that the α-string containing α is α —rαj, ···,α, ···, α, + qαj is known. We have
hence q can be determined by the Cartan matrix. Since α + αj is a root if and only if q > 0 this gives a method of ascertaining whether or not α + αj is a root.
Example. Suppose the Cartan matrix is
that is,
Since α1 — α2 is not a root these relations imply that the α1-string containing α2 and the α2-string containing α are, respectively,
The only positive root of level two is α1 + α2. Since α2 + 2αi is not a root the only positive root of level three is α1 + 2α2,. Since 2α1 + 2α2 is not a root the only positive root of level four is α1 + 3α2. We have 2(α1+ 3α2, α1)/(α1, α1 = 2 – 3 = – 1, which implies that (α1 + 3α2) + α1 = 2α1 + 3α2 is a root. Since α1 + 4α2 is not a root 2α1 + 3α2 is the only positive root of level five. Since (2α1 + 3α2) + α1 = 3(α1 + α2) and (2α1 + 3α2) + α2 = 2(α1 + 2α2) are not roots there are no roots of level six or higher. Hence the roots are
A simple induction on levels shows that any positive root β can be written as
αij ε π in such a way that every partial sum
is a root. The number k in (β2) is the level of β. We shall now abbreviate Then the result IX implies that
hence this element spans β. It follows also that
and this element spans –β. since the αi ε π torm a basis for, the
*, the hαi and hence the hi form a basis for
. Since
=
+ Σ (
β +
–β) summed over the positive roots, this proves
XVII. Let π — (α1, α2, …, αl) be a simple system of roots for relative to
and let ei, fi, hi be as in (27). Then the 3l elements ei, fi, hi generate
. For each positive root β we can select a representation of β = αi1 + αi2, + … + αik so that αi1 + · … + αim is a root for every m
k. Then the elements
determined by the positive roots β form a basis for .
We shall be interested in the multiplication table of the basis (36). For this we require
XVIII. Let β be a positive root and let the sequence i1, i2, …, ik be determined by β as in XVII. Let 1′, 2′, …, k′ be a permutation of 1, 2, … k. Then [ei1.ei2, … eik,] is a rational multiple of [ei1.ei2, … eik,], the multiplier being determined by the Aij. A simliar statement holds for the f’s.
Proof: This is clear for k = 1, so we use induction and assume the result for positive roots of level k — 1. If ik = ik, = j, then we may assume that the sequence chosen for the root β — αj is i1, i2 … ik–1 Then the result for k — l implies that [ei1, … ei(k–1),] = t[ei1 … eik–1] where t is a rational number determined by the Aij. Then Next suppose ik = j ≠ ik and write
where the displayed ej is the last one occurring in the expression. If any of the partial sums αi1 + … + a im, is not a root, then [ei1 … eik,] = 0. Since this fact can be ascertained from the knowledge of the Cartan matrix the result holds in this case. Now suppose every αi1, + … + αim, is a root. Then [ei1, … eik,] ≠ 0. Since and
By (19), , is a root ≠ 0) where the q and r are integers, q > 0, r
0 and q and r are determined by the αj string containing β = αi1+ … + air′. This string is known from (Aij). A similar argument shows that
where s is a non-zero integer which can be determined from the Aij. It follows that
Thus where t is a rational number which can be determined from the Aij. This reduces the discussion to the first case. The f’s can be treated similarly.
We can now prove the following basic theorem.
THEOREM 2. Let π = (α1 α2, … αl) be a simple system of roots for a split semi-simple Lie algebra relative to a splitting Cartan subalgebra
. Let the ei, fi, hi i = 1, 2, …, l, generators of
defined in (27) and let the basis hi, [ei1 … eik], [fi1 … fik] for
be as specified in XVII. Then the multiplication table for this basis has rational coefficients which are determined by the Cartan matrix (Aij).
Proof: We have [hihj] = 0 and by (28). similarly,
. It remains to consider products of e terms and f terms. Since [ei1, … eik =
and
. This can be obtained by operating on x with a certain (non-commutative) polynomial in ad ei1, …, ad eik. It follows from this and a similar argument for the f’s that it suffices to show that [[ei1 … eik]ej], [[fi1 … fik]ej], [[ei1 … eik]fj], [[fi1 … fik]fj] are rational combinations of our base elements where the coefficients can be determined by the Aij. The argument is the same for the last two as for the first two so we consider the first two only. To evaluate [[ei1 … eik]ej] we ascertαi n first whether or not β + αj, β = αi1 + … + αik is a root. If not, then the product is 0. On the other hand, if γ = β + αj is a root then, by XVIII, [[ei1 … eik]ej] is a rational multiple of the e-base element associated with γ, the multiplier determined by the Cartan matrix. Next consider [[fi1 … fik]ej]. If k = 1 the product is 0 unless i1 = j, in which case [fi1ej] = — hj. If k
2 we shall show by induction that the product is a rational linear combination of f-base elements. Thus, if k = 2 the product is 0 unless j = i1 or j = i2. For [[fi1fj]ej] we have [[fi1fj]ej] = [fi1[fjej]] + [[fi1ej]fj] – [fi1hj] = Aji1fi1, since i1 + j and so [fi1ej] = 0. The relation just derived implies also that [[fi,fi1]ej] = — Aji1fi1. Now assume k > 2. If no ir = j, then the product is 0. Otherwise, let ir+l be the last index in [fi1, ··· fik] which equals j. Then
The first term is a rational multiple determined by the Aij of an f-base element. The induction hypothesis establishes the same claim for the second element. This completes the proof.
SUMMARY. Before continuing our analysis, it will be well to summarize the results which we have obtained. For any split semi-simple Lie algebra with splitting Cartan subalgebra
we have obtained a canonical set of generators ei, fi, hi, i = 1, 2, …, l satisfying the defining relations (28). These were obtained by choosing a simple system π of roots. The characteristic property of π is that every root
, where the ki are all either non-negative or non-positive integers. Such systems are obtained by introducing lexicographic orderings in the space
*0 of rational linear combinations of the roots and selecting the positive roots which are not of the form a + β, α, β > 0, in such an ordering. A canonical set of generators associated with π is hi = 2hα/(αi, αi), e = eαi, fi = 2e-αi/(αi, αi) where eαi is any non-zero element of
βi and e-αi is chosen in
-αi so that [eαi e-αi] = hαi. We observe that hi is uniquely determined by αi while ei can be replaced by any μiei, μi ≠ 0, in Φ. Then fi will be replaced by
. The elements ei, fi, hi constitute a canonical basis for a split three-dimensional simple Lie algebra
i. Moreover, it is easy to check that if
is any canonical basis for
i such that
If ei, fi, hi are canonical generators we obtain a canonical basis for in the following manner: the basis hi for
and for each non-zero root β a base element eβ = [ei1 … eik] or [fi1 … fik] according as β > 0 or β < 0 where (i1 …, ik) is a sequence such that αi1 + … + αik = ± β and every partial sum αi1 … αim + a im is a root.
The multiplication table for the canonical basis is rational and is determined by the Cartan matrix (Aij), Aij = 2(αi, αj)/(αi, αi). The Aij are integers, Aij = 2 and if i ≠ j, then either Aij = 0 = Aji or one of the numbers Aij, Aji is —1 and the other is —1, —2 or —3. We observe also that the group of orthogonal linear transformations generated by the reflections
which is a subgroup of the Weyl group W is finite. (Later we shall see that this subgroup coincides with W.) The reflection Si can be described by the Cartan matrix. Thus if we take the basis (α1, α2, · αl) for *0, then Si is completely described by
The conditions we have noted on the Aij are redundant. However, sometimes one subset of these conditions will be used while at other times another will be used.
Theorem 2 of the last section makes the following isomorphism theorem almost obvious.
THEOREM 3.Let ,
′ be split semi-simple Lie algebras over a field Φ of characteristic 0 with splitting Cartan subalgebras
,
′ of the same dimensionality l. Let (α1, α2, … αl), (α′1, α′2, …, α′l) be simple systems of roots for
and
′ respectively. Suppose the Cartan matrices (2(αi, αi))/(2(α′i, α′j)/(α′i, α′i)) are identical. Let ei, fi, hi, e′i, f′i, h′i, i = 1,2, ···,l, be canonical generators for
and
′ as in (27). Then there exists a unique isomorphism of
onto
′ mapping e¿ on ei, on e′i, fi on f′i, hi on h′i.
Proof: By XVI we know that Σkiαi is a root for if and only if Σkiα′i is a root for
′. If β is a positive root for
we write β = αi1 + … αik so that αi1 + … αim is a root, m
k. Then β′ = α′i1 + … α′ik and every α′i1 + … + α′im is a root. We can choose as base elements in the canonical basis (36) for
and
′ the elements
Then Theorem 2 shows that the coefficients in the multiplication table for the basis for
and
′ are identical. Hence the linear mapping which matches these base elements is the required isomorphism. The uniqueness is clear since the ei, fi, hi are generators.
The result we have just proved is basic for the problem of determining the simple Lie algebras. It is useful also for the study of automorphisms (which we shall consider later) of a single Lie algebrα. We note here that the result implies that there exists an automorphism of mapping ei →, fi, fi → ei, hi → —hi. This follows by observing that α′i = —αi, i = 1, ···,l, is a simple system and e′i = fi, f′i = ei, h′i = —hi is a corresponding set of generators. We shall need this in § 7.
DEFINITION 2. A simple system of roots π = (α1, α2, …, αl) is called indecomposable if it is impossible to partition π into non- vacuous non-overlapping sets π′, π′′ such that Aij = 0 for every αi ε π′, αj ε π′′.
THEOREM 4. is simple if and only if the associated simple system, π of roots is indecomposable.
Proof. Suppose first that π = (α1, … αk) ∪ (αk+1, ···, αl) 1 k < l, so that Aij = 0, i
k, j > k. Choose canonical generators ei, fi, hi and let
1 denote the subalgebra generated by the ej, fj, hj, j
k. It is readily seen that
where
1 is the subspace of
spanned by the hj and the summation is taken over the roots γ which are linearly dependent on the αj,. Hence 0 ⊂
1 ⊂
. If r > k, j
k, then Ajr = 0 and since αj — αr is not a root, this implies that αj + αj is not a root. Hence [ejer] = 0 as well as [fj, er] = 0. Also [hjer] = 0 and consequently er is in the normalizer of
i. similarly, fr is in the normalizer of
1. Since
1 is contained in its own normalizer it follows that
is the normalizer of
1. Hence
1 is an ideal and
is not simple. Conversely, suppose
is not simple. Then
=
1 ⊕
2 where the
i are non-zero ideals. Let α be a non-zero root and let eα ε
α. Then
and
Since
α is one dimensional this implies that either
α
1 or
α
2. Since [
1
2] = 0 and [
α
-α] ≠ 0 we have either
. In particular, we may order the canonical generators and ei, fi so that e1 f1, …, ek, fk ε
1, ek+1, fk+1, …, el, fl ε
2. Since the
i are non-zero ideals, l
k < l and 0 = [ej[erfr]] = [ejhr] = Arjej if j
k and r > k. Hence Ajr = Arj = 0 and the simple system of roots is decomposable.
The results of the last section reduce the problem of determining the simple Lie algebras to the following two problems: (1) determination of the Cartan matrices (Aij) corresponding to indecomposable simple systems of roots and (2) determination of simple algebras associated with the Cartan matrices. We consider (1) here and (2) in the next section. We observe first that the indecomposability condition amounts to saying that it is impossible to order the indices (or the αi) so that the matrix has the block form
where B, C are not vacuous.
We now associate a diagram—the Dynkin diagram — with the Cartan matrix Aij. We choose l points α1, α2, ··· αl and we connect αi to α1, i ≠ j, by AijAji lines. Also we attach to each point αi the weight (αi, αi). If Aij = 0 = Aji, αi and αj, are not connected and if Aij ≠ 0, Aji ≠ 0, then Aij/ Aji = (αi, αi)/(αj, αj). Hence Aji/Aij and AijAji can be determined from the diagram. Since Aij is nonpositive this information determines Aij and Aji. Thus the matrix can be reconstructed from the diagram of points, lines and the weights. We consider two examples:
For G2 we have A21/A12 = (α1 α1)/(α2α2) = 3, A12A21 = 3 which implies thatA12 = –1, A21 = –3. Hence the Cartan matrix is the matrix in (29). For Ai we have and all the other Aij = 0. Hence the matrix is
In determining the Dynkin diagrams we at first drop the weights (αi, αi) on the points and consider only the collection of points and the lines joining these. We have l points α1, α 2, …, αi, and αi and αj, i ≠ j, are not connected if AijAji = 0 and are connected by AijAji = 1, 2, or 3 lines if AijAji ≠ 0. The αi are linearly independent vectors in a Euclidian space E0 over the rationals. This can be imbedded in the Euclidian space E = E0R over the field R of real numbers. If θij denotes the angle between αi and αj then AijAji = 4 cos2 θij and cos θij 0.
Any finite set α1 α2, …, αl of linearly independent vectors in a Euclidian space (over the reals) will be called an allowable configuration (a.c.) if 4 cos2 θij = 4(αi, αj)2 / (αi, αi) (αj, αj) = 0, 1, 2 or 3 and cos θij 0 for every i, j, i ≠ j. Thus cos
or
and accordingly θij = 90°, 120°, 135°, or 150°. We may replace αi by the unit vector ui which is a positive multiple of αi. Then the conditions are
The Dynkin diagram (without weights) of an α.c. π is a collection of points ui, i = 1, ···, l, and lines connecting these according to the rule given before: ui and uj are not connected if (ui, uj) = 0 and ui and uj are connected by 4(ui, uj)2 = 1, 2 or 3 lines otherwise. An α.c. is indecomposable if it is impossible to partition π into non-overlapping non-vacuous subsets π′, π′′ such that (ui, uj) = 0 if ui ε π′, uj ε π′′. The corresponding condition on the Dynkin diagram is connectedness′. If u, v ε π, then there exists a sequence ui1 = u, ui2, …, uik = v such that uij and uij+1 are connected in the diagram. If the Dynkin diagram is known, then all (ui, uj) will be known. We shall determine the Dynkin diagrams for all the connected α.c. after a few simple observations as follows.
1. If S is a Dynkin diagram, the diagram obtained by suppressing a number of points and the lines incident with these is the Dynkin diagram of the α.c. obtained by dropping the vectors corresponding to the points.
2. If l is the number of vertices (points) of a Dynkin diagram, then the number of pαi rs of connected points (u, v, (u, v) ≠ 0) is less than l.
Proof. Let u = Σl1ui.Then
If (ui, uj) ≠ 0, then 2(uiuj) — 1. Hence the inequality shows that the number of pairs ui, uj with (ui, uj) ≠ 0 is less than l.
3. A Dynkin diagram of an α.c. contains no cycles. (A cycle is a sequence of points u1 ···, uk such that ui is connected to ui+1, i k — 1 and uk is connected to ui.)
Proof. The subset forming a cycle is a diagram of an α.c. violating 2.
4. The number of lines (counting multiplicities) issuing from a vertex does not exceed three.
Proof. Let u be a vertex, v1, v2, …, vk the vertices connected to u. No two vi are connected since there are no cycles. Hence (vi, vj) = 0, i ≠ j. In the space spanned by u and the vi we can choose a vector v0 such that (v0, v0) = 1 and v0, v1 …, vk are mutually orthogonal. Since u and the vi, i 1, are linearly independent, u is not orthogonal to v0 and so (u, v0) ≠ 0. Since u = ∑k0(u, vj)vj,
Hence . Since 4(u, vi)2 is the number of lines connecting u and vi we have our result.
5. The only connected α.c. containing a triple line is
This is clear from 4.
6. Let π be an a.c. and let v1, v2, …, vK be vectors of π such that the corresponding points of the diagram form a simple chain in the sense that each one is connected to the next by a single line. Let π′ be the collection of vectors of π which are not in the simple chain, v1, …, vk together with the vector . Then π′ is an a.c.
Proof: We have 2(vi, vi+1) = −1, i = 1, …, k− 1. Hence (v, v) =
k + 2 ∑i<j(vi, vj). Since there are no cycles (vi, vj) = 0 if i < j unless j = i + 1. Hence (v, v) = k − (k − 1) = 1 and v is a unit vector. Now let u ∈ π, u ≠ vi. Then u is connected with at most one of the vi, say vj since there are no cycles. Then and 4(u, v)2 = 4(u, vj)2 = 0,1, or 3 as required.
The diagram of π′ is obtained from that of π by shrinking the simple chain to a point: Thus we replace all the vertices vi by the single vertex v and we join this to any u ∈ π, u ≠ vi by the total number of lines connecting u to any one of the vj in the original diagram. Application of this to the following graphs
reduces these respectively to
Since the center vertex in each of these has four lines from this vertex, these cannot be diagrams of a.c., by 4. Hence we have
7. No Dynkin diagram contains a subgraph of the form (40).
8. The only possible connected Dynkin diagrams have one of the following forms
Proof: If a connected Dynkin diagram S contains a triple line then it must by G2 by 5. If S contains a double line it contains only one such line and it contains no node, that is, graph of the form This is clear from 7. Also it is clear that S cannot contain two nodes. This reduces the possibilities to those of (42).
We now investigate the possibilities for p, q, r in the second and third types in (42). For the second type set Since 2(ui, ui+1) = − 1 and 2(vi, vj+1) = −1 we have
Since pq > 0 this gives (p + 1) (q + 1) >2pq which is equivalent to (p − 1) (q − 1) < 2. Hence the only possibilities for the positive integers p, q are
The first two cases differ only in notation. Hence we have
9. The only connected Dynkin diagrams of the second type in (42) are
Finally we consider the third case in (42). Set The vectors u, v, w are mutually orthogonal and z is not in the space spanned by these vectors. Hence if θ1, θ2, θ3 respectively, are the angles between z and u, v and w then cos2 θ1 + cos2 θ2 + cos2 θ3 < 1 (cf. the proof paragraph 4 above). Now
Similarly so that we have
We may suppose p q
r (
2). Then p−1
q−1
r−1
and (49) implies that 3r−1 > 1. Since r
2. this gives r = 2. Then (49) gives p−1 + q−1
. Hence 2q−l >
and q < 4. Hence 2
q < 4. If q = 2, then the condition is that p−1 which holds for all p. If q = 3, then the condition is p−1 > 1/6 and p < 6. Hence in this case p = 3, 4, 5. Thus the solutions for p, q, r are
10. The only connected Dynkin diagrams of the third type in (42) are
We have now completed the proof of the following
THEOREM 5. The only connected Dynkin diagrams are Al l 1, Bl = Cl, l
2, D, l
4 and the five “exceptional” diagrams G2, F4, E6, E7, E8, given in (42), (48), (51), and (52).
We now re-introduce the weights on the diagrams. This will give all the possible Cartan matrices: We recall that in the Dynkin diagram obtained from the simple system π = (α1, α2, …, α1), Aij Aji, i ≠ j, is the number of lines connecting αi and αj. If Aij ≠ 0, Aji ≠ 0, then Aji/Aij = (αi, αi)/(αj, αj) and Aij or Aji = −1 while the other of these is −1, −2 or −3. Since nothing is changed in multiplying all the αi by a fixed non-zero real number, we may take one of the αi to be a unit vector. If the diagram has only single lines, then all the (αi, αi) = 1 since the diagram is connected. Hence the weighted diagrams for Al, Dl, E6, E7, E8 are
For G2 we have chosen the notation so that
For F4 we may take the weights as follows:
For Bl and Cl we take the following diagrams:
These diagrams give the possible Cartan matrices.
The time has now come to reveal the identity of the principal characters of our story—the split simple Lie algebras. With every connected Dynkin diagram of an α.c. which we determined in § 5 we obtain a corresponding Cartan matrix (Aij) and there exists for this matrix a split simple Lie algebra with canonical generators ei, fi, hi, i = 1,2, …,l such that [eihj] = Ajiei We shall give the simplest (linear) representation of the algebras corresponding to diagrams Al, Bl, Cl, Dl, G2, F4 and E6. Later (Chapter VII) another approach will be used to prove the existence of split simple Lie algebras corresponding to the diagrams E7 and E8.
We recall that if is an irreducible Lie algebra of linear transformations in a finite-dimensional vector space over a field of characteristic 0, then
where
1 is semi-simple and
is the center. Hence such an algebra is semi-simple if and only if
= 0. We note also that if
is semi-simple and
contains an abelian subalgebra
such that
… where α, β, … are non-zero mappings of
into Φ such that [eαh] = α(h)eαh ∊
is a splitting Cartan subalgebra for
and
is split. We shall use these facts in our constructions.
Let be the algebra of linear transformations in an (l + 1)- dimensional vector space
over
, l
1. It is well known (and easy to prove) that
acts irreducibly on
. We have
where
is the derived algebra and is the set of linear transformations of trace 0. Evidently, any
-invariant subspace is
-invariant. Hence
is irreducible. Also the decomposition
shows that the center of
is 0. Hence
is semisimple.
We now identify with the algebra
l+1 of (l + 1) × (l + 1) matrices with entries in
,
with
the set of matrices of trace 0. We introduce the usual matrix basis (eij), i,j = 1, …, l + 1, in
l+1 so that
A basis for is
Set . Then the set of h′s is an abelian subalgebra
of l dimensions and
The l2 + l linear functions are distinct and are non-zero weights of ad
. We have
where α runs over these weights. It follows that
is a splitting Cartan subalgebra and the α are the non-zero roots. Set
This shows that every root has the form ∑kiαi where the ki are integers and ki 0 for all ki
0 for all l. Hence the αi form a simple system of roots for
relative to
. Equations (57) and (58) show that αiΛ+j; is not a root, 1
i
l − 1, αi + 2αi + 1 is not a root, and αi + αj is not root if j > l + 1. This means that the Cartan integers Aij have the following values:
Hence the Dynkin diagram is the connected simple chain
It follows that is simple of type Al (with Dynkin diagram Al). Hence we have
THEOREM 6. Let be the Lie algebra of linear transformations of trace zero in an (l + 1)-dimensional vector space over
. Then
is a split simple Lie algebra of type Al.
Assume next that is n-dimensional over
and is equipped with a non-degenerate bilinear form (x, y) which is either symmetric or skew. Let
be the Lie algebra of linear transformations A which are skew relative to (x,y), that is, (xA, y) = −(x, yA) (cf. § 1.2). We require the following
LEMMA 2. is irreducible if n
3.
Proof: Let be a non-zero invariant subspace relative to
and let z be a non-zero vector in
. Let u be any vector in the orthogonal complement
. Choose
and consider the linear transformation A: x →(x, y) − (v, x)u One checks that A ε
. Moreover, zA = –(v, z)u ≠ 0 is in
. Hence
contains
for every z ε
. It follows that dim
n − 1 and dim
= n unless
for every z ε
. Then
is totally isotropic and dim
[n/2], (cf. the author’s Lectures in Abstract Algebra II, p. 170). This implies that n = 2, contrary to our assumption.
We shall now distinguish the following three cases: B. (x, y) is symmetric and n = 2l + 1, l 1. C. (x, y) is skew so necessar2y n is even dimensional, say, n = 2l, l > 1. D. (x,y) is symmetric, n = 2l, > 1. Moreover, we shall assume that in the symmetric cases (B and D) the bilinear form has maximal Witt index. This means that
contains a totally isotropic subspace
of l dimensions.
B. Let (u1, u2, …, un) be a basis for and let (ui, uj) = σij, s = (σij). A linear transformation A ε
if and only if (uiA, uj) = −(ui,ujA) for i,j = l, …,n. If uiA = ∑kαik,ujk then these conditions are that ∑kαik,σkj = –∑kαik,σjkor in matrix form,
Since the form is of maximal Witt index the basis can be chosen so that
where 1l denotes the identity matrix of l rows and columns and σ is a non-zero element in (cf. the author [2], vol. II, p. 168). Since nothing is changed by reβlacing the form by a non-zero multiple of the form we may assume that
If we partition a in the same way as s:
where aij ε Фl, u1 u2 are l × 1 matrices and v1,v2 are 1 × l matrices, then a simple computation shows that (60) holds if and only if
These conditions imply that the following set of elements is a basis for identified with the algebra of matrices satisfying (60)
where i, j = 1, …, l. The set is an abelian subalgebra of
. The linear forms which are subscripts can be identified with the linear mappings h → α(h), where h = ∑ ωihi and we have [eαh] = αeα, α = ωi − ωj, ωi + ωj etc. It follows that
is a splitting Cartan subalgebra and the α are the non-zero roots.
acts irreducibly on
and if z = h0 + Σραeα is a center element, then [zh] = 0 gives Σ ραα(h)eα = 0. Since the eα are linearly independent this implies that ραα(h) = 0 for all h. Since α ≠ we have ρα = 0. Hence z = h0. But then [zeα] = 0 gives α(h0) = 0 for all α. Since there are l linearly indeβendent α (e.g. the ωi, l = 1, …) we have h0 = 0. Hence the center is 0 and
is semi-simple and split.
We now assume l 12 and we set
One checks that this is a simple system of roots with Dynkin diagram Bi:
Hence is simple of type Bi. We therefore have the following
THEOREM 7. Let be the Lie algebra of linear transformations in a (2l + 1)-dimensional space, l
2, which are skew relative to a nondegenerate symmetric linear form, of maximal Witt index. Then
is a split simple Lie algebra of type Bi.
C. Let be of dimension 2l, l
1, (x, y) non-degenerate skew linear form in
. Let
be the Lie algebra of linear transformations which are skew relative to (x, y). We can choose a basis (u1, u2, …, u2l) so that the matrix q = ((ui, uj)) is
As in B, one sees that can be identified with the Lie algebra of matrices α ε
2l such that aq = −qa′. This implies that
where
Hence has the basis
where i, j = 1, 2, …,l. As in B, one proves that = {∑ ωi hi} is a Cartan subalgebra and the subscripts in (68) define the roots relative to
. Also
has center 0 and so, by Lemma 2,
is semi-simple. The roots
form a simple system with Dynkin diagram Cl if l 3. This proves
THEOREM 8. Let be the Lie algebra of linear transformations in a 2l-dimensional space, l
3, which are skew relative to a nondegenerate skew linear form (the symplectic Lie algebra.) Then
is a split simple Lie algebra of type Cl.
D. , 2l-dimensional, l
2, (x, y) symmetric of maximal Witt index. Here we can choose the basis (ui) so that t = ((ui, uj)) has the form
and can be identified with the Lie algebra of matrices a satisfying at = −ta′. These are the matrices
such that
Then has the basis
where i,j = 1, 2, …,l. = {Σ ωihi} is a splitting Cartan subalgebra and the subscripts in (73) define the roots. The center of
is 0; hence
is semi-simple. Set
Then these αi form a simple system of roots which has the Dynkin diagram Dl if l 4. We therefore have
THEOREM 9. Let be the Lie algebra of linear transformations in a 2l-dimensional space, l
2, which are skew symmetric relative to a non-degenerate symmetric linear form of maximal Witt index. Then if l
4,
is a split simple Lie algebra of type Dl.
The four classes of Lie algebras α, B, C and D are called the “great” classes of simple Lie algebras. These correspond to the linear groups which Weyl has called the classical groups in his book with this title. It is easy to see directly, or from the bases, that we have the following table of dimensionalities
The determination of the simple systems and the general isomorphism theorem (Theorem 3) and the criterion for simplicity (Theorem 4) yield a number of isomorphisms for the low dimensional orthogonal and symplectic Lie algebras. The verifications are left to the reader. These are
1. The orthogonal Lie algebra in 3-space (three-dimensional space ) defined by a form of maximal Witt index and the symplectic Lie algebra in 2-space are split three-dimensional simple and so are isomorphic to the algebra of matrices of trace 0 in 2-space.
2. The orthogonal Lie algebra of a form of maximal Witt index in 4-space is a direct sum of two ideals isomorphic to split threedimensional simple Lie algebras.
3. The symplectic Lie algebra in 4-space is isomorphic to the orthogonal Lie algebra in 5-space defined by a symmetric form of maximal Witt index.
4. The orthogonal Lie algebra in 6-space of a form of maximal Witt index is isomorphic to the Lie algebra of linear transformations of trace 0 in 4-space.
The remaining split simple Lie algebras: types G2, F4, E6, E7, and E8, are called exceptional. We shall give irreducible reβresentations for G2, F4, E8 but we shall be content to state the results without proofs, even though some of these are not trivial. A complete discussion can be found in a forthcoming article by the author ([11]).
Our realizations of G2 and G4 whill be as the derivation algebras of certain non-associative algebras, namely, an algebra of Cayley numbers and an exceβtional simple Jordan algebra
.
Following Zorn, the definition of the split Cayley algebra or vector-matrix algebra is as follows. Let V be the three-dimensional vector algebra over
. Thus V has basis i, j, k over
and has bilinear scalar multiplication and skew symmetric vector multiplication × satisfying: i,j,k are orthogonal unit vectors and
Let be the set of 2 × 2 matrices of the form
Addition and multiplication by elements of are as usual, so that
is eight-dimensional. We define an algebra product in
by
The split Cayley algebra is defined to be together with the vector space operations defined before and the multiplication of (78).
is not associative but satisfies a weakening of the associative law called the alternative law:
Let be the Lie algebra of derivations in
. The unit matrix 1 is the identity in
and since 12 = 1, 1D = 0 for every derivation D. The space
0 of elements of trace 0 (α + β. = 0) coincides with the space spanned by the commutators [xy] = xy − yx, x, y ε
. Hence
0D
0 for a derivation. Thus
0 is a seven-dimensional subspace of
which is invariant under
The reβresentation in
0 is faithful and irreducible.
If T is a linear transformation of trace 0 in V, and T* is its adjoint relative to the scalar multiplication, then it can be verified that
is a derivation in . The set of these derivations is a subalgebra
0 isomorphic to
In any alternative algebra any mapping of the form Dα, b = [aLbL] + [aLbR] + [aRbR], where α, b are in the algebra and aL, aR denote the left and the right multiplications (x → ax, x → xa) determined by α, is a derivation. In
any derivation has the form De1.a12 + De2.b21 + D0 where
and D0 ε 0. If
is a Cartan subalgebra of
0,
is a Cartan subalgebra of
. If we identify
0 with
, we can take
to be the set of matrices of the form ω1h1 + ω2h2, h1 = e11 − e33, h2 = e22 − e33. Then
is a splitting Cartan subalgebra of
and the roots of
in
are:
The center of
is 0 and since
acts irreducibly in
0,
is semisimple. The roots α1 = ω1 − ω2, α2 = ω2 form a simple system with Dynkin diagram G2. Hence
is split simple of type G2.
The Cayley algebra has an anti-automorphism
of period two such that
. Let
denote the space of 3 × 3 hermitian Cayley matrices defined by this anti-automorphism. Thus
is the set of matrices of the form
If X, Y ε , then
where the product XY is the usual matrix product. Then is a non-associative algebra relative to the usual vector space operations and the multiplication (83). Moreover, the multiplication in
satisfies
These identities are the defining properties of a class of algebras called Jordan algebras. A consequence of these identities is that if Rα denotes the mapping X → X·A = A·X then
is a derivation in the algebrα.
Let denote the Lie algebra of derivations in
and let
0 denote the subalgebra of those derivations which map the elements
into 0. Then 0 is isomorphic to an orthogonal Lie algebra of a symmetric form of maximal Witt index in 8-space. Let
be a Cartan subalgebra of
0 corresponding to the Cartan subalgebra selected above for the orthogonal Lie algebra (type D4). Then
is a splitting Cartan subalgebra of
.
The derivation algebra maps the 26-dimensional space J0 of matrices of of trace 0 into itself. The representation in J0 is faithful and irreducible. The center of
is 0, so
is semi-simple. The roots of
in
are
form a simple system with the diagram of type F4. Hence
is split simple of type F4.
We now use the notation for the derivation algebra of
and we let
denote the set of linear transformations in
which have the form
This is clearly a subspace of the space of linear transformations in Moreover, if tr B = 0 and E ε
, then
Since these relations imply that
is a Lie algebra of linear transformations. Now
acts irreducibly in
and has 0 center. Hence
is semi-simple. Let
denote the Cartan subalgebra of
defined above and let
. Then
is a Cartan subalgebra of
with roots
where Ai and Mi are as in (87). The roots
form a simple system of type E6. Hence is simple of type E6.
An enumeration of the roots gives the following table of dimensionalities:
We state also (without proof) that E7 and E8 exist and have dimensionalities 133 and 248 respectively.
Conclusion. Every split simple Lie algebra over an arbitrary field of characteristic 0 is isomorphic to one of the Lie algebras Al, l 1, Bl, l
2, Cl, l
3, Dl, l
4 constructed above or to an exceβtional Lie algebra G2, F4, E6, E7 or E8. We shall see later (Chapter IX) that the algebras listed here are not isomorphic. Hence the results give a complete classification of split simple Lie algebras over any field of characteristic 0. In particular, we have a complete classification of simple Lie algebras over any algebraically closed field of characteristic 0.
One of the most fruitful and profound ideas in the theory of Lie groups is the method introduced by Weyl for transferring problems on representations of Lie groups and algebras to the case of compact Lie groups. Weyl gave this method the striking title of the “unitary trick” and he used it to give the first proof of complete reducibility of the representations of semi-simple Lie algebras over algebraically closed fields of characteristic 0. He used it also for studying the irreducible representations, determining the characters and dimensionalities of these representations. We shall consider this later on. Weyl’s method permits the application of analysis−via the theory of compact groups−to Lie algebras. The essence of the method has been formalized by Chevalley and Eilenberg in the following form.
A property P of Lie algebras is called a linear property if (1) P holds for implies that P holds for
Ω, Ω any extension field of the base field of L and (2) P holds for
Ω for some extension field Ω implies that P holds for
. A trivial example is the statement that dim
= n. Cartan’s criterion implies that the property of semi-simplicity is a linear property for Lie algebras of characteristic 0. Also it can be proved that the property that the finite-dimensional representations of such algebras are completely reducible is a linear property.
We now introduce a certain class of semi-simple Lie algebras over the field R of real numbers and we shall see that the validity of a linear property P for all of these algebras implies the validity of P for all semi-simple Lie algebras of characteristic zero. The class we require is given in the following
DEFINITION 3. A Lie algebra over the field R of real numbers is called compact if its K2ling form is negative definite. (This implies semi-simplicity.)
The importance of compact Lie algebras is that the groups associated with these algebras in the Lie theory are compact groups. The following result is an immediate consequence of the classification-due to Cartan−of simple Lie algebras over the field of real numbers. The proof we shall give, which does not make use of this structure theory, is due to Weyl.
THEOREM 10. Let be a semi-simple Lie algebra over the field C of complex numbers. Then there exists a compact Lie algebra
u over the field of real numbers such that the “complexification”
.
Proof: Since C is algebraically closed our structure theory applies. Let ei, fi, hi be canonical generators for and let the canonical basis be chosen as in §4. We have seen that there exists an automorphism σ of
such that
. Then
and since the ei fi generate
, σ2 = 1. If we recall the form of the canonical basis determined by the ei, fi, hi it is clear that
for every non-zero root α. Hence if we choose any eα ≠ 0 in
α then
. If we replace
we obtain
and, since we are working in an algebraically closed field, we may choose λα so that
. Hence we may suppose that
for the positive roots α. If we choose
for the positive roots α (relative to some ordering in
. We now choose as basis for
: (h1, h2, …, hl, eα, e−α, …) where
and these elements satisfy
for every non-zero root α. If α and β are non-zero roots and β ≠ ± α then
if α + β is a root and [eαeβ] = 0 otherwise. If we apply σ to this relation we obtain
We know also that (eα, e−α) = − 1 implies that [eαe−α] = hα and this element is a rational linear combination of the hi. Now if α + β is also a root, then
On the other hand,
by (20). This can be continued to
Since α and β are linearly indeβendent, hα and ββ are linearly indeβendent. Hence the last relation can be compared with (95) to see that is a positive rational number. It follows that all the Nαβ, α, β roots, are real numbers. Since [eαhl = α(h)eα and α(hi) is rational it is now clear that the multiplication table for the basis we have chosen has real coefficients. Hence the set of real linear combinations of this basis is an algebra
1 over the real field R such that
1c =
. Since
induces an automorphism of period two in
1. We now modify this automorphism by intertwining it with the standard automorphism
of the complex field. Thus if we denote the chosen basis by (ui) we let τ be the mapping:
This is a semi-automorphism of
in the sense that τ is a semi- linear transformation and [xy] τ = [xτyr], In fact, we have (x +yτ) = xτ + yτ,
. The fact that τ satisfies [xy]τ = [xτyτ] is clear from the reality of the multiplication table of the ui. It is clear also that τ is an automorphism of
considered as an algebra over R; hence the set
u of fixed points under τ is an R- subalgebra of
. We shall show that
u is the required compact form of
. It is clear that τ2 = 1. Hence
. follows that every element of
u is a real linear combination of the elements
. If we use the form of the
we see that every element of
u is a real linear combination of the elements
where α ranges over the positive roots. These elements form a basis for over C and this implies that
. Also we see that any element of
u has the form
where the ξj are real, ρα complex, the conjugate. Since
=
, the Killing form for
u is the restriction to
u of the Killing form in
. Hence we may use the orthogonality properties and the relations (eα, e−α) = − 1 to calculate
where h = ∑ ξihi. Since (h, h) > 0 unless h = 0, it is clear that (x, x) < 0 unless x = 0. Hence the Killing form is negative definite.
Now let P be a linear property of Lie algebras which is valid for every compact Lie algebra over the field R of real numbers. Let be a semi-simple Lie algebra over a field
of characteristic 0 and let Ω be the algebraic closure of
. Then
Ω is split and so this has the form
0Ω
where
0 is a Lie algebra over the field Q of rational numbers. Then
0c is a semi-simple Lie algebra over the complex field C; hence, by Theorem 10, there exists a compact Lie algebra
u over the reals such that
. Since ρ is satisfied by
u, it holds for
hence for
0 and for
. It follows that P holds for
.
In these exercises all base fields are of characteristic 0 and all algebras and modules are finite-dimensional. Unexplained notations are as in the text.
1. Show that is a Cartan subalgebra of a semi-simple Lie algebra
if and only if: (1)
is maximal commutative and (2)
is reductive in
, that is, ad
is completely reducible.
2. Let be a Cartan subalgebra of the semi-simple Lie algebra
. Show that if R is any finite-dimensional reβresentation of
, then hR is semisimple for every h ε
.
3. Let 1 be a semi-simple subalgebra of the semi-simple Lie algebra
and let
1 be a Cartan subalgebra of
1. Show that
1 can be imbedded in a Cartan subalgebra of
.
4. Use the method of § 3 to obtain a canonical basis and multiplication table for A2 and G2.
5. Use the method of § 3 to obtain the roots for the split Lie algebra of type E7 (assuming it exists) and show in this way that the dimensionality of this Lie algebra is 133.
6. Without using the determination of the connected Dynkin diagrams show that cannot be the Dynkin diagram of a Cartan matrix. (Hint: Show that the set of roots determined by this contains a positive root β such that 2β is a root.)
7. Show that the constants of multiplication of a canonical basis (§ 3) of a split semi-simple Lie algebra are rational numbers with denominators of the form 2k,3l, k, l integers.
8. Prove that any split semi-simple Lie algebra can be generated by two elements.
9. Let , the algebra of linear transformations in an (l + 1)-dimensional vector space and let (A, B) be the Killing form in
. Show that
(Hint: Use Exercise 3.9)
10. Express the Killing forms of the Lie algebras of types Bi, Ci and Di in terms of the trace form tr AB where the reβresentation is the one given in §6.
11. Show that if is split,
a splitting Cartan subalgebrα, then
is a maximal solvable subalgebra and
, is a maximal n2potent subalgebra of
.
12. Determine the Weyl group for the split Lie algebra of type G2 and the split Lie algebra of type Al.
13. Let αl be a simple root and let α be a positive root ≠ αi. Show that every root of the form α + kai, k an integer, is positive. Show that where the summation is taken over the k such that α + kai is a root and α is any positive root. Hence prove that
.