実直苦闘 ジッチョククトウ、ケイジンゴクエツ

Root System And Its Classification

"You are researching platypus."

数学

In this post, we are going to study the structure of the thing so called root system, and give a classification of it. The root system is a powerful tool to classify complex semisimple Lie algebras, but can be abstracted to a more general version.

Definition, Examples and Irreducibility

We suppose ERrE\cong \mathbb{R}^r, and (,)(\cdot,\cdot) is a positive definite inner product on it. A finite subset RE{0}R\subset E\setminus\{0\} is called an abstract root system, if

  1. (Root.A) EE is spanned by RR;
  2. (Root.B) α,βR\forall\alpha,\beta\in R, nαβ=2(α,β)(α,α)n_{\alpha\beta}=2\frac{(\alpha,\beta)}{(\alpha,\alpha)} is an integer;
  3. (Root.C) If α,βR\alpha,\beta\in R, then sβ(α)=αnβαβRs_\beta(\alpha)=\alpha-n_{\beta\alpha}\beta\in R.

Its elements are called roots, and the rank of E=Span(R)E=\operatorname{Span}(R) is also said the rank of RR. A thing shall be remarked is that the symbol nαβn_{\alpha\beta} may be written as β,α\langle\beta,\alpha\rangle.

The condition 2 basically means the projection relationship between the roots are limited, which will be shown later. The condition 3 gives the reflection of α\alpha as the following image shows. Especially, if you take α=β\alpha=\beta, this means αR-\alpha\in R, which overall implies R=RR=-R.

reflection

We may operatorize nαβn_{\alpha\beta} by letting α(x)=2(α,x)(α,α)\alpha^\vee(x)=\frac{2(\alpha,x)}{(\alpha,\alpha)}. Hence nαβn_{\alpha\beta} now can be written as α(β)\alpha^\vee(\beta). We call this functional, by the view of duality, the coroot of the root α\alpha.

There are of course infinitely many root systems. However, we can add a restriction to make it a little bit simpler. A root system is called reduced, if there is no constant ratio between two roots, which formally means α,cαRc=±1\alpha,c\alpha\in R\Rightarrow c=\pm1. Later, all the systems we are talking about are assumed to be reduced. (Actually, in some textbooks like Erdmann and Wildon’s Introduction to Lie Algebras, this is directly set as an axiom of root systems.)

Two roots determine a root system of rank 22, therefore, it’s important to classify the root systems of rank 22 at first.

For two different roots, let θ\theta represent the angle between α\alpha and β\beta. Then we have (α,β)=2αβcosθ(\alpha,\beta)=2\|\alpha\|\|\beta\|\cos\theta by Ancient Egyptian mathematics. Hence nαβ=2αβcosθn_{\alpha\beta}=2\frac{\|\alpha\|}{\|\beta\|}\cos\theta, nαβnβα=4cos2θ[0,4]n_{\alpha\beta}n_{\beta\alpha}=4\cos^2\theta\in[0,4] and finally only 0,1,2,30,1,2,3. By analysis of each case, we have all the possibilities of the two roots:

  1. (A1) θ=π/2\theta=\pi/2, nαβ=nβα=0n_{\alpha\beta}=n_{\beta\alpha}=0;
  2. (A2) θ=π/3\theta=\pi/3, nαβ=nβα=1n_{\alpha\beta}=n_{\beta\alpha}=1, α2=β2\|\alpha\|^2=\|\beta\|^2;
  3. (A2’) θ=2π/3\theta=2\pi/3, nαβ=nβα=1n_{\alpha\beta}=n_{\beta\alpha}=-1, α2=β2\|\alpha\|^2=\|\beta\|^2;
  4. (B2) θ=π/4\theta=\pi/4, nαβ=1n_{\alpha\beta}=1, nβα=2n_{\beta\alpha}=2, α2=2β2\|\alpha\|^2=2\|\beta\|^2;
  5. (B2’) θ=3π/4\theta=3\pi/4, nαβ=1n_{\alpha\beta}=-1, nβα=2n_{\beta\alpha}=-2, α2=2β2\|\alpha\|^2=2\|\beta\|^2;
  6. (G2) θ=π/6\theta=\pi/6, nαβ=1n_{\alpha\beta}=1, nβα=3n_{\beta\alpha}=3, α2=3β2\|\alpha\|^2=3\|\beta\|^2;
  7. (G2’) θ=5π/6\theta=5\pi/6, nαβ=1n_{\alpha\beta}=-1, nβα=3n_{\beta\alpha}=-3, α2=3β2\|\alpha\|^2=3\|\beta\|^2.

The label of these situations, are actually from the classification of root systems. By applying the definition of root systems, we will get the following picture of root systems of rank 22.

root systems of rank 2

One should realize that the first picture is actually not A1. The root system of A1 is only one axis, which is the unique root system of dimension 1. However, when we study A2, we will easily find out that it can not be written as a union of two 1 dimensional root systems because of the restriction of rule (Root.C).

This gives us the idea of irreducible root system. More formally, we say a root system RR is irreducible if it cannot be expressed as a disjoint union of two non-empty subsets R1R2R_1\sqcup R_2 where R1R_1 and R2R_2 are two root systems and for all αR1\alpha\in R_1 and βR2\beta\in R_2, we have (α,β)=0(\alpha,\beta)=0 (By this, we say R1R_1 and R2R_2 are orthogonal, denotes R1R2R_1\perp R_2).

For each root system, it can be written as a union of irreducible systems as

R=R1R2Rk.R=R_1\sqcup R_2\sqcup\cdots\sqcup R_k.

And the decomposition is unique modulo ordering. To see this, we define a relation between roots. For two roots α,βR\alpha,\beta\in R, we say αβ\alpha\sim\beta if there is a sequence

α=γ0,γ1,,γn=β\alpha=\gamma_0,\gamma_1,\cdots,\gamma_n=\beta

where for each i<ni<n, we have γi⊥̸γi+1\gamma_i\not\perp\gamma_{i+1}. It’s easy to check that this relation is an equivalence relation. Each equivalence class forms an irreducible root system, and this is a decomposition of RR. For arbitrary irreducible root system in arbitrary decomposition of RR, we can find it in the decomposition we constructed, by applying the relation we defined to some roots, and vice versa. Therefore, such decomposition is unique.

The irreducible root systems are the simplest ones, and it is actually what is going to be classified. Each root system is corresponding to a complex simple Lie algebra, we will show this in another article.

Polarization, Basis and Recovery

From now on, we suppose RR is an irreducible root system.

Before classifying, we still need some preparations, which is used to simplify the irreducible root systems further. Firstly, we know that R=RR=-R, therefore we can study a half part of the root system. Formally, we choose arbitrary hyperplane with codimension 11 which doesn’t contain any root, This separates the root system into two parts. We can choose one side of it and say it is positive, while the other part of it is called negative. This is called a polarization of the root system.

Formally, we find a vector tt which is not orthogonal to any root, and calculate (t,α)(t,\alpha) for arbitrary root α\alpha. If (t,α)>0(t,\alpha)>0, we say α\alpha is positive, otherwise say negative. By Riesz-Fréchet representation theorem, this is equivalent to choose a tEt\in E^* and calculate t(α)t(\alpha). The positive roots are denoted by R+R_+ and negative ones by RR_-.

Once a polarization of a root system is chosen, we can find simple roots on it. A positive root is called simple, if it can not be written as a sum of two other positive roots. We might see that every positive root can now be written as sum of simple roots. Given a positive root which is not simple, we have α=β+γ\alpha=\beta+\gamma, where β,γR+\beta,\gamma\in R_+. Since tEt\in E^*, we have 0<t(β),t(γ)<t(α)0<t(\beta),t(\gamma)<t(\alpha). We know that tt has only finite values on R+R_+, therefore such process will terminate somewhere, and the decomposition is found.

By classification of root systems of rank 22, it’s easy to find out that α+βR\alpha+\beta\in R if (α,β)<0(\alpha,\beta)<0. Therefore for two simple roots α,βR+\alpha,\beta\in R_+, we have (α,β)<0(\alpha,\beta)<0, since otherwise we have (α,β)<0(-\alpha,\beta)<0 so γ=βα\gamma=\beta-\alpha is a root then make either β\beta or α\alpha not simple.

It’s obvious that simple roots span the space EE. Actually, they forms a basis of EE. Assume converse condition, then we can obtain a nontrivial relation that

iIcivi=jJcjvj\sum_{i\in I}c_iv_i=\sum_{j\in J}c_jv_j

where I,JI,J are disjoint and ci>0c_i>0. However,

0<iIcivi2=(iIcivi,iIcivi)=(iIcivi,jJcjvj)0,0<\|\sum_{i\in I}c_iv_i\|^2=(\sum_{i\in I}c_iv_i,\sum_{i\in I}c_iv_i)=(\sum_{i\in I}c_iv_i,\sum_{j\in J}c_jv_j)\le0,

which is a contradiction. Hence, the simple roots are always a basis of EE.

Actually, this leads to the following concept, which is the axiomatic version of simple roots and arranged as an alternative in some textbooks. A subset BRB\subset R is called a base for RR if

  1. (Base.A) BB is a basis of EE.
  2. (Base.B) every βR\beta\in R can be written as β=αBkαα\beta=\sum_{\alpha\in B}k_\alpha \alpha with kαZk_\alpha\in\Z, and kαk_\alpha have the same sign.

The existence of it is shown by constructing simple roots. The corresponding definition of positive roots can also be given after this proposition. A root is said to be positive if the coefficients in decomposition of (Base.B) are not negative.

The classification of irreducible root systems, are actually classifying bases of it. To show this, we left to show the unique rebuild of irreducible root systems.

Intuitively, we shall construct a root system by repeatedly acting sαs_\alpha on the bases. We can see sαs_\alpha are permutations on RR. Therefore, it generates a group W=sααRW=\langle s_\alpha\rangle_{\alpha\in R} on RR, and we call it the Weyl group. Naturally, we can define a subgroup of it as WB=sγγBW_B=\langle s_\gamma\rangle_{\gamma\in B}. To obtain the rebuild, we only need to prove WB(B)=RW_B(B)=R. More precisely, we need to prove that for each βR\beta\in R, there is some αB\alpha\in B and gWBg\in W_B such that β=gα\beta=g\alpha.

The following is the proof.

We only need to consider R+BR^+\supset B. Let βR+\beta\in R^+, we have the decomposition β=γBkγγ\beta=\sum_{\gamma\in B}k_\gamma\gamma, where Zkγ0\Z\ni k_\gamma \ge 0. The following concept is important in the proof which gives an index to use induction on it.

ht(β):=γBkγ.\operatorname{ht}(\beta):=\sum_{\gamma\in B}k_\gamma.

This definition seems to be summoned from void, but is actually unavoidable. I don’t like this proof and I want to find a more general and more natural version, but has not been established.

We start from the situation when ht(β)=1\operatorname{ht}(\beta)=1, then βB\beta\in B, we only take α=β\alpha=\beta and gg is the identity map.

Now we suppose the proposition holds for all β\beta that ht(β)<n\operatorname{ht}(\beta)<n. Then for ht(β)=n2\operatorname{ht}(\beta)=n\ge2, at least two of kγk_\gamma are strictly positive. There will always be some γ0B\gamma_0\in B such that (β,γ0)>0(\beta,\gamma_0)>0. If unfortunately otherwise, we have

(β,β)=(β,γBkγγ)=γkγ(β,γ)0,(\beta,\beta)=(\beta,\sum_{\gamma\in B}k_\gamma\gamma)=\sum_{\gamma}k_\gamma(\beta,\gamma)\le0,

which contradicts to the positive definiteness of β\beta. Now we consider sγ0(β)s_{\gamma_0}(\beta). Since βγ0\beta\not=\gamma_0, there is some αB\alpha\in B such that kα0k_\alpha\not=0 and αγ0\alpha\not=\gamma_0. Now by sγ0(β)=βnβαγ0s_{\gamma_0}(\beta)=\beta-n_{\beta\alpha}\gamma_0, we shall see the coefficient of α\alpha in the decomposition of sγ0(β)s_{\gamma_0}(\beta) is also kαk_\alpha, then sγ0(β)R+.s_{\gamma_0}(\beta)\in R^+. Now,

ht(sγ0(β))=ht(β)nβγ0<ht(β)=n.\operatorname{ht}(s_{\gamma_0}(\beta))=\operatorname{ht}(\beta)-n_{\beta\gamma_0}<\operatorname{ht}(\beta)=n.

Hence, we can find αB\alpha\in B and hWBh\in W_B making sγ0(β)=hαs_{\gamma_0}(\beta)=h\alpha, then β=(sγ0h)α\beta=(s_{\gamma_0}h)\alpha.

Now we proved.

Dynkin Diagram and Restrictions

We now need to classify bases of simple root systems. However, they are still too abstract for us to see, although they are simple enough. A direct idea of realizing the classification, is find some invariants of bases, and here it is.

Given a root system (α1,,αp)(\alpha_1,\cdots,\alpha_p), the Cartan matrix of it is written as (nij)1i,jp(n_{ij})_{1\le i,j\le p}. We shall see this is a perfect encoding of bases.

Another encoding of bases is a diagram. We draw roots as vertices, and draw dαβ:=nαβnβα=:dβαd_{\alpha\beta}:=n_{\alpha\beta}n_{\beta\alpha}=:d_{\beta\alpha} lines between vertices α\alpha and β\beta. Finally, we draw an arrow on edge from α\alpha to β\beta if α>β\|\alpha\|>\|\beta\|. By the construction of irreducible root systems, we shall see that every Dynkin diagram of the bases is a connected graph.

Dynkin diagram: simplification

Now, by classifying Dynkin diagrams, we can give a classification of bases. To make it simpler, we can ignore the arrows on it, by normalizing the roots.

Given a base AA with dimension nn, we consider the set

Aˉ={αˉ:=α(α,α);αA}.\bar A=\{\bar \alpha:=\frac{\alpha}{\sqrt{(\alpha,\alpha)}};\alpha\in A\}.

Then we shall see for each α,βAˉ\alpha,\beta\in \bar A,

dαβ:=4(αˉ,βˉ)2=4(α(α,α),β(β,β))2=4(1(α,α)(β,β))2(α,β)2=4(α,α)(β,β)(α,β)2=2(α,β)(α,α)2(α,β)(β,β)=nαβnβα.\begin{aligned} d_{\alpha\beta}:=4(\bar\alpha,\bar\beta)^2 &=4(\frac{\alpha}{\sqrt{(\alpha,\alpha)}},\frac{\beta}{\sqrt{(\beta,\beta)}})^2\\ &=4(\frac{1}{\sqrt{(\alpha,\alpha)}\sqrt{(\beta,\beta)}})^2(\alpha,\beta)^2\\ &=\frac{4}{(\alpha,\alpha)(\beta,\beta)}(\alpha,\beta)^2\\ &=\frac{2(\alpha,\beta)}{(\alpha,\alpha)}\frac{2(\alpha,\beta)}{(\beta,\beta)}=n_{\alpha\beta}n_{\beta\alpha}. \end{aligned}

Then the Dynkin diagram of Aˉ\bar A is actually a graph forgetting arrows on the Dynkin diagram of AA. And the vertices are corresponding.

We can give an axiomatic description of such sets. A set Aˉ\bar A is said to be admissible if

  1. (vi,vi)=1(v_i,v_i)=1 for all ii and (vi,vj)0(v_i,v_j)\le 0 if iji\not= j;
  2. If iji\not=j, then 4(vi,vj)2{0,1,2,3}4(v_i,v_j)^2\in\{0,1,2,3\}.

Later, a lot of lemmas will be derived only from these two conditions. We use the notation Aˉ\bar A to denote the simplified set a base AA of a root system, which is admissible and has Aˉ=A=n|\bar A|=|A|=n.

Dynkin diagram: admissibility

In this section, all we suppose is that the set is admissible. We now study the restrictions an admissible set can give.

Firstly, the numbers of pairs of vertices joined by at least one edge is at most n1n-1. (Res.A) To prove this, we consider v=v1++vnv=v_1+\cdots+v_n, then

0<(v,v)=(v1+v2++vn,v1+v2++vn)=(v1,v1+v2++vn)++(vn,v1+v2++vn)=(v1,v1)+(v1,v2)++(v1,vn)+(v2,v1)+(v2,v2)++(v2,vn)++(vn,v1)+(vn,v2)++(vn,vn)=n+2i<j(vi,vj).\begin{aligned} 0<(v,v)&=(v_1+v_2+\cdots+v_n,v_1+v_2+\cdots+v_n)\\ &=(v_1,v_1+v_2+\cdots+v_n)+\cdots+(v_n,v_1+v_2+\cdots+v_n)\\ &=(v_1,v_1)+(v_1,v_2)+\cdots+(v_1,v_n)\\ &+(v_2,v_1)+(v_2,v_2)+\cdots+(v_2,v_n)\\ &+\cdots\\ &+(v_n,v_1)+(v_n,v_2)+\cdots+(v_n,v_n)\\ &=n+2\sum_{i<j}(v_i,v_j). \end{aligned}

Therefore, n>2i<j(vi,vj)=i<jdijn>-2\sum_{i<j}(v_i,v_j)=\sum_{i<j}\sqrt{d_{ij}}.

Now, the number of pairs of vertices {vi,vj}\{v_i,v_j\} such that dij1d_{ij}\ge1 is less or equal to i<jdij\sum_{i<j}\sqrt{d_{ij}}, which is smaller than nn. Hence the claim proved.

By this, we can easily see that there will be no cycle in a diagram. (Res.B)

What’s more, no vertex of Γ\Gamma is incident to four or more edges. (Res.C) We take vv is an arbitrary vertex and there are kk vertices v1,,vkv_1,\cdots,v_k incident to it. Since there is no cycles, we have (vi,vj)=0(v_i,v_j)=0 for 0i<jk0\le i<j\le k. Now, consider the subspace Span(v,v1,,vk)=:U\operatorname{Span}(v,v_1,\cdots,v_k)=:U. Then there must be an orthogonal basis of UU like v0,v1,,vkv_0,v_1,\cdots,v_k. Where v0=0\|v_0\|=0 and (v,v0)0(v,v_0)\not=0.

The projection decomposition is v=0ik(v,vi)viv=\sum_{0\le i\le k}(v,v_i)v_i. Then we have

1=(v,v)=(0ik(v,vi)vi,0jk(v,vj)vj)=0ik0jk(v,vi)(v,vj)(vi,vj)=0ik(v,vi)2.\begin{aligned} 1=(v,v)&=(\sum_{0\le i\le k}(v,v_i)v_i,\sum_{0\le j\le k}(v,v_j)v_j)\\ &=\sum_{0\le i\le k}\sum_{0\le j\le k}(v,v_i)(v,v_j)(v_i,v_j)\\ &=\sum_{0\le i\le k}(v,v_i)^2. \end{aligned}

Hence 1ik(v,vi)2<1\sum_{1\le i\le k}(v,v_i)^2<1. By the previous knowledge of bases, 4(v,vi)2>14(v,v_i)^2>1, then (v,vi)2>14(v,v_i)^2>\frac{1}{4}. It’s so obvious now that kk should be no more than 33.

Then obviously we know that, if a graph has a triple edge, then it will be only two vertices like below. (Res.D)

triple edge

An interesting and useful lemma can be given here, which is called the Shrinking Lemma. Given a subset AˉAˉ\bar A'\subset \bar A, and the graph Γ\Gamma' of Aˉ\bar A' is a subgraph of Γ\Gamma, where Γ\Gamma is the graph of AA. The lemma claims that, if Γ\Gamma' is a line, then (AˉAˉ){vAv}(\bar A\setminus \bar A')\cup\{\sum_{v\in A'}v\} is still an admissible set. That is to say, intuitively, we can shrink a line in a graph to a point.

We only have to verify the conditions of admissible sets. We enumerate Aˉ\bar A' as Aˉ={vi}i=1k\bar A'=\{v_i\}_{i=1}^k and define u=i=1kviu=\sum_{i=1}^kv_i.

(u,u)=(i=1kvi,i=1kvi)=k+2i=1k1(vi,vi+1)=k(k1)=1.\begin{aligned} (u,u)&=(\sum_{i=1}^kv_i,\sum_{i=1}^kv_i)\\&=k+2\sum_{i=1}^{k-1}(v_i,v_{i+1})=k-(k-1)=1. \end{aligned}

Choose some vAˉAˉv\in \bar A\setminus\bar A', since vv can only be incident to at most one viv_i, then either (v,u)=0(v,u)=0 or (v,vi)<0(v,v_i)<0. The same thing holds for 4(v,u)24(v,u)^2. Now we proved.

A simple conclusion (Res.E) can now be given that a graph Γ\Gamma has

  1. no more than one double edge;
  2. no more than one branch vertex;
  3. not both a double edge and a branch vertex.

A vertex is said to be a branch vertex if, it is incident to three or more vertices, and it’s always 33 in the context of Dynkin diagrams. We only apply the Shrinking Lemma, then by (Res.C), it holds.

Now, we can study the situation of double edge and branch vertex respectively.

We know that if a graph Γ\Gamma has a double edge, then it should have the form below.

double edge?

And without loss of generality, pqp\ge q. Now, we let v=i=1piviv=\sum_{i=1}^piv_i, then

(v,v)=(i=1pivi,i=1pivi)=i=1pj=1pij(vi,vj)=i=1pi2i=1p1i(i+1)=p(p+1)(2p+1)6p(p+1)(p1)3=p(p+1)2.\begin{aligned} (v,v)&=(\sum_{i=1}^piv_i,\sum_{i=1}^piv_i)\\ &=\sum_{i=1}^p\sum_{j=1}^pij(v_i,v_j)\\ &=\sum_{i=1}^pi^2-\sum_{i=1}^{p-1}i(i+1)\\ &=\frac{p(p+1)(2p+1)}{6}-\frac{p(p+1)(p-1)}{3}\\ &=\frac{p(p+1)}{2}. \end{aligned}

Similarly, we let w=i=1qiwiw=\sum_{i=1}^qiw_i, then (w,w)=q(q+1)/2(w,w)=q(q+1)/2. Now,

(v,w)2=(pq(vp,wq))2=p2q22.(v,w)^2=(pq(v_p,w_q))^2=\frac{p^2q^2}{2}.

By the Cauchy-Schwarz inequality, (v,w)2<(v,v)(w,w)(v,w)^2<(v,v)(w,w), then 2pq<(p+1)(q+1)2pq<(p+1)(q+1), which implies pq<p+q+1pq<p+q+1. Hence,

(p1)(q1)=pqpq+1<p+q+1p1+1=2.\begin{aligned} (p-1)(q-1)&=pq-p-q+1\\ &<p+q+1-p-1+1=2. \end{aligned}

Either q=1q=1 or p=q=2p=q=2. Hence, if Γ\Gamma has a double edge, it can only be one of below. (Res.F)

double edge!

Now, we can study the situation where there is a branch vertex.

branch vertex

As you know, without loss of generality, we take pqrp\ge q\ge r.

We now let u=i=1piuiu=\sum_{i=1}^piu_i, v=i=1qiviv=\sum_{i=1}^qiv_i, and w=i=1riwiw=\sum_{i=1}^riw_i. We easily see that u,v,wu,v,w are pairwisely orthogonal. Let u^=u/u\hat u=u/\|u\|, v^=v/v\hat v=v/\|v\| and w^=w/w\hat w=w/\|w\|. By the Gram–Schmidt process, we have an orthogonal basis of Span(u^,v^,w^,z)\operatorname{Span}(\hat u,\hat v,\hat w,z) as {u^,v^,w^,z0}\{\hat u,\hat v,\hat w,z_0\} and (z,z0)>0(z,z_0)>0. Therefore we can decompose zz as

z=(z,u^)u^+(z,v^)v^+(z,w^)w^+(z,z0)z0,z=(z,\hat u)\hat u+(z,\hat v)\hat v+(z,\hat w)\hat w+(z,z_0)z_0,

So,

(z,u^)u^+(z,v^)v^+(z,w^)w^=z(z,z0)z0(z,u^)2+(z,v^)2+(z,w^)2<11u2(z,u)2+1v2(z,v)2+1w2(z,w)2<12p24p(p+1)+2q24q(q+1)+2r24r(r+1)<1p2(p+1)+q2(q+1)+r2(r+1)<1p(q+1)(r+1)+q(p+1)(r+1)+r(p+1)(q+1)<2(p+1)(q+1)(r+1)pqr<p+q+r+2pqr+[pq+pr+qr]+[p+q+r]+1<[pq+pr+qr]+2(p+q+r)+3(p+1)(q+1)(r+1)<(p+1)(q+1)+(p+1)(r+1)+(q+1)(r+1)1p+1+1q+1+1r+1>1.\begin{aligned} &\quad(z,\hat u)\hat u+(z,\hat v)\hat v+(z,\hat w)\hat w=z-(z,z_0)z_0\\ \Rightarrow&\quad(z,\hat u)^2+(z,\hat v)^2+(z,\hat w)^2<1\\ \Leftrightarrow&\quad\frac{1}{\|u\|^2}(z,u)^2+\frac{1}{\|v\|^2}(z,v)^2+\frac{1}{\|w\|^2}(z,w)^2<1\\ \Leftrightarrow&\quad\frac{2p^2}{4p(p+1)}+\frac{2q^2}{4q(q+1)}+\frac{2r^2}{4r(r+1)}<1\\ \Leftrightarrow&\quad\frac{p}{2(p+1)}+\frac{q}{2(q+1)}+\frac{r}{2(r+1)}<1\\ \Leftrightarrow&\quad p(q+1)(r+1)+q(p+1)(r+1)+r(p+1)(q+1)\\ &\qquad<2(p+1)(q+1)(r+1)\\ \Leftrightarrow&\quad pqr<p+q+r+2\\ \Leftrightarrow&\quad pqr+[pq+pr+qr]+[p+q+r]+1\\ &\qquad<[pq+pr+qr]+2(p+q+r)+3\\ \Leftrightarrow&\quad (p+1)(q+1)(r+1)\\ &\qquad<(p+1)(q+1)+(p+1)(r+1)+(q+1)(r+1)\\ \Leftrightarrow&\quad\frac{1}{p+1}+\frac{1}{q+1}+\frac{1}{r+1}>1. \end{aligned}

A thing has to be said here is that the way I calculated this is way more complicated than it could have been. Actually, we only have to notice that

p2(p+1)=12(11p+1),\frac{p}{2(p+1)}=\frac{1}{2}(1-\frac{1}{p+1}),

then we will have

p2(p+1)+q2(q+1)+r2(r+1)<112[3(1p+1+1q+1+1r+1)]<11p+1+1q+1+1r+1>1.\begin{aligned} &\quad\frac{p}{2(p+1)}+\frac{q}{2(q+1)}+\frac{r}{2(r+1)}<1\\ \Leftrightarrow&\quad\frac{1}{2}\left[3-\left(\frac{1}{p+1}+\frac{1}{q+1}+\frac{1}{r+1}\right)\right]<1\\ \Leftrightarrow&\quad\frac{1}{p+1}+\frac{1}{q+1}+\frac{1}{r+1}>1. \end{aligned}

This is given by Claude opus. THANKS TO IT.

Since 1p+11q+11r+112\frac{1}{p+1}\le\frac{1}{q+1}\le\frac{1}{r+1}\le\frac{1}{2}, we have 1r+1>13\frac{1}{r+1}>\frac{1}{3} and hence r<2r<2, so we must have r=1r=1. Now, 1p+1+1q+1>12\frac{1}{p+1}+\frac{1}{q+1}>\frac{1}{2}, then 1q+1>14\frac{1}{q+1}>\frac{1}{4}. Hence q<3q<3.

  1. If q=2q=2, we have 1p+1>1213=16\frac{1}{p+1}>\frac{1}{2}-\frac{1}{3}=\frac{1}{6}. Hence p<5p<5.
  2. If q=1q=1, we only have 1p+1>0\frac{1}{p+1}>0. Then it seems to have no constraints.

Conclude the study of branch vertex, we have shown that either q=r=1q=r=1 or q=2q=2, r=1r=1 and p<5p<5. (Res.G)

Dynkin diagram: constraints

We have studied the restriction on admissible sets, and it is indeed restrictions on bases. We now return to the study of the Dynkin diagram of bases, which may have arrows on it.

Now consider two incident root α\alpha and β\beta, we shall see they have the same length if and only if they have a single edge. Actually, nαβnβα=2(α,β)2(α,α)(β,β)=1n_{\alpha\beta}n_{\beta\alpha}=\frac{2(\alpha,\beta)^2}{(\alpha,\alpha)(\beta,\beta)}=1, if and only if nαβ=nβα=1n_{\alpha\beta}=n_{\beta\alpha}=-1 (they are all integers and obtuse). This implies (α,α)=(β,β)(\alpha,\alpha)=(\beta,\beta). Conversely, if we have α=β\|\alpha\|=\|\beta\|, by the study of the root system of rank 22, we have nαβ=nβα=1n_{\alpha\beta}=n_{\beta\alpha}=-1.

Now, we can give the classification of Dynkin diagrams. There are only these Dynkin diagrams which are possible that the corresponding root system exists. The letters of the families are given by Cartan.

Dynkin diagrams

A thing can be remarked here, is that B1=A1B_1=A_1, C2=B2C_2=B_2 and D3=A3D_3=A_3. Therefore, we always start BnB_n from n=2n=2, CnC_n from n=3n=3 and DnD_n from n=4n=4 to make the classification without repetition.

Recovery of Bases

We have already given all the possibilities of Dynkin diagrams. But, is there really a root system exists? We still have to show the bases corresponding to all the graphs. The derivation of these results are quite complicated, and I don’t really think it’s necessary, so they will be omitted here.

In this chapter, we write {ei}i=1n\{e_i\}_{i=1}^n as the orthonormal basis of the space Rn\mathbb{R}^n.

An(n1)A_n(n\ge1)

A

Here, αi=eiei+1\alpha_i=e_i-e_{i+1}.

Bn(n2)B_n(n\ge2)

B

Here, αi=eiei+1\alpha_i=e_i-e_{i+1} for 0in10\le i\le n-1, and αn=en\alpha_n=e_n.

Cn(n3)C_n(n\ge3)

C

Here, αi=eiei+1\alpha_i=e_i-e_{i+1} for 0in10\le i\le n-1, and αn=2en\alpha_n=2e_n.

Dn(n4)D_n(n\ge4)

D

Here, αi=eiei+1\alpha_i=e_i-e_{i+1} for 0in10\le i\le n-1, and αn=en1+en\alpha_n=e_{n-1}+e_n.

E6E_6

E6

Here,

α1=12(e1+e8e2e3e4e5e6e7),α2=e1+e2,αi=ei1ei2(3i6).\begin{aligned} &\alpha_1=\frac{1}{2}(e_1+e_8-e_2-e_3-e_4-e_5-e_6-e_7),\\ &\alpha_2=e_1+e_2,\\ &\alpha_i=e_{i-1}-e_{i-2}(3\le i\le6). \end{aligned}

E7E_7

E7

Here,

α1=12(e1+e8e2e3e4e5e6e7),α2=e1+e2,αi=ei1ei2(3i7).\begin{aligned} &\alpha_1=\frac{1}{2}(e_1+e_8-e_2-e_3-e_4-e_5-e_6-e_7),\\ &\alpha_2=e_1+e_2,\\ &\alpha_i=e_{i-1}-e_{i-2}(3\le i\le7). \end{aligned}

E8E_8

E8

Here,

α1=12(e1+e8e2e3e4e5e6e7),α2=e1+e2,αi=ei1ei2(3i8).\begin{aligned} &\alpha_1=\frac{1}{2}(e_1+e_8-e_2-e_3-e_4-e_5-e_6-e_7),\\ &\alpha_2=e_1+e_2,\\ &\alpha_i=e_{i-1}-e_{i-2}(3\le i\le8). \end{aligned}

F4F_4

F4

α1=e2e3,α2=e3e4,α3=e4,α4=12(e1e2e3e4).\begin{aligned} &\alpha_1=e_2-e_3,\\ &\alpha_2=e_3-e_4,\\ &\alpha_3=e_4,\\ &\alpha_4=\frac{1}{2}(e_1-e_2-e_3-e_4). \end{aligned}

G2G_2

G2

Here, α1=e1e2\alpha_1=e_1-e_2, and α2=e2+e32e1\alpha_2=e_2+e_3-2e_1.

Wait…

Wait, is everything done?

Everything seems to be perfect. We have simplified the problem and found an object simple enough to be classified, and we have found a good way to encoding them. We have discussed all the possibilities, and shown that there actually are corresponding objects of them.

However, what if an irreducible root system has two different bases, while each of them has a Dynkin diagram respectively? If such thing happens, they will live in different classes. Although this is no biggie, we still do not want such thing happens, since such mathematics is not “beautiful”, just like the classification of Platypus.

Therefore, it remains to prove that, two bases of one root system, have the same Dynkin diagram, which means, they are geometrically equivalent.

Previous contents basically refer to [Erdmann–Wildon] and [Etingof], and the contents following is based on [Humphreys].

We know that, every base is a simple roots live in some corresponding polarization of a root system. Although a base lives in many polarizations, one polarization has only one base. Therefore, we can establish an equivalence relation between polarizations, and corresponding equivalence classes appear. Actually, if we take the perspective of geometry, all the tt define the polarizations in an equivalence class are residing in a cone, and this is exactly what we call a Weyl Chamber. Therefore, strictly speaking, each base is corresponding to a Weyl Chamber uniquely.

Weyl Chamber

Then, for sake of convenience, we define some notations. Given a root system RR, BB is a base of it. Then we use CBC_B to denote the Weyl Chamber of it. Apparently, there are many polarizations in CBC_B, then we write CB+C_B^+ for the equivalence class of positive sides, and conversely, we say CBC_B^-.

Given a root rr, we say rCB+r\in C_B^+ abusively, if and only if for any R+CB+R^+\in C_B^+, rR+r\in R^+.

Now, let’s consider two bases AA and BB of a root system RR. We then show that there always be some ωW\omega\in W makes ωA=B\omega A=B, where WW is the Weyl group of RR.

Define δA=12rCA+r\delta_A=\frac{1}{2}\sum_{r\in C_A^+}r and δB=12rCB+r\delta_B=\frac{1}{2}\sum_{r\in C_B^+}r. Now, for each αA\alpha\in A, we have

sα(δA)=12rCA+sα(r)=12(sα(α)+rCA+{α}sα(r))=12(α+rCA+{α}r)=12(rCA+r2α)=δAα.\begin{aligned} s_\alpha(\delta_A)&=\frac{1}{2}\sum_{r\in C_A^+}s_\alpha(r)\\ &=\frac{1}{2}\left(s_\alpha(\alpha)+\sum_{r\in C_A^+\setminus\{\alpha\}}s_\alpha(r)\right)\\ &=\frac{1}{2}\left(-\alpha+\sum_{r\in C_A^+\setminus\{\alpha\}}r\right)\\ &=\frac{1}{2}\left(\sum_{r\in C_A^+}r-2\alpha\right) =\delta_A-\alpha. \end{aligned}

The third equals sign is given because the simple roots permute CA+{α}C_A^+\setminus\{\alpha\}, which has been proved in the recovery of root system from a base.

Now, we iterate the elements in the group WW, and then choose some ωW\omega\in W to make (ω(δA),δB)(\omega(\delta_A),\delta_B) maximal. Then we compute (ω(sα(δA)),δB)(\omega(s_\alpha(\delta_A)),\delta_B),

(ω(sα(δA)),δB)=(ω(δA),δB)(ω(α),δB).(\omega(s_\alpha(\delta_A)),\delta_B)=(\omega(\delta_A),\delta_B)-(\omega(\alpha),\delta_B).

By the maximality of (ω(δA),δB)(\omega(\delta_A),\delta_B), we have (ω(α),δB)0(\omega(\alpha),\delta_B)\ge0. Now, if ω(α)CB\omega(\alpha)\in C_B^-, we have (ω(α),δB)<0(\omega(\alpha),\delta_B)<0, this is a contradiction. Hence, we have ω(α)CB+\omega(\alpha)\in C_B^+. Since ω\omega keeps addition, we know that ω(α)\omega(\alpha) is a simple root. This implies ω(A)=B\omega(A)=B by iterating αA\alpha\in A.

Finally, let’s compute

(sα(α1),sα(α2))=(α12(α1,α)(α,α)α,α22(α2,α)(α,α)α)=(α1,α2)2(α1,α)(α,α)(α,α2)2(α2,α)(α,α)(α1,α)+4(α1,α)(α2,α)(α,α)2(α,α)=(α1,α2).\begin{aligned} &(s_\alpha(\alpha_1),s_\alpha(\alpha_2))\\=&(\alpha_1-\frac{2(\alpha_1,\alpha)}{(\alpha,\alpha)}\alpha,\alpha_2-\frac{2(\alpha_2,\alpha)}{(\alpha,\alpha)}\alpha)\\ =&(\alpha_1,\alpha_2)-\frac{2(\alpha_1,\alpha)}{(\alpha,\alpha)}(\alpha,\alpha_2)-\frac{2(\alpha_2,\alpha)}{(\alpha,\alpha)}(\alpha_1,\alpha)\\ &+\frac{4(\alpha_1,\alpha)(\alpha_2,\alpha)}{(\alpha,\alpha)^2}(\alpha,\alpha)\\ =&(\alpha_1,\alpha_2). \end{aligned}

This implies elements in the Weyl group preserve inner product. And hence it is an isometric map. Therefore, their Dynkin diagram shall be the same.

Now, everything we want has been proved, and this is the end of this article.