Processing math: 100%

2. Algebraic root systems

This construction, which is my own, is an alternative axiomatization of the root systems from Lie theory. This is not completely unparalleled; there is a similar construction by Dyer (in the paper On rigidity of abstract root systems of Coxeter systems), though it is a bit more complicated. These are a combination of standard results about Coxeter groups and standard results about root systems combined in one abstract framework. For more traditional reading try Combinatorics of Coxeter groups by Björner and Brenti or Reflection Groups and Coxeter Groups by Humphreys.

A pre-root system is defined as a set R with a binary operation :R×RR satisfying the following axioms:

Axiom 1: x(xy)=y

Axiom 2: x(yz)=(xy)(xz)

Note that this operation is is virtually never associative. One might suspect that this leads to pathologies but with the third axiom below quite the opposite is true.

For convenience in notation we make the following definitions.

Definition 1: Let R be a pre-root system. If xR, define a function sx:RR by sx(y)=xy for all yR. sx is the reflection corresponding to x. Let W(R) be the group generated by the reflections in R; W(R) is called the Weyl group of R. Define also x=xx=sx(x). Then we have the following identities (proof left to the reader):
  1. s2x=1, the identity element of W(R), and in particular (x)=x.
  2. sx(y)=sx(y)
  3. sx(y)=sx(y)
  4. sx(yz)=sx(y)sx(z)=ssx(y)(sx(z))


Definition 2: Let BR be a subset of a pre-root system R and define R+(B) to be the smallest subset of R satisfying
  1. BR+(B).
  2. If yR+(B), xB, and xy, then xyR+(B).

Now we can state the third axiom. A pre-root system R is called a root system if the following additional axiom holds:

Axiom 3: There exists a subset BR such that for all xR we have xR+(B) if and only if xR+(B).

A subset BR satisfying Axiom 3 is called a basis of R, the elements of B are called simple roots, and R+(B) is called the set of positive roots. The subset R+(B)={yR|yR+(B)} is called the set of negative roots. The reflections sx for xB are called the simple reflections and the set of these will be denoted by S(B). These definitions are dependent on the choice of basis, so most of the time we will be fixing a basis of R and considering pairs (R,B) where R is a root system and B is a basis. It will therefore be convenient to refer to the pair (R,B) as a root system.

The inevitable question is, "Why should I care about this?" Let's construct an example. Let V be the real vector space spanned by the indeterminates t1,t2,,tn+1 and let RnA be the set of all differences titj with ij. For positive integers a,b with ab define a permutation sa,b, the transposition exchanging a and b, by sa,b(i)={b if i=aa if i=bi otherwise Then we define a product :RnA×RnARnA by (tatb)(tctd)=tsa,b(c)tsa,b(d) I leave it to you to prove that RnA satisfies Axioms 1 and 2. To convince you that our notation is consistent, note that (tatb)(tatb)=tbta=(tatb) where the right-hand side of the equation can be interpreted either in a root-system-theoretic sense or in the vector space sense; the two are equivalent.

For Axiom 3 we need to choose a basis. Define BnA={titi+1|1in} Theorem 2.1: (RnA,BnA) is a root system.

Proof: I claim that R+(BnA)={titj|i<j} To see that R+(BnA) must contain all differences titj with i<j we use induction on ji, the result being by definition for ji=1. Suppose then that tatb for all ba<k are contained in R+(BnA). Then (tbtb+1)(tatb)=tatb+1, so the result follows by induction. To see that R+(BnA) must be exactly equal to the set of these differences, note that if tatb is such that a<b, then if ca or c+1b then (tctc+1)(tatb)=tsc,c+1(a)tsc,c+1(b) satisfies sc,c+1(a)<sc,c+1(b), so since R+(BnA) is the smallest subset with the property that for all yR+(BnA) and xBnA with xy we have that xyR+(BnA) we have the result.

Now, ask yourself the following question: "What is the Weyl group of RnA (that is to say W(RnA))?" It is the symmetric group Sn+1! To see this, extend the action of tatb to the elements ti, that is (tatb)ti=tsa,b(i). Where an element of W(RnA) sends any root is completely determined by where it sends all ti, so an element of W(RnA) is uniquely determined by the bijection it induces of {t1,,tn+1} with itself. This gives us a permutation representation of W(RnA). This representation contains all of the transpositions, so it must be the entirety of Sn+1. The theory below will allow us to prove in short order highly nontrivial results about elements of Sn+1.

Pre-root systems that do not satisfy Axiom 3 are of little interest to us. However, they are far from useless. Structures that satisfy only Axiom 2 are called quandles, and structures satisfying both Axioms 1 and 2 are called involutory quandles. Both have applications in knot theory.


Now we turn to the development of the theory of root systems.

Proposition 2.2: Let (R,B) be a root system. Then
  1. For all wW(R) and xR we have wsxw1=sw(x).
  2. For all yR there exist wW(R) and xB such that y=w(x).
  3. The set of simple reflections S(B) is a generating set of W(R).
Proof of 1: Let wW(R) and xR. We prove the result by induction on the minimal number of reflections x1,,xk such that w=sx1sxk. If k=0 this is clear. If the result holds for k1, note that sx1w=sx2sxk. Then sx1wsx(sx1w)1=ssx1w(x) by the induction hypothesis. Thus wsxw1(y)=sx1ssx1w(x)sx1(y)=x1(sx1w(x)(x1y))=(x1sx1w(x))(x1(x1y))=w(x)y=sw(x)(y) and the result follows by induction.

Proof of 2: First assume yR+(B). Set B0=B, and for i>0 define Bi={xy|xB,xy,yBi1} We claim that BiR+(B) for all i. We prove this by induction, the base case B0R+(B) being clear by definition. Suppose yBi, i>1. Then there exists xBi such that xyBi1 (note we are using the fact that x(xy)=y). Since by the induction hypothesis xyR+(B), and xB, by definition of R+(B) we must have that x(xy)=yR+(B), so the result follows by induction.

It follows that i=0BiR+(B) We also have that if yi=0Bi, xB, and xy then xyi=0Bi. By definition, R+(B) is the smallest subset satisfying this property, so the opposite inclusion R+(B)i=0Bi holds. For each yBi we have that there exists xB and wW(R) such that w(x)=y; namely, if we set yi=yBi and let xiB be such that xiyiBi1, then y=sxisxi1sx1(x0) where x0B satisfies x1x0=y1, so we may take w=sxisxi1sx1 and x=x0, hence the result follows for positive roots. For negative roots y, note that y=sy(y) is positive by Axiom 3, so there exists wW(R) and xB such that w(x)=sy(y); thus syw(x)=y, so we may take w=syw.

Proof of 3: It suffices to show that each reflection sy for yR+(B) can be written as a product of simple reflections since sy=sy. We know from the proof of part 2 that there exist simple roots x1,,xiB and xB such that sx1sxi(x)=y. Set sx1sxi=w. Then wsxw1=sw(x)=sy by part 1, so the result follows.


Definition 3: Let (R,B) be a root system and let wW(R). A sequence of simple reflections (sx1,,sxn) is called a word for w if w=sx1sxn. We know by the previous proposition that since S(B) generates W(R), every element of W(R) has at least one word. If (sx1,,sxn) is a word for w such that the length n is as small as possible, then the word is called a reduced word. Define (w) to be the length of a reduced word for w. Define also the inversion set I(w) of w by I(w)={yR+(B)|w(y)R+(B)} A positive root yR+(B) such that w(y)R+(B) will correspondingly be called an inversion.


Theorem 2.3: Let (R,B) be a root system and let wW(R). Let yR+(B). Then (wsy)<(w) if and only if yI(w). If yI(w) and (sx1,,sxn) is a (possibly unreduced) word for w, then there is an index i such that (sx1,,^sxi,,sxn) is a word for wsy.

Proof: Suppose yI(w) and let (sx1,,sxn) be a word for w. Let i be the maximal index such that sxisxn(y)R+(B). We have that xisxi+1sxn(y)R+(B), hence sxi+1sxn(y)=xi because xiB. Thus y=sxnsxi+1(xi), hence sy=sxnsxi+1sxisxi+1sxn. Thus (sx1,,^sxi,,sxn) is a word for wsy, where the caret indicates omission. If we assume that the word is reduced, then since wsy has a word that is shorter than a reduced word for w we must have that (wsy)<(w).

Now note that if yR+(B)I(w), then wsy(y)=w(y)=w(y)R+(B), hence yI(wsy). Thus (w)=((wsy)sy)<(wsy), so the result follows.


The previous theorem is more important than it looks. It implies that (W(R),S(B)) is a Coxeter system and W(R) is a Coxeter group. Coxeter groups have extremely nice properties and there is a huge amount of literature on them. We're assuming no prior knowledge in this blog, so we will be proving the required results about Coxeter group theory as they come up.


Proposition 2.4: Let (R,B) be a root system and let wW(R).
  1. (w)=|I(w)|; that is, the length of a reduced word for w is the number of inversions of w.
  2. If xI(w)B, then (wsx)=(w)1.
Proof: We prove part 1 by induction on |I(w)|, the result being true for the identity, which has empty inversion set. Let (sx1,,sxn) be a reduced word for w. Then xnI(w)B since (wsxn)<(w). Let yR+(B). We have that wsxn(y)R+(B) if and only if w(xny)R+(B). Thus I(wsxn)=x(I(w){xn}). In particular, |I(wsxn)|=(wsxn)=|I(w)|1. We can append sxn to any reduced word for wsxn to obtain a word for w, which must be reduced since (w)>(wsxn). Thus (w)=(wsxn)+1=|I(wsxn)|+1=|I(w)|, so both results follow.


As a final note, I promised you nontrivial results about Sn+1, so here is one that comes for free. Note that the set of simple reflections in Sn+1 is the set of all si,i+1, that is to say the adjacent transpositions.

Corollary 2.5: If wSn+1, then the minimal number of adjacent transpositions required to express w (meaning (w)) is exactly the number of pairs i<j such that w(i)>w(j).

Proof: We know from the previous proposition that (w) is equal to the number of inversions of w. What is an inversion of an element of Sn+1? It is a root titj with i<j such that tw(i)tw(j) is a negative root, meaning w(i)>w(j). This is exactly what the statement of the corollary claims.

1. Introduction

In this blog we will be exploring Schubert calculus without any algebraic geometry or algebraic topology. Our method of investigation is just plain algebra. With just some undergraduate level abstract algebra under your belt, you can go from zero to the very core of the deep combinatorics of Schubert calculus in minutes. Allow me to demonstrate.

For this introduction we will be focusing only on type A, and the main object of study is the symmetric group Sn. Sn for our purposes will be the group of all bijective functions from the set [n]={1,2,,n} to itself. There are natural injective homomorphisms SnSn+1 for all n obtained by considering [n] as a subset of [n+1]; the image of a function in this homomorphism SnSn+1 will fix n+1 and do to the rest of the elements what it is doing already. It will be convenient to consider all of these groups to be sitting inside S, which is the group of bijective functions f:NN such that f(i)=i for all but finitely many i. Essentially, S=n=1Sn

Let F be a field of characteristic zero. We turn to the polynomial ring F[x1,x2,] in countably many indeterminates, which we will for convenience refer to as R. S acts on R by F-automorphisms (that is, automorphisms that are F-linear). Specifically, for fS the action of f on xi is fxi=xf(i) There is only one way to extend this to an F-automorphism of R. Essentially f's action is the substitution xixf(i) in the usual sense. This action, while natural, is not commonly used in the literature because while it works perfectly well as a left action on the polynomials themselves, meaning f(gp)=(fg)p, it does not come from a left action on the variables. If you move the action inside the parentheses of the polynomial p(x1,), viewed as a function, the order of multiplication in S is reversed. Despite this, this will be the most convenient action for our purposes.

Now we will need to refer to specific elements of S, namely the adjacent transpositions si defined by si(j)={i+1 if j=ii if j=i+1j otherwise The elements si generate S as a Coxeter group, which is very important and will be discussed later, but we will not dwell on it now. Now we are interested in defining the divided difference operators i for iN. i takes elements of R to elements of R and is F-linear, and is defined abstractly as follows: i=1xixi+1(1si) If this definition is too abstract for you, that's no problem. It is sufficient to know that i acts as follows: i(p)=psipxixi+1 That is, we take the polynomial, switch two of the variables, subtract from the original, and divide by the difference of the two variables we swapped. It is not obvious that this always yields a polynomial, and you may want to convince yourself of that before moving on.

Now we define the operators w for all wS built from the i. Namely, if w can be expressed as a specific product of adjacent transpositions w=si1sik with k as small as possible (so the expression is as short as possible) then we define w=i1ik If we take such a product where the expression is not as short as possible, we instead get 0.

It is not difficult to work out a product rule for i. You may want to try it yourself, but in any case the formula is i(pq)=i(p)q+pi(q)+(xi+1xi)i(p)i(q) We can iterate this to obtain a product rule for w for arbitrary permutations w. We end up with a formula of the form w(pq)=u,vScwu,vu(p)v(q) where the coefficients cwu,v are polynomials depending on u, v, and w.

You may be surprised to know that by now I have already made good on my promise of bringing you from zero to Schubert calculus in minutes. The coefficients cwu,v, the equivariant Littlewood-Richardson coefficients, are one of the central objects of study in Schubert calculus, and finding a formula of a certain type for them is a hopelessly difficult unsolved problem. While irrelevant to our geometry-free investigation, it is interesting to note that the polynomials cwu,v are the structure constants in the equivariant cohomology rings of complete flag varieties over the complex numbers (rings we will construct combinatorially later), and whenever cwu,v is an integer it counts the number of points in transverse triple intersections of Schubert varieties. It is known via geometric proofs (see Graham, Duke Math. J. Volume 109, Number 3 (2001), 599-614) that cwu,v is a polynomial in the differences xi+1xi with nonnegative integer coefficients. This has not yet been paralleled combinatorially except in special cases, but we're working on it.

In the next post we will build our general framework for equivariant Schubert calculus.