Here we define subsystems of root systems and prove necessary facts about them. Unlike for groups, rings, etc. it's pretty tough to prove a subset closed under the operation is itself a root system. Though an equivalent statement is given as an exercise in the text Combinatorics of Coxeter groups, so are some unsolved problems, and the equivalent statement was the main result of Matthew Dyer's dissertation.
Definition 1: Let (R,B) be a root system. A subset U⊆R is called a subsystem if it is closed under the operation ∗ in R. Define
BR(U)={x∈R+(B)∩U|x∗y∈R+(B) for all y∈R+(B)∩U with x≠y}
We will be showing below that BR(U) forms a basis for U, though it's not even clear that BR(U) is nonempty. Define also for x∈R and U⊆R
x∗U={x∗y|y∈U}
Intuitively this is a "translation" of a subsystem by a root. We've been using this notation already for inversion sets. Note that y↦x∗y is a bijection between U and x∗U, and since (x∗y)∗(x∗z)=x∗(y∗z) we have that x∗U is a subsystem of R. You may wish to verify that if x∈U then x∗U=U.
Lemma 4.1: If (R,B) is a root system, U⊆R is a subsystem, and x∈B−U then
BR(x∗U)=x∗BR(U)
and
R+(BR(x∗U))=x∗R+(BR(U))
Proof: First we prove that BR(x∗U)=x∗BR(U). Note that x∉x∗U. Suppose y∈BR(x∗U). Then for all z∈R+(B)∩(x∗U) such that y≠z we have that y∗z∈R+(B)∩U. Since x∈B and x∉U, x≠y,z and hence x∗(y∗z)=(x∗y)∗(x∗z)∈R+(B)∩U. Thus x∗y∈BR(U), hence y∈x∗BR(U). Thus BR(x∗U)⊆x∗BR(U).
Suppose conversely that y∈x∗BR(U). Then for all z∈R+(B)∩U with z≠x∗y we have that (x∗y)∗z∈R+(B)∩U, hence x∗((x∗y)∗z)=y∗(x∗z)∈R+(B)∩(x∗U). Thus y∈BR(x∗U), and the result follows.
Now note that R+(BR(x∗U)) is the smallest set satisfying BR(x∗U)=x∗BR(U)⊆R+(BR(x∗U)) and if y∈R+(BR(x∗U)) and x′∈x∗BR(U) are such that x′≠y, then x′∗y∈R+(BR(x∗U)). Since x∗(y∗z)=(x∗y)∗(x∗z), we see that the bijection y↦x∗y preserves these properties, thus x∗R+(BR(x∗U))=R+(BR(U)). The result follows.
◻
The following theorem is essentially the main result of Dyer's dissertation (Reflection subgroups of Coxeter systems, Journal of Algebra 135 (1), 1990, 57-73).
Theorem 4.2: Let (R,B) be a root system and let U⊆R be a subsystem. Then (U,BR(U)) is a root system and R+(BR(U))=R+(B)∩U
Proof: The theorem follows if we prove that R+(BR(U))=R+(B)∩U. It is an easy consequence of the definition of BR(U) that R+(BR(U))⊆R+(B)∩U. We prove that the reverse inclusion holds. Suppose y∈R+(B)∩U. Let y=w(x) for w∈W(R) and x∈B with ℓ(w) minimal. We prove by induction on ℓ(w) that y∈R+(BR(U)). If ℓ(w)=0, then y∈B, hence y∈BR(U), so y∈R+(BR(U)). Otherwise let x′∈B be such that ℓ(sx′w)<ℓ(w). If x′∈U, then since x′∗y=sx′w(x) and ℓ(sx′w)<ℓ(w) we have by the induction hypothesis that x′∗y∈R+(BR(U)), hence the same is true of y=x′∗(x′∗y). If x′∉U, then x′∗y∈x′∗U and by the induction hypothesis x′∗y∈R+(BR(x′∗U))=R+(x′∗BR(U))=x′∗R+(BR(U)) (the equalities by Lemma 4.1), so y=x′∗(x′∗y)∈R+(BR(U)). The result follows by induction. ◻
We need to know now that W(U) can be naturally treated as a subgroup of W(R). For x∈U define sUx:U→U to be the corresponding reflection in U, and sRx to be the corresponding reflection in R. Define WR(U) to be the subgroup of W(R) generated by all sRx.
Theorem 4.3: Let (R,B) be a root system and let U⊆R be a subsystem. Then the group homomorphism WR(U)→W(U) determined by sRx↦sUx is an isomorphism.
Proof: The reflections sRx for x∈BR(U) generate WR(U) by Theorem 2.2. Let w∈WR(U) be a nonidentity element; we claim that the image of w is not the identity. Let ℓU(w) be the minimal length of a word for w in terms of elements of S(BR(U)). Let y∈R+(B)∩U. We claim that ℓU(wsy)<ℓU(w) if and only if y∈I(w). The proof of this is similar to an earlier proof (Theorem 2.3), so we omit it. By exactly the same line of reasoning as the proof of Proposition 2.4, it follows that ℓU(w)=|R+(B)∩U|. Thus if w fixes R we must have that ℓU(w)=0, so w is the identity. This proves injectivity of the homomorphism. Surjectivity is clear. ◻
We therefore do not distinguish reflections in a subsystem from the corresponding reflections in the larger root system, and we identify W(U) with the subgroup WR(U) of W(R).
Definition 2: If (R,B) is a root system, U⊆R is a subsystem, and w∈W(U) define
IU(w)=I(w)∩U
so that I(w)=IR(w). We use this notation to distinguish inversion sets of elements in W(U) in the subsystem U from the inversion sets of the elements in the larger root system R.
Lemma 4.4: Let (R,B) be a root system, U⊆R a subsystem, x∈B−U, and w∈W(U). Then
IU(w)=x∗Ix∗U(wsx)
Proof: If y∈IU(w), then w(y)∉R+(B) and y∈U. Then x∗y∈(x∗U)∩R+(B), and wsx(x∗y)=w(y)∉R+(B). Thus IU(w)⊆x∗Ix∗U(wsx). Conversely, if y∈Ix∗U(wsx), then y∈x∗U, so that x∗y∈U, and wsx(y)=w(x∗y)∉R+(B). Since x∉x∗U we have that x∗y∈IU(w) and the result follows. ◻
Now we define a function ϕU:W(R)→W(U) that allows us to "project" elements of W(R) onto W(U).
Theorem-Definition 4.5 If (R,B) is a root system and U⊆R is a subsystem, then there is a unique function ϕU:W(R)→W(U) such that for each w∈W(R) we have that
I(w)∩U=IU(ϕU(w))
Proof: We prove existence by induction ℓ(w), the result being clear if ℓ(w)=0. Let x∈I(w)∩B. If x∈U, then since I(w)={x}∪(x∗I(wsx)) we have that
I(w)∩U=({x}∪(x∗I(wsx)))∩U={x}∪(x∗IU(ϕU(wsx)))=IU(ϕU(wsx)sx)
(using Lemma 3.2), so
ϕU(w)=ϕU(wsx)sx
If x∉U, we note that
I(w)∩U=({x}∪(x∗I(wsx)))∩U=x∗(I(wsx)∩(x∗U))=x∗(Ix∗U(ϕx∗U(wsx)))=x∗(x∗IU(ϕx∗U(wsx)sx))=IU(ϕx∗U(wsx)sx)
(using Lemma 3.2 and Lemma 4.4), hence
ϕU(w)=ϕx∗U(wsx)sx
and the result follows by induction. Uniqueness follows by uniqueness of inversion sets (Proposition 3.5). ◻
The following theorem is quite deep and is related to the theory of permutation patterns.
Theorem-Definition 4.6: Let (R,B) be a root system, let w∈W(R), and let U⊆R be a subsystem. Then there exist unique elements wU∈W(R) and wU∈W(U) such that I(wU)∩U=∅ and
w=wUwU
In particular, wU is the unique element of minimal length in the coset wW(U).
Proof: I claim that wU=ϕU(w) satisfies the condition, with wU=ww−1U. We prove by induction on ℓU(ϕU(w)) that I(wϕU(w)−1)∩U=∅. Set v=ϕU(w). If ℓU(v)=0, then wU=w and wU=1=ϕU(w) will work because I(wU)∩U=∅ by definition. Otherwise let x∈BR(U)∩IU(v). Then ℓ(wsx)<ℓ(w) because x∈I(w), and by the induction hypothesis since ϕU(wsx)=vsx we have that I((wsx)U)∩U=I((wsx)(vsx)−1)∩U=∅. Since (wsx)(vsx)−1=wv−1, the result follows by induction.
To prove uniqueness of wU, and hence of wU, suppose u∈wW(U) is an element of minimal length. Then I(u)∩U=∅ because ℓ(usy)>ℓ(u) for all y∈R+(B)∩U. Any element in wW(U) is equal to uv for some v∈W(U), and if y∈IU(v) then −v(y)∈R+(B)∩U, hence −uv(y)∈R+(B)∩U, hence y∈I(uv). This proves uniqueness of the minimal element as well as the element u such that I(u)∩U=∅, because any other element in the coset contains a root in R+(B)∩U in its inversion set.
◻
3. Roots, words and inversion sets
In this post we gather some technical results relating roots, words and inversion sets. This is all standard Coxeter group theory.
Definition 1: Let (R,B) be a root system and let w=(sx1,…,sxn) be a word. For each index i≤n we assign a root ri(w)=sxnsxn−1⋯sxi+1(xi)
The proof of Lemma 3.1 below is actually quite easy and we leave it to the reader.
Lemma 3.1: Let (R,B) be a root system and let u=(sx1,…,sxn) be a word. Let x∈B and let v=(sx1,…,sxn,sx) Then for each 1≤i≤n we have ri(v)=x∗ri(u) ◻
The next lemma allows us to build inversion sets root by root.
Lemma 3.2: Let (R,B) be a root system, w∈W(R), and x∈B. If x∉I(w), then I(wsx)={x}∪(x∗I(w)) If x∈I(w), then I(wsx)=x∗(I(w)−{x})
Proof: Suppose x∉I(w). Then since wsx(x)=w(−x)=−w(x)∈R−(B) we have that x∈I(wsx). If y∈I(wsx), then wsx(y)=w(sx∗y)∉R+(B) Since y≠x, x∗y∈R+(B). Thus x∗y∈I(w). Conversely, if y∈I(w), then w(y)=wsx(x∗y)∉R+(B), hence x∗y∈I(wsx). Thus y↦x∗y is a bijection of I(w) with I(wsx)−{x}, and we are done. The second part is left to the reader. ◻
Theorem 3.3 tells us why the roots we assign to these indices are important.
Theorem 3.3: Let (R,B) be a root system and let w=(sx1,…,sxn) be a reduced word for the element w=sx1⋯sxn Then I(w)={ri(w)|1≤i≤n}
Proof: We prove this by induction on ℓ(w). The base case is ℓ(w)=0, in which case a reduced word for w is empty and hence the result is true. For the induction step, we assume the result is true for w with an arbitrary word w=(sx1,…,sxn), let x∈B−I(w), and prove the result for the element wsx and the word w′=(sx1,…,sxn,sx) We define unambiguously Iwsx={ri(w′)|1≤i≤n+1} By Lemma 3.1 above and the induction hypothesis, we have that Iwsx={x}∪(x∗I(w)) Hence by Lemma 3.2, Iwsx=I(wsx) and the result follows. ◻
Definition 2: Given an element w∈W(R), denote by R(w) the set of reduced words for w. Given a reduced word w=(sx1,…,sxn) for the element w, we define a bijection bw:I(w)→[ℓ(w)] by declaring that bw(rj(w))=j. Then define B(w)={bw|w∈R(w)} These can be seen as linear orderings on the inversion set. These correspond bijectively to reduced words as we prove now.
Theorem 3.4: Let (R,B) be a root system and let w∈W(R). Then the map R(w)→B(w) given by w↦bw is a bijection.
Proof: Clearly the map is surjective. We prove it is injective by induction on ℓ(w). This is clear if ℓ(w)=0. Otherwise, suppose u≠v are reduced words for w. If uℓ(w)≠vℓ(w), then there exist distinct x,x′∈B∩I(w) such that bu(x)=bv(x′)=ℓ(w), hence bu≠bv. If instead uℓ(w)=vℓ(w), then there exists x∈B∩I(w) such that bu(x)=bv(x)=ℓ(w). Define u′=(ui)i<ℓ(w) and v′=(vi)i<ℓ(w) Then u′≠v′ and if y≠x then bu(y)=bu′(x∗y) and bv(y)=bv′(x∗y) By the induction hypothesis, bu′(x∗y)≠bv′(x∗y) for some y∈I(w), hence bu≠bv and the result follows by induction. ◻
Uniqueness of inversion sets is a very useful fact that we now prove.
Proposition 3.5: Let (R,B) be a root system and let u,v∈W(R). If I(u)=I(v), then u=v.
Proof: We prove this by induction on ℓ(u)=ℓ(v). If ℓ(u)=ℓ(v)=0 this is clear. Otherwise let x∈I(u)∩B. Then I(usx)=I(vsx)=x∗(I(u)−{x}) By the induction hypothesis, usx=vsx, hence u=(usx)sx=(vsx)sx=v by induction. ◻
Definition 1: Let (R,B) be a root system and let w=(sx1,…,sxn) be a word. For each index i≤n we assign a root ri(w)=sxnsxn−1⋯sxi+1(xi)
The proof of Lemma 3.1 below is actually quite easy and we leave it to the reader.
Lemma 3.1: Let (R,B) be a root system and let u=(sx1,…,sxn) be a word. Let x∈B and let v=(sx1,…,sxn,sx) Then for each 1≤i≤n we have ri(v)=x∗ri(u) ◻
The next lemma allows us to build inversion sets root by root.
Lemma 3.2: Let (R,B) be a root system, w∈W(R), and x∈B. If x∉I(w), then I(wsx)={x}∪(x∗I(w)) If x∈I(w), then I(wsx)=x∗(I(w)−{x})
Proof: Suppose x∉I(w). Then since wsx(x)=w(−x)=−w(x)∈R−(B) we have that x∈I(wsx). If y∈I(wsx), then wsx(y)=w(sx∗y)∉R+(B) Since y≠x, x∗y∈R+(B). Thus x∗y∈I(w). Conversely, if y∈I(w), then w(y)=wsx(x∗y)∉R+(B), hence x∗y∈I(wsx). Thus y↦x∗y is a bijection of I(w) with I(wsx)−{x}, and we are done. The second part is left to the reader. ◻
Theorem 3.3 tells us why the roots we assign to these indices are important.
Theorem 3.3: Let (R,B) be a root system and let w=(sx1,…,sxn) be a reduced word for the element w=sx1⋯sxn Then I(w)={ri(w)|1≤i≤n}
Proof: We prove this by induction on ℓ(w). The base case is ℓ(w)=0, in which case a reduced word for w is empty and hence the result is true. For the induction step, we assume the result is true for w with an arbitrary word w=(sx1,…,sxn), let x∈B−I(w), and prove the result for the element wsx and the word w′=(sx1,…,sxn,sx) We define unambiguously Iwsx={ri(w′)|1≤i≤n+1} By Lemma 3.1 above and the induction hypothesis, we have that Iwsx={x}∪(x∗I(w)) Hence by Lemma 3.2, Iwsx=I(wsx) and the result follows. ◻
Definition 2: Given an element w∈W(R), denote by R(w) the set of reduced words for w. Given a reduced word w=(sx1,…,sxn) for the element w, we define a bijection bw:I(w)→[ℓ(w)] by declaring that bw(rj(w))=j. Then define B(w)={bw|w∈R(w)} These can be seen as linear orderings on the inversion set. These correspond bijectively to reduced words as we prove now.
Theorem 3.4: Let (R,B) be a root system and let w∈W(R). Then the map R(w)→B(w) given by w↦bw is a bijection.
Proof: Clearly the map is surjective. We prove it is injective by induction on ℓ(w). This is clear if ℓ(w)=0. Otherwise, suppose u≠v are reduced words for w. If uℓ(w)≠vℓ(w), then there exist distinct x,x′∈B∩I(w) such that bu(x)=bv(x′)=ℓ(w), hence bu≠bv. If instead uℓ(w)=vℓ(w), then there exists x∈B∩I(w) such that bu(x)=bv(x)=ℓ(w). Define u′=(ui)i<ℓ(w) and v′=(vi)i<ℓ(w) Then u′≠v′ and if y≠x then bu(y)=bu′(x∗y) and bv(y)=bv′(x∗y) By the induction hypothesis, bu′(x∗y)≠bv′(x∗y) for some y∈I(w), hence bu≠bv and the result follows by induction. ◻
Uniqueness of inversion sets is a very useful fact that we now prove.
Proposition 3.5: Let (R,B) be a root system and let u,v∈W(R). If I(u)=I(v), then u=v.
Proof: We prove this by induction on ℓ(u)=ℓ(v). If ℓ(u)=ℓ(v)=0 this is clear. Otherwise let x∈I(u)∩B. Then I(usx)=I(vsx)=x∗(I(u)−{x}) By the induction hypothesis, usx=vsx, hence u=(usx)sx=(vsx)sx=v by induction. ◻
Subscribe to:
Posts (Atom)