1 Introduction
In [Reference Steinberg9, Sections 14.1, 14.3], Steinberg proved the following theorem (see also [Reference Bourbaki1, Ch. V, Section 2, Exercise 3], [Reference Curtis, Iwahori and Kilmoyer3, Theorem 9.13], [Reference Geck and Pfeiffer5, Theorem 5.1.4] and [Reference Kane7, Section 24-3]).
Theorem 1.1 (Steinberg).
Let V be a finite-dimensional vector space endowed with an inner product (for example, a Euclidean space or a complex Hilbert space). Let $\{v_1, \dots , v_n\}$ be a basis of V and $W \subseteq \operatorname {GL}(V)$ be the group generated by (orthogonal) reflections with respect to these basis vectors. Suppose V is a simple W-module. Then the W-modules $\{\bigwedge ^d V \mid 0 \le d \le n\}$ are simple and pairwise nonisomorphic.
The proof relies on the existence of an inner product which stays invariant under the W-action. With the help of this inner product, the vector space $\bigwedge ^d V$ is decomposed into a direct sum $\bigwedge ^d V^\prime \bigoplus (v \wedge \bigwedge ^{d-1} V^\prime )$ , where $V^\prime $ is a subspace of V of codimension one and a simple module of a subgroup generated by fewer reflections, and v is a vector orthogonal to $V^\prime $ . The theorem is proved by induction on the number of reflections.
We extend this result to a more general context, where the W-invariant inner product may not exist. The following is the main theorem.
Theorem 1.2. Let $\rho :W \to \operatorname {GL}(V)$ be an n-dimensional representation of a group W over a field $\mathbb {F}$ of characteristic 0. Suppose that $s_1, \dots , s_k \in W$ satisfy:
-
(1) for each i, $s_i$ acts on V by a (generalised) reflection with reflection vector $\alpha _i$ of eigenvalue $\lambda _i$ (see Definition 2.1 for related notions);
-
(2) the group W is generated by $\{s_1, \dots , s_k\}$ ;
-
(3) the representation $(V,\rho )$ is irreducible;
-
(4) for any pair $i,j$ of indices, $s_i \cdot \alpha _j \ne \alpha _j$ if and only if $s_j \cdot \alpha _i \ne \alpha _i$ .
Then the W-modules $\{\bigwedge ^d V \mid 0 \le d \le n\}$ are irreducible and pairwise nonisomorphic.
Remark 1.3. The condition (4) in Theorem 1.2 is a technical condition (automatically satisfied in the setting of Theorem 1.1). However, it is not that strict. For example, if $s_i$ and $s_j$ are both of order $2$ (so that they generate a dihedral subgroup), and if $s_i \cdot \alpha _j = \alpha _j$ while $s_j \cdot \alpha _i \ne \alpha _i$ , then the order of $s_is_j$ in W must be $\infty $ . Moreover, since $\langle \alpha _i,\alpha _j \rangle $ forms a subrepresentation of the subgroup $\langle s_i, s_j \rangle $ , there are uncountably many two-dimensional representations of the infinite dihedral group $\langle s_i, s_j \rangle $ , but only two of them invalidate the condition (4) (see [Reference Hu6, Section 2.2]).
In the paper [Reference Hu6], we construct and classify a class of representations of an arbitrary Coxeter group of finite rank, where the defining generators of the group act by (generalised) reflections. In view of the previous remark, most of these reflection representations satisfy the conditions of Theorem 1.2. Thus, our result applies to them giving many irreducible representations of the Coxeter group. (Note that in [Reference Hu6], we have seen that only a few reflection representations admit a nonzero bilinear form which is invariant under the group action. Consequently, the module $\bigwedge ^d V$ usually fails to decompose into the form $\bigwedge ^d V^\prime \bigoplus (v \wedge \bigwedge ^{d-1} V^\prime )$ as it did in proving Theorem 1.1. Even if we have such a decomposition, the subspace $V^\prime $ may not be a simple module of a suitable subgroup. Therefore, the arguments in proving Theorem 1.1 usually fail in the context of Theorem 1.2.)
This paper is organised as follows. In Sections 2–4, we revisit basic concepts and provide the background we need concerning (generalised) reflections, exterior powers and graphs. In Section 5, we prove our main theorem. In Section 6, we present some byproducts, including a description of the subspace of an exterior power that is fixed pointwise by a set of reflections, and a Poincaré-like duality on exterior powers. In Section 7, we raise several interesting questions that have yet to be resolved.
2 Generalised reflections
Let $\mathbb {F}$ be a field and V be a finite-dimensional vector space over $\mathbb {F}$ .
Definition 2.1.
-
(1) A linear map $s:V \to V$ is called a (generalised) reflection if s is diagonalisable and $\operatorname {\mathrm {rank}}(s - \mathrm {Id}_V) = 1$ .
-
(2) Suppose s is a reflection on V. The hyperplane $H_s : = \ker (s - \mathrm {Id}_V)$ , which is fixed pointwise by s, is called the reflection hyperplane of s. Let $\alpha _s$ be a nonzero vector in $\operatorname {\mathrm {Im}}(s - \mathrm {Id}_V)$ . Then, $s \cdot \alpha _s = \lambda _s \alpha _s$ for some $\lambda _s \in \mathbb {F} \setminus \{1\}$ and $\alpha _s$ is called a reflection vector of s.
Note that if s is an invertible map, then $\lambda _s \ne 0$ .
Lemma 2.2. Let s be a reflection on V. Then there exists a nonzero linear function $f : V \to \mathbb {F}$ such that $s \cdot v = v + f(v) \alpha _s$ for any $v \in V$ .
Proof. Note that ${V = H_s \bigoplus \mathbb {F} \alpha _s}$ . Any vector v can be written in the form ${v = v_s + c_v \alpha _s}$ where $v_s \in H_s$ and $c_v \in \mathbb {F}$ . Then
The linear function $f: v \mapsto (\lambda _s - 1) c_v$ is the desired function.
3 Exterior powers
In this section, let W be a group and $\rho : W \to \operatorname {GL}(V)$ be a representation of W, where V is an n-dimensional vector space over the base field $\mathbb {F}$ . Let $\bigwedge ^d V$ ( $0 \le d\le n$ ) be the dth exterior power of V. The representation $\bigwedge ^d \rho $ of W on $\bigwedge ^d V$ is given by
By convention, $\bigwedge ^0 V$ is the one-dimensional W-module with trivial action. However, $\bigwedge ^n V$ carries the one-dimensional representation $\det \circ \rho $ . For more details, one may refer to [Reference Fulton and Harris4]. The following is a well-known fact.
Lemma 3.1. Suppose $\{\alpha _1, \dots , \alpha _n\}$ is a basis of V. Then,
is a basis of $\bigwedge ^d V\ (0 \le d \le n)$ . In particular, $\dim \bigwedge ^d V = \binom {n}{d}$ .
Suppose an element s of W acts on V by a reflection, with reflection hyperplane $H_s$ and reflection vector $\alpha _s$ of eigenvalue $\lambda _s$ (see Definition 2.1). Note that W is a group and s is invertible. Thus, $\lambda _s \ne 0$ . We define
to be the eigen-subspaces of s in $\bigwedge ^d V$ , for the eigenvalues $1$ and $\lambda _s$ , respectively.
Lemma 3.2 (See [Reference Geck and Pfeiffer5, Lemma 5.1.2] and the proof of [Reference Curtis, Iwahori and Kilmoyer3, Proposition 9.12]).
Let $W, s, V$ be as above. Suppose $\{v_1, \dots , v_{n-1}\}$ is a basis of $H_s$ . Then, $V_{d,s}^+\ (0 \le d \le n)$ has a basis
and $V_{d,s}^-$ has a basis
In particular, $\dim V_{d,s}^+ = \binom {n-1}{d}$ , $\dim V_{d,s}^- = \binom {n-1}{d-1}$ and $\bigwedge ^d V = V_{d,s}^+ \bigoplus V_{d,s}^-$ . (Here we regard $\binom {n-1}{n} = \binom {n-1}{-1} = 0$ .)
Proof. Note that $\{\alpha _s, v_1, \dots , v_{n-1}\}$ is a basis of V. Denote by $B^+$ and $B^-$ the two sets of vectors in (3.2) and (3.3), respectively. Then the disjoint union $B^+ \cup B^-$ is a basis of $\bigwedge ^d V$ by Lemma 3.1. Clearly, $B^+ \subseteq V_{d,s}^+$ and $B^- \subseteq V_{d,s}^-$ . Therefore, ${\bigwedge ^d V = V_{d,s}^+ \bigoplus V_{d,s}^-}$ and the result follows.
Corollary 3.3. Let $W, s, V$ be as above.
-
(1) We have $V_{d,s}^+ = \bigwedge ^d H_s$ . Here, $\bigwedge ^d H_s$ is regarded as a subspace of $\bigwedge ^d V$ naturally.
-
(2) Extend $\alpha _s$ arbitrarily to a basis of V, say, $\{\alpha _s, \alpha _2, \dots , \alpha _n\}$ . Then, $V_{d,s}^-\ (0 \le d \le n)$ has a basis
(3.4) $$ \begin{align} \{ \alpha_s \wedge \alpha_{i_1} \wedge \dots \wedge \alpha_{i_{d-1}} \mid 2 \le i_1 < \dots < i_{d-1} \le n \}. \end{align} $$
Proof. Point (1) in Corollary 3.3 is directly derived from Lemma 3.2. For (2), suppose ${s \cdot \alpha _i = \alpha _i + c_i \alpha _s}$ ( $c_i \in \mathbb {F}$ ) for $i = 2, \dots , n$ (see Lemma 2.2). Then,
Thus, $\alpha _s \wedge \alpha _{i_1} \wedge \dots \wedge \alpha _{i_{d-1}} \in V_{d,s}^-$ . Note that by Lemma 3.1, the $\binom {n-1}{d-1}$ vectors in (3.4) are linearly independent, and that $\dim V_{d,s}^- = \binom {n-1}{d-1}$ by Lemma 3.2. Thus, the vectors in (3.4) form a basis of $V_{d,s}^-$ .
For a subset $B \subset V$ , we denote by $\langle B \rangle $ the linear subspace spanned by B. By convention, $\langle \emptyset \rangle = 0$ . The following lemma is evident by linear algebra.
Lemma 3.4. Let $B \subseteq V$ be a basis of V and $B_i \subseteq B\ (i \in I)$ be a family of subsets of B. Then, $\bigcap _{i \in I} \langle B_i \rangle = \langle \bigcap _{i \in I} B_i \rangle $ .
Using the lemmas above, we can deduce the following results which will be used in the proof of our main theorem.
Proposition 3.5. Let $W, V$ be as above. Suppose $s_1, \dots , s_k \in W$ such that, for each i, $s_i$ acts on V by a reflection with reflection vector $\alpha _i$ of eigenvalue $\lambda _i$ . Suppose $\alpha _1, \dots , \alpha _k$ are linearly independent. We extend these vectors to a basis of V, say, $\{\alpha _1, \dots , \alpha _k, \alpha _{k+1}, \dots , \alpha _n\}$ .
-
(1) If $0 \le d < k$ , then $\bigcap _{1 \le i\le k} V_{d, s_i}^- = 0$ .
-
(2) If $k \le d \le n$ , then $\bigcap _{1 \le i\le k} V_{d, s_i}^-$ has a basis
$$ \begin{align*} \{\alpha_1 \wedge \dots \wedge \alpha_k \wedge \alpha_{j_{k+1}} \wedge \dots \wedge \alpha_{j_d} \mid k+1 \le j_{k+1} < \dots < j_d \le n\}. \end{align*} $$In particular, if $d = k$ , then $\bigcap _{1 \le i\le k} V_{k, s_i}^-$ is one-dimensional with a basis vector $\alpha _1 \wedge \dots \wedge \alpha _k$ .
Proof. By Corollary 3.3(2), $V_{d,s_i}^-$ ( $1 \le i \le k$ ) has a basis
Note that by Lemma 3.1, the ambient space $\bigwedge ^d V$ has a basis
and $B_i \subseteq B$ , for all $i = 1, \dots , k$ . By Lemma 3.4,
If $0 \le d < k$ , then $\bigcap _{1 \le i \le k} B_i = \emptyset $ and $\bigcap _{1 \le i\le k} V_{d, s_i}^- = 0$ . If $k \le d \le n$ , then
and $\bigcap _{1 \le i \le k} B_i$ is a basis of $\bigcap _{1 \le i\le k} V_{d, s_i}^-$ .
Proposition 3.6 (See the proofs of [Reference Curtis, Iwahori and Kilmoyer3, Theorem 9.13] and [Reference Geck and Pfeiffer5, Theorem 5.14]).
Let $W, V$ be as above. Suppose there exists $s \in W$ such that s acts on V by a reflection. If $0 \le k,l \le n$ are integers and $\bigwedge ^k V \simeq \bigwedge ^l V$ as W-modules, then $k = l$ .
Proof. If $\bigwedge ^k V \simeq \bigwedge ^l V$ , then $\dim \bigwedge ^k V = \dim \bigwedge ^l V$ and $\dim V_{k,s}^+ = \dim V_{l,s}^+$ . By Lemmas 3.1 and 3.2, this is equivalent to
The two equalities force $k = l$ .
The following lemma is due to Chevalley [Reference Chevalley2, page 88] (see also [Reference Milne8, Corollary 22.45]).
Lemma 3.7 (Chevalley).
Let $\mathbb {F}$ be a field of characteristic 0. Let W be a group and V, U be finite-dimensional semisimple W-modules over $\mathbb {F}$ . Then $V \bigotimes U$ is a semisimple W-module.
Note that the W-module $\bigwedge ^d V$ can be regarded as a submodule of $\bigotimes ^d V$ via the natural embedding
(The notation $\mathfrak {S}_d$ denotes the symmetric group on d elements.) Therefore, we have the following corollary.
Corollary 3.8. Let $\mathbb {F}$ be a field of characteristic 0. Let W be a group and V be a finite-dimensional simple W-module over $\mathbb {F}$ . Then the W-module $\bigwedge ^d V$ is semisimple.
4 Some lemmas on graphs
By definition, an (undirected) graph $G = (S,E)$ consists of a set S of vertices and a set E of edges. Each edge in E is an unordered binary subset $\{s,t\}$ of S. For our purpose, we only consider finite graphs without loops and multiple edges (that is, S is a finite set, there is no edge of the form $\{s,s\}$ and each pair $\{s,t\}$ occurs at most once in E).
A sequence $(s_1, s_2, \dots , s_n)$ of vertices is called a path in G if $\{s_i, s_{i+1}\} \in E$ , for all i. In this case, we say that the two vertices $s_1$ and $s_n$ are connected by the path. A graph G is called connected if any two vertices are connected by a path.
Definition 4.1. Let $G = (S,E)$ be a graph and $I \subseteq S$ be a subset. We set ${E(I) := \{\{s,t\} \in E \mid s,t \in I\}}$ to be the set of edges with vertices in I, and call the graph ${G(I) : = (I, E(I))}$ the subgraph of G spanned by I.
Definition 4.2. Let $G = (S,E)$ be a graph and $I \subseteq S$ be a subset. Suppose there exists vertices $r \in I$ and $t\in S \setminus I$ such that $\{r,t\} \in E$ is an edge. Let $I^\prime : = (I \setminus \{r\} ) \cup \{t\}$ . Then we say $I^\prime $ is obtained from I by a move.
Intuitively, $I^\prime $ is obtained from I by moving the vertex r to the vertex t along the edge $\{r,t\}$ . In particular, $\lvert I \rvert = \lvert I^\prime \rvert $ .
We shall need the following lemmas in the proof of our main theorem.
Lemma 4.3. Let $G = (S, E)$ be a connected graph. Let $I, J \subseteq S$ be subsets with cardinality $\lvert I \rvert = \lvert J \rvert = d$ . Then J can be obtained from I by finite steps of moves.
Proof. We do induction downwards on $\lvert I \cap J \rvert $ . If $\lvert I \cap J \rvert =d$ , then $I=J$ and there is nothing to prove.
If $I \ne J$ , then there exist vertices $r \in I \setminus J$ and $t \in J \setminus I$ . Since G is connected, there is a path connecting r and t, say,
Let $0 = i_0 < i_1 < \dots < i_k < l$ be the indices such that $\{r_{i_0}, r_{i_1}, \dots , r_{i_k} \}$ is the set of vertices in I on this path, that is, $\{r_{i_0}, \dots , r_{i_k} \} = \{r_i \mid 0 \le i < l, r_i \in I\}$ .
Clearly, $r_{i_k+1}, r_{i_k+2}, \dots , r_l \notin I$ . So beginning with I, we can move $r_{i_k}$ to $r_{i_k+1}$ , then to ${r_{i_k+2}}$ , and finally to t. Therefore, the set $I_1 := (I \setminus \{r_{i_k}\}) \cup \{t\}$ can be obtained from I by finite steps of moves. Similarly, from $I_1$ , we can move $r_{i_{k-1}}$ to $r_{i_k}$ so that we obtain
Do this recursively, and finally we get $I_{k+1} := (I \setminus \{r_{i_{0}}\}) \cup \{t\} = (I \setminus \{r\}) \cup \{t\}$ from I by finite steps of moves.
Moreover, we have $\lvert I_{k+1} \cap J \rvert = \lvert I \cap J \rvert + 1$ . By the induction hypothesis, J can be obtained from $I_{k+1}$ by finite steps of moves. It follows that J can be obtained from I by finite steps of moves.
Lemma 4.4. Let $G = (S, E)$ be a connected graph. Suppose $I \subseteq S$ is a subset such that $\lvert I \rvert \ge 2$ , and for any $t \in S \setminus I$ , either one of the following conditions is satisfied:
-
(1) for any $r \in I$ , $\{r,t\}$ is not an edge;
-
(2) there exist at least two vertices $r,r^\prime \in I$ such that $\{r,t\}, \{r^\prime ,t\} \in E$ .
Then there exists $s \in I$ such that the subgraph $G(S \setminus \{s\})$ is connected.
Proof. For two vertices $r,t \in S$ , define the distance $\operatorname {d}(r,t)$ in G to be
Let $m = \max \{\operatorname {d}(r,t) \mid r,t \in I\}$ , and suppose $s, s^\prime \in I$ such that $\operatorname {d}(s, s^\prime ) = m$ .
We claim first that
Otherwise, any path in G connecting r and $s^\prime $ (note that G is connected) must pass through s. It follows that $\operatorname {d}(r, s^\prime )> \operatorname {d}(s, s^\prime ) = m$ , which contradicts our choice of m.
Next we claim that
and therefore, $G(S \setminus \{s\})$ is connected.
In (4.2), the case where $t \in I \setminus \{s\}$ has been settled in the claim (4.1). Thus, we may assume $t \in S \setminus I$ . Let $(t = t_0, t_1, \dots , t_p = s)$ be a path of minimal length in G connecting t and s. Then, $t_1, \dots , t_{p-1} \in S \setminus \{s\}$ . In particular, t is connected to $t_{p-1}$ in $G(S \setminus \{s\})$ . If $t_{p-1} \in I$ , then by the claim (4.1), $t_{p-1}$ is connected to $s^\prime $ in $G(S \setminus \{s\})$ . Thus, t is connected to $s^\prime $ in $G(S \setminus \{s\})$ as desired. If $t_{p-1} \notin I$ , then, since $\{t_{p-1}, s\} \in E$ , there is another vertex $r \in I \setminus \{s\}$ such that $\{t_{p-1},r\} \in E$ . By the claim (4.1) again, r is connected to $s^\prime $ in ${G(S \setminus \{s\})}$ , and so is t. See Figure 1 for an illustration.
Remark 4.5. Applying Lemma 4.4 to the trivial case $I = S$ , we recover the following simple fact (see also the hint of [Reference Bourbaki1, Ch. V, Section 2, Exercise 3(d)]): if $G = (S,E)$ is a connected graph, then there exists $s \in S$ such that $G(S \setminus \{s\})$ is connected.
5 Proof of Theorem 1.2
This section is devoted to proving Theorem 1.2.
By Proposition 3.6, we see that the W-modules $\{\bigwedge ^d V \mid 0 \le d \le n\}$ are pairwise nonisomorphic. Thus, to prove Theorem 1.2, it suffices to show that $\bigwedge ^d V$ is a simple W-module for each fixed d.
By Corollary 3.8, the W-module $\bigwedge ^d V$ is semisimple. Therefore, the problem reduces to proving
Let $S := \{1, \dots , k\}$ and $E := \{\{i, j \} \mid s_i \cdot \alpha _j \ne \alpha _j\}$ . Then $G = (S, E)$ is a graph in the sense of Section 4.
Claim 5.1. G is a connected graph.
Proof. Otherwise, suppose $S = I \sqcup J$ such that for any $i \in I$ and $j \in J$ , $\{i, j\}$ is never an edge, that is, $s_j \cdot \alpha _i = \alpha _i$ . If $V = V_I :=\langle \alpha _i \mid i \in I \rangle $ , then for any $j \in J$ , $\alpha _j$ is a linear combination of $\{\alpha _i \mid i \in I\}$ . It follows that $s_j \cdot \alpha _j = \alpha _j$ , which is absurd. Therefore, ${V_I \ne V}$ .
By Lemma 2.2, $s_i \cdot V_I \subseteq V_I$ for any $i \in I$ . However, for any $j \in J$ , $s_j$ acts trivially on $V_I$ . Since W is generated by $\{s_1, \dots , s_k\}$ , $V_I$ is closed under the action of W, that is, $V_I$ is a proper submodule. This contradicts the assumption that V is a simple W-module.
Claim 5.2. V is spanned by $\{\alpha _1, \dots , \alpha _k\}$ . In particular, $n \le k$ (where $n = \dim V$ ).
Proof. Let $U = \langle \alpha _1, \dots , \alpha _k \rangle \subseteq V$ . By Lemma 2.2, $s_i \cdot U \subseteq U$ for any $i \in S$ . Thus, U is a W-submodule. However, V is a simple W-module. So $U = V$ .
Claim 5.3. There exists a subset $I \subseteq S$ , such that:
-
(1) $\{\alpha _i \mid i \in I\}$ is a basis of V;
-
(2) the subgraph $G(I)$ (see Definition 4.1) is connected.
Proof. Suppose we have found a subset $J \subseteq S$ such that:
-
(a) V is spanned by $\{\alpha _i \mid i \in J\}$ ;
-
(b) the subgraph $G(J)$ is connected.
For example, S itself is such a subset by Claims 5.1 and 5.2. If the vectors ${\{\alpha _i \mid i\in J\}}$ are linearly independent, then we are done.
Now suppose $\{\alpha _i \mid i\in J\}$ are linearly dependent. By a permutation of indices, we may assume $J = \{1, \dots , h\}$ , $h \le k$ , and
If there exists $j \in J$ such that $j \ge l+1$ and $s_j \cdot \alpha _i \ne \alpha _i$ for some $i \le l$ , then
Here, $\widehat {c_i \alpha _i}$ means this term is omitted. Thus, there is an index $i^\prime $ with $i^\prime \le l$ and $i^\prime \ne i$ such that $s_j \cdot \alpha _{i^\prime } \ne \alpha _{i^\prime }$ . In other words, if $l+1 \le j \le h$ , then one of the following is satisfied:
-
(1) for any $i \le l$ , $\{i,j\}$ is never an edge;
-
(2) there exist at least two indices $i, i^\prime \le l$ such that $\{i,j\}, \{i^\prime , j\} \in E$ .
Applying Lemma 4.4 to the subset $\{1, \dots , l\} \subseteq J$ , we see that there is an index $i_0 \le l$ such that the subgraph $G(J \setminus \{i_0\})$ is connected. Moreover, V is spanned by ${\{\alpha _i \mid i \in J \setminus \{i_0\}\}}$ by our assumption (5.2). Thus, $J \setminus \{i_0\}$ satisfies the conditions (a) and (b), and $J \setminus \{i_0\}$ has a smaller cardinality than J.
Apply the arguments above recursively. Finally, we will obtain a subset $I \subseteq S$ as claimed.
Now suppose $I = \{1, \dots , n\} \subseteq S\ (n = \dim V)$ is the subset obtained in Claim 5.3. By Lemma 3.1, $\bigwedge ^d V$ has a basis
Note that for any such basis vector $\alpha _{i_1} \wedge \dots \wedge \alpha _{i_d}$ , the vectors $\alpha _{i_1}, \dots , \alpha _{i_d}$ of V are linearly independent.
Claim 5.4. For any indices $1 \le i_1 < \dots < i_d \le n$ , the subspace $\bigcap _{1 \le j \le d} V_{d, s_{i_j}}^-$ is one-dimensional with a basis vector $\alpha _{i_1} \wedge \dots \wedge \alpha _{i_d}$ (the subspace $V_{d,s_i}^-$ is defined in (3.1)).
Proof. Apply Proposition 3.5(2) to $s_{i_1}, \dots , s_{i_d} \in W$ .
Now suppose $\varphi : \bigwedge ^d V \to \bigwedge ^d V$ is an endomorphism of the W-module. For any ${i \in I}$ and any $v \in V_{d,s_i}^-$ , we have $s_i \cdot \varphi (v) = \varphi (s_i \cdot v) = \lambda _i \varphi (v)$ . Thus, $\varphi (v) \in V_{d,s_i}^-$ . Therefore, for any indices $1 \le i_1 < \dots < i_d \le n$ , we have $\varphi (\bigcap _{1 \le j \le d} V_{d, s_{i_j}}^-) \subseteq \bigcap _{1 \le j \le d} V_{d, s_{i_j}}^-$ . By Claim 5.4,
To prove the statement (5.1), it suffices to show that the coefficients $\gamma _{i_1, \dots , i_d}$ are constant among all choices of $i_1, \dots , i_d$ . We may assume $d \le n-1$ .
Claim 5.5. Let $I_1 = \{1 \le i_1 < \dots < i_d \le n\}, I_2 = \{1 \le j_1 < \dots < j_d \le n\}$ be two subsets of I. Suppose $I_2$ can be obtained from $I_1$ by a move (see Definition 4.2) in the graph $G(I)$ . Then, $\gamma _{i_1, \dots , i_d} = \gamma _{j_1, \dots , j_d}$ .
Proof. To simplify notation, we assume $I_1 = \{1, \dots , d\}$ , $I_2 = (I_1 \setminus \{d\}) \cup \{d+1\}$ and $\{d, d+1\} \in E$ is an edge. In view of Lemma 2.2, for $i = 1, \dots , d$ , we assume
Then, $c_d \ne 0$ . We have
Hence,
and also equals
Note that $c_d \ne 0$ , and that the vectors involved in the equation above are linearly independent. Thus, we have $\gamma _{1,\dots , d} = \gamma _{1,\dots , d-1, d+1}$ which is what we want.
Now apply Lemma 4.3 to the connected graph $G(I)$ . Then by Claim 5.5, we see that the coefficients $\gamma _{i_1, \dots , i_d}$ are constant among all choices of $i_1, \dots , i_d \in I$ . As we have pointed out, this means that the statement (5.1) is valid.
The proof is completed.
6 Some other results
Lemma 6.1. Let $H_1, \dots , H_k \subseteq V$ be linear subspaces of a vector space V. Regard $\bigwedge ^d H_i$ as a subspace of $\bigwedge ^d V$ for $0 \le d \le n$ . Then, $\bigcap _{1 \le i \le k} (\bigwedge ^d H_i) = \bigwedge ^d (\bigcap _{1 \le i \le k} H_i)$ .
Proof. We do induction on k and begin with the case $k = 2$ . Let $I_0$ be a basis of $H_1 \cap H_2$ . Extend $I_0$ to a basis of $H_1$ , say, $I_0 \sqcup I_1$ , and to a basis of $H_2$ , say, $I_0 \sqcup I_2$ . Then $I_0 \sqcup I_1 \sqcup I_2$ is a basis of $H_1 + H_2$ . Further, extend $I_0 \sqcup I_1 \sqcup I_2$ to a basis of V, say, $I_0 \sqcup I_1 \sqcup I_2 \sqcup I_3$ .
We define a total order $\le $ on the set of vectors $I_0 \sqcup I_1 \sqcup I_2 \sqcup I_3$ , and we write $v_1 < v_2$ if $v_1 \le v_2$ and $v_1 \ne v_2$ . By Lemma 3.1, B, $B_1$ , $B_2$ , $B_0$ are bases of $\bigwedge ^d V$ , $\bigwedge ^d H_1$ , $\bigwedge ^d H_2$ , $\bigwedge ^d (H_1 \cap H_2)$ , respectively, where
(The sets $B_1, B_2, B_0$ may be empty.) Moreover, $B_1, B_2 \subseteq B$ , and $B_0 = B_1 \cap B_2$ . Apply Lemma 3.4 to the vector space $\bigwedge ^d V$ . We obtain
For $k \ge 3$ , by the induction hypothesis,
as desired.
The following proposition, which is derived from Lemma 6.1, recovers [Reference Steinberg9, Section 14.2] in a more general context.
Proposition 6.2. Let $\rho :W \to \operatorname {GL}(V)$ be a finite-dimensional representation of a group W. Suppose $s_1, \dots , s_k \in W$ such that, for each i, $s_i$ acts on V by a reflection with reflection hyperplane $H_i$ . Then, $\{v \in \bigwedge ^d V \mid s_i \cdot v = v, \mbox { for all } i\} = \bigcap _{1 \le i \le k} V_{d,s_i}^+ = \bigwedge ^d (\bigcap _{1 \le i \le k} H_i)$ for $0 \le d \le n$ .
Proof. By Corollary 3.3(1), $V_{d,s_i}^+ = \bigwedge ^d H_i$ for all i. Therefore,
The last equality follows from Lemma 6.1.
The next result is a Poincaré-like duality on exterior powers of a representation.
Proposition 6.3. Let $\rho : W \to \operatorname {GL}(V)$ be an n-dimensional representation of a group W. Then, $\bigwedge ^{n-d} V \simeq (\bigwedge ^d V)^* \bigotimes (\det \circ \rho )$ as W-modules for all $d = 0,1, \dots , n$ . Here we denote by $(\bigwedge ^d V)^*$ the dual representation of $\bigwedge ^d V$ .
Proof. Fix an identification of linear spaces $\bigwedge ^n V \simeq \mathbb {F}$ . For any $d = 0, 1, \dots , n$ , we define a bilinear map
Clearly, this induces an isomorphism of linear spaces
Note that for any $w \in W$ , $u \in \bigwedge ^{n-d} V$ , $v \in \bigwedge ^{d} V$ , we have
Therefore, for $w \in W$ , $u \in \bigwedge ^{n-d} V$ ,
This implies $u \mapsto f(u,-) \otimes 1$ is an isomorphism $\bigwedge ^{n-d} V \xrightarrow {\sim } (\bigwedge ^d V)^* \bigotimes (\det \circ \rho )$ of W-modules.
7 Further questions
Question 7.1. Can we remove the technical condition (4) in the statement of Theorem 1.2?
Question 7.2. Is it possible to find two nonisomorphic simple W-modules $V_1, V_2$ satisfying the conditions of Theorem 1.2, and two integers $d_1, d_2$ with $0 < d_i < \dim V_i$ , such that $\bigwedge ^{d_1} V_1 \simeq \bigwedge ^{d_2} V_2$ as W-modules?
If the answer to this question is negative for reflection representations of Coxeter groups, then the irreducible representations obtained in the way described in Section 1 are nonisomorphic to each other.
Question 7.3. What kinds of simple W-modules V have the property that the modules $\bigwedge ^d V$ , $0 \le d \le \dim V$ , are simple and pairwise nonisomorphic? Can we formulate any other sufficient conditions (in addition to Theorem 1.2) or any necessary conditions?
Acknowledgements
The author is deeply grateful to Professors Ming Fang and Tao Gui for enlightening discussions. The author would also like to thank the anonymous referee for suggestions, which helped to improve this paper.