Part III Representation Theory Lectured by Stuart Martin, Lent Term 2013 Please send any comments and corrections to [email protected]. Contents 0 Preliminaries 2 1 Semisimple algebras 4 2 Irreducible modules for the symmetric group 8 3 Standard basis of Specht modules 12 4 Character formula 14 5 The hook length formula 21 6 Multilinear algebra and algebraic geometry 24 Interlude: some reminders about affine varieties 29 7 Schur–Weyl duality 31 8 Tensor decomposition 34 9 Polynomial and rational representations of GL(V ) 37 Interlude: some reminders about characters of GLn (C) 40 10 Weyl character formula 41 11 Introduction to invariant theory and first examples 44 12 First fundamental theorem of invariant theory 50 Example sheet 1 53 Example sheet 2 57 Last updated Thursday 21 March 2013 1 0 Preliminaries In this course we will study representations of (a) (finite) symmetric groups (b) (infinite) general linear groups all over C, and apply the theory to “classical” invariant theory. We require no previous knowledge, except for some commutative algebra, ordinary representation theory and algebraic geometry, as outlined below. Commutative algebra References: Part III course, Atiyah–Macdonald. It is assumed that you will know about rings, modules, homomorphisms, quotients and chain conditions (the ascending chain condition, ACC, and descending chain condition, DCC). Recall: a chain, or filtration, of submodules of a module M is a sequence (Mj : 0 ≤ j ≤ n) such that M = M0 > M1 > · · · > Mn > 0 The length of the chain is n, the number of links. A composition series is a maximal chain; equivalently, each quotient Mj−1 /Mj is irreducible. Results: 1. Suppose M has a composition series of length n. Then every composition series of M has length n, and every chain in M can be extended to a composition series. 2. M has a composition series if and only if it satisfies ACC and DCC. Modules satisfying both ACC and DCC are called modules of finite length. By (1), all composition series of M have the same length, which we denote by `(M ) and call the length of M . 3. (Jordan–Hölder) If (Mi ) and (Mi0 ) are composition series then there exists a one-to-one cor0 respondence between the set of quotients (Mi−1 /Mi ) and the set of quotients (Mi−1 /Mi0 ) such that the corresponding quotients are isomorphic. Remark. For a vector space V over a field K, the following are equivalent: (i) finite dimension (ii) finite length (iii) ACC (iv) DCC and if these conditions are satisfied then `(V ) = dim(V ). (Ordinary) representation theory (of finite groups) References: Part II course, James–Liebeck, Curtis–Riener Vol. 1, §1 below. Work over C or any field K where char(K) - |G|. Let V be a finite-dimensional vector space. 2 GL(V ) is the general linear group of K-linear automorphisms of V . If dimK (V ) = n, choose a basis e1 , . . . , en of V over K to identify it with K n . Then θ ∈ GL(V ) ↔ Aθ = (aij ) ∈ Mn (K), where θ(ej ) = n X aij ei i=1 and, in fact, Aθ ∈ GLn (K). This gives rise to a group isomorphism GL(V ) ∼ = GLn (K). Given G and V , a representation of G on V is a homomorphism ρ = ρV : G → GL(V ) G acts linearly on V if there is a linear action G × V → V , given by (g, v) 7→ g · v, where (a) (action) (g1 g2 ) · v = g1 · (g2 · v) and e · v = v (b) (linearity) g · (v1 + v2 ) = g · v1 + g · v2 , g · (λv) = λ(g · v) If G acts linearly on V then the map G → GL(V ) given by g 7→ ρg , where ρg (v) = g · v, defines a representation of G on V . Conversely, given a representation ρ : G → GL(V ) we have a linear action of G on V given by g · v = ρ(g)v. In this case, we say V is a G-space or G-module. By linear extension, V is a KG-module, where KG is the group algebra – more to come later. For finite groups you should know If S is an irreducible KG-module then S is a composition factor of KG (Maschke’s theorem) If G is finite and char(K) - G, then ever KG-module is completely reducible; or equivalently, every submodule is a summand. The number of inequivalent (ordinary) irreducible representations of G is equal to the number of conjugacy classes of G. If S is an irreducible CG-module and M is any CG-module, then the number of composition factors of M isomorphic to S is equal to dim HomCG (S, M ). Algebraic geometry References: Part II course, Reid, §6 below. You should know about affine varieties, polynomial functions, the Zariski topology (in particular, Zariski denseness) and the Nullstellensatz. 3 1 Semisimple algebras Conventions All rings are associative and have an identity, which we denote by 1 or 1R . Modules are always left R-modules. (1.1) Definition A C-algebra is a ring R which is also a C-vector space, whose notions of addition coincide, and for which λ(rr0 ) = (λr)r0 = r(λr0 ) for all r, r0 ∈ R and λ ∈ C. There are obvious notions of subalgebras and algebra homomorphisms. Usually, our algebras will be finite-dimensional over C. (1.2) Examples and remarks (a) C is a C-algebra, and Mn (C) is a C-algebra. P If G is a finite group thenPthe group P algebraPCG = { g αg g : αg ∈ C} is a C-algebra with pointwise addition: g αg g + g βg g = g (αg + βg )g P P 0 multiplication: g g hh0 =g αh βh Recall the correspondence between representations of G over C and finite-dimensional CG-modules. (b) If R, S are C-algebras then the tensor product R ⊗ S = R ⊗C S is a C-algebra. [More on ⊗ can be found in Atiyah–MacDonald p. 30.] (c) If 1R = 0R then R = {0R } is the zero ring. Otherwise, for λ, µ ∈ C, we have λ1R = µ1R if and only if λ = µ, so we can identify λ ∈ C with λ1R ∈ R. With the identification C ,→ R we can view C as a subalgebra of R. (d) An R-module M becomes a C-space via λm = (λ1R )m for λ ∈ C, m ∈ M . Given any R-module N , HomR (M, N ) is a C-subspace of HomC (M, N ). In particular, it is a C-space, with structure (λϕ)(m) = λϕ(m) = ϕ(λm) ∀λ ∈ C, m ∈ M, φ ∈ HomR (M, N ) (e) If M be an R-module. Then EndR (M ) = HomR (M, M ) is a C-algebra with multiplication given by composition of maps: (ϕψ)(m) = ϕ(ψ(m)) ∀m ∈ M, φ, ψ ∈ EndR (M ) In particular, if V is a C-space then EndC (V ) is an algebra. Recall that EndC (V ) ∼ = Mn (C), where n = dimC (V ). (f) Let R be a C-algebra and X ⊆ R be a subset. Define the centraliser of X in R by cR (X) = {r ∈ R : rx = xr ∀x ∈ X} It is a subalgebra of R. The centre of R is cR (R) = Z(R). (g) Let R be C-algebra and M be a R-module. Then the map αR → EndC (M ) given by α(r)(m) = rm for r ∈ R, m ∈ M is a map of C-algebras. Now, EndR (M ) = cEndC (M ) (αR), and so αR ⊆ cEndC (M ) (EndR (M )). (h) An R-module M is itself naturally an EndR (M )-module, where the action is evaluation, and the action commutes with that of R. The inclusion α as in (g) says that elements of R act as EndR (M )-module endomorphisms of M . Given an R-module N , HomR (M, N ) is also an EndR (M )-module, where the action is composition of maps. 4 (1.3) Lemma Let R be a finite-dimensional C-algebra. Then there are only finitely many isomorphism classes of irreducible R-modules. Moreover, all the irreducible R-modules are finite-dimensional. Proof. Let S be an irreducible R-module and pick 0 6= x ∈ S. Define a map R → S by r 7→ rx. This map has nonzero image, so its image is S by irreducibility. In particular, dimC (S) ≤ dimC (R) < ∞ But S must occur in any composition series of R, so by the Jordan–Hölder theorem, there are only finitely many isomorphism classes of irreducible modules. (1.4) Lemma (Schur’s lemma) Let R be a finite-dimensional C-algebra. Then (a) If S T are nonisomorphic irreducible R-modules then HomR (S, T ) = 0. (b) If S is an irreducible R-module then EndR (S) ∼ = C. Proof (a) If f : R → S is a map of algebras then ker(f ) ≤ R and im(f ) ≤ S. By irreducibility, ker(f ) = 0 or R and im(f ) = 0 or S. If ker(f ) = 0 then R ∼ = im(f ) = S, contradicting nonisomorphism; so we must have ker(f ) = R and im(f ) = 0, i.e. f = 0. (b) Clearly D = EndR (S) is a division ring. Since S is finite-dimensional, so is D. Thus, if d ∈ D, then the elements 1, d, d2 , . . . are linearly dependent. Let p ∈ C[X] be a nonzero polynomial with p(d) = 0. Since C is algebraically closed, p factors as p(X) = c(X − a1 ) · · · (X − an ) for some 0 6= c ∈ C and ai ∈ C. Hence (d − a1 1D ) · · · (d − an 1D ) = 0 But, being a division ring, D has no zero divisors, so one of the terms must be zero. Thus d = aj 1D ∈ C1D . But d ∈ D was arbitrary, so D = C1D ∼ = C. (1.5) Definition An R-module is semisimple (or completely reducible) if it is a direct sum of irreducible submodules. A finite-dimensional C-algebra R is semisimple if it is so as an R-module. (1.6) Proposition (a) Every submodule of a semisimple module is semisimple, and every such submodule is a direct summand. (b) Every quotient of a semisimple module is semisimple. (c) Direct sums of semisimple modules are semisimple. In fact, M is semisimple if and only if every submodule of M is a direct summand. The proofs of this and of (1.6) are left as an exercise. Remarks 1. R a semisimple C-algebra ⇒ any R-module is semisimple. 2. G a finite group ⇒ CG is semisimple. (This is precisely Maschke’s theorem.) (1.7) Definition If R and S are C-algebras, M is an R-module and N is an S-module, then M ⊗ N has the structure of an R-module via r(m ⊗ n) = rm ⊗ n, r ∈ R, m ∈ M, n ∈ N 5 and the structure of an S-module via s(m ⊗ n) = m ⊗ sn, s ∈ S, m ∈ M, n ∈ N and these actions commute: r(s(m ⊗ n)) = r(m ⊗ sn) = rm ⊗ sn = s(rm ⊗ n) = s(r(m ⊗ n)) hence the images of R, S in EndC (M ⊗ N ) commute. If N has basis {e1 , . . . , ek } then the map M ⊕k → M × N given by (m1 , . . . , mk ) 7→ m1 ⊗ e1 + · · · + mk ⊗ ek is an isomorphism of R-modules. Similarly, if M has basis {f1 , . . . , f` } then the map N ⊕` → M ⊗ N given by (n1 , . . . , n` ) 7→ n1 ⊗ f1 + · · · + n` ⊗ f` is an isomorphism of S-modules. (1.8) Lemma Let M be a semisimple R-module. Then the evaluation map M S ⊗ HomR (S, M ) → M S is an isomorphism of R-modules and of EndR (M )-modules. Here S runs over a complete set of nonisomorphic irreducible R-modules and we use the action of R on S and of EndR (M ) on HomR (S, M ). Proof Check that the map is a map of R-modules and of EndR (M )-modules. It’s an isomorphism of vector spaces, by observing that we can reduce to the case where M is irreducible and apply Schur’s lemma (1.4). (1.9) Lemma Let M be a finite-dimensional semisimple R-module. Then Y EndC HomR (S, M ) EndR (M ) ∼ = S where S runs over a complete set of nonisomorphic irreducible R-modules. Q Proof Observe that the product E = S EndC HomR (S, M ) acts naturally on S ⊗ HomR (S, M ) and, since this action commutes with that of R, there exists a homomorphism E → EndR (M ), which is injective. Now count dimensions by (1.8), M is isomorphic to a direct sum of dimC HomR (S, M ) copies of each irreducible module S, so by (1.4), P dimC EndR (S, M ) = S (dimC HomR (S, M ))2 = dimC E (1.10) Theorem (Artin–Wedderburn Theorem) Any finite-dimensional semisimple C-algebra is isomorphic to a product k Y R= EndC (Vi ) i=1 where Vi are finite-dimensional vector spaces. Conversely, if R has this form then it is semisimple, the nonzero Vi s form a complete set of nonisomorphic irreducible R-modules and, as an R-module, R is isomorphic to a direct sum of dimC Vi copies of each Vi . Proof We’ll take this in steps. Step 1 (Deal with the product) Without loss of generality, assume each Vi is nonzero. The Vi are naturally R-modules, with the factors other than EndC (Vi ) acting as 0. 6 Now EndC , and hence R, has transitively on Vi − {0}, thus the Vi are irreducible R-modules. If Vi has basis {ei1 , . . . , eimi } then the map R → V1 ⊕ · · · ⊕ V1 ⊕ · · · ⊕ Vk ⊕ · · · ⊕ Vk | {z } | {z } m1 copies given by mk copies (f1 , . . . , fk ) 7→ (f1 (e11 ), . . . , f1 (e1m1 ), . . . , fk (ek1 ), . . . , fk (ekmk )) is an injective map of R-modules, and hence is an isomorphism by counting dimensions. Thus R is semisimple. The Vi form a complete set of irreducible R-modules (by Jordan–Höldeer, as in (1.3)), and they are nonisomorphic: if i 6= j then (0, . . . , 0, 1, 0, . . . , 0) annihilates Vj but not Vi . Step 2 Let R be any ring. Now the natural map α : R → EndEndR (R) (R) sending r to the map x 7→ rx is an isomorphism. It’s injective, and if θ ∈ EndEndR (R) (R) then it commutes with the endomorphisms αr ∈ EndR (R), where αr (x) = xr. Now if r ∈ R then θ(r) = θ(1r) = θ(αr (1)) = αr (θ(1)) = θ(1)r so θ acts by left-multiplication. Hence θ = α(θ(1)) ∈ α(R), so α is surjective. Assume R is a semisimple C-algebra. Then EndR (R) is semisimple by (1.9), and so R is a semisimple EndR (R)-module. By (1.9) and the isomorphism α, R has the required form. (1.11) Lemma If R is semisimple and M is a finite-dimensional R-module then the natural map α : R → EndEndR (M ) (M ) sending r to the map m 7→ rm is surjective. Proof Let I = M ◦ = {r ∈ R : rM = 0} be the annihilator of M ; then ker α = I. Now, R/I is semisimple and M is an R/I-module, and EndR (M ) = EndR/I (M ). So we can replace R by R/I and suppose that M is faithful (and α is injective). Q By (1.9), EndR (M ) = S EndC HomR (S, M ) and, since M is faithful, all the spaces HomR (S, M ) are nonzero because they L are precisely the simple EndR (M )-modules. By (1.8), M is isomorphic (as an EndR (M )-module) to S S ⊗ HomR (S, M ), so it’s the direct sum of dimC S copies of the simple module HomR (S, M ) for each S. As in (1.9), this implies that dimC EndEndR (M ) (M ) = X (dimC S)2 S But this is equal to dimC R, so α is an isomorphism. Finally, it’s worth noting: (1.12) Lemma Let R be a C-algebra and let e ∈ R be ‘almost idempotent’ in the sense that e2 = λe for some 0 6= λ ∈ C. Then for any R-module M , we have an isomorphism of EndR (M )-modules HomR (Re, M ) ∼ = eM Proof eM . eM is an EndR (M )-submodule of M : if ϕ ∈ EndR (M ) and em ∈ eM then ϕ(em) = eϕ(m) ∈ Replacing e by e λ, we may suppose that e is idempotent. Now ther exists a map of EndR (M )-modules Hom(Re, M ) → eM given by ϕ 7→ ϕ(e), whose inverse sends m to the map r 7→ rm. 7 Chapter I: Representation theory of the symmetric group 2 Irreducible modules for the symmetric group Recall the correspondence between representations ρ : G → GL(V ) and CG-modules given by setting g · v = ρ(g)v for g ∈ G and v ∈ V . The trivial representation G → C× = GL(C) is given by g 7→ 1, and the corresponding CG-module is C, sometimes written CG to emphasise the context. If X ⊆ {1, . . . , n}, write SX for the subgroup of Sn fixing every number outside of X. Let ε : Sn → {±1} be the sign representation, and write Y εσ = 1≤i<j≤n σ(i) − σ(j) i−j Throughout this section, write A = CSn ; this is a finite-dimensional semisimple C-algebra. (2.1) Definition For λ i ∈ N0 , we say λ = (λ1 , λ2 , λ3 , . . . ) is a partition of n, and write λ ` n, if P λ1 ≥ λ2 ≥ λ3 ≥ · · · and i λi = n. Write (λa1 1 , λa2 2 , . . . , λakk ) to denote (λ1 , . . . , λ1 , λ2 , . . . , λ2 , . . . , λk , . . . , λk , 0, 0, . . . ) | {z } | {z } | {z } a1 with λ1 > λ2 > · · · > λk . a2 ak Partitions of n correspond with conjugacy classes of Sn , e.g. (5, 22 , 1) ↔ (• • • • •)(• •)(•) If λ, µ ` n, write λ < µ if and only if there is some i ∈ N with λj = µj for all j < i and λi < µi . This is the lexicographic (or dictionary) order on partitions. It is a total order on the set of partitions of n. E.g. (5) > (4, 1) > (3, 2) > (3, 12 ) > (22 , 1) > (2, 13 ) > (15 ) (2.2) Definition Partitions λ ` n may be geometrically represented by their Young diagram [λ] = {(i, j) : i ≥ 1, 1 ≤ j ≤ λi } ⊆ N × N If (i, j) ∈ [λ] then it is called a node of [λ]. The k th row (resp. column) of a diagram consists of those nodes whose first (resp. second) coordinate is k. Example If λ = (4, 22 , 1) then [λ] = • • • • • • • • • ← λ1 ← λ2 ← λ3 The height (or length) of λ is the length of the first column of [λ], i.e. max{j : λj 6= 0}, and is denoted by ht(λ). (2.3) Definition A λ-tableau tλ is one of the n! arrays of integers obtained by replacing each node in [λ] by one of the integers 1, . . . , n, with no repeats. E.g. the following are (4, 3, 1)-tableaux t1 = 1 3 8 2 6 4 7 5 , t2 = 4 2 6 5 1 7 8 3 Equivalently, a λ-tableau can be regarded as a bijection [λ] → {1, 2, . . . , n}. 8 Sn acts on the set of [λ]-tableaux, e.g. (1 4 7 8 6)(2 5 3) sends t1 to t2 . Formally, given a tableau tλ and π ∈ Sn , the composition of the functions π and tλ gives a new λ-tableau πtλ , where (πtλ )(x) = π(tλ (x)). Given a partition λ ` n, we want to pick one representative of all the corresponding λ-tableaux. Let t◦λ be the standard λ-tableau, i.e. the tableau numbered in order from left to right and from top to bottom, e.g. 1 2 3 4 5 6 7 ◦ t(5,22 ,1) = 8 9 10 Define the row stabilizer Rt and column stabilizer Ct of a tableau t by σ ∈ Rt σ ∈ Ct ⇔ ⇔ each i ∈ {1, . . . , n} is in the same row of t and σt each i ∈ {1, . . . , n} is in the same column of t and σt For instance, if t = t1 as in the examples after (2.3) above, then Rt1 Ct1 (2.4) Definition = = S{1,2,4,5} × S{3,6,7} × S{8} S{1,3,8} × S{2,6} × S{4,7} × S{1} Let tλ be a λ-tableau. The Young symmetrizer is X h(tλ ) = εc rc ∈ A r∈Rtλ , c∈Ctλ Write hλ = h(t◦λ ). Examples (a) If λ = (n), then h = h(t(n) ) = P σ∈Sn σ, called the symmetrizer in A. Since ρ · h = h for all ρ ∈ Sn , Ah = Ch is the trivial representation. P (b) If λ = (1n ), then h = h(t(1n ) ) = σ∈Sn εσ σ, called the alternizer in A. Since ρ · h = ερ h, Ah = Ch is the sign representation. Goal The left-ideals in A of the form Ahλ for λ ` n (known as Specht modules) form a complete set of nonisomorphic irreducible A-modules: see (2.14) below. (2.5) Lemma If λ ` n, t = tλ is a tableau, σ ∈ Sn . (a) Rt ∩ Ct = 1; (b) The coefficient of 1 in h(tλ ) is 1; (c) Rσ(t) = σRt σ −1 and Cσ(t) = σCt σ −1 ; (d) h(σtλ ) = σh(tλ )σ −1 ; (e) Ah(tλ ) ∼ = Ahλ as A-modules. Proof (a)–(d) are clear. For (e), tλ = σt◦λ for some permutation σ ∈ Sn . Postmultiplication by σ then gives an isomorphism Ah(tλ ) ∼ = Ahλ . The following lemma is sometimes called the basic combinatorial lemma. (2.6) Lemma true: Let λ, µ ` n with λ ≥ µ, and let tλ and tµ be tableaux. Then one of the following is 9 (A) There exist distinct integers i and j occurring in the same row of tλ and the same column of tµ ; (B) λ = µ and tµ = rctλ for some r ∈ Rtλ and c ∈ Ctλ . Proof Suppose (A) is false. If λ1 6= µ1 then λ1 > µ1 , so [µ] has fewer columns than [λ]. Hence two of the numbers in the first row of tλ are in the same column of tµ , so (A) holds, contradicting our assumption. So λ1 = µ1 . Since (A) fails, some c1 ∈ Ctµ forces c1 tµ to have the same members in the first row of tλ . Now ignore the first rows of tλ and c1 tµ . The same argument implies that λ2 = µ2 and there is c2 ∈ Ctµ such that tλ and c2 c1 tµ have the same numbers in each of their first two rows. Continuing in this way we see that λ = µ and there exists c0 ∈ Ctλ such that tλ and c0 tµ have the same numbers in each row. Then rtλ = c0 tµ for some r ∈ Rtλ . Now define c = r−1 (c0 )−1 r ∈ r−1 c0 Ctµ (c0 )−1 r = r−1 Cc0 tµ r = Cc0 tµ = Ctλ Then tµ = rctλ . (2.7) Lemma If σ ∈ Sn cannot be written as rc for any r ∈ Rtλ and c ∈ Ctλ then there exist transpositions u ∈ Rtλ and v ∈ Ctλ such that uσ = σv. Proof Since (2.6)(B) fails for tλ and σtλ , there exist integers i 6= j in the same row of tλ and the same column of σtλ . Take u = (i j) and v = σ −1 uσ. (2.8) Lemma Given a tableau tλ and a ∈ A, the following are equivalent: (1) rac = εc a for all r ∈ Rtλ and c ∈ Ctλ ; (2) a = kh(tλ ) for some k ∈ C. Proof (2)⇒(1) As r0 runs through Rtλ so does rr0 , and as c runs through Ctλ so does c0 c with εc0 c = εc0 εc . Calculation reveals that rh(tλ )c = εc h(tλ ). P (1)⇒(2) Let a = σ∈Sn aσ σ. If σ is not of the form rc then by (2.7) there exist transpositions u ∈ Rtλ and v ∈ Ctλ such that uσv = σ. By assumption, uav = εv a, and the coefficient of σ gives auσv = εv aσ , so aσ = −aσ , and hence aσ = 0. The coefficient of rc in (1) gives a1 = εc arc ∈ C. Thus X X a= arc rc = εc a1 rc = a1 h(tλ ) r,c rc as required (2.9) Lemma Proof If a ∈ A then h(tλ )ah(tλ ) = kh(tλ ) for some k ∈ C. x = h(tλ )ah(tλ ) satisfies (2.8)(1). (2.10) Proposition (a) h(tλ )2 = Define fλ = dimC Ahλ . Then n! h(tλ ) fλ (b) fλ | n! (c) h(tλ )Ah(tλ ) = Ch(tλ ) In particular, fλ h(tλ ) is idempotent. n! Proof 10 (a) Let h = h(tλ ). We already know that h2 = kh for some k ∈ C. Right-multiplication by h induces a linear map ĥ : A → A. For a ∈ A, (ah)h = k(ah), so k̂|Ah is multiplication by k. Extend a basis of Ah to a basis of A. With respect to this basis, kIfλ ∗ ĥ ∼ 0 0 and so tr(ĥ) = kfλ . With respect to the basis Sn of A, ĥ has matrix H whose entries are Hστ = coefficient of σ in τ h But Hστ = 1, so tr(ĥ) = n!. n! Hence k = . fλ (b) The coefficient of 1 in h2 = kh is c1 , c2 ∈ Ctλ such that r1 c1 r2 c2 = 1. P εc1 εc2 ∈ Z where the sum runs over all r1 , r2 ∈ Rtλ and (c) By (2.9), the only other possibility is hAh = 0; but h2 6= 0. (2.11) Proposition Ah(tλ ) is an irreducible A-module. Proof Put h = h(tλ ). Then Ah 6= 0 and A is semisimple, so it is enough to show that Ah is indecomposable. Suppose Ah = U ⊕ V . Then Ch = hAh = hU ⊕ hV , so one of hU and hV is nonzero; without loss of generality, hU 6= 0. Then hU = Ch, so Ah = AhU ⊆ U , so U = Ah and V = 0. (2.12) Lemma If λ > µ are partitions and tλ , tµ are tableaux then h(tµ )Ah(tλ ) = 0. Proof By (2.6) there exist integers in the same row of tλ and the same column of tµ . Let τ ∈ Rtλ ∩ Ctµ be the corresponding transposition. Then h(tµ )h(tλ ) = h(tµ )τ τ h(tλ ) = −h(tµ )h(tλ ) so h(tµ )h(tλ ) = 0. Applying this to (σtλ , tµ ) for σ ∈ Sn gives 0 = h(tµ )h(tλ ) = h(tµ )σh(tλ )σ −1 and so h(tµ )σh(tλ ) = 0. Thus h(tµ )Ah(tλ ) = 0. (2.13) Lemma If λ 6= µ and tλ , tµ are tableaux, then Ah(tλ ) Ah(tµ ). Proof Assume without loss of generality that λ > µ. If there exists an A-module isomorphism f : Ah(tλ ) ∼ = Ah(tµ ) then (2.10) (2.12) f (Ch(tλ )) = f (h(tλ )Ah(tλ )) = h(tλ )f (Ah(tλ )) = h(tλ )Ah(tµ ) = which is a contradiction. (2.14) Theorem A-modules. Proof 0 The left-ideals {Ahλ : λ ` n} form a complete set of nonisomorphic irreducible They are irreducible by (2.11) and nonisomorphic by (2.13) and, since |{λ : λ ` n}| = |{ccls of Sn }| = |{irreducible CSn -modules}| they form a complete set. Further reading Chapter 4 of James’s lecture notes (Springer) – alternative proof using ‘polytabloids’ and certain permutation modules. 11 3 Standard basis of Specht modules We first redefine what we mean by a ‘standard’ tableau. (3.1) Definition t is a standard tableau if the numbers increase along the rows (left to right) and columns (top to bottom) of t. The standard tableaux of shape λ are ordered such that tλ < t0λ if tλ is smaller than t0λ in the first place they differ when you read [λ] ‘like a book’. E.g. the standard (3, 2)-tableaux are 1 4 2 5 3 < 1 3 2 5 4 1 3 < 2 4 5 < 1 2 3 5 4 < 1 2 3 4 5 Let Fλ be the number of standard λ-tableaux. Fλ = fλ ; recall from (2.10) that fλ = dimC Ahλ . X First we show Fλ2 = n!. Goal λ`n Notation λ/µ means there is some m with λ ` m, µ ` m − 1 and [µ] ⊆ [λ] ⊆ N × N. This idea will be used in induction arguments. X (3.2) Lemma Let λ ` m. Then Fλ = Fµ . µ : λ/µ Proof If tλ is standard then t−1 λ ({1, 2, . . . , m − 1}) is the diagram of a partition µ ` m − 1; and tλ |[µ] must be standard. The converse is similar. (3.3) Lemma If λ 6= π are partitions of m then |{ν : ν/λ and ν/π}| = |{τ : λ/τ and π/τ }| ∈ {0, 1} Proof If ν/λ and ν/π then [ν] ⊇ [λ] ∪ [π]. Since [λ] and [π] differ in at least one place, we must in fact have [ν] ⊇ [λ] ∪ [π]. Likewise, if λ/τ and µ/τ then [τ ] ⊇ [λ] ∩ [π] and we must have [τ ] = [λ] ∩ [π]. Now [λ] ∪ [π] and [λ] ∩ [π] are partitions, so ν exists ⇔ |[λ] ∪ [π]| = m + 1 ⇔ |[λ] ∩ [π]| = m − 1 ⇔ τ exists This completes the proof. (3.4) Lemma If λ ` m then (m + 1)Fλ = X Fν . ν : ν/λ Proof by induction on m. The case m = 1 is clear. Suppose the statement holds true for partitions of m − 1. Then X X X (3.2) X Fν = Fπ = |{ν : ν/λ}|Fλ + Fπ ν : ν/λ ν : ν/λ π : ν/π ν,π : ν/λ,π6=λ Using the fact that |{ν : ν/λ}| = |{τ : λ/τ }| + 1 together with (3.3) gives that our expression is equal to X (|{τ : λ/τ }| + 1)Fλ + Fπ τ,π : λ/τ,π/τ,π6=λ = Fλ + X τ,π : λ/τ,π/τ Fπ = Fλ + X τ : λ/τ 12 mFτ = Fλ + mFλ = (m + 1)Fλ as required. X (3.5) Lemma Fλ2 = m! λ`m Proof by induction on m. The case m = 1 is clear. Suppose the statement holds true for partitions of m − 1. Then X λ`m Fλ2 (3.2) = X Fλ Fτ X (3.4) = τ `m−1 λ`m,λ/τ mFτ2 = m · (m − 1)! = m! as required. (3.6) Lemma If tλ > t0λ are standard then h(tλ )h(t0λ ) = 0. Proof It suffices to show that there are integers i 6= j lying in the same row of t0λ and the same column of tλ . For then the corresponding transposition τ = (i j) ∈ Rt0λ ∩ Ctλ gives the result as in the proof of (2.12). It remains to find such integers i and j. Let x ∈ [λ] be the first place where tλ and t0λ first differ. Let t0λ = i and y = (t0λ )−1 (i) ∈ [λ]. Then y must be below and to the left of x since tλ is standard. In particular, x cannot lie in the first column or the last row. Let z ∈ [λ] be in the same row as x and the same column of y, and let j = tλ (z) = t0λ (z) be the common value of tλ and t0λ at z. It is now elementary to verify that i and j satisfy the above assumption. L (3.7) Theorem CSn = CSn h(tλ ), where the direct sum runs over all standard tableaux tλ for all partitions λ ` n. P Proof The sum is direct: suppose we have a nontrivial relation a(tλ )h(tλ ) with a(tλ ) ∈ CSn are not all zero. Choose µ maximal (in the ordering on partitions) such that some a(tµ )h(tλ ) 6= 0, and for this µ pick t0µ minimal (in the ordering on standard µ-tableaux) such that a(t0µ )h(t0µ ) 6= 0. Multiplying the relation on the right by h(t0µ ) gives a(t0µ )h(t0µ )2 = 0 by (3.6) and (2.12), and hence a(t0µ )h(t0µ ) = 0, contradicting our assumption. L L fλ CSn hλ as CSn -modules, It is clear that the CSn h(tλ ) ≤ CSn . By Artin–Wedderburn, CSn ∼ = while in our direct sum there are F copies of Ch . By Jordan–Hölder, F ≤ λ λ λ P 2 P Lfλ . But fλ ≤ Fλ since, by (3.5), Fλ = n! = fλ2 . Hence Fλ = fλ for each λ ` n, and so CSn ≤ CSn h(tλ ). (3.8) Corollary The number of standard λ-tableaux is equal to dimC (CSn hλ ). L h(tλ )M , where the direct (3.9) Lemma Let M be a finite-dimensional CSn -module. Then M ∼ = sum runs over all standard tableaux tλ for all partitions λ ` n. P Proof If m(tλ ) = 0 is a nontrivial relation with m(tλ ) ∈ h(tλ )M then choose µ minimal and t0µ maximal with m(t0µ ) 6= 0. Premultiply the relation by h(t0µ ) to contradict (3.6) and (2.12). Finally, Proof Hλ = hλ as in the proof of (3.7). M ∼ = HomCSn (1.12) h(tλ )M M ∼ = M HomCSn (CSn h(tλ ), M ) CSn h(tλ ), M ∼ = HomCSn (CSn , M ) ∼ =M 13 4 Character formula Given a finite-dimensional CG-algebra M , its character is defined by M →M χM (g) = tr m 7→ gm and χM is a class function G → C, i.e. a function that is constant on the conjugacy classes of G. If α is a conjugacy class in G, write χM (α) for the common value of χM on α. Given λ ` n, denote the character of the irreducible CSn -module CSn hλ by χλ . We will calculate χλ (α) and later use the formula to derive Weyl’s character formula for GLn (C). Now Sn -class α comprises all permutations with fixed cycle type nαn · · · 2α2 1α1 , i.e. having αn n-cycles, . . . , α2 2-cycles and α1 1-cycles. Let nα = |α|. (4.1) Lemma Proof nα = n! 1α1 2α2 . . . nαn · α1 !α2 ! . . . αn ! Any permutation in α is one of the n! of the form n n n z }| { z }| { z }| { (∗)(∗) · · · (∗) (∗∗)(∗∗) · · · (∗∗) · · · (∗ · · · ∗)(∗ · · · ∗) · · · (∗ · · · ∗) | {z }| {z } | {z } α1 α2 αn αn α1 Each such permutation can be represented in 1 . . . n · α1 ! . . . αn ! ways: each of the αk k-cycles can be represented in k ways by cycling their components, and each block of αk k-cycles can be represented in αk ! ways by permuting the position of the factors. (4.2) Theorem (orthogonality Then ( X n! λ µ (a) nα χ (α)χ (β) = 0 ccls α ( X n!/nα (b) χλ (α)χλ (β) = 0 λ`n Proof relations) Let λ, µ ` n and let α and β be conjugacy classes in Sn . if λ = µ if λ = 6 µ if α = β if α = 6 β Every element of Sn is real, i.e. conjugate to its inverse, and so χλ (σ) = χλ (σ −1 ) = χλ (σ) so χλ (σ) ∈ R for all σ ∈ Sn . These relations are then just the usual ones for general finite groups. Notation Given x = (x1 , . . . , xn ) ∈ Cm , `1 , . . . , `m ∈ Z, define ` |x`1 , . . . , x`m | = det(xi j ) Usually `j ≥ 0, in which case this is a homogeneous polynomial of degree m X `j in x1 , . . . , xm . j=1 Example The Vandermonde determinant is given by V (x) = |xm−1 , . . . , 1| (4.3) Lemma Proof V = V (x) = Exercise. Q i<j (xi − xj ) In fact, if `i ≥ 0 for each i then |x`1 , . . . , x`m | is divisible by V . We define the Schur polynomial s` by |x`1 , . . . , x`m | s` (x) = m−1 |x , . . . , 1| 14 This is indeed a polynomial in x1 , . . . , xm , provided that `i ≥ 0. Example s(1,2,...,n) (x) = x1 · · · xm . If xi , yj ∈ C for 1 ≤ i, j ≤ m and xi yj 6= 1 for all i, j, then (4.4) Lemma (Cauchy’s lemma) det 1 1 − xi yj Proof by induction on m. m Y = xm−1 , . . . , 1 y m−1 , . . . , 1 1 1 − xi yj i,j=1 The case m = 1 is clear. Let m > 1. Now 1 xi − x1 yj 1 − = · 1 − xi yj 1 − x1 yj 1 − x1 yj 1 − xi yj So subtracting the first row from each other row in the determinant, one can remove the factor (xi −x1 ) 1 from each row i 6= 1 and from each column. So 1 − x1 yj det 1 1 − xi yj = ! Y m Y 1 (xi − x1 ) 1 − x1 yj i>1 j=1 1 1 y1 1−x y12 y1 det 1−x3 y1 | y2 1−x2 y2 y2 1−x3 y2 .. . .. . y2 1−xn y2 y1 1−xn y1 {z ··· ··· ··· .. . ··· 1 yn 1−x2 yn yn 1−x3 yn .. . yn 1−xn yn =∆ } Now subtract the first column in the matrix above from the other columns and use the fact that y1 yj − y1 1 yj − = · 1 − xi yj 1 − xi y1 1 − xi y1 1 − xi yj This gives 1 ∗ ! m Y Y 1 ∆ = (yj − y1 ) det ∗ . 1 − x y i 1 j>1 i=2 .. ∗ 0 0 1 1−x2 y2 1 1−x3 y2 1 1−x2 y3 1 1−x3 y3 1 1−xn y2 1 1−xn y3 .. . .. . ··· ··· ··· .. . ··· 0 1 1−x2 yn 1 1−x3 yn .. . 1 1−xn yn The result now follows by applying the induction hypothesis to this last determinant. (4.5) Lemma Choose xi , yj ∈ C for 1 ≤ i, j ≤ n with modulus < 1. Then X ` 1 x 1 , . . . , x `m y `1 , . . . , y `m det = 1 − xi yj `1 >`2 >···>`m ≥0 Proof We use the expansion X 1 = (xi yj )` 1 − xi yj `≥0 The determinant is thus X σ∈Sm εσ m Y i=1 1 + xi yσ(i) + (xi yσ(i) )2 + · · · and the monomial x`11 · · · x`mm (`i ∈ N) occurs with coefficient m X Y σ∈Sm i=1 `i yσ(i) = y `1 , . . . , y `m 15 In particular, it this term is zero unless the `i are all distinct. So X 1 det = x`11 . . . x`mm y `1 , . . . , y `m 1 − xi yj `1 ,...,`m distinct X = X `1 >···>`m ≥0 π∈Sm X X = `1 >···>`m ≥0 X = `1 >···>`m ≥0 which is what we needed. ` y π(1) , . . . , y `π(m) ` x1π(1) ` . . . xmπ(m) ` x1π(1) ` . . . xmπ(m) επ π∈Sm ` y 1 , . . . , y `m ! ! ` x 1 , . . . , x`m y `1 , . . . , y `m Notation If λ = (λ1 , . . . , λm ) ∈ Zm , let `i = λi + m − i, i.e. `1 = λ1 + m − 1, . . . , `m = λm . Sometimes in the literature, the `i are denoted by βi and are thus called ‘β-numbers’. Note that (1) λ1 ≥ · · · ≥ λm if and only if `1 > · · · > `m ` x 1 , . . . , x `m P (2) If i λi = n and λi ≥ 0 then m−1 is a polynomial of degree m in the xi s. |x , . . . , 1| Write Λ+ (m, n) for the set of all partitions λ of n into ≤ m parts. (4.6) Definition Given x1 , . . . , xm , y1 , . . . , ym ∈ C, then for any i ∈ N, set si = xi1 + · · · + xim , ti = y1i + · · · + tim These are called the power sums (sometimes referred to in the literature as Newton sums). ` x 1 , . . . , x `m y `1 , . . . , y `m X 1 X αn α1 αn 1 (4.7) Lemma = nα sα 1 . . . sn t1 . . . tn |xm−1 , . . . , 1| |y m−1 , . . . , 1| n! + ccls α λ∈Λ (m,n) Note that both quotients on the left-hand side genuinely are polynomials (in fact, Schur polynomials), so they make sense even if the xi , yi are not distinct. Proof Both sides are polynomials, so without loss of generality we may take all the xi , yi to have modulus < 1. Then n n Y X xi yj (xi yj )2 (xi yj )3 1 = + + + ··· log 1 − xi yj 1 2 3 i,j=1 i,j=1 = s1 t1 s2 t2 s3 t3 + + + ··· 1 2 3 and hence n Y s1 t1 s2 t2 s3 t3 + + + ··· 1 2 3 n ∞ X 1 s1 t1 s2 t2 s3 t3 = + + + ··· n! 1 2 3 n=0 α α X 1 n! s1 t1 1 s2 t2 2 = · ··· n! α1 !α2 ! · · · 1 2 1 = exp 1 − xi yj i,j=1 (∗) where (∗) denotes the sum over all n and all sequences α1 , α2 , . . . of nonnegative integers with only finitely many nonzero terms and α1 + α2 + · · · = n. But then n Y X sα1 sα2 · · · tα1 tα2 · · · 1 1 2 1 2 = α1 2α2 · · · α !α ! · · · 1 − x y 1 i j 1 2 i,j=1 (∗) 16 By (4.4)(Cauchy) and (4.5), we have ` m x 1 , . . . , x`m y `1 , . . . , y `m X sα1 sα2 · · · tα1 tα2 · · · X Y 1 1 2 1 2 = = α1 2α2 · · · α !α ! · · · |xm−1 , . . . , 1| |y m−1 , . . . , 1| 1 − x y 1 1 2 i j i,j=1 `1 >`2 >···>`m ≥0 (∗) so we can equate terms which are of degree n in the xi to obtain the desired result. (4.8) Definition For λ = (λ1 , . . . , λm ) ∈ Zm and α a conjugacy class in Sn , let ψλ (α) be the αn 1 coefficient of the monomial xλ1 1 . . . xλmm in sα 1 . . . sn , i.e. X αn 1 ψλ (α)xλ1 1 . . . xλmm sα 1 . . . sn = λ∈Zm Remarks (1) ψλ is a class function for Sn . (2) ψλ (α) = 0 if any λi is negative or if m X i=1 λi 6= n. (3) ψλ is a symmetric function of the λi s. Notation Set ωλ (α) = X επ ψ(`π(1) +1−m, `π(2) +2−m, ... , `π(m) ) (α) π∈Sm The idea now is to show that ωλ = χλ whenever λ ` n. (4.9) Lemma (Character formula) αn m−1 1 sα x , . . . , 1 = 1 . . . sn X λ∈Λ+ (m,n) ωλ (α) x`1 , . . . , x`m Proof The left-hand side is equal to X X 0 ψλ (α)xλ1 1 . . . xλmm ετ xm−1 τ (1) . . . xτ (m) λ∈Zm τ ∈Sm which, by combining terms, is equal to X X λ τ (1) ετ ψλ (α)xτ (1) +m−1 λ τ (m) . . . xτ (m) λ∈Zm τ ∈Sm Let `i = λτ (i) +m−i—note that this is not our usual convention—then since the ψλ (α) are symmetric, our expression is equal to X X ψ(`1 +1−m, ... , `m ) (α)x`τ1(1) . . . x`τm(m) λ∈Zm τ ∈Sm = X ψ(`1 +1−m, ... , `m ) (α) λ∈Zm ` x 1 , . . . , x`m Whenever the `i are not distinct, the corresponding term gives zero. So setting λi = `i + i − m and using the fact that `1 > · · · > `m if and only if λ1 ≥ · · · ≥ λm , our expression is equal to X X ψ(`π(1) +1−m, ... , `π(m) ) επ x`1 , . . . , x`m λ1 ≥···≥λm π∈Sm The terms for which λ is not a partition of n are zero, for if λm < 0 then `m + π −1 (m) − m < 0. Thus we have the desired equality. 17 Let λ, λ0 ` n be partitions of ≤ m parts. Then ( X n! if λ = λ0 nα ωλ (α)ωλ0 (α) = 0 if λ 6= λ0 ccls α (4.10) Lemma (Orthogonality of ω) Proof X By (4.7) we have λ∈Λ+ (m,n) X ` αn α1 αn m−1 1 x 1 , . . . , x`m y `1 , . . . , y `m = 1 x , . . . , 1 y m−1 , . . . , 1 nα sα 1 . . . sn t1 . . . tn n! ccls α By (4.9) this is equal to X 1 n! ccls α X λ,λ0 ∈Λ+ (m,n) 0 ` 0 nα ωλ (α)ωλ0 (α) x 1 , . . . , x`m y `1 , . . . , y `m where `i come from λi and `0i come from λi in the usual way. 0 0 As λ, λ0 vary, the polynomials x`1 , . . . , x`m and y `1 , . . . , y `m are linearly independent in the polynomial ring C[x1 , . . . , xm , y1 , . . . , ym ], so the result follows. Now that we have trudged through lots of combinatorics we will reintroduce the symmetric group. (4.11) Lemma If λ ∈ Λ+ (n, m) then ψλ (α) is the character of the CSn -module CSn rλ , where X rλ = σ σ∈Rt◦ λ Proof (X-rated) Take the character χ and of CSn rλ and suppose σ ∈ α, where α is a conjugacy class in Sn . Write R = Rt◦λ , and find a coset decomposition Sn = N [ gi R i=1 for the transversal {1 = g1 , g2 , . . . , gN }. (Recall: a transversal is a minimal set of coset representatives.) Then CSn rλ has basis {gi rλ }1≤i≤N , which we now use to calculate the traces. The CSn -action is given by σgi rλ = gj rλ for σ ∈ Sn , where σgi ∈ gj R. Thus χ(α) = {1 ≤ i ≤ N : gi−1 σgi ∈ R} We will now take this and bash it very hard lots of times until enough juice comes out to give us what we want. Qm Now g −1 σg ∈ R if and only if g ∈ gi R, where gi−1 σgi ∈ R, and |R| = 1 λj !, so 1 χ(α) = Qm 1 λj ! {g ∈ Sn : g −1 σg ∈ R} Now since g −1 σg = h−1 σh if and only if gh−1 ∈ CSn (σ), and each value taken by g −1 σg for g ∈ Sn is taken by |CSn | elements. By (4.1) we have 1 χ(α) = Qm 1α1 2α2 . . . nαn α1 !α2 ! . . . αn ! |α ∩ R| 1 λj ! since |CSn (σ)| = n!/nα . Now a permutation τ ∈ α ∩ R restricts to a permutation of the numbers in the ith row of t◦λ . If this restriction involves, say, αij j-cycles, then the αij satisfy αi1 + 2αi2 + · · · + nαin = λi (1 ≤ i ≤ m) (∗) α1j + α2j + · · · + αmj = αj (1 ≤ j ≤ n) 18 The number of permutations in R of this type is λ2 ! λm ! λ1 ! · · · 1α11 . . . nα1n α11 ! . . . α1n ! 1α21 . . . nα2n α21 ! . . . α2n ! 1αm1 . . . nαmn αm1 ! . . . αmn ! and so X χ(α) = α1 ! 1α11 . . . nα1n α11 ! . . . α1n ! ··· αm ! 1αm1 . . . nαmn αm1 ! . . . αmn ! (†) where the sum runs over all αij satisfying (∗). By the multinomial theorem, α1m 11 X nm . . . xm α1 !xα αm !x1αn1 . . . xα m αn 1 1 . . . s = sα · · · n 1 1α11 . . . nα1n α11 ! . . . α1n ! 1αm1 . . . nαmn αm1 ! . . . αmn ! where the sum runs over all αij ∈ N satisfying α1j + α2j + · · · + αmj = αj for all 1 ≤ j ≤ n. So ψλ (α), which is the coefficient of xλ1 1 . . . xλmm , is equal to the right-hand side of (†), and hence it equals χ(α). (4.12) Lemma Let λ, µ ` n with µ ≤ λ. The irreducible module CSn hµ is isomorphic to a submodule of CSn rλ if and only if λ = µ. Proof With µ < λ and σ ∈ Sn , by (2.6) there exist two integers in the same row of t◦λ and the same column of σ −1 t◦µ , so if τ is their transposition then στ σ −1 ∈ Ct◦µ . And hµ σrλ = hµ στ σ −1 στ rλ = −hµ σrλ ⇒ hµ σrλ = 0 Thus 0 = hµ CSn rλ ∼ = HomCSn (CSn hµ , CSn rλ ) by (1.12). Conversely suppose µ = λ. Then hλ rλ X σ∈Ct◦ εσ σ = h2λ 6= 0 λ so 0 6= hλ CSn rλ ∼ = HomCSn (CSn hλ , CSn rλ ). Note that, in general, CSn hλ is not a submodule of CSn rλ . (4.13) Lemma Proof Step 1 If λ ∈ Λ+ (n, m) then ωλ = χλ . We’ll take this in steps. ωλ is a Z-linear combination of ψν with ν ≥ λ and with coefficient of ψλ equal to 1. If π ∈ Sn and µπ is the partition with parts `π(1) + 1 − m, . . . , `π(m) i.e. with parts λi + π −1 (i) − 1. Since ψλ (α) is symmetric in the λi , we can write X ωλ = επ ψµπ π∈Sn If π = 1 then µπ = λ, so we’re done. If π 6= 1, then λ1 +π −1 (1)−1 ≥ λ (with equality if and only if π −1 (1) = 1; and then λ2 +π −1 (2)−2 ≥ λ2 with equality if and only if π −1 (2) = 2; and so on. Thus µπ > λ. / X Step 2 ωλ = kλν χν with kλν ∈ Z, kλλ > 0 and kλν = 0 if ν < λ. ν`n By (4.11) and (4.12), ψλ is an N-linear combination χµ s with µ ≥ λ and with nonzero coefficient of χλ . Thus ωλ is a Z-linear combination of χν s with ν ≥ λ and with positive coefficient of χλ . / 19 Step 3 The result follows. We know by (4.10) that ( n! if λ = µ nα ωλ (α)ωµ (α) = 0 if λ 6= µ ccls α X 2 In the case λ = µ, orthogonality of χλ implies kλν = 1, and thus kλν = 0 if λ 6= µ and kλλ = 1. / X ν`n So we’re done. The kλν s in the above proof are sometimes referred to as Kostka numbers in the literature. At long last, take m ∈ N, x1 , . . . , xm ∈ C arbitrary, `i = λi + m − i for λ ` n, and si = xi1 + · · · + xim . X αn m−1 1 x , . . . , 1 = χλ (α) x`1 , . . . , x`n (4.14) Theorem sα 1 . . . sn λ∈Λ+ (m,n) Proof (4.9) and (4.13). Remarks (1) With m ≥ n we can ensure that the right-hand side of the equation in (4.14) involves all partitions of n. (2) If λ ∈ Λ+ (m, n) then χλ (α) is the coefficient of the monomial x`11 . . . x`mm in the expansion of n s1α1 . . . sα n . 20 5 The hook length formula We already know by (3.8) that the dimension of an irreducible CSn -module is the number of standard tableaux. We give some results which allow for easy calculation of this dimension. If λ ∈ Λ+ (n, m) then Q 1≤i<j≤n (`i − `j ) Qn fλ = deg χλ = n! · i=1 `i ! (5.1) Theorem (Young–Frobenius formula) Proof deg χλ = χλ (1) is the coefficient of x`11 . . . x`mm in the expansion of X τ (1)−1 (x1 + · · · + xm )n ετ x1 . . . xτm(m)−m (x1 + · · · + xm )n xm−1 , . . . , 1 = τ ∈Sn By the multinomial theorem, this coefficient equals X n! (`1 + 1 − τ (1))! . . . (`m + 1 − τ (m))! τ ∈Sn 1 1 1 = n! ,..., , (` − m + 1)! (` − 1)! `! 1 = n! · Qm |. . . , `(` − 1), `, 1| j=1 `j ! m−1 1 ` , . . . , `2 , `, 1 = n! · Qm j=1 `j ! This determinant is a Vandermonde determinant, so applying (4.3) gives the result. (5.2) Definitions (a) Let λ ` n. The (i, j)-hook of λ is the intersection of an infinite Γ–shape having (i, j) as its corner with [λ]. For example, the (2, 2)–hook of (42 , 3) is shown as stars (?) in the following tableau: • • • • • • ? ? ? ? • (b) The (i, j)-hook of [µ] comprises the (i, j)-node along with the (µi − j) nodes to the right of it (the arm) and the (µ0j − i) nodes below it (the leg). [Here, [µ] has column lengths µ01 , µ02 , . . . .] The length of the (i, j)-hook is hij = µi + µ0j + 1 − i − j. (c) Replacing the (i, j)-node of [µ] by hij for each hook graph of (42 , 3) is 6 5 5 4 3 2 node gives the hook graph. For example, the 4 3 1 2 1 (d) A skew hook is a connected part of the rim of [µ] which can be removed to leave a proper diagram. For example, the skew 4-hooks of (42 , 3) are • • • • • • • ? ? ? ? and • • • • • • • ? ? ? ? There is a natural one-to-one correspondence between hooks and skew hooks in a given tableau. For each node (i, j) there is a unique skew hook starting in the ith row and ending in the j th column, and it corresponds with the (i, j)-hook. 21 Theorem (5.3) (hook length formula) [Frame, Robinson & Thrall, 1954] n! fλ = Q (i,j)∈[λ] Proof hij Suppose λ has m parts. By (5.1) we have to show that m Y (`i − `k ) k=i+1 λi Y hij = `i ! for each i j=1 The (combined) product on the left-hand side is a product of λi + m − 1 = `i terms, so it suffices to show that they are precisely 1, 2, . . . , `i in some order. Now `i − `m > `i − `m−1 > · · · > `i − `i+1 λ01 hi1 > hi2 > · · · > hiλi Since λ has m parts and is the length of the first column, λ01 = m, and hi1 = λi . So each term is ≤ `i . Thus it is enough to show that hij 6= `i − `k for any j, k. However if r = λ0j then λr ≥ j and λr+1 < j, so hij − `i + `r+1 = (λi + r − 1 − j + 1) − (λi + m − i) + (λr + m − r) = λr + 1 − j > 0 hij − `i + `r+1 = (λi + r − 1 − j + 1) − (λi + m − i) + (λr+1 + m − r − 1) = λr+1 − j < 0 So `i − `r < hij < `i − `r+1 . This theorem is often referred to as the FRT theorem in the literature. Example When n = 11 and λ = (6, 3, 2), the hook graph is given by 8 4 2 and so fλ = 7 3 1 6 1 5 3 2 1 11! = 990 8.7.5.4.32 .22 .13 Remarks (1) The determinantal form of the Young–Frobenius formula is given by 1 fλ = n! (λi − i + j)! where λ ∈ Λ+ (n, m) and the determinant on the right is of an m × m matrix and we adopt the 1 convention = 0 if m < 0. For a proof see e.g. Sagan or James §19. m! X (2) The equation fλ2 = n! can be proved as a purely combinatorial statement without reference to λ`n representations, using the so-called ‘RSK algorithm’ (for Robinson–Schensted–Knuth). (3) The Murnaghan–Nakayama rule (e.g. example sheet 1, question 9) states that if πρ ∈ Sn with ρ an r-cycle and π a permutation of the remaining n − r numbers then X χλ (πρ) = (−)i χν (π) ν where ν runs over partitions for which [λ] − [ν] is a skew r-hook of leg length i. (Note: leg length refers to the leg of the hook with which the skew hook corresponds.) 22 For example, conj. class 2 χ(5,4 z }| { 2 ) ((5, 4, 3, 1)) = −χ(5,3) + χ(3 ,2) = (22 ) (3,1) − (χ χ | | {z } from (5,3) (1) = −χ +0+0 on (4, 3, 1) +χ {z (2,12 ) from (32 ,2) = −1 23 ) } on (3, 1) on (1) Chapter II: Representation theory of general linear groups 6 Multilinear algebra and algebraic geometry Let V and W be finite-dimensional CG-modules, where G is any (possibly infinite) group. Recall the following: Tensor products: V ⊗C W is a CG-module with the diagonal action g(v ⊗ w) = gv ⊗ gw. Basic properties: – V ⊗C∼ =C – V ⊗W ∼ =W ⊗V ∼ V ⊗ (W ⊗ Z) – (V ⊗ W ) ⊗ Z = – If θ : V → V 0 and ϕ : W → W 0 are CG-maps then there is a CG-map defined by θ⊗ϕ : V ⊗W v⊗w → 7→ V 0 ⊗ W0 θ(v) ⊗ ϕ(w) Hom spaces: HomC (V, W ) is a CG-module via ! X X αi gi (f )(v) = αi gi (f (gi−1 v)) i i Dual space: V ∗ = HomC (V, C). Basic properties: – If V is 1-dimensional then V ∗ ⊗ V ∼ = C. – If θ : V → W is a CG-map then the dual θ∗ : W ∗ → V ∗ is too. – The map V∗⊗W f ⊗w is a CG-isomorphism. → HomC (V, W ) 7→ (v 7→ f (v)w) Tensor powers: The nth tensor power of V is defined by T n V = V ⊗n = V ⊗ · · · ⊗ V , | {z } T 0V = C n copies Basic properties: – If V has basis {ei : 1 ≤ i ≤ m} then V ⊗n has basis {ei1 ⊗ · · · ⊗ ein : 1 ≤ i1 , . . . , in ≤ m} and hence dim(V ⊗n ) = mn . – V ⊗n is naturally a left CSn -module via σ(v1 ⊗ · · · ⊗ vn ) = vσ−1 (1) ⊗ · · · vσ−1 (n) – The actions of CSn and CG commute, i.e. σgx = gσx for all g ∈ G, σ ∈ Sn , x ∈ V ⊗n (6.1) Definition The nth exterior power of V is Λn V = V ⊗n /X, where X = spanC {x − εσ σx : x ∈ V ⊗n , σ ∈ Sn } The image of v1 ⊗ · · · ⊗ vn in Λn V under the quotient map will be denoted by v1 ∧ · · · ∧ vn . Define the antisymmetric tensors by T n Vanti = {x ∈ V ⊗n : σx = εσ x for all σ ∈ Sn } (6.2) Lemma 24 (i) v1 ∧ · · · ∧ vn = εσ vσ−1 (1) ∧ · · · ∧ vσ−1 (n) for all σ ∈ Sn . This tells us, by considering transpositions, that v1 ∧ · · · ∧ vn = 0 whenever vi = vj for some i 6= j. (ii) T n Vanti and Λn V are CG-modules. (iii) Λn V has basis {ei1 ∧ · · · ∧ ein : 1 ≤ i1 < · · · < in ≤ n}, and so dim Λn V = In particular, Λm V is one-dimensional and Λm+1 V = Λm+2 V = · · · = 0. Proof m . n An easy exercise. Remark If V is m-dimensional over a field k then usually we define Λn V = V ⊗n /X, where X = spanC {v1 ⊗ · · · ⊗ vn : vi = vj for some i 6= j} If char k 6= 2 then this definition reduces to (6.2). X εσ σ is the alternizer [cf. (2.4)], and the natural (6.3) Lemma T n Vanti = aV ⊗n , where a = σ∈Sn map T n Vanti → Λn V is an isomorphism of CG-modules. Proof If x is antisymmetric then X X X εσ σx = εσ (εσ x) = x = n! x ax = σ∈Sn σ∈Sn σ∈Sn and so T n Vanti ⊆ aV ⊗n . Conversely, since σa = εσ a for all σ ∈ Sn , any element of aV ⊗n must be antisymmetric. We have a map aV ⊗n ,→ V ⊗n → Λn V which is a CG-homomorphism and whose kernel is X ∩ aV ⊗n ≤ aX, since x = metric. But for y ∈ V ⊗n , 1 ax for x antisymn! a(y − εσ σy) = ay − εσ (εσ a)y = ay − ay = 0 so ax = 0 (since x ∈ X). Hence the map is injective. And it is surjective: any x ∈ Λn V is the image 1 of some y ∈ V ⊗n , but then x is the image of ay ∈ aV ⊗n . So the map is an isomorphism. n! (6.4) Lemma Proof Λn (V ∗ ) ∼ = (Λn V )∗ . The quotient map V ⊗n Λn V is surjective, so induces an inclusion (Λn V )∗ ,→ (T n V )∗ = T n (V ∗ ) Now by the universal property of Λn V (i.e. any multilinear map V × · · · × V → C factors through | {z } n copies Λn V ), the image of the inclusion is T n (V ∗ )anti , which is isomorphic to Λn (V ∗ ). Now prepare for some déjà vu in (6.5)–(6.8). . . (6.5) Definition The nth symmetric power of V is Sn V = V ⊗n /Y , where Y = spanC {x − σx : x ∈ V ⊗n , σ ∈ Sn } The image of v1 ⊗ · · · ⊗ vn in Sn V under the quotient map will be denoted by v1 ∨ · · · ∨ vn . Define the symmetric tensors by T n Vsym = {x ∈ V ⊗n : σx = x for all σ ∈ Sn } (6.6) Lemma 25 (i) v1 ∨ · · · ∨ vn = vσ−1 (1) ∨ · · · ∨ vσ−1 (n) for all σ ∈ Sn . (ii) T n Vsym and Sn V are CG-modules. (iii) Sn V has basis {ei1 ∨ · · · ∨ ein : 1 ≤ i1 ≤ · · · ≤ in ≤ n}, and so dim Sn V = Proof An easy exercise. T n Vsym = sV ⊗n , where s = (6.7) Lemma m+n−1 . n X σ is the symmetrizer [cf. (2.4)], and the natural σ∈Sn map T n Vsym → Sn V is an isomorphism of CG-modules. Proof If x is symmetric then sx = X σx = σ∈Sn and so T n Vsym ⊆ sV ⊗n . X x = n! x σ∈Sn Conversely, since σs = s for all σ ∈ Sn , any element of sV ⊗n must be symmetric. We have a map sV ⊗n ,→ V ⊗n → Sn V which is a CG-homomorphism and whose kernel is Y ∩ aV ⊗n ≤ sY , since x = But for y ∈ V ⊗n , 1 sx for x symmetric. n! s(y − σy) = sy − sy = 0 so sx = 0 (since x ∈ Y ). Hence the map is injective. And it is surjective: any x ∈ Sn V is the image 1 of some y ∈ V ⊗n , but then x is the image of sy ∈ sV ⊗n . So the map is an isomorphism. n! (6.8) Lemma Proof Sn (V ∗ ) ∼ = (Sn V )∗ . The quotient map V ⊗n Sn V is surjective, so induces an inclusion (Sn V )∗ ,→ (T n V )∗ = T n (V ∗ ) Now by the universal property of Sn V (i.e. any multilinear map V × · · · × V → C factors through | {z } n copies Sn V ), the image of the inclusion is T n (V ∗ )sym , which is isomorphic to Sn (V ∗ ). Polynomial maps between vector spaces (6.9) Definition Let V and W be finite-dimensional C-spaces with bases {ei : 1 ≤ i ≤ m} and {fr : 1 ≤ r ≤ h}, respectively. A function ϕ : V → W is a polynomial map (resp. n-homogeneous map) if, for all X1 , . . . , Xm ∈ C, ~ 1 + · · · + ϕh (X)f ~ h ϕ(X1 e1 + · · · + Xm em ) = ϕ1 (X)f ~ abbreviates for some polynomials (resp. homogeneous polynomials of degree n) ϕ1 , . . . , ϕh . [Here X the m-tuple (X1 , . . . , Xm ).] (6.10) Lemma The definitions in (6.9) do not depend on the choice of bases. That is, if there exist such functions ϕi with respect to some bases {ei }, {fr }, then there exist such functions with respect to any bases {e0i }, {fr0 }. Proof Write e0i = m X pji ej and fr = h X s=1 j=1 26 qsr fs0 Then m X ϕ Xi e0i i=1 ! = ϕ m X i,j=1 = h X Xi pji ej m X ϕr r=1 = h X ϕr r,s=1 and the h X r=1 Xi1 p1i1 , . . . , i1 =1 m X Xim pmim im =1 m X Xi1 p1i1 , . . . , i1 =1 m X ! Xim pmim im =1 fr ! qsr fs0 ϕr · · · qsr are polynomials (resp. n-homogeneous polynomials) like the ϕr . Let PC (V, W ) and HC,n (V, W ) denote the spaces of polynomial maps and n-homogeneous maps, respectively, from V to W . They are C-spaces, as can be easily verified. (6.11) Lemma The composite of polynomial maps X → W → Z is a polynomial map X → Z. Likewise, the composite of an n-homogeneous map X → W and a n0 -homogeneous map W → Z is an nn0 -homogeneous map X → Z. Proof An easy exercise. Basic examples (1) HC,0 (V, W ) ∼ =W (2) HC,1 (V, W ) = HomC (V, W ) (3) ∆ ∈ HC,n (V, Sn V ), where ∆ : v 7→ v ∨ · · · ∨ v. For ! X X ∆ Xi ei = Xi1 . . . Xin ei1 ∨ · · · ∨ ein = i ~i X i1 ≤···≤in λ~i Xi1 . . . Xin ei1 ∨ · · · ∨ ein for suitable λ~i ∈ C. [Here ~i abbreviates the n-tuple (i1 , . . . , in ).] (6.12) Theorem For C-spaces V and W , the map ψ 7→ ψ ◦ ∆ induces an isomorphism ∼ = HomC (Sn V, W ) −→ HC,n (V, W ) called restitution. Proof It’s injective: suppose ψ ◦ ∆ = 0. We use descending induction on i to show that ψ(v1 ∨ · · · ∨ vn ) = 0 whenever i of the terms are equal. When i = n this is precisely our assumption. Suppose it’s true for i + 1, then for any α ∈ C we have 0 = ψ((x + αy) ∨ · · · ∨ (x + αy) ∨vi+2 ∨ · · · ∨ vn ) {z } | i+1 = i+1 X j=0 αj i+1 ψ(x · · ∨ x} ∨ y ∨ · · · ∨ y ∨vi+2 ∨ · · · ∨ vn ) | ∨ ·{z | {z } j i+1−j j Since this holds true for any α ∈ C, each term must be zero. In particular i+1 ψ(x · · ∨ x} ∨y ∨ vi+2 ∨ · · · ∨ vn ) = 0 | ∨ ·{z 1 i If dim V = m and dim W = h, say, then m+n−1 dim HomC (Sn V, W ) = h = dim HC,n (V, W ) n 27 So the map is an isomorphism. (6.13) Corollary Sn V = spanC {v ∨ · · · ∨ v : v ∈ V } Proof Take W = C in (6.12). If the elements v ∨ · · · ∨ v do not span then there exists a nonzero map Sn V → C whose composition with ∆ is zero, contradicting (6.12). Remark We can construct the inverse explicitly. Let ϕ ∈ HC,n (V, W ), so ϕ(X1 e1 + · · · + Xm em ) = ϕ1 (X1 , . . . , Xm )f1 + · · · + ϕk (X1 , . . . , Xm )fh with ϕi homogeneous. The total polarisation P ϕ ∈ HomC (Sn V, W ) of ϕ is P ϕ(ei1 ∨ · · · ∨ ein ) = h X ∂ n ϕj fj ∂Xi1 . . . ∂Xin j=1 | {z } ∈C (The partial derivative is formal, is symmetric in the Xik , and lies in C since ϕj has degree N .) Now X (P ϕ)(∆(X1 e1 + · · · + Xm em )) = Xi1 . . . Xin (P ϕ)(ei1 ∨ · · · ∨ ein ) i1 ,...,in = X X j = X Xi1 . . . Xin i1 ,...,in ∂ n ϕj fj ∂Xi1 . . . ∂Xin n! ϕj (X1 , . . . , Xm )fj j = n!ϕ(X1 e1 + · · · + Xm em ) So P ϕ ◦ ∆ = n! ϕ. TheP penultimate equality follows from Euler’s theorem, which states that if F ∂F Xi = rF . is r-homogeneous then i ∂X i Example Let ϕ : V → C be a quadratic form: ! m X X ϕ Xi ei = aij Xi Xj , i=1 aij = aji i,j hP P i P Then (P ϕ) ( i Xi ei ) ∨ Y e = 2 i,j aij Xi Yj is twice the corresponding symmetric bilinear j j J form. 28 Interlude: some reminders about affine varieties Let K be an infinite field. Then R = K[X1 , . . . , Xn ] a Noetherian ring by Hilbert’s basis theorem. Define An K = K n , the n-dimensional affine space over K. If f ∈ R and p = (a1 , . . . , an ) ∈ An then the element f (p) = f (a1 , . . . , an ) ∈ K is the evaluation of f at p. Define a map {ideals I of R} I ←→ 7−→ {subsets of An } V (I) where V (I) = {p ∈ An : f (p) = 0 for all f ∈ I}. A subset X ⊆ An is an algebraic set if X = V (I) for some ideal I. But I is finitely generated, say I = (f1 , . . . , fm ), so V (I) = {p ∈ An : fj (p) = 0 for each 1 ≤ j ≤ m} is a locus of points satisfying a finite number of polynomial equations. Algebraic subsets of An form the closed sets of a topology on An , called the Zariski topology. The Zariski topology on An induces a topology on any algebraic set X ⊆ An : the closed subsets of X are the algebraic subsets of X. Example Zariski-closed subsets of A1 are A1 and finite subsets of A1 ; that is, the Zariski topology on A1 is the cofinite topology. Returning to our correspondence, we have a map in the other direction {ideals I of R} I(X) ←→ ←− {subsets of An } X where I(X) = {f ∈ R : f (p) = 0 for all p ∈ X}. An algebraic set X is irreducible if there does not exist a decomposition X = X1 ∪ X2 with X1 , X2 ( X proper algebraic subsets. Note X is irreducible ⇔ I(X) is a prime ideal If I E R then the radical √ I of I is given by √ I = {f ∈ R : f m ∈ I for somem} √ and I is a radical ideal if I = I. Note that prime ideals are automatically radical. Theorem (Nullstellensatz) Let K be algebraically closed. Then (a) Maximal ideals of R are of the form mj = (X1 − a1 , . . . , Xn − an ) for some p = (a1 , . . . , an ) ∈ An . (b) If (1) 6= I E R then V (I) 6= ∅. √ (c) I(V (I)) = I for all ideals I. Proof Omitted. Let W a finite-dimensional K-space. A map f : W → K is regular (or polynomial) if it is defined by a polynomial in the coordinates with respect to a given basis of W . (Check: it doesn’t matter which basis we pick.) 29 Define the coordinate ring (or ring of regular functions) of W , denoted K[W ], to be the Kalgebra of polynomial functions on W . If e1 , . . . , em is a basis of W and x1 , . . . , xm is the corresponding dual basis of W ∗ then K[W ] = K[x1 , . . . , xm ] We say f ∈ K[W ] is d-homogeneous if f (tw) = td f (w) for all d ∈ C and w ∈ W . Write K[W ]d for the subspace of K[W ] of all d-homogeneous polynomials. Then M K[W ] = K[W ]d d≥0 is a graded K-algebra, i.e. if f, g are homogeneous of degree df , dg , respectively, then f g is homogeneous of degree df + dg . The monomials xd11 · · · xdmm with d1 + · · · + dm = d form a basis of K[W ]d . In particular, K[W ]1 = W ∗ . This extends to a canonical identification of K[W ]d with Sd (W ∗ ), so that M M K= K[W ]d = Sd (W ∗ ) = S(W ∗ ) d≥0 d≥0 We call S(W ∗ ) the symmetric algebra of W ∗ . Zariski-dense subsets A subset X ⊆ W , with W finite-dimensional, is Zariski-dense if every function f ∈ K[W ] vanishing on X is the zero function. So every polynomial function f ∈ K[W ] is completely determined by its restriction f |X to a Zariski-dense subset X ⊆ W . Now let X be an arbitrary subset of W . As above, the ideal of X I(X) = {f ∈ K[W ] : f (a) = 0 for all a ∈ X} = \ a∈X where ma = I({a}) is the maximal ideal of functions vanishing at a, i.e. eva : K[W ] → C ma = ker f 7→ f (a) We say X ⊆ Y is Zariski-dense in Y if I(X) = I(Y ). Further reading Chapter 4 of Reid’s book. 30 ma 7 Schur–Weyl duality Let V be an m-dimensional C-space and let n ∈ N. V ⊗n is a right CSn -module via the action1 (v1 ⊗ · · · ⊗ vn )σ = vσ(1) ⊗ · · · ⊗ vσ(n) So there is a map CSn σ → EndC V ⊗n 7→ (v 7→ vσ) —(1) Regarding V as a representation of GL(V ) in the natural way, V ⊗n is a (left) CGL(V )-module via the action g(v1 ⊗ · · · ⊗ vn ) = gv1 ⊗ · · · ⊗ gvn So there is a map CGL(V ) → EndC V ⊗n g 7→ (v 7→ gv) —(2) The actions (1) and (2) commute, so by abusing notation slightly, (1) induces a map ϕ : CSn → EndCGL(V ) (V ⊗n ) and (2) induces a map ψ : CGL(V ) → EndCSn (V ⊗n ) Schur–Weyl duality asserts that the images of CSn and CGL(V ) in EndC (V ⊗n ) are each other’s centralizers. (7.1) Theorem (Schur–Weyl duality) [Schur, 1927] ψ(CGL(V )) = EndCSn (V ⊗n ) and ϕ(CSn ) = EndCGL(V ) (V ⊗n ) Such an assertion looks innocuous, but it is fundamental in explaining why the symmetric group and the general linear group are so closely related. (7.2) Definition The C-algebra SC (m, n) is the subalgebra of EndC (V ⊗n ) consisting of endomorphisms which commute with the image of CSn , i.e. SC (m, n) = EndCSn (V ⊗n ) It is called the Schur algebra. [For a definition over arbitrary fields, see J.A. Green’s book Polynomial representations of GLn , SLN 830.] CSn is semisimple and V ⊗n is a finite-dimensional CSn -module, so by (1.9), SC (m, n) is a semisimple C-algebra. Put W = EndC (V ). This is a CGL(V )-module under the conjugation action (check this). (7.3) Lemma There is a map α : W ⊗n → EndC (V ⊗n ) given by f1 ⊗ · · · ⊗ fn 7→ v1 ⊗ · · · ⊗ vn 7→ f1 (v1 ) ⊗ · · · ⊗ fn (vn ) which is an isomorphism both of CSn -modules and of CGL(V )-modules. 1 Note that, because this is a right-action, σ1 σ2 denotes the effect of first applying σ1 and then applying σ2 , which is ‘backwards’ compared to normal. 31 Proof W ⊗n = W ⊗ · · · ⊗ W ∼ = (V ⊗ V ∗ ) ⊗ · · · ⊗ (V ⊗ V ∗ ) ∼ (V ⊗ · · · ⊗ V ) ⊗ (V ∗ ⊗ · · · ⊗ V ∗ ) = ∼ V ⊗n ⊗ (V ⊗n )∗ = ∼ = EndC (V ⊗n ) It is left as an exercise to check that this isomorphism is the map we think it is, that W ⊗n has the natural structure of a CSn -module and EndC (V ⊗n ) inherits structure from V ⊗n via θ · σ = ϕ(σ)θϕ(σ −1 ), and finally that α is a CSn -map. (7.4) Lemma SC (m, n) = α(T n Wsym ) Proof SC (m, n) is the set of all x ∈ EndC (V ⊗n ) fixed under the action of Sn . By (6.5), T n Wsym is the set of all y ∈ W ⊗n fixed under the action of Sn (which is induced by α, i.e. y ·σ ∈ W ⊗n corresponds with ϕ(σ)α(y)ϕ(σ −1 ) ∈ EndC (V ⊗n ). Recall affine n-space An from §6. (7.5) Lemma An is irreducible; i.e. if An = X ∪Y with X, Y Zariski-closed then X = An or Y = An . Proof The ring of regular functions on An is R = C[X1 , . . . , Xn ]. If X and Y are zero-sets of ideals I, J E R then, by assumption, any maximal ideal contains either I or J. If I, J are both nonzero then pick 0 6= f ∈ I and 0 6= g ∈ J. Any maximal ideal contains f g, so By Nullstellensatz, f g ∈ p for all a1 , . . . , an ∈ C f g(a1 , . . . , an ) = 0 {0} = {0}, contradicting the fact that R is an integral domain. (7.6) Lemma Let Y ≤ Cd be a subspace. Then Y is Zariski-closed, where Cd is identified with Ad in the canonical way. Proof Chose a basis ϕ1 , . . . , ϕh of HomC (Cd /Y, C) and view them as maps Cd → C. Then Y is the zero set of the ϕi . n (7.7) Lemma Proof n T WSym = spanC z }| { {ϕ ⊗ · · · ⊗ ϕ : ϕ ∈ GL(V )} Define a subspace X of T n Wsym by n X = spanC z }| { {ϕ ⊗ · · · ⊗ ϕ : ϕ ∈ GL(V )} n z }| { There is a map α : W → W given by α(ϕ) = ϕ ⊗ · · · ⊗ ϕ, which is a regular map of the affine 2 2n spaces Am ∼ = W and Am ∼ = W ⊗n . ⊗n Now X is a subspace, so is Zariski-closed by (7.6), and hence α−1 (X) is Zariski-closed. Thus, W = α−1 (X) ∪ {endomorphisms with zero determinant} is a union of Zariski-closed subsets. 2 Since Am is irreducible (by (7.5)), α−1 (X) = W . Thus X contains all maps of the form ϕ ⊗ · · · ⊗ ϕ for ϕ ∈ W . But they span T n Wsym since the images ϕ ∨ · · · ∨ ϕ span Sn W by (6.13). In other words. . . n (7.8) Theorem z }| { SC (m, n) = spanC {ϕ ⊗ · · · ⊗ ϕ : ϕ ∈ GL(V )} The assertion that the image of CGL(V ) is the centralizer of CSn is a reformulation n z }| { of the assertion that SC (m, n) = spanC {ϕ ⊗ · · · ⊗ ϕ : ϕ ∈ GL(V )} , which is (7.8). Proof of (7.1) 32 Conversely, SC (m, n) is a semisimple C-algebra. By (1.11), CSn maps onto EndSC (m,n) (V ⊗n ), and since the image of GL(V ) spans SC (m, n), we must have EndSC (m,n) (V ⊗n ) = EndCGL(V ) (V ⊗n ) Thus CSn maps onto EndCGL(V ) (V ⊗n ), i.e. the image of CSn in EndC (V ⊗n ) is the centralizer of the image of CGL(V ). 33 8 Tensor decomposition Let V be an m-dimensional C-space. We know from (3.9) that L V ⊗n = h(tλ )V ⊗n —(1) where the direct sum runs over standard tableaux tλ with λ ` n. Examples If λ = (1n ) then hλ V ⊗n ∼ = T n Vanti ∼ = Λn V . If λ = (n) then hλ V ⊗n ∼ = T n VSsym ∼ = Sn V . In particular,when n = 2, this tells us that V ⊗V ∼ = T 2 Vanti ⊕ T 2 Vsym ∼ = Λ2 V ⊕ Sn V Now the actions of Sn and GL(V ) on V ⊗n commute. So if λ ` n and tλ is a λ-tableau, then h(tλ )V ⊗n is a CGL(V )-submodule of V ⊗n . Observe that h(tλ )V ⊗n ∼ = hλ V ⊗n as CGL(V )-modules since σh(tλ )σ −1 = hλ for some σ ∈ Sn . So premultiplication by σ induces an isomorphism ∼ = h(tλ )V ⊗n −→ hλ V ⊗n (8.1) Lemma The nonzero hλ V ⊗n with λ ` n are nonisomorphic irreducible CGL(V )-modules. Moreover, if M is an SC (m, n)-module and M is viewed as a CGL(V )-module by restriction via the natural map CGL(V ) → SC (m, n) then M is isomorphic to a direct sum of copies of the hλ V ⊗n . Proof Recall that SC (m, n) = EndCSn (V ⊗n ) and hλ V ⊗n (1.12) ∼ = HomCSn (CSn hλ , V ⊗n ) By (1.9) and (1.10)(Artin–Wedderburn), the nonzero spaces hλ V ⊗n are a complete set of nonisomorphic irreducible SC (m, n)-modules. Now SC (m, n) is semisimple, so the result follows from (2) and (3) below, both of which derive from the fact (at the end of §7) that the map CGL(V ) → SC (m, n) is surjective. (2) If M is an SC (m, n)-module and N is a CGL(V )-module of M then N is an SC (m, n)-submodule of M . (3) If M and N are SC (m, n)-modules and θ : M → N is a CGL(V )-module map then it is an SC (m, n)-map. This implies that (1) is a decomposition of V ⊗n into CGL(V )-submodules which are either zero or irreducible. Which of these submodules are nonzero? We’ll prove: (8.2) Theorem Let λ ` n and m = dimC V . Then if λm+1 6= 0 0 ⊗n dimC hλ V = 1 if λm+1 = 0 andλ1 = · · · = λm ≥ 2 otherwise First we have a slightly technical result. Take a partition λ ` n such that [λ] is partitioned into two nonempty parts, say of sizes i and j = n − i, by a vertical wall. For example, if λ = (5, 3, 2, 1) ` 11 we could have • • • • • • • • ←− j = 4 i = 7 −→ • • • 34 For λ ` n, let tλ be a tableau whose entries in the left-hand part are {1, . . . , i}. Let µ ` i be the corresponding partition, so that tµ = tλ ↓[µ] . Let ν ` j correspond to the right-hand part, with tableau tν . This is a map tν : [ν] → {10 , . . . , j 0 } (8.3) Lemma where k 0 = k + i for 1 ≤ k ≤ j There exists a surjective CGL(V )-module map ((h(tµ )V ⊗i ) ⊗ (h(tν )V ⊗j ) → h(tλ )V ⊗n Proof —(∗) Let Si = Aut({1, . . . , i}) and Sj = Aut({10 , . . . , j 0 }). Si and Sj are embedded in Sn and are disjoint, so we regard CSi and CSj as subsets of CSn which commute. Also Ctλ = Ctµ × Ctν and H = Rtµ × Rtν ≤ Rtλ S and H permutes each side of the wall. Take a transversal Rtλ = i ri H. Then (2.4) h(tλ ) = X X r∈Rtλ c∈Ctλ = X X i = X εc rc X X X εc0 εc00 ri r0 r00 c0 c00 r 0 ∈Rtµ r 00 ∈Rtν c0 ∈Ctµ c00 ∈Ctν ri h(tµ )h(tν ) i So h(tλ )h(tµ )h(tλ ) = X ri h(tµ )2 h(tν )2 (2.10) = kh(tλ ) i where k = i!j! . We thus have a CGL(V )-map fµ fν V ⊗i ⊗ V ⊗j → h(tλ )V ⊗n given by premultiplication by h(tλ ). Restriction of this map to the left-hand side of (∗) is surjective, since h(tλ )(x ⊗ y) = h(tλ ) · 1 (h(tµ )x ⊗ h(tν )y) k (8.4) Lemma (1) If λm+1 = 0 then hλ V ⊗n 6= 0 (2) If n > 0 and λm = 0 then dimC hλ V ⊗n ≥ 2 Proof (1) Let ij be the row in which j occurs in t◦λ and let x = ei1 ⊗ · · · ⊗ ein . For σ ∈ Sn σx = x ⇔ ij = iσ−1 (j) ∀j ⇔ j, σ −1 (j) are in the same row ⇔ σ ∈ Rt◦λ ⊗n So is the coefficient of x in the decomposition of hλ x with respect to the standard basis of V Rt◦ 6= 0. So hλ x 6= 0. λ (2) Let y = e1+i1 ⊗ · · · ⊗ e1+in and adapt the argument above, to see that hλ x and hλ y are linearly independent. (8.5) Lemma If λm+1 = 0 and λm > 0 then hλ V ⊗n ∼ = Λm V ⊗ h(λ1 −1,...,λm −1) V ⊗(n−m) 35 Proof Put a wall in [λ] between the first column and the rest. Let tλ be a tableau whose first column consists of {1, . . . , m}. By (8.3) there exists a surjection Λm V ⊗ h(tν )V ⊗(n−m) h(tλ )V ⊗n where ν = (λ1 − 1, . . . , λm − 1). Hence, with the usual identifications, there exists a map Λm V ⊗ hν V ⊗(n−m) hλ V ⊗n Both hν V ⊗(n−m) and hλ V ⊗n are nonzero, hence they are irreducible CGL(V )-modules by (8.1). Since Λm V is one-dimensional, both sides are irreducible and the map is an isomorphism. Proof of (8.2) If λm+1 6= 0 then [λ] has i > m rows and as in (8.5) there exists a surjection Λi V ⊗ hν V ⊗(n−i) hλ V ⊗n but Λi V = 0, so hλ V ⊗n = 0. If λm+1 = 0 then iterating (8.5) we have dimC hλ V ⊗n = dimC h(λ1 −λm ,...,λm−1 −λm ) V ⊗(n−mλm ) ( 1 if λ1 = · · · = λm = ≥ 2 if not, by (8.4) 36 9 Polynomial and rational representations of GL(V ) Let V be an m-dimensional C-space with basis e1 , . . . , em . (9.1) Definition A finite-dimensional CGL(V )-module M , with basis w1 , . . . , wn , is rational (resp. polynomial, n-homogeneous) if there exist rational (resp. polynomial, n-homogeneous) functions fij (Xrs ) for 1 ≤ i, j ≤ h in m2 variables Xrs for 1 ≤ r, s ≤ m such that the map representation basis e basis wj i GL(V ) −−−−−−−−−→ EndC (M ) −−−−−→ Mh (C) GLm (C) −−−−−→ sends (Ars )1≤r,s≤m 7→ (fij (Ars ))1≤i,j≤h . Exercise and M . Check that the definitions in (9.1) are independent of the choice of bases ei and wj of V Examples (0) The natural representation of GL(V ) on V ⊗n is polynomial. In fact. . . (1) V ⊗n is n-homogeneous; but. . . (2) C ⊕ V is polynomial but not homogeneous; and. . . (3) V ∗ is rational but not polynomial (or homogeneous). (9.2) Lemma (a) Submodules, quotients and direct sums of rational (resp. polynomial, n-homogeneous) modules are rational (resp. polynomial, n-homogeneous). (b) If M is rational then so is M ∗ . (c) If M and N are rational (resp. polynomial, n- and n0 -homogeneous) then M ⊗ N is rational (resp. polynomial, (n + n0 )-homogeneous). (d) If n 6= n0 and M is both n- and n0 -homogeneous then M = 0. (e) if M is rational (resp. polynomial) then Λn M and Sn M are rational (resp. polynomial). Proof Easy exercise. (9.3) Definition For λ1 ≥ · · · ≥ λm ≥ 0 with P i λi = n, set Dλ1 ,...,λm (V ) = h(λ1 ,...,λm ) (V ⊗n ) This is sometimes referred to in the literature as the Weyl module or Schur module. (9.4) Theorem Every n-homogeneous CGL(V )-module is a direct sum of irreducible submodules. Moreover, the modules of (9.3) form a complete set of nonisomorphic irreducible n-homogeneous CGL(V )-modules. Proof By (8.1) it suffices to show that any n-homogeneous CGL(V )-module is obtained from some S(m, n)-module (by restriction). Let M be an n-homogeneous CGL(V )-module. M corresponds to a map ρ0 : GL(V ) → EndC (M ) Consider the following diagram 37 GL(V ) ρ0 EndC (M ) ,→ Sn (EndC V ) EndC (V ) ρ1 ∼ = T n (EndC V )sym ρ2 EndC (M ) EndC (M ) ∼ = S(m, n) ρ4 ρ3 EndC (M ) EndC (M ) In the top row all the maps are natural, and the composite γ : GL(V ) → S(m, n) is the natural map we use for restricting S(m, n)-modules to CGL(V )-modules. Claim There exist ρ1 , ρ2 , ρ3 , ρ4 as above making the diagram commute. Since M is n-homogeneous, we can extend the domain of definition of ρ0 to obtain an n-homogeneous map ρ1 . For ρ2 , use the universal property of symmetric powers. Then ρ3 , ρ4 are transported by the isomorphisms. We check that ρ4 is a map of C-algebras. Well ρ4 (1) = ρ4 (γ(1)) = ρ0 (1) = 1 and ρ4 (γ(gg 0 )) = ρ0 (gg 0 ) = ρ0 (g)ρ0 (g 0 ) = ρ4 (γ(g))ρ4 (γ(g 0 )) for g, g 0 ∈ GL(V ). But γ(GL(V )) spans S(m, n) by (7.8). So M is a CS(m, n)-module and the restriction to GL(V ) is the module we started with. (9.5) Lemma Every polynomial module for GL(C) = C× decomposes as a direct sum of submodules on whih g ∈ C× acts as multiplication by g r for some r. Proof If ρ : C× → GL(M ) ∼ = GLh (C) is a polynomial representation then each ρ(g)ij is polynomial in g. Choose N ∈ N such that each ρ(g)ij has degree < N . Then M is a CG-module by restriction, where ij G = {e2π N : 0 ≤ j ≤ N } ⊆ C× is a cyclic group of order N . Now M = M1 ⊕ · · · ⊕ Mh as a CG-module with each Mj one-dimensional, and g ∈ G acts as multiplication by g nj on Mj for 0 ≤ j < N . (These are the possible irreducible CG-modules.) Choosing nonzero elements of the Mj s gives a basis for M , and if (ρ(g)ij ) is the matrix of ρ(g) with respect to this basis then ( g ni if i = j ρ(g)ij = 0 if i 6= j ij for g ∈ e2π N (0 ≤ j < N ), and hence for all g ∈ C× since the ρ(g)ij are polynomials of degree < N in g. (9.6) Lemma Every polynomial decomposes as a direct sum of n-homogeneous modules. Proof Tke a representation ρ : GL(V ) → GL(U ) be a representation of GL(V ) on a C-space U . We have an inclusion C× ,→ GL(V ), so that one can regard U as a polynomial representation of C× , so by (9.5) U will split as U = U0 ⊕ · · · ⊕ UN with α1 ∈ GL(V ) acting as multiplication by αn on Un . Let u ∈ Un and let g ∈ GL(V ). Write gu = u0 + · · · + uN 38 where ui ∈ Ui for 0 ≤ i ≤ N . Now (α1)gu = g(α1)u for α ∈ C× , so u0 + αu1 + · · · + αN uN = αn u0 + · · · + αn uN and hence ui = 0 for i 6= n. Thus gu ∈ Un and the Un are CGL(V )-submodules of U . Since α1 acts as multiplication by αn on Un , it follows that Un is an n-homogeneous CGL(V )-module. (9.7) Theorem Every polynomial CGL(V )-module is a direct sum of irreducible submodules. The modules Dλ1 ,...,λm (V ) with λ1 ≥ · · · ≥ λm ≥ 0 are a complete set of nonisomorphic irreducible polynomial CGL(V )-modules. (9.8) Definition sentation If n ∈ Z then the one-dimensional CGL(V )-module corresponding to the repreGL(V ) → C⊗ g 7→ (det g)n is denoted by detn . Note that det1 ∼ = Λm V, (9.9) Definition detn ∼ = (det1 )⊗n if n ≥ 0, detn ∼ = (det−n )∗ if n ≤ 0 If λ1 ≥ · · · ≥ λm but λm < 0 then define Dλ1 ,...,λm = Dλ1 −λm ,...,λm−1 −λm ,0 (V ) ⊗ detλm Remark If λ1 ≥ · · · ≥ λm > 0 then by iterating (8.5) (as in the proof of (8.2)) we know that Dλ1 ,...,λm (V ) ∼ = Dλ1 −λm ,...,λm−1 −λm ,0 (V ) ⊗ detλm (9.10) Theorem Every rational CGL(V )-module is a direct sum of irreducible submodules. The modules Dλ1 ,...,λm (V ) for λ1 ≥ · · · ≥ λm form a complete set of nonisomorphic irreducible (rational) CGL(V )-modules. p with p a polynomial function. deti Thus if M is a rational CGL(V )-module then M1 = M ⊗ detN is a polynomial CGL(V )-module for some N . Since M1 decomposes as a direct sum of irreducibles, so does M . If M is irreducible then so is M1 , since detN is one-dimensional, and thus M1 is of the form Proof The rational functions f : GL(V ) → C are of the form f = M1 ∼ = Dµ1 ,...,µm (V ) for some µ1 ≥ · · · ≥ µm Finally, M = Dµ1 −N,...,µm −N (V ) by the above remark. (9.11) Theorem Proof The one-dimensional rational CGL(V )-modules are precisely the detn for n ∈ Z. As above, pass to polynomial modules and then apply (8.2). Remark The characterisation of irreducible CGL(V )-modules can be summarised by: n-homogeneous polynomial rational λ1 ≥ · · · ≥ λm > 0 with λ1 ≥ · · · ≥ λm ≥ 0 λ1 ≥ · · · ≥ λm 39 P i λi = n (9.4) (9.7) (9.10) Interlude: some reminders about characters of GLn (C) Let Tn ≤ GLn (C) be the n-dimensional torus, i.e. the subgroup of GLn (C) consisting of diagonal matrices. If ρ : GLn (C) → GL(W ) is a rational representation then the character of ρ is the function λ1 .. χρ : (x1 , . . . , xn ) 7→ tr ρ . λn sometimes denoted by χW . Basic properties Let ρ, ρ0 be rational representations of GLn (C) associated with W, W 0 . −1 (a) χρ ∈ Z[x1 , x−1 1 , . . . , xn , xn ], and if ρ is polynomial then χρ ∈ Z[x1 , . . . , xn ]. (b) χρ is a symmetric function under the conjugation action of Sn on Tn , i.e. under permutations of the diagonal entries. (c) ρ ∼ = ρ0 ⇒ χρ = χρ0 ; ⇐ will be (10.6). (d) χW ⊕W 0 = χW + χW 0 and χW ⊗W 0 = χW χW 0 . (e) If W is an irreducible polynomial representation of degree m then χW is a homogeneous polynomial of degree m. −1 (f) χW ∗ (x1 , . . . , xn ) = χW (x−1 1 , . . . xn ). (g) When V = Cn we have χS2 V χV ⊗n = (x1 + · · · + xn )m , X X = xi xj , χΛ2 V = xi xj , i<j i≤j χdet = x1 . . . xn , χΛn−1 V = n X i=1 x1 . . . xbi . . . xn , χV ∗ = x−1 1 + · · · + x−1 n , χSj V = hj (x1 , . . . , xn ) = X xi1 . . . xij i1 ≤···≤ij The hj are called complete symmetric functions. We could (but won’t) study the theory of symmetric functions and apply it here; see R. Stanley’s book for more. 40 10 Weyl character formula Let V be an m-dimensional C-space. If λ = (λ1 , . . . , λm ) ∈ Zm and P λ1 ≥ · · · ≥ λm then the character of the CGL(V )-module Dλ1 ,...,λm (V ) (= hλ V ⊗n if λm ≥ 0 and n = i λi ) will be denoted by φλ . (10.1) Lemma Proof If ξ ∈ EndC (V ) then φλ (ξ) is a symmetric rational function of the eigenvalues of ξ. The function P (x1 , . . . , xm ) = φλ (diag(x1 , . . . , xm )) is a rational function of the x1 , . . . , xm , and it is symmetric since diag(x1 , . . . , xm ) is conjugate to diag(xτ (1) , . . . , xτ (m) ) for any τ ∈ Sm . Now choose a basis for V such that the matrix A1 of ξ is in Jordan normal form, and for t ∈ C let At be the matrix obtained from A1 by replacing the upperdiagonal 1s by ts. Let ξt be the endomorphism corresponding to At . For t 6= 0, At is conjugate to A1 , and so φλ (ξt ) = φλ (ξ). Since φλ is a rational function it is continuous (wherever it is defined), and so φλ (ξ) = lim φλ (ξt ) = φλ (lim ξt ) = φλ (ξ0 ) = P (x1 , . . . , xm ) t→0 t→0 which is a symmetric function of the eigenvalues of ξ. (10.2) Lemma Let α be an Sn -conjugacy class with cycle type nαn , . . . , 1α1 , and let ξ be an endomorphism of V with eigenvalues x1 , . . . , xm . If si = xi1 + · · · + xim then X αn 1 sα χλ (α)φλ (ξ) 1 . . . sn = λ∈Λ+ (m,n) (Compare this with (4.14).) Proof Pick g ∈ α. Suppose ξ has matrix diag(x1 , . . . , xm ) with respect to the standard basis e1 , . . . , em of V . We compute the trace of the endomorphism of V ⊗n sending x 7→ (gξ)x = ξgx in two ways. Way 1 V ⊗n is a CSn -module, so by (1.8), M CSn hλ ⊗ HomCSn (CSn hλ , V ⊗n ) V ⊗n ∼ = λ`n which, by (1.12), becomes V ⊗n ∼ = M (CSn hλ ⊗ hλ V ⊗n ) λ`n This is an isomorphism both of CSn - and EndCSn (V ⊗n )-modules. Since the action of GL(V ) on V ⊗n commutes with that of Sn , the corresponding action of gξ on the right-hand side is given by the action of g on CSn hλ and of ξ on hλ V ⊗n . So the trace of the action is P λ —(1) λ∈Λ+ (n,m) ξ (α)φλ (ξ) Way 2 (direct computation) gξ(ei1 ⊗ · · · ⊗ ein ) = xi1 . . . xin eig−1 (1) ⊗ · · · ⊗ eig−1 (n) and so tr(gξ) = X xi1 . . . xin D where D = {(i1 , . . . , in ) : 1 ≤ i1 , . . . , in ≤ m and ig−1 (j) = ij ∀j}. The condition ig−1 (j) = ij for all j is equivalent to requiring that the function j 7→ ij be constant on the cycles involved in g. Hence αn 1 tr(gξ) = sα 1 . . . sn 41 —(2) Equating (1) and (2) gives the result. (10.3) Theorem Let λ ∈ Λ+ (m, n) and let ξ ∈ EndC (V ) with eigenvalues x1 , . . . , xm . If si = xi1 + · · · + xim , then s αn X χλ (α) s1 α1 n ··· φλ (ξ) = α1 ! . . . αn ! 1 n ccls α Proof Take the formula of (10.2), multiply it by nα χµ (α) and sum over α; then use orthogonality of the χs. (10.4) Theorem (Weyl’s character formula for GLn ) Let ξ ∈ GL(V ) with eigenvalues x1 , . . . , xm . Then ` x 1 , . . . , x `m φλ (ξ) = m−1 |x , . . . , 1| where `i = λi + m − i. Proof (a) Suppose λi ≥ 0 for all i, so that λ ∈ Λ+ (n, m). By (10.2) and the character formula for the symmetric group (4.14) we know that ` x 1 , . . . , x`m X X α1 αn λ s1 . . . sn = χ (α) m−1 = χλ (α)φλ (α) |x , . . . , 1| + + λ∈Λ (n,m) λ∈Λ (n,m) Then the orthogonality of χλ s allows us to equate coefficients. (b) For general λ, since (9.9) Dλ ∼ = Dλ1 −λm ,...,λm −λm (V ) ⊗ detλm it follows that φλ (ξ) = ` −λ x 1 m , . . . , x`m −λm |xm−1 , . . . , 1| (x1 . . . xm ) λm ` x 1 , . . . , x`m = m−1 |x , . . . , 1| Remark This short proof uses the character formula of Sn from §4. One can do this in reverse, using integration to compute the Weyl formula for the compact subgroup Un of unitary matrices in GLm (C), and then translate that to GLm (C). [See also Telemann’s notes for the 2005 Part II course.] You then use (10.2) to pass to Sn . See also Weyl’s book. (10.5) Theorem (Weyl’s dimension formula) Q 1≤i<j≤m deg φλ = dim Dλ1 ,...,λm (V ) = Q Proof (`i − `j ) 1≤i<j≤m (j − i) For t ∈ C set xm = 1, xm−1 = et , . . . , x1 = e(m−1)t . Then Y ` x 1 , . . . , x`m = (e`i t − e`j t ) 1≤i<j≤m since it’s the transpose of a Vandermonde matrix. The term of lowest degree is Q i<j ((`i − `j )t). Q Q Also xm−1 , . . . , 1 = i<j (e(m−i)t − e(m−j)t ) and the term of lowest degree in t is i<j ((j − i)t). If ξt = diag(x1 , . . . , xm ) then by (10.4), deg φλ = φλ (1) = limt→0 φλ (ξt ) = 42 Q i<j (`i −`j ) Q i<j (j−i) (10.6) Theorem phic. If two rational CGL(V )-modules have the same characters then they are isomor- Proof As λ varies, the rational functions in Weyl’s character formula are linearly independent elements of C(x1 , . . . , xm ). (10.7) Lemma Dλ1 ,...,λm (V )∗ ∼ = D−λm ,...,−λ1 (V ) Let ψ be the character of Dλ1 ,...,λm (V ). Then −` −` x m , . . . , x−`1 x 1 , . . . , x−`m −1 = ψ(ξ) = φλ (ξ ) = −(m−1) x , . . . , 1 1, . . . , x−(m−1) m−1−` m−1−` m 1 x ,...,x = = φµ (ξ) |xm−1 , . . . , 1| Proof where µ = (−λm , . . . , −λ1 ). Hence we have the required isomorphism by (10.6). Finally we have the famous Clebsch–Gordan formula. (10.8) Theorem (Clebsch–Gordan formula) Let V = C2 and p, q ∈ N. Then min{p,q} Dp,0 (V ) ⊗ Dq,0 (V ) ∼ = Proof M Dp+q−r,r (V ) r=0 Exercise; write down the characters and use (10.6). (10.9) Examples of rational CGL(V )-modules C∼ = D0,0,...,0,0 (V ) deti ∼ = Di,i,...,i,i (V ) V ∼ = D1,0,...,0,0 (V ) Sn V ∼ = Dn,0,...,0,0 (V ) Λn V ∼ = D1,...,1,0,...,0 (V ) where n ‘1’s and m − n ‘0’s appear, and n ≤ m V∗ ∼ = D0,0,...,0,−1 (V ) (Sn V )∗ ∼ = D0,0,...,0,−n (V ) (Λn V )∗ ∼ = D0,...,0,−1,...,−1 (V ) where m − n ‘0’s and n ‘−1’s appear, and n ≤ m 43 Chapter III: Invariant theory 11 Introduction to invariant theory and first examples Let W be a finite-dimensional C-space and C[W ] be its coordinate ring (cf. §6). Facts about C[W ] C[W ] is a commutative C-algebra, which is infinite-dimensional if W 6= 0, with (λf )(w) = λf (w), (f + g)(w) = f (w) + g(w), (f g)(w) = f (w)g(w), 1C[W ] (w) = 1 for all λ ∈ C, f, g ∈ C[W ] and w ∈ W . C[W ] is the ring of regular functions of the affine variety Adim W ∼ = W. C[W ] ∼ = M (6.12) HC,n (W, C) ∼ = n≥0 M Sn (W ∗ ) = S(W ∗ ) n≥0 This latter isomorphism is known as polarisation. If W ∗ has basis w1∗ , . . . , wh∗ then the map C[X1 , . . . , Xh ] → Xi 7→ C[W ] wi∗ is a C-algebra homomorphism. If W is a CG-module then C[W ] is a CG-module via the action gf (w) = f (g −1 w) for g ∈ G, f ∈ C[W ] and w ∈ W . For g ∈ G, the map C[W ] → C[W ] is a C-algebra automorphism, since f 7→ gf g(f f 0 )(w) = (f f 0 )(g −1 w) = f (g −1 w)f 0 (g −1 w) = (gf )(w)(gf 0 )(w) = ((gf )(gf 0 ))(w) and g1C[W ] (w) = 1C[W ] (g −1 w) = 1 = 1C[W ] (w) (11.1) Definition If V, W are CG-modules then f : W → V is concomitant if f (gw) = gf (w) for all w ∈ W, g ∈ G A G-invariant for the CG-module W is a concomitant map f : W → C. Remarks (1) f is invariant if and only if it is a fixed point under the G-action, i.e. gf = f for all g ∈ G. (2) linear concomitant ≡ CG-homomorphism (3) The map ∆ : V w → Sn W is an n-homogeneous concomitant. 7 → w ∨ ··· ∨ w (4) C[W ]G = {f ∈ C[W ] : gf = f for all g ∈ G} is a C-subalgebra of C[W ], called the polynomial invariant ring. It is a graded C-algebra: M C[W ]G = C[W ]G d d≥0 with grading inherited by that of C[W ]. 44 (11.2) Problem Compute C[W ]G , describing its generators and relations. What follows are some classical results which we will state but not prove. [Noether] If G is finite then C[W ]G is finitely generated as a C-algebra. [Noether, 1916] If char K = 0 then for any representation W of the finite group G, the invariant ring K[W ]G is generated by the invariants of degree ≤ |G|. [Hilbert] If W is a rational CGL(V )-module then C[W ]GL(V ) and C[W ]SL(V ) are finitely-generated C-algebras. In fact [Hochster–Roberts, 1974] they are also Cohen–Macaulay rings. Also, C[W ]SL(V ) is a unique factorisation domain, and hence it is a Gorenstein ring (i.e. it has finite injective dimension). [Popov, Astérisque, 1988] There is an explicit bound on the number of generators for C[W ]SL(V ) . Examples (11.3) Symmetric polynomials W = Cn with basis e1 , . . . , en , is naturally a CSn -module. If ξ1 , . . . , ξn is the corresponding dual basis for W ∗ , then the isomorphism ∼ = → C[W ] C[X1 , . . . , Xn ] − Xi 7→ ξi allows an identification C[W ] ↔ C[X1 , . . . , Xn ]. The polynomial invariant ring of the CSn -module W is C[X1 , . . . , Xn ]Sn = {p ∈ C[X1 , . . . , Xn ] : p symmetric in the Xi s} We have elementary symmetric functions ej (X1 , . . . , Xn ) ∈ C[X1 , . . . , Xn ]Sn such tha ~ n−1 + · · · + en (X) ~ (t + X1 ) · · · (t + Xn ) = tn + e1 (X)t ~ = (X1 , . . . , Xn ). That is, where X ~ = X1 + · · · + Xn , e1 (X) ~ = ei (X) X Xj1 Xj2 . . . Xji , ~ = X1 . . . Xn en (X) j1 <j2 <···<ji The fundamental theorem of symmetric functions computes the polynomial invariants for W ; namely, it shows that the map C[Y1 , . . . , Yn ] → C[X1 , . . . , Xn ]Sn ~ Yi 7→ ei (X) is a C-algebra isomorphism. The standard way to prove this is by showing that C[X1 , . . . , Xn ]Sn = C[e1 , . . . , en ] by induction on n. Recall for p(t) ∈ C[t] monic of degree n, say p(t) = tn + a1 tn−1 + · · · + an = its discriminant is disc(p) = Y i<j n Y (t + λi ) i=1 (λi − λj )2 2 Q Now, since the polynomial X n−1 , . . . , 1 = i<j (Xi − Xj )2 is symmetric in the Xi s, it may be represented as a polynomial in the ei s, say Y ~ . . . , en (X)) ~ (Xi − Xj )2 = D(e1 (X), i<j 45 Also, disc(p) is a polynomial in the coefficients of p, for example disc(t2 + bt + c) = b2 − 4c, disc(t3 + bt + c) = −4b3 − 27b2 (11.4) Alternating groups Restriction in (11.3) enables consideration of W as a representation of An (n ≥ 2). The ring of invariants is C[X1 , . . . , Xn ]An . Consider the Vandermonde determinant n−1 Y X , . . . , 1 = (Xi − Xj ) i<j n−1 2 X , . . . , 1 is Sn -invariant since it is equal to D(e1 , . . . , en ) as above. Claim The C-algebra map θ : C[Y1 , . . . , Yn , Z] Yi Z is surjective, and ker(θ) = (Z 2 − D(Y1 , . . . , Yn )). → C[X1 , . . . , Xn ]An 7→ n−1ei X 7→ , . . . , 1 Proof We prove the claim in three steps. Step 1 θ is surjective. Take a transposition τ ∈ Sn . Let f ∈ C[X1 , . . . , Xn ]An . Then fsym = f + τ f is a symmetric polynomial, since Sn is generated by τ and An , and τ fsym = τ f + τ 2 f = τ f + f = fsym and if σ ∈ An then σfsym = σf + τ (τ −1 στ )f = f + τ f = fsym Similarly, falt = f − τ f is an alternating polynomial. Observe that falt (X1 , . . . Xn ) = 0 whenever Xi = Xj for i 6= j. Hence by Nullstellensatz q falt ∈ (Xi − Xj ) = (Xi − Xj ) and so (Xi − Xj ) | falt for all i 6= j. In particular, falt = X n−1 , . . . , 1 g for some polynomial g. Now g is symmetric, and hence f= 1 1 (fsym + falt ) = (fsym + X n−1 , . . . , 1 g) 2 2 Since fsym and g lie in im(θ), so does f . ~ )) ⊆ ker(θ). This is clear since Step 2 (Z 2 − D(Y ~ )) = X n−1 , . . . , 12 − D(e1 , . . . , en ) = D(e1 , . . . , en ) − D(e1 , . . . , en ) = 0 θ(Z 2 − D(Y ~ )). Step 3 ker(θ) ⊆ (Z 2 − D(Y ~ , Z) is of the form Any polynomial P (Y ~ , Z) = Q(Y ~ , Z)(Z 2 − D(Y ~ )) + A(Y ~ ) + B(Y ~ )Z P (Y | {z } | {z } quotient remainder ~ ) + B(Y ~ )Z then P = 0. We show that if P ∈ ker(θ) has the form A(Y Suppose a1 , . . . , an ∈ C, and let f (t) = tn + a1 tn−1 + · · · + an = (t + λ1 ) · · · (t + λn ) 46 Then P (a1 , . . . , an , λn−1 , . . . , 1) = θ(P )(λ1 , . . . , λn ) = 0 since P ∈ ker(θ). Exchanging λ1 and λ2 changes the sign of the Vandermonde determinant, so 1 P (a1 , . . . , an , ±δ) = 0, where δ = disc(tn + a1 tn−1 + · · · + an ) 2 . Using the given form for P , A(a1 , . . . , an ) ± B(a1 , . . . , an )δ = 0 and hence A = B = 0 on the Zariski-dense open subset {(a1 , . . . , an ) ∈ Cn : disc(tn + a1 tn−1 + · · · + an ) 6= 0} Then A = B = 0 everywhere on Cn , so P = 0. (11.5) Characteristic polynomial Let V be an m-dimensional C-space. GL(V ) acts naturally on U = EndC (V ) by conjugation, i.e. (g · θ)(v) = g · θ(g −1 · v) (or, in shorthand, g · θ = gθg −1 ). Suppose χθ (t) = det(t1V + θ) is the characteristic polynomial of θ and cn (θ) is the coefficient of tm−n in χθ (t). Then χg·θ (t) = det(t1V + gθg −1 ) = det(g(t1V + θ)g −1 ) = χθ (t) so cn : U → C is an (n-homogeneous) invariant. If θ has eigenvalues lambda1 , . . . , λm , then putting θ in Jordan normal form gives χθ (t) = m Y (t + λi ) i=1 so that cn (θ) = en (λ1 , . . . , λm ); in particular c1 (θ) = λ1 + · · · + λm = tr(θ) Claim The C-algebra map and cm (θ) = λ1 . . . λm = det(θ) C[Y1 , . . . , Ym ] → C[U ]GL(V ) is an isomorphism. Yi 7→ ci Sketch proof If f is a polynomial invariant for U then f (θ) is a symmetric polynomial function of the eigenvalues of θ. Mimic the proof that the characters of rational CGL(V )-modules are symmetric rational functions of the eigenvalues of g ∈ GL(V ) as done in (10.1). (11.6) Discriminant of quadratic forms Let V be an m-dimensional C-space and let U = HC,2 (V, C) = {quadratic forms on V } By polarisation (see end of §6) identify U with the symmetric m × m matrices via f 7→ A, where f (x) = xt Ax for x ∈ Cm a column vector. Define disc(f ) = det(A). GLm (C) acts on U by t (gf )(x) = f (g −1 x) = xt (g −1 Ag −1 )x, so t f ∈ U, g ∈ GLm (C), x ∈ Cm disc(gf ) = det(g −1 Ag −1 ) = (det g)−2 disc(f ) 47 and therefore disc : U → C is an SLm (C)-invariant (but not a GLm (C)-invariant). Claim The C-algebra map C[X] → X 7→ C[U ]SLm (C) is an isomorphism. disc Proof It is injective: if there is a polynomial P such that P (disc f ) = 0 for all f ∈ U then P (λ) = 0 for all λ ∈ C, since the quadratic form fλ (x1 , . . . , xm ) = x21 + · · · + x2m−1 + λx2m has discriminant λ, so P = 0. Let θ : U → C be a polynomial SLn (C)-invariant and define F : C λ → C 7 → θ(fλ ) This is a polynomial function since θ is a polynomial map. We prove that θ(f ) = F (disc f ). It is enough to show this for f with disc f 6= 0. For, if this is case is known, then U = {f : disc f = 0} ∪ {f : θ(f ) = F (disc f )} is a union of Zariski-closed subsets, but U is irreducible, so that U = {f : θ(f ) = F (disc f )}. Recall that any symmetric matrix is congruent (over C) to a matrix of the form 1 .. . 1 diag(1, . . . , 1, 0, . . . , 0) = 0 . .. 0 Thus if f ∈ U corresponds to the matrix A and λ = disc f = det A 6= 0 then there exists B ∈ GLm (C) such that B t AB = I. Now if C = diag(1, . . . , 1, (det B)−1 ), then BC ∈ SLm (C), and (BC)t A(BC) = C t B t ABC = diag(1, . . . , 1, (det B)−2 ) = diag(1, . . . , 1, λ) and so θ(f ) = θ(fλ ) = F (λ), as required. (11.7) Discriminant of binary forms Define U = HC,n (C2 , C) = Take f ∈ U . Then f has the form homogeneous polynomials of degree n in two variables X1 , X2 f (X1 , X2 ) = a0 X1n + a1 X1n−1 X2 + · · · + an X2n = b(λ1 X1 + µ1 X2 )(λ2 X1 + µ2 X2 ) · · · (λn X1 + µn X2 ) and define disc(f ) = b2n−2 Y i<j (λi µj − λj µi )2 This is well-defined since it is unchanged if two terms are exchanged or if one term is enlarged and another is reduced by the same factor. Example (n = 3) disc(f ) = −27a20 a23 + 18a0 a1 a2 a3 − 4a0 a32 − 4a31 a3 + a21 a22 Claim disc : U → C is a (2n − 2)-homogeneous SL2 (C)-invariant. Proof Exercise for the daring. (11.8) Definition) U ⊕ V → C. A covariant for a CSL(V )-module U is a polynomial invariant of the form Examples 48 Every invariant θ for U gives rise to a covariant U ⊕V (u, v) → C 7 → θ(u) If U = HC,n (V, C) then there is an ‘evaluation covariant’ ev : U ⊕V (f, v) → C . 7→ f (v) (11.9) The Hessian If f is a function of X1 , . . . , Xm then the Hessian of f is ∂2f H(f ) = det ∂Xi ∂Xj It is itself a function of X1 , . . . , Xm . Let U = HC,n (Cm , C), the n-homogeneous polynomials in X1 , . . . , Xm . The Hessian defines a polynomial map U ⊕ Cm → C (f, v) 7→ H(f )(v) Now U is naturally a CGLm (C)-module via the action (gf )(v) = f (g −1 v), f ∈ U, g ∈ GLm (C), v ∈ Cm By the chain rule, H(gf )(gv) = (det g)−2 H(f )(v) and so H is an SL2 (C)-covariant. 49 12 First fundamental theorem of invariant theory The first fundamental theorem of invariant theory gives generators for the set of polynomial invariants when the module is the direct sum of copies of V and V ∗ . The hope was then in principle to compute invariants of an arbtirary rational module, but in practice this proved to be impossible. We will state and prove the first fundamental theorem for GL(V ) and state it for SL(V ). As usual, V is an m-dimensional C-space with basis e1 , . . . , em . (12.1) Theorem (multilinear first fundamental theorem for the general linear group) For n, r ∈ N, (i) HomCGL(V ) (T n (V ∗ ) ⊗ T r (V ), C) = 0 if n 6= r; (ii) HomCGL(V ) (T n (V ∗ ) ⊗ T n (V ), C) = spanC {µσ : σ ∈ Sn }, where µσ (ϕ1 ⊗ · · · ⊗ ϕn ⊗ v1 ⊗ · · · ⊗ vn ) = ϕσ(1) (v1 ) · · · ϕσ(n) (vn ) Remark Proof A proof in arbitrary characteristic can be found in de Cocini and Procesi (1976). Let X, Y be CG-modules, where G is some group. We know that HomC (X ∗ ⊗ Y, C) = (X ∗ ⊗ Y )∗ ∼ = X ∗∗ ⊗ Y ∗ ∼ =X ⊗Y∗ ∼ = HomC (Y, X) Taking G-fixed points, we obtain HomCG (X ∗ ⊗ Y, C) ∼ = HomCG (Y, X) With this and the isomorphism T n (V ∗ ) ∼ = (T n V )∗ we have an isomorphism ∼ = π : HomCGL(V ) (T r V, T n V ) −→ f 7−→ HomCGL(V ) (T n (V ∗ ) ⊗ T r (V ), C) ϕ1 ⊗ · · · ⊗ ϕn ⊗ ~v 7→ (ϕ1 ⊗ · · · ⊗ ϕn )(f (v)) Now (i) If n 6= r then HomCGL(V ) (T r V, T n V ) = 0 because HomCGL(V ) (S 0 , S) = 0 if S 0 (resp. S) is an irreducible r-homogeneous (resp. n-homogeneous) CGL(V )-module. (ii) is just a restatement of Schur–Weyl duality. To see this, recall that EndCGL(V ) (T n V ) is spanned by λσ for σ ∈ Sn , where λσ (v1 ⊗ · · · ⊗ vn ) = vσ−1 (1) ⊗ · · · ⊗ vσ−1 (n) and now π(λσ ) = µσ . Notation For CG-modules U and W , let HomCG,n (U, W ) denote the vector space of all n-homogeneous concomitants U → W (cf. (11.1)). Note that HomCG,n (U, W ) is a CG-module with conjugation action (gf )(u) = gf (g −1 u), g ∈ G, u ∈ U, f ∈ HomCG,n (U, W ) Note also that HomCG,n (U, W ) = HC,n (U, W )G . (12.2) Lemma If U is a CG-module then C[U ]G ∼ = M HomCG (U, C) n≥0 Proof We know that C[U ] ∼ = M HC,n (U, C) n≥0 50 is an isomorphism of CG-modules; now take G-fixed points. (12.3) Lemma If U and W are CCG-modules then there exist invrse isomorphisms of vector spaces HomCG (Sn U, W ) → ψ 7→ where ∆ : U u HomCG,n (U, W ) → HomCG,n (U, W ) and ψ◦∆ f 7→ HomCG (Sn U, W ) 1 n! P f 7 → Sn U and P f is the total polarisation of f (see after (6.13)). 7→ u ∨ · · · ∨ u Clearly the maps induce isomorphisms between HomC (Sn U, W ) and HC,n (U, W ), so we’ll Proof prove: (i) ψ concomitant ⇒ ψ ◦ ∆ concomitant; (ii) f concomitant ⇒ 1 n! P f concomitant. So. . . (i) is clear since ∆ is a concomitant. (ii) To prove this, there are (at least) three possible approaches. Use the formula for P f (not recommended). 1 P f ◦ ∆, so Approach 2 If f ∈ HomCG,n (U, W ) then f = n! 1 1 P f (g(u ∨ · · · ∨ u)) = P f (g∆(u)) n! n! 1 P f ∆(g(u)) = n! Approach 1 = f (g(u)) = gf (u) 1 =g P f (u ∨ · · · ∨ u) n! Approach 3 HomC (Sn U, W ) and HC,n (U, W ) are CG-modules, with G acting by conjugation, 1 and ψ 7→ ψ ◦ ∆ is a CG-map and an isomorphism. So its inverse f 7→ n! P f is also a CG-map. These maps restrict to isomorphisms between sets of G-fixed points. Remark δ natural ∆ factors as U − → T n U −−−−→ Sn U , where δ(u) = u ⊗ · · · ⊗ u | {z } n so composition with δ induces a surjection HomCG (T n U, W ) −→ HomCG (U, W ) We’ll use this reformulation of (12.3) to prove (12.4) Theorem (first fundamental theorem for the general linear group) Let U = V ⊕ · · · ⊕ V ⊕ V ∗ ⊕ · · · ⊕ V ∗ . Then C[U ]GL(V ) is generated as a C-algebra by elements ρij | {z } | {z } p q for 1 ≤ i ≤ p, 1 ≤ j ≤ q, where ρij (v1 , . . . , vp , ϕ1 , . . . , ϕq ) = ϕj (vi ) Proof Clearly the ρij are polynomial invariants, since gϕj (gvi ) = ϕj (g −1 gvi ) = ϕj (vi ) 51 By (12.2) we must show that any n-homogeneous invariant f is a linear combination of products of the ρij . By (12.3) there exists a surjection HomCGL(V ) (T n U, C) −→ HomCGL(V ),n (U, C) Lp+q Put V1 = · · · = Vp = V and Vp+1 = · · · = Vp+q = V ∗ , so that U = i=1 Vi . This induces a decomposition of T n U as T nU = p+q M p+q M i1 =1 i2 =1 Thus HomCGL(V ) (T n U, C) = ··· p+q M in =1 M i1 ,...,in Vi1 ⊗ Vi2 ⊗ · · · ⊗ Vin HomCGL(V ) (Vi1 ⊗ · · · ⊗ Vin , C) Focus on one of the summands. By (12.1), either HomCGL(V ) (Vi1 ⊗ · · · ⊗ Vin , C) = 0, or n = 2k, k of the ij are ≤ p and k of the ij are > p. In the latter case, say ij ≤ p for j = α1 , . . . , αk and ij > p for j = β1 , . . . , βk . In this case, HomCGL(V ) (Vi1 ⊗ · · · ⊗ Vin , C) is spanned by the k! maps τσ , for σ ∈ Sk , given by τσ (v1 ⊗ · · · ⊗ vn ) = ϕβσ(1) (vα1 ) · · · ϕβσ(k) (vαk ) The n-homogeneous invariant of U corresponding to τσ is just ρiα1 ,iβσ(1) −p · · · ρiαk ,iβσ(k) −p and we’re done. (12.5) Theorem If r ∈ N and U = EndC V ⊕ · · · ⊕ EndC V then C[U ]GL(V ) is generated as a {z } | r C-algebra by invariants ti1 ,...,ik : where k ≥ 1 and 1 ≤ i1 < · · · < ik ≤ r. U → C (θ1 , . . . , θr ) 7→ tr(θi1 · · · θik Remarks (1) [Procesi, 1976] In fact, C[U ]GL(V ) is generated by the ti1 ,...,ik with 1 ≤ k ≤ 2m − 1. (2) Procesi also shows that (12.5) is equivalent to the first fundamental theorem. This links invariant theory to the representation theory of bouquets, a certain type of quiver. ∼ Proof By (12.2) and (12.3) we must compute HomCGL(V ) (T n (EndC V, C) which, since EndC V = V ∗ ⊗V , is isomorphic to HomC (T n V ∗ ⊗T n V, C). By the multilinear first fundamental theorem (12.1), this is spanned by the µσ for σ ∈ Sn , where µσ (ϕ1 ⊗ ϕn ⊗ v1 ⊗ · · · ⊗ vn ) = ϕσ(1) (v1 ) · · · ϕσ(n) (vn ) We now compute the corresponding map νσ : T n (EndC V ) −→ C Let V have basis e1 , . . . , em and V ∗ have dual basis η1 , . . . , ηm . Let (Aij ) be the matrix of θ ∈ EndC V with respect to these bases, i.e. X θ(v) = Aij ηj (v)ei i,j ∗ so that the corresponding element of V ⊗ V is P i,j Aij ηj ⊗ ei . If θk has matrix Akij then θ1 ⊗ · · · ⊗ θn corresponds with X X A1a1 b1 · · · Anan bn ηb1 ⊗ · · · ⊗ ηbn ⊗ ea1 ⊗ · · · ⊗ ean ∈ T n V ∗ ⊗ T n V a1 ,...,an b1 ,...,bn 52 so we have νσ (θ1 ⊗ · · · ⊗ θn ) = = X ~ a,~b X ~b A1a1 b1 · · · Anan bn ηbσ(1) (ea1 ) · · · ηbσ(n) (ean ) A1bσ(1) b1 · · · Anbσ(n) bn [The latter equality comes from the definition of dual basis.] Writing σ = (i1 · · · ik )(j1 · · · j` ) · · · as a product of disjoint cycles, we reorder this to get X νσ (θ1 ⊗ · · · ⊗ θn ) = Aib1i bi Aib2i bi · · · Aibki bi Ajb1j bj · · · Ajb`j bj · · · 2 1 3 2 1 k 2 1 1 ` = tr(θik · · · θi1 ) tr(θj` · · · θj1 ) · · · which is what we required. Theorem (first fundamental theorem for the special linear group) Let V be an m-dimensional C-space with basis e1 , . . . , em and let V ∗ have dual basis η1 , . . . , ηm . If U = V ⊕ ··· ⊕ V ⊕V ∗ ⊕ ··· ⊕ V ∗ {z } | {z } | p q then C[U ]SL(V ) is generated as a C-algebra by the polynomial invariants which send (v1 , · · · , vp , ϕ1 , · · · , ϕq ) 7→ ϕj (vi ) for 1 ≤ i ≤ p, 1 ≤ j ≤ q; (v1 , · · · , vp , ϕ1 , · · · , ϕq ) 7→ |vi1 , . . . , vim | for 1 ≤ i1 < · · · < im ≤ p (v1 , · · · , vp , ϕ1 , · · · , ϕq ) 7→ |ϕj1 , . . . , ϕjm | for 1 ≤ j1 < · · · < jm ≤ q where |v| is the determinant of the matrix whose nth column is the coordinate of the vin with respect to e1 , . . . , em and |ϕ| likewise with respect to ηj1 , . . . , ηjm . End of course. 53 MATHEMATICAL TRIPOS PART III, 2013 REPRESENTATION THEORY This sheet deals with the representation theory and combinatorics of Sn and of GLn . Work over C unless told otherwise. 1 Let H be a subgroup of Sn ; define the element H − = algebra CSn . Prove that and P h∈H εh h which lies in the group πH − = H − π = επ H − for any π ∈ H, πH − π −1 = (πHπ −1 )− for any π ∈ Sn . 2 Let r be a natural number. Let λ = (λ1 , . . . λ` ) be a sequence of positive integers. We P call λ a composition of r into ` parts if `i=1 λi = r. The elements λi are called the parts of λ. (a) Find a formula for the number of compositions of r. (b) Find a formula for the number of compositions of r with at most k parts. 3 Let r, s be non-negative integers. A partition of the form λ = (r, 1s ) is called a hook partition; we will write H(n) for the set of all hook partitions of n. Let f λ be the number of standard tableaux of shape λ. (a) Find a formula for the number of standard tableaux of shape λ = (r, 1s ) where λ is a partition of n. Justify P your answer. (b) Prove that λ∈H(n) f λ = 2n−1 . P (c) Prove that λ∈H(n) (f λ )2 = 2(n−1) . n−1 4 Let n ∈ N, and let λ = (λi )i∈N and µ = (µj )j∈N be partitions of n. The conjugate partition λ0 is defined to be the partition λ0 = (λ01 , λ02 , . . .) where λ0i = |{j|λj > i}|. The dominance order on the set of partitions of n is defined as follows: λ µ if and only if for all i ∈ N: X X λj 6 µj . j6i j6i 0 0 (a) Prove that λ µ if and only if µ λ . (b) We say that a partition λ is self-conjugate if λ = λ0 . Show that there are as many self-conjugate partitions of n as there are partitions of n with distinct odd parts (i.e. all λi are odd and distinct). (a) Assuming k > 3, prove that then f (3 ) > 3k. (b) Let λ = (l, . . . , l) = (lk ) be a partition of n = kl and λ̃ = (l + 1, . . . , l + 1) be a partition of n + k. Prove (by using the Young-Frobenius formula (5.1)) that 5 k f λ̃ (n + k)! l! = · . λ f n! (l + k)! (c) Let k, l > 3 with kl = n. Prove that then f (l ) > n. k 1 6 Let Ck be the set of all standard tableaux of shape (k, k) whose entry in box (2,1) is some odd number. (a) Find |Ck |. (b) For a partition λ = (r, s) with two parts, write down f (r,s) (e.g. from the YoungFrobenius formula). Can you find limk→∞ |Ck |/f (k,k) ? 7 In this question λ = (λ1 λ2 , . . . , λk ) is a partition of n, and M λ is the permutation module with basis the λ-tabloids (a λ-tabloid is an equivalence class of numberings of a Young diagram (with distinct numbers 1,...,n), two being equivalent if the corresponding rows containing the same entries.) (a) Show that M λ is isomorphic to the permutation module C∆ where ∆ = cos(Sn : Sλ ), (the coset space with Sλ the Young subgroup). (b) Show that n! dim M λ = Qk . i=1 λi ! (c) In the case λ = (n − 2, 2) has two parts, show that M λ ∼ = CΩ(2) where Ω(2) is the set of 2-element subsets of Ω = {1, . . . , n} (with the natural action). P 8 For a λ-tableau T with n boxes, define the signed column sum κT = σ∈CT εσ σ . Let eT = κT {T } be the λ-polytabloid. Let 1 denote the identity element of Sn . (a) List the columns of T as C1 , C2 , . . . , Cs . Prove that P κT = κC1 · · · κCs . − (b) Let H 6 Sn and let (a b) ∈ H. Define H = h∈H εh h, as in Q.1. Prove that H − = s · (1 − (a b)) for some s ∈ CSn . (c) Let π ∈ Sn and σ ∈ CT . Show that κπT = πκT π −1 , κT σ = σκT = εσ κT and eπT = πeT ; ; σeT = εσ eT . 9 *[The Murnaghan-Nakayama Rule.] Part (a) below is an optional task for enthusiasts only. You will need to read about James’ method (lecture notes pp 79–83) using the Littlewood-Richardson Rule. A good account appears in Sagan’s Springer GTM, p.182. This beautiful result has generalisations to the irreducible characters of exotic structures called Iwahori-Hecke algebras of type An−1 due to Ram (who appealed to Schur-Weyl duality), and it appears to have been known to Young in his work on hyperoctahedral groups (the so-called Weyl groups of type Bn ). (a) Consider the character χλ (w) =: χλµ where w ∈ Sn has cycle type (µ1 , µ2 , . . . , µ` ). Suppose there are precisely t partitions λ(1), . . . , λ(t) which are obtained by removing a rim λ(i) hook of length µ1 from [λ]. Find an expression for χλµ in terms of χ(µ2 ,µ3 ...,µ` ) . (b) Let n > 6. Prove that (n−3,3) χ(2,1n−2 ) = (n − 3)(n − 4)(n − 5)/6. 10 Given some CSn -module M , with action written as π · m (π ∈ Sn , m ∈ M ). Define a new CSn -module M ε as the module whose underlying vector space is M and with new action π ∗ m := επ π · m (π ∈ Sn , m ∈ M ). (a) Verify that M ε is a CSn -module. (b) Prove that M is irreducible if and only if M ε is irreducible. n (c) Let χ be the character of M , and let χ(1 ) be the character of the Specht module n n S (1 ) . Prove that the character χε of M ε is given by χε (π) = χ(1 ) (π) for π ∈ Sn . n [Why do you now know that the trivial module S (n) and the sign module S (1 ) are the only 1-dimensional CSn -modules.] 11 Let ρ :GL(V ) →GLn (C) be an irreducible polynomial representation. Show that the matrix entries ρij ∈ C[GL(V )] are homogeneous polynomials of the same degree. More generall, show that every polynomial representation ρ is a direct sum of representations ρ(i) whose matrix entries are homogeneous polynomials of degree i. 12 Give a proof of the result proved in lectures that affine n-space An is irreducible without appealing to the Nullstellensatz. 13 Prove that the 1-dimensional rational representations of C∗ =GL1 (C) are power maps of the form u 7→ ur where r ∈ Z. 14 (a) Show that the set of diagonalisable matrices is Zariski dense in Mn (C) (hence every invariant function on Mn (C) is completely determined by its restriction to Tn , the diagonal matrices. [Actually this is nothing more than Jordan decomposition; if you work a bit harder you can show this is true over any field, not necessarily algebraically closed]. (b) Let ρ :GLn →GL(W ) be a rational representation. Show that ρ is polynomial if and only if ρ|Tn is polynomial. Show also that ρ is polynomial if and only if its character χρ is a polynomial. (c) Deduce that every 1-dimensional rational representation of GL(V ) is of the form r det :GL(V ) →GL1 , r ∈ Z. 15 Consider the linear action of GL(V ) on End(V ) by conjugation. (a) What are the orbits of thisP action? Any element α ∈End(V ) has a characteristic n polynomial of the form pα (t) = t + ni=1 (−1)i ei (α)tn−i , where n = dim V , showing that the ei are invariant polynomial functions on End(V ). (b) [HARD] Find a proof of the fact that the ring of invariants for the conjugation action is generated by the si without using properties of symmetric functions. (c) Show that the functions Tr1 . . . ,Trn generate the invariant ring C[End(V )]GL(V ) . Here Trk :End(V ) → C is the map A 7→TrAk , (k = 1, 2, . . .). SM, Lent Term 2013 Comments on and corrections to this sheet may be emailed to [email protected] MATHEMATICAL TRIPOS PART III, 2013 REPRESENTATION THEORY This sheet deals mostly with polynomial invariant theory. Work over C. 1 Let W be finite-dimensional over C. Check that the coordinate functions x1 , . . . , xm ∈ C[W ] are algebraically independent (equivalently, if f (a1 , . . . , am ) = 0 for a polynomial f and all a = (a1 , . . . , am ) ∈ Cm then f is the zero polynomial). Hint: use induction on the number of variables and the fact that a non-zero polynomial in one variable has only finitely many zeroes. 2 (a) Show that the natural representation of SL2 (C) on C2 is self-dual (i.e. equivalent to its dual representation). Hint: find a non-singular matrix P such that (A−1 )t = P AP −1 . (b) Show that this representation has two orbits. 3 Consider GL2 acting on C[X, Y ] (induced by the natural representation of GL2 on C2 ). (a) What is the image of X and Y under the action of any g ∈GL2 ? (b) Calculate C[X, Y ]GL2 and C[X, Y ]SL2 . (c) Calculate C[X, Y ]U where U is the subgroup of upper triangular unipotent matrices. 4 Consider the representation of SLn (C) in its action on Mn (C), as discussed in lectures. Prove that the invariant ring is generated by the determinant. You may need the fact that GLn is Zariski-dense in Mn . 5 Determine the invariant rings of C[M2 (C)]U and C[M2 (C)]T under left multiplication by the subgroup U of upper triangular unipotent matrices and the 2-torus T =diag(t, t−1 ) of diagonal matrices of SL2 . 6 Let Tn be the subgroup of GLn of non-singular diagonal matrices. Choose the standard basis in W = Cn and the dual basis in W ∗ and thus identify the coordinate ring C[W ⊕ W ∗ ] with C[x1 , . . . xn , ζ1 , . . . , ζn ]. Show that C[W ⊕ W ∗ ]Tn = C[x1 ζ1 , . . . , xn ζn ]. What happens if you replace Tn by the subgroup Tn0 of diagonal matrices with determinant 1? 1 7 Recall that for each j > 1 the symmetric functions nj (x) = xj1 + xj2 + · · · xjn are known as power sums (or Newton functions). As before the ej are the elementary symmetric functions. (a) Show that (−1)j+1 jej = nj − e1 nj−1 + e2 nj−2 − · · · + (−1)j−1 ej−1 n1 for all jQ= 1 . . . , n. This is known P as the Newton-Girard identity. [Hint: let j = n, consider f (t) = i (t − xi ) and calculate i f (xi ). For j < n what can you say about the right-hand side?] (b) Here is Weyl’s original proof1 of the identity in (a). Define the polynomial n Y ψ(t) = (1 − txi ) = 1 − e1 t + e2 t2 − · · · + (−1)n en tn , i=1 where the ei are the elementary symmetric functions. Determine its logarithmic derivative 0 (t) − ψψ(t) as a formal power series. Deduce (a). (c) Show that (in characteristic 0) the power sums generate the symmetric functions. 8 (a) Let A be a commutative algebra and let G be a group algebra of automorphisms of A. Assume the representation of G on A is completely reducible. Show that the subalgebra AG of invariants has a canonical G-stable complement and the corresponding G-equivariant projection π : A → AG satisfies π(hf ) = hπ(f ) for h ∈ AG , f ∈ A. This projection is sometimes called the Reynolds operator. (b) Let A = ⊕j>0 Aj be a graded K-algbbra (meaning Ai Aj ⊂ Ai+j ). Assume the idea A+ = ⊕j>0 Aj is finitely-generated. Show that A is finitely-generated as an algebra over A0 . (In fact if the ideal A+ is generated by the homogeneous elements a1 , . . . , an , then A = A0 [a1 , . . . , an ].) (c) Deduce Hilbert’s theorem: if W is a G-module and the representation of G on K[W ] is completely reducible then the invariant ring K[W ]G is finitely-generated. [Hint: you will need to use Hilbert’s Basis Theorem.] This shows also that K[W ]G is noetherian. (d) Hilbert’s Theorem is applicable to finite groups provided Maschke’s Theorem holds (i.e we require that |G| is invertible in K). In the special case of characteristic 0, show that for any representation W of G, the ring of invariants K[W ]G is generated by invariants of degree less than or equal to |G| (Noether’s Theorem). Schmidt introduced a numerical invariant β(G) for every finite group G. It is defined to be the minimal number m such that for every representation W of G, the invariant ring K[W ]G is generated by the invariants of degree 6 m. With this definition, Noether’s Theorem asserts that β(G) 6 |G|. In fact Schmidt showed that β(G) = |G| if and only if G is cyclic2. It is also a fact that if G is actually abelain then β(G) coincides with the so-called Davenport constant. In the next few questions we’ll write V = C2 , G =SL(V ) =SL2 (C), and Cn = HC,n (V, C) ∼ = (S n V )∗ ∼ = S n (V ∗ ), which can be identified with the set of homogeneous polynomials of degree n in variables X1 , X2 . Note that {Cn : n ∈ N} are the non-isomorphic simple CG-modules. For a fixed n, recall that a covariant for Cn is a polynomial CSL(V )-invariant Cn ⊕ V → C. The following questions sketch a classification of the generators for C[Cn ⊗ V )]G . 1Weyl, The Classical Groups, II.A.3 Finite groups and invariant theory in Séminaire d’Algèbre Paul Dubreil et M.-P. Malliavin (Springer Lecture Notes 1478, Springer 1991) 2see Suppose f, g are functions of X1 , X2 and r ∈ N. Define the rth transvectant of f, g as r X (−1)j ∂rf ∂rg τr (f, g) = r−j j j r−j . j!(r − j)! ∂X ∂X ∂X ∂X 1 2 1 2 j=0 r 1 ∂ ∂ ∂ ∂ = − (f (X1 , X2 )g(Y1 , Y2 ) |Y1 =X1 ,Y2 =X2 . r! ∂X1 ∂Y2 ∂X2 ∂Y1 You should observe τ0 (f, g) = f g, τ1 (f, g) is the Jacobian of f, g and τ2 (f, f ) is the Hessian of f . Show that (a) if f, g are homogeneous polynomials of degrees p, q then τr (f, g) = 0 unless r 6 min{p, q}, in which case it is a homogeneous polynomial of degree p + q − 2r; (b) if r 6 min{p, q} then the map τr : Cp ⊗ Cq → Cp+q−2r sending f ⊗ g 7→ τr (f, g) is a non-zero map of CG-modules; (c) any CG-module map Cp ⊗ Cq → Cp+q−2r is a multiple of τr . [Hint: use ClebschGordan]; (d) the τr give an isomorphism of CG-modules 9 min{p,q} Cp ⊗ Cq → M Cp+q−2r . r=0 10 Let p, q, n, r ∈ N and r 6 min{q, n}. Set N = max{0, r − p}. Show that there are scalars λN , . . . , λr ∈ C, with λr 6= 0, such that r X λj τj (τr−j (f, g), h), f τr (g, h) = j=N for all f ∈ Cp , g ∈ Cq , h ∈ Cn . [HINT: Schur’s Lemma.] Thus the τj are trying hard to be associative. 11 Fix n ∈ N. Let R = C[Cn ⊕ V ], so that RG is the set of covariants of Cn . If ϕ ∈ R and f ∈ Cn , define ϕ(f ) ∈ C[V ] by ϕ(f )(x) = ϕ(f, x). If r ∈ N and φ, ψ ∈ R define τr (φ, ψ) as the map Cn ⊕ V → C by (f, x) 7→ τr (φ(f ), ψ(f ))(x). Finally if d, i ∈ N, let Rdi be the set of all φ ∈ R which are homogeneous, of degree d in Cn and degree i in V . Show that G (a) Rdi are CG-submodules of R, and R is graded, R = ⊕∞ d,i=0 Rdi , hence so too is R ; (b) Rdi Rej ⊆ Rd+e,i+j and τr (Rdi , Rej ) ⊆ Rd+e,i+j−2r for r 6 min{i, j}; (c) the assignment φ 7→ (f 7→ φ(f )) induces an isomorphism RG ∼ = homCG,d (Cn , Ci ), di and deduce that G (d) any covariant θ ∈ Rdi (d > 1) can be expressed as a linear combination min{n,i} θ= X τn−r (φr , E) r=0 G with φr ∈ Rd−1,n+i−2r and E is the evaluation map E(f, v) = f (v) [you should note that (c) G says that R1,i = C.E when i = n and 0 if i 6= n.] 12 [Gordan’s Theorem3 - a weak form] If S is a C-subalgebra of RG with the property that τr (φ, E) ∈ S whenever r ∈ N and φ ∈ S, then S = RG . [Remark: after the last question the proof should reduce to a two- line argument.] 13 [REALLY HARD] This uses Gordan’s Theorem to find a set of generators of RG in the case n = 3. Given n = 3, show that RG is generated by the covariants G ; E ∈ R13 G ; H = τ2 (E, E), the Hessian, ∈ R22 G t = τ1 (H, E) ∈ R33 ; G D = τ3 (t, E), the discriminant (×48), ∈ R40 14 [REALLY REALLY HARD4 ] Take n = 4 and repeat the last question to generate RG using quartic forms. Classically, the (still unsolved) problem of computing all covariants of binary forms of degree n was tackled by something called the symbolic method (using polarisation to reduce it to the FFT). Associated to the symbolic method is the symbolic notation, designed to make the calculations easier, but still non-trivial. You can consult any old text on Invariant Theory, such as Grace and Young if interested. SM, Lent Term 2013 Comments on and corrections to this sheet may be emailed to [email protected] 3Paul Gordan, dubbed the ‘King of Invariant Theory’ is perhaps better known for being Emmy Noether’s thesis advisor. The result appeared in 1868 in Crelle’s Journal as Beweis, dass jede Covariante und Invariante einer binären Form eine ganze Funktion mit numerischen Coeffizienten einer endlichen Anzahl solcher Formen ist. 4J.J. Sylvester’s collected works (four volumes, edited by H.F. Baker) include tables with details about higher degree forms. However his published tables for forms of degree larger than 6 appear to be totally wrong. Corrections appear in von Gall (1888) and Dixmier/Lazard (1986), Shioda (1967) and Brouwer/Popoviciu (2010).
© Copyright 2026 Paperzz