Spec. Matrices 2016; 4:80–100 Research Article Open Access Ajda Fošner* and Bojan Kuzma Preserving zeros of Lie product on alternate matrices DOI 10.1515/spma-2016-0009 Received July 8, 2015; accepted December 13, 2015 Abstract: We study continuous maps on alternate matrices over complex field which preserve zeros of Lie product. Keywords: Alternate matrix, Lie product, general preserver MSC: 15A86, 15A99, 15B99 1 Introduction and the main theorem Linear preserver problem is the problem which concerns the characterization of linear operators on matrix algebras as well as on more general rings and operator algebras that leave certain functions, subsets, relations, etc., invariant. It is one of the most active subjects in the last few decades (for surveys of the topic we refer the reader to the papers [13, 14, 24]). This kind of problems arise in most parts of mathematics because in many cases the corresponding results provide important information on the automorphisms of the underlying structures. In the last few decades a lot of results on linear preservers and also general (non-linear) preservers have been obtained (see [16]). One among the important preserver problems is classifying maps that preserve commutativity (see, for example, [2, 3, 5, 18–22, 25–27] and references therein). Part of the rationale behind studying this kind of problems is the fact that in associative algebras quite a few elements do commute (for example, every polynomial in x commutes with x). Even more importantly, the assumption of preserving commutativity can be considered as the assumption of preserving zero Lie products. Because of applications in quantum mechanics it is also interesting to study the problem of characterizing general commutativity preserving maps. These are the maps that preserve zeros of Lie product and are not assumed to satisfy any additional algebraic assumption like additivity (see e.g. [17]). Here, let us also mention two papers [4] and [6] dealing with maps preserving zeros of a usual product. In the present paper we will be interested in describing commutativity preserving maps on the algebra of all n × n alternate complex matrices. We will only assume that the map is injective and continuous but will not presume it is linear. Let us remark that our results are in the spirit of Šemrl [26] who studied injective continuous maps which preserve zeros of Lie product on complex matrices. Moreover, we emphasize that without imposing some additional regularity conditions on the map, like continuity, we cannot hope for a nice structural result (see also [25, 26]). Let us list some mostly standard notation. Throughout, n ≥ 3 will be an integer and M n will be the algebra of all n × n matrices over the field of complex numbers. Let E ij be the standard basis of M n . By Altn ⊂ M n we will denote the subspace of all alternate matrices (i.e., matrices with the property A t = −A). Here A t denotes *Corresponding Author: Ajda Fošner: Ajda Fošner, University of Primorska, Faculty of Management, Cankarjeva 5, SI-6000 Koper, Slovenia, E-mail: [email protected] Bojan Kuzma: Bojan Kuzma, University of Primorska, FAMNIT, Glagoljaška 8, SI-6000 Koper and Institute of Mathematics, Physics and Mechanics, Department of Mathematics, Jadranska 19, SI-1000 Ljubljana, Slovenia, E-mail: [email protected] © 2016 Ajda Fošner and Bojan Kuzma, published by De Gruyter Open. This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivs 3.0 License. Preserving zeros of Lie product on alternate matrices | 81 the transpose of a matrix A ∈ M n . For A = (a ij ) ∈ M n we will denote by A = (a ij ) = (a ij ) the conjugate of A. As usual, we use the notation diag(a1 , . . . , a k ) to denote the k × k diagonal matrix with diagonal entries a1 , . . . , a k . We will study injective continuous maps ϕ with the property AB = BA ϕ(A)ϕ(B) = ϕ(B)ϕ(A). =⇒ This condition can be written as [A, B] = 0 [ϕ(A), ϕ(B)] = 0, =⇒ where [A, B] = AB − BA is the Lie product. Note that alternate matrices are closed under the Lie product. We will represent vectors x ∈ Cn as n × 1 complex matrices. Note that the standard basis e1 , e2 , . . . , e n of n C is the set of all n × 1 matrices having all entries equal to zero but one that is equal to one. If x, y ∈ Cn are two linearly independent nonzero vectors, then xy t − yx t = x ∧ y is a rank two alternate matrix. Every rank two alternate matrix can be written in this form. A matrix A is said to be nonderogatory if every eigenvalue of A has geometric multiplicity one. Let us point out that a matrix B ∈ M n commutes with a nonderogatory matrix A ∈ M n if and only if there is a complex polynomial p such that B = p(A) (see [12, p. 135]). The main idea behind our proof is to utilize the Brouwer’s invariance of domain theorem [10, p. 344] stating that if U is an open subset of Rm and F : U → Rm is a continuous injective map, then F(U) is open. In particular, there is no injective continuous map from Rk into Rm whenever m < k. We acknowledge that the same idea was already used before, say in Petek and Šemrl’s characterization of continuous maps which preserve adjacency of matrices in one direction (see [23]). Later, the same idea was used in [26] and [9]. The following is a basis for our arguments in Section 2 and Section 3. If x t Ax = 0 for every column vector x ∈ Cn , a complex matrix A is alternate. Each alternate matrix is congruent to a block-diagonal matrix. More precisely, there exists an invertible matrix S such that SAS t = m ⨁︁ J ⊕ 0n−2m , i=1 where (︃ J := 0 −1 )︃ 1 . 0 Consequently, the rank of each alternate matrix is an even integer and a minimal possible nonzero rank is 2. ∑︀ ∑︀m Moreover, since SAS t = m i=1 E (2i−1) (2i) − E (2i) (2i−1) = i=1 e 2i−1 ∧ e 2i , we have A= m ∑︁ xi ∧ yi i=1 for some linearly independent vectors x i = S−1 e2i−1 and y i = S−1 e2i , i = 1, . . . , m. We will need some further properties of alternate matrices over the complex field. We borrow the following notations and facts from the book by Gantmacher, [11, p. 14–21]. In matrices below, all non-specified entries are zero. Firstly, given an integer k ≥ 2, let ⎛ ⎞ 1 k ∑︁ ⎜ ⎟ . ⎟ ∈ Mk V k := E(k−i+1) i = ⎜ (1) .. ⎝ ⎠ i=1 1 be an anti-diagonal k × k matrix and let J k := k−1 ∑︁ i=1 E i (i+1) ⎛ 0 ⎜ ⎜ ⎜ =⎜ ⎜ ⎝ ⎞ 1 .. . .. . .. . ⎟ ⎟ ⎟ ⎟ ⎟ 1⎠ 0 82 | Ajda Fošner and Bojan Kuzma be an elementary k × k Jordan upper-triangular nilpotent. For odd integers q we also define an alternated version of J q , that is, a matrix q−1 J (q) := 2 ∑︁ E i (i+1) − q−1 ∑︁ E i (i+1) i= q+1 2 i=1 ⎛ 0 ⎜ ⎜ ⎜ ⎜ ⎜ ⎜ ⎜ =⎜ ⎜ ⎜ ⎜ ⎜ ⎜ ⎝ 1 .. ⎞ . .. . .. . 1 0 −1 .. . .. . .. . ⎟ ⎟ ⎟ ⎟ ⎟ ⎟ ⎟ ⎟ ∈ Mq . ⎟ ⎟ ⎟ ⎟ ⎟ −1⎠ 0 Note that J (q) is an upper-triangular q × q nilpotent having precisely q−1 2 of +1 and the same number of −1 on the first upper-diagonal. Note also that J (q) and J q are similar. Moreover, let Idk denotes the k × k identity matrix and, given an integer p and a scalar λ, we define the (2p) × (2p) matrix K (pp) by λ K (pp) λ := 1 2 (︀ (︃ √ )︀ Id2p − −1V2p · λ Idp +J p 0 0 −λ Idp −J p )︃ √ (︀ )︀ · Id2p + −1V2p . (2) (pp) Notice that K (pp) is similar to K−λ , hence by [11, Theorem 4, p. 9] they are also orthogonally similar, that is, λ there exists a matrix Q such that (pp) (pp) −1 K−λ = QK−λ Q and Q−1 = Q t . (3) Secondly, given an odd integer q we define K (q) := 1 2 (︀ √ √ )︀ (︀ )︀ Idq − −1V q · J (q) · Idq + −1V q . (4) When q = 1 we let K (1) be a zero 1 × 1 matrix. Example 1.1. The exact appearance of matrices defined by formulas (3)–(4) will play no role in the present paper. For convenience we only show K (44) and K (7) , here i2 = −1. λ ⎛ K (44) λ 0 ⎜ i ⎜ ⎜ ⎜ 0 ⎜ 0 1⎜ = ⎜ 2⎜ ⎜ 0 ⎜ ⎜ 0 ⎜ ⎝ −1 2iλ ⎛ K (7) −i 0 i 0 0 −1 2iλ −1 0 ⎜ −1 ⎜ ⎜ ⎜0 1⎜ = ⎜ 0 2⎜ ⎜0 ⎜ ⎜ ⎝ −i 0 0 −i 0 i −1 2iλ −1 0 1 0 −1 0 −i 0 −i 0 0 −i 0 2iλ −1 0 0 0 1 0 −1 − i 0 −i 0 0 0 1 −2iλ 0 −i 0 0 0 0 1+i 0 1−i 0 0 0 1 −2iλ 1 i 0 −i 0 0 i 0 −1 + i 0 1 0 1 −2iλ 1 0 0 i 0 −i i 0 i 0 −1 0 1 ⎞ −2iλ 1 ⎟ ⎟ ⎟ 0 ⎟ ⎟ 0 ⎟ ⎟, 0 ⎟ ⎟ ⎟ 0 ⎟ ⎟ i ⎠ 0 ⎞ 0 ⎟ i ⎟ ⎟ 0⎟ ⎟ 0⎟ ⎟ 0⎟ ⎟ ⎟ −1⎠ 0 Let d ≥ 1 be an integer. By induction on d, it is easy to prove the following lemma. Preserving zeros of Lie product on alternate matrices | 83 Lemma 1.2. Let q = 2k + 1 be an odd integer. Then (︃ )︃ 0(q−d)×d D (q) d (J ) = , 0d×d 0d×(q−d) where D= ⎧ 2 d−1 d ⎪ ⎪ ⎨Idk−d+1 ⊕ diag(−1, (−1) , . . . , (−1) ) ⊕ (−1) Idk−d+1 , (−1) ⎪ ⎪ ⎩0 , d−k+1 2 2k−d+1 diag(−1, (−1) , . . . , (−1) ), d≤k k < d ≤ 2k . d ≥ 2k + 1 √ √ √ √ )︀t )︀ )︀2 One easily verifies that (Idk − −1V k = (Idk − −1V k is symmetric with (Idk − −1V k = −2 −1V k and √ √ (︀ )︀ (︀ )︀ (q) 1 is similar to J q because they are both nilpotents with 2 Idk − −1V k · Idk + −1V k = Idk . Moreover, J maximal nilindex. Hence, it is easy to see that K (pp) and K (q) are alternate matrices which are similar to (J p ⊕ λ J p ) and to J q , respectively. In a sense, these are the only possible cases. Namely, by [11, Corollary on p. 21], ⨁︀ 1 (p i p i ) any alternate matrix is orthogonally similar to a block-diagonal alternate matrix of the form ni=1 K λi ⊕ ⨁︀n2 (q j ) t K . More precisely, if A ∈ M is alternate, then there exists a matrix Q ∈ M with Q Q = Id and there n n n j=1 exist integers n1 ≥ 0, n2 ≥ 0, integers p1 , . . . , p n1 and q1 , . . . , q n2 , and scalars λ1 , . . . , λ n1 such that QAQ−1 = QAQ t = n1 ⨁︁ i pi ) K (p ⊕ λi i=1 n2 ⨁︁ K (q j ) . (5) j=1 If n1 = 0 or n2 = 0, we omit the corresponding summand. We are now ready to state our main result. Recall that a polynomial p is called odd if p(−λ) = −p(λ) for each λ. Main Theorem. Let n ≥ 15 and let ϕ : Altn → Altn be an injective continuous map that preserves zeros of Lie product. Then there exists an orthogonal matrix Q ∈ M n and for each A ∈ Altn there is an odd polynomial p A (λ) such that either ϕ(A) = Qp A (A)Q t for all A ∈ Altn , or ϕ(A) = Qp A (A)Q t for all A ∈ Altn . Remark 1.3. We believe that the above theorem is true also for matrices of lower dimensions n = 5, . . . , 14, but we have not proved this yet. Note also that in the case n = 4 we have at least one additional map. Namely, it can be shown that a map Φ : Alt4 → Alt4 , which swaps positions (1, 4), (2, 3) and (4, 1), (3, 2) and leaves the other entries intact, preserves zeros of Lie product. 2 Preliminaries Let us denote by A′ = {X ∈ Altn : AX = XA} the alternate commutant (a.k.a. a centralizer) of a matrix A ∈ Altn and by Sp(A) the spectrum of a matrix A. We will start with a well-known Lumer–Rosenblum theorem [15] (see also [1, p. 2] for a short proof valid in general Banach algebras). Lemma 2.1. Suppose that A ∈ M n and B ∈ M m . If 0 ∉ Sp(A) + Sp(B) = {α + β : α ∈ Sp(A), β ∈ Sp(B)}, then the equation AX + XB = 0 has only a zero solution. We proceed with a series of lemmas. 84 | Ajda Fošner and Bojan Kuzma Lemma 2.2. If A = K (q) with q odd, then dim A′ = (q−1) 2 . Proof. Note that A is nonderogatory. Hence, its ordinary commutant (i.e., {X ∈ M q : AX − XA = 0}) equals to Poly(A), i.e., the set of all polynomials in A [12, Theorem 3.2.4.2.]. For the alternate commutant, we must pick out those X ∈ Poly(A) which are alternate. Now, if X = p(A) = α0 Id +α1 A + · · · + α q−1 A q−1 is alternate, then ∑︀ q−1 2 clearly Y = α1 A + α3 A3 + α5 A5 + · · · + α q−2 A q−2 is also alternate and so is X − Y = i=0 α2i A2i . But note that q−1 q−1 ∑︀ ∑︀ 2 2 (X − Y)t = i=0 α2i (A t )2i = i=0 α2i (−A)2i = X − Y. So, α0 = α2 = · · · = α q−1 = 0. But then dim A′ = (q−1) 2 . Lemma 2.3. Let p ≥ 1 be any integer. Then the following holds. (︀ )︀′ = p. (i) If λ ≠ 0, then dim K (pp) λ (︀ )︀′ = 2p. (ii) If λ = 0 and p is even, then dim K (pp) λ (︀ (pp) )︀′ = 2p − 1. (iii) If λ = 0 and p is odd, then dim K λ Proof. Suppose that X ∈ Alt2p satisfies K (pp) X − XK (pp) = 0. Using the definition of K (pp) , we arrive at λ λ λ (︀ (︀ )︀ (︀ )︀ (λ Idp +J p ) ⊕ (−λ Idp −J p ) · S−1 XS − S−1 XS · (λ Idp +J p ) ⊕ (−λ Idp −J p ) = 0 √ (︃ −1 )︀ for S = Id2p − −1V2p . We now decompose S XS = X11 X21 X12 X22 (6) )︃ into a 2 × 2 block matrix. Comparing the individual blocks of (6) we derive four equations (λ Idp +J p )X11 − X11 (λ Idp +J p ) = 0, (λ Idp +J p )X12 + X12 (λ Idp +J p ) = 0, (7) (λ Idp +J p )X21 + X21 (λ Idp +J p ) = 0, (λ Idp +J p )X22 − X22 (λ Idp +J p ) = 0. The first and the last equations simplify into J p X11 = X11 J p and J p X22 = X22 J p , respectively. Since J p is nonderogatory we see that X11 and X22 have to be polynomials in J p . This yields that (︃ )︃ f11 (J p ) X12 −1 S XS = , X21 f22 (J p ) where f1 , f2 are complex polynomials and X12 , X21 satisfy (7). Now, recall that X is alternate. Due to S t = S, √ (︀ )︀2 this gives (S−1 XS)t = S t · (−X) · (S−1 )t = −SXS−1 = −S2 (S−1 XS)S−2 . On the other hand, S2 = Id2p − −1V2p = √ −2 −1V2p , and S−2 = √ −1 2 V 2p . (︃ With this in hand we establish )︃ (︃ )︃t t X21 f11 (J p ) X12 = f22 (J p )t X21 f22 (J p ) (︃ )︃ f (J p ) X12 = −V2p 11 V2p X21 f22 (J p ) (︃ )︃ (︃ )︃ (︃ 0 Vp f11 (J p ) X12 0 =− · · Vp 0 X21 f22 (J p ) Vp (︃ )︃ V p f22 (J p )V p V p X21 V p =− . V p X12 V p V p f11 (J p )V p f11 (J p )t t X12 Vp 0 )︃ (8) Comparing the (1, 1) block, we get that f11 (J p ) = −V p f22 (J p )t V p . Since Z ↦→ V p f22 (Z)t V p is a flip map (i.e., reflection over anti-diagonal), we easily conclude V p f22 (J p )t V p = f22 (J p ). Hence, f11 (J p ) = −f22 (J p ). In the case of off-diagonal blocks we are facing two options. Preserving zeros of Lie product on alternate matrices | 85 Case 1. Assume λ ≠ 0. Then, by Lemma 2.1, the second equation and the third equation of (7) have only zero solution. So, X12 = 0 = X21 and, therefore, {︃ (︃ )︃ }︃ 0 (pp) ′ −1 f 11 (J p ) (K λ ) = S S : f11 ∈ C[λ] (9) 0 −f11 (J p ) is of dimension p. Case 2. Assume λ = 0. Then, the second equation of (7) simplifies into J p X12 + X12 J p = 0 with a solution X12 = n−1 ∑︁ αj n−j ∑︁ j=0 (−1)i E i (i+j) . i=1 However, X being alternate further forces −X12 = (V p X12 V p )t , by comparing the (2, 1) entries of (8). Note that the right-hand side is a flip map. It is now easy to verify that, with p even, one has α1 = 0 = α3 = · · · = α p−3 = α p−1 , and, with p odd, one has α0 = 0 = α2 = · · · = α p−3 = α p−1 . Therefore, with p even, at block position (1, 2) of X we have 2p linearly independent matrices, while with p odd we have p−1 2 linearly independent matrices. We can duplicate the calculations also for X21 . We already know that the diagonal blocks are of the form described in (9). Combined, we have ⎧ ⎨p + 2 · (︀ p )︀ = 2p , p is even (pp) ′ (︁2 )︁ . dim(K λ ) = ⎩p + 2 · p−1 = 2p − 1 , p is odd 2 Let n ≥ 3 and A ∈ Altn . We would like to estimate the dimension of A′ . Using orthogonal similarity, we may assume that A is alternate and block-diagonal, say A= n1 ⨁︁ K (q i ) ⊕ i=1 n2 ⨁︁ (p j p j ) K0 j=1 ⊕ n3 ⨁︁ k pk ) K (p , λ k k=1 (︀ )︀ with q i ≥ 1 odd, p j , p k ≥ 1, and λ1 , . . . , λ n3 nonzero. Let us write X = X ij ∈ A′ into block form. Then it is (︀ )︀′ (︀ (p p ) )︀′ (︀ k p k ) )︀′ easy to see that the diagonal blocks satisfy X ii ∈ K (q i ) , X jj ∈ K0 j j , and X kk ∈ K (p . By the previous λk lemmas we have qi − 1 , 2 {︃ 2p j , (p p ) dim(K0 j j )′ = d(p j ) = 2p j − 1 , dim(K (q i ) )′ = p j even p j odd , k pk ) ′ dim(K (p ) = pk . λ k Thus, the dimension of diagonal blocks of matrices from A′ equals n3 n2 n1 ∑︁ ∑︁ q i − 1 ∑︁ + d(p j ) + pk . 2 i=1 j=1 k=1 It remains to calculate the dimension of off-diagonal blocks of matrices from A′ . First, consider X ii′ , 1 ≤ i < i′ ≤ n1 . This block must satisfy the equation K (q i ) X ii′ = X ii′ K (q i′ ) . Note that (q i ) K is similar to an elementary Jordan upper-triangular matrix J q i . It is easy to see that the dimension of the solutions of the above equation equals the dimension of the solutions of a similar equation J q i Y = YJ q i′ . Also, one easily computes that this dimension is min{q i , q i′ }. (p p ) Next, consider X ij , 1 ≤ i ≤ n1 , 1 ≤ j ≤ n2 . This block must satisfy the equation K (q i ) X ij = X ij K0 j j . (p p ) Note that K0 j j is similar to J p j ⊕ J p j , so the dimension of the solutions of the above equation equals the dimension of the solutions of the equation J q i Y = Y(J p j ⊕ J p j ). Writing Y = (Y1 |Y2 ) we easily reduce this case to the previous case and find that the dimension of the solutions is 2 min{q i , p j }. 86 | Ajda Fošner and Bojan Kuzma (p ′ p ′ ) (p p ) Further, consider X jj′ , 1 ≤ j < j′ ≤ n2 . This block must satisfy the equation K0 j j X ij = X ij K0 j j . Arguing as above, we reduce this equation to the equation (J p j ⊕ J p j )Y = Y(J p j′ ⊕ J p j′ ), which easily establishes that the dimension of the solutions equals 4 min{p j , p j′ }. Now, consider X ik or X jk , 1 ≤ i ≤ n1 , 1 ≤ j ≤ n2 , 1 ≤ k ≤ n3 . These blocks must be zero because in the equation they are multiplied with a nilpotent matrix on the left and with an invertible matrix on the right side. By the theory of elementary operators, this yields that X ik = 0 = X jk . At the end, consider X kk′ , 1 ≤ k < k′ ≤ n3 . If λ k = λ k′ , then we easily reduce this case to the previous cases to find that the dimension of these blocks equals 4 min{k, k′ }. On the other hand, if λ ≠ λ′ , then X kk′ = 0. Consequently, dim A′ ≤ n3 n2 n1 ∑︁ ∑︁ ∑︁ ∑︁ q i − 1 ∑︁ + d(p j ) + pk + min{q i , q i′ } + 2 min{q i , p j } 2 ′ 1≤i≤n j=1 i=1 + ∑︁ k=1 4 min{p j , p j′ } + 1≤j<j′ ≤n2 ∑︁ 1≤i<i ≤n1 1 1≤j≤n2 (10) 4 min{p k , p k′ }. 1≤k<k′ ≤n3 Lemma 2.4. Let n ≥ 3 and A ∈ Altn . Then A′ = Altn if and only if A = 0. Proof. If A = 0, then, of course, A′ = Altn . Now, suppose that A ≠ 0. According to (10), we can easily see that and, therefore, A′ ≠ Altn . in this case dim A′ < n(n−1) 2 Lemma 2.5. Let n ≥ 9 and A ∈ Altn be a nonzero matrix. Then A is of minimal rank if and only if dim A′ ≥ (n−2)(n−3) + 1. 2 Proof. Let A ∈ Altn be a nonzero alternate matrix. Using orthogonal similarity, we may write A in a blockdiagonal form as in the previous investigations. Now, if rk A = 2 (here, rk A denotes the rank of the matrix A), then either A = K (3) or A = K (11) , λ ≠ 0, or λ (22) A = K0 . In the first case we have n1 = n − 2, q1 = 3, q2 = · · · = q n1 = 1, while n2 = 0 = n3 . By the remarks + 1. Likewise we argue in the second case, where above, it is easy to see that in this case dim A′ = (n−2)(n−3) 2 (n−4)(n−5) ′ dim A′ = (n−2)(n−3) + 1, and in the last case, where dim A = + 2(n − 4) + 4 = (n−4)(n−1) + 4. 2 2 2 It remains to show that if rk A > 2, then the dimension of A′ is strictly smaller than (n−2)(n−3) + 1. 2 Let m be a dimension of all nilpotent blocks in A. Then we can write A = A1 ⊕ A2 ∈ Altm ⊕ Altn−m , where A1 is nilpotent and A2 is invertible. It is easy to show that A′ = A′1 ⊕ A′2 ⊆ Altm ⊕ Altn−m which gives dim A′ ≤ dim Altm + dim Altn−m = m(m − 1) (n − m)(n − m − 1) + . 2 2 This is strictly below (n−2)(n−3) + 1 whenever 3 ≤ m ≤ n − 3 because n ≥ 9. Recall that if m = 2, then dim A′ ≤ 2 (n−2)(n−3) dim Alt2 +(dim Altn−2 −1) < + 1 since A′2 = Altn−2 is possible only when A2 = 0. Note also that we 2 cannot have m = n − 1 since A is alternate. Hence, the equality holds only if m ∈ {0, 1, n − 2, n}. Case m = 0. Note that in this case A is invertible. Let us fix an eigenvalue λ ∈ Sp(A) and write A block⨁︀ p i p i diagonally as A = A1 ⊕ A2 , where A1 = K λ is the sum of alternate blocks with eigenvalues −λ, λ, while Sp(A2 ) ∩ {−λ, λ} = ∅. Let m1 be the dimension of A1 . Observe that m1 ∉ {0, 1, n − 1} because A is invertible. + 1, as desired. The same is true if If 3 ≤ m1 ≤ n − 3, then, arguing as above, we get that dim A′ < (n−2)(n−3) 2 either m1 = 2 (by Lemma 2.4, 0 ≠ A2 is of dimension n − 2 > 2) or m1 = n − 2 (in this case we just change ⨁︀s ∑︀ (p k p k ) the roles of A1 and A2 ). At the end, if m1 = n, then we can write A = with sk=1 2p k = n and k=1 K λ 1 ≤ p1 ≤ · · · ≤ p s . We already know (according to previous observations) that the k-th diagonal block of A′ has dimension p k and the (k, h), k < h, off-diagonal block of A′ has dimension min{2p k , 2p h }. Therefore, dim A′ = s ∑︁ k=1 pk + ∑︁ 1≤k<h≤s min{2p k , 2p h } = 2sp1 + 2(s − 1)p2 + · · · + 2p s − n . 2 Preserving zeros of Lie product on alternate matrices | 87 Note that sp1 ≤ p1 + · · · + p s = 2n and (s − 1)p2 ≤ p2 + · · · + p s = 2n − p1 ≤ 2n − 1 and, inductively, (s − k + 1)p k ≤ p k + · · · + p s = 2n − p1 − · · · − p k−1 ≤ 2n − (k − 1) for every 1 ≤ k ≤ s. Therefore, (︀ )︀ n dim A′ =2 sp1 + (s − 1)p2 + · · · + p s − 2 (︀ n )︀ n n n ≤2 2 + ( 2 − 1) + · · · + ( 2 − s + 1) − 2 n 2 )︁ n n2 (︁∑︁ k − = . ≤2 2 4 k=1 But this is strictly below (n−2)(n−3) 2 + 1 whenever n ≥ 9. Case m = 1. In this case we can write A = 0 ⊕ A2 ∈ 0 ⊕ Altn−1 , where A2 is invertible. As above, we easily see that dim A′ = dim A′2 < ((n−1)−2)((n−1)−3) + 1 ≤ (n−2)(n−3) + 1, as desired. 2 2 Case m = n − 2. In this case we can write A = A1 ⊕ K (11) , where A1 is nonzero. Observe that A′ ⊂ Altn−2 ⊕ Alt2 . λ Moreover, the inclusion is strict because A1 ≠ 0. Hence, dim A′ < dim Altn−2 +1 = (n−2)(n−3) + 1, as claimed. 2 ⨁︀n1 (q i ) ⨁︀n2 (p j p j ) Case m = n. In this case A is nilpotent and we can decompose it into A = ⊕ with i=1 K j=1 K 0 q1 ≥ q2 ≥ · · · ≥ q n1 odd and p1 ≥ p2 ≥ · · · ≥ p n2 . (q1 ) ⊕ A2 , where A2 contains all other nilpotent blocks of A. If X = (︃ If q1 ≥ 5, )︃ then we have A = K X11 X12 ∈ A′ , then X11 ∈ (K (q1 ) )′ , X22 ∈ A′2 ⊆ Altn−q1 , and K (q1 ) X12 = X12 A2 . To compute the diment −X12 X22 ⨁︀ 1 (q i ) sion of (1, 2) block in A′ , we decompose X12 into blocks according to the block structure of A2 = ni=2 K ⊕ ⨁︀n2 (p j p j ) ∑︀n1 K . By previous observations, we see that the dimension of (1, 2) block equals min { q 1 , qi } + i=2 0 ∑︀nj=1 ∑︀n1 ∑︀n2 ′ 2 j=1 2 min{ q 1 , p j } ≤ i=2 q i + j=1 2p j = n − q 1 . Hence, we may estimate the dimension of A as follows q1 − 1 (n − q1 )(n − q1 − 1) + + (n − q1 ). 2 2 dim A′ ≤ dim(K (q1 ) )′ + dim Altn−q1 +(n − q1 ) = This is a strictly decreasing function of q1 ≤ n and it is strictly smaller than (n−2)(n−3) + 1 whenever q1 > 3. So, 2 we are done in this case. If p1 ≥ 3, then we can write A = K0(p1 p1 ) ⊕ A2 , where A2 contains all other nilpotent blocks of A. Proceeding as above, we see that n1 n2 ∑︁ ∑︁ dim A′ ≤ dim(K0(p1 p1 ) )′ + dim Altn−2p1 +( 2 min{p1 , q i } + 4 min{p1 , p j }) i=1 ≤2p1 + =(2p1 + (n − 2p1 − 1)(n − 2p1 ) +( 2 n2 ∑︁ 2p j + n1 ∑︁ i=1 j=2 =n + (n − 2p1 ) + n1 ∑︁ j=2 2q i + i=1 n2 ∑︁ 4p j ) j=2 n2 n1 ∑︁ ∑︁ (n − 2p1 − 1)(n − 2p1 ) 2p j + qi ) + qi ) + ( 2 j=2 i=1 (n − 2p1 − 1)(n − 2p1 ) . 2 Note that this is a strictly decreasing function of p1 ∈ [3, 2n ] and it takes a maximum at p1 = 3. Since n ≥ 9 it follows that the maximum is strictly smaller than (n−2)(n−3) + 1. Hence, dim A′ < (n−2)(n−3) + 1. 2 2 Now, suppose that there are at least two nonzero blocks in the decomposition of A. By the above, we may assume that all blocks in A are of dimension at most 4. First, let A be orthogonal similar to A = K0(22) ⊕ K0(22) ⊕ A2 . Then dim A′ ≤4 + 4 + n1 n2 n1 n2 ∑︁ ∑︁ ∑︁ (n − 8)(n − 9) ∑︁ +( 2 min{2, q i } + 4 min{2, p j }) + ( 2 min{2, q i } + 4 min{2, p j }) 2 i=1 ≤(4 + 4 + n1 ∑︁ i=1 qi + n2 ∑︁ j=2 j=2 i=1 n1 n2 n1 n2 ∑︁ ∑︁ ∑︁ ∑︁ (n − 8)(n − 9) 2p j ) + ( qi + 2p j ) + ( 2q i + 4p j ) + 2 i=1 j=3 i=1 j=2 j=3 88 | Ajda Fošner and Bojan Kuzma =n + (n − 8) + 2(n − 4) + (n − 8)(n − 9) . 2 It is easy to see that this is strictly smaller than (n−2)(n−3) + 1 for n ≥ 9. 2 Next, let A be orthogonally similar to K (3) ⊕ K0(22) ⊕ A2 . In this case n ∑︁ (︀ )︀ dim A′ ≤ dim K (3) + dim K0(22) + dim Altn−7 + 2 min{3, 2} + min{3, 1} + 2(n − 7) i=7 (n − 7)(n − 8) + 4 + 3(n − 7) =1 + 4 + 2 (n − 7)(n − 2) +9 = 2 and it is easy to see that this is strictly smaller than (n−2)(n−3) + 1 for n ≥ 9. 2 Assume lastly that A is orthogonally similar to K (3) ⊕ K (3) ⊕ A2 . In this case n n ∑︁ (︀ )︀ ∑︁ dim A′ ≤ dim K (3) + dim K (3) + dim Altn−6 + min{3, 3} + min{3, 1} + min{3, 1} i=6 i=6 (n − 6)(n − 7) =1 + 1 + + 3 + 2(n − 6) 2 (n − 6)(n − 3) = +5 2 and this is strictly smaller than (n−2)(n−3) 2 + 1 for n ≥ 9. The proof is completed. We will also require an upper bound for dimension of the alternate commutant of rank-four alternate matrices. Lemma 2.6. Let n ≥ 15 and let A ∈ Altn be of rank-four. Then, dim A′ (n−3)(n−4) 2 Proof. If A ∈ Altn , then, from its Jordan decomposition and the fact that each block of a block-diagonal alternate matrix is itself alternate and hence of even rank, we see that rk A = 4 if and only if, up to orthogonal similarity, (i) A = K (22) ⊕ 0n−4 for λ ≠ 0, (ii) A = K0(33) ⊕ 0n−6 , (iii) A = K (5) ⊕ 0n−5 , (iv) A = K0(22) ⊕ K0(22) ⊕ 0n−8 , λ (11) (22) (v) A = K λ ⊕ K0 ⊕ 0n−6 with λ ≠ 0, (vi) A = K (11) ⊕ K (11) ⊕ 0n−4 for λ, µ ≠ 0, (vii) A = K (3) ⊕ K (3) ⊕ 0n−6 , µ λ (2,2) (11) (3) (3) (viii) A = K ⊕ K0 ⊕ 0n−7 , or (ix) A = K ⊕ K λ ⊕ 0n−4 with λ ≠ 0. Now, it is easy to see that in each of the cases (i)–(ix), the dimension of an alternate commutant is a quadratic polynomial in n. (︃ )︃ Namely, it is an elementary that, with an alternate block matrix X kept fixed, ′ X * (X ⊕ 0n−i )′ = , where the dimension of a rectangular space of matrices * is a linear function of −*t Altn−i n. So, after some calculations we see that under (i), dim A′ = (n−9)n+24 , under (ii), dim A′ = (n−9)n+28 , under 2 2 (n−9)n+24 (n−9)n+40 (n−9)n+28 ′ ′ ′ (iii), dim A = , under (iv), dim A = , under (v), dim A = , under (vi), dim A′ ≤ 2 2 2 (11) ′ (11) dim(K λ ⊕ K λ ) = (n−9)n+28 , under (vii), dim A′ = (n−9)n+28 , under (viii), dim A′ = (n−9)n+32 , and under (ix), 2 2 2 (n−9)n+24 ′ dim A = . From here, one easily finds that the matrix from the case (iv) has the maximal dimension 2 of its alternate commutant among all matrices in (i)–(ix). This dimension equals (n−9)n+40 and, since n ≥ 15, 2 this is smaller than (n−3)(n−4) , as claimed. 2 Lemma 2.7. The set of all diagonalizable alternate matrices is dense in Altn . Proof. Given an odd integer q, the matrix L(q) = √ √ 1 (Idq − −1V q )(−J (q) )t (Idq + −1V q ) ∈ M q 2 is alternate. Namely, if we multiply alternate matrix K (q) = √ √ 1 (Idq − −1V q )J (q) (Idq + −1V q ) 2 Preserving zeros of Lie product on alternate matrices | 89 with symmetric matrix V q on the left and on the right, and use V q J (q) V q = −(J (q) )t we find L(q) = V q K (q) V q . Moreover, also X = (L(q) )q−2 is alternate because alternate matrices are closed under forming odd powers. Hence, given ε > 0, note that a matrix K (q) + εX is alternate and it is similar to J (q) + ε(−(J (q) )t )q−2 = J (q) + q+1 (−1) 2 ε(−E(q−1) 1 + E q2 ). By expanding the determinant on the first column, we see that the characteristic polynomial of this matrix is equal to −2ελ − λ q . In particular, for a nonzero ε, the alternate matrix K (q) + εX is diagonalizable because it has pairwise distinct eigenvalues and approaches K (q) as ε → 0. Similar, given any integer p, the matrix L(pp) := √ √ 1 (Id − −1V2p ) (J p ⊕ (−J p ))t (Id2p + −1V2p ) 2 2p (11) is alternate. If p is even, then the matrix X = (L(pp) )p−1 is alternate as well. For any scalar λ0 it is easy to see that the characteristic polynomial of K (pp) + εX is the same as the characteristic polynomial of (λ0 Idp +J p + λ0 t p−1 t p−1 ε(J p ) ) ⊕ (−λ0 Idp −J p − ε(J p ) ), which is equal to ((λ0 − λ)p − ε) · ((λ0 + λ)p − ε). Since the eigenvalues of the first block are equal to the zeros of the polynomial (λ0 − λ)p − ε, which are pairwise distinct if ε ≠ 0, we see that the first block is diagonalizable for every ε > 0. The same is true for the second block. Hence, K (pp) + εX λ0 is diagonalizable alternate matrix whenever ε > 0. If p is odd, then the matrix X = (L(pp) )p−2 is alternate because it is an odd power of an alternate matrix (︀ )︀ (pp) L . As above, we see that the characteristic polynomial of K (pp) + εX is equal to (λ0 − λ) (λ0 − λ)p−1 − 2ε · λ 0 (︀ )︀ (−1)(λ0 + λ) (λ0 + λ)p−1 − 2ε (decompose the determinant of each block in the matrix (λ0 Idp +J p + ε(J pt )p−2 ) ⊕ (−λ0 Idp −J p − ε(J pt )p−2 ), which is similar to K (pp) + εX, by the first column). Hence, we deduce that the matrix λ0 K (pp) + εX is diagonalizable and it approaches K (pp) as ε → 0. Consequently, the set of all diagonalizable λ0 λ0 alternate matrices is dense in Altn . For a nonzero vector x ∈ Cn we denote L x := {x ∧ v : v ∈ Cn }. It is easy to see that L x = L y precisely when x, y are linearly dependent. Namely, otherwise we could assume x = e1 and y = e2 . But then e1 ∧ e3 ∈ L e1 is not in L e2 . It is also easy to show that L x ∩ L y = C(x ∧ y) whenever L x ≠ L y . In the proof of the main theorem we will also need the following lemmas. Lemma 2.8. If A, B ∈ L x are linearly independent, then dim(A′ ∩ B′ ) ≥ (n−3)(n−4) . 2 Proof. Write A = x ∧ y and B = x ∧ z. Since A, B are linearly independent, so are the vectors x, y, z. Thus, there exists invertible P such that Px = e1 , Py = e2 , and Pz = e3 . Now, pick any alternate X ∈ A′ . Then, (Xx)y t − (Xy)x t = XA = AX = x(X t y)t − y(X t x)t = −x(Xy)t + y(Xx)t . If we multiply this equation on the right by P and on the left by P t , we get (PXx)e2t − (PXy)e1t = −e1 (PXy)t + e2 (PXx)t . Comparing both sides, we easily obtain PXx = αe1 + βe2 and PXy = 𝛾 e1 − αe2 for some scalars α, β, 𝛾 . Therefore, Xx = αx + βy, Xy = 𝛾 x − αy. Xx = α′ x + β′ z, Xz = 𝛾 ′ x − α′ z Likewise we derive for any alternate X ∈ B′ . Since x, y, z are linearly independent, X ∈ A′ ∩ B′ implies that Xx = αx, Xy = 𝛾 x − αy, Xz = 𝛾 ′ x − αz. Inversely, given X with the above property, X clearly belongs to A′ ∩ B′ , provided it is alternate. To check the alternateness, we let Q = P−1 . Then x = Qe1 . We can claim similar for y and z. Thus, the above equations reduce into Q t XQe1 = αQ t Qe1 , Q t XQe2 = 𝛾 Q t Qe1 − αQ t Qe2 , 90 | Ajda Fošner and Bojan Kuzma Q t XQe3 = 𝛾 ′ Q t Qe1 − αQ t Qe3 . (︀ )︀ Note that X is alternate precisely when Q t XQ is. So, writing the symmetric matrix Q t Q as Q t Q = s ij , the above three conditions give ⎛ ⎞ αs11 𝛾 s11 − αs12 𝛾 ′ s11 − αs13 *1 ⎜ ⎟ ⎜αs12 𝛾 s12 − αs22 𝛾 ′ s12 − αs23 *2 ⎟ ⎟ Q t XQ = ⎜ ′ ⎜αs13 𝛾 s13 − αs23 𝛾 s13 − αs33 *3 ⎟ , ⎝ ⎠ .. .. .. . . . F where *i represents the remainder of the i-th row. Clearly, this is an alternate matrix precisely when αs11 = 0 = 𝛾 s12 − αs22 = 𝛾 ′ s13 − αs33 and 𝛾 s11 − αs12 = −αs12 , etc. So, α = 0 = 𝛾 = 𝛾 ′ , unless the upper-left 3 × 3 block of Q t Q is zero. However, we are free to choose the alternate (n − 3) × (n − 3) lower-right block, represented by linearly independent alternate matrices in this block, which gives our lower F. There are precisely (n−3)(n−4) 2 estimate. Remark 2.9. The upper bound for dim A′ ∩ B′ equals 3 + (n−3)(n−4) . This is possible precisely when the upper2 left 3 × 3 block of Q t Q is zero. Then, X ∈ A′ ∩ B′ is equivalent to ⎛ ⎞ 0 0 0 −αs14 ... ⎜ 0 0 0 −(𝛾 s14 − αs24 ) . . .⎟ ⎜ ⎟ ⎜ ⎟ ′ t 0 0 0 −( 𝛾 s − αs ) . . . ⎜ ⎟. 14 34 Q XQ = ⎜ ⎟ ′ ⎜αs14 𝛾 s14 − αs24 𝛾 s14 − αs34 ⎟ ⎝ ⎠ F .. .. .. . . . This can happen, for example, if A = (e1 + ie2 ) ∧ (e3 + ie4 ) and B = (e1 + ie2 ) ∧ (e5 + ie6 ) (i.e., A2 = 0 = B2 = AB = BA). Lemma 2.10. Let Ω ⊂ Altn be a subset of alternate matrices such that rk A, rk B, rk(A + B) ≤ 2 for every A, B ∈ Ω. Then either Ω ⊆ L ^x for some nonzero vector ^x or there exists invertible P ∈ M n such that Ω ⊆ P(Alt3 ⊕0n−3 )P t lies in at most three dimensional subspace. Proof. Let us choose two linearly independent matrices A1 , A2 ∈ Ω and write them as A i = x i ∧ y i , i = 1, 2. Since rk(A1 + A2 ) ≤ 2 we may assume that x1 = x2 := ^x. Namely, at least one among the triples x1 , y1 , x2 or x1 , y1 , y2 must be linearly dependent. Without loss of generality, assume the first one is. Write x2 = αx1 + βy1 and note that x1 ∧ y1 = (αx1 + βy1 ) ∧ ( 1α )y1 = x2 ∧ ( 1α )y1 , if α ≠ 0. However, if α = 0, then we may as well assume that β = 1 and we have x1 ∧ y1 = y1 ∧ (−x1 ) = x2 ∧ (−x1 ). In any case, A1 , A2 ∈ L x2 . Moreover, with the help of some invertible matrix P we may further assume that ^x = e1 , y1 = e2 , and y2 = e3 , that is, A1 = e1 ∧ e2 ∈ L e1 and A2 = e1 ∧ e3 ∈ L e1 . Suppose that there exists a matrix A3 = x3 ∧ y3 ∈ Ω \ L e1 . Due to rk(A1 + A3 ) ≤ 2, either x3 or y3 is a linear combination of e1 , e2 . Without loss of generality we may assume the former and write x3 = αe1 + βe2 . Clearly, β ≠ 0, otherwise A3 ∈ L e1 . Similarly, rk(A3 + A2 ) ≤ 2 implies that at least one among x3 and y3 is a linear combination of e1 , e3 . But for x3 this is already impossible, so y3 = 𝛾 e1 + δe3 . Consequently, A3 ∈ Alt3 ⊕0n−3 whenever A3 ∈ Ω \ L e1 . It remains to show that Ω ⊆ Alt3 ⊕0n−3 . In fact, pick any matrix A4 ∈ Ω and assume erroneously that it is linearly independent of A1 , A2 , A3 . Then, as before, due to rk(A2 + A4 ), rk(A1 + A4 ) ≤ 2 we must have A4 ∈ L e1 . So, A4 = e1 ∧ y4 , where y4 is linearly independent of e1 , e2 , e3 . But then, A3 + A4 = ( αβ e1 + e2 ) ∧ (β𝛾 e1 + βδe3 ) + e1 ∧ y4 is of rank four, a contradiction. For nonzero vectors x, y ∈ Cn , we denote [x] = {αx : α ∈ C} and [x] + [y] = {αx + βy : α, β ∈ C}. Lemma 2.11. If x, y are linearly independent vectors, then [z] ⊆ [x] + [y] precisely when the intersection (L x ∩ L z ) ∩ (L y ∩ L z ) is nonzero. Preserving zeros of Lie product on alternate matrices | 91 Proof. Suppose [z] ⊆ [x] + [y]. Then z = αx + βy for some scalars α, β. If β = 0 then L z = L x and (L x ∩ L z ) ∩ (L y ∩ L z ) = L x ∩ L y = C(x ∧ y) ≠ 0. The same is true when α = 0. Now, if α and β are both nonzero, then (L x ∩ L z ) = C(x ∧ z) = Cx ∧ (αx + βy) = C(x ∧ y) and (L y ∩ L z ) = Cy ∧ (αx + βy) = C(y ∧ x) = C(x ∧ y) ≠ 0. Now, assume that (L x ∩ L z ) ∩ (L y ∩ L z ) ≠ 0. Then λ(x ∧ z) = µ(y ∧ z) for some nonzero scalars λ, µ. If x, y, z were linearly independent, there would exist an invertible P such that Px = e1 , Py = e2 , and Pz = e3 . Multiplying the above identity with P on the left and P t on the right side yields that λ(e1 ∧ e3 ) = µ(e2 ∧ e3 ), a contradiction. 3 Proof of the main theorem Let ϕ : Altn → Altn be an injective continuous map that preserves zeros of Lie product. Then for every A ∈ Altn , we have (︀ )︀′ ϕ(A′ ) ⊆ ϕ(A) . In particular, ϕ(Altn ) = ϕ(0′ ) ⊆ ϕ(0)′ ⊆ Altn . Since ϕ is injective and continuous, it follows by the invariance of domain theorem that ϕ(0)′ cannot be contained inside a proper linear subspace of Altn . This yields that ϕ(0)′ = Altn and, consequently, ϕ(0) = 0. The remainder of the proof is divided into several steps. Step 1. The set Ξ2 ⊆ Altn of rank two alternate matrices is invariant under ϕ. Moreover, ϕ(CA) ⊆ Cϕ(A) for every A ∈ Ξ2 . Let 0 ≠ A ∈ Altn . By Lemma 2.5 dim A′ ≥ (n−2)(n−3) + 1 if and only if rk A = 2. This yields that every 2 A ∈ Altn of rank two is mapped by ϕ into an alternate matrix of rank two. Namely otherwise, ϕ would map A′ , which is a linear subspace of dimension at least (n−2)(n−3) + 1, continuously and injectively into ϕ(A)′ , 2 (n−2)(n−3) whose dimension is strictly smaller than + 1. But this is impossible by the invariance of domain 2 theorem. Therefore, for every A ∈ Ξ2 of rank two there exists B ∈ Ξ2 such that ϕ(A) = B. It remains to prove ϕ(CA) ⊆ CB. If there was a nonzero complex number λ such that ϕ(λA) ∉ CB, then ϕ would map A′ = (λA)′ injectively and continuously into B′ ∩ ϕ(λA)′ . Now, any X ∈ B′ ∩ ϕ(λA)′ also commutes with B + ϕ(λA). Hence, if rk(B + ϕ(λA)) ≥ 4, then, by Lemma 2.5, dim B′ ∩ ϕ(λA)′ ≤ dim(B + ϕ(λA))′ (n−2)(n−3) + 1 ≤ dim A′ . However, if rk(B + ϕ(λA)) = 2, then Remark 2.9 implies dim(B′ ∩ ϕ(λA)′ ) dim A′ . 2 Each of these two cases contradicts Brouwer’s invariance of domain theorem. Step 2. Let x be a nonzero vector. Then, rank(ϕ(x ∧ y1 ) + ϕ(x ∧ y2 )) ≤ 2 for each x ∧ y1 , x ∧ y2 ∈ L x . Indeed, write ϕ(x ∧ y1 ) = x̃1 ∧ ỹ1 and ϕ(x ∧ y2 ) = x̃2 ∧ ỹ2 . Note that (︀ )︀ ϕ (x ∧ y1 )′ ∩ (x ∧ y2 )′ ⊆ (x̃1 ∧ ỹ1 )′ ∩ (x̃2 ∧ ỹ2 )′ . Now, suppose erroneously that rk(x̃1 ∧ ỹ1 + x̃2 ∧ ỹ2 ) > 2. Then, by Lemma 2.8, the space (x ∧ y1 )′ ∩ (x ∧ y2 )′ would be mapped injectively and continuously into a subspace (x̃1 ∧ ỹ1 )′ ∩ of dimension at least (n−3)(n−4) 2 (x̃2 ∧ ỹ2 )′ ⊆ (x̃1 ∧ ỹ1 + x̃2 ∧ ỹ2 )′ which by Lemma 2.6 has a strictly smaller dimension, a contradiction to the invariance of domain theorem. Therefore, rk(x̃1 ∧ ỹ1 + x̃2 ∧ ỹ2 ) ≤ 2. Step 3. For each L x there exists some L x̃ such that ϕ(L x ) ⊆ L x̃ . In fact, if this would not be the case for some x, then the previous step together with Lemma 2.10 would imply that ϕ maps the n −1 dimensional space L x continuously and injectively into at most three dimensional subspace, a contradiction. Step 4. Let x be a nonzero vector such that ϕ(L x ) ⊆ L x̃ ∩ L x̃′ for some vectors x̃ and x̃′ . Then x̃ and x̃′ must be linearly dependent. To see this, note that ϕ(L x ) ⊆ L x̃ implies dim(Lin ϕ(L x )) = dim L x̃ = n − 1 by the invariance of domain theorem (here, Lin ϕ(L x ) denotes the linear span of the set ϕ(L x )). But then ϕ(L x ) ⊆ L x̃ ∩ L x̃′ implies that L x̃ = Lin ϕ(L x ) = L x̃′ which is possible precisely when x̃, x̃′ are linearly dependent. 92 | Ajda Fošner and Bojan Kuzma We can hence introduce a well-defined map φ on a projective space PCn by letting φ : [x] ↦→ [x̃], where x̃ satisfies ϕ(L x ) ⊆ L x̃ . To prove that this is a projectivity, assume that [z] ⊆ [x] + [y]. We can clearly suppose that x, y are linearly independent, otherwise [x] = [y] = [z] and, hence, there is nothing to prove. By Lemma 2.11, we must have (L x ∩ L z ) ∩ (L y ∩ L z ) ≠ {0}. Since ϕ is injective and ϕ(0) = 0, this further implies (︀ )︀ {0} ≠ ϕ (L x ∩ L z ) ∩ (L y ∩ L z ) ⊆ (ϕ(L x ) ∩ ϕ(L z )) ∩ (ϕ(L y ) ∩ ϕ(L z )). Thus, by Lemma 2.11, φ([z]) ⊆ φ([x]) + φ([y]). Step 5. The map φ is injective. Let x1 , x2 be two linearly independent vectors and suppose that φ([x1 ]) = φ([x2 ]) = [x̃] for some x̃. Then, by the definition of φ and the invariance of domain theorem, ϕ(L x1 ) and ϕ(L x2 ) are open subsets of L x̃ containing ϕ(0) = 0. Let x1 ∧ y ∈ L x1 ∖L x2 . As limλ→0 ϕ(λx1 ∧ y) = ϕ(0) = 0, we can find a small enough nonzero complex number λ such that ϕ(λx1 ∧ y) ∈ ϕ(L x2 ), contradicting the injectivity of ϕ. Thus, φ is injective. Step 6. The span of lines in Im φ equals PCn . Hence, Im φ is not contained in a projective line. To prove this, denote [ẽ1 ] := φ([e1 ]). Note that L e1 and L ẽ1 are both n−1 dimensional linear subspaces and ϕ maps continuously and injectively the first one into the second one. So, by the invariance of domains, ϕ(L e1 ) is an open subset inside L ẽ1 that contains a zero matrix. Therefore, given any vector z, linearly independent of ẽ1 , there exists some y such that ϕ(e1 ∧ y) = λẽ1 ∧ z for small enough nonzero scalar λ. Clearly, e1 ∧ y ≠ 0, because ϕ(0) = 0. Therefore, e1 , y are linearly independent. Now, e1 ∧ y ∈ L e1 ∩ L y . Thus, λẽ1 ∧ z = ϕ(e1 ∧ y) ∈ ϕ(L e1 ∩ L y ) ⊆ ϕ(L e1 ) ∩ ϕ(L y ) ⊆ L ẽ1 ∩ L ỹ . But this is possible only when ỹ = αẽ1 + βz for some scalars α, β. Moreover, β ≠ 0 because linear independence of e1 , y and injectivity of φ forces L ỹ ≠ L ẽ1 . We may now assume that β = 1. So, given any z, linearly independent from ẽ1 , there is some y ≠ 0 and a scalar α such that φ([y]) = [αẽ1 + z]. But then the image of φ spans PCn : it clearly contains [ẽ1 ] and it also contains [α i ẽ1 + e i ] for every i (except for one possible exception, when ẽ1 , e i are linearly dependent). But then this n lines (or n + 1 lines, if ẽ1 , e i are always independent) span the whole projective space. ^ and a field homomorphism σ : C → C such that for each Step 7. There exists an invertible matrix Q ^ σ ∧ y σ )Q ^ t for appropriate scalar λ x∧y ∈ C. nonzero alternate x ∧ y we have ϕ(x ∧ y) = λ x∧y Q(x By a nonsurjective version of the fundamental theorem of projective geometry (see e.g. [7, Theorem 3.1]) there exists a field homomorphism σ : C → C and a σ-linear map T : Cn → Cn such that φ([x]) = [Tx] for every line [x] ∈ PCn . Consequently, ϕ(L x ) ⊆ L Tx . Thus, from x ∧ y ∈ L x ∩ L y we derive that ϕ(x ∧ y) ∈ ϕ(L x ∩ L y ) = ϕ(L x ) ∩ ϕ(L y ) ⊆ C (Tx ∧ Ty). It follows that ϕ(x ∧ y) = 1β (Tx) ∧ Ty, where a scalar β depends ^ σ for some matrix Q ^ ∈ M n . Since Lin(Im φ) = Cn , the on the operator x ∧ y. It is also easy to see that Tx = Qx σ σ σ σ t 1 1 ^ is invertible. Moreover, ϕ(x ∧ y) = Qx ^ ∧ Qy ^ = Q(x ^ ^ . matrix Q ∧ y )Q β β Step 8. Homomorphism σ is either a complex conjugation or the identity. ^ 1 ∧ (σ(α)e2 + e3 ))Q ^ t for appropriate nonzero scalars µ α . To see this, consider ϕ(e1 ∧ (αe2 + e3 )) = µ α Q(e The continuity of ϕ forces the continuity of the functions µ α σ(α) and µ α . This gives that σ is a continuous homomorphism of C. Thus, it is either a complex conjugation or the identity. Replacing ϕ by the map X ↦→ ϕ(X), X ∈ Altn , if necessary, we may assume in the sequel that σ is the identity. ^ tQ ^ is a scalar multiple of the identity matrix. Step 9. The matrix Q To see this, choose any orthogonal matrix W. Given any distinct i, j, k, l, observe that W(e i ∧ e j )W t commutes with W(e k ∧ e l )W t . Hence, 0 = ϕ(W(e i ∧ e j )W t )ϕ(W(e k ∧ e l )W t ) − ϕ(W(e k ∧ e l )W t )ϕ(W(e i ∧ e j )W t ) (︀ t ^t ^ t ^t t ^t ^ t ^ t )︀ ^ ^ = µ QW(e i ∧ e j )W Q QW(e k ∧ e l )W Q − QW(e k ∧ e l )W Q QW(e i ∧ e j )W Q , where µ = λ W(e i ∧e j )W t λ W(e k ∧e l )W t ≠ 0. In particular, t ^t ^ ^ t QW(e ^ (e i ∧ e j )W t Q k ∧ e l ) = (e k ∧ e l )W Q QW(e i ∧ e j ). Preserving zeros of Lie product on alternate matrices | 93 ^ t QW ^ corresponding to Comparing the (i, k), (i, l), (j, k), and (j, l) positions, we derive that the entries of W t Q t ^t ^ those positions vanish. Due to arbitrariness of i, j, k, l, we see that W Q QW is diagonal for (︃ every (︃ orthogonal )︃)︃ 1 1 t 1 ^ Q ^ is diagonal. Furthermore, choosing W = √ W. Choosing W = Idn , we conclude that Q ⊕ 2 −1 1 ^t ^ Idn−2 , we (︃ see that (︃ the entries )︃)︃ at the positions (1, 1) and (2, 2) of Q Q are equal. Choosing in this succession 1 1 ^ tQ ^ are equal. ⊕ Idn−2−i for i = 1, . . . , n − 2, we derive that all diagonal entries of Q W = 0i ⊕ √1 2 −1 1 ^ tQ ^ = λ Idn for some nonzero scalar λ, as desired. Therefore, Q := √1 Q ^ is an orthogonal This implies that Q λ ^ Q ^ t = µ X QXQ t for every rank two alternate matrix X. matrix with ϕ(X) = µ X QX λ Step 10. According to above observations, we may replace ϕ by the map X ↦→ Q t ϕ(X)Q. The new map is still continuous, preserves zeros of Lie product, and fixes every alternate matrix of minimal rank modulo scalar multiples. Step 11. If D is a diagonalizable alternate matrix then there exists an odd polynomial p D (λ) (i.e., p D (λ) = −p D (−λ)) such that ϕ(D) = p D (D) (i.e., ϕ acts locally polynomially on diagonalizable alternate matrices). To see this, let D be a diagonalizable alternate matrix. It follows from (5) that there exists an orthogonal matrix W such that {︃ W(α1 J) ⊕ (α2 J) ⊕ · · · ⊕ (α k J)W t ; n even D= (︀ )︀ t W (α1 J) ⊕ (α2 J) ⊕ · · · ⊕ (α k J) ⊕ 0 W ; n odd (︃ )︃ 0 1 for some scalars α i , where J is a 2 × 2 matrix of the form J = . Clearly, D commutes with a minimal −1 0 rank matrix A i = W(02i ⊕ J ⊕ 0n−2i−2 )W t for every i = 1, . . . , ⌊n/2⌋. Consequently, also ϕ(D) commutes with ϕ(A i ) = 𝛾i A i ≠ 0, which forces ϕ(D) = W((β1 J) ⊕ (β2 J) ⊕ · · · ⊕ (β k J))W t , respectively, (︃ )︃ ϕ(D) = W((β1 J) ⊕ J J (β2 J) ⊕ · · · ⊕ (β k J) ⊕ 0)W t for some scalars β i . Take a block alternate matrix X = ⊕ 0 and note that this J J matrix is of minimal rank. If, say α1 = α2 , then X commutes with D. Hence, also its image, ϕ(X) = 𝛾X X ≠ 0, commutes with ϕ(D) which is possible only when β1 = β2 . Likewise we show that α i = α j forces β i = β j , i, j = 1, . . . , k. It is known that there exists a polynomial p D (λ) with p D (α i ) = β i for i = 1, . . . , n. Hence, ϕ(D) = p D (D). In other words, ϕ acts locally as polynomial on diagonalizable alternate matrices. Moreover, we will show that these polynomials can be chosen to be odd (recall that a polynomial f (λ) is odd if f (−λ) = −f (λ), or ∑︀ equivalently, if f (λ) = i a2i+1 λ2i+1 for some scalars a2i+1 ∈ C). In fact, suppose that A and p(A) are both alternate for some polynomial p(λ). Decompose p(λ) = p+ (λ)+ p− (λ) into a sum of even and odd parts. Clearly, A2i are symmetric matrices for each nonnegative integer i. So, p+ (A) is a linear combination of symmetric matrices and, hence, it is symmetric. Likewise one can show that p− (A) is alternate. So, the symmetric matrix p+ (A) = p(A)−p− (A) is a difference of two alternate matrices p(A) and p− (A), whence p+ (A) = 0 = p(A)−p− (A). Thus, we can assume that p(λ) = p− (λ) is an odd polynomial, i.e., p(−λ) = −p(λ). Step 12. ϕ(X) ∈ X ′ for every alternate X. In fact, by Step 11 we know that ϕ is a locally polynomial map on the set of all diagonalizable alternate matrices. In particular, every diagonalizable alternate matrix A commutes with ϕ(A). According to Lemma 2.7, the set of all diagonalizable alternate matrices is dense in Altn . Thus, by the continuity of ϕ, every X ∈ Altn commutes with ϕ(X). In particular ϕ(X) ∈ X ′ , and, if alternate X is nonderogatory, then ϕ(X) ∈ X ′ = Poly(X) ∩ Altn [12, Theorem 3.2.4.2.]. Step 13. Action of ϕ on a general alternate A. As explained in (5), there exists an orthogonal Q such that A = Qt n1 (︁⨁︁ i=1 i pi ) K (p ⊕ λi n2 ⨁︁ )︁ (︀ )︀ K (q j ) Q = Q t A1 ⊕ A2 Q. j=1 To ease the proof, we will continue in substeps. Also, we will temporarily modify ϕ into a map ϕ Q (X) := Qϕ(Q t XQ)Q t , (12) 94 | Ajda Fošner and Bojan Kuzma which clearly shares the same properties as ϕ. In particular, all steps until the present one hold for ϕ Q as well. The benefit of introducing ϕ Q is that ϕ(A) is a polynomial in A if and only if ϕ Q (A1 ⊕ A2 ) is a polynomial in block-diagonal A1 ⊕ A2 . Substep 13.a. There exist odd polynomials f1 (λ), . . . , f n1 (λ) and odd polynomials h1 (λ), . . . , h n2 (λ) such that n1 (︁⨁︁ ϕ Q (A1 ⊕ A2 ) = i pi ) f i (K (p )⊕ λi i=1 n2 ⨁︁ )︁ h j (K (q j ) ) (13) j=1 To see this, choose continuous functions λ1 (ε), . . . , λ n1 (ε), nonzero and in absolute value pairwise distinct at each fixed ε > 0, such that limε↘0 λ i (ε) = λ i for i = 1, . . . , n1 . Define A1 (ε) := n1 ⨁︁ i pi ) , K (p λ (ε) (︀ )︀ A(ε) := A1 (ε) ⊕ A2 . i i=1 By its definition (2) we see that, at each fixed ε > 0, every diagonal block of A1 (ε) is a nonderogatory matrix with eigenvalues {−λ i (ε), λ i (ε) : i = 1, . . . , n1 }. Hence, due to 0 ≠ |λ i (ε)| ≠ |λ j (ε)| for i ≠ j, we derive that A(ε) is nonderogatory at each fixed ε > 0 and so by Step 12 and the fact that (X1 ⊕ X2 )′ = X1′ ⊕ X2′ whenever block matrices X1 , X2 have no eigenvalue in common (see Lemma 2.1 or [8, Theorem 2 and Remark on p. 127]), we have: n1 )︁ (︀ )︀ (︁⨁︁ ′ i pi ) ′ ϕ Q (A(ε)) ∈ A(ε)′ = A1 (ε)′ ⊕ A′2 = (K (p ) ⊕ A (14) 2 . λ (ε) i i=1 i pi ) ′ i pi ) ′ i pi ) Since λ i (ε) ≠ 0 for ε > 0, the equation (9) gives (K (p ) = (K (p ) . Also, the matrices K (p are nonderogaλ=1 λ=1 λ i (ε) tory, and therefore i pi ) ′ i pi ) i pi ) (K (p ) = Poly(K (p ) ∩ Altn = Poly(K (p ) ∩ Altn . λ=1 λ=1 λi Combined with (14) we derive ϕ Q (A(ε)) ∈ n1 (︁⨁︁ )︁ ′ i pi ) Poly(K (p ) ⊕ A ; 2 λi ε > 0. i=1 Hence, by continuity and the fact that the set on the right side of the above equation is closed, ϕ Q (A1 ⊕ A2 ) = lim ϕ Q (A(ε)) ∈ n1 (︁⨁︁ ε↘0 )︁ i pi ) Poly(K (p ) ⊕ A′2 . λi (15) i=1 Since ϕ Q (A1 ⊕A2 ) is alternate, there are hence odd polynomials (see last paragraph of Step 11) f1 (λ), . . . , f n1 (λ) such that n1 (︁⨁︁ )︁ i pi ) ϕ Q (A1 ⊕ A2 ) ∈ f i (K (p ) ⊕ A′2 . (16) λi i=1 ⨁︀ 2 (q j ) Likewise, it follows from (4) that the block constituents of A2 = nj=1 K , i.e., matrices K (q j ) , are non(︀ )︀ ⨁︀ ′ derogatory. Hence, if X ∈ A′2 = K (q j ) is partition into blocks conformally to A2 then its j-th diagonal block satisfies X jj ∈ (K (q j ) )′ = Poly(K (q j ) ) ∩ Altq j . (17) Moreover, recall from the proof of Lemma 2.7 that the matrices A ε := n1 (︁⨁︁ i=1 where i pi ) X (p ⊕ ε n2 ⨁︁ (q j ) Xε )︁ , ⎧ ⎪ X (1) ⎪ ε := 0, ⎪ ⎪ √ ⎪ ⎨X (q j ) := K (q j ) + ε (Id −√−1V)(︀E qj ε (q j −1) 1 − E q j 2 )(Idq j + −1V), 2 i pi ) i pi ) ⎪ X (p := K (p + ε(L(pp) )p−1 , for p even ⎪ ε ⎪ λ i +ε ⎪ ⎪ ⎩X (p i p i ) := K (p i p i ) + ε(L(pp) )p−2 , for p odd ε λ i +ε (18) j=1 qj ≥ 3 , Preserving zeros of Lie product on alternate matrices | 95 (L(pp) is defined by equation (11)) are alternate, diagonalizable, and converge to A1 ⊕ A2 as ε → 0. By Step 11, ∑︀ ϕ Q (A ε ) are odd polynomials (i.e., p(λ) = a2i+1 λ2i+1 ) in A ε and, by the continuity of ϕ Q , they converge to ϕ Q (A1 ⊕ A2 ) as ε → 0. Note that a polynomial in a block-diagonal matrix A ε is itself a block-diagonal matrix. Therefore, ϕ Q (A1 ⊕ A2 ) is also a block-diagonal alternate matrix. It then follows from (16)–(17) that there exist odd polynomials h1 (λ), . . . , h n2 (λ) which satisfied the desired equation (13). Substep 13.b. To finish the proof that ϕ Q (A1 ⊕ A2 ) is a polynomial in A1 ⊕ A2 we have to show that whenever in equation (13) we have (a) λ i = ±λ i′ or (b) λ i = 0 then there exists a common polynomial f (λ) such that, under (a), (p p ) (p p ) i pi ) i pi ) f i (K (p ) ⊕ f i′ (K λ ′ i′ i′ ) = f (K (p ) ⊕ f (K λ ′ i′ i′ ) λi λi i i respectively, under (b) i pi ) i pi ) f i (K (p ) ⊕ h j (K (q j ) ) = f (K (p ) ⊕ f (K (q j ) ). λi λi Also, we have to show that there exists a common polynomial h(λ) such that this is shown the proof ends because it is well-known that ⨁︀ h j (K (q j ) ) = ⨁︀ h(K (q j ) ). Once Poly(X1 ⊕ X2 ) = Poly(X1 ) ⊕ Poly(X2 ) whenever matrices X1 , X2 have no eigenvalue in common. We thus have to observe three cases. (p p i′ ) i pi ) Substep 13.c. Assume first that K (p and K λ ′ i′ λi i (p i p i ) K−λ i is orthogonally similar to i pi ) K (p . λi whenever share the same eigenvalue, that is, λ i = ±λ i′ . By (3) the block Hence we may assume that in (12), λ i = ±λ i′ then actually λ i = λ i′ . For the sake of simplicity, we will also assume that p i ≤ p i′ and that i = 1 and i′ = 2, that is, the first two 1 p1 ) diagonal blocks in (13) share the same eigenvalue. Then, by the definition of K (p (see (2)), we see that λ1 A1 ⊕ A2 commutes with alternate block matrices (︃ )︃ 0 Xk Tk = ⊕ 0n−2p1 −2p2 , k = 1, 2, −X kt 0 where √ (︃ [0|I] X1 = Id2p1 − −1V2p1 · 0p1 ×p2 (︃ √ (︀ )︀ [0|0] X2 = 12 Id2p1 − −1V2p1 · 0p1 ×p2 1 2 (︀ )︀ 0p1 ×p2 [0|0] )︃ √ )︀ (︀ · Id2p2 + −1V2p2 , 0p1 ×p2 [0|I] )︃ √ (︀ )︀ · Id2p2 + −1V2p2 , and I denotes the identity matrix of the size p1 × p1 . Assume that p1 = p(︃ 2 . In this case, let )︃ X = X1 + X2 = √ √ (︀ )︀ (︀ )︀ 0 Id2p1 1 ⊕ 0n−4p1 . Then 2 Id2p1 − −1V 2p1 Id2p1 Id2p1 + −1V 2p1 = Id2p1 and T = T 1 + T 2 = − Id2p1 0 T 3 = −T. In particular, T is diagonalizable. Thus, ϕ Q maps T into an odd polynomial in T. But due to T 3 = −T it is fixed by ϕ Q , modulo multiplication by a nonzero scalar. We infer that ϕ Q (A1 ⊕ A2 ) must also commute with T. Hence, the first two diagonal blocks of ϕ Q (A1 ⊕ A2 ) are intertwined with a matrix X. Since X = Id2p1 1 p1 ) 2 p2 ) we get that, in identity (13), f1 (K (p ) = f2 (K (p ). Since p2 = p1 and λ1 = λ2 we can hence take f (λ) = f1 (λ) λ1 λ2 in this case. On the other hand, if p1 p2 , then we have T13 = −T1 and T23 = −T2 so, by the above, T1 and T2 are both fixed by ϕ Q , modulo scalars. Since ϕ Q preserves commutativity, ϕ Q (A1 ⊕ A2 ) commutes with every linear combination of ϕ Q (T1 ) ∈ CT1 and ϕ Q (T2 ) ∈ CT2 . In particular, ϕ Q (A1 ⊕ A2 ) also commutes with 96 | Ajda Fošner and Bojan Kuzma (︀ T = (T1 + T2 ). By taking the first two blocks from (13) and using definition (2) we see that f1 (λ1 Idp1 +J p1 ) ⊕ )︀ (︀ )︀ (−λ1 Idp1 −J p1 ) ⊕ f2 (λ2 Idp2 +J p2 ) ⊕ (−λ2 Idp2 −J p2 ) commutes with (︃ )︃⎞ ⎛ [0|I] 0 0 ⎟ ⎜ 0 [0|I] ⎟ ⎜ ⎟ (︃ )︃ T1 + T2 = ⎜ t ⎜ ⎟ [0|I] 0 ⎝ ⎠ − 0 0 [0|I] (︃ )︃ (︀ )︀ [0|I] 0 Again, intertwines blocks f1 (λ1 Idp1 +J p1 ) ⊕ (−λ1 Idp1 −J p1 ) and 0 [0|I] (︀ )︀ f2 (λ2 Idp2 +J p2 ) ⊕ (−λ2 Idp2 −J p2 ) . It is straightforward that f1 (λ1 Idp1 +J p1 ) [0|I] = [0|I] f2 (λ2 Idp2 +J p2 ); λ2 = λ1 . Since polynomials f1 (λ), f2 (λ) are odd and since we can assume that the degree of f i (λ) is smaller than the degree of minimal polynomial of corresponding block, i.e., smaller than p i , we can write ⎛ ⎞ f1 (λ1 ) 0 α1 0 α2 . . . . . . . ⎜ ⎟ f1 (λ1 ) 0 α1 0 α2 . . .⎟ f1 (λ1 Idp1 +J p1 ) = ⎜ ⎝ ⎠ .. .. .. . . . for suitable scalar α i and likewise for f2 (λ1 Idp2 +J p2 ). It then easily follows that f1 (λ) and f2 (λ) coincide in the first min{p1 , p2 } = p1 monomials. Hence, the present substep is done by taking f (λ) = f2 (λ). Substep 13.d. From Substep 13.c it follows that ϕ fixes every square-zero alternate matrix, modulo scalars, (︀⨁︀ (22) )︀ because such matrices take the form O K λ=0 ⊕ 0 O t for some orthogonal O. Substep 13.e. Next, assume that A1 ⊕ A2 contains two blocks K (q j ) and K (q j′ ) . As above, we may assume that q j ≤ q j′ and that j = 1 and j′ = 2. Moreover, for simplicity we will temporarily permute the diagonal blocks so that, in the present substep only, A1 ⊕ A2 = K (q1 ) ⊕ K (q2 ) ⊕ . . . . If q1 = q2 , then A1 ⊕ A2 commutes with an alternate block matrix )︃ (︃ 0 Idq1 ⊕ 0n−q1 −q2 T= 0 − Idtq1 which satisfies T 3 = −T. Hence, T is fixed by ϕ Q , modulo multiplication by a nonzero scalar. As in the first case, it is easy to see that the first two blocks B1 = h1 (K (q1 ) ) and B2 = h2 (K (q2 ) ) of ϕ Q (A1 ⊕ A2 ) coincide. Hence, B1 ⊕ B2 = h(K (q1 ) ) ⊕ h(K (q2 ) ) where h(λ) = h1 (λ). (︃ )︃ 0 D1 (q1 ) d On the other hand, assume q2 = q1 + 2d for some integer d ≥ 1. Then, by Lemma 1.2, (J ) = , 0 0 where D1 is a diagonal (q1 − d) × (q1 − d) matrix with ±1 on the diagonal. Clearly, if d ≥ q1 , then (J (q1 ) )d = 0. Similarly, (︃ )︃ 0 D2 (q2 ) d (J ) = , (19) 0 0 where D2 is a diagonal (q2 − d) × (q2 − d) matrix with ±1 on the diagonal. Obviously, D2 is a nonzero matrix since q2 > d. Moreover, if we write q1 = 2k + 1, then q2 = q1 + 2d = 2(k + d) + 1, and, by formula in Lemma 1.2, (︀ )︀ D2 = Idk+1 ⊕ diag −1, (−1)2 , . . . , (−1)d−1 ⊕ (−1)d Idk+1 . (20) Now it easily follows that, if we remove from D2 the first d and the last d rows and columns, we obtain D1 . More precisely, D1 is a compression of the matrix D2 onto the subspace generated by {e d+1 , ..., e q1 } and D1 = 0 if d ≥ q1 . Preserving zeros of Lie product on alternate matrices | If d is odd, then we define the matrix (︃(︃ √ (︀ )︀d − −1 J (q1 ) −1 T=S ^ t V q1 −V q2 X ^ X √ (︀ (q ) )︀d −1 J 2 )︃ 97 )︃ ⊕ 0n−q1 −q2 S, √ (︀ )︀ ^ = 0q ×2d |D ^ ,D ^ is the compression of −D2 onto the last q1 basis vectors, and S = (Idq1 + −1V q1 ) ⊕ where X 1 √ ^ = −P(J (q2 ) )d , where P : Cq2 → Cq1 is a projection (Idq2 + −1V q2 ) ⊕ Idn−q1 −q2 . Observe that (19) implies X ^ (q2 ) = P(J (q2 ) )d+1 maps basis vector e i into 0 for onto coordinates (d + 1) through (q2 − d). By Lemma 1.2, XJ q1 i = 1, . . . , (2d + 1) and into −(±)i e i−2d−1 ∈ C for i = (2d + 2), . . . , (2d + q1 ) = q2 , where ⎧ d+1 ⎪ i > q2 − k ⎪ ⎨(−1) ; (±)i = (−1)d+k+1−i ; ⎪ ⎪ ⎩1; q2 − (d + k) < i ≤ q2 − k . otherwise ̂︁ e ^ i = 0 for 1 ≤ i ≤ 2d and Xe ^ i = −(±) On the other hand, Xe i i−2d for i ≥ 2d + 1, where ⎧ d ⎪ i > q2 − k − 1 ⎪ ⎨(−1) ; ̂︁ d+k−i−1 (±)i = (−1) ; q2 − (d + k) − 1 < i ≤ q2 − k − 1 . ⎪ ⎪ ⎩1; otherwise Apply J (q1 ) and use the fact that J (q1 ) e i−2d = e i−2d−1 if i − 2d ≤ q12−1 and J (q1 ) e i−2d = −e i−2d−1 if i − 2d > is then a straightforward calculation that ^ = XJ ^ (q2 ) . J (q1 ) X q1 −1 2 . It (21) Next, it is easy to see that T is alternate. Assume first d < q1 . Then the block structure of STS−1 equals d ⎛ q1 − d 0 ⎜ 0 d ⎜ ⎜ ⎜ (D′′ )f d ⎜ 2 ⎜ q1 − d ⎜ 0 ⎜ ⎝ 0 d d 0 q1 − d d d ^2 − −1D 0 0 0 0 0 √ 0 ^ 2 )f (D 0 0 0 0 0 0 q1 − d ^2 −D 0 √ −1D′2 0 0 0 0 ^2 −1D 0 0 √ d ⎞ 0 −D′′2 ⎟ ⎟ ⎟ 0 ⎟ ⎟ ⊕ 0n−q1 −q2 . ⎟ 0 ⎟ √ ⎟ −1D′′2 ⎠ 0 ∑︀ ^2 ∈ Here, X f = VX t V, V = i E i (n−i+1) , is a flip map, i.e., reflection over anti-diagonal, and D′2 , D′′2 ∈ M d , D ′ ′′ ^ M q1 −d are appropriate diagonal matrices such that D2 = D2 ⊕ D2 ⊕ D2 . It follows from (20) that ^ 2 )f = − D ^ 2. (D We claim that T 2 = 0. In fact, STS−1 maps the basis vectors e q1 +1 , . . . , e q1 +d already into zero. It further maps the vectors e k , k = 1, . . . , d, into ±e q1 +k which are annihilated by another application of STS−1 . Moreover, √ it maps e q1 +d+k into ± −1e q1 +k for k = 1, . . . , d, which are annihilated by the second application of STS−1 . √ Next, STS−1 maps the vectors e d+k , k = 1, . . . , q1 − d into ±( −1e k + e q1 +d+k ), which are further mapped by STS−1 into √ √ √ ±( −1STS−1 e k + STS−1 e q1 +d+k ) = ±(− −1e k+q1 + −1e q1 +k ) = 0. √ √ √ Finally, the vectors e q1 +2d+k , k = 1, . . . , q1 are mapped to STS−1 e q1 +2d+k = ±(e k − −1e q1 +d+k ) = ± −1( −1e k + e q1 +d+k ), which, as we showed above, are annihilated by the second application of STS−1 . This shows T 2 = 0 if d < q1 . By the similar analysis we can show that T 2 = 0 also when d ≥ q1 (the block structure of STS−1 is ^ 2 = 0). Consequently, by Substep 13.d, same as before except that now, D ϕ Q (T) ∈ CT. 98 | Ajda Fošner and Bojan Kuzma (q1 ) ⊕ J (q2 ) commutes with (︃ We next prove )︃ that A1 ⊕ A2 commutes with T. It suffices to show that J ^ 0 X ^ = XJ ^ (q2 ) and J (q2 ) V q2 X ^ t V q1 = V q2 X ^ t V q1 J (q1 ) . The . This is equivalent to the fact that J (q1 ) X t ^ −V q2 X V q1 0 first equality follows from (21) while the second one is equivalent to the first one which is seen by applying the bijective operation Z ↦→ (V q1 ZV q2 )t to the first equality, and in the resulting ^ q2 )t = (V q1 (XJ ^ (q2 ) )V q2 )t (V q1 (J (q1 ) X)V ^ q2 )t = (V q1 J (q1 ) V q1 · V q1 XV ^ q2 )t = (V q1 XV ^ q2 )t · (V q1 J (q1 ) V q1 )t = rewrite the left-hand side into (V q1 J (q1 ) XV t t (q1 ) ^ (V q2 X V q1 ) · (−J ) and likewise rewrite the right-hand side. It follows that ϕ Q (A1 ⊕ A2 ) commutes Poly(T) = CT as well. This is equivalent to the fact (︃ with ϕ Q (T) ∈ )︃ ^ 0 X that h1 (J (q1 ) ) ⊕ h2 (J (q2 ) ) commutes with for appropriate odd polynomials h1 (λ) and h2 (λ) ^ t V q1 0 −V q2 X ^ = of degrees smaller than q1 or q2 respectively. Actually, this is further equivalent to the fact that h1 (J (q1 ) )X (q2 ) (q1 ) ^ (q2 ) ^ ^ Xh2 (J ). Recall that J X = XJ , hence it easily follows that for each s = 0, 1, . . . , q1 , we have (︀ )︀ ^ = 0q ×2d |W s = X(J ^ (q2 ) )s , (J (q1 ) )s X 1 where W s ∈ M q1 is strictly upper-triangular matrix with ±1 on s-th super diagonal and zeros elsewhere. Thus, writing h1 (x) = α1 x + α3 x3 + · · · + α q1 −2 x q1 −2 , h2 (x) = β1 x + β3 x3 + · · · + β q2 −2 x q2 −2 ∑︀ ∑︀ ^ = Xh ^ 2 (J (q2 ) ) implies α i W i = then h1 (J (q1 ) )X β j W j . Since matrices W1 , W3 , . . . , W q1 −2 are linearly independent while W q1 = W q1 +2 = · · · = 0 we deduce α i = β i for i = 1, . . . , q1 − 2. This shows that h1 (J (q1 ) ) = h2 (J (q1 ) ) and we are done as before. The case when d is even can be treated in a similar way. We just have to take the block matrix )︃ )︃ (︃(︃ √ (︀ )︀d+1 ^ X − −1 J (q1 ) −1 T=S √ (︀ (q ) )︀d+1 ⊕ 0n−q1 −q2 S, ^ t V q1 −V q2 X −1 J 2 ^ = where S is similar as above except that now X (︃ ⃒ 0 (︁ ⃒ (q1 −1)×1 0q1 ×2d ⃒ 01×1 ^′ D 01×(q1 −1) )︃ )︁ with diagonal matrix ^ ′ being a compression of −D′2 onto the last (q1 − 1) basis vectors, where diagonal matrix D′2 comes from D (︃ )︃ 0 D′2 (q2 ) d+1 (J ) = . Similarly as in the previous case, we derive that T 2 = 0 and that T commutes with 0 0 A1 ⊕ A2 . So, we are done as above. Substep 13.f. Before continuing with the last case we will show that every cube-zero alternate matrix is fixed by ϕ Q , modulo scalars. So, let T be an alternate matrix with T 3 = 0. Since every cube-zero alternate matrix consists of square-zero blocks and cube-zero blocks we may write without loss of generality that T = (K (2,2) ⊕ . . . ⊕ K (2,2) ) ⊕ (K (3) ⊕ . . . ⊕ K (3) ) ⊕ 0. Now, by subssteps 13.c and 13.e, we know that ϕ Q (T) has a similar blockdiagonal structure with each block being an odd polynomial of the corresponding block. That is, ϕ Q (T) = (α1 K (2,2) ⊕ . . . ⊕ α r K (2,2) )⊕(β1 K (3) ⊕ . . . ⊕ β s K (3) )⊕0. Furthermore, from substep 13.c we know that α1 = . . . = α r and from subsetep 13.e we learn that β1 = . . . = β s . Hence, ϕ Q (T) = α(K (2,2) ⊕ . . . ⊕ K (2,2) ) ⊕ β(K (3) ⊕ . . . ⊕ (2,2) K (3) ) ⊕ 0. It only remains to show that α =⎛β. To this end, we may ⊕ K (3) ⊕ 0 and ⎞ assume that T = K (︃ )︃ 0 −1 0 i 0 X ⎜ ⎟ (2,2) (3) 1 i ϕ Q (T) = αK ⊕ βK ⊕ 0. Using X = ⎝0 0 . Since 0⎠, define alternate Y = 2 + 2 −X t 0 0 0 − 21 + 2i 0 (rank Y , rank Y 2 , rank Y 3 ) = (4, 2, 0), it is easy to see that Y is similar to K (3) ⊕ K (3) ⊕ 0 and, hence, fixed by ϕ Q , modulo scalars. Moreover, Y commutes with T. Therefore, Y has to commute with ϕ Q (T) as well. But then scalars α and β have to be equal. Hence, every cube-zero alternate matrix is fixed by ϕ Q , modulo scalars. Preserving zeros of Lie product on alternate matrices | 99 i pi ) Substep 13.g. Suppose lastly that K (p and K (q j ) share the same eigenvalue, i.e, λ i = 0. Again for simplicity λi we temporarily permute the diagonal blocks so that, in the present substep only, A1 ⊕ A2 = K (q j ) ⊕ K0(p i p i ) ⊕ . . . . Define an alternate block matrix (︃ T= 0 −X t X 0 )︃ ⊕ 0n−q j −2p i , where √ √ )︀ (︀ )︀ ^ · Id2p i + −1V2p i Idq j − −1V q j · X (︁ ^ is a q j × 2p i matrix defined as follows. If q j ≤ p i , then we take X ^ = 0q ×(p −q ) and X j i j X= 1 2 (︀ (︀ )︀ D = Idk0 ⊕ diag (−1), (−1)2 , . . . , (−1)k0 −1 , k0 = D )︁ 0q j ×p i , where q j +1 2 . )︃ [D|0p i ×p i ] , where 0(q j −p i )×2p i (︃ ^= If q j > p i , then we take X D= {︃ Idp i ; p i ≤ k0 := Idk0 ⊕ diag (−1), (−1)2 , . . . , (−1) (︀ )︀ p −k0 i ; p i > k0 := q j +1 2 q j +1 2 . (︀ )︀ ^=X ^ J p i ⊕ (−J p i ) , so A1 ⊕ A2 commutes with T. Let us show that An easy exercise gives that J (q j ) X T 3 = 0. Firstly, recall that T 2 = −(XX t ⊕ X t X) is block-diagonal. We claim that XX t = 0. To see this, note that √ √ (Id2p i − −1V2p i )2 = −2 −1V2p i and that V q2j = Idq j . This yields that XX t = √ √ √ ^ 2p i X ^ t V q j · V q j (Idq j − −1V q j ). −1(Idq j − −1V q j ) · XV ^ · V2p i X ^ t V q j = 0 which gives the claimed XX t = 0. Then also X t XX t = 0, so T 3 = 0. Now, it is easy to see that X Consequently, by substep 13.f, T is fixed by ϕ Q , modulo a scalar. Since A1 ⊕ A2 commutes with T, then ϕ Q (A1 ⊕ A2 ) commutes with ϕ Q (T) ∈ CT as well. Recall that (︀ (︀ )︀ ϕ Q (A1 ⊕ A2 ) = h j K (q j ) ) ⊕ f i K0(p i p i ) ⊕ . . . where h j and f i are odd polynomials of degrees less that q j and p i , respectively. It follows easily from ϕ Q (A1 ⊕ A2 )T = Tϕ Q (A1 ⊕ A2 ) that h j and f i coincide in the first min{q j − 1, p i − 1} monomials, as desired (here, we ^ = X(J ^ p i ⊕ (−J p i ))k ≠ 0 for k = 1, . . . , min{q j − 1, p i − 1}). The proof is completed. use that (J (q j ) )k X References [1] [2] [3] [4] [5] [6] [7] [8] R. Bhatia, P. Rosenthal, How and why to solve the operator equation AX − XB = Y, Bull. London Math. Soc. 29 (1997), 1–21. M. Brešar, Commuting traces of biadditive mappings, commutativity preserving mappings, and Lie mappings, Trans. Amer. Math. Soc. 335 (1993), 525–546. M. Brešar, C. R. Miers, Commutativity preserving mappings of von Neumann algebras, Canad. J. Math. 45 (1993), 695–708. J.-T. Chan, C.-K. Li, N.-S. Sze, Mappings on matrices: invariance of functional values of matrix products, J. Aust. Math. Soc. 81 (2006), 165–184. M. D. Choi, A. Jafarian, H. Radjavi, Linear maps preserving commutativity, Linear Algebra Appl. 87 (1987), 227–241. J. Cui, J. Hou, Maps leaving functional values of operator products invariant, Linear Algebra Appl. 428 (2008), 1649–1663. C.-A. Faure, An elementary proof of the fundamental theorem of projective geometry, Geom. Dedicata 90 (2002), 145–151. P.A. Fillmore, D.A. Herrero, W.E. Longstaff, The hyperinvariant subspace lattice of a linear transformation, Linear Algebra Appl. 17 (1977), 125–132. 100 | Ajda Fošner and Bojan Kuzma [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] A. Fošner, B. Kuzma, T. Kuzma, N.-S. Sze, Maps preserving matrix pairs with zero Jordan product, Linear Multilinear Algebra 59 (2011), 507–529. W. Fulton, Algebraic topology: a first course, Springer, Graduate Texts in Mathematics 153, New York (1995). F. R. Gantmacher, Applications of the theory of matrices, Interscience Publishers, Inc., New York (1959). R. A. Horn, C. R. Johnson, Matrix analysis, Cambridge University Press, Cambridge (1985). C.-K. Li, S. Pierce, Linear preserver problems, Amer. Math. Monthly 108 (2001), 591–605. C.-K. Li, N.-K. Tsing, Linear preserver problems: A brief introduction and some special techniques, Linear Algebra Appl. 162164 (1992), 217–235. G. Lumer, M. Rosenblum, Linear operator equations, Proc. Amer. Math. Soc. 10 (1959), 32–41. L. Molnár, Selected preserver problems on algebraic structures of linear operators and on function spaces, Springer, Lecture Notes in Mathematics 1895, Berlin (2007). L. Molnár, P. Šemrl, Nonlinear commutativity preserving maps on self-adjoint operators, Q. J. Math. 56 (2005), 589–595. M. Omladič, H. Radjavi, P. Šemrl, Preserving commutativity, J. Pure Appl. Algebra 156 (2001), 309–328. T. Petek, Additive mappings preserving commutativity, Linear Multilinear Algebra 42 (1997), 205–211. T. Petek, Mappings preserving spectrum and commutativity on Hermitian matrices, Linear Algebra Appl. 290 (1999), 167– 191. T. Petek, A note on additive commutativity-preserving mappings, Publ. Math. (Debr.) 56 (2000), 53–61. T. Petek, H. Sarria, Spectrum and commutativity preserving mappings on H2 , Linear Algebra Appl. 364 (2003), 317–319. T. Petek, P. Šemrl, Adjacency preserving maps on matrices and operators, Proc. R. Soc. Edinb. 132 (2002), 661–684. S. Pierce et.al., A survey of linear preserver problems, Linear Multilinear Algebra 33 (1992), 1–192. P. Šemrl, Non-linear commutativity preserving maps, Acta Sci. Math. (Szeged) 71 (2005), 781–819. P. Šemrl, Commutativity preserving maps, Linear Algebra Appl. 429 (2008), 1051–1070. W. Watkins, Linear maps that preserve commuting pairs of matrices, Linear Algebra Appl. 14 (1976), 29–35.
© Copyright 2026 Paperzz