2009 American Control Conference Hyatt Regency Riverfront, St. Louis, MO, USA June 10-12, 2009 FrA14.6 An Algorithm for the Long Run Average Cost Problem for Linear Systems with Non-observed Markov Jump Parameters Carlos A. Silva and Eduardo F. Costa Abstract— This paper addresses the problem of long run average cost for linear systems with non-observed Markov jump parameters. We present an algorithm that relies on the approximation of the (infinite horizon) cost via its finite horizon version and uses an evolutionary-based algorithm for the finite horizon cost. A numerical example illustrates the proposed algorithm. I. Introduction There has been a great deal of attention to linear systems with Markov jump parameters (LSMJP). LSMJP constitutes a well known class of system that can be successfully employed in applications featuring random failures, environmental changes and other phenomena that lead to random, abrupt changes of behaviour. There exist numerous results, spanning from notions of stability [13], stabilization [16], [9], basic features such as stabilizability and detectability [3], to optimal solution to finite and infinite quadratic costs [14], [7], [8] and filtering [10]. The available results for LSMJP are strong enough to allow for parallels with standard linear systems, mainly in scenarios with complete state observation (which includes the observation of the Markov state), and we mention as illustration the existence of generalized Riccati equations that characterize the optimal solution to the long run average cost (LRAC), even in the presence of additive Gaussian noise, see [6]. However, the parallel with standard linear systems is weaker when dealing with incomplete or no observation of the Markov state. For instance, in [11] authors need to consider a variationallike approach to the finite horizon quadratic cost problem with partial observation. In the context of LRAC that we are interested in this paper, there only exists a result on bounds for the cost [19], which allows to study existence of the LRAC and convergence of the finite horizon average cost to the LRAC, assuming controls in static state feedback form. Direct extension of the above mentioned variationallike approach seems not to be viable, as it typically provides non-stationary controls, whereas the LRAC, as an infinite horizon problem, requires stationary controls. In fact, the LRAC problem for LSMJP with unobserved Markov state can be interpreted in simple terms as how to obtain a static gain that is not a function of the Markov state and minimizes the average cost incurred by the controlled Markov jump system. Such controls can be implemented when the Markov variable is not accessible and are easier to implement than non-stationary controls, and they are stable (in a certain stochastic sense) provided some mild detectability-like conditions hold, making them of much appeal for applications. However, to the extent of our knowledge there is no available study on methods for providing solutions to the LRAC problem. In this paper, we present a simple algorithm, based on the idea of approximation via the finite horizon cost (FHC). The FHC problem with horizon is dealt with a genetic algorithm, aiming at a suboptimal static gain in the form uk = Kxk , where xk is the so-called continuous state and u is the control. Next, we employ the algorithm for the FHC with increasing horizon T , for approximating the LRAC. The LRAC is more complex as it involves issues of stability and sensitivity to the initial condition of the system. To handle these difficulties, we employ a modified cost (by including a parameter ε > 0), and analyse the impact on the original cost. We also show that stabilizing gains K are associated to finite LRAC with ε > 0. These contributions are presented in a quite simple and comprehensive manner, and are fundamental for the approximation of the LRAC via the FHC and, hence, for our algorithm to provide a solution to the LRAC. Apart from these theoretical results, we employ some new strategies in the genetic algorithm for the FHC problem: (i) use of solutions given by coupled Riccati equations from the LSMJP theory to initialize the population and (ii) representation of each element of the gain in a polynomial form to increase the number of genes of each element of the population, as detailed in Section III. The paper is organized as follows. In Section II we formalize the LRAC and FHC problems for LSMJP. The algorithm for the FHC is detailed in Section III, and in Section IV we obtain some results that allow us to use this algorithm to approximate the LRAC. Section V presents an illustrative example. We finish with some concluding remarks. C. A. Silva and E. F. Costa are with Depto. de Matemática Aplicada e Estatı́stica, Universidade de São Paulo, C.P. 668, 13560-970, São Carlos, SP, Brazil. [email protected], ef- [email protected] Research supported by CNPq Grant 304429/2007-4, FAPESP Grants 2008/02035-8 and 2003/06736-7. 978-1-4244-4524-0/09/$25.00 ©2009 AACC II. Problem Formulation and Preliminary Results One simple way to describe the discrete-time LSMJP considered in this paper is to take into account, initially, 4434 Proposition 1. Let X ∈ Mn , Q ∈ Mn and Σ ∈ Mn be defined by Xi = x0 x′0 πi , Qi = Q and Σi = E, ∀i ∈ S. Then, a standard time-varying linear system in the form xk+1 = Sk xk + Fk uk + wk , k ≥ 0, the initial condition x0 ∈ R n , where x ∈ R n is the state, u ∈ R r is the input and wk is a zero-mean independent Gaussian random variable with covariance matrix E. Consider the cost function T WT = ∑ x′k Qxk + u′k Ruk , k=0 Q′ with Q = ≥ 0 and R = R′ > 0. However, instead of the standard hypothesis that the values of Sk and Fk are known a-priori for all time instants k ≥ 0, we consider that they assume values from the collections A = {A1 , . . . , AS } B = {B1 , . . . , BS } respectively, accordingly to a Markov chain {θ0 , θ1 , . . .}: Sk = Aθk , Fk = Bθk , k ≥ 0. The transition probabilities are P(θk+1 = j|θk = i) = pi j with i, j ∈ S = {1, . . . , S}, and the initial distribution is π = [P(θ0 = 1), . . . , P(θ0 = S)] = [π1 , . . . , πS ]. We assume that the matrix P = [pi j ] and the vector π are known. In this situation, the system is inherently stochastic and xk forms a stochastic process, in such a manner that W as defined above is a random variable, and we consider its expected value for optimization purposes, Y T = E {W T }. It is a well known fact in the LSMJP literature that Y T can be represented in terms of certain linear operators (denoted here by T) involving the conditional second moment matrices of the system, as we present in the sequel. Let Mr,s denote the linear space formed by a number S of r × s-dimensional matrices, Mr,s = {U = (U1 , . . . ,US )}. Also, Mr ≡ Mr,r . For U,V ∈ Mr , U ≥ V signifies that Ui − Vi ∈ Rr0 for each i ∈ N , and similarly for other mathematical relations. Consider tr{·} as the trace operator. It is known that Mr,s equipped with the inner product S hU,V i = ∑ i=1 ∑ p jiU jV jU j′ , T 0 (V ) ∀i ∈ N , E D k−1 TAk (X) + ∑ TAl (Σ) , Q . (2) l=0 The collection of matrices X in Proposition 1 represents the conditional second moment of the initial condition, Xi = E {x0 x′0 |θ0 = i}P(θ0 = i). We write Y (X) to emphasize the dependence of Y on X. Observation and Control Structures We assume that only the quantity xk is available to the controller at each time instant k. θk is not observed. In connection, we consider a linear state feedback control in the form uk = Kxk . (3) K is referred to as a static gain (in opposition to timedependent gains of the form Kk ). Static gains are among the simplest to implement controls, thus being of interest for many applications. For the complete observation case (xk and θk observed), with controls in the form uk = Kθk xk , (4) there are strong results in LSMJP literature, parallelling the deterministic linear systems theory. One result that shall be useful to us is the solution of finite horizon quadratic cost problems via coupled algebraic Riccati equations, allowing to compute optimal gains Ki in a simple way, as presented in the Appendix. For a given gain K (or a collection of gains K = {Ki , i ∈ S} in the complete observation case), we have a closed loop form xk+1 = (Sk + Fk K) (respectively, xk+1 = (Sk + Fk Kk )), giving rise to collections of “closed loop” matrices AK = (A1 + B1 K, . . . , AS + BS K) (respectively, AK = (A1 + B1 K1 , . . . , AS + BS KS )), operators TAK and costs given by (2). We denote YKT and YKT (X) to emphasize the dependence on K and on X, E k−1 T D (5) YKT (X) = ∑ TAk K (X) + ∑ TAl K (Σ), Q . l=0 k=0 Problem Formulation We are interested in the optimization problems S j=1 T ∑ k=0 tr{Ui′Vi } forms a Hilbert space. Following the notation of [3] we define, for U ∈ Mn , V ∈ Mn , the operator TU : Mn → Mn by TU,i (V ) := YT = (1) P1 : minGYGT (X), in the scenario with T < ∞, and P2 : minG ZG , when T = ∞, where ZG = lim sup YGT (X)/T. and we define for convenience = V , and for t ≥ 1, Tt (V ) = T(Tt−1 (V )). It is simple to check that T is linear. The next result is an adaptation of the results in [6, ch 3], see also [19]. T →∞ T P2 is referred to as the LRAC problem. The scenario with T = ∞ is more complex in the sense that finite Z does not ensure that the controlled system is stable, as we shall see later. 4435 TABLE I the infinite horizon scenario involves the question of stability and related issues of sensitivity to the initial condition. For instance, we may have limT →∞ (1/T )YGT (X) < ∞ and limT →∞ (1/T )YGT (X) = ∞, W 6= X. Moreover, lack of positiveness of Σ and Q may lead to finite cost controls that are not stabilizable, provided certain conditions of detectability and stabilizability are not satisfied. In order to overcome these difficulties, we consider the standard hypothesis that the Markov chain is ergodic, in such a manner that the limiting distribution P∞ π is unique [1]. We also consider the modified cost Genetic Algorithm for the FHC problem P1 1) Set the GA and system parameters 2) Initialize the population K ℓ as in (6) and (7) 3) Perform genetic and evolutionary operators to obtain a new population and the gains K ℓ (7) 4) For each K ℓ , calculate Y ℓ (X) via (5) 5) If the stop criterion is not satisfied, return to (3) III. A Genetic Algorithm for the Problem P1 One important issue that arises when dealing with GAs is how to create an adequate initial population. In principle, one could simply take randomly generated gains K 1 , . . .. However, we have observed in many numerical examples that the cost associated to many of these gains are extremely high, leading to a high probability of being excluded in the selection process. As a result, the “genetic variety” of the the second generation of the population decays, and the algorithm performance is poor, see e.g. [18], [15], [17]. In order to overcome this difficulty, we initially solve the optimization problem within the class of controls in the form (4) via the recursive coupled algebraic Riccati equations described in the Appendix, leading to a collection of gains L1 , . . . , LS that are optimal for the complete state observation problem, and then we use these gains to obtain the initial population as follows: K ℓ = α1 L1 + . . . + αS LS , YKε (X) = T ∑ k=0 D E k−1 TAk K (X) + ∑ TAl K (Σ + ε I), Q + ε I , l=0 where I = (I, . . . , I) ∈ Mn , and the related LRAC ZGε , with the property that ZKε is stabilizing if and only if the gain K is stabilizing, and we evaluate the impact on the original costs. Let us start formalizing the stabilizability notion. Definition 1. We say that a static feedback gain K (respectively, a collection of gains K ∈ Mr,n in the complete observation scenario) is mean square (MS) stabilizing when, for each Σ ∈ Mn satisfying Σ = Σ′ ≥ 0, there exists Γ ∈ Mn , Γ = Γ′ ≥ 0, such that k−1 TAk K (X) + ∑ TAl K (Σ) < Γ (8) l=0 ℓ = 1, . . . , q (6) where L1 , . . . , LS are given by (14) (or by (15) in the context of the LRAC problem P2), q is the size of the initial population and αi , i ∈ S, are zero-mean independent Gaussian variables with some arbitrary covariance matrix P. The static gain K (or, in the complete observation case, each element the collection of static gains K) may be low dimensional, as in Example 1, where K = [K(1, 1) K(1, 2) K(1, 3)]. In these situations, if we simply identify each element of K as one gene of each element of the population, we will have a few genes, and we have observed a poor performance of the GA. We employ the polynomial representation for each element of K = [K(i, j)], K(i, j) = β1 (i, j)1 + β2 (i, j)2 + . . . + β p(i, j) p . (7) For the initialization of each element of the population, ℓ (i, j) K ℓ , we employ randomly generated β1ℓ (i, j), . . . , β p−1 ℓ ℓ ℓ 1 ℓ and define β p (i, j) = K (i, j)− β1 (i, j) + . . .+ β p−1(i, j) p−1 ; the GA will determine the β ’s of the following generations. The Algorithm is presented in Table I. IV. An Algorithm for the Problem P2 (LRAC) The problem P2 (LRAC problem for LSMJP) is far more complex than the finite horizon problem P1 because for each X ∈ Mn , X = X ′ ≥ 0, provided k ≥ k̄ for some k̄ (possibly dependent on X). Remark 1. MS-stabilizability as defined above and the related notion of MS-stability are equivalent to the standard MS notions in the LSMJP literature, e.g. K is MSstabilizing if and only if k ∑ TAl K (X) < ∞, l=0 ∀X = X ′ ≥ 0. (9) For the equivalences, we refer to [6]. Proposition 2. If K is not MS-stabilizing, then for each M ∈ Mn , M = M ′ ≥ 0, there exists a sufficiently large integer T̄ such that ∑T̄k=0 TAk K (I) ≥ M. Proof: From (9), if K is not stabilizable we have that there exists X such that, for each Γ, ∑kl=0 TAk K (X) ≥ Γ for some k. The linearity of T and the fact that υ I ≥ X for some υ ≥ 0, allow to write ∑kl=0 TAl K (I) ≥ υ −1 ∑kl=0 TAl K (X) ≥ υ −1 Γ. Lemma 1. K is MS-stabilizing if and only if ZKε < ∞. 4436 Proof: Necessity. Replacing Σ by Σ+ ε I in Definition Proof: The statement that ZK ≤ ZKε is trivial. Let us denote the quantities Γ and k̄ in Definition 1 by ΓΣ and k̄X to emphasize the dependence on Σ and X respectively. Let k̃ = max(k̄X , k̄0 ). It is straightforward to check from (5) and the linearity of T that 1 and employing Proposition 1 we write E k−1 YKε ,T (X) 1 T −1 D = ∑ TAk K (X) + ∑ TAl K (Σ + ε I) , Q + ε I T T k=0 l=0 E k−1 k̄−1 D 1 = ∑ TAk K (X) + ∑ TAl K (Σ + ε I) , Q + ε I T k=0 l=0 E k−1 T −1 D 1 + ∑ TAk K (X) + ∑ TAl K (Σ + ε I) , Q + ε I T k=k̄ l=0 ∆ 1 + T T T −1 ∑ lim sup T →∞ YKε ,T (X) T ≤ Γ, Q + ε I Sufficiency. Let Γ = ZKε < ∞. Let us deny the statement by assuming that K is not stabilizing. In Proposition 2, set M = ε −2 (ΓI + 1) and consider the corresponding T̄ . We can write E ℓ−1 1 T −1 D Γ = lim sup ∑ TAℓ K (X) + ∑ TAl K (Σ + ε I) , Q + ε I T →∞ X T l=0 ℓ=0 E T −1 D ℓ−1 1 ≥ ε 2 lim sup ∑ ∑ TAl K (I) , I T →∞ X T ℓ=0 l=0 E T −1 D ℓ−1 1 = ε 2 lim TAl K (I) , I ∑ ∑ T →∞ T ℓ=0 l=0 E T −1 D T̄ 1 ≥ ε 2 lim TAl K (I) , I ∑ ∑ T →∞ T ℓ=T̄ l=0 1 T −1 T − T̄ M, I ∑ M , I ≥ ε 2 Tlim T →∞ T →∞ T ℓ=T̄ ≥ ε 2 ε −2 (ΓI + 1) , I ≥ Γ + 1, ≥ ε 2 lim which is an absurd. The algorithm we propose in this section seeks for the LRAC using the approximation via Y ε ,T (X)/T . Thus, it is important to show that this quantity converges to Z ε as T → ∞. This issue was studied and solved in [19, Corollary 2]; for convenience, we present an adaptation to the present context, next. Proposition 3. Assume the Markov chain θ is ergodic and that K is MS-stabilizing. Then, there exist scalars α , β ≥ 0 such that Y ε ,T (X) α kX(ℓ)k + β (11) ZK − ≤ T T Next, the impact of the scalar ε in ZKε is evaluated in terms of the original LRAC ZK , showing that ZKε converges to ZK as ε tends to zero, provided the gain K is stabilizing. Lemma 2. If K is MS-stabilizing, then there exists ξ ≥ 0 such that ZK ≤ ZKε ≤ ZK + εξ . ∑ TAl K (I) ≤ ε ΓI (12) l=0 l=0 ∆ T − k̄ Γ, Qε I = + Γ, Q + ε I , T T k=k̄ (10) k k−1 l k̄−1 T (X) + T (Σ + where we set ∆ = ε I) , Q+ ∑l=0 AK ∑k=0 AK ε I . This leads to ≤ k−1 k−1 TAk K (0) + ∑ TAl K (ε I) = ε for k ≥ k̄0 . One can check that E T −1 D k−1 YKε ,T (X) − YKT (X) = ∑ ∑ TAl K (ε I) , Q + ∑ k=0 k=0 k−1 l=0 E T −1 D k−1 E TAk K (X) + ∑ TAl K (Σ) , ε I + ∑ ∑ TAl K (ε I) , ε I T −1 D l=0 k=0 l=0 (13) Now, evaluations similar to the one in (10) for the terms on the right-hand side of (13) provides YKε ,T (X) − YKT (X) T ∆ T − k̃ ε ΓΣ , I + ε ΓI , Q + ε 2 ΓI , I , ≤ + T T which leads to YKε ,T (X) − YKT (X) ≤ εξ T T →∞ where we set ξ = ΓΣ , I + ΓI , Q + ε ΓI , I . In the algorithm for the LRAC, we use the coupled algebraic Riccati equation in (15) (in the Appendix) to initialize the population instead of the coupled recursive Riccati equations in (14). The solution of (15) with positive weighting matrices is connected to stabilizing solutions of the LRAC problem, see e.g. [12], thus providing stabilizing initial gains Li for (6). We include in the Appendix a simple method for solving the coupled algebraic Riccati equation, for ease of reference. lim sup V. Numerical Example Example 1 (Magnetic suspension system). We take into account the model of a magnetic suspension system, as presented in [5]. For ease of reference, next we present the data of the system, 0 1 0 0 . 0 A1 = 1750 0 −34.07 , B1 = 0 0 −0.0383 1.9231 Here we consider failures in the control action u. Let the transition probability matrix P be given by 0.9 0.1 P= , 0.5 0.5 and assume that whenever θk = 2 the system is not affected by the controller, and we set A2 = A1 and B2 = 0 accordingly. We also consider zero-mean additive noise with covariance matrix GG′ , with 1 0 G = 1e−6 1 0 . 0 1 4437 and the fact that perfect observation of the jump variable θ may be difficult in many practical situations. In order to overcome some intrinsic difficulties of the problem, we have proposed a modification on the original cost, parameterized by a scalar ε , and we show that the modified cost ZKε converges to the original one when ε tends to zero, provided K is MS-stabilizing. We also shown that finite ZKε is strongly connected with MSstabilizing K, in such a manner that, when the method converges, the obtained solution is stabilizing. 8.0e+007 Y ε ,T (X) T 7.5e+007 7.0e+007 6.5e+007 6.0e+007 5.5e+007 0 5 10 15 20 T 25 30 References Fig. 1. Cost Y ε ,T (X)/T with T = 10 versus the number of iterations t of the GA. 1.2e+008 Y ε ,T (X) T 1.1e+008 1.0e+008 9.0e+007 8.0e+007 7.0e+007 0 100 200 300 T 400 500 600 700 Fig. 2. Y ε ,T (X)/T converging to the LRAC as the horizon T increases. The weighting matrices are R = 1 and Q = I. We consider initial condition X = I, and we set ε = 10−3 . For the initialization of the algorithm, we employ the coupled Riccati equations (15) (solved using the method presented in the Appendix), which yields L1 ≈ [2797 66.87 − 53.71]; L2 = [0 0 0]. We start with horizon length T = 10. Figure 1 shows the behaviour of the best average cost Y ε ,T (X)/T as a function of the number of iterations (number of population) of the method of Table 1. Figure 2 shows how the average cost evolves as T increases, illustrating the convergence to the LRAC, Z ε (X). The obtained static gain K and the attained LRAC are given by K ≈ [2324 57.91 − 44.99]; Z ε (X) ≈ 117569345. VI. Conclusions In this paper we have presented a genetic algorithm for the problems of finite horizon cost and of long run average cost, with static feedback controls, for discretetime linear systems with Markov jump parameters and partial state observation. This is a relevant problem for applications, taking into account the simplicity of the controller (in the form uk = Kxk ) and its implementation, [1] E. Cinlar. Introduction to Stochastic Processes. Prentice-Hall, 1975. [2] E. F. Costa and J. B. R. do Val. Weak detectability and the linear quadratic control problem of discrete-time Markov jump linear systems. International Journal of Control. Special issue on Switched and Polytopic Linear Systems, 75:1282– 1292, 2002. [3] E. F. Costa, J. B. R. do Val, and M. D. Fragoso. A new approach to detectability of discrete-time infinite Markov jump linear systems. SIAM Journal on Control and Optimization, 43:2132–2156, 2005. [4] E. F. Costa, A. L. P. Manfrim, and J. B. R. do Val. Weak controllability and weak stabilizability concepts for linear systems with Markov jump parameters. In Proc. ACC’06 American Control Conference, Minneapolis,USA, 2006. [5] E. F. Costa, V. A. Oliveira, and J. B. Vargas. Digital implementation of a magnetic suspension control system for laboratory experiments. IEEE Transactions on Automatic Control, 42:315 – 322, 1999. [6] O. L. V. Costa, M. D. Fragoso, and R. P. Marques. DiscreteTime Markovian Jump Linear Systems. Springer-Verlag, New York, 2005. [7] O. L. V. Costa and M.D. Fragoso. Discrete-time LQ-optimal control problems for infinite Markov jump parameter systems. IEEE Transactions on Automatic Control, AC-40:2076–2088, 1995. [8] O. L. V. Costa and E. F. Tuesta. Finite horizon quadratic optimal control and a separation principle for Markovian jump linear systems. IEEE Trans. Automat. Control, 48, 2003. [9] C. E. de Souza and D. F. Coutinho. Robust stability of a class of uncertain Markov jump nonlinear systems. IEEE Transactions on Automatic Control, 51(11):1825–1831, 2006. [10] C. E. de Souza and M. D. Fragoso. H∞ filtering for discretetime linear systems with Markovian jumping parameters. International Journal of Robust and Nonlinear Control, 13:1299– 1316, 2003. [11] J. do Val and T. Basar. Receding horizon control of jump linear systems and a macroeconomic policy problem. Journal of Economic Dynamics & Control, 23:1099–1131, 1999. [12] J. B. R. do Val and E. F. Costa. Stabilizability and positiveness of solutions of the jump linear quadratic problem and the coupled algebraic Riccati equation. IEEE Transactions on Automatic Control, 50:691–695, 2005. [13] M. D. Fragoso and O. L. V. Costa. A unified approach for stochastic and mean square stability of continuous-time linear systems with Markovian jumping parameters and additive disturbances. SIAM Journal on Control and Optimization, 44(4):1165–1191, 2005. [14] Z. Gajic and R. Losada. Solution of the state-dependent noise optimal control problem in terms of Lyapunov iterations. Automatica, 35:951–954, 1999. [15] G. P. H. Chou and C.-H. Chu. Genetic algorithms for communications network design-an empirical study of the factors that influence performance. volume 5, pages 612–619. IEEE Transactions on Evolutionary Computation, June 2001. [16] Y. Ji and H. J. Chizeck. Jump linear quadratic Gaussian control: Steady-state solution and testable conditions. Control Theory and Advanced Technology, 6(3):289–319, 1990. [17] J. Liu S. P. M. Choi and S. P. Chan. A genetic agent-based negotiation system. Computer Networks, 37(2):195–204, 2001. 4438 [18] A. B. Simões and E. Costa. Transposition versus crossover: An empirical study. pages 612–619, Orlando, Florida, USA, July 1999. Proceedings of the Genetic and Evolutionary Computation Conference. [19] A. N. Vargas, E. F. Costa, and J. B. R. do Val. Bounds for the finite horizon cost of Markov jump linear systems with additive noise and convergence for the long run average cost. In Conference on Decision and Control 2006, 2006. Appendix Consider the following coupled recursive Riccati equation, with initial condition Xi0 = Qi , i ∈ S, Xik+1 = A′i Eki Ai − (A′i Eki Ai Bi )(R + B′i Eki Bi )−1 (B′i Eki Ai ) + Qi (14) where Eki = ∑Sj=1 pi j X jk is the coupling term. Consider the associated coupled algebraic Riccati equation in the variables Xi = Xi ≥ 0, i ∈ S, Xi = A′i Ei Ai − (A′i Ei Ai Bi )(R + B′i Ei Bi )−1 (B′i Ei Ai ) + Qi where Ei = ∑Sj=1 pi j X j . (14) can be solved recursively. We present here, for ease of reference, the following algorithm [2], [4], which converges to the solution of (15) if and only if a solution exists. Method for solving (15) Step 1. Set κi ≤ 1 as the largest integer for which √ κi pii Ai is stable. Step 2. Set X 0 = (X10 , ..., XN0 ) ∈ Mn+ . Step 3. For k = 1, 2, .. and i = 1, 2, ..., N solve the standard algebraic Riccati equations: −Xik +κi pii A′i Xik Ai + A′i Ẽki Ai − (κi pii A′i Xik Bi + A′i Ẽki Bi ) × (R + κi pii B′i Xik Bi + B′i Ẽki Bi )−1 × (κi pii B′i Xik Ai + B′i Ẽki Ai ) + Qi where (15) i Ẽki = ∑ pi j X jk + (1 − κi)pii Xik . j=1 4439 =0 (16)
© Copyright 2026 Paperzz