2 SIEGMUND DUALITY FOR MARKOV CHAINS ON PARTIALLY ORDERED STATE SPACES 3 PAWEŁ LOREK 4 Mathematical Institute University of Wrocław 1 5 Abstract. For Markov chains on finite partially ordered state space we show that the Siegmund dual exists if and only if the chain is Möbius monotone, in which case we give formula for the transitions of this dual. This is an extension of Siegmund’s result for totally ordered state spaces in which case the existence of the dual is equivalent to usual stochastic monotonicity (which is - for total ordering - equivalent to Möbius monotonicity). Exploiting the relation between stationary distribution of ergodic chain and absorption probabilities of its Siegmund dual we show three applications: calculating absorption probabilities of chain with two absorbing states knowing stationary distribution of other chain; calculating stationary distribution of ergodic chain knowing absorption probabilities of other chain; providing stable simulation scheme for stationary distribution of a chain provided we can simulate its Siegmund dual. These are accompanied by concrete examples: some nonsymmetric game; some extension of birth and death chain and nonstandard tandem network of two servers respectively. We also show that one can construct a Strong Stationary Dual chain by performing appropriate Doob transform of the Siegmund dual of time-reversed chain. Keywords: Markov chains; Siegmund duality; Strong Stationary Duality; Möbius monotonicity; partial ordering; Doob h-transform. 6 1. Introduction 7 Siegmund [19] introduced a notion of duality (today called Siegmund duality) for Markov chains with discrete time and general state space with total ordering. This duality is intended to relate the stationary distribution of a process with the absorption probability of its dual. In case of total ordering, Siegmund showed that dual chain exists if and only if the original chain is stochastically monotone (w.r.t. total ordering). The aim of this note is a generalization of this result to state spaces which are only partially ordered. Thomas Liggett in his famous book [12] writes (about Siegmund’s result for total ordering): 8 9 10 11 12 13 14 15 E-mail address: [email protected]. Work supported by NCN Research Grant DEC-2013/10/E/ST1/00359. 1 16 “This result depends heavily on the fact that the state space of the processes is totally ordered. In the particle system context, the state spaces are only partially ordered, and unfortunately the analogous result fails. [...] having a (reasonable) dual is a much more special property then being monotone, when the state space is not totally ordered.“ 17 As the main result we show that the existence of the Siegmund dual is equivalent to Möbius monotonicity of the chain (introduced in Lorek and Szekli [15]) w.r.t. fixed partial ordering. In this case we give transitions of the dual. It turns out that the usual stochastic monotonicity of a chain does not imply existence of Siegmund dual for general partial orderings (it does for linear ordering, since then stochastic and Möbius monotonicities are equivalent). In general, the required Möbius monotonicity is not stronger, but different than stochastic monotonicity. There is a one-to-one correspondence (the relation is given in (2.3)) between the stationary distribution of ergodic chain and absorption probabilities of its Siegmund dual. We will show three potential applications of this relation. First application: given chain with two absorbing states (gambler’s ruin-like) we can find ergodic chain for which the original one is a Siegmund dual. Then finding stationary distribution of this chain answers the question about the absorption probabilities. As an example for this application we present some nonsymmetric game, which turns out to be a Siegmund dual to a nonsymmetric random walk on cube considered in Lorek and Szekli [15]. Second application: given ergodic chain, the task is to find its stationary distribution. We can calculate its Siegmund dual, then finding the absorption probabilities of the dual answers the question about the stationary distribution of the ergodic chain. As an example we present ergodic chain for which Siegmund dual corresponds to ”Gambler’s Ruin with Catastrophes and Windfalls“ considered in Hunter [11] for which absorption probabilities are calculated therein explicitly. Third application: This is similar to the second one, we are given an ergodic chain, the task is again to find its stationary distribution. Calculating Siegmund dual can be relatively easy, but finding such probabilities is often a challenging task. However if we are able to simulate the dual, then we can have stable simulation scheme for the stationary distribution of the ergodic chain via simply estimating absorption probabilities. In contrast to most of Monte Carlo Markov Chains (MCMC) we do not require a priori any knowledge about, e.g., rate of convergence, mean absorption time etc. Besides, MCMC are designed to provide a sample from stationary distribution, whereas stable simulation scheme lets us estimate the stationary distribution at the specific state. As the example we present (including calculations of Siegmund dual) some tandem network of two servers. This is some modification of the standard tandem (for which 2 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 stationary distribution is known): when one of the servers is empty, the rates of the other are changed. It turns out that then the stationary distribution is not known (it is surely not of product form as for standard tandem). Foley and McDonald [7] considered similar modified tandem (if the server is idle it helps the other) showing large deviations and rough asymptotics for the stationary distribution. In particular they showed that such changes on borders imply dramatic changes in the stationary distribution. Siegmund dualities appear in various contexts. Original paper Siegmund [19] deals with stochastically monotone chains w.r.t a total ordering. Many authors focused on birth and death chains, e.g., Cox and Rösler [2], Diaconis and Fill [4] (where connections to Strong Stationary Duality is also given), Dette et al. [3] (where the Siegmund duality is related to Wall duality) or Huillet [9] (where a specific birth and death chain, Moran model, was considered). The observation, that such duality holds for some random walks on {0, 1, . . .} goes back to Lindley [13]. The duality was also studied in a financial context, where the probability that the steady-state queue length exceeds level k equals the probability that a dual risk process starting at level k is ruined in a finite time. There were some approaches to extend this to nonlinear state spaces. Błaszczyszyn and Sigman [1] considered Rd -valued Markov processes (their Siegmund dual was set-valued). Recently, Huillet and Martínez [10] considered dualities related to Möbius matrix for other nonlinear state spaces, namely for Markov chains on partitions and sets. In Lorek and Szekli [15] we consider general partial orderings and show that Möbius monotonicity of a time-reversed chain implies existence of a Strong Stationary Dual (SSD) chain on the same state space (more details in Section 5). Recently, in Fill and Lyzinski [6] and in Miclo [17] dualities for onedimensional diffusions were studied. Further research include studying existence of Siegmund dual in d-dimensional diffusions, e.g., the ones studied in Harrison and Williams [8], as well as developing the theory for Möbius monotone processes for continuous-time and general state space chains. The rest of the note is organized as follows. In Section 2 we describe Siegmund duality and relations between the stationary distribution of a chain and the absorption probability of its dual. In Section 3 we shortly recall Möbius monotonicity. In Section 4 we present our main result (Theorem 4.1) on equivalence of Möbius monotonicity and Siegmund duality for general partial orderings. Section 5 gives a connection to Strong Stationary Duality. Finally, in Section 6 we present three applications: in Section 6.1 we give winning/losing probabilities for some nonsymmetric game, in Section 6.2 we give stationary distribution of some extension of birth and death chain and in Section 6.3 we present stable simulation scheme for some nonstandard tandem queue system. 3 96 2. Siegmund duality 97 Let X ∼ (ν, PX ) be a discrete-time Markov chain with initial distribution ν, transition matrix PX and finite state space E = {e1 , . . . , eM } partially ordered by . Throughout the paper we assume it is ergodic with the stationary distribution π. We assume, that there exist unique minimal state, say e1 ,Pand unique maximal state, say eM . P For A ⊆ E define PX (e, A) := e0 ∈A PX (e, e0 ) and similarly π(A) := e∈A π(e). Define also {e}↑ := {e0 ∈ E : e e0 }, {e}↓ := {e0 ∈ E : e0 e} and δ(e, e0 ) = 1(e, e0 ). Recall that Markov chain X is stochastically monotone, if ∀(e1 e2 )∀(D − down set)PX (e1 , D) ≥ P(e2 , D), where D is a down set if (e e0 , e0 ∈ D) =⇒ e ∈ D. We say that Markov chain Z with transition matrix PZ is the Siegmund dual of X if 98 99 100 101 102 103 104 105 106 107 108 (2.1) 109 110 111 112 113 Siegmund [19] showed that for total ordering (then, let us denote elements of E as numbers {1, 2, . . . , M }) such dual exists if and only if X is stochastically monotone. The main thing then is to show that (2.1) holds for one-step transitions. Since the main part of the proof is one line long, we include it here. We want to have PX (i, {j}↓ ) = PZ (j, {i}↑ ). We can calculate (2.2) 114 115 116 117 118 119 120 121 ∀(ei , ej ∈ E) ∀(n ≥ 0) PnX (ei , {ej }↓ ) = PnZ (ej , {ei }↑ ). PZ (j, i) = PZ (j, {i}↑ ) − PZ (j, {i + 1}↑ ) = PX (i, {j}↓ ) − PX (i + 1, {j}↓ ). The latter is nonnegative if and only if X is stochastically monotone. Note, that P (2.2) does not have to define transition matrix, since we may have i PZ (j, i) < 1. Siegmund [19] adds then P one extra absorbing state, say M + 1, and defines PZ (j, M + 1) = 1 − i PZ (j, i). In a similar way, for general partial ordering, if we are able to find a subprobability kernel fulfilling (2.1) for all ei , ej ∈ E, we may be forced to add one extra absorbing state, say eM +1 . Note that (2.1) implies that eM is an absorbing state, thus Z has two absorbing states. Taking limits as n → ∞ on both sides of (2.1) we have 122 (2.3) π({ej }↓ ) = limn→∞ PnZ (ej , {ei }↑ ) = P (Z is absorbed in eM |Z0 = ej ). 124 This way the stationary distribution of X is related to the absorption of its Siegmund dual Z. 125 3. Möbius monotonicity 126 Let C(ei , ej ) = 1(ei ej ). We can always rearrange states so that ei ej implies i ≤ j. Then the matrix C is 0-1 valued, upper-triangular, thus invertible. The inverse C−1 is often denoted as µ and is called Möbius function. The famous Möbius inversion formula (see, e.g., Rota [18]) states: 4 123 127 128 129 130 (3.1) Let F̄ (e) = X f (e0 ), then f (e) = e0 e 131 132 133 134 135 136 137 138 X µ(e, e0 )F̄ (e0 ). e0 e We say that function F̄ : E → RM is Möbius monotone if the resulting f (e) in (3.1) is nonnegative for all e ∈ E. For all states e2 ∈ E denote: F̄e2 (e0 ) = PX (e0 , {e2 }↓ ). We say that chain X is Möbius monotone, if functions F̄e2 are Möbius monotone for all e2 ∈ E. Equivalently, the definition can be stated in matrix form: Definition 3.1. Markov chain X with transition matrix PX is Möbius monotone w.r.t partial order if C−1 PX C ≥ 0 (i.e., each entry nonnegative). In terms of transition probabilities, it means that X ∀(ei , ej ∈ E) µ(ei , e)PX (e, {ej }↓ ) ≥ 0. eei Remark 3.2. In a similar way to (3.1), we have the following version of Möbius inversion formula: X X Let F (e) = f (e0 ), then f (e) = µ(e0 , e)F (e0 ). e0 e 139 140 e0 e This way,X once π({ej }↓ ) is calculated for all ej ∈ E in (2.3), we can calculate π(ej ) = µ(e, ej )π({e}↓ ). eej 141 For more details on Möbius monotonicity see Lorek and Szekli [15]. 142 4. Main result 143 In this section we will show the existence of the Siegmund dual on state space E∗ = E∪{eM +1 }, where we add one extra absorbing state eM +1 (coffin state). At each step, the process can be possibly killed, i.e., absorbed in this state. 144 145 146 147 148 149 150 151 Theorem 4.1. Let X ∼ (ν, PX ) be a Markov chain on E = {e1 , . . . , eM } with partial ordering . Assume e1 is a unique minimal state and eM is a unique maximal state. Then, there exists the Siegmund dual Z of X on E∗ = E ∪ {eM +1 } if and only if X is Möbius monotone. The transitions of the dual are as follows: PZ (ej , ei ) = X µ(ei , e0 )PX (e0 , {ej }↓ ), ei , ej ∈ E, e0 ei (4.1) PZ (eM +1 , ej ) = δ(eM +1 , ej ), X PZ (ek , eM +1 ) = 1 − PZ (ek , e), e∈E 5 ej ∈ E∗ , ek ∈ E. 152 153 154 155 156 Proof. Note that the state space of dual is enriched with one extra absorbing state called eM +1 . We need equation (2.1) to hold for all states from E. One can think that eM +1 is incomparable to all other states in E∗ , what we assume, since it simplifies some notation. First, we show that (2.1) holds for n = 1, what can be rewritten as X PZ (ej , e). PX (ei , {ej }↓ ) = eei 157 Using Möbius inversion formula (3.1) we have X (4.2) PZ (ej , ei ) = µ(ei , e)PX (e, {ej }↓ ), eei what is nonnegative if and only if X is Möbius monotone. For P partial order with unique minimal element e1 the Möbius function fulfills: ei ∈E µ(ei , e) = 1(e = e1 ) (to see this consider first row of C−1 after applying first elementary row operation of Gauss-Jordan elimination). This implies that rhs of (4.2) is not greater than 1, since X X X PZ (ej , ei ) = µ(ei , e)PX (e, {ej }↓ ) ei ∈E ei ∈E eei 158 (∗) = XX µ(ei , e)PX (e, {ej }↓ ) ei ∈E e = X e 159 160 161 162 163 164 165 166 167 X µ(ei , e) = PX (e1 , {ej }↓ ) ≤ 1, ei ∈E where in (∗) we used the fact that µ(ei , e) = 0 for ei e (see, e.g., proof of Proposition 1 in Rota [18]). Note that submatrix PZ with column and row corresponding to coffin state eM +1 excluded, can be written as (C−1 PX C)T , i.e. we have PZ (ej , ei ) = (C−1 PX C)T (ej , ei ) for ej , ei ∈ E. Chapman-Kolmogorov equations allow to extend the transition probabilities to PnZ (ej , ei ) for all ej , ei ∈ E∗ and n ≥ 0. Then, the dual chain is defined. To see that (2.1) holds for ej , ei ∈ E note that since eM +1 is absorbing state, we must have PnZ = (C−1 PnX C)T , i.e. (4.3) 168 PX (e, {ej }↓ ) C(PnZ )T = PnX C, which is (2.1) written in a matrix form. 169 170 Remark 4.2. Note that the Siegmund dual chain will have two absorbing states. Beside the extra state eM +1 , the state which is maximal w.r.t is 6 171 172 also absorbing. Indeed PZ (eM , ei ) = = X 0 e eX i µ(ei , e0 )PX (e0 , {eM }↓ ) µ(ei , e0 ) = 1(ei = eM ). e0 ei 173 174 175 176 Remark 4.3. The assumption on existence of the relaxed, as the following example shows: 4 1 1 1 6 6 6 4 1 1 PX = 6 6 6 , C = 0 1 1 4 0 6 6 6 minimal state cannot be 0 1 1 1 . 0 1 In other words, we have three states, say e1 , e2 , e3 . The state e3 is a maximal one and states e1 , e2 are incomparable. Then, we can calculate 1 1 1 2 2 (C−1 PX C)T = 0 0 0 6 1 6 0 1 . 177 Thus, PZ given in (4.2) does not define a subprobability kernel. 178 Remark 4.4. For given X we can distinguish two Siegmund dual chains. The one defined in (2.1) can be called Siegmund↓ dual, say Z ↓ . The other, say Z ↑ , called Siegmund↑ dual, is the one fulfilling 179 180 ∀(ei , ej ∈ E) ∀(n ≥ 0) PnX (ei , {ej }↑ ) = PnZ (ej , {ei }↓ ). 181 182 183 184 185 The monotonicity defined in Lorek and Szekli [15] was actually called ↓ Möbius monotonicity, whereas ↑ -Möbius monotonicity was defined as (CT )−1 PCT ≥ 0. In a similar way one can have a version of Theorem 4.1, showing that the Siegmund↑ dual exists if and only if X is ↑ -Möbius monotone. We skip the details, noting that these monotonicities are not equivalent. 186 5. Siegmund duality and Strong Stationary Duality 187 In this Section we point out a connection to Strong Stationary Duality. Diaconis and Fill [4] show that their Strong Stationary Dual chain can be constructed in three steps, where one step involves calculating the Siegmund dual w.r.t. to total ordering. It turns out that Strong Stationary Dual chain given in Lorek and Szekli [15] can be constructed in similar three steps, where one involves calculating the Siegmund dual w.r.t. fixed partial ordering. Recall that X ∼ (ν, P) is an ergodic Markov chain on finite state space E = {e1 , . . . , eM } with the stationary distribution π. Let E∗ = {e∗1 , . . . , e∗N } be, possibly different, state space of absorbing Markov chain X ∗ ∼ (ν ∗ , P∗ ) whose unique absorbing state is denoted by e∗N . Define Λ, a matrix of size 7 188 189 190 191 192 193 194 195 196 197 198 N ×M , to be a link if it is a stochastic matrix with the property: Λ(e∗N , e) = π(e). We say that X ∗ is a Strong Stationary Dual (SSD) of X with link Λ if ν = ν ∗ Λ and ΛP = P∗ Λ. 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 Diaconis and Fill [4] prove that then the absorption time T ∗ of X ∗ is so called Strong Stationary Time for X. This is such a random variable T that XT has distribution π and T is independent from XT . The main application is in studying the rate of convergence of an ergodic chain to its stationary distribution, since for such random variable we always have: dT V (νPk , π) ≤ sep(ν, Pk , π) ≤ P (T > k), where dT V stands for total variation distance, and sep stands for separation ”distance“. For details see Diaconis and Fill [4]. In general, there is no recipe on how to find SSD, i.e., a triple E∗ , P∗ , Λ. In Diaconis and Fill [4] authors give a recipe for a dual on the same state space ← − E∗ = E provided that time reversed chain X is stochastically monotone with respect to total ordering. For given P let h be its harmonic function, i.e., such a nonnegative function h that Ph = h. By bold h denote the vector h = (h(e1 ), . . . , h(eM )). Then, Ph defined as Ph (e, e0 ) = P(e, e0 )h(e0 )/h(e) on {e : h(e) > 0} is a transition matrix and is often called Doob h-transform. The SSD, in case when time reversed chain is stochastically monotone, is then given by authors in three steps (Theorem 5.5 in Diaconis and Fill [4]): ← − i) Calculate time reversal P of P. ← − ← − ii) Calculate the Siegmund dual ( P )Z of P . ← − ← − iii) Calculate the Doob H-transform P∗ = (( P )Z )H of ( P )Z , where H = πC with π = (π(e1 ), . . . , π(eM )). In above procedure the Siegmund dual must be calculated w.r.t. total ordering, and then we have C(e, e0 ) = 1(e ≤ e0 ). In Lorek and Szekli [15] we give a recipe for SSD on the same state space E∗ = E for chains, whose time-reversal is Möbius monotone with respect to some partial ordering expressed by matrix C(e, e0 ) = 1(e e0 ). In matrix form, the transitions are given by: ← − P∗ = diag(H)−1 (C−1 P C)T diag(H), P where H = πC (i.e., H = (H(e1 ), . . . , H(eM )) with H(e) = e0 :e0 e π(e)) and diag(H) denotes diagonal matrix with vector H on the diagonal. Thus, ← − P∗ is a Doob H-transform of (C−1 P C)T what is exactly the Siegmund dual ← − of P from Theorem 4.1. In other words, SSD given in Lorek and Szekli [15] results from exactly the same steps i)–iii) as in Theorem 5.5 in Diaconis and Fill [4]. Remark 5.1. For non-linear ordering Möbius monotonicity and stochastic monotonicity are, in general, different. In particular, we can have a chain which is not stochastically monotone, but for which we can construct both, 8 241 Siegmund dual and Strong Stationary Dual, since it can be Möbius monotone. According to our knowledge Falin [5] was the first who observed that these are two different notions of monotonicity (however, connections to daulities were not given there). In a subsequent paper Lorek and Markowski [14] we give more details on relations between Möbius, realizable, weak and usual (strong) stochastic monotonicities in chains on partially ordered state spaces. 242 6. Applications 235 236 237 238 239 240 Let us recall the relation between the stationary distribution of X and absorption probabilities of its Siegmund dual Z (i.e., (2.3)) π({ej }↓ ) = P (Z is absorbed in eM |Z0 = ej ). 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 This relation can have three potential applications: A1) Suppose we are given Markov chain Z with two absorbing states (winning and losing). To calculate the probability of being absorbed in one of them we can find an ergodic chain X for which Z is its Siegmund dual. Then the task reduces to calculating the stationary distribution of X. A2) Suppose we are given ergodic Markov chain X, the task is to find its stationary distribution. Then we can calculate its Siegmund dual Z and if we are able to find absorption probabilities, then we have the stationary distribution of X A3) Similarly to A2): We are given ergodic chain X. Assume that the approach given in b) is infeasible, we can find Siegmund dual but we are unable to calculate its absorption probabilities. However, we can have so-called stable simulation scheme for estimating the stationary distribution. Say we want to estimate π(e). Then we calculate Siegmund dual Z of X and we simulate it several times always starting the chain at e and simply estimate the winning/losing probability. Note that (in contrast to most Monte Carlo Markov Chains methods) we do not need any knowledge about, e.g. the rate of convergence or time to absorption, we simply run the chain Z until it is absorbed. In this Section we provide three examples for A1, A2 and A3 types of applications. In the first two we exploit some knowledge from literature on the stationary distribution (A1) or absorption probabilities (A2) to solve the problem, whereas for application A3 we present all the calculations needed for stable simulation scheme for some nonstandard tandem network of two stations. 6.1. Application A1: Some nonsymmetric game: explicit formula for probabilities of winning and losing. Consider the following game. The states of the game are (0, W ON, LOST ). The state 0 means we did not yet win nor lose, the names of other states speak for themselves. We 9 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 start at state 0. At each step we can win with probability p > 0, lose with probability q > 0 or nothing can happen with probability 1 − (p + q). Of course, eventually we will end up in W ON or LOST state with probabilities p/(p + q) and q/(p + q) respectively. Now consider the following generalization. We are given d games G1 , . . . , Gd Pd and parameters pi , qi , i = 1, . . . , d such that i=1 (pi + qi ) ≤ 1. We win the whole game if we win all the games G1 , . . . , Gd and we lose the whole game if we lose at least one game Gi , i ∈ {1, . . . , d}. If we win game Gi we will not play it anymore. The generic state is either eL := LOST or e = (e(1), . . . , e(d)), e(i) ∈ {0, 1}, i = 1, . . . , d, where e(i) = 1 means that we have already won the game Gi . Denote also eW := (1, . . . , 1). At each step we can win new game Gi with probability pi (provided we have not won it already) or lose it with probability qi (in which case we automatically lose P the whole game), or we can do nothing with the remaining probability 1 − i:e(i)=0 (pi + qi ). The described dynamics is a Markov chain, say Z on state space E∗ = {0, 1}d ∪ {eL } with the following transitions: pi if e0 = e + si , X qi if e0 = eL , e 6= eL , i:e(i)=0 X PZ (e, e0 ) = 1 − (pi + qi ) if e0 = e 6= eL , i:e(i)=0 1 if e0 = e = e , L 290 291 292 293 294 295 296 297 298 where si = (0, . . . , 0, 1, 0, . . . , 0) with 1 at the i-th coordinate. The chain has two absorbing states: eL and eW (we lose or we win). Eventually, the chain will be absorbed in one of them. A natural question arises: What is the probability of winning the whole game starting with arbitrary set of already won games e0 ∈ E? In other words, we want to calculate the following probabilities: ρ(e0 ) = P (τeM < τe0 | Z0 = e0 ), where τe := inf{n ≥ 0 : Zn = e}. To answer the question consider first another chain, the nonsymmetric random walk X on E = {0, 1}d , defined for the same parameters pi , qi , i = 1, . . . , d, with transitions: if e0 = e + si , qi p if e0 = e − si , i 0 PX (e, e ) = X X 1− qi − pi if e0 = e. i:e(i)=0 i:e(i)=1 Under mild assumption that at least for one state e we have that PX (e, e) > 0, the chain is ergodic with the stationary distribution: Y Y pi qi . π(e) = p i + qi pi + qi i:e(i)=0 i:e(i)=1 10 299 300 301 302 P Let |e| = di=1 e(i) (called a level of e). We consider coordinate-wise ordering, i.e. e e0 if e(i) ≤ e0 (i) for all i = 1, . . . , d. The state e1 = (0, . . . , 0) is a unique minimal and eM = (1, . . . , 1) is a unique maximal one. Then, (E, ) is a Boolean lattice with the following Möbius function: 0 µ(e, e0 ) = (−1)|e |−|e| 1(e e0 ). The chain was considered in Lorek and Szekli [15], where we calculated its Strong Stationary Dual w.r.t to above partial order. The chain is Möbius P monotone if and only if di=1 (pi + qi ) ≤ 1. ”In between the lines“ we also calculated its Siegmund dual (the reader is referred to Lorek and Szekli [15] for calculations). It turns out, that PZ is the Siegmund dual of PX and thus, via (2.3), we have X Y Y qi pi ρ(e0 ) = . p i + qi p i + qi 0 ee i:e(i)=1 303 i:e(i)=0 For example, if pi = p and qi = q for all i = 1, . . . , d, then we have X 1 ρ(e0 ) = q |e| pd−|e| (p + q)d ee0 d−|e0 | |e0 | 0 X 1 |e | k d−k p = q p = . k p+q (p + q)d k=0 304 305 306 307 308 309 310 311 312 313 314 315 316 Note that then of course the probability ρ(e0 ) depends only on level |e0 |. In 0 particular, for p = q we have that ρ(e0 ) = 2d−|e | . Remark 6.1. Matrix PZ can be written as upper-triangular and thus P we can read the eigenvalues from the diagonal. These are equal λA = 1 − i∈A (pi + qi ) for all A ⊆ {1, . . . , d}. Because of relation (4.3), these are also the eigenvalues of PX . 6.2. Application A2: Finding stationary distribution of some extension of birth and death chain. Consider classical birth and death chain on state space E = {1, 2, . . . , N } with constant birth rate q > 0 and death rate p > 0. Assume p 6= q and p + q < 1. Let X be the following modification of this birth and death chain: in addition, there is extra probability 1 − p − q of going from any state to maximal state N . Formally, the transitions of X are given by (6.1) q if j = i + 1, i = 1, . . . , N − 2, 1−p if (i = N − 1, j = N ) or (i = j = N ), p if (j = i − 1, i = 2, . . . , N ) or (i = j = 1), PX (i, j) = 1 − p − q if i = 1, . . . , N − 2, j = N, 0 otherwise. 11 317 318 319 Theorem 6.2. Consider Markov chain X on E = {1, . . . , N } with transitions given in (6.1). Assume that p 6= q and p + q < 1. Then, the stationary distribution of the chain is given by b k2 c−1 b k−1 c 2 X X 1 k k − 1 γj − γj 2p 2j + 1 2j + 1 j=0 j=0 (6.2) π(k) = 2(2p)N −k+2 , b N 2−1 c X N γj 2j + 1 j=0 320 where γ = (1 − 4pq). 321 Proof. First, we will calculate Siegmund dual of X. Consider linear ordering :=≤ on E. The transitions of Siegmund dual are then calculated using in (2.2), i.e., PZ (j, i) = PX (i, {j}↓ ) − PX (i + 1, {j}↓ ). Considering all the cases we have 322 323 324 PZ (j, j − 1) = PX (j − 1, {j}↓ ) − PX (j, {j}↓ ) = q + p − p = q, j = 2, . . . , N, PZ (j, j + 1) = PX (j + 1, {j}↓ ) − PX (j + 2, {j}↓ ) = p − 0 = p, j = 1, . . . , N − 1, PZ (N − 1, N ) = PX (N, {(N − 1)}↓ ) = p, = PX (N, {N }↓ ) = 1. P For every j = 1, . . . , N − 1 we have that i PZ (j, i) < 1, more precisely PZ (N, N ) 325 N X i=1 N X PZ (1, i) = p < 1, P(j, i) = p + q < 1, j = 2, . . . , N − 1. i=1 326 327 Thus we add one extra state, call it 0 obtaining transition matrix of Siegmund dual Z on E∗ = {0, 1, . . . , N }: p if j = i + 1, i = 1, . . . , N − 1, q if j = i − 1, i = 2, . . . , N − 1, 1−p if i = 2, j = 1, (6.3) PZ (i, j) = 1 − p − q if i = 2, . . . , N − 1, j = 0, 1 if (i = j = 0) or (i = j = N ), 0 otherwise. 12 These are exactly the transitions corresponding to Gambler’s Ruin with Catastrophes considered in Hunter [11]. The ruin probability is given therein in eq. (2.6), the winning probability reads (with ρ(0) = 0) k k √ √ 1 + 1 − 4pq 1 − 1 − 4pq − 2p 2p ρ(k) = N N . √ √ 1 + 1 − 4pq 1 − 1 − 4pq − 2p 2p For linear ordering the relation (2.3) reads π({k}↓ ) = ρ(k), thus π(k) = ρ(k) − ρ(k − 1). Above relation and some elementary calculations using binomial expansion k−1 (1 + √ x)k − (1 − √ b 2 c √ X k k x) = 2 x xj 2j + 1 j=0 328 yield (6.2) and thus complete the proof. 329 6.3. Application A3: Stable simulation scheme for some nonstandard tandem network of two servers. In this Section we will present some two-node closed tandem network with the unknown stationary distribution (which is surely not of product form). Studying the absorption probability in the resulting Siegmund dual chain is a rather complicated task. We will present stable simulation scheme and will estimate the stationary distribution for some concrete parameters. It is also worth noting that the structure of the resulting Siegmund dual is interesting, e.g., once it hits some barriers, it will not leave them, see details below. Consider a tandem network of two stations, each with finite buffer of size N . Denote the chain by X ≡ (X)n≥0 with state space E = {0, . . . , N }2 , where state (x, y) denotes that there are x customers at server 1 and y customers at server 2. When x > 0 and y < N the system operates as the following classical tandem: with probability λ1 there is arrival to station 1 (if x < N ); with probability λ2 there is arrival to station 2 (if y < N ); with probability µ1 customers traverse from server 1 to server 2; with probability µ2 the customer from server 2 leaves the network; with the remaining probability nothing happens. Without loss of generality we assume that λ1 + µ1 + λ2 + µ2 = 1. When x = 0 then the probability of arrival to station 2 is λ2 + µ1 (if y < N ) and when y = N there is a departure from server 1 with probability µ1 (if x > 0). This can be seen as a modification of standard Gordon–Newell network. Similar modification (of Jackson network, i.e., with countable state space) where considered in Foley and McDonald [7], where the rough asymptotics (and large deviations) were derived, showing that the modification (only on borders) changes the stationary distribution dramatically. 13 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 Transitions of the chain under consideration are as follows: 0 0 (6.4) PX ((x, y), (x , y )) = 356 357 λ1 if x0 = x + 1, y 0 = y, λ2 if y 0 = y + 1, x0 = x > 0, λ2 + µ1 if y 0 = y + 1, x0 = x = 0, µ2 if y 0 = y − 1, x0 = x, µ1 if x0 = x − 1, y 0 = y + 1, µ1 if x0 = x − 1, y 0 = y = N, µ2 if x0 = x < N, y 0 = y = 0, µ2 + λ1 if x0 = x = N, y 0 = y = 0, µ1 + λ2 if x0 = x = 0, y 0 = y = N, λ1 if x = x0 = N, 0 < y 0 = y < N, λ2 if 0 < x0 = x < N, y 0 = y = N, λ1 + λ2 if x0 = x = y 0 = y = N. Consider coordinate-wise ordering, i.e., (x, y) (x0 , y 0 ) ⇐⇒ x ≤ x0 and y ≤ y 0 . Then, the Möbius function is following: (6.5) 1 if −1 if µ((x, y), (x0 , y 0 )) = 0 otherwise. 358 359 (x0 = x and y 0 = y) or (x0 = x + 1 and y 0 = y + 1), (x0 = x + 1 and y 0 = y) or (x0 = x and y 0 = y + 1), We will calculate transition matrix PZ of the Siegmund dual Z directly from Theorem 4.1: PZ ((x1 , y1 ), (x2 , y2 )) = X µ((x2 , y2 ), (x, y))PY ((x, y), (x1 , y1 )↓ ), (x,y)(x2 ,y2 ) 360 where PY ((x, y), (x1 , y1 )↓ ) := 361 function µ we have X PY ((x, y), (x0 , y 0 )). With our Möbius (x0 ,y 0 )(x1 ,y1 ) PZ ((x1 , y1 ), (x2 , y2 )) 14 362 (6.6) = PY ((x2 , y2 ), (x1 , y1 )↓ ) − PY ((x2 + 1, y2 ), (x1 , y1 )↓ ) −PY ((x2 , y2 + 1), (x1 , y1 )↓ ) + PY ((x2 + 1, y2 + 1), (x1 , y1 )↓ ), 363 364 365 366 where we should understand that for example, for y2 = N the corresponding terms, in this case PY ((x2 , y2 +1), (x1 , y1 )↓ ) and PY ((x2 +1, y2 +1), (x1 , y1 )↓ ) are equal to 0. Let us start with the case: x1 , x2 , y1 , y2 > 1 and x1 , x2 , y1 , y2 < N . Then, using (6.6) we have for example: PZ ((x1 , y1 ), (x1 − 1, y1 )) 367 = PY ((x1 − 1, y1 ), (x1 , y1 )↓ ) − PY ((x1 , y1 ), (x1 , y1 )↓ ) −PY ((x1 − 1, y1 + 1), (x1 , y1 )↓ ) + PY ((x1 , y1 + 1), (x1 , y1 )↓ ) = λ1 + µ2 − µ2 − µ2 + µ2 = λ1 . 368 369 We have also PZ ((x1 , y1 ), (x1 − k, y1 )) = 0 for k = 1, . . . , x1 − 1. Considering some possible transitions to border of the state space: PZ ((x1 , y1 ), (0, y1 − 1)) 370 = PX ((0, y1 − 1), (x1 , y1 )↓ ) − PX ((1, y1 − 1), (x1 , y1 )↓ ) −PX ((0, y1 ), (x1 , y1 )↓ ) + PX ((1, y1 ), (x1 , y1 )↓ ) = 1 − 1 − (λ1 + µ2 ) + (µ2 + λ1 ) = 0. 371 372 373 374 We will skip the rest of (lengthy) calculations, but considering case-bycase, X turns out to be Möbius monotone. We add one extra (coffin) state, let us name it (+∞, +∞). The transitions of the Siegmund dual are as follows: 15 375 376 PZ ((x, y), (x0 , y 0 )) = (6.7) where η(x, y) = (x2 ,y2 377 378 379 380 381 382 383 384 385 386 387 λ1 if N − 1 > x0 = x − 1, y 0 = y, λ2 if N − 1 > y 0 = y − 1, x0 = x < N, λ2 + µ1 if N − 1 > y 0 = y − 1, x0 = x = N, µ1 if x0 = x + 1, N − 1 > y 0 = y − 1, µ2 if x0 = x, y 0 = y + 1, λ1 + λ2 + µ1 if x = 0, y = 0, x0 = +∞, y 0 = +∞, , λ2 + µ1 if x > 0, y = 0, x0 = +∞, y 0 = +∞, λ1 if x = 0, y > 0, x0 = +∞, y 0 = +∞, 1 if x0 = x = y 0 = y = N or x0 = x = y 0 = y = +∞, 1 − η(x, y) X )∈E∗ :(x if x0 = x, y 0 = y. PX ((x, y), (x2 , y2 )). Note that outside 2 ,y2 )6=(x,y) the borders the transitions look like some reversed network (however, note that this is not the usual time reversal). The behavior on the borders is different. First, the chain can go to (+∞, +∞) only from states of form (0, y) or (x, 0). Second, once the process is on ”upper“ ((x, N )) or ”right“ ((N, y)) border, it behaves like a usual, absorbing birth and death chain with probabilities of being killed only possible on (0, N ) and (N, 0) respectively. On ”upper“ border, birth rate is µ1 and death rate is λ1 , whereas on ”right“ border the birth rate is µ2 and death rate is µ1 + λ2 . The similar behaviour (i.e., not leaving the borders) we also noticed in Strong Stationary Dual for two-node network representing two independent servers, see Lorek and Szekli [16] for details. Stable simulation scheme. To estimate π({(x, y)}↓ ) we simply estimate winning probability in Siegmund dual Z by running the chain K times independently, each time starting from (x, y). Denote the output of i-th run as Yi ∈ {0, 1}, where 0 means that the chain was absorbed in +∞ and 1 means that the chain was absorbed in (N, N ). Thus, we can estimate the winning probability by K 1 X Yi . ρ̂(x, y) = K i=1 By relation relation (2.3) we thus estimated stationary distribution: π̂({(x, y)}↓ ) = ρ̂(x, y). If we wish to estimate π(x, y), then we can use Remark 3.2 to get 16 (for x > 0, y > 0) π̂(x, y) = = X µ((x0 , y 0 ), (x, y))π̂({(x0 , y 0 }↓ ) (x0 ,y 0 )(x,y) π̂({(x, y)}↓ ) − π̂({(x − 1, y)}↓ ) − π̂({(x, y − 1)}↓ ) + π̂({(x − 1, y − 1)}↓ ), 388 389 390 391 392 393 394 thus we need separately to estimate π({(x, y)}↓ ), π({(x − 1, y)}↓ ), π({(x, y − 1)}↓ ) and π({(x − 1, y − 1)}↓ ). Numerical results. We estimated the stationary distribution of the tandem for N = 5 (thus the state space is E = {0, . . . , 5}2 with |E| = 36) with 3 1 6 parameters λ1 = 16 , λ2 = 16 , µ1 = µ2 = 16 . The simulations were performed in The Julia Language. For each (x, y) ∈ E we performed independently K = 1 mln iterations. Results are presented in Fig. 1. y\x 0 1 2 3 4 5 0 0.029739 0.028980 0.022189 0.012979 0.008949 0.006902 1 0.048597 0.034663 0.021324 0.013645 0.006901 0.002193 2 0.067564 0.04128 0.020593 0.010742 0.005404 0.003982 3 0.087365 0.046606 0.024072 0.009349 0.005434 0.001226 4 0.11586 0.051627 0.020465 0.010822 0.003319 0.000404 5 0.159256 0.050266 0.018471 0.006062 0.001679 0.001091 Figure 1. Estimated π̂(x, y), x, y ∈ {0, . . . , N }, N = 5, λ1 = 1 6 3 16 , λ2 = 16 , µ1 = µ2 = 16 398 Performing the simulations using 2 mln iterations for each (x, y) ∈ E yields in results agreeing with above estimation on the first three digits after comma. The L2 distance between π̂ estimated using 1 mln iterations and π̂ estimated using 2 mln iterations was less then 2 · 10−5 . 399 Acknowledgments 400 403 Author thanks Bartłomiej Błaszczyszyn and Zbigniew Palmowski for helpful discussions. Partial support by the project RARE-318984, a Marie Curie IRSES Fellowship within the 7th European Community Framework Programme, is kindly acknowledged. 404 References 395 396 397 401 402 405 406 407 408 409 410 411 [1] B. Błaszczyszyn and K. Sigman. Risk and duality in multidimensions. Stochastic processes and their applications, 83:331–356, 1999. [2] J. T. Cox and U. Rösler. A duality relation for entrance and exit laws for Markov processes. Stochastic processes and their applications, 16:141—-156, 1984. [3] H. Dette, J. A. Fill, J. Pitman, and W. J. Studden. Wall and Siegmund Duality Relations for Birth and Death Chains with Reflecting Barrier. Journal of Theoretical Probability, 10(2):349–374, Apr. 1997. 17 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 [4] P. Diaconis and J. A. Fill. Strong stationary times via a new form of duality. The Annals of Probability, 18(4):1483–1522, 1990. [5] G. Falin. Monotonicity of random walks in partially ordered sets. Russian Mathematical Surveys, 167, 1988. [6] J. A. Fill and V. Lyzinski. Strong Stationary Duality for diffusion processes. arXiv, pages 1–33, 2014. [7] R. D. Foley and D. R. McDonald. Large deviations of a modified Jackson network: Stability and rough asymptotics. The Annals of Applied Probability, 15(1B):519–541, Feb. 2005. [8] J. M. Harrison and R. J. Williams. Multidimensional Reflected Brownian Motions Having Exponential Stationary Distributions. The Annals of Probability, 15(1):115– 137, Jan. 1987. [9] T. Huillet. Siegmund duality with applications to the neutral Moran model conditioned on never being absorbed. Journal of Physics A: Mathematical and Theoretical, 43(37), Sept. 2010. [10] T. Huillet and S. Martínez. On Möbius Duality and Coarse-Graining. Journal of Theoretical Probability, 29(1):143—-179, July 2016. [11] B. Hunter. Gambler’ s Ruin with Catastrophes and Windfalls. Journal of Statistical Theory and Practice, 2(2):199—-219, 2008. [12] T. M. Liggett. Interacting Particle Systems. 2004. [13] D. V. Lindley. The theory of queues with a single server. Mathematical Proceedings of the Cambridge Philosophical Society, 48(02):277–289, Oct. 1952. [14] P. Lorek and P. Markowski. Monotonicity requirements for exact sampling with Markov chains. Technical report, 2016. [15] P. Lorek and R. Szekli. Strong stationary duality for Möbius monotone Markov chains. Queueing Systems, 71(1-2):79–95, Mar. 2012. [16] P. Lorek and R. Szekli. Strong Stationary Duality for Möbius monotone Markov chains: examples. Probability and Mathematical Statistics, 36(1):75—-97, 2016. [17] L. Miclo. Strong stationary times for one-dimensional diffusions. arXiv preprint arXiv:1311.6442, pages 1–40, 2013. [18] G.-C. Rota. On the foundations of combinatorial theory I. Theory of Möbius functions. Probability theory and related fields, 368:340–368, 1964. [19] D. Siegmund. The equivalence of absorbing and reflecting barrier problems for stochastically monotone Markov processes. The Annals of Probability, 1976. 18
© Copyright 2026 Paperzz