Journal of Forecasting J. Forecast. 25, 129–152 (2006) Published online in Wiley InterScience (www.interscience.wiley.com) DOI: 10.1002/for.978 Autoregressive Gamma Processes CHRISTIAN GOURIEROUX1 AND JOANN JASIAK2* 1 2 CREST, CEPREMAP and University of Toronto, Canada York University, Canada ABSTRACT We introduce a class of autoregressive gamma processes with conditional distributions from the family of noncentred gamma (up to a scale factor). The paper provides the stationarity and ergodicity conditions for ARG processes of any autoregressive order p, including long memory, and closed-form expressions of conditional moments. The nonlinear state space representation of an ARG process is used to derive the filtering, smoothing and forecasting algorithms. The paper also presents estimation and inference methods, illustrated by an application to interquote durations data on an infrequently traded stock listed on the Toronto Stock Exchange (TSX). Copyright © 2006 John Wiley & Sons, Ltd. key words intertrade durations; autoregressive gamma; CIR; high frequency INTRODUCTION The liquidity of a financial asset on a given day and market depends on how frequently that asset is traded, or equivalently on the times between consecutive transactions, called intertrade durations (Gourieroux et al., 1999). The times to trade recorded sequentially in time form time series which possess interesting dynamic properties, such as time-varying moments and serial correlation. The duration analysis of financial assets is a relatively new topic of research, motivated by the interest in assessing future asset liquidity, in the sense of time it can take to purchase or sell a given quantity of shares. While some stocks are traded instantaneously, others require a considerable amount of time to trade, which varies randomly throughout a trading day. In the literature, there exist few models designed for intertrade durations. The most popular among these is the autoregressive conditional duration (ACD) model, introduced by Engle and Russell (1998) [see also Bauwens, Giot (1997)]. The conditional hazard function of the ACD suggests classifying this model as an accelerated hazard model with lagged durations as explanatory variables. While the ACD is quite easy to estimate, it relies on a strong and unrealistic assumption of path-independent overdispersion. This means that the conditional standard deviation exceeds the conditional mean by a fixed amount, unrelated to the history of the process. Another limitation of the ACD is the difficulty in deriving the stationarity and ergodicity conditions for models with temporal dependence beyond lag one. An alter* Correspondence to: Joan Jasiak, York University. Department of Economics, 4700 Keele Street, Toronto, ON M3J 193. Canada. E-mail: [email protected] Copyright © 2006 John Wiley & Sons, Ltd. 130 C. Gourieroux and J. Jasiak native model, called the stochastic volatility duration (SVD), was developed by Ghysels et al. (2003). The SVD model improves upon the ACD in that it allows for independent variation of the conditional mean and variance and yields a time-varying over- or underdispersion. The SVD is a nonlinear factor model, with two dynamic heterogeneity factors that determine the mean and dispersion, respectively. Therefore, the estimation is technically difficult, and requires advanced statistical tools. The aim of this paper is to introduce a class of models, which accommodate various nonlinearities in inter-durations dynamics and have easily verifiable stationarity conditions and straightforward estimation procedures. Our approach borrows from the early literature on dynamic processes for gamma distributed variables developed for applications to ruin processes, such as the gamma autoregressive (GAR) [see, e.g., Gaver and Lewis (1980); the survey by Grundwald et al. (2000);1 the model of Sim (1990, 1994); the work by Griffiths (1970) and Kotz and Adams (1964)]. The new model, called the autoregressive gamma model (ARG), is designed specifically for intertrade duration analysis and forecasting, but can also be used in application to asset volatility and traded volumes, for example. In contrast to the ACD, and similar to the SVD, an ARG can fit any data that feature time-varying over- or underdispersion and can accommodate complex nonlinear dynamics. Finally, an ARG can represent processes with long memory more parsimoniously than ACD and SVD. The paper is organized as follows. In the next section, we introduce the autoregressive gamma process of order one (ARG(1)), with a conditional distribution from the family of noncentred gamma (up to a scale factor), and the noncentrality parameter written as a linear function of lagged durations. This section provides the forecast function and discusses the stationarity. The temporal dependence is examined in the third section by investigating the conditional moments, the first- and second-order autocorrelograms and nonlinear canonical correlations. The ARG model is viewed as a nonlinear state space model, with an integer-valued latent state variable in the fourth section. The ARG-based state space representation allows us to provide the forecasting, filtering and smoothing algorithms. The relationship between the discrete time ARG process and the continuous time Cox–Ingersoll–Ross process is covered in the fifth section. The sixth section shows the ARG models of autoregressive order higher than one (ARG(p)) and long memory ARG models. In the seventh section, the ARG is compared to other autoregressive processes for gamma distributed variables existing in the literature. Statistical inference is discussed in the eighth section. The application to intertrade durations is presented in the ninth section. Contrary to empirical literature on intertrade durations, which is focused on very frequently traded assets and time deformation, we consider an infrequently traded stock for which the risk of low liquidity, i.e. of a long waiting time to trade, is significant. A final section concludes the paper. Proofs are gathered in the appendices. THE AUTOREGRESSIVE GAMMA PROCESS The autoregressive gamma processes are Markov processes with conditional distributions from the family of noncentred gamma (up to a scale factor) and with path-dependent noncentrality coefficients. Let us recall that a variable Y follows the distribution g (d, b, c), if and only if Y/c follows the noncentred gamma distribution g (d, b ) characterized by three nonnegative parameters: the degree of freedom d, noncentrality parameter b and scale parameter c. The noncentred gamma distribution 1 The use of GAR processes for the analysis of financial data is limited due to the assumption of conditional homoscedasticity (constant conditional variance), which underlies the model. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 131 arises as a Poisson mixture of gamma distributions: we say that Y follows the distribution g (d, b) if, and only if, there exists a Poisson variable Z ~ P(b), such that the conditional distribution of Y given Z is g (d + Z ). The density of the g (d, b, c) distribution is: y • È yd + k -1 exp- bb k ˘ fd ,b ,c ( y ) = 1y > 0 expÊ - ˆ Â Í d + k Ë c ¯ k = 0 Î c G (d + k ) k! ˚˙ (1) and the two first moments of a noncentred gamma distributed variable are: EY = cd + cb , VY = c 2d + 2c 2 b (2) ARG(1) Definition 1 The process (Yt) is an autoregressive gamma process of order one (ARG(1)) if, and only if, the conditional distribution of Yt given Yt-1 is g (d, btYt-1, ct), where bt, ct are deterministic functions of time. Definition 1 is wide in the sense that it comprises heterogeneous Markov ARG(1) processes and allows for variation in time of all parameters, except for the degree of freedom. In particular, the scale ct and the parameter bt appearing in the noncentrality coefficient are time-varying, although path-independent. The history of the process determines the entire noncentrality coefficient, btYt-1, which is written as a linear function of the lagged value of the process. Let us recall that all parameters are assumed nonnegative and, for convenience, define a new parameter rt = btct, which will later be interpreted as an autoregressive coefficient. From the mixture of gamma interpretations given at the beginning of this section, it follows that the dynamics of the ARG process can be written as: Zt Yt =  Wi ,t + e t i =1 where et, Zt, Wi,t are independent conditional on Yt-1 and distributed as g (n, ct), P(btyt-1), g (1, ct), respectively (see Sim, 1990). This dynamic representation of an ARG is very convenient for simulations of the process trajectories. A closed-form expression of the conditional distribution of the current value of an ARG(1) process Yt given Yt-h is available for any h. Proposition 1 The conditional distribution of Yt given Yt-h is noncentred gamma up to a scale factor: g (d, bt|t-hYt-h, ct|t-h), where t r t t - h = b t t - h ct t - h = ’t =t - h +1 r t t ct t - h =  (ct r t +1 . . . rt ) t = t - h +1 Proof. see Appendix A. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 132 C. Gourieroux and J. Jasiak Stationarity condition for ARG(1) To derive the stationarity condition, let us consider a homogeneous Markov ARG(1) process, with time-invariant parameters bt = b, ct = c and rt = bc = r. By Proposition 1, the parameters of the conditional distribution of Yt given Yt-h become: rt t -h = r h ct t - h = c(1 + r + . . . + r h -1 ) = c 1- rh 1- r with the limiting values: lim r t t - h = 0, lim ct t - h = hÆ• hÆ• c , if r < 1 1- r It follows that: Proposition 2 An ARG(1) process is stationary, when the conditional distribution of Yt given Yt-1 is g (d, btYt-1, c) with r = bc < 1. (ii) The conditional distribution of Yt given Yt-h is: (i) h 1- rh ˘ È r (1 - r ) Yt - h , c g˜ Íd , , "h h 1 - r ˙˚ Î c(1 - r ) (iii) The marginal invariant distribution is gamma, up to a scale factor: c ˆ g˜ Ê d , 0, Ë 1- r¯ The limiting case of an autoregressive coefficient equal to one will be discussed later. SERIAL DEPENDENCE There exist various methods for examining serial dependence in stationary ARG processes. In this section, we consider the first- and second-order conditional moments, autocorrelograms and nonlinear canonical decomposition. Conditional moments From Proposition 2 it follows that: Proposition 3 The first two conditional moments of a stationary ARG(1) process are: E(Yt Yt -1 ) = cd + rYt -1 V (Yt Yt -1 ) = c 2d + 2 rcYt -1 Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 133 and more generally: E(Yt Yt -1 ) = V (Yt Yt - h ) = c(1 - r h ) d + r h Yt - h 1- r c 2 (1 - r h ) (1 - r ) 2 2 d + 2r h c(1 - r h ) Yt - h 1- r We see that the stationary ARG(1) process is a weak AR(1) with conditional heteroscedasticity specified as a linear function of the lagged value.2 The expressions of conditional moments indicate that an ARG(1) can be under- or overdispersed. The conditional overdispersion exists if and only if: V (Yt Yt -1 ) > E(Yt Yt -1 ) 2 ¤ 0 > c 2d (d - 1) + 2 rcYt -1 (d - 1) + r 2 Yt 2-1 The discriminant equation of the polynomial on the right-hand side of the last expression is: D ¢ = - r 2 c 2 (d - 1) It is negative for d > 1, and positive for d < 1. When d > 1, we have V(Yt|Yt-1) < E(Yt|Yt-1)2, "Yt-1. 2 ) ( When d < 1, the product of the roots c d d - 1 is negative, while the two roots are real. Proposi2 r tion 4 below, summarizes the results on dispersion of an ARG(1). Proposition 4 When d > 1, the stationary ARG(1) process features conditional and marginal underdispersion. When d < 1, the stationary ARG process features marginal overdispersion. In parallel, the process may feature either conditional under- or overdispersion, depending on the value of Yt-1. First- and second-order autocorrelograms From the weak AR(1) interpretation of an ARG process, it follows that: r (h) = Corr(Yt , Yt - h ) = r h (3) The second-order autocorrelogram represents the serial dependence in squared values of the process. It is proven in Appendix B that: r ( 2 ) (h) = Corr(Yt 2 , Yt -2 h ) = r h r h + 2(d + 1) 2d + 3 (4) Nonlinear canonical decomposition The nonlinear canonical decomposition (see Barret, Lampard (1955), Lancaster (1958, 1963), Wong (1962), Eagleson (1964)) allows for the analysis of nonlinear time dependence in Markov processes 2 However, the process doesn’t admit a strong AR(1) representation with i.i.d. innovations as in the general class of models considered by Grunwald et al. (2000). Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 134 C. Gourieroux and J. Jasiak to which belong the ARG. The formula of canonical decomposition is derived in Appendix C using the relationship between the ARG and the continuous-time Cox–Ingersoll–Ross process [see also Sim (1990, p. 329) for an approach based on the joint Laplace transform]. Let us denote by: d d -1 È (1 - r)yt ˘ È1 - r ˘ yt f (yt ; d , c, r) = 1yt >0 expÍ ˙˚ ÍÎ c ˙˚ G (d ) Î c (5) the marginal p.d.f. of Yt, and by: f1 ( yt yt -1 ; d , c, r ) = 1yt > 0 exp( - yt c) k • È ytd + k -1 ( ryt -1 c) ˘ X Â Í d +k exp( - ryt -1 c) ˙ k! G (d + k ) k =0 Î c ˚ (6) the conditional p.d.f. of Yt given Yt-1. Proposition 5 The conditional p.d.f. of Yt given Yt-1 admits the following nonlinear canonical decomposition: • ¸ Ï f1 (yt yt -1 ; d , c, r) = f (yt ; d , c, r)Ì1 +  r ny n (yt , d , c, r)y n (yt -1 , d , c, r)˝ ˛ Ó n =1 where: G (d )G (n + 1) ˆ y n ( y; d , c, r ) = Ê Ë G (d + n) ¯ 12 L(nd -1) Ê Ë (1 - r)y ˆ c ¯ and Ln is a generalized Laguerre polynomial: n k L(nd -1) (z) =  (-1) k =0 G (d + n) zk G (d + k )G (n - k + 1) k! The polynomials yn, n varying, satisfy the conditions: Ey n (Yt ) = 0, Cov[y n (Yt ), y m (Yt )] = 0, "n π m Vy n (Yt ) = 1 Proposition 5 leads directly to the following corollaries. Corollary 1 The conditional p.d.f. of Yt given Yt-h admits the nonlinear canonical decomposition: • ¸ Ï fh (yt yt -h ; d , c, r) = f (yt ; d , c, r)Ì1 +  r hny n (yt ; d , c, r)y n (yt -h ; d , c, r)˝ ˛ Ó n =1 Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 135 The canonical decomposition allows for easy forecasting of any polynomial transformation of Yt. Corollary 2 We get E[yn(Yt)|Yt-1] = rnyn(Yt-1), "n. Thus, there exist several nonlinear transformations of the ARG(1) process that follow a weak AR(1) process. For instance, y 1 (Yt ) = (1 - r )Yt ˘ 1 È d12 Í ˙˚ c d Î is AR(1) with autoregressive coefficient r, and 12 2 (1 - r)Yt (1 - r) Yt 2 ˆ È 2 ˘ Ê d (d + 1) ( ) + + 1 y 2 (Yt ) = Í d Á ˜ Î d (d + 1) ˚˙ Ë 2 c 2c 2 ¯ is AR(1) with autoregressive coefficient r2. Corollary 3 The joint distribution of (Yt, Yt-1) is symmetric, which means that the process is timereversible. THE STATE SPACE REPRESENTATION The dynamics of (Yt) can be written in the form of a nonlinear state space representation, with the mixing variable (Zt) as the state variable and the transition equation: Zt Yt -1 ~ P [ bYt -1 ] The measurement equation is: Yt Zt ~ g [d + Zt , c] The latent variable Zt has an interesting interpretation in finance. In some applications, it may be interesting to rank the stocks traded on a given market with respect to their risk of liquidity, that is, to distinguish stocks which trade quickly from those with long waiting times to trade. For example, if the liquidity risk is measured in terms of intertrade durations (which are the waiting times to trade), the integer-valued Zt can provide a liquidity rating of a stock. Then, by imitating the Moody’s system of risk rating, we can assign the highest rating to the stock with the lowest liquidity risk, such as Zt = 0 = AAA, a lower one to the stock with Zt = 1 = AA, etc. The ARG model can yield the current rating Zt given the lagged observed liquidity Yt-1, and provide a forecast of future liquidity Yt+h given the current rating Zt. In the next subsection, we focus on the dynamics of the latent mixing variable (Zt). The first part concerns the smoothing procedure and the conditioning variable is the observable process (Yt). The second part shows the forecasts of Zt based on its own past values. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 136 C. Gourieroux and J. Jasiak Smoothing The smoothing algorithm allows us to recover the path of the latent process (Zt). Proposition 6 (i) The values Z1, . . . , ZT of the state variable are independent conditional on Y0, Y1, . . . , YT. (ii) The conditional distribution of Zt given Y0, Y1, . . . , YT depends on the observable variables Yt-1, Yt only. (iii) This conditional p.d.f. is given by: zt 1 È byt yt -1 ˘ G (d + zt )G ( zt + 1) ÍÎ c ˙˚ l( zt yt -1 , yt ) = • z Ï 1 Ê byt yt -1 ˆ ¸ Ì ˝  ) ( )Ë c ¯ ˛ z = 0 Ó G (d + z G z + 1 Proof. See Appendix D. Marginal distribution of the state variable Proposition 7 The latent process (Zt) is a Markov process such that the transition density: z l( zt zt -1 ) = (bc) t d + zt + zt -1 (bc + 1) G (d + zt + zt -1 ) G ( zt + 1)G (d + zt -1 ) is the density of a negative binomial distribution with parameters: n = d + zt -1 , p= 1 1 + bc Proof. See Appendix E. The transition density can be used for examining the probability of an asset migrating from one liquidity rating to another, by analogy to the migration of bond issuers between various classes of risk rating in finance. CONTINUOUS TIME LIMIT OF ARG(1) The continuous time counterpart of the ARG(1) process is the Cox–Ingersoll–Ross (CIR) model, commonly used in theoretical finance to represent the volatility of financial assets. In order to prove that the CIR model is a limit of ARG when the time increments are allowed to shrink to zero, we need to consider the infinitesimal drift and volatility functions. This means that we need to study the behaviour of a prediction of Yt at horizon h, when h tends to zero. We get: Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 137 1 lim hÆ 0 [ E(Yt Yt - h = y) - y] h 1 È c(1 - r h ) ˘ d + r h y - y˙ = lim hÆ 0 Í h Î 1- r ˚ = lim hÆ 0 1 - r h È cd - y˘˙ Í h Î1 - r ˚ cd = - log r Ê - yˆ Ë1- r ¯ lim hÆ 0 1 V (Yt Yt - h = y) h 2 Ï 1 (1 - r h ) r h c(1 - r h ) ¸ = lim hÆ 0 Ì c 2 d +2 y˝ 2 h 1- r (1 - r ) Ó2 ˛ 2c log r =y 1- r This leads us to the conclusion that the limiting continuous time process associated with the ARG(1) process is the CIR process (Cox et al., 1985): dYt = a(b - Yt )dt + s Yt dWt (7) where: a = - log r > 0, b = cd -2 log r , s2 = c 1- r 1- r A more general result can be established (see, e.g., Lamberton and Lapeyre, 1992) as well. Proposition 8 The stationary ARG process is a discretized version of the CIR process. The link between the ARG and CIR processes points to the importance of the ARG model for applications to intertrade durations. Indeed, it is well-known that asset volatility and liquidity are strongly related. Therefore, it comes as no surprise that the basic model for asset volatility is related to the model for intertrade durations. ARG(p) The ARG can be extended to an autoregressive order strictly greater than one. We will see that the forecast function and ergodicity conditions for an ARG(p) can easily be obtained. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 138 C. Gourieroux and J. Jasiak Definition of ARG( p) Definition 2 The process (Yt) is an autoregressive gamma process of order p (ARG(p)), if, and only if, the conditional distribution of Yt given Yt-1 = (Yt-1, Yt-2, . . . , Yt-p)¢ is g (d, b¢Yt-1, c), where b¢Yt-1 = b1Yt-1 + . . . + bpYt-p, and bj ≥ 0, j = 1, . . . , p. The main difference between the ARG(1) and ARG( p) is the form of the noncentrality coefficient, which becomes a linear function of past values up to and including lag p. We can rewrite the ARG( p) easily in the form of a p-dimensional process Yt = (Yt, . . . , Yt-p+1)¢, which is a Markov process of order one, and has the conditional Laplace transform given by: E[exp( -u ¢Yt ) Yt -1 ] = E[exp( -u1Yt . . . - u p Yt - p+1 ) Yt -1 ] = exp[ -u2 Yt -1 . . . - u p Yt - p+1 ]E[exp( -u1Yt ) Yt -1 ] - cu1 -d = [1 + cu1 ] exp( -u2 Yt -1 . . . - u p Yt - p+1 ) expÈÍ b ¢Yt -1 ˘˙ ˚ Î1 + cu1 (from Lemma 1 in Appendix A) [ = [1 + cu1 ] exp - A(u)¢ Yt -1 -d (8) ] where: cu1 cu1 cu1 È cu1 ˘ A(u) = Í b1 + u2 , b 2 + u3 ; . . . , b p-1 + u p , bp Î1 + cu1 1 + cu1 1 + cu1 1 + cu1 ˙˚ (9) Stationarity condition The stationarity of ARG( p) is determined by the function A(u) (Darolles et al., 2002). Proposition 9 The ARG(p) process is stationary if limhÆ0 Aoh(u) = 0, for any u with nonnegative components, where Aoh denotes the function A compounded h times with itself. As the corollary to Proposition 9 we get: Corollary 3 The stationarity condition is: c(b1 + b2 + . . . + bp) < 1. Proof. See Appendix F. Long memory The ARG(1) can feature long memory, under conditions similar to the standard AR(1), that is, when the autoregressive coefficient r is equal to one, and when r is stochastic. (i) Autoregressive coefficient r = 1 Let us fix the scale parameter at c = 1. When r = b = 1, the conditional Laplace transform of the process becomes (see Appendix A, Lemma 1): Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 139 E[exp( -uYt ) Yt -1 ] = exp[ - a(u)Yt -1 + b(u)] where a(u) = u/(1 + u), b(u) = -d log(1 + u). The conditional Laplace transform at horizon h is: E[exp( -uYt + h -1 ) Yt -1 ] = exp[ - a oh (u)Yt -1 + bh (u)] (say) where the compounded function aoh satisfies: a oh (u) = u 1 + hu We see that limh-• aoh(u) = 0, which implies that the process is weakly ergodic (Darolles et al., 2002). Moreover, the conditional expectations E[exp(-uYt+h-1)|Yt-1] tend to their limiting value at a hyperbolic rate in h. This suggests that the process has an autocorrelation function with a hyperbolic rate of decay, that it is features long memory. At the same time, Yt is a stationary Markov process. The long memory property is of relevance for applications to intertrade durations of frequently traded stocks, in which a long range of persistence is generally observed. The long-range persistence can be captured parsimoniously by an ARG(1), while it requires the presence of a fairly large number of lags in the competing ACD or SVD models (Jasiak, 1998). (ii) Stochastic autoregressive coefficient Let us consider an ARG(1) process parameterized by d, r and c. When r is fixed, the bivariate density of yt, yt-h can be written as (see Corollary 1): fh ( yt , yt - h ) = f [ yt ; d , c (1 - r )] f [ yt - h ; d , c (1 - r )] • ¸ Ï ¥ Ì1 +  r hn Yn [ yt ; d , c (1 - r )]Yn ( yt - h ; d , c (1 - r )]˝ ˛ Ó h =1 where g = c/(1 - r). In this expression, parameters d and g appear in the marginal distribution and canonical variates, while parameter r determines the canonical correlations and temporal dependence. The first- and second-order marginal moments of the model E(Yt d , r, g ) = m(d , g ), V (Yt d , r, g ) = s 2 (d , g ) depend on d and g only. The autocovariance function is given by: Cov(Yt , Yt - h d , r, g ) = r hs 2 (d , g ) and is shown to include r in its expression. Let us now assume that the autoregressive parameter r is stochastic and takes values from the interval [0, 1]. Also, let us suppose that r has the probability density p, and that d and g remain fixed. By integrating the stochastic parameter r out of the first- and second-order moments, we obtain: Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 140 C. Gourieroux and J. Jasiak E(Yt d , p , g ) = m(d , g ), V (Yt d , p , g ) = s 2 (d , g ) and the autocovariance function Cov(Yt , Yt - h ) = E[ Cov(Yt , Yt - h d , r, g ) d , p , g )] + Cov[ E(Yt d , r, g ), E(Yt - h d , r, g ) d , p , g )] = E[ Cov(Yt , Yt - h d , r, g ) d , p , g ] = E[ r hs 2 (d , g ) d , p , g )] = Ep ( r h )s 2 (d , g ) The autocorrelation function is no longer a power function of r, but instead an expected value of r power h: Corr(Yt , Yt - h d , p , g ) = Ep ( r h ) Consequently, the autocorrelation function features hyperbolic decay when the distribution p of coefficient r assigns sufficiently large probabilities to values close to one. For example, this is the case when the distribution of r is beta (Granger and Joyeux, 1980). ARG AND GAR: COMPARISON In the time series literature, there exist a variety of dynamic models for gamma processes. [See e.g. Jacobs, Lewis (1978), Lawrance, Lewis (1981), Smith, Miller (1986)]. So far, none of them has been used in an application to intertrade durations. As mentioned earlier in the text, the dynamic properties of financial duration data are quite particular, and the available dynamic gamma models don’t account for features such as conditional heteroscedasticity, for example. In this section, we discuss the gamma autoregressive process introduced by Gaver and Lewis (1980). Gaver and Lewis considered the following linear autoregression: Yt = rYt -1 + e t where the innovations are i.i.d. random variables. They proved that for 0 £ r < 1, there exists a distribution of the error term et such that the marginal distribution of Yt is exponential (or more generally a gamma). When the marginal distribution of Yt is exponential g(1, l), the innovation distribution is a mixture of a point mass at zero with probability r and an exponential distribution g (1, l) with probability 1 - r. There are three main differences between the GAR and the ARG processes: (i) By construction, the GAR is conditionally homoscedastic and doesn’t allow for independent variation of the conditional mean and variance of Yt. In application to intertrade durations, the assumption of constant duration volatility is very strong and unrealistic. (ii) The conditional distribution of GAR admits a discrete component. This means that the GAR process can become deterministic Yt = rYt-1 with probability r. (iii) The functional dependence between serial correlation and the distribution of innovations is hard to interpret. For example, in the unit root case r = 1, the GAR process (Yt) takes value zero, at all t, and remains stationary. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 141 Table I. Autoregressive models for gamma processes a(u) = ru Gaver and Lewis (1980): GAR Gourieroux and Jasiak (2003): ARG Sim (1990, 1994) a(u) = ru/(1 + cu) Let us point out that the ARG and GAR processes are special cases of Car (compound autoregressive) processes. Indeed, an autoregressive process with marginal gamma distribution is easy to construct by considering a conditional Laplace transform of the type: E[exp( -uYt ) Yt -1 ] = exp[ - a(u)Yt -1 + c(u) - c[ a(u)]] where c is the log-Laplace transform of the marginal gamma distribution. The models differ essentially by the form of function a, which determines serial dependence, as shown in Table I. STATISTICAL INFERENCE Let us consider a sample of discrete time observations y1, . . . , yT , of size T. It will be shown that standard estimation methods such as GMM and the maximum likelihood can be used for estimation of ARG processes. Moment estimators The parameters of the ARG( p) process can be estimated from the first- and second-order conditional moments, since: E(Yt Yt -1 ) = cd + cb ¢Yt -1 V (Yt Yt -1 ) = c 2d + 2c 2 b ¢Yt -1 Consistent estimators of cd and cbj, j = 1, . . . , p, are obtained by regressing Yt on 1, Yt-1, . . . , Yt-p. The efficiency of these estimators can be improved by taking into account conditional hereroscedasticity. For instance, we can apply a pseudo-maximum likelihood method based on a Gaussian pseudo-family. The PML estimators are the solutions of: (cˆ, dˆ , bˆ )¢ = arg max c ,d ,b 2 Ï 1 1 ( yt - cd - cb ¢ yt -1 ) ¸ 2 2 Ì- log(c d + 2c b ¢ yt -1 ) ˝  2 c 2d + 2c 2 b ¢ yt -1 ˛ t = p +1 Ó 2 T Maximum likelihood estimators The log-likelihood function: T LT (d , c, b ) =  log f1 ( yt yt -1 ; d , c, b ) (10) t =1 is of a complicated form, and needs to be approximated in practice. The approximations of the loglikelihood function for ARG(1) given below are based on three different methods. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 142 (i) C. Gourieroux and J. Jasiak Truncating the series expansion of conditional density: k T T ÈK (ryt -1 c) ˘ 1 ytd + k -1 L(T1) (d , c, r) = -  (yt + ryt -1 ) +  log Í Â d + k ˙ k! t =1 c t =1 Î k =0 c G (d + k ) ˚ (11) (ii) Truncating the nonlinear canonical decomposition: T T N È ˘ L(T2 ) (d , c, r ) =  log f ( yt ; d , c, r ) +  log Í1 +  r ny n ( yt ; d , c, r )y n ( yt -1 ; d , c, r )˙ Î n=0 ˚ t =1 t =1 m Intuitively this is a method of moments involving all marginal and cross moments E(Y tmY t-1 ), for m, n £ N. A drawback of this approach is the possibility of obtaining negative values of the log-likelihood function due to truncation. (iii) Using the simulated maximum likelihood method based on artificial drawings of the latent Poisson state variable: T L(T3) (d , c, r ) = -  t =1 s T S yt ytd + zt -1 È ˘ +  log Í d + z s s ˙ t c t =1 ) ( c G d + z Î s =1 t ˚ where S is the number of drawings and (z1t, . . . , zSt) are independently drawn in the Poisson distribution P(ryt-1/c) = P(byt-1). The advantage of the last approach is that it can easily be extended to any autoregressive order p higher than one. This can be done by performing the drawings of the Poisson state variable in the Poisson distribution P(b¢ yt-1). APPLICATION TO FINANCIAL DURATION SERIES In empirical literature on high-frequency data, the analysis of intertrade durations typically concerns very frequently traded stocks, which feature long memory. However, as the liquidity risk is much higher in infrequently traded (illiquid) stocks, accurate forecasts of future times to trade may become valuable inputs to any trading strategy. The aim of this section is to examine the dynamics of an infrequently traded stock with long waiting times to the next upcoming quote or transaction. Therefore, the number of observations available in our sample is considerably lower than in other applications with a sampling period of similar length. The ARG model is estimated from data on interquote durations of the Dayton Mining stock traded on the Toronto Stock Exchange in October 1998. The choice of a stock from the mining industry is not arbitrary. In the past, a false announcement of a gold mine discovery has undermined the credibility of the whole mining sector, and diminished the frequency at which those stocks were traded. Prior to estimation, durations between quotes with the same time stamp were aggregated.3 The times between market openings and closures were deleted from the sample. We also removed from 3 In particular, we disregard simultaneous trades at either market opening or during the day, when a large buy (sell) order is filled by several sell (buy) orders. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 143 0.0004 0.0 0.0002 frequency 0.0006 Autoregressive Gamma Processes 0 5000 10000 15000 20000 durations Figure 1. Dayton Mining, durations the sample all observations on quotes recorded during market preopenings. Thus, we are mainly concerned with the intraday dynamics of interquote durations. Dayton Mining belongs to infrequently traded stocks, and is not frequently quoted either. The sample consists of 381 observations, with mean 1162.5 sec. [approx 20 min]4 and variance 4,762,851 [standard error of about 35 min]. The kernel smoothed marginal density of the Dayton interquote durations is displayed in Figure 1. It features a typical form of a gamma density function with a small degree of freedom. To choose the best fit, several ARG models with different numbers of lags were estimated by OLS in the first step. The Dayton Mining series displays a rather short range of serial dependence in interquote durations, typical for infrequently traded stocks. In this, it differs from very frequently traded assets which feature long memory. Since the coefficients on lags greater than two were found insignificant, we retained for further analysis the ARG(1) and ARG(2) models. Both models were then estimated by QMLE. Table II summarizes the results obtained from fitting the ARG(1) and ARG(2) models to the sample of Dayton Mining using the regression approach and QMLE. We observe that the estimated coefficients r1 and r2 are positive, less than one, and sum up to less than one in the ARG(2) model. This means that both models are stationary. Also, both models fit the data fairly well, and all their coefficients are significant. While the mean log-likelihood of ARG(2), equal to (-8.154), shows little improvement compared to that of ARG(1), equal to (-8.175), the R-squared doubles when the second lag is added (0.030 compared to 0.018).5 In the next step, the parameters of the ARG(2) are used in a simulation experiment. It consists in fixing the vector of initial durations at two consecutive values observed at the end of the sample: 4 5 To compare with an average intertrade duration less than 1 min for a frequently traded stock, such as the IBM. The value of R2 is small, since it measures linear serial dependence in a nonlinear dynamic framework. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 144 C. Gourieroux and J. Jasiak Table II. Estimation of ARG(1) and ARG(2) Parameter ARG(1) OLS 10.024E2 (1.25E2) 0.132 (0.050) — — cd r1 r2 QMLE 10.527E2 (0.26E2) 0.0794 (0.057) — — ARG(2) OLS 8.925E2 (1.35E2) 0.1196 (0.051) 0.1104 (0.051) QMLE 9.396E2 (0.24E2) 0.0201 (1.262) 0.1612 (0.393) Table III. One-step predictions from ARG(2) Mean Variance y(T + 1) 2 quotes 3 quotes 4 quotes 1,793.93 8,929,084.7 1,985.58 10,497,642 1,815.43 12,056,015 1,940.20 13,455,554 y(T) = 5226 and y(T - 1) = 7512, and simulating the next four out-of-sample observations y(T + 1), y(T + 2), y(T + 3), y(T + 4). We replicate this experiment 1000 times. Next we compute the average predicted waiting times for 1, 2, 3 and 4 quotes, given by y(T + 1), [y(T + 1) + y(T + 2)]/2, [y(T + 1) + y(T + 2) + y(T + 3)]/3 and [y(T + 1) + y(T + 2) + y(T + 3) + y(T + 4)]/4, consecutively. These forecasts provide a natural measure of future liquidity. More precisely, if we assume that the quotes result in trades of the same volume V, say, we can claim that our forecasts represent the average time necessary to trade the basic quantity V of assets when the volumes V, 2V, 3V, 4V , etc. are to be traded (Gourieroux et al., 1999). By comparing the distributions of forecasted durations, we see how the liquidity cost measured by time necessary to trade a given volume without causing a change in the price of asset, depends on the traded quantity. The means and variances of forecasted durations are reported in Table III. CONCLUSIONS This paper introduced the autoregressive gamma processes ARG for high-frequency financial data. The ARG can represent the dynamics of various positively valued time series such as intertrade (interquote) durations or series of squared returns, which approximate asset volatility in the same way intertrade durations approximate asset liquidity. Alternatively, an ARG model can fit a series of volumes per trade, which is an alternative proxy for liquidity. The ARG is sufficiently flexible to accommodate short as well as long memory. The stationarity and ergodicity conditions and forecast functions are available for ARG processes of any autoregressive order. The estimation relies on a variety of methods, ranging from a simple regression to more sophisticated simulation-based estimators. Finally note that the ARG specification for positively valued univariate processes can easily be extended to a multivariate framework. This approach leads to the Wishart autoregressive (WAR) processes for sequences of symmetric, positive definite stochastic matrices (Gourieroux, 2004; Gourieroux et al., 2003). Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 145 APPENDIX A: CONDITIONAL DISTRIBUTION AT ANY HORIZON Let us prove the result for h = 2 by using the conditional Laplace transforms. The same argument holds for any horizon H. Lemma 1 b t ct u ¸ -d E[exp( -uYt ) Yt -1 ] = (1 + ct u) expÏÌ-Yt -1 ˝ 1 + ct u ˛ Ó Proof. We have: E[exp( -uYt ) Yt -1 ] = E[[ E exp( -uYt ) Zt ] Yt -1 ) [ = E (1 + ct u) - (d + Zt ) ] Yt -1 , where Zt Yt -1 ~ P [ b t Yt -1 ] 1 ˆ¸ -d = (1 + ct u) expÏÌ- b t Yt -1 Ê1 ˝ Ë 1 + ct u ¯ ˛ Ó b t ct u ¸ -d = (1 + ct u) expÏÌ-Yt -1 ˝ 1 + ct u ˛ Ó QED Let us now consider the conditional Laplace transform at horizon h = 2. We get: E[exp(-uYt ) Yt -2 ] = E[ E[exp(-uYt ) Yt -1 ] Yt -2 ] b t ct u ¸ -d È ˘ = (1 + ct u) E ÍexpÏÌ-Yt -1 ˝ Yt -2 ˙ Î ˚ 1 + ct u ˛ Ó b t ct u ˆ = (1 + ct u) Ê1 + ct -1 Ë 1 + ct u ¯ -d -d b t ct u ¸ Ï b t -1ct -1 Ô 1 + ct u Ô expÌ-Yt -2 ˝ b t ct u Ô Ô 1 + ct -1 Ó 1 + ct u ˛ b t ct b t -1ct -1u ¸ -d = [1 + (ct + ct -1 b t ct )u] expÏÌ-Yt -2 ˝ 1 + (ct + ct -1 b t ct )u ˛ Ó b t t -2 ct t -2 u ¸ -d = (1 + ct t -2 u) expÏÌ-Yt -2 ˝ 1 + ct t -2 u ˛ Ó APPENDIX B: SECOND-ORDER AUTOCORRELOGRAM We have: g ( 2 ) (h) = Cov(Yt 2 , Yt 2-h ) 2 = E(Yt 2 Yt -2 h ) - (EYt 2 ) 2 = E[ E(Yt 2 Yt -h )Yt 2-h ] - (EYt 2 ) Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 146 C. Gourieroux and J. Jasiak [ ] ) + E[(c 2 2 = E[V (Yt Yt -h )Yt 2-h ] + E E(Yt Yt -h ) Yt 2-h - (EYt 2 ) ( 2 = E (ct2t -hd + rt t -h ct t -h Yt -h ) Yt 2-h 2 ] 2 d + rt t -h Yt -h ) Yt 2-h - (EYt 2 ) t t -h = ct2t -hd (1 + d )m2 + 2 rt t -h ct t -h (1 + d )m3 + rt2t -h m4 - m22 where: m j = E(Yt j ) = cj (1 - r ) j d (d + 1) . . . (d + j - 1) It follows that: 2 g ( 2 ) (h ) = c 2 (1 - r h ) 2 c2 2 d (d + 1) 2 2 (1 - r ) (1 - r ) + 2r hc + r 2h = c c3 1- rh (1 + d )d (d + 1)(d + 2) 3 1- r (1 - r ) c4 (1 - r ) 4 (1 - r ) 4 d (d + 1)(d + 2)(d + 3) 4 [ c4 (1 - r ) 4 d 2 (d + 1) 2 2 d (d + 1) (1 - r h ) d (d + 1) + 2 r h (1 - r h )(d + 1)(d + 2) + r 2 h (d + 2)(d + 3) - d (d + 1)] = c4 (1 - r ) 4 d (d + 1)[ r 2 h [d (d + 1) - 2(d + 1)(d + 2) + (d + 2)(d + 3)] + r h ( -2d (d + 1) + 2(d + 1)(d + 2)]] = = c4 (1 - r ) 4 d (d + 1)[ 2 r 2 h + 4(d + 1)r h ] 4 2 r hd (d + 1)[ r h + 2(d + 1)] c4 (1 - r ) = g ( 2 ) (0 )r h r h + 2(d + 1) 2d + 3 APPENDIX C: THE NONLINEAR CANONICAL ANALYSIS Since an ARG process is a discretized version of the CIR process (see Proposition 9), we can derive the nonlinear canonical decomposition of the conditional distribution of Yt given Yt-1 from the spectral analysis of the infinitesimal generator associated with the CIR process. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 147 (1) Spectral decomposition of the infinitesimal generator The infinitesimal generator A associated with the diffusion equation: dYt = a(b - Yt )dt + s Yt dWt is given by: y Æ Ay ( y) = a(b - y) dy 1 d 2y ( y) + s 2 y 2 ( y) dy 2 dy Let us verify whether the eigenfunctions of the infinitesimal generator are polynomials. (i) Let us first consider a polynomial of degree n, yn say, satisfying the condition: a( b - y ) 1 fd 2y n dy n ( y) = l ny n ( y) ( y) + s 2 y 2 dy dy 2 where ln is the corresponding eigenvalue. By identifying the coefficients of yn, we see that: ln = -na = n log r. This implies the necessary form of the eigenvalue. (ii) Next we need to solve the differential equations: a (b - y ) dy n 1 d 2y n (y) + s 2 y ( y ) + nay n ( y ) = 0 dy 2 dy 2 or: (b - y) dy n 1 s 2 d 2y n ( y) + (y) + ny n (y) = 0 y dy 2 a dy 2 Let us introduce the change of variable: z= 2a 2a y, y n ( y) = y *n Ê 2 yˆ = y *n ( z ) 2 Ë s s ¯ (C.1) where: 2a/s 2 = (1 - r)/c. The differential equation becomes: 2 2a dy n* 1 s 2 (2 a) d 2y n* ( ) ( ) (z) + ny n (z) = 0 + b y z y dz 2 a (s 2 )2 dz 2 s2 2 Ê 2 ab - zˆ dy n* (z) + z d y n* (z) + ny (z) = 0 n Ë s2 ¯ dz dz 2 (C.2) 2 ab = d . It is known that the solution of differential equation (C.2) is proportional to s2 the generalized Laguerre polynomial (Abramowitz and Stegun, 1970, formula 22.6.15): where: Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 148 C. Gourieroux and J. Jasiak n L(nd -1) ( z ) =  ( -1) k =0 k G (d + n) zk G (d + k )G (n - k + 1) k! (C.3) Upon integrating with respect to a g (d) distribution, the polynomials satisfy (Abramowitz and Stegun, 1970, formula 22.2.12): Ú • 0 2 [ L(nd -1) (z)] 1 G (d + n) exp(- z)z d -1dz = G (d ) G (d )G (n + 1) We derive the standardized polynomials: 12 G (d )G (n + 1) ˘ (d -1) ( ) L*n d -1 ( z ) = ÈÍ Ln ( z ) Î G (d + n) ˙˚ (C.4) Finally the eigenfunctions of the initial differential equation are derived by applying the change of variable (C.1): 12 G (d )G (n + 1) ˘ (d -1) Ê (1 - r ) y ˆ y n ( y) = ÈÍ Ln Ë c ¯ Î G (d + n) ˚˙ (C.5) Therefore yn, n = 0, 1, . . . forms a basis of eigenfunctions of the infinitesimal generator associated with the eigenvalues ln= log rn. (iv) It is known that for a univariate diffusion equation the functions yn, n varying, are also eigenfunctions of the conditional expectation operator: y Æ E(y(Yt)|Yt-1 = y), associated with the eigenvalues: exp ln= rn. Moreover, the process is time-reversible. The conditional distribution of Yt given Yt-1, f1(yt|yt-1; d, r, c) (say) can be written as (see Lancaster, 1963): • ¸ Ï f1 ( yt yt -1 ; d , r, c) = f ( yt ; d , r, c)Ì1 +  r ny n ( yt )y n ( yt -1 )˝ ˛ Ó n =1 (C.6) c ˆ Ê . where f(yt; d, r, c) is the marginal distribution of Yt, i.e. the distribution g˜ Ë d , 0, 1- r¯ APPENDIX D: PROOF OF PROPOSITION 7 Let us denote by g(yt|zt) the conditional p.d.f. of Yt given Zt and by h(zt|yt-1) the conditional p.d.f. of Zt given Yt-1. Then the conditional distribution of Y1, . . . , YT, Z1, . . . , ZT given Y0 is: t l( y1 , . . . , yT , z1 , . . . , zT yo ) =  [ g( yt zt )h( zt yt -1 )] t =1 We find that: Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 149 l(z1 , . . . , zT y0 , y1 , . . . , yT ) l(y1 , . . . , yT , z1 , . . . , zT yo ) = l(y1 , . . . , yT yo ) t ’ [g(y z )h(z y = ’  [g(y z)h(z y t t =1 t t =1 t t t -1 )] • t t -1 )] z =0 È ˘ Í g(y z )h(z y ) ˙ t t t t t -1 ˙ = C l(zt yt , yt -1 ) (say) = ’t =1 Í • t =1 Í ˙ ÍÎ Â g(yt z)h(z yt -1 ) ˙˚ z =0 t Therefore we note that: (i) Z1, . . . , ZT are independent, conditional on Yo, Y1, . . . , YT. (ii) The conditional distribution: l( zt yo , y1 , . . . , yT ) = l( zt yt , yt -1 ) depends on yt-1, yt only. g(y z )h(zt yt -1 ) (iii) l(zt yt , yt -1 ) = • t t  {g(yt z)h(z yt -1 )} z =0 z d + zt -1 (byt -1 ) t Ê - yt ˆ y ( ) b exp exp y t 1 Ë c ¯ G (d + zt ) G (zt + 1) cd + zt = z • d + z -1 Ï 1 (byt -1 ) ¸ Ê - yt ˆ yt ( ) y exp b exp Ì d +z ˝ t -1 Â Ë c ¯ G (d + z) G (z + 1) ˛ z =0 Ó c 1 zt 1 È byt yt -1 ˘ G (d + zt )G (zt + 1) ÍÎ c ˙˚ = z • Ï 1 Ê byt yt -1 ˆ ¸ Ì Â Ë c ¯ ˝˛ z =0 Ó G (d + z )G (z + 1) APPENDIX E: TRANSITION DENSITY OF THE STATE VARIABLE We get: l( zt zt -1 ) = Ú l( zt , yt -1 zt -1 )dyt -1 = Ú h( zt yt -1 )g( yt -1 zt -1 )dyt -1 Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 150 C. Gourieroux and J. Jasiak = Ú exp( - byt -1 ) b zt ytz-t 1 ytd-+1zt -1 -1 yt -1 ˆ expÊ dyt -1 d + zt -1 Ë c ¯ G ( zt + 1) c G (d + zt -1 ) 1 1 b zt expÈÍ - yt -1 Ê b + ˆ ˘˙ ytd-+1zt -1 -1dyt -1 Ë c¯˚ G ( zt + 1)G (d + zt -1 ) cd + zt -1 Ú Î zt G (d + zt + zt -1 ) b 1 = d + zt -1 d + zt + zt -1 G ( zt + 1)G (d + zt -1 ) c Ê b + 1ˆ Ë ¯ c = z = (bc) t d + zt + zt -1 (bc + 1) G (d + zt + zt -1 ) G ( zt + 1)G (d + zt -1 ) APPENDIX F: STATIONARITY CONDITION We use Proposition 9 to find the conditions, which ensure that the solution of the p-dimensional recursive system: X1,t = cX1,t -1 b1 + X2 ,t -1 1 + cX1,t -1 M X p-1,t = X p ,t = cX1,t -1 b p-1 + X p ,t -1 1 + cX1,t -1 cX1,t -1 bp 1 + cX1,t -1 tends to (0, . . . 0)¢, when t tends to infinity, for any nonnegative initial value (X1,0, . . . , Xp,0)¢. The system is equivalent to: X1,t = X p ,t b1 + X2 ,t -1 bp M X p-1,t = X p ,t X p ,t = b p - b p-1 + X p ,t -1 bp bp 1 + cX1,t -1 It follows that Xj,t, j = 1, . . . , p take nonnegative values, and that Xp,t is always less than bp. Moreover, the sequence (Xp,t) satisfies the nonlinear recursive equation: X p ,t = b p - bp b p-1 b1 + X p ,t - p ˘˙ 1 + c ÈÍ X p ,t -1 + . . . + X p ,t - p+1 bp bp Î ˚ Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for Autoregressive Gamma Processes 151 A possible limiting value l of this sequence satisfies: l = bp - bp b p-1 ˘ È b1 + 1˙ 1 + cl Í + . . . + bp Îbp ˚ 1 È ˘ Therefore, the admissible values are l = 0 and l = b p Í1 . Î c(b1 + . . . + b p ) ˙˚ If c(b1 + . . . + bp) < 1, the sequence (Xp,t) takes values in the compact set [0, bp], with a unique admissible limiting value l = 0. We infer its convergence to zero. 1 È ˘ by selectIf c(b1 + . . . + bp) > 1, we can get the convergence of (Xp,t) to b p Í1 Î c(b1 + . . . + b p ) ˚˙ ing the initial value appropriately. ACKNOWLEDGEMENTS The authors gratefully acknowledge financial support of the Natural Sciences and Engineering Research Council of Canada. REFERENCES Abramowitz M, Stegun I. 1970. Handbook of Mathematical Functions, 9th edn. John Wiley: Dover. Barret J, Lampard D. 1955. An expansion for some second order probability distributions and its applications to noise problems. IRE Trans Information Theory IT-1: 10–15. Bauwens L, Giot P. 1997. The logarithmic ACD model: an application to the bid–ask quote process of the NYSE stocks. Annales d’Economie et de Statistique 60: 117–149. Cherian K. 1941. A bivariate gamma type distribution function. Journal of the Indian Mathematical Society 5: 133–144. Cox J, Ingersoll J, Ross S. 1985. A theory of the term structure of interest rates. Econometrica 53: 385–408. Darolles S, Gourieroux C, Jasiak J. 2002. Structural Laplace transform and compound autoregressive models. CREST DP. Eagleson G. 1964. Polynomial expansions of bivariate distributions. Annals of Mathematical Statistics 35: 1208–1215. Engle R, Russell J. 1998. The autoregressive conditional duration model: a new model for irregularly spaced data. Econometrica 66: 1127–1162. Gaver D, Lewis P. 1980. First order autoregressive gamma sequences and point processes. Advances Applied Probability 12: 727–745. Ghysels E, Gourieroux C, Jasiak J. 2003. Stochastic volatility duration models. Journal of Econometrics 119: 413– 433. Gourieroux C. 2004. The Wishart process for stochastic risk. Econometric Reviews, forthcoming. Gourieroux C, Jasiak J, Le Fol G. 1999. Intraday market activity. Journal of Financial Markets 2: 193–226. Gourieroux C, Jasiak J, Sufana R. 2003. The Wishart autoregressive process of multivariate stochastic volatility. Working Paper, University of Toronto. Granger C, Joyeux R. 1980. An introduction to long memory time series models and fractional differencing. Journal of Time Series Analysis 1: 15–29. Griffiths R. 1970. Infinitely divisible multivariate gamma distributions. Sankhya A32: 393–404. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for 152 C. Gourieroux and J. Jasiak Grunwald G, Hyndman R, Tedesco L, Tweedie R. 2000. Non-Gaussian conditional linear AR(1) models. Australian and New Zealand Journal of Statistics 42: 479– 495. Jacobs R, Lewis P. 1978. Discrete time series generated by mixtures I: correlational and runs properties. JRSS B 40: 94–105. Jasiak J. 1998. Persistence in intertrade durations. Finance 19: 166–195. Kotz S, Adams J. 1964. Distribution of a sum of identically distributed exponentially correlated gamma variables. Annals of Mathematical Statistics 35: 277–283. Lamberton D, Lapeyre B. 1992. Introduction au calcul stochastique appliqué à la Finance. In Mathematique et Applications. Ellipses-Editions: Paris. Lancaster H. 1958. The structure of bivariate distribution. Annals of Mathematical Statistics 29: 719–736. Lancaster H. 1963. Correlations and canonical forms of bivariate distributions. Annals of Mathematical Statistics 34: 532–538. Lawrance A, Lewis P. 1981. A new autoregressive time series model in exponential variables. Advances in Applied Probability 13: 826–845. Lawrance A, Lewis P. 1985. Modelling and residual analysis of nonlinear autoregressive time series in exponential variables. JRSS B 47: 165–202. Lloyd E, Saleem S. 1979. A note on seasonal Markov chains with gamma or gamma-like distributions. Journal of Applied Probability 16: 117–128. Lunde A. 1997. A conjugate gamma model for duration in transaction data. DP University of Aarhus. Phatarfod R. 1971. Some approximate results in renewal and dam theories. Journal of Australian Mathematical Society 12: 425–432. Sim C. 1990. First order autoregressive models for gamma and exponential processes. Journal of Applied Probability 27: 325–332. Sim C. 1994. Modelling Nonnormal first order autoregressive time series. Journal of Forecasting 13: 369–381. Smith S, Miller J. 1986. A non-Gaussian state space model and application to prediction of records. Journal of the Royal Statistical Society, Senes B 48: 79–88. Tong H. 1990. Nonlinear Time Series: A Dynamic System Approach. Oxford University Press: Oxford. Wong E. 1964. The construction of a class of stationary markov processes. In Sixteenth Symposium on Applied Mathematics, American Mathematical Society, pp. 264–276. Wong E, Thomas J. 1962. On polynomial expansions of second order distributions. SIAM Journal of Applied Mathematics 10: 507–516. Authors’ biographies: Joann Jasiak is Associate Professor of Economics at York University, Toronto, Associate Fellow at Centre Interuniversitaire de Recherche en Economie Quantitative (CIREQ) and Member of Centre de Recherche en E-Finance (CREF). Her research interest include time series econometrics, financial econometrics, nonlinear models and high-frequency data. Christian Gourieroux is Professor of Economics at University of Toronto, Canada, and ENSAE (Ecole Nationale de Statistique et Economie) in Paris, France. He is director of the Finance and Insurance section of CREST, member of CEPREMAP, France, and CREF, Canada, and Associate Fellow at CIREQ. His research interest include time series econometrics, financial econometrics, nonlinear models and high-frequency data. Authors’ addresses: Christian Gourieroux, University of Toronto, Toronto, Canada. Joan Jasiak, Department of Economics, York University, 4700, Keele Street, Toronto, ON, M3J 1P3, Canada. Copyright © 2006 John Wiley & Sons, Ltd. J. Forecast. 25, 129–152 (2006) DOI: 10.1002/for
© Copyright 2026 Paperzz