July 10, 1984 Autoregressive Conditionally ~eteroscedastic Models Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC 27695-8203 Key Words and Phrases: Autoregression; heteroscedasticity; maximum likelihood and generalized least squares estimators; asymptotic normality. Abstract Two linear regression models, where the independent variables are (4) fixed and bounded, and (b) the lagged values of the dependent variable, with autoregressive conditionally heteroscedastic (ARCH) errors are considered. A series representation and some ergodic properties of the first order ARCH errors are derived. The strong consistency and the asymptotic normality of the maximum likelihood estimators are established. Asymptotic distributions of the least squares estimator and an estimated generalized least squares estimator are also derived. Autoregressive Conditionally Heteroscedastic Models Sastry G. Pantula North Carolina State University 1. Introduction: Consider a process {Y } satisfying the equation t (1.1) Yt • X a + e ~t~ t where X : 1 x p is a known row vector and ~: p ~t parameters. X 1 is a vector of unknown Traditional econometric models assume that e t is a sequence of 2 uncorrelated (0,0 ) random variables and the conditional mean of Y give~ the t past is !t~' where !t may contain lagged values of Yt • function, the best forecast of Y , based on the past t Under a symmetric loss in~ormation, is the con- ditional mean of Y given the past and is denoted by E[YtIFt_lJ, where t F - • information up to time (t-l). t l Therefore for the model (1.1), the one step forecast error is The unconditional variance of the one period forecast is given by 0 2 . Note that and For the conventional econometric models, however, the conditional variance does not depend on F t- 1. For some processes one might expect better forecast intervals if additional information from the past were allowed to affect the forecast variance. A model which allows the conditional variance to depend on the past -2- realization is the bilinear model described by Granger and Anderson (1978). Jones (1965), Granger and Anderson (1978), and Priestly (1978) considered nonlinear time series models. Engle (1982) proposed a class of models, called autoregressive conditionally heteroscedastic (ARCH) models, for which both the conditional mean and the . conditional variance of a time series are functions of the past behavior of the time series. If the conditional density is normal, then a general expression for the ARCH model is where g and h are measurable. A special case we consider assumes that the mean can be expressed as a linear combination of variables in the information set, while the variance is a qth order weighted average of the squares of past di.sturbances. More precisely, where and e Assume that t = Y - X a t -t .. aI' a2 , •.. , aq are nonnegative and aO is positive. Although Yt is conditionally normal, Engle (1982) established that the Yt are not jointly normal and the marginal distribution of Y is not normal. t restrict to the case q = 1. In this manuscript w" -3- Engle (1982) derived the moments of the let} process for q = 1. In section 2 we derive a representation for the let} process and use it to derive the moments of the let} process. let} process. We also derive the ergodic properties of the Engle (1982) also considered the maximum likelihood estimation of the parameters. He established that the information matrix is block diagonal, . indicating that the maximum likelihood estimators of ~ and~are independent. He also indicated that the maximum likelihood estimators are asymptotically normal. In section 3 we formally derive the asymptotic distribution of the maximum likelihood estimators. We also derive the asymptotic properties of the least squares and estimated generalized least squares estimators of 2. ~ and ~. Properties of the ARCH Models The simplest ARCH model is the first order linear model given by y t =X a -t- (2.1) where (2.2) and Ft - l = a-field generated by Y and e ' sSt-I, t • 1, 2, .•. s s It is assumed that So is strictly positive and 8 In the 1 is nonnegative. following theorem we obtain a representation for the {e~} sequence which is useful in deriving the properties of the model (2.1). Theorem 2.1: Let let} be a sequence of random variables satisfying (2.2). Assume that So >0 and 0 ~ 81 < 1. It is assumed that the process began indefinitely in the past with a finite initial variance. e where {Zt; variables. t 2 t ~E Bt ( tw = So t=O 1 i=O Z2 .) t-~ a.s. , Then, (2.3) = 0, ±l, ±2, ... } is a sequence of normal independent (0,1) -4- Proof: Since e is conditionally normal, we get t where Zt is a standard normal random variable independent of e - l . t e Therefore, 2 = Z2( 13 2 t 0 + Sle t _ l ) t 2 2 2 = Zt [SO + SlZt_l(SO + 6l e t _ 2 )] j-l l l 2 j j-l 2 2 I: 6 ( 'IT Zt-i) + 6 ( 'IT Zt-i)e = 6 0 l=O 1 i=O 1 i=O t j Note that, l Z2 .)] = (1 - 1>1 " )-1 < (i'IT l=o 1 =O t-l. E[ 00 I: 6l 00. Therefore, by result (xi) of Chung (1974, p. 42), co I: l=O l l 2 13 ('IT Z ) < 1 i=O t-i a.s . . co Also, assuming that the process was initiated in the infinite past with a finite initial variance, co E [ I: j=O Using the same argument, co "j (j; 1 I: j=O 1>1 i=O 2 Z . )e t-1. 2 < co . t-J a. s. , and j j -1 Z2 6 ( 11' 1 i=O ) 2 t-i e t-j - 0 a •s ., as j + co • Therefore, 2 e t = 60 co I: l=O l l Z2 .) 61 ('IT t-l. i=O A similar expression for e 2 t a.s. 0 can be obtained in case q > 1. Note that e t has a sYmmetric distribution and that the sequence {e } is strictly t stationary. The probability density function f of e t can be obtained by solving -5x ~ f f(x) • 2 2(SO+B u 2 ) l e -00 We have not been able to obtain a solution for (2.4). obtain the moments of the {e } process. t Theorem 2.2: We use (2.3) to See also Engle (1982). Let {e } be a sequence of random variables satisfying the t conditions of Theorem 2.1. m 2r Then, = E[e 2t r ] < 00 i f and only i f er • Br1 Also, if e r r ~ (2j-l) j=l <1 • < 1, then r-j Proof: Note that, . and by Monotone convergence theorem (see Chung, 1974, p. 42), Using Minkouski's inequality, E[e~r] :a 13 0r .tim f . n~ n-- . J=O • a0r < Also, 00 .tim n-if [n. ~ J=O j B1 {E 13 er < 1 j i~O Z 2r 1 } r t-i -1 e;j+l)/rr 1 . r ] -6- E[e;r] Therefore, M Zr sjr E( 'II'j zZr ) 0 j=O 1 i=O t-i co -j = Sr0 r Sl 6( j+l) r j=O ~ Sr = co co r if 6 is finite iff 6 r r ~ 1 < l. Now, = E[e;r] M Zr .. E[ZZr<S + Sl e2 )r] t 0 t-1 r r-j sj -r = r (~) m 6 Sl So J 1 Zj r j=O r-l ,. 6 (1-6 )-1 r (~) r r j=O J Note that if 3S 2 1 < 1, So Sl r-j m 2j 0 then v(e~) .. 204(1-3Si)-1 and where 2 is a sequence uncorrelated (0,0 ) random variables. We now give two lemmas that are useful in obtaining the ergodic properties of the {e } process. t Lemma 2.1: F n ~ F + . n l Let {U } be a sequence of F measurable random variables and n n Suppose there exists an integrable random variable U and a constant c such that p[lu I >xl ;:icp[lul >xl. n Then -7- n -1 n k=l[U k - E(UkIFk_l)1 <~ If E[IUllog+lul1 -E,. .0. or if {Uk,k ~ I} and {E(UkIFk_l),k ~ I} are strictly stationary sequences then the covergence is almost sure. Proof: See Hall and Heyde (1980, p. 36). . Lemma 2.2: 0 A covariance stationary process {X } obeys the mean law of large t numbers, i.e., X n =n -1 n 1: t=l X t converges in mean square to a square integrable random variable X. process is st~ictly If the stationary and integrable, then X converges almost surely n to an integrable random variable. Proof: See Reve~sz 0 (1968, p. 99). In the following theorem we obtain the ergodic properties of {e } • t Theorem 2.3: er < 1, Consider {e } satisfying the conditions of Theorem 2.1. t then n n If 38~ < -1 a.s. , -1 n 1: tal e 2r-l -0 t a. s. . 1, then for a fixed j ; 0, n -1 n 1: tal e e .--+ t t-J 0 a.s. and a. s. . Proof: From Lemma 2.1 and Lemma 2.2, If -8-1 n E e 2t - M 2 n (say) , a.s. tal and a. s. , where a.s • . Therefore, a. s . . Also, - a. s. , and hence or .. S .. E[e~] • -1 n 2 M 2 0 S (1 - 1 )-1 Therefore n For r ~ a.s. 2, we use induction procedure to prove the result. n for s .. 1, 2, -1 n I' -1 n - Assume that a. s. , E tOIl ... , n and E tal e t - 1. 2r E tOIl e t - We know from Lemma 1 and Lemma 2 that a.s. (say), -9- Note that E[e 2rl F _ ]" t t 1 er r r (j) J'''0 cor; _ B1 2. e J t-1 Therefore, M 2r .. er r-1 r j=O (~) J c:~) or MOle (1- e )-1 r;:l 2r r r j-O r-j 2j + M er M 2r (~)(60)r-j M . 2J 61 J a.s. a.s. _ E[e 2r ] t Therefore, a. s. . Note that, if er < 1, then a.s. and hence n -1 n 2r-1 r e -- 0 tOIl t a.s. as n-+ ~. Similarly, and we g'!t n -1 n r tOIl et et - . J 0 a.s. as n- ~. Consider, -- 6 .. 0 2 4 E(e _ ) + 6 E(e _ ), 1 t 1 t 1 E(e~ e~_l)' Therefore, by Lemma 1, n -1 n 2 r e t e 2t-1-tOIl a. s. . a.s . -10- Similarly, n as n -1 n 2 2 1.: e e t=l t t-j 2 2 E(e t et_Jo) a.s. + .... In the next section we consider estimation of the parameters a and 3. §. Estimation of ARCH Models Given Y ' Y ' ... , Y satisfying (2.U, we desire to estimate ~ and § = (6 0 ,6 1 )'. n l 2 Engle (1982) used the method of scoring to obtain the maximum likelihood estimates, but did not formally derive the asymptotic properties of the maximum likelihood estimators. Engle (1982) indicated that if the conditions of Crowder (1976) are satisfied then the maximum likelihood estimates are asymptotically normal. We derive the limiting distribution of the maximum likelihood estimators by verifying the conditions of Hall and Heyde (1980, p. 174). We also consider an estimated generalized least squares estimator and derive its limiting distribution. We consider two particular choices for (a) ~t ~t is fixed and bounded; and For case (b), we also assume that ~l consider the likelihood given ~l' a (YO' Y- l , ••• , Y-P+l) is given and we We will also indicate how the results may be extended for the case where X consists of both fixed and lagged values and -t also for the case with X -t 3.1. that are not necessarily bounded. Maximum Likelihood Estimation Consider the log likelihood conditional on L (y) n - = -n -1 t=l ~ [in f (y t t IF t- l~:; ~l ' -11- where f (y t t IF t- 1) X' .. (~', ~' ) . is the conditional density of Y given the past and t From (2.2), (3.1) where Therefore, L (y) = constant n _ + (2n)-1 ~ bih \+ (2n)-1 ~ h-l(Y - tal \ ;I t=l t t X a)2 -t- (3.2) The maximum likelihood estimator in of X is the value of X that minimizes Ln(X)' Let X~ a (a~,~') be the true value of X. All probabilities and expectations are taken with respect to the true value XO. We assume that and (3.3) so that E[e 12 ] t <m (If X is fixed and bounded then we may assume that t 8 o S 6 S 0.3, so that E[e ] < m.) We also assume that ~O is in the interior 1 t of a compact set L. Therefore Xo is Let as~umed to be in the interior of f. r. defined below will be taken to be contained in All neighborhoods For 0 > 0, and IIx - xoll we obtain from the Taylor series expansion that Ln(r) • Ln(rO) + (r - ro)' ~:~~ + ~(y - y ~ ~o )' H (y-y ) n - ~o (3.4) < 0, -12- where and y* is a point between y and y o (not necessarily the same at each occurrence). We include a result from Hall and Heyde (1980) that we use to obtain the asymptotic properties of y -n Theorem 3.1: • Suppose that lim s~ n-+a> <5 * <5 -1/ T(y)I n .. J.] < co a.s. , 1 S i S P + 2 1 S j ~ p + 2 (3.5) a. s. , (3.6) and (3.7) . where ~ is a positive definite matrix of constants. of estimators fin} such that in converges to there is an event E with P(E) >1 - € Yo Then, there exists a sequence almost surely, and for and nO such that on E, for n € >0 > nO' Yn satisfies and L (y) attains a relative minimum at in • n - n -~ (aaLynj - L N(Q,~) y""yO where W is a positive definite matrix, then If, in addition, (3.8) -13- n Proof: ~(- rn rO ) -L - N(O _'!! -1 ~!! -1) . 0 See Hal! and Heyde (1980, p. 174). We now compute the partial derivatives of Ln(r). Note that, (y aL -aB n t-1 - X a)X' -t-1- -t-1 ' . --12n 1 n .. - r X' X t-1-t-1 n t=l (y n 1 - 2B 1 -n r t=l [(y t X )2 X' X t-! - -t-!~ -t-1-t-! - Xta) - - ] (y + X' X] - X a) [X'X t-l -t-l-t-t-1 -t-l-t h2 t 2 n [2(Y t -X-t-a) - h t ] 1 =- r 2n t:l1 a~a~' h3 t a2L n [(Yt-1 ~ -X a) 2 t-l -t-l- 4] (Yt-l-!t-l~) (y X a)2 -t-l- and a2 Ln a§a~' ¥ [(Yt-!t~)2- = -n1 t:ll ht ] h2 t [~] (y t-l - X a)X -t-l- -t-l 2 n [2(Y t -X-t-a) - h t ] r n t=l h2 t 1 [(Yt - 1- X a)] ;t-l:)3 (y t-l -t-!- X -t-1 X -t- l · -14- Recall that 1 e t = Zt v~t t = eO where v 2 + e l e t _l e ' ,. (eO,e ) is the true value of ~' = (SO,Sl) and Zt is a sequence of l independent N(O,l) variables. Therefore, 1 nE -~ + _1 e nE (Z~-l) e .~ Z X' X' tool-tool nt-I v t toot n 1 tal = and 1 n - - 2n tal E In the following theorem we verify the conditions (3.5) - (3.8) for the case (a) where we assume that X is fixed and bounded. -t Theorem 3.2: n -1 where ~ Assume that X -t ¥ is fixed and bounded. X'X __ A , tal -t .. t .. as n - "" , is a positive definite matrix. Then, the conditions (3.5) - (3.8) are satisfied with H - W,. [!!ll Q Q]., !!22 where !!U !!22 c ,. A(c .. ,. . ~ 2 1 + 2e l c 3 ) r c2 .. c 3 c3 ] c4 -1 c2 - E[v 2 ] -2 - E[v 2 ] c = E[etvi 2] = l 3 Also, assume that eil(c l - eOc 2) , -15- and c Proof: 4 = For a fixed i, let U n =v -~ n Z -1 n M X ., n, 1 where M" sup {X .}. i,n n,l Then, E[U n IFn- 1] = 0 2 and E(U ) is finite. n n -1 n ~ t=l U - t a.s. Therefore, by Lemma 2.1, 0 a.s. , as n~ = . Similarly, a.s. for k Also, and = 0, 1, 2, 3, 4 and t,s = 0,1 . -16- Note also that, =0 a.s. and Therefore, lim n n~ -1 n -1 = 0~ k v e - X t=l t t l -t-l a. s. , and a.s Now consider, coV(V~l, v~:j) = E[v~lv~:j] {E[v~1]}2 _ -1 { j-l l l 2 }-l v-I] :a E[v t ] E[ 6 0 + 6 1 So l;O 6 1 i~O Zt-l-i - t ~ 6- 3 6 j + l E( 2 .. 0 1 e t - l _j ) Therefore, as n - "" and since v -1 is bounded the convergence is almost sure. t Similarly, lim n n~ -1 n k t=l V -IX' t -t X = A c ~t l and lim n n~ Therefore, -1 n -2 -2 k v = E[V ] t t=l t a.s. a. s. , -17- a. s. , and Since and we observe that is positive definite. ~ Now we establish (3.5). Note that IT n (x)I 1J .. is a linear combination of terms of the form = h -t k ft(y; a, b k) -' = 0, where a 1, 2; b II X - Cons ider for = 0, Xo II < (Y t - X a)b - X a)a(y -tt-l -t-l- t - where Y* = A.y. + (l-A.)Y i 1 1 i, 1 2, 3. 15 , Ift(y;a, b, k) - f (yo;a, b, k)1 ~ 15 - = 1, 1, 2, 3, 4; and k 0 and 0 p+2 r i-l ~ Note that, - y.1, 01 Iy~ 1 ~ A.1 Iy.1 - y.1, 0 1 < 15 'X* < gt(a,b,k) , • We will show that af I for all X* and n -1 n r t=l ay: , lit - xoll < 15 , gt(a,b,k) converges almost surely to a finite limit. -18- For example, consider Note that, .. - 3h~4(Yt - Oft aS l 2 a) ?St~) (Y t - l - X -t-l- 6 Now, .. e Y - X -tooa t t - at where a t .. X (a - a ) • -t - -0 Then, -1 n 8 2 6 n L ete Note that, since we assumed that E[e ] is finite, t _l t t=l -1 n -1 n 6 2 L e - converge almost surely to finite constants. n L e and n t t=l t 1 t-1 other terms of IT (x)I .. can be bounded to establish (3.5). n 1.J , Now, using Scott's martingale central limit theorem (see Scott (1973» we Similarly, 1 (aL~ will show that n~ ~ is asymptotically normal. aX y"y - -0 Define, x-xo where D' = (Db, Sn" 11 1 , 11 ) is an arbitrary column vector such that 2 ~ a1 t~l -1(Z2 1) vt tOO 'X' ~ -~ Z n'X' et-lDO-t-l - t~l v t t~O-t D'n ~ O. Note that, -19- is a martingale. It is clear that {S ,F } n n Let ] V2 _ E[S2/ F n n n-l and s . 2 _ E[S2] n n Then, V2 - 2 n +~ n -1 -2 2 v t e - n'X' X n !)O + E v t n'X'X l -O-t-t-O -O-t-l-t-l t t-l t-l n e12 E n E t=l 2 -2 v t (n l + n 2e t _ l ) 2 and s 2 2 n n - 26 c E n' X' X n + c n' E X' X nO n 1 3 tal -O-t-l-t-l -0 1 -0 t-l -t-t - Note that, P as n - -1 co and o ~im n n-- -1 2 s n Since we assumed that Elv- 3 / 2 t z3 1 t , D~ D . E[e~] < co, < it follows that co ' and the Lindeberg condition is satisfied. s and -1 L S n n N(O,l) Therefore, -20- o Now we consider the case ~t = (Y t - l , Yt - 2 , .•• , Yt - p ), We assume that the roots of the characteristic equation p-l mP - a m 1 - a - lie inside the unit circle. p = 0 Then, Y can be written as an infinite moving t average as, co 1: j=O (3.10) w.e . J t-J where {w.l satisfy J =1 , j • 0 = 0 , j <0 In the following theorem we obtain the asymptotic properties of the maximum likelihood estimator. -Theorem 3.3: Assume that X -t = (y t- l' Yt- 2' •.. , Yt-p ) and that the roots of the equation (3.9) lie inside the unit circle. are satisfied with H = W= - 2) G -~l !!22 0 where and ~22 is as defined in Theorem 3.2. Proof: For a fixed i, let U n = v n-1 e n Y n-i Then the conditions (3.5) - (3.8) -21- Then, 2 and E[U ] is finite. Therefore, n n -1 n I: t=l a.s. U - 0 -2 2 (e -v ) e t "Also, n -1 n I: t=l v t t t t- lX 1 . t- ,4 -0 a.s. , and as n + 00 Using the arguments similar to those in Theorem 3.2, we get a. s. , = 0 = !!22 a.s . . = n Also, r =y-0 tim n+OO -1 n I: t=l n + tim 26 12 n -1 I: v -2 e 2 X' X n+OO t=l t t-l-t-l-t-l Consider, for fixed i and j n -1 n -1 I: v Y .Y . = t-4 t-J t=l t Now for fixed q and r, consider n -1 n I: t=l -1 vee t t-q t-r 00 I: k=O 00 w w n s=o k s I: -1 n I: t=l v -1 t e t - i - k e t-j-s -22I f q=r=l, then n -1 n v -1 2 tOIl t e t - 1 1: = (n6 ) 1 -1 n 1: tOIl -E[v -1 e 2 _ J t t 1 v -1 t (v -6 ) t 0 a.s. .If q"r>l, then it can be shown that for j > q. Therefore, a.s. Now if q ~ -1 r, then v e e are uncorrelated and hence t t-q t-r e -1 e .. E[V e e J t-q t-r t t -q t -r .. 0 a.s. Using Lemma 6.3.1 of Fuller (1976), we get lim n n~ -1 n -1 -1 1: v Y .Y . " E[V Y .Yt.J t t -1 -J tOIl t t-1 t-J a.s . . Similarly, lim n n~ and hence Note that -1 n -2 -2 1: v Y Y . " E[V Y 'Y .J t t -1 t -J tOIl t t-i t-J a.s. , -23- for any arbitrary ~O Now. E[v • -1 vp+1(Y p ' Yp - l ' .... Yl )· -1 , !t!t 1 is the variance covariance matrix of t If then or ~o !~l • 0 a.s . . However. we know that the variance covariance matrix of !p+l = (Y p .Y p- l •..•• Y1 ) is positive definite. Therefore. ~tl is positive definite. Now to establish (3.5). note that IT (Y)l , is a linear combination of terms -n - i J of the form -k a X a)b ql q2 q3 q4 ft(v;a,b.k.a) • h t (Y t - Xta) Xt -1, i Xt-.J l ' Xt.r Xt.~0 ~ ~ - - (y t- 1 - -t- 1 where qi =0 or 1 and a. band k range from 0 to 4. Again, for example. consider f (y_;2.4.3.0_) t = h-t 3 (y t - X a)2(y - X a)4 -tt-1 -t-1- Then :ii constant [e t 6 2 2 6 2 06 p2 y2 . e 6 _ + Y 1 .e + 0 e t t 1 i=l t- -1 t i=l t-1 t 1 + and 0 8 p 2 t ~ t i=l j=l y2 y2 . ] t-i t-J-l -24- 3 a t- 1) y t-1.. 11 ·Since E[e~2] is assumed to be finite, we get lim sup 0n~ 1 0+0 IT-n (x*)I l.J .. < Now we verify (3.8). ~ a.s. Let S = n' (aLaX ') n where n\ _ n' = (~~, n1 , n2 ) is an arbitrary vector of constants such that n'n ; Then, n r _!.< v"Z n'X' t=l t t_O-t vn2 = E[S2n IFn- 1] = 2 8 and Note that 2 ~ w 1 t=l v -2 e 2 n'X' X n + t t-l.0·t-l·t-l.0 v -1 n'X'X n t .O-t·t.O °. -25- as n - 1 , Since we assumed E[el~J t < 00 00 , and Therefore, by Scott's martingale central limit theorem, s -1 n S n L N(O,l) , and hence - YO) - L N(Q,~ -1 ) It is easy to see that if ~t o = (l,Y t - l , Yt - 2 , ... , Yt-p+l) the maximum likelihood estimator is still consistent and asymptotically normal. is fixed but not necessarily bounded then than n -~ . ~ may converge to a For example, if X = t, then (a - a ) t O X -t at a rate faster is 0p(n -1 ). Now we consider the least squares estimation of 3.2. O If y. Least Squares Estimation: The maximum likelihood estimates considered in 3.1 do not have explicit expressions and are estimated using iterative procedures. We now consider the ordinary and estimated generalized least squares estimates of y. The least squares estimates are obtained as follows: Step 1: of ~. Regress Y on X to obtain the ordinary least squares estimator a -t -t - -26- -2 -2 to get -6 and 6 , Let Regress e on a column of ones and e 0 1 t l t -2 v 6 + 6 e t 0 1 t l --1 -2 --1 --1 -2 e on v and v e - to get an estimated generalized Step 3: Regress v t t t t t l -2 + 6 e _ least squares estimates and • Let v .. l t l l t Step 4: Regress v-~ Y on v-~ X to get an estimated generalized least squares t t t -t Step 2: - - .. . eO estimate a - of e -t - Let {!t} and {Y } satisfy the conditions of Theorem 3.2. t YO be in the interior of f. ~ - (r - YO) - L Then N(2, ~O) and where y'" (~',§') 2 -1 B .. -1 o - cr A B .. o We first consider the case is fixed and bounded. Theorem 3.4: n A a We now study the properties of ~, ~, ~ and~. where X eO [ o -1 2 ~l -1] , ~l ~l Let -27- -1 - o -1 ] 2 !!22 ~l !!22 A .. E -1 and c l and !!22 as defined in Theorem 3.2. Proof: Note that ... , x' ) -n We know that Consider, S n .. n' _0 X'e n = tOIl 1: bte t where b t .. ~ n. i-l~' 0 X . t,~ and ~O is an arbitrary vector of constants with a martingale with, v2 .. E[S2 IF n n n- 1] a. s. , nb nO ~ O. Note that {S ,F } is n n -28- and s 2 n 2 '" (J Therefore, s L -2 2 V n n 1 and -1 n , 2 !lO sn - 2 ~!lO (J Since E[e:l is finite, the Lindeberg condition is satisfied and Therefore, ~ L 2 -1 n (':-':O)-N(Q,(J~ ). Now consider, n _ =[ ~ E t=2 (n-1) n E n -2 Ee t=2 t-1 t=2 where e and Note that, t = Y t - - X a -t- -29n -1 n -2 r et-la t - l + n t=2 r a t-l t=2 Similarly, n -1 n -4 t=2 t-l r = e and n -1 n -2 -2 r e e t=2 t-l t n = n -1 t=2 t 2 2 0 (n-~) et_le t + P Therefore, - n~ (a-a) - -- n 1 n [ -1 2 nEe t-2 t n -1 n r e2 t-2 t -1 n 4 r e t=2 t r [n-~ n r ta 2 dt ] n n -~ r dte~_l t=2 + o p (n-~) where d t - (Z2 - l)v t t Note that, n -1 n r t=2 e 2 t -1 n 4 n r e t=2 t as n + 00 r - A -I a. s. , • Con!3 ider , where nl and n2 are two arbitrary constants. Then S n is a martingale with -30- and Using the usual arguments we get s -1 S - L n N(O,1) , n and hence ~ L -1-1 n (~-~)--+ N(Q, ~l ~l ~l ) • Z Note that (Zt-l) and e t are uncorrelated and hence ~ and e are asymptotically Now we consider ~ obtained by step 3. independent. Using arguments as above, we get n n -e -1 v--2 e -2 ] t l t taZ n --Z -4 1: v t=Z t e t - l v- -2 2 t v--2 -2 t-Z t e t - l [ti = 1: and t=2 -1 + 0 p (1) n 1: t=Z Consider, n is a martingale with v 2 .. Z n n 1: (n t=Z Z Z n e _ ) 1 + Z t 1 and s Z n 1: v e --1 -2 ] t=Z t t n --1 -Z -2 1: v e e t=Z t t-l t n 1: !!Z2 Then, S [n = Z(n-l) E(n l 2 Z + n e _ ) Z t 1 a. s. , • -31- Since E[e~] is finite, Sn satisfies the Lindeberg condition and -1 L s n Sn - N(O,l) Therefore, A-l Now we consider the regression of v t A-l Y on v X t -t t where Let Then, as n -+ ... , and Consider, = ¥( ~ n X .)v·~ Z t=l i=l i,O t,1 t t Note again that S n is a martingale with and The Lindeberg condition is clearly satisfied and hence Note that, -32- s-IS n ~N(O,l) n and 2 Since (Z t -1) and Z t are uncorrelated . a . and 8 are asymptotically independent. .. A 0 Note that and the equality hold only.if 8 1 = O. Therefore, the maximum likelihood estimator is asymptotically the best among the three estimators considered and the estimated generalized least squares estimator is asymptotically better than the ordinary least squares estimator. Here we have assumed that X is fixed and bounded. .. t Suppose X is fixed .. t and satisfies - a -h,i,j n _ r lim co n- tal and lim n- nr G tal 2 ~ -1 X. 2 X. = 0 , t,1 n,1 for i = 1, 2, ..• , p; j - 1, 2, ••• , p , where o = ..n diag{(~ x2 .\ ~ \(=1 t~ Then, it can be shown that , ... , ~n r X 2)~} t=l t,p -33- and Now we obtain the asymptotic properties of the least squares estimators for = (Y t - 1 • the case when ~t Theorem 3.5: Yt-2' ...• Yt - p )· Assume that X saisfies the conditions in Theorem 3.3. ~t and where B ,. ~3 [9 o ~ Q ~ for j ~ 1 = a 2 9-1 + 6 Q-1 Q Q-1 1~1 ~2~1 i. 2 (Q2)·· = E[Y .Y .e 1] ~ 1J t -1 t -J t CD k+i-l = y e 2(0)k=O 1: wkwk +· .6 1 J-1 Then. -34- and -1 = E[v t Yt -1.'Y t -].J ~ -1 2 = k~O wk wk+i - 1 E[v t et_i_kJ . Proof: Note that, -1 n E Y n .Y a.s. . tal t-1. t-J and hence a.s. Consider, S n a n'X'e -0- ... n = t=l Eb e t t where b a t n. OY . i=l 1., t-1. t and n_ is an arbitrary vector with n'n ~ O. Then S is a martingale with, -0n n n' X' X e 2 _ ~O vn2 = 60 -0 n' x'x nO + 6 a.s. , E -0 1 -t -t t 1 t=l O and s 2 n = n 6 n' gl~O + n 6 n' 1 -0 0 -0 Note that, for j n -1 n E Y .Y ~ E[~~~te~_lJ ~O i, e 2 t=l t-1. t-j t-l ~ = k=O E -1 n 2 E wkw s n E e . k e .e 1 s=O t=l t-1.- t-S-] t~ -35- Therefore, and n -1 2 sn - Since E[e:J is finite, the Lindeberg condition is satisfied and s -1 n L S -N(O,I) . n Therefore, Using the arguments similar to those of Theorem 3.4, it follows that n~(~-~)~(2' 2 ~~l~l~~l) n~(~-~)~(2' 2~;~ ~l~;~) , and a and ~ are asymptotically independent. Now, to obtain the limiting distribution of A ~, consider = where gn and v2 n ~O are as defined in Theorem 3.4. = nO' X'G-1Xn - -n --0 a. s., and From Theorem 3.3, we know that Then, S n is a martingale with -36- n -1 n -1 -1 r v Y .Y . -- E[v t Yt -L.Y~.J t=l t t -L t -J ~-J a.s. Therefore, s -2 2 n Ll V n . and n -1 2 n' E[v-1X'X J sn -- -0 t -t-t ~O· Using Scott's martingale central limit theorem, we get s -1 S n L -N(O,l) n and Note also that ~ and ~ are asymptotically independent. 0 If X involves both fixed and lagged variables then one can obtain results -t si.milar to those of Fuller, Hasza and Goebel(l981). Also, if Yt process has a unit root, we can obtain the asymptotic distribution of the least squares estimator. Consider, for example, and {e } satisfies the conditions of Theorem 3.4. t of a l The least squares estimator is given by, a l = [t~2 Y;-l] -1 Then, n 2 -1 n = [n - 2 t=2 r Y J [n- l r Y e J t-l t=2 t-l t -37- If 6 = 0, Dickey and Fuller (1979) obtained the asymptotic distribution of 1 n(a 1 - 1). We now show that even if 6 1 ~ 0, n(a l - 1) has the same limiting distribution. We know that .tim n n-- -1 n 2 r e t=2 t a.s. Now consider, T n = n -~ Y = n-l r n-1 a. z* 1.,n i,n i=l and -2 n r n ,. n r y2t-l t=2 where z* = (Z*l ,n ' ..• ,Z*n- 1 ,n ) , = M e -n "'n mit(n) = (i,t) - th element of = 2(2n-1) -!.. 2 ~n 1 Cos[4n-2)- (2t-l)(2i-1)'lf] and l a i, n = ith element of n -~(l , ... , l)M-n -38- Using Scott's martingale central theorem, it follows that, for any fixed k, where !k is an k x k identity matrix. Now using the arguments similar to Hasza (1977) and Pantula (1982) it follows that n(~l - 1) has the same limiting distribution as that obtained by Dickey and Fuller (1979). Similarly, the results for pth order ARCH models may be obtained. 4. Summary: We have considered linear regression models with autoregressive conditionally heteroscedastic errors, introduced by Engel (1982). representation for the first order ARCH errors. We have obtained a series We hav.e used the representation to derive the ergodic properties of the errors. Similar representation can be obtained for the qth (q> 1) order ARCH errors but are not presented here. A special 2 case where the conditional error variance is of the form 80 + 8 a.e . 1 j=l J t-J -1 -1 [q+l-j] will be considered elsewhere. a. = q or a. = 2[q(q+1)] t J , where J We have considered the maximum likelihood estimation of ARCH regression models. The maximum likelihood estimators do not have explicit algebraic form and are computed using iterative methods. We have shown that the maximum likelihood estimators are strongly consistent and asymptotically normal. We have also shown that the least squares estimator and an estimated generalized least squares estimator are asymptotically normal. (Y t For a random walk model = alY - + et,a = 1) with ARCH errors, we have shown that the asymptotic l t l distribution of the least squares estimator of a l is the distribution obtained by Dickey and Fuller (1979) for the homoscedastic case. Acknowledgements I wish to express my thanks to Professor Wayne A. Fuller for helpful suggestions. -39- BIBLIOGRAPHY Chung, K. L. (1974). A Course in Probability Theory. Crowder, M. J. (1976). Observations, Academic Press, NY. Maximum Likelihood Estimation for Dependent Journal of the Royal Statistical Society, Series B, 45-53 • . Dickey, D. A. and W. A. Fuller (1979). Distribution of the Estimators for Autoregressive Time Series with a Unit Root, Journal of American Statistical Association, 427-531. Z~, Engle, R. F. (1982). Autoregressive Conditional Heteroscedasticity with Estimates of United Kingdom Inflation, Econometrica, Fuller, W. A. (1976). 2Q, 987-1007. Introduction to Statistical Time Series. Fuller, W. A., D. P. Hasza, and J. J. Goebel (1981). Wiley, NY. Estimation of the Parameters of Stochastic Difference Equations, The Annals of Statistics, 2, 531-543. Granger, C. W. J. and A. Anderson (1978). An Introduction to Bilinear TimeSeries Models. Vandenhoeck and Ruprecht, Gtlttingen. Hall, P. and C. C. Heyde (1980). Martingale Limit Theory and Its Application. Academic Press, NY. Hasza, D. P. (1977). Estimation in Nonstationary Time Series. Unpublished Ph.D. Thesis, Iowa State University, Ames, Iowa. Jones, R. H. (1965). An Experiment in Nonlinear Prediction, Applied Meterology, Pantula, S. G. (1982). Time Series. ~, Journal of 701-705. Properties of Estimator of the Parameters of Autoregressive Unpublished Ph.D. Thesis, Iowa State University, Ames, Iowa. Priestly, M. B. (1978). The Statistician, Nonlinear Models in Time Series Analysis, ~Z' 159-176. -40- Revesz, P. (1968). Scott, D. J. (1973). The Laws of Large Numbers. Academic Press, NY. Central Limit Theorems for Martingales and Processes with Stationary Increments Using a Skorokhod Representation Approach, Advances in Applied Probability, ~, 119-137.
© Copyright 2024 Paperzz