Testing for Pairwise Serial Independence via the Empirical Distribution Function Author(s): Yongmiao Hong Source: Journal of the Royal Statistical Society. Series B (Statistical Methodology), Vol. 60, No. 2 (1998), pp. 429-453 Published by: Wiley for the Royal Statistical Society Stable URL: http://www.jstor.org/stable/2985949 . Accessed: 22/11/2013 14:15 Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at . http://www.jstor.org/page/info/about/policies/terms.jsp . JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. . Wiley and Royal Statistical Society are collaborating with JSTOR to digitize, preserve and extend access to Journal of the Royal Statistical Society. Series B (Statistical Methodology). http://www.jstor.org This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions J.R. Statist. Soc. B (1998) 60, Part2,pp.429-453 Testing forpairwise serial independence via the empiricaldistributionfunction Yongmiao Hongt CornellUniversity, Ithaca,USA [ReceivedJuly1996. RevisedMay1997] Summary.Builton Skaug and Tj0stheim's approach,thispaperproposesa newtestforserial independenceby comparing the pairwiseempirical distribution functions of a timeseries with the productsof its marginals forvariouslags, wherethe numberof lags increaseswiththe samplesize anddifferent lagsare assigneddifferent weights. themorerecentinformation Typically, receivesa largerweight. The testhas some appealingattributes. Itis consistent againstall pairwise dependencesand is powerful againstalternatives whose dependencedecays to zero as thelag increases.Although theteststatistic is a weighted sumofdegenerateCram6r-von Mises ithas a nullasymptotic statistics, N(O,1) distribution. The teststatistic and itslimit distribution are invariant to anyorderpreserving transformation. The testappliesto timeserieswhosedistributionscan be discreteor continuous, withpossiblyinfinite moments. theteststatistic Finally, only involvesranking theobservations and is computationally simple.Ithas theadvantageofavoiding smoothednonparametric A simulation estimation. is conductedto studythe finite experiment sampleperformance oftheproposedtestin comparison withsome relatedtests. Keywords: Asymptotic normality; Cram6r-von Misesstatistic; Empirical diitribution function; Hypothesis testing; Serialindependence; Weighting 1. Introduction The detectionof serialdependenceis importantfornon-Gaussianand non-lineartimeseries. Tests forserialindependenceare usefuldiagnostictools in fitting non-lineartimeseriesand identifying appropriatelags, especiallywhen the time serieshas zero autocorrelation(see Grangerand Terasvirta(1993)). For instance,to detectnon-linearity we can testwhetherthe residualsfroma linearfitare independent. We can also testtherandomwalk hypothesis fora timeseriesbytesting whether itsfirst differences are seriallyindependent. Testsforindependence are usefulin othercontextsas well (see Robinson (1991a)). Independenceis stilla testable in populationswithinfinite hypothesis second-order moments. In practice,correlationtests(e.g. Box and Pierce (1970) and Ljung and Box (1978)) are widelyused to testserialindependence,but theyare not consistentagainstalternatives with zero autocorrelation. The lack of consistencyis unsatisfying fromboth theoreticaland practical viewpoints.Also, thesetestsrequiresecond-or higherordermoments,excludingtheir applicationsto timeserieswithinfinite second-ordermoments. There have been manynonparametric testsfor serialindependence,e.g. Chan and Tran (1992), Delgado (1996), Hjellvikand Tj0stheim(1996), Pinkse(1997), Robinson(1991a) and tAddressfor correspondence: Departmentof Economics and Departmentof StatisticalSciences,College of Arts and Sciences,Uris Hall, CornellUniversity, Ithaca, NY 14853-7601,USA. E-mail:[email protected] ? 1998 RoyalStatistical Society This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 1369-7412/98/60429 430 Y. Hong Skaug and Tj0stheim(1993a, b, 1996). These testsare consistentagainstserialdependences densityestimation.Both up to a finiteorder.Some of themrequiresmoothednonparametric theoryand simulationstudies(e.g. Skaug and Tj0stheim(1993b, 1996)) suggestthatthefinite testsmay dependheavilyon the choice of of smootheddensity-based sampleperformances methods thesmoothingparameters.Resamplingmethodssuchas bootstrapand permutation have been used to obtainaccuratesizes (e.g. Chan and Tran (1992), Delgado (1996), Hjellvik and Tj0stheim(1996) and Skaug and Tj0stheim(1993b, 1996)). is Skaugand Tj0stheim's estimation testthatavoidssmoothednonparametric One important whichextendsthetestof Blumet al. function, (1993a) testbased on theempiricaldistribution way. The teststatisticis invariantto any order (1961) to a timeseriessettingin a significant The distributiongeneratingthe data can be continuousor dispreservingtransformation. crete;when it is continuous,the test is distributionfree.No momentis required;this is as oftenarisesin economicsand high attractivefortimeserieswhose variancesare infinite, frequencyfinancialtimeseries(e.g. Fama and Roll (1968) and Mandelbrot(1967)). Finally, thestatisticdependsonlyon rankingtheobservationsand is simpleto compute.We notethat functionvia an alternativeapproach. Delgado (1996) also used the empiricaldistribution Builton Skaug and Tj0stheim's(1993a) approach,thispaper proposesa new testforserial functionsof a timeserieswith independence.It comparesthepairwiseempiricaldistribution the productsof its marginalsforvarious lags. The testhas some new attributes.First,it is consistentagainstall pairwisedependencesforcontinuousrandomvariables.Thus, the test is available. It may also be expectedto have good maybe usefulwhenno priorinformation power against long memoryprocessesbecause a long lag is used (see Robinson (1991b)). largerweightsare givento lower lags; typically weightsare givento different Second,different orderlags. Non-uniform weightingis expectedto givebetterpowerthanuniformweighting whosedependencedecaysto zero as thelag increases,as is oftenobserved againstalternatives forseasonallyadjustedstationarytimeseries.Third,althoughour statisticis a weightedsum of Cramer-vonMises statistics,it has a null asymptoticone-sidedN(O, 1) distributionno matterwhetherthe data are generatedfroma discreteor a continuousdistribution.We consider some 'leave-one-out'test statistics,which improve size performancesin small samples.We also presenta slightlymodifiedversionof Skaug and Tj0stheim's(1993a) test thatimprovesthe size in small sampleswhen relativelymanylags are used. We emphasize thatthenew testshouldbe viewedas not competingwithbut as a complementto Skaug and regimesand have theirown merits. Tj0stheim's(1993a) test,because theyworkin different fewlags are tested,whereasthe Skaug and Tj0stheim's(1993a) testis relevantwhenrelatively testproposedappliesforrelativelymanylags. In addition,thepresenttestis builton Skaug and Tj0stheim's(1993a) approach,and one mayperformbetterin some situationswhilethe betterin othersituations.Simulationshowsthatthenewtesthas good power otherperforms againstlinear- bothshortand longmemory- processesand somenon-linearprocesses,but like othertestsbased on the empiricaldistributionfunctionsit has relativelylow power against Engle's (1982) autoregressiveconditionalheteroscedasticprocess. For this class of tests(e.g. Skaug and Tj0stheim(1996)) maybe expected smootheddensity-based alternatives, to performbetter. Section 2 introducesthe test statisticand derivesits asymptoticnormality.Section 3 establishesconsistencyagainst all pairwisedependencesand derivesan optimalweighting schemeforvarious lags. In Section 4, we use simulationmethodsto compare the new test withSkaug and Tj0stheim's(1993a) test,and Skaug and Tj0stheim's(1996) density-based sketchedin AppendixA; detailsare available fromtheauthoron test.The proofsare briefly request. This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions PairwiseSerialIndependence 431 2. The approach and test statistics Consider a real-valuedstrictlystationarytime series {X,}I-,? with marginaldistribution Fj(x,y) = P(X, < x, X,- <, y), (x,y) E R', function G(x) = P(X, < x) and pairwisedistribution Suppose thata randomsample{X,},= of size n is observed.We are interested j = 0, 1,. in testingthe null hypothesisthat {X,} is seriallyindependent. Existingtests for serial independenceare all based on some measure of dependence. Because X, and X,< are independentifand onlyifg(X,) and g(X,-) are independentforany continuousmonotonicfunctiong, it is desirableto use a measurethatis invariantto g (see discussion).Hoeffding(1948) proposedthemeasure Skaug and Tj0stheim(1996) forfurther D2fj)=J {F)(u, v) - G(u) G(v)}2 dFj(u, v). (1) By Hoeffding's(1948) theorem3.1, transformation. This is invariantto any orderpreserving D2(j) = 0 ifand onlyif functions, density whenFj(u,v) has continuouspairwiseand marginal consistentagainstall are measure (1) In on based this case, tests X, and X,< are independent. is thatD2(j) = 0 but it If possible however, is discontinuous, pairwisedependences. Fj(u, v) In thiscase, for an example. 548, p. Hoeffding (1948), See X, and X,< are not independent. all dependences. pairwise against not consistent on measure are (1) testsbased Hoeffding (1948) used a UMeasure(1) or itsanaloguehas been used to testindependence. betweentwoidentically statistic estimatorforan analogueof measure(1) to testindependence randomvariables.Blum et al. (1961) used an empiricaldistridistributed and independently estimatorforan analogue of measure(1) to testindependenceamong butionfunction-based distributedrandom vector.Also see the componentsof an identicallyand independently Deheuvels(1981) and Carlstein(1988). For j = 0, 1, . . . define the pairwise empirical distributionfunction of (X,, X,<) Fj(x, y) = (n _])' , 1(X,< x) 1(X,< < y), I=j+l where1(.) is the indicatorfunction.A consistentestimatorformeasure(1) can be givenby (j)= ( _j)_ , 1=1+1 {F(X,, X,<)-F (X,, oo) Ft(oo,X,..) }2. (2) Skaugand Tj0stheim(1993a) werethefirstto use measure(2) to testforserialindependence. Theyconsideredtheteststatistic STI a= (n - (3) 1) Fn(j). j=1 and identicallydistributed, When {X,}InAare independently ST I a ZE XZJ Aij 00 00 i=1j=1 (P) x2 randomvariableswithp degrees in distribution as n -* oo, wheretheX (p) are independent random For continuous variables,Aij= (ij2)-2 , and the the are of freedomand weights. A, the discrete random free.For variables, Ai dependon thedata-generating testis distribution free. distribution test is not processand mustbe estimated;the extensionof Box and Pierce's natural Statistic(3) can be viewedas an appropriateand This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 432 Y. Hong (1970) correlationtest.It can detectpairwisedependencesup to lag p, includingthosewith Some evidence(Skaug and Tj0stheim(1993a), p. 600) suggeststhatthis zero autocorrelation. approachthattestsjointindependence pairwiseapproachhas betterpowerthanan alternative of (X,-1,. . ., X,_p).It also gives a much simplerlimitdistributionthan the joint testing assetofstatistic approach.Skaug and Tj0stheim(1993a) documentedthatthemostsignificant (3) is that it is very close to correlationtests in power against linear processes and is considerablybetterthancorrelationtestsin poweragainsta varietyof non-linearprocesses, although it has relativelylow power against Engle's (1982) autoregressiveconditional process. In addition,the size of statistic(3) is quite accurateformoderate heteroscedastic sample sizes comparedwithexistingsmootheddensity-basedtestsfor serial independence, especiallywhenp is small. For largep, Skaug and Tj0stheim(1993a) pointedout that the lead to a testwitha levelthatis too high criticalvaluesofstatistic(3) progressively asymptotic methods,advocatedby Skaug and as p increasesforsmalln. The bootstrapand permutation tests,can be Tj0stheim(1993b, 1996)to producethecorrectlevelsforsmootheddensity-based applied to statistic(3) as well. theorythatgivesa simpleand reasonableapproxIn thispaper,we developa distribution imationforstatistic(3) whenp is large.We show thatforlargep it is possibleto obtain an forstatistic(3), afterproperstandardization.However,forpower, N(0, 1) limitdistribution we maywantto checkmorelags as n increases.This ensuresconsistencyagainstall pairwise weightscan be givento dependencesforcontinuousrandomvariables.In addition,different i.e. to different lags. In particular,more weightscan be givento more recentinformation, lowerorderlags. This may improvepoweragainststationarytimeserieswhose dependence decays to zero as the lag increases.These considerationssuggestthe statistic = n-I k -_j) 23(j), k2(j/p)(n j=l (4) the followingassumption. wherek is a kernelfunctionsatisfying continuousat 0 and all excepta finitenumber 1. k: R -+ [-1, 1] is symmetric, Assumption < = dz oo and < of points,withk(O) 1, Jfok2(z) Ik(z)l CIzI-b as z -* oo forsome b > ^ and 0 < C< oo. This assumptionhelpsto ensurethatstatistic(4), afterdivisionby n,convergesin probability a consistenttestagainstall pairwisedependencesforcontinuous to Zj??1D2(j), thusdelivering of k withk(O) = 1 ensuresthatthebias of statistic(4) randomvariables.Here,thecontinuity vanishes.The conditionJork2(z)dz < oo impliesthatk(z) - 0 as z -* oo. This ensuresthat Daniell,Parzen,quadBartlett, varianceofstatistic(4) vanishes.The truncated, theasymptotic raticspectraland Tukeykernels(e.g. Priestley(1981), pages 441-442) all satisfyassumption weights.Note that all n - 1 lags 1. Except for the truncatedkernel,all have non-uniform are used in statistic(4) if k has unboundedsupport. Statistic(4) is a weightedsum of von Mises statistics.When the truncatedkernel(i.e. k(z) = 1(IzJ< 1)) is used, we obtain ST2a = L (n-]) n( j) (5) j=l whichis asymptotically equivalentto statistic(3). Thus, we can see thatboth statistic(3) and statistic(5) are based on the truncatedkernelor uniformweighting.The use of n -j in statistic(5) is similarin spiritto Ljung and Box's (1978) modifiedversionof Box and Pierce's This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions PairwiseSerialIndependence 433 (1970) test.Our simulationstudylatershowsthatstatistic(5) has bettersize thanstatistic(3) forlargep in small samples. We also considera leave-one-outempiricaldistribution function (XS< E ,s#t s=j+l Fj(,)(X,,X,-j)=(n-j-l)-' < X,<j), X1)l(XS_ wherethe sum excludesthe tthobservation.This yieldsthealternativestatistic n-2 V* = S:k j=I (n -j -1)n 2(jlp) (6) 2(j), where b*2(j) = (n _j)_ {t](,)(X,,K,.)- F()(X,, co)FP(,)(oo, X_j)}2- t=j+l but asymptoticanalysis Both statistic(4) and statistic(6) have the same limitdistribution, shows that statistic(6) has a smallerapproximationerrorfor the limitdistribution. Thus, in our simulationstudy. statistic(6) may give bettersizes in small samples,as is confirmed Similarly,we can consider STlb =(n - ST2b = (n -j E j=l l) j=1 - (7) n)2() (8) 1) Jn*2(j) as statistics(3) and (5). These two testshave the same limitdistribution We firstgiveour genericteststatistics. Theorem1. Suppose that assumption(1) holds, and p 0 < c < oo. Define n-In2 = cnV for some 0 < v < 1 and 2o E k 4( M =Zk2(J ) {(2Inj)b2(j)AO}/A k4(i;)} Mb = Zk {(n--)2b*2(j) (ip) where Ao= [n A0= n (n-2 Om (X,){1- n Z VE[{min(X,, t=l s=1 ) Xs)}-G(X,) (Xs)] \2 with (u) = n-1E 1(X,< u). t=1 This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions , 434 Y. Hong If {X,}'I are identicallyand independently then Ma distributed, - and Ma -* N(O, 1) and Mb -* N(O, 1) in distribution. Mb -* 0 in probability, Theorem 1 applies to discrete,continuousor mixeddistributions. If G(x) is knownto be continuous,we can use the followingsimplified statistics. Theorem2. Suppose thattheconditionsof theorem1 hold. Define Ma = 90 E k2Q) Mb= 90Z k2 (J){(n {(nD-j)J 3(j) _j l}/{2 - 1) f)*2( j) 2 k4(J ) , k j /2 _ }1/ If {X,}', are identicallyand independently distributedcontinuousrandomvariables,then Ma - Mb -* 0 in probability, and Ma -* N(0, 1) and Mb -* N(0, 1) in distribution. We note that theorem2 does not apply to discretedistributions, because the mean and asymptoticvarianceof (n -j) DnL ) cannotbe computedin discretecases. Althoughthevon Mises statistic (n -j) Dn(j) does notfollowan asymptotic x2-distribution, theorems1 and 2 showthata weightedsum of (n _j) i52 (j) has an asymptotic N(0, 1) distributionforlargep aftercentring and scaling.To obtainan intuitive idea, considerforexample theuse of thetruncatedkernel.In thiscase, Ma of theorem2 becomes Ma = E {(n-_j) 2l(j)-3 }/ 2 ) As shownin Skaug and Tj0stheim(1993a), (n-j) 12(j) E k=1 1=1 (kir)2(lr)2 X21(1) in distribution as n -* oo, thushavingmean 1/36 and asymptoticvariance2/902.In addition,cov{(n i)12(i), (n -_j) J2(j)} -_ 0 for i 7'-j as n -* oo, suggestingthat(n - i)13)(i) and (n -j) I5 (j) are asymptotically uncorrelatedor independent.Therefore,{ (n -j) 32(]j), j = 1, . . ., p} can be viewedas an asymptotically identicallyand independently distributed sequencewithmean 1/36and variance2/902.The sum of thissequence,afterdifferencing the mean and dividingby the standarddeviation,will convergeto N(0, 1) in distribution as p becomeslarge.Our proof,ofcourse,does not dependon thissimplisticintuition.Instead,we developa dependentdegenerateV-statistic projectiontheoryto approximatestatistic(4) as a weightedsum of degenerateV-statistics overlags,which,afterstandardization, is thenshown to be asymptotically N(0, 1) by usingan appropriatemartingalelimittheorem(e.g. Brown (1971)). See AppendixA formoredetails.For degenerateV-statistics (or relatedU-statistics) of dependentprocesses,see (forexample)Carlstein(1988) and Sen (1963). The N(0, 1) approximation is convenient forinference. For smalln or smallp, however,itmay notbe accurate.As a practicalalternative, thebootstrapor permutation, as advocatedin Skaug and Tj0stheim(1993b,1996),can be used as a remedyforobtainingtherightlevel.Indeed,the currenttestsituationis ideallysuitedto theseresamplingmethods,whichcan be expectedto yieldas good a levelas thebestasymptotic In particular, approximation. permutation givesthe exactlevel.However,it is impossibleto computethelevelexactlyin practiceforall exceptvery smallsamplesizes.Hence,Monte Carlo methodsmustbe used to obtainan approximation. This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 435 Pairwise Serial Independence 3. Consistency conditions. To stateour consistencytheorem,we impose some additionalregularity Assumption 2. Jz12k2(z)dz < oo. 00 0 Assumption3. Put U, = G(X,), where G is the continuousmarginaldistributionof the strictlystationarymixingprocess {X,} with strongmixingcoefficient a(j) = 0(yj--6) as -* The of oo for some 6 0. distribution has continuous > a joint (U,, U,1) joint density j functionon [0, 1]2 thatis bounded by some A < oo, whereA does not dependon j. Theorem3. Suppose that assumptions1-3 hold, and p = cnVfor some 0 < v < 3 and 0 < c < oo. Let Ma and Mb be definedas in theorem2. Then,withprobability approaching1 as n -+ oo, pl2nMD(j)/{2J (P'/2 /n)Mb 00 k4(z)z 90 E D2(j) 1/2 /fg 00 2 k4(z)dz) Theorem 3 impliesthat limn,o{P(Ma > C,G) = I for any non-stochasticsequence {Cn= of Ma againstall pairwisedependencesforcontinuous o(n/pl/2)1, thusensuringconsistency randomvariables.It suggeststhat whenmore data become available the testsMa and Mb Thus, Ma and Mb are useful largerclass of alternatives. have poweragainstan increasingly about possible alternatives.Because negativevalues of whenwe have no priorinformation Ma and Mb are oneMa and Mb can occur onlyunderthe null hypothesisasymptotically, of sided tests;appropriateupper-tailedcriticalvalues N(0, 1) should be used. Althoughthe temporalconditionon a(j) is mild,it rulesout strongdependences.However,it is reasonablethat any independencetestshould be expectedto have strongpower againststrongdependences.In particular,Ma and Mb may be expectedto have good power againststrongdependencesbecause a long lag is used. We shall investigatethepowerof Ma and Mb againstlongmemoryprocessesvia simulation.It shouldalso be notedthattheorem3 holds onlyforcontinuousrandomvariables.For discreterandomvariables,Ma and Mb are notconsistentagainstall pairwisedependences,because D2(j) can be 0 evenifthetimeseries like otherasymptoticnotions,consistency against is not pairwiseindependent.Furthermore, all pairwisedependencesseems to be mostlyof theoreticalinterest.In practice,p is always finitegivenany samplesize n. When a kernelwithboundedsupport(i.e. k(z) = 0 if lzl > 1) is p lags are tested.Whena kernelwith used,p is thelag truncationnumber,and onlythefirst unboundedsupportis used,p is not a lag truncationnumber.In thiscase, all n - 1 lags are fromlags muchlargerthanp are negligible. used, but thecontributions An important forstationary timeserieswhosedependence issueis thechoiceofk. Intuitively, to give as measuredby statistic(1) decaysto zero as thelag increases,it seemsmoreefficient moreweightsto lowerorderlags. The asymptotic analysisbelow showsthatthisis indeedthe relativeefficiency betweentwokernels,say k, and k2,we use case. To comparetheasymptotic whichis suitableforlargesampletestsunderfixed Bahadur's(1960) asymptotic slopecriterion, alternatives. Bahadur'sasymptotic p-valueof thetest slope is therateat whichtheasymptotic This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 436 Y. Hong statistic its goesto 0 as n -* oo. BecauseMa is asymptotically N(0, 1) underthenullhypothesis, asymptotic p-valueis 1 - '1(Ma), where 1 is thecumulativedistribution functionof N(0, 1). Define SA(k) = -21n{1 Given ln{1-?_(z)} =I -Iz2{1 + o(1)} as z (p/n2)Sn(k)-* 902{ - -* oo, (M,) we have by theorem3 that D2(j)} /{2 J k4(z)dz} in probability.FollowingBahadur (1960), we call 902{ 5j??D2(j)}2/2 Jo k4(z)dz theasymptotic slope of Ma. Under the conditionsof theorem3, it is straightforward to show that Bahadur's asymptoticrelativeefficiency of k2 to k, is ARE(k2:kl) = { 00 dz/ k14(z) 00 o 1/(2-v) k2(z)dz} Thus, k2 is moreefficient thank, if Jk4(z)dz< J 4(Z)dz 2 o For example,Bahadur'sasymptotic efficiency oftheBartlett kernel(kB(z) = (1-IZI) (IzI <, 1)) to thetruncated kernel(kT(z) = (IzOI< 1)) is ARE(kB :kT) = 511(2-v)> 2.23. Followingreasoningthatis analogousto Hong's (1996a,b) local poweranalysisforsomecorrelation tests,we can obtainthattheDaniell kernel k(z) = sin(rzV/3) rz /3 ZElR (9) maximizesthe Bahadur asymptoticslope over theclass of kernels K(r) = {k satisfiesassumptions1, 2, k(2)= r2/2> 0, K(A) > 0 forA E R}, (10) where k(r)= lim[{I - k(z)}/IZIr] is called the 'characteristic exponent'of functionk and K(A) = (2r)-' J k(z) exp(-iAz) dz is the Fouriertransform of k. This class includesthe Daniell, Parzen and quadraticspectral kernelsbut rulesout the truncated(r = oo) and Bartlett(r = 1) kernels.Mainly because of assumption2, K(r) is more restrictive than a class of kernelsoftenconsideredto derive optimalkernelsusingappropriatecriteria(see Andrews(1991) and Priestley(1962)) in the spectralanalysisliterature. Hong (1996a, b) showedthattheDaniell kernel(9) maximizesthe local powerof some correlationtestsovera class of kernelsslightly moregeneralthanK(T). The resultobtainedheresuggeststhattheoptimalityof theDaniell kernel(9) carriesoverto thenon-linearHoeffding measure(1). However,theoptimality of theDaniell kernel(9) seems to be only of theoreticalinterest.Simulationstudieslater show that commonlyused non- This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions PairwiseSerialIndependence 437 uniformkernelshave similarpowersforMa and Mb, but theyare all morepowerfulthanthe truncatedkernelin mostcases. One advantageof Skaug and Tj0stheim's(1993a) testis thatit does not requirechoosing any kernelfunctionand p as a functionof n. For Ma and Mb, p behaveslike a smoothing parameteras we requirep -+ oo as n -+oo to ensureasymptoticnull normalityand confromsmoothed sistencyagainstall pairwisedependences.Of course,Ma and Mb are different estimationfor testsbecauseMa and Mb do notrequiresmoothednonparametric density-based testwould also need to each lag. To testall pairwisedependences,a smootheddensity-based letp -> oo as n -+ oo, in additionto theoriginalsmoothingparameterfordensityestimation. In thepresentcontext,the optimalp forMa and Mb dependson thealternativedependence structure.Skaug and Tj0stheim(1993a) noted that choosingp too large will decrease the powerof statistic(3). The same is trueof Ma and Mb as wellin general.Skaug and Tj0stheim lag includedinthemodel. choosingp largerthanthesmallestsignificant (1993a) recommended It is reasonableto use a 'ruleof thumb'consistingof takingp equal to thelargestsignificant lag includedin the model. Of course,thisrule may not make sense forcertainalternatives of thechoiceofp on bothsize theeffect dependentprocesses.We investigate suchas strongly and powervia simulation.Our simulationfindsthatthisrule of thumbapplies well to the kernelsmaximalpoweris oftenachievedforp larger truncatedkernel,but fornon-uniform lag includedin themodel.In addition,theuse of thenon-uniform thanthelargestsignificant kernelalleviatesthe loss fromchoosingp too largebecause the kerneldiscountstheloss of degreesoffreedomforhigherorderlags; thismakesthepowerless sensitiveto thechoiceofp. We deferthiscomplicatedissue theissueofchoosingan optimalp is important. Nevertheless, to otherwork. 4. MonteCarlo evidence of thetestsproposed.In additionto an identically We now studyfinitesampleperformances AR(l), distributed and independently process,we also considerfollowingalternatives: X, = 0.2X,_1+,El; ARFIMA(0, d, 0), X, =(1-L) 026,, LX, =X,; ARCH(l), X, = ,(l ?0 5X2 )1/2; GARCH(1, 1), h, = 1 +O.1X,2_ + 0.8h,l1; X, = c,h1/2, TAR(l), X, = (-0.5X,_ if X,1 > 1, + c, if X,_1 < 1; 0.4X,t1+c,, NMA, X, = 0.8,1E,-2 + El; This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 438 Y. Hong NAR(3) X, = 0.25X,2 - 0.4E,_1X,I + 0.2ct-2Xt-2 + 0.5Eu.3X,_3 + E,; NAR(5) XI = O.3(ct,1Xt-1-Et -2 + c,.3X,3 -c,4X,.4 -c,_5X,.5) + l,; EXP(3) 3 X,=0.2 Z X1j exp(- IX2 )+E,; i=l EXP(1 0), 10 X,= 0.8 E X,< exp(- -X2 ) + ,. j=, Here, AR(1) is a first-order autoregressive process,ARFIMA(0, d, 0) is a fractionally differencedintegratedprocess,a popular long memorymodel (see Robinson (199ib, 1994)), ARCH(1) is a first-order autoregressive conditionalheteroscedastic process,GARCH(1, 1) is a generalizedARCH processof order(1, 1), TAR(1) is a first-order thresholdautoregressive process, NMA is a non-linearmoving average process, NAR(3) and NAR(5) are nonlinearautoregressive processesof orders3 and 5 respectively, and EXP(3) and EXP(10) are non-linearexponentialautoregressive processesof orders3 and 10 respectively. These models are a fairlyrepresentative selectionof both linearand non-lineartimeseries.Exceptforthe ARFIMA(0, d, 0) process,all have been used in a varietyof existingsimulationstudiesfor independencetests.We considertwo typesof c,: normaland log-normalwithzero mean and unit variance,forn = 100 and n = 200. The process ARFIMA(0, d, 0) is generatedusing Davies and Harte's(1987) fastFouriertransform algorithm.ExceptfortheARFIMA(0, d, 0) process,we generatedn + 100 observationsforeach n and thendiscardedthefirst100 to reduce the effectof initialvalues. For the ARFIMA(O, d, 0) process,we generatedn + 156 = 28observationsand discardedthefirst156forn = 100,and generatedn + 312 = 29observations and discardedthe first312 forn = 200. The simulationexperiments were conducted usingthe GAUSS 386 randomnumbergeneratoron a Cyrixpersonalcomputer. To examineeffects ofvariousweightson thetestsproposed,we consideredfivekernels:the truncated,Bartlett,Daniell, Parzen and quadratic spectralkernels.The last threekernels belongto theclass K(er/V/3) in expression(10). To investigate theeffectof choosingdifferent values of p, we consideredp from1 to 20; this covers a sufficiently large rangeof p given n = 100 and n = 200. We also studied Skaug and Tj0stheim's(1993a) test STia and its variantsSTib, ST2a and ST2b, as well as Skaug and Tj0stheim's(1996) smootheddensitybased test p J= J,i(), i=l where This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions Pairwise Serial Independence . 0.30 . .,.. 0.30 0.27 0.27 0.24 0.24 ST1a ST1b, 0.21 0.21 ,Ma-Dan 0.18 STa 0.18 Ma-Trun STbb 01 0.15 0.06 439 SMb-Trun 0ST2c,2 0.03 0.09 P P (a) 0.16 0.20 . '. . . ' . 'M'D / 0.20 0.18 0.18 0.14 0.14 STia STlb 0.12 / 0.12 Ma-Dan 0.10 (b) 0.16 'Ma-Trun o.1o 0.08 ST1aa Mb-Dan STIb' ,Mb-Trun ?a 0 Ma-Trun 0.06~~~~~~~~~~~~S2 0.04 ,=1 o 0o0 1 b .0 0.0 " Mb-Trun ST2b, 0.02 0.2 0.00 o ST2aoS 3 5 J-Test 0.02 00 7 9 I1 1 13 15 17 19 0.00 1 3 5 7 9 1 13 I1 p p (c) (d) 15 17 19 Fig. 1. Size underthe normalprocess: (a) empiricalsize at the 10% level; (b) bootstrapsize at the 10% level; (c) empiricalsize at the 5% level; (d) bootstrapsize at the 5% level = (n - })f,1Qj) S ii {Jj(1)(X1, X,..1)-J,()(X,)A,.. X,..) = {(n -j)hn}-lt=j+l~~~~~~~t fJ(,)(X,, E K{Xs=j+I f(l)( = s$I { (n - I)hn } E s=I,sot J)(X,..J)}, X)/h,1} K{ (X,..1-X)/,} K{ (X,-Xs)/hn } withK: R -R+ a kernelfunctionand hna bandwidth.Here, 'leave-one-out'bivariateand are used. As in Skaug and Tj0stheim(1996),we firststandardized marginaldensityestimators This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 440 Y. Hong 10 r .0 0.8 0.8 0.6 ,Mo-Don 0.6 ,,Mo-Trun ,Mo-Dan , ,Mo-Trun 0.404 o.2 0.2 3 1 5 7 11 9 i3 15 17 1 19 3 5 7 9 p 11 13 15 17 __ _Mo- Don 19 p (a) (b) 1.0 1..... Don ~~~~~Mo- > 0.404 o.o .T2 .. . 3 5 7 . . 11 13 o. . 0.202 1 9 1S 17 19 1 3 5 7 9 11 p p (c) (d) 13 15 17 19 power, AR(1), (a) size-corrected d, 0) processes: Power at the 5% level under AR(1) and ARFIMA(O, power, AR(1), log-normal error; (d) error; (b) bootstrap power, AR(1), normal error; (c) size-corrected d, 0), normal error; (f) bootpower, ARFIMA(O, bootstrap power, AR(1), log-normal error; (e) size-corrected d, 0), log-normal error; (h) power, ARFIMA(O, d, 0), normal error; (g) size-corrected strap power, ARFIMA(O, d, 0), log-normal error bootstrap power, ARFIMA(O, Fig. 2. normal thedata by thesampledeviationand used h = n-116withthe Gaussian kernel.This testhas been provento be morepowerfulthanor comparablewithmanyexistingsmootheddensitybased testsagainst a varietyof non-linearprocesses(see Skaug and Tj0stheim(1993a, b, as n -+ oo forsome a 2 -+ N(O, 42) in distribution (n/p)l12J 1996)). Under independence, we also used a bootstrapprocedureforall theteststo In additionto asymptoticinference, studybootstrapsize and power.To describethe procedure,we firstconsiderMa. Given a sample {X,},n1, we generate m bootstrap samples {X*jL'1, 1= 1, . . ., m. For each bootstrap This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions Pairwise Serial Independence t.0 , , . . 0.8 . . .Ma- Don 1.0 I 0.8 M -Dan ,Mo-Trun , -Ma-Trun 0.6 -- , 441 I STlo ,'ST1oaST T Test 0.404 0.2 0.2 0.0 1 3 . 7 5 11 9 13 15 17 19 0.0 1 3 5 7 9 13 15 17 19 17 19 (f) (e) , . 1.0 11 p p , 1.0 0.8 , ,. . 0.60. " \Mo-Trun \STlao 0.4 0.0 *1 "0ST2 0ST2.4 //J-Test 0.2 'Mo-Trun ST a -JTs /0.2 3 5 7 11 9 13 15 17 19 . * 3 5 7 11 9 p p (g) (h) 13 t5 Fig.2. (continued) ofthe function Mt*.To estimatethedistribution sample{X*/},1,wecomputea bootstrapstatistic if of function distribution the empirical {MII}I bootstrapstatisticM*a we could use I m is a = 50. For we choose m is rather simulation extensive, experiment large.Becauseour sufficiently the estimate to section and bootstrap 3.3, Tj0stheim(1996), sucha smallm,we followHjellvik kernelestimator function distribution F*(z) = P(M* < z) by a smoothednonparametric ft(z) = rsni J [rn-ijE {h,n,\/(2ir)F-' exp{-(y m FW=-00 /= 1/1hl - M* )2/2h, dy In = M-1E (D{ (z -M*)Ih l=l This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 442 Y. Hong , 1.0 , . 1.0 0.8 0.8 0.8 0.6 ~~~~, \ ,' ,Mo~~-Don \ 0.404 o.* \ \,J-Test \,,M-Don J-Test ,' , \ \< ,',~~Mo-Trun . ,',',ST1o ' ',, T20X o ,,',' ST2o 0.2 0.2 p p (a) (b) 1.0 , 1.0 0.8 0.8 0.6 0.6 , J-Test \ 0.4 ,J-Test Mo-Don 0.4 , ' , ',Mo-Trun \ , \ 3 S 7 9 ll ,', ,' , ',',ST1 a/ 13 15 17 19 , / / 0.2 02 ' 0.2 * \,,'Mo-Trun \,,',Stlo . 1 3 5 7 / / / ', 11 9 p p (c) (d) ' 13 Mo-Don Mo-Trun STio ST2a 15 17 19 Fig. 3. Power at the 5% level underARCH(1) and GARCH(1, 1) processes: (a) size-correctedpower,ARCH(1), normalerror;(b) bootstrappower,ARCH(1), normalerror;(c) size-correctedpower,ARCH(1), log-normalerror; (d) bootstrap power, ARCH(1), log-normalerror;(e) size-correctedpower, GARCH(1, 1), normal error; (f) bootstrappower,GARCH(1, 1), normalerror;(g) size-correctedpower,GARCH(1, 1), log-normalerror;(h) bootstrap power,GARCH(1, 1), log-normalerror of N(O,1), h,,,= Sn,,m"5is thebandwidth function distribution where'1 is thecumulative For moredetails,see Hjellvikand and S,A, is the samplestandarddeviationof {Ma*!}72= if ofindependence Tj0stheim (1996).HavingobtainedPn,wecan rejectthenullhypothesis based on the leveland Ma is theteststatistic F*t(Ma) > 1 - a, whereaois thesignificance is appliedto theothertests. The sameprocedure originalseries{XJ,},1. criticalvalues.For STia, and bootstrap both asymptotic sizes using consider by first We of values 10000simulations critical using their tabulate we asymptotic and ST2b, STIb,ST2a This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions Pairwise Serial Independence 1.0 - 1fX--rl.0 - - -s r--- 0.8 0.8 0.6 0.6 0.4 , , J-Test ,',Mo-Don ,',Mc-Trun , J-Test ,Mo-Dan Ma-Trun 0., ST2a 0.2 __: 0.0 1 _ 5 3 ~~~~ST2a 7 9 11 0.2 13 15 17 19 . 0fX(_.0._. OO_ ' 3 5 7 / 9 11 p 0.83 0.6 0.6 A-Test I ' ,' /,Mc04 / zSTlo/ 5 7 9 STia 0.2 * 1 5 Mo-Trun STlo , 3 19 , J-Test ,'MoDo ST2o~~~~~~~/ 0.0 17 1.0 . 0.81 0.2 . 15 (f) , 040g 13 p (e) 1.0 0 443 71 9 0. / 3 a5 p 13 17 15 17 1 P (g) (h) Fig.3. (continued) 200 200 EE i=l j~=] - 1)-I X2J(P)' a truncated versionofthelimitdistribution ofSTia. For theJ-test, we onlystudyitsbootstrapsize.Sizes forasymptotic and bootstrap criticalvaluesare based on 5000and 1000 replications respectively. Fig. I reports sizesofMa, Mb, STI a, ST Ib, ST2a,ST2bandtheJ-test underthenullhypothesis ofindependence at the10% and5% levels,forn = 100andnormal innovations. We onlyreporttheDanielland truncated kernels forMa and Mb,becausethe fournon-uniform kernels perform We havethefollowing similarly. observations. In (a) terms ofasymptotic critical values,Ma andMb havereasonable sizes,although they This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions Y. Hong 444 .0 . 1.0 0.8 0.8 0.6 0.6 ,Mo-Don , 0.2 0.0 5 3 1 | e i\ - ( ,'/,S2 15 I 17 19 0.0 t 3 5 7 J-Test 11 9 p / 13 ''S2 17 15 19 JTs (b) p s a 0. \\0 0.6 0.48STla 0.6 13 11 9 7 ,,'~~ p (a) ,l .e Mo-Trun Mo-Trun ' ' i p - ,',ST2o 0.8 _ nm1.0 M ,MoMa-Da . b e p M norma i .SW 0.6 (c)~~~~~~~~~~~~~~~~~~~~~~S I(d) error power,NMA,log-normal error;(h) bootstrap power,NMA,log-normal error;(g) size-corrected at the 5%/level.Both theDaniell and thetruncatedkernels show slightoverrejections have similarsizes.Sizes are robustto thechoiceofp. The leave-one-outversionMb has slightlybettersizes than Ma has, suggestingsome gain fromusing the leave-one-out statistics.The testsSTi a' ST2a, STi b and ST2b have precisesizes forsmallp. For large but themodifiedstatisticsST2a and p, STi a and STi b tendto givesome overrejection, ST2b continueto have good sizes. The leave-one-outversionST2b has slightlybetter sizes thanST2a. (b) All the testshave good bootstrapsizes. In particular,bootstrapsizes are betterthan This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions Pairwise Serial Independence 1.0 1.0 0.8 0.8 ~~~~~~~~~~~~~~~~0.8 0.6 A , ~~~J-Test 0-6 t/\ ,,~~~~Mo-Don \ Mo-Trun 0.4 ,J-Test 'MoDon Mo-Trun 0.4 0.2 0.2 0.0 445 1 5 3 13 11 9 7 15 17 19 0.0 1.0 1.0 0.8 0.8 \ Test ,J- _--~~ \ \ 0 4 \5 , __- ' - S o' ' 3 5 7 9 11 13 15 17 19 (0 (e) 0.6 o. 1 p p 0.6 0.6 ,STl1o ,J-Test ,,M-Don ,', Mo-Trun , ,Mo-Don , ' Mo-Trun o. ',,STl1a 0.2 0.2 oo~~~~~~~~~~~~~~~~~~ P~~~~~~~~~~~~~~~. o.o 0.0 Fig. 4. PL p p (g) (h) (continued) at the5% levelforall thetests. approximation thosegivenbytheasymptotic gives approximation We also considersizesat the1% level.At thislevel,theasymptotic to variousdegreesforall thetestsin mostcases. The bootstrapsizesare overrejections havingthebestsizeundernormalinnovawiththeJ-test andreasonable, better significantly sizesat all the haveinvariant function distribution tions.The testsbasedon theempirical For n = 200,we studiedsizes innovations. threelevelsunderbothnormaland log-normal for forall thetests,especially values.The sizesareimproved critical onlyusingasymptotic and ST2a. STIa powerat the5% levelforn = 100.To comparethetestson an equalbasis, Figs2-6 report of underthenullhypothesis from5000replications valuessimulated critical weuseempirical This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions Y. Hong 446 , , , , , 1.0 1.0 , , ., 0.8 0.81 0.6 - - ---J-Test SMa-Dan \ - , \ \,, 7 5 x . P 1a 15 Ma17 19 t error;(b) bootstrap 5% l , N u + 0.8- t 3 < < 7 5 11 / 9 ST2a15 13 17 19 ~~~~~~~~~~~~~~~~0.6 (b) (a) a 0.2 o4t p p 0.4 1. 11 9 _J-Test , Ma-Don - , -Mo-Trun , - /'STla 2 _ ~ _-- \ ,Mo-Trun ,',STla 4 < ) 0o2t 0.6 N ) nom-al N ; () s 1.0 p , ( e p r , n J-Test ~~~~~~~~~~~~~~~~0.2 0.0 (ecess:() lognoma ne h %leeerror; pi.5 ower,a NAR(3), normalero;f)btsapow,NA() power, NAR(5), size-corrected error log-normal power,NAR(5), error;(h) bootstrap log-normal power,NAR(5), error;(g) size-corrected normal independence.We also considerpower by using bootstrapcriticalvalues. Powers by using empiricaland bootstrapcriticalvalues are based on 500 and 200 replicationsrespectively. a werealso studied,and similarrankingpatternswerefound. Powersat the10% and 1 levels Both Ma and Mb have similarpowers.Also, STia and ST2a have powersimilarto STib and For clarity,we only reportresultsfor M STi ST2 and the J-test. ST2b respectively. and thetruncatedkernelsforMa are reported,because thefournonDaniell the only Again, powers. (In some cases, the Bartlettkernelis slightlymore similar have uniformkernels kernels,but this is not inconsistentwith the non-uniform three powerfulthan the other This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions Pairwise Serial Independence 1.0 I .0 447 I. ,J-Test , ,Ma-Dan 1,J-Test ,,',a-Dan 0.4~~~~~~~~~0 o.o 1 ST. .. .ST..l.o. 3 5 7 9 11 13 15 17 1 19 3 5 7 . , os * * * , . . ,M-TDon ? | . . . . Fig. 5. 19 13 1S 17 19 ,. | oo0.6 . ~~~~~~~~~~~~~~~~0.4 0.4 1 17 * 1.0. ~ ~ Ma- Dan < .o o.o __- k 15 (f) X ~\ 13 p (e) , 1.0 11 9 p 3 5 7 9 11 13 15 17 19 *1 3 5 7 9 11 p p (g) (h) (continued) of theDaniell kernelbecause theBartlettkernelis outsidetheclass of kernelsover optimality whichtheDaniell kernelis optimal.)We can make thefollowingobservations. (a) For Ma, the Daniell kernelis more powerfulthan or comparablewiththe truncated weighting.For the kernelin mostcases, suggestingthe gains fromusingnon-uniform alternativesNAR(3), NAR(5), EXP(3) and EXP(10), the truncatedkernelis more powerfulthan or comparablewiththe Daniell kernelfor smallp, but as p becomes largetheDaniell kernelbecomesdominant.The testsSTIa and ST2a have roughlythe same power as the truncatedkernel-basedMa-test.(The truncatedkernel-basedMatest and ST2a have identicalpower because thereis an exact relationshipbetween them.) This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 448 Y. Hong 1.0 . . . . 1.0 0.8 0.8 , Ma-Dan , ,Ma-Trun 0.6 , ,Ma -Dan Ma -Trun 13 17 0.6 0.404 0.202 1 3 . 1.0 os5 5 . h 7 9 I I11 13 is 17 19 5 *1 9 p p (a) (b) , . * * 1.0 * 010.8 \ % \Ma-Don \ 0.6 ,Mo-Don ',Ma-Trun \Ma-Trun 0.0 *1 1 o / ',',ST2o 0.6 % Ma0 0.4 0.4 0.2 21 . 20 0, 3 5 7 II 9 13 15 17 19 . *1 , _ 5 9 13 p p (c) (d) 17 21 Fig.6. Powerat the5% levelunderEXP(3)and EXP(10)processes:(a) size-corrected power,EXP(3),normal error; (b) bootstrap power,EXP(3),normal error; (c) size-corrected power,EXP(3),log-normal error; (d) bootstrap power,EXP(3),log-normal error; (e) size-corrected power,EXP(10),normal error; (f)bootstrap power,EXP(10), normal error; (g) size-corrected power,EXP(10), log-normal error; (h) bootstrap power,EXP(10), log-normal error (b) For the alternatives AR(1), TAR(1), NAR(3), NAR(5), EXP(3) and EXP(10), STia, ST2a and thetruncatedkernel-basedMa-testachievemaximalpowerswhenp is equal to thelargestsignificant lag includedin themodel. This is consistentwiththe findings of Skaug and Tj0stheim(1993a). For the ARFIMA(O, d, 0) process,STia, ST2a and the truncatedkernel-basedMa-testachieve maximal powers at p = 1 under normal innovations,but theirpowersgrowwithp underlog-normalinnovations.In contrast, the Daniell kernelgenerallyachieves the maximalpower whenp is largerthan the largestsignificant lag. This is trueevenfortheAR(1) process.A possiblereasonis that This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions PairwiseSerialIndependence 0.8 0.8 0.6 0.6 0.4 04MD \XSTl a \ST2a \XSTl a \ST2a 5 3 7 .. 9 - J T s Te t0.2 ------------ 0.2 0.0 * .I 13 II I __ .0 17 t5 I 19 1 3 5 7 9 p * , * * * . 17 19 1.0 . 0.8 0.6 0.6 0.6 ; = 0.40.4 \ s s\"Mo-Dan s \\ Ma-Trun \\STl1a | [ . . 04S2 -J-Test 0.2 Fig. 6. I 15 (f) * 0.8 0.0 0.03 13 II p (e) 1.0 449 S (continued) 7 11 9 13 15 ~~~~~~~~~~~,',Mo-Trun , , 'STlo ''S2 ',J -Test 0.2 17 19 0.C, 0.03 5 7 tl 9 p p (g) (h) 13 15 17 19 D2(j) may not vanish whenj is largerthan the largestsignificant lag, althoughits magnitudemay be smaller.Thus, to include some extra termsbeyond the largest significant lag may enhancethepowerifD2(j) is sufficiently largeto offsetthe loss of additionaldegreesof freedom.For the ARFIMA(O, d, 0) process,the Daniell kernel has the maximalpower whenp = 3 under normal innovations,and its power also growswithp underlog-normalinnovations.Note thattheDaniell kerneloftenmakes power less sensitiveto the choice of p, because it discountshigherorderlags, which usuallycontributeless in power. (c) None of the testsuniformly dominatesthe othersagainst all the alternativesunder study.The J-testis morepowerfulagainstARCH(1), GARCH(1) and NMA processes withnormalinnovations,whereasthetestsbased on theempiricaldistribution function This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 450 Y. Hong are more powerfulthan the J-testagainst the AR(1), ARFIMA(O, d, 0), TAR(l), EXP(3) and EXP(10) processes.For NAR(3), NAR(5) and NMA with log-normal innovations,theJ-testis morepowerfulwhenp is small,whereasthetestsbased on the functionare more powerfulforlargep. It seems thatthe tests empiricaldistribution functionare morepowerfulagainstthealternatives based on theempiricaldistribution whosedependencesare mostlycapturedby theconditionalmean,whereastheJ-testis whose dependencesare mostlycapturedby the morepowerfulagainstthe alternatives powers conditionalsecond moments.Almost all the testshave bettersize-corrected under log-normalinnovationsthan under normal innovations;only the J-testhas betterpowers against the ARFIMA(O, d, 0) and ARCH(1) processesunder normal innovationsthanunderlog-normalinnovations. (d) The bootstrappowersforall thetestsare similarto thoseusingempiricalcriticalvalues is small. undernormalinnovations.In thiscase, theloss ofpowerdue to bootstrapping For log-normalinnovations,however,the bootstrappowersare substantiallysmaller powersin quite a fewcases. An exceptionis the J-testagainst thanthe size-corrected the ARFIMA(O, d, 0) process,whose bootstrappower is much betterthan its sizecorrectedpower. powerswereconsidered.All the testshave betterpowers For n = 200, onlysize-corrected to variousdegrees,buttherelativerankingsremainunchanged.The againstall thealternatives functionstillhave relativelylow powersagainstthe testsbased on theempiricaldistribution ARCH(1), GARCH(1, 1) and NMA processes.However,thisdoes notmeanthattheyshould it is If interestindeedis in detectingthesealternatives, not be used to detectsuchalternatives. functionto the square of the sensibleto apply the testsbased on the empiricaldistribution originalseries.This can be expectedto yieldgood poweragainstARCH-type alternatives. Acknowledgements I would like to thanktwo referees,the JointEditor,N. Kieferand T. Lyons for helpful of thispaper. improvement commentsand suggestionsthatled to a significant Appendix A constant finite 0 < A < oo denotesa generic 1-3 here.Throughout, We sketchtheproofoftheorems in different places. thatmaydiffer A.1. Proofof theorem1 1-5 1is basedon propositions Theproofoftheorem WeproveforMa only;theproofforMb is similar. areavailablefromtheauthoron request. statedbelow.Theproofsofthesepropositions 1 showsthattheweighted sumofvonMisesstatistics V, in equation(4) can be approxProposition relatedto variouslags. V-statistics sumofthird-order imatedbya weighted Proposition1. Put h(x, y) = I (x < y) - G(y). For 0 < j fI(j) = (n -])' < n, define E Af](XI, XIj), I=j+l where 11j(x,y) = (n -])' n E t=j+l h(X,,x) h(X,1, y). This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions Pairwise Serial Independence Then n-I Z k2(j/p)(n -j) j=l DJ2(j)= n-I ?,k2(j/p)(n -j) j=lI 451 fIt(i) + Op(p/n'12). The nextresultis a projectiontheoryfordependentdegenerateV-statistics. Proposition2. For 0 < j < n, define y) dG(x) dG(y), ftj2(x, = fi2(j) wherefIn(j)(x,y) is definedin proposition1. Then n-I n-I E k2(j/p)(n -j) fI2(j) = E k2(j/p)(n -j) FI2(j) + Op(pln'2). J To statesubsequentresults,put A,,(j) = h(X,,x) h(Xs,x) h(X,1, y) h(X51, y) dG(x) dG(y). Note thatA,s(j) = As,(j). In addition,E{A,s(j)[F,_} = 0 almostsurelyforall t > s and allj > 0, where and hereafter{F,} is the sequenceof a-fieldsconsistingof {X,, -r< t}. We can now write n-1 jp Zk2(]/p)(n _j)f2(j)= j=l n-2 n 2jl k2(ip)(n-j)1 -j)-< E A1(j) +2 n-1 jp Zk2(/p)(n I j=l I=j+1 n 1-1 t=j+I s=j+l As(j) =An + 4n,say. term. Proposition3 showsthatAuican be approximatedby a non-stochastic Proposition3. LetA^nbe definedas above, and define AO= [JG(x) { 1-G(x) } dG(x)1 Then n-I An =-AOE k(j/p) j=l The statistic4nis a weightedsum of degeneratesecond-orderU-statisticswith mean 0. By rearrangingthe summationindices,we can write n f -1 s-1 (n _j) AI,(j) -2k Bn= Z I 1=3 tvs=2j=1 { sequence. Because E{A,s(j) JF,_ I} = 0 forall t > s and all j > 0, Anis a sum of a martingaledifference By applyingBrown's(1971) martingalelimittheorem,we can showthata properlystandardizedversion to N(O, 1), as statedbelow. of Bn convergesin distribution Proposition4. Define Bo = (|[G {min(x,y)} - G(x) G(y)]2dG(x) dG(y)) Then {2Bo n-2 k4(j/p)} ~~~1/2 4n N(O, 1). Both Ao and Bo mustbe estimatedif we do not know whetherG(x) is continuousor discrete.The followingresultshowsthatAoand Bo definedin theorem1 are consistentforAo and Bo. Proposition5. LetAOand Bo be definedas in theorem1. Then This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions 452 Y. Hong AO- AO = Op(n 1/2), ho- Bo = op(l) Now, combiningpropositions1-3 yields n-I n-I j=l j=1 E k2(j/p)(n-j) DJn(j)= Ao E k2(j/p) + Bn+ Op(p/n'12) n-I =Ao E k2(j/p) + B? + Op(p/n'12), j=l wherethe secondequalityfollowsfromproposition5 and p -+ oo. It followsthat 2Bo k4(E j}] /p) = {(n-j) Dn(j)-Ao) 1=1-I j=1 2Bo = k4( /p)} n + Op(p2/nl2 N(O ) by proposition4 and p = cn' for0 < v < 1, wherewe have also made use of the factthat n-2 k4( j/p) ?. j=l 00 =p o k4(z) dz{I1+ o(l)}. By replacingBo by A0,we have d -* Ma by Slutsky'stheoremand ho -B N(O, 1) = op(l). This completestheproofforMa. D A.2. Proofof theorem2 on (0, 1). This factsuggeststhatwe can For a continuousdistribution distributed G, G(X,) is uniformly computeAo and Bo directly: Ao={ u(l-u) du}= and Bo= [JJ{min(ul,u2)-uiu2}2du2duu,=9d. Hence, we need not useAOand Bo. The desiredresultfollowsimmediately. A.3. Proofof theorem3 O The consistencyof Ma followsfromthe factthat E kr(j/p)= p j=lI f kr(z)dz{I+ o(l)} forr > 2,p = cn' for0 < v < 3, and twopropositionsstatedbelow.The proofsof thesepropositionsare available fromtheauthoron request. Proposition6. Put Dj(x, y) = Fj(x, y) - G(x) G(y). Define D2( j) =n_ 2 Dj(XI, XIs) j)_, 1=j+l Then n-I n-I n-I j=I j=1 Ek2(j/p)(n _j)J25(j) = n- Ek2(j/p)(n _j)D2(j) + op(l). This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions PairwiseSerialIndependence 453 Proposition7. Let D2(j) be definedas in equation(1) and D2(j) be definedas in proposition6. Then -I E k2(j/p)(n_j) j=l D2(j) = E D2(j) + op(l). j=l References and autocorrelationconsistentcovariancematrixestimation.EconoAndrews,D. W. K. (1991) Heteroskedasticity metrica,59, 817-858. Bahadur,R. R. (1960) Stochasticcomparisonof tests.Ann.Math. Statist.,31, 276-295. Blum,J.,Kiefer,J. and Rosenblatt,M. (1961) Distributionfreetestsof independencebased on the sample distributionfunction.Ann.Math. Statist.,32, 485-498. movingaverage integrated in autoregressive Box, G. and Pierce,D. (1970) Distributionof residualautocorrelations timeseriesmodels.J. Am. Statist.Ass., 65, 1509-1527. Brown,B. M. (1971) Martingalecentrallimittheorems.Ann.Math. Statist.,42, 59-66. observations.CalcuttaStatist.Ass. Bull.,37, based on non-independent Carlstein,E. (1988) DegenerateU-statistics 55-65. Chan, N. H. and Tran, L. T. (1992) Nonparametrictestsforserialdependence.J. TimeSer. Anal., 13, 102-113. 74, 95-101. Davies, R. and Harte,D. (1987) Tests forHursteffect.Biometrika, J. Multiv. testsof independence. distribution-free decompositionformultivariate Deheuvels,P. (1981) An asymptotic Anal., 11, 102-113. function.J. TimeSer. Anal., 17, Delgado, M. A. (1996) Testingserialindependenceusingthe sample distribution 271-285. withestimatesof the varianceof United Kingdom conditionalheteroskedasticity Engle, R. (1982) Autoregressive 50, 987-1007. inflation.Econometrica, J. Am.Statist.Ass.,63, 817-836. stabledistributions. Fama, E. F. and Roll, R. (1968) Some propertiesof symmetric New York: OxfordUniversity Granger,C. and Terasvirta,T. (1993) ModellingNonlinearEconomicRelationships. Press. Hjellvik,V. and Tj0stheim,D. (1996) Nonparametricstatisticsfortestingof linearityand serial independence.J. Nonparam.Statist.,6, 223-251. W. (1948) A nonparametric testof independence.Ann.Math. Statist.,19, 546-557. Hoeffding, 64, 837-864. Hong, Y. (1996a) Consistenttestingforserialcorrelationof unknownform.Econometrica, 83, 615-625. (1996b) Testingforindependencebetweentwo covariancestationarytimeseries.Biometrika, 65, 297-303. Ljung, G. M. and Box, G. E. P. (1978) A measureof lack of fitin timeseriesmodels.Biometrika, Mandelbrot,B. (1967) The variationof some otherspeculativeprices.J. Bus., 40, 393-413. testingforserialindependence.J. Econometr.,to be published. Pinkse,J. (1997) Consistentnonparametric 4, 551-564. Priestley,M. (1962) Basic considerationsin theestimationof spectra.Technometrics, (1981) SpectralAnalysisand TimeSeries,vol. 1. London: AcademicPress. testing.Rev. Econ. Stud.,58, 437-453. entropy-based Robinson,P. M. (1991a) Consistentnonparametric in multiplere(1991b) Testingfor strongserial correlationand dynamicconditionalheteroskedasticity 47, 67-84. gression.J. Econometr., (ed. C. Sims), (1994) Time serieswithstrongdependence.In Proc. 6th WrldCongr.Advancesin Econometrics Press. vol. 1, pp. 47-95. Cambridge:CambridgeUniversity whentheobservationsare not independent.CalcuttaStatist.Ass. Sen, P. K. (1963) On thepropertiesof U-statistics Bull., 12, 69-92. Skaug, H. J. and Tj0stheim,D. (1993a) Nonparametrictest of serial independencebased on the empirical function.Biometrika, 80, 591-602. distribution in Time Series Analysis:thePriestley (1993b) Nonparametrictestsof serialindependence.In Developments BirthdayVolume(ed. T. Subba Rao), pp. 207-229. London: Chapman and Hall. (1996) Measures of distancebetweendensitieswithapplicationto testingforserial independence.In Time Series Analysisin Memoryof E. J. Hannan(eds P. M. Robinson and M. Rosenblatt),pp. 363-377. New York: Springer. This content downloaded from 128.84.125.184 on Fri, 22 Nov 2013 14:15:47 PM All use subject to JSTOR Terms and Conditions
© Copyright 2026 Paperzz