Economics Letters 155 (2017) 104–110 Contents lists available at ScienceDirect Economics Letters journal homepage: www.elsevier.com/locate/ecolet Tests for serial correlation of unknown form in dynamic least squares regression with wavelets✩ Meiyu Li, Ramazan Gençay * Department of Economics, Simon Fraser University, 8888 University Drive, Burnaby, BC, V5A 1S6, Canada article info Article history: Received 12 March 2017 Accepted 18 March 2017 Available online 29 March 2017 JEL classification: C1 C2 C12 C22 C26 a b s t r a c t This paper extends the multi-scale serial correlation tests of Gençay and Signori (2015) for observable time series to unobservable errors of unknown forms in a linear dynamic regression model. Our tests directly build on the variance ratio of the sum of squared wavelet coefficients of residuals over the sum of squared residuals, utilizing the equal contribution of each frequency of a white noise process to its variance and delivering higher empirical power than parametric tests. Our test statistics converge to the standard normal distribution at the parametric rate under the null hypothesis, faster than the nonparametric test using kernel estimators of the spectrum. © 2017 Elsevier B.V. All rights reserved. Keywords: Dynamic least squares regression Serial correlation Conditional heteroscedasticity Maximum overlap discrete wavelet transformation 1. Introduction Existing residual based serial correlation tests for dynamic models fall into two categories, the parametric tests in the time domain and the nonparametric tests in the frequency domain. The residual based parametric tests estimate autocorrelation coefficients directly, delivering easy implementation and desirable finite sample performance (Hayashi, 2000; Box and Pierce, 1970; Godfrey, 1978; Wooldridge, 1990, 1991), whereas their power relies on the choice of lag length heavily. In contrast, the residual based nonparametric tests compare the kernel-based spectrum estimator with the white noise spectrum and have more power (Hong, 1996; Hong and Lee, 2003). Nevertheless, the nonparametric estimation of the underlying spectrum sacrifices the convergence rate, which is slower than the parametric rate and often associated with poor finite sample performance (Chen and Deo, 2004a, b). Based on the fact of the equal contribution of each frequency of a white noise process to its variance, Gençay and Signori ✩ We gratefully acknowledge Bertille Antoine for providing helpful comments and suggestions. All errors are attributed to the authors. Corresponding author. E-mail addresses: [email protected] (M. Li), [email protected] (R. Gençay). * http://dx.doi.org/10.1016/j.econlet.2017.03.021 0165-1765/© 2017 Elsevier B.V. All rights reserved. (2015) proposes a family of wavelet-based serial correlation tests for observable time series, GS and GSM, utilizing only wavelet coefficients of the observed time series data. This paper extends the GS and GSM tests into a linear regression framework, particularly when lagged dependent variables are included in regression equations. Our modified tests are robust to models with conditionally heteroscedastic errors of unknown form. Inheriting the test design of Gençay and Signori (2015) by using the additive variance decomposition of the wavelet and the scaling coefficients, instead of any nonparametric estimation of the underlying spectrum, our test statistics converge to the normal distribution at the parametric rate under the null hypothesis (faster than the nonparametric test) and display higher power than the parametric test. In addition, contrary to the sensitiveness of finite sample performance to the choice of lag length for the parametric test, and the choice of bandwidth for the nonparametric test, our tests are rather stable when different wavelet decomposition levels are utilized. This paper proceeds as follows. Section 2 illustrates the modeling environment for our residual-based tests. Section 3 proposes The corresponding modified test statistics. Section 4 reports comprehensive Monte Carlo simulations for finite sample performance of our tests in comparison to commonly used tests. Section 5 contains concluding remarks of this paper. M. Li, R. Gençay / Economics Letters 155 (2017) 104–110 2. Model setting Consider the dynamic regression model yt = Yt′ α + Wt′ γ + ut = Xt′ β + ut (1) where: Yt = (yt −1 , . . . , yt −P ) with P ≥ 1; α = (α1 , . . . , αP ) having the values such that the roots of z P − α1 z P −1 − · · · − αP = 0 are all strictly inside the unit circle; Wt is an L-vector of regressors which may include lagged exogenous variables and constants; Xt′ ≡ (Yt′ , Wt′ ); and β ′ ≡ (α ′ , γ ′ ). Let K = P + L denotes the number of regression coefficients, and T denotes the number of observations available for estimation. All of the data generation processes are subject to the following two assumptions. ′ 105 (2015) distribute more tightly around zero than a standard normal distribution in dynamic models, more difficult to reject under the null hypothesis and less powerful. For dynamic linear models, we modify the statistics to restore the asymptotic distributions and improve efficiency in small samples. ′ Assumption 2.1 (Ergodic Stationary). The (K + 1) dimensional vector stochastic process {yt , Xt } is jointly stationary and ergodic. Assumption 3.1 (Finite Fourth Moments for Regressors). E(u2t ut −j ut −k ), E(ut ut −j ut −k xt −l,i ), E(ut ut −j xt −k,i xt −l,n ), E(ut −j xt ,i xt −k,n xt −l,s ), E(xt ,i xt ,n xt −j,s xt −k,v ) exist and are finite for all j, k, l = 0, 2, . . . , Lm − 1 and i, n, s, v = 1, 2, . . . , K . Theorem 3.1. If Assumptions 2.1, 2.2 and 2.3 are satisfied, then the sample analogue Êm,T constructed with residuals ût from the model in Eq. (1) is a consistent estimator of the wavelet variance ratio Em ∑T wvarm (û) Êm,T = = ∑t =T 1 var(û) Assumption 2.1 implies the stationarity and ergodicity of error process, {ut }. Nevertheless, the unconditional homoscedasticity does not exclude the models with a conditional heteroscedastic error term. To be specific, for any stationary and ergodic process, Assumption 2.1 allows any form of unconditional higher order moment dependence which disappears as time lag increases, and any form of time varying conditional higher order moments. We consider tests for the following null hypothesis: Assumption 2.3 (Stronger form of Predeterminedness). All regressors are predetermined such that: E(ut |ut −1 , ut −2 , . . . , Xt , Xt −1 , . . .) = 0.1 Throughout this paper, all the statistics constructed with wavelets are based on the Maximum Overlap Discrete Wavelet 3 Transform (MODWT) Define the mth level of wavelet coef∑Lfilter. m −1 m ficients as wm,t = l=0 h̃m,l ut −l mod T , where Lm := (2 − 1)(L − 1) + 1, L is the length of the initial MODWT filter, and {h̃m,l } is the mth level MODWT filter. Assumption 2.3 indicates that cov(Xt −j , ut ) = 0, while it allows cov(Xt +j , ut ) ̸ = 0 for some j > 0. The correlation between the future lagged dependent variable and current error term makes the asymptotic variance estimator in the GS and GSM tests inconsistent when the realizations of error terms {ut } are replaced with residuals {ût } directly. As a result, the test statistics of Gençay and Signori 1 Assumption 2.3 is a sufficient condition for g ≡ [ut −j ut , ut Xt′ ]′ to be a jt martingale difference sequence, stronger than we need. We maintain this stronger assumption for the convenience of interpretation. 2 The notation a − b mod T stands for ‘a − b modulo T ’. If j is an integer such that 1 ≤ j ≤ T , then j mod T ≡ j. If j is another integer, then j mod T ≡ j + nT where nT is the unique integer multiple of T such that 1 ≤ j + nT ≤ T . The periodic boundary conditions have trivial impact on the finite sample distribution due to the short length of most discrete wavelet filters. 3 For details of MODWT filter, see Gençay et al. (2001) and Percival and Walden (2000). 2m . Further, if Assumption 3.1 is satisfied, then √ LGm = ( Ts4 Êm,T − m m Ψ̂m Ĉm Hm 4H ′ Ĉ ′ 1 ) 2m d − → N(0, 1), where {h̃m,l } is the wavelet filter used in the construction of Êm,T and ⎡ h̃m,1 h̃m,2 h̃m,3 ⎡ ⎤ h̃m,2 h̃m,3 h̃m,3 ··· ··· .. . ··· .. . h̃m,Lm −1 h̃m,Lm −1 0 h̃m,Lm −1 0 0 . ··· ⎢ ⎢ ⎢ Hm ≡⎢ ⎢ ((Lm −1)×1) ⎣ .. . h̃m,Lm −1 0 ⎥ ⎥ 0 ⎥ ⎥ .. . .. ⎥ ⎦ 0 ⎤ h̃m,0 ⎢ h̃m,1 ⎢ ⎢ × ⎢ h̃m,2 ⎢ . ⎣ .. Assumption 2.3 is stronger than the assumption of white noise and needed for the derivation of the null distribution of our test statistic. If Assumptions 2.1, 2.2 and 2.3 are satisfied, the OLS coefp ficient estimator in Eq. (1) is consistent, namely β̂ − → β (Hayashi, 2000 Proposition 2.1(a), page 124). Throughout the paper, we assume Assumptions 2.1 and 2.2 hold and impose periodic boundary conditions on all time series, like {ut }, where ut ≡ ut mod T .2 3. Asymptotic null distribution 1 p − → 2 t =1 ût ′ Assumption 2.2 (Rank Condition). The K × K matrix E(Xt Xt ) is nonsingular (and hence finite), which is denoted by ΣXX . 2 ŵm ,t ⎥ ⎥ ⎥ ⎥, ⎥ ⎦ h̃m,Lm −2 ⎤ ⎡ ĉ1 Ĉm ((Lm −1)(K +1)×(Lm −1)) .. ⎢ ≡⎣ ⎥ ⎦, . ĉLm −1 ⎡ T 1∑ s2 ≡ 1 T −K T ∑ û2t , SXX = t =1 T 1∑ T T t =1 T Xt Xt′ , µ̄j ≡ t =1 −1 −SXX µ̄j , ⎤ Xt ût ût −j ⎥ t =1 T û2t Xt Xt′ ⎥ ⎥, ⎥ ⎦ t =1 T 1∑ T ] 1 ′ 2 1∑ Xt û2t ût −k T 1∑ ≡ ((K +1)×1) û2t ût −j ût −k ⎢ ⎢ T t =1 the (j, k) block of Ψ̂m ≡ ⎢ T ⎢ ∑ ((K +1)×(K +1)) ⎣ 1 T [ ĉj Xt ût −j . t =1 Combining these multiple-scale tests to gain power against a wide range of alternatives, we derive the asymptotic joint distribution of these tests as follows. Theorem 3.2. If Assumptions 2.1, 2.2, 2.3 and 3.1 are satisfied, then the vector √ Ts4 4 ( 1 Ê1,T − 21 , Ê2,T − where the mth column of 0 . ((LN −Lm )(K +1)×1) 1 22 , . . . , ÊN ,T − Υ ((LN −1)(K +1)×N) is 1 2N )′ d − → N(0, Υ ′ ΨN Υ ), C m Hm ((Lm −1)(K +1)×1) followed by 106 M. Li, R. Gençay / Economics Letters 155 (2017) 104–110 Further, a feasible multivariate test statistic is 4 LGMN = Ts ( 1 Ê1,T − 4 21 , . . . , ÊN ,T − ) 1 2N (Υ̂ ′ Ψ̂N Υ̂ )−1 ( )′ 1 1 d × Ê1,T − 1 , . . . , ÊN ,T − N − → χN2 . 2 2 Large sample inference on the values of the vector (LG1 , . . . , LGN ) can be handily implemented using the χ 2 distribution. Indeed, it is a standard result (Bierens, 2004 Theorem 5.9, page 118) that for a multivariate normal N-dimensional vector X and the corresponding non-singular N × N variance–covariance matrix Σ , X ′ Σ −1 X is distributed as a χN2 . Theorems 3.1 and 3.2 accommodate models with conditional heteroscedastic error terms, such as ARCH and GARCH, due to the estimation of Ψ . The fourth moments are easily estimated with sample analogues but may slow the convergence of our tests for very small samples. When errors satisfy conditional homoscedasticity, a more efficient test statistic is available to accelerate the convergence rate in finite samples. Assumption 3.2 (Stronger form of Conditional Homoscedasticity). E(u2t |ut −1 , ut −2 , . . . , Xt , Xt −1 , . . .) = σ 2 > 0. Theorem 3.3. If Assumptions 2.1, 2.2, 2.3 and 3.2 are satisfied, then √ LGhm = ( T ′ (I 4Hm Lm −1 − Φ̂m )Hm Êm,T − 1 ) 2m d − → N(0, 1), where {h̃m,l } is the wavelet filter used in the construction of Êm,T and ∑ −1 Φ̂m ≡ (φ̂jk ), φ̂jk ≡ µ̄′j SXX µ̄k /s2 , µ̄j ≡ 1/T Tt=1 Xt ût −j . In addition, the vector √ ( T 4 Ê1,T − 1 1 2 2 2N , . . . , ÊN ,T − 2 d − → N(0, H ′ (ILN −1 − ΦN )H), where the mth column of H ((LN −1)×N) is followed by 0 . Hm ((LN −Lm )×1) ((Lm −1)×1) Further, a feasible multivariate test statistic is LGMNh = T 4 ( Ê1,T − 1 1 2 2N , . . . , ÊN ,T − 1 ) (H ′ (ILN −1 − Φ̂N )H)−1 )′ ( 1 1 d → χN2 . × Ê1,T − 1 , . . . , ÊN ,T − N − 2 2 4. Monte Carlo simulations Consider the following linear dynamic regression model, M: Yt = 1 + 0.5Yt −1 + 0.5W1,t + 1.5W2,t + ut , where the exogenous variables W1,t = 0.01 + 0.4W1,t −1 + ν1,t and W2,t = 0.2W1,t −1 + 0.01W2,t −1 + ν2,t , with innovations [ All the wavelet based test statistics from T1 to T4 are computed with the MODWT Haar wavelet. The robustness of our modified tests to different wavelet filters is discussed at the end of Section 4. In addition, both T5 and T6 are computed with a quadratic norm and the Daniell kernel function, k(z) = sin(π z)/π z, z ∈ (−∞, ∞), which maximizes the power of the kernel test over a wide class of the kernel functions when p → ∞ (Hong, 1996, 2010). To study the empirical size of each test statistic, we consider sample sizes of 50, 100, and 500 observations. For each of the following null hypotheses and sample sizes, we compute the rejection rates from 5000 replications. i.i.d (1) Normal process ut ∼ N(0, 9). )′ 1 , Ê2,T − 1 T4: The estimated Gençay and Signori (2015) tests EGSm and EGSMN computed with estimated heteroscedastic variance via Newey–West estimator at scale m = 1, 2, 3 or level N = 2. T5: The Hong (1996) test Kph with bandwidth p. Same as Hong (1996), we use three rates: p1 = [ln(n)], p2 = [3n0.2 ] and p3 = [3n0.3 ], where [a] denotes the integer closest to a. T6: The Hong and Lee (2003) test Kp with same bandwidth p as Kph . Kp is computed following the formula suggested in Hong (2010). T7: The Breusch (1978) and Godfrey (1978) test LMk with lag lengths of k = 5, 10, 20. T8: The Wooldridge (1990, 1991) test Wk with lag lengths of k = 5, 10, 20. Wk is computed following the two step regressions suggested in Hong (2010): (i) Regress (ût −1 , . . . , ût −k ) on Xt and save the estimated k × 1 residual vector v̂t ; (ii) Regress 1 on v̂t ût and obtain SSR. Without serial correlation, N − SSR, robust to conditional heteroscedasticity, follows χk2 asymptotically. ] ([ ] [ ν1,t 0 1 ∼N , ν2,t 0 0.3 0.3 3 ]) . We examine finite sample performance of our test statistics in comparison to some commonly used residual based tests for serial correlations, which are detailed as follows. T1: The modified conditional homoscedasticity constrained tests LGhm and LGMNh at scale m = 1, 2, 3 or level N = 2. T2: The modified conditional heteroscedasticity robust tests LGm and LGMN at scale m = 1, 2, 3 or level N = 2. T3: The original Gençay and Signori (2015) tests GSm and GSMN computed with constrained homoscedastic variance at scale m = 1, 2, 3 or level N = 2. i.i.d (2) ARCH(2) process ut = σt ϵt , ϵt ∼ N(0, 1), σt2 = 0.001 + 0.2ϵt2−1 + 0.4ϵt2−2 . i.i.d (3) N(0,1)-GARCH(1, 1) process ut = σt ϵt , ϵt ∼ N(0, 1), σt2 = 0.001 + 0.1ϵt2−1 + 0.7σt2−1 . i.i.d (4) t5 -GARCH(1, 1) process ut = σt ϵt , ϵt ∼ t5 , σt2 = 0.001 + 0.1ϵt2−1 + 0.7σt2−1 , where t5 stands for a Student’s t with 5 degrees of freedom. i.i.d (5) EGARCH(1, 1) process ut = σt ϵt , ϵt ∼ N(0, 1), ln(σt2 ) = 0.001 + 0.4|ϵt −1 | + 0.1ϵt −1 + 0.7 ln(σt2−1 ). Hypothesis (1) permit us to examine size performances under the conditional homoscedastic white noise errors. For dynamic models, the larger the variance of the innovations, the more severe the underrejections of the tests T3 and T4 are. To highlight the impact of lagged dependent variables on these inconsistent tests, we choose a large variance. Hypotheses (2), (3), (4) and (5) allow us to focus on the circumstances of conditional heteroscedastic white noise errors. For conditional homoscedastic white noise errors, Table 1 illustrates the rejection rate of all tests under the null hypothesis (1) at 10%, 5% and 1% levels of significance. The empirical sizes of the LGhm and LGMNh tests match their nominal sizes even with sample size as small as 50 observations. The LGm and LGMN tests converge relatively slower, with sizes being viable with sample size 100, and accurate with sample size 500. All of our modified statistics are stable for varied wavelet decomposition levels and model specifications. The performance of the LMk test is acceptable but relies heavily on the choice of lag lengths. Analogous results hold for Wk with slight underrejections in all circumstances. Both Kph and Kp significantly underreject at all levels of significance, indicating the slow convergence rate and right skewness in finite samples. In addition, the performance of the two kernel tests are vulnerable to the choice of bandwidth. 3.66 10.82 7.18 5.58 9.32 11.74 11.38 10.34 10.92 8.12 3.26 GS1 GS2 GS3 GSM2 LGh1 LGh2 LGh3 LGM2h LM5 LM10 LM20 2.80 4.50 5.34 10% Level h Kp1 h Kp2 h Kp3 50 T 1.74 2.56 3.44 5.52 2.96 0.72 4.76 6.26 6.76 5.06 1.18 5.64 3.46 2.50 5% 0.74 1.10 1.34 0.90 0.16 0.00 0.78 1.48 2.36 1.44 0.10 1.22 0.84 0.62 1% 3.96 5.60 6.48 10.20 9.16 5.14 9.56 10.54 10.22 9.58 3.54 10.02 7.10 5.26 10% 100 2.24 3.26 4.02 4.68 3.86 2.04 5.04 5.36 5.52 4.96 1.34 4.98 3.16 2.30 5% 0.78 1.48 1.66 0.82 0.52 0.10 1.00 1.22 1.88 1.30 0.14 1.14 0.76 0.42 1% M : Yt = 1 + 0.5Yt −1 + 0.5W1,t + 1.5W2,t + ut and ut ∼ N(0, 9) i.i.d. 5.22 6.60 7.90 9.68 10.10 9.48 9.52 9.94 10.08 9.68 3.52 9.80 7.38 4.72 10% 500 3.30 4.24 4.66 4.68 5.08 4.50 4.96 4.80 5.32 4.58 1.38 4.66 3.12 1.98 5% 1.40 1.60 1.88 0.82 0.94 0.82 1.02 0.90 1.10 1.08 0.08 0.90 0.62 0.34 1% Kp1 Kp2 Kp3 W5 W10 W20 LG1 LG2 LG3 LGM2 EGS1 EGS2 EGS3 EGSM2 3.60 4.76 5.42 9.24 4.42 0.00 14.42 15.54 15.18 18.20 7.98 19.82 16.68 23.84 10% 50 2.08 2.76 3.30 3.40 0.86 0.00 8.00 8.92 8.20 10.18 4.20 12.48 11.28 16.98 5% 0.82 1.16 1.24 0.14 0.02 0.00 1.60 2.38 1.60 2.08 1.32 4.18 5.50 8.68 1% 4.16 5.66 6.50 10.54 7.46 2.90 11.74 12.72 11.72 12.82 6.04 15.84 15.28 15.14 10% 100 Table 1 Rejection probabilities in percentages of tests at 10%, 5% and 1% levels of significance under the null hypothesis (1). All size simulations are based on 5000 replications. 2.52 3.54 3.82 4.70 3.06 0.70 6.26 6.08 5.98 6.66 2.92 9.28 10.40 9.44 5% 0.92 1.44 1.54 0.58 0.30 0.00 1.26 1.30 1.00 1.48 0.44 2.66 4.94 3.50 1% 500 5.06 6.48 7.66 9.42 9.08 7.90 9.94 10.04 10.68 10.48 4.22 11.78 9.94 7.64 10% 5% 3.26 3.90 4.64 5.04 4.38 3.62 5.28 5.18 5.40 5.06 1.50 6.48 4.98 3.58 1% 1.26 1.58 1.80 0.94 0.86 0.60 1.16 1.02 1.00 1.08 0.08 1.36 1.32 0.84 M. Li, R. Gençay / Economics Letters 155 (2017) 104–110 107 108 M. Li, R. Gençay / Economics Letters 155 (2017) 104–110 Table 2 Rejection probabilities in percentages of tests at 10%, 5% and 1% levels of significance under the null hypothesis (2), (3), (4) and (5). All size simulations are based on 5000 replications. M : Yt = 1 + 0.5Yt −1 + 0.5W1,t + 1.5W2,t + ut N(0, 1) − GARCH(1, 1) ARCH(2) T 100 500 100 Level 10% 5% 1% 10% 5% 1% 10% 5% 1% 10% 5% 1% LG1 13.24 7.04 1.40 10.36 5.00 1.00 12.92 6.40 1.58 10.20 4.98 1.00 LG2 12.56 5.92 0.78 11.10 5.50 1.00 12.44 6.54 1.48 10.16 5.04 0.88 LG3 12.94 6.42 1.22 10.52 5.08 0.94 12.18 6.14 1.20 10.46 5.60 0.98 LGM2 13.66 6.86 1.06 10.92 5.18 0.94 13.82 7.60 1.46 10.36 5.30 1.10 Kp1 9.48 6.30 2.92 9.00 6.12 2.86 10.68 7.32 3.58 9.32 6.54 2.90 Kp2 9.50 6.00 2.36 8.80 6.14 2.52 10.34 7.06 3.14 9.38 6.50 2.88 Kp3 9.08 5.82 2.32 9.04 5.62 2.34 10.30 6.92 2.96 10.02 6.38 2.42 W5 8.02 3.00 0.34 9.08 3.78 0.44 9.64 4.14 0.54 10.46 5.22 1.16 W10 6.02 1.88 0.08 8.30 3.64 0.44 7.20 2.68 0.30 10.00 4.60 0.84 W20 2.86 0.66 0.00 7.76 3.38 0.46 3.04 0.66 0.00 8.46 3.60 0.42 t5 − GARCH(1, 1) 500 EGARCH(1, 1) LG1 12.32 5.82 1.06 10.56 5.16 0.76 12.40 6.52 1.46 9.58 4.58 0.96 LG2 12.18 6.24 1.04 10.40 5.30 1.00 12.22 6.56 1.16 10.28 5.56 1.26 LG3 11.38 5.80 1.24 10.18 4.74 0.82 12.00 6.34 1.26 10.26 5.58 1.18 LGM2 12.80 5.92 1.10 10.86 5.20 0.92 14.00 6.52 1.04 9.82 5.18 1.10 Kp1 8.82 6.12 2.88 9.36 6.22 2.78 8.06 5.50 2.70 8.10 5.42 2.46 Kp2 8.62 5.58 2.28 9.06 5.62 2.56 8.08 5.34 2.64 8.12 5.22 2.08 Kp3 8.44 5.20 2.02 8.56 5.26 2.10 8.44 5.56 2.38 8.44 5.00 1.72 W5 8.50 3.30 0.42 9.98 4.50 0.66 9.08 3.72 0.44 9.58 4.50 0.66 W10 6.20 2.60 0.16 9.00 4.00 0.56 6.64 2.32 0.22 9.28 4.28 0.64 W20 3.14 0.74 0.06 7.20 3.16 0.48 2.96 0.50 0.00 7.98 3.20 0.50 Table 3 Size adjusted power in percentages of tests at 1% level of significance under the alternative hypothesis (a), AR(2)-N(0,1) process. Simulations are carried out for a set of alternatives obtained varying α1 and α2 in interval [−0.4, 0.4] in increments of 0.1. All rejection rates are based on 5000 replications with 1% nominal size for sample size of 100 observations. M. Li, R. Gençay / Economics Letters 155 (2017) 104–110 109 Table 4 Size adjusted power in percentages of tests at 1% level of significance under the alternative hypothesis (b), AR(2)-GARCH(1,1) process. Simulations are carried out for a set of alternatives obtained varying α1 and α2 in interval [-0.4, 0.4] in increments of 0.1. All rejection rates are based on 5000 replications with 1% nominal size for samples of 100 observations. Table 5 Size and power in percentages for various wavelet families at 5% level of significance in model Yt = 1 + 0.5Yt −1 + 0.5W1,t + 1.5W2,t + ut , against the following hypotheses, i.i.d i.i.d AR(2)-N(0,1): ut = α1 ut −1 + α2 ut −2 + ϵt and ϵt ∼ N(0, 1); AR(2)-GARCH(1,1): ut = α1 ut −1 + α2 ut −2 + σt ϵt , ϵt ∼ N(0, 1) and σt2 = 0.001 + 0.1ϵt2−1 + 0.7σt2−1 . All the rejection rates are based on 5000 replications with sample sizes of 100, 500 and 1000 observations. Model Panel A: AR(2)-N(0,1), LGM2h T α1 α1 α1 α1 α1 α1 α1 α1 α1 = 0, α2 = 0 = 0, α2 = 0 = 0, α2 = 0 = 0.1, α2 = 0 = 0.1, α2 = 0 = 0.1, α2 = 0 = 0, α2 = 0.1 = 0, α2 = 0.1 = 0, α2 = 0.1 Haar D(4) Panel B: AR(2)-GARCH(1,1), LGM2 D(6) LA(8) BL(14) C(6) T Haar D(4) D(6) LA(8) BL(14) C(6) 100 500 1000 5.20 5.04 4.94 5.18 5.32 4.84 4.96 5.34 4.76 4.86 5.38 4.74 4.80 5.42 4.98 5.14 5.34 4.84 100 500 1000 6.94 4.84 5.58 6.84 4.70 5.30 6.86 4.78 5.50 7.02 4.92 5.48 7.12 5.04 5.58 6.84 4.72 5.34 100 500 1000 10.64 44.10 75.88 11.54 44.40 76.04 11.62 44.28 75.94 11.60 44.24 75.78 11.74 43.94 75.04 11.62 44.40 76.08 100 500 1000 10.62 38.80 69.34 10.56 38.46 69.12 10.22 37.98 68.98 10.18 37.76 68.94 10.02 37.56 68.18 10.46 38.36 69.16 100 500 1000 10.18 38.62 69.18 9.06 33.80 63.84 8.48 30.80 59.34 8.32 29.18 56.68 7.92 26.06 51.44 9.06 33.56 63.52 100 500 1000 11.16 34.42 61.20 10.78 31.10 56.68 10.16 29.28 53.36 10.10 27.96 51.12 9.80 25.50 47.14 10.62 31.00 56.34 Table 2 illustrates the rejection rate for conditional heteroscedastic white noise errors under the null hypotheses (2), (3), (4) and (5). We only focus on the robust test statistics LGm , LGMN , Kp and Wk .4 LGm and LGMN deliver accurate sizes in sample with 500 observations under all hypotheses. For a small sample, they slightly overreject but are rather stable over different levels of decompositions. The Kp tests have great difficulty of getting it close to their corresponding nominal sizes at all levels of significance with all sample sizes, underrejecting at 10% level of significance but significantly overreject at both 5% and 1% levels of significance. The Wk tests underreject in all models and rely on the choice of lag lengths heavily. 4 The simulation results for other test statistics are available upon request. To examine the empirical size adjusted power, we consider the following alternative hypotheses of errors in model M. All the simulation results are based on 5000 replications with samples of 100 observations.5 i.i.d (a) AR(2)-N(0,1) process ut = α1 ut −1 + α2 ut −2 + ϵt and ϵt ∼ N(0, 1). (b) AR(2)-GARCH(1, 1) process ut = α1 ut −1 + α2 ut −2 + σt ϵt , i.i.d ϵt ∼ N(0, 1) and σt2 = 0.001 + 0.1ϵt2−1 + 0.7σt2−1 . Table 3 illustrates the empirical size adjusted power of h and LM5 against the alternative hypothesis (a) tests LGM2h , Kp1 5 Size adjusted power is computed with, for a given sample size, the empirical critical values obtained from Monte Carlo simulations with 50,000 replications. 110 M. Li, R. Gençay / Economics Letters 155 (2017) 104–110 AR(2)-N(0,1) process by varying α1 and α2 in interval [-0.4, 0.4] in increments of 0.1.6 The rejection rates are computed with respect to 1% nominal size for sample size of 100 observations. Bandwidth p1 is chosen for the Kph test to guarantee its larger power, since a faster p gives better size while a slower p gives better power (Hong, 1996). We choose k = 5 for the LMk test for the same reason. The first column of the table contains the size adjusted power of each test for various alternatives. In the second column, we report the relative power gains of the conditional homoscedasticity h constrained test LGM2h with respect to the Kp1 test and the LM5 test. Against the great majority of the alternatives, LGM2h outperforms the LM5 test a large amount. Nevertheless, there is no clear edge h h among the LGM2h test, the Kp1 test. Specifically, the Kp1 test has higher power under slightly more number of alternatives with the highest power gain around 46%. Yet the highest power gain of LGM2h reaches 174% in other alternatives. In Table 4, we repeat the previous power analysis for the tests LGM2 , Kp1 and W5 against the alternative hypothesis (b), AR(2)GARCH(1, 1) process. Roughly speaking, the corresponding conditional heteroscedasticity robust tests inherit the relative power performance in the previous alternative (a) process. The LGM2 test performs neck to neck with the Kp1 test, leaving W5 dominated significantly in all alternatives. The Kp1 test has higher power against slightly more number of alternatives. Nevertheless, the highest power gain of the Kp1 test is 41% while the relative power gain of the LGM2 test achieves 310% against other alternatives. Finally, we examine the effect of various wavelet families on our modified tests, LGM2h and LGM2 , in Table 5. The wavelet filters investigated include Daubechies filters Haar, D(4), D(6), Least Asymmetric filter LA(8), Best Localized filter BL(14) and Coiflet filter C(6). Overall, the size and power of wavelet tests do not exhibit significant differences across various wavelet filters, although the shorter filters displays larger power in small samples. 5. Conclusions This paper proposes a test for serial correlation of unknown form for the residuals from a linear dynamic regression model. Our test possesses the same strong power as the nonparametric test in the frequency domain and the convergence rate identical to the parametric test in the time domain simultaneously. Our tests build on the characteristics of the spectral density function of the underlying process compared to that of the white noise process. There is no intermediate step such as the kernel estimation of the spectral density functions in our approach. With the additive feature of the variance decomposition by the wavelet and the 6 Simulation results for the moving average processes are similar to those for hypotheses (a) and (b), autoregressive processes. We omit them in the interest of space and they are available upon request. scaling coefficients, our test converges to the standard normal distribution at the parametric rate, avoiding the issue of slower convergence rate caused by nonparametric spectrum estimators. Our test is easily implementable and calculated fast. Acknowledgments Ramazan Gençay gratefully acknowledges financial support from the Natural Sciences and Engineering Research Council of Canada and the Social Sciences and Humanities Research Council of Canada. The software for the tests are available at www.sfu.ca/ ~rgencay/. References Bierens, H.J., 2004. Introduction to the Mathematical and Statistical Foundations of Econometrics. Cambridge University Press. Box, G.E.P., Pierce, D.A., 1970. Distribution of residual autocorrelations in autoregressive-integrated moving average time series models. J. Amer. Statist. Assoc. (ISSN: 01621459) 65 (332), 1509–1526. http://www.jstor.org/stable/ 2284333. Breusch, T.S., 1978. Testing for autocorrelation in dynamic linear models. Aust. Econ. Papers 17 (31), 334–355. http://dx.doi.org/10.1111/j.1467-8454.1978.tb00635. x. Chen, W.W., Deo, R.S., 2004a. A generalized portmanteau goodness-of-fit test for time series models. Econometric Theory 20 (2), 382–416. http://www.jstor.org/ stable/3533385. Chen, W.W., Deo, R.S., 2004b. Power transformations to induce normality and their applications. J. R. Stat. Soc. Ser. B Stat. Methodol. 66 (1), 117–130. http://www. jstor.org/stable/3647630. Gençay, R., Selçuk, F., Whitcher, B., 2001. An Introduction to Wavelets and Other Filtering Methods in Finance and Economics. Academic Press, San Diego. Gençay, R., Signori, D., 2015. Multi-scale tests for serial correlation. J. Econometrics 184, 62–80. http://dx.doi.org/10.1016/j.jeconom.2014.08.002. Godfrey, L.G., 1978. Testing against general autoregressive and moving average error models when the regressors include lagged dependent variables. Econometrica 46 (6), 1293–1301. http://www.jstor.org/stable/1913829. Hayashi, F., 2000. Econometrics. Princeton University Press, Princeton, New Jersey. Hong, Y., 1996. Consistent testing for serial correlation of unknown form. Econometrica 64 (4), 837–864. http://www.jstor.org/stable/2171847. Hong, Y., 2010. Serial Correlation and Serial Dependence. In: Durlauf, S.N., Blume, L.E. (Eds.), Macroeconometrics and Time Series Analysis. Palgrave Macmillan UK, London, ISBN: 978-0-230-28083-0, pp. 227–244. http://dx.doi.org/10.1057/ 9780230280830_25. Hong, Y., Lee, Y.J., 2003. Consistent Testing for Serial Correlation of Unknown form Under General Conditional Heteroskedasticity. Cornell University, Preprint. Percival, D.B., Walden, A.T., 2000. Wavelet Methods for Time Series Analysis. Cambridge University Press, Cambridge. Wooldridge, J.M., 1990. An encompassing approach to conditional mean tests with applications to testing nonnested hypotheses. J. Econometrics 45, 331–350. Wooldridge, J.M., 1991. On the application of robust , regression-based diagnostics to models of conditional means and conditional variances. J. Econometrics 47, 5–46.
© Copyright 2026 Paperzz