Bootstrapping realized volatility and realized beta

Bootstrapping realized volatility and realized beta under a local
Gaussianity assumption∗
Ulrich Hounyo
†
Aarhus University and CREATES
May 4, 2016
Abstract
The main contribution of this paper is to propose a new bootstrap method for statistics based
on high frequency returns. The new method exploits the local Gaussianity and the local constancy
of volatility of high frequency returns, two assumptions that can simplify inference in the high
frequency context, as recently explained by Mykland and Zhang (2009).
Our main contributions are as follows.
First, we show that the local Gaussian bootstrap is
rst-order consistent when used to estimate the distributions of realized volatility and realized
betas.
Second, we show that the local Gaussian bootstrap matches accurately the rst four cu-
mulants of realized volatility up to
o (h)
(where
h−1
denotes the sample size), implying that this
method provides third-order renements. This is in contrast with the wild bootstrap of Gonçalves
and Meddahi (2009), which is only second-order correct. Third, we show that the local Gaussian
bootstrap is able to provide second-order renements for the realized beta, which is also an improvement of the existing bootstrap results in Dovonon, Gonçalves and Meddahi (2013) (where
the pairs bootstrap was shown not to be second-order correct under general stochastic volatility).
Lastly, we provide Monte Carlo simulations and use empirical data to compare the nite sample
accuracy of our new bootstrap condence intervals for integrated volatility with the existing results.
JEL Classication: C15, C22, C58
Keywords: High frequency data, realized
volatility, realized beta, bootstrap, Edgeworth expan-
sions.
1
Introduction
Realized measures of volatility have become extremely popular in the last decade as higher and higher
frequency returns are available. Despite the fact that these statistics are measured over large samples,
their nite sample distributions are not necessarily well approximated by their asymptotic mixedGaussian distributions.
This is especially true for realized statistics that are not robust to market
microstructure noise since in this case researchers usually face a trade-o between using large sample
sizes and incurring in market microstructure biases. This has spurred interest in developing alternative
∗
I would like to thank Nour Meddahi for very stimulating discussions, which led to the idea of this paper.
I am
especially indebted to Sílvia Gonçalves for her valuable comments. I am also grateful to Asger Lunde, Peter Exterkate
and Matthew Webb, the co-editor Eric Renault and anonymous referees for helpful advice, comments and suggestions. I
acknowledge support from CREATES - Center for Research in Econometric Analysis of Time Series (DNRF78), funded
by the Danish National Research Foundation, as well as support from the Oxford-Man Institute of Quantitative Finance.
†
Department of Economics and Business, Aarhus University, 8210 Aarhus V., Denmark. Email: [email protected].
1
approximations based on the bootstrap. In particular, Gonçalves and Meddahi (2009) have recently
proposed bootstrap methods for realized volatility whereas Dovonon, Gonçalves and Meddahi (2013)
have studied the application of the bootstrap in the context of realized regressions.
The main contribution of this paper is to propose a new bootstrap method that exploits the local
Gaussianity framework described in Mykland and Zhang (2009). As these authors explain, one useful
way of thinking about inference in the context of realized measures is to assume that returns have
constant variance and are conditionally Gaussian over blocks of consecutive
M
observations. Roughly
speaking, a high frequency return of a given asset is equal in law to the product of its volatility (the
spot volatility) multiplied by a normal standard distribution. Mykland and Zhang (2009) show that
this local Gaussianity assumption is useful in deriving the asymptotic theory for the estimators used
in this literature by providing an analytic tool to nd the asymptotic behaviour without calculations
being too cumbersome. This approach also has the advantage of yielding more ecient estimators by
varying the size of the block (see Mykland and Zhang (2009) and Mykland, Shephard and Sheppard
(2012).
The main idea of this paper is to see how and to what extent this local Gaussianity assumption
can be explored to generate a bootstrap approximation. In particular, we propose and analyze a new
bootstrap method that relies on the conditional local Gaussianity of intraday returns. The new method
(which we term the local Gaussian bootstrap) consists of dividing the original data into non-overlapping
blocks of
M
observations and then generating the bootstrap observations at each frequency within a
block by drawing a random draw from a normal distribution with mean zero and variance given by the
realized volatility over the corresponding block. Using Mykland and Zhang's (2009) blocking approach,
one can act as if the instantaneous volatility is constant over a given block of consecutive observations.
In practice, the volatility of asset returns is highly persistent, especially over a daily horizon, implying
that it is at least locally nearly constant.
We focus on two realized measures in this paper: realized volatility and realized regression coecients. The latter can be viewed as a smooth function of the realized covariance matrix. Our proposal
in this case is to generate bootstrap observations on the vector that collects the intraday returns that
enter the regression model by applying the same idea as in the univariate case. Specically, we generate
bootstrap observations on the vector of variables of interest by drawing a random vector from a multivariate normal distribution with mean zero and covariance matrix given by the realized covariance
matrix computed over the corresponding block.
Our ndings for realized volatility are as follows. When
M
is xed, the local Gaussian bootstrap
is asymptotically correct but it does not oer any asymptotic renements. More specically, the rst
four bootstrap cumulants of the
t-statistic
estimator that is based on a block size of
based on realized volatility and studentized with a variance
M
do not match the cumulants of the original
to higher order (although they are consistent). Note that when
M = 1,
the new bootstrap method
coincides with the wild bootstrap of Gonçalves and Meddahi (2009) based on a
2
t-statistic
N (0, 1) external random
variable. As Gonçalves and Meddahi (2009) show, this is not an optimal choice, which is in line with
our results. Therefore, our result generalizes that of Gonçalves and Meddahi (2009) to the case of a
xed
M > 1.
However, if the block length
M → ∞,
as
h→0
such that
M h → 0,
(where
h−1
denotes
the sample size), then the local Gaussian bootstrap is able to provide an asymptotic renement. In
particular, we show that the rst and third bootstrap cumulants of the
corresponding cumulants at the rate
o h
1/2
t-statistic
converge to the
, which implies that the local Gaussian bootstrap oers
a second-order renement. In this case, the local Gaussian bootstrap is an alternative to the optimal
two-point distribution wild bootstrap proposed by Gonçalves and Meddahi (2009). More interestingly,
we also show that the local Gaussian bootstrap is able to match the second and fourth order cumulants
up to
o (h),
small enough error to yield a third-order asymptotic renement.
This is contrast to
the optimal wild bootstrap methods of Gonçalves and Meddahi (2009), which can not deliver thirdorder asymptotic renements. We also show that the local Gaussian bootstrap variance is a consistent
estimator of the (conditional) asymptotic variance of realized volatility. Then, we provide a proof of the
rst-order asymptotic validity of the local Gaussian bootstrap method for unstudentized (percentile)
intervals. This is contrast to the best existing bootstrap methods of Gonçalves and Meddahi (2009),
which can not consistently estimate the asymptotic covariance matrix of realized covolatility.
For the realized regression estimator proposed by Mykland and Zhang (2009), the local Gaussian
bootstrap matches the cumulants of the
t-statistics
through order
o h1/2
when
M → ∞.
Thus, this
method can promise second-order renements. This is contrast with the pairs bootstrap studied by
Dovonon, Gonçalves and Meddahi (2013), which is only rst-order correct.
Our Monte Carlo simulations suggest that the new bootstrap method we propose improves upon
the rst-order asymptotic theory in nite samples and outperforms the existing bootstrap methods.
The rest of this paper is organized as follows. In the next section, we rst introduce the setup, our
assumptions and describe the local Gaussian bootstrap. In Sections 3 and 4 we establish the consistency
of this method for realized volatility and realized beta, respectively. Section 5 contains the higher-order
asymptotic properties of the bootstrap cumulants. Section 6 contains simulations, while the results of
an empirical application are presented in Section 7. We conclude in Section 8 by describing how the
local Gaussian bootstrap can be extended to a setting which allow for irregularly spaced data, nonsynchronicity, market microstructure noise and/or jumps in the log-price process. Three appendices are
provided. Appendix A contains the tables with simulation results whereas Appendix B and Appendix
C contain the proofs.
2
Framework and the local Gaussian bootstrap
The statistics of interest in this paper can be written as smooth functions of the realized multivariate
volatility matrix. Here we describe the theoretical framework for multivariate high frequency returns
and introduce the new bootstrap method we propose.
3
Sections 3 and 4 will consider in detail the
theoretical properties of this method for the special cases of realized volatility and realized beta,
respectively.
We follow Mykland and Zhang (2009) and assume that the log-price process
of a
d-dimensional vector of assets is dened on a probability space (Ω, F, P ) equipped with a ltration
(Ft )t≥0 .
We model
X
as a Brownian semimartingale process that follows the equation,
t
Z
Xt = X0 +
Z
where
µ = (µt )t≥0
is a
d-dimensional
dimensional Brownian motion and
Σt ≡ σt σt0
t
σs dWs , t ≥ 0,
µs ds +
0
that
(1)
(d) 0
Xt = Xt , . . . , Xt
(1)
0
predictable locally bounded drift vector,
σ = (σt )t≥0
is an adapted càdlàg
is the spot covariance matrix of
X
W = (Wt )t≥0
d-
is
d × d locally bounded process such
t.
at time
We follow Barndor-Nielsen et al. (2006) and assume that the matrix process
σt
is invertible and
satises the following assumption
t
Z
σt = σ0 +
as ds +
and
Z
a, b,
and
v
Z
t
bs dWs +
0
where
t
Z
vs dZs ,
0
a
are all adapted càdlàg processes, with
is a vector Brownian motion independent of
(2)
0
also being predictable and locally bounded,
W.
The representation in (1) and (2) is rather general as it allows for leverage and drift eects. Assumption 2 of Mykland and Zhang (2009) or equation (1) of Mykland and Zhang (2011) also impose a
Brownian semimartingale structure on the instantaneous covolatility matrix
σ.
Equation (2) rules out
jumps in volatility, but this can be relaxed (see Assumption H1 of Barndor-Nielsen et al. (2006) for
a weaker assumption on
Suppose we observe
from which we compute
σ ).
X
over a xed time interval
1/h
[0, 1]
intraday returns at frequency
Z
ih
Z
yi ≡ Xih − X(i−1)h =
µt dt +
(i−1)h
where we will let
yki
at regular time points
to denote the
i-th
ih,
for
i = 0, . . . , 1/h,
h,
ih
1
σt dWt , i = 1, . . . , ,
h
(i−1)h
k , k = 1, . . . , d.
intraday return on asset
(3)
The parameters of
interest in this paper are functions of the elements of the integrated covariance matrix of
X,
i.e., the
process
Z
t
Γt ≡
Σs ds, t ∈ [0, 1] .
0
Without loss of generality, we let
over the period
[0, 1] ,
we let
Γkl
t=1
and dene
denote the
Γ = Γ1 =
(k, l)-th
As equation (3) shows, the intraday returns
yi
R1
0
element of
Σs ds
as the integrated covariance of
Γ.
depend on the drift
out inference for observations in a xed time interval the process
For most purposes it is only a nuisance parameter.
X
µt
µ,
unfortunately when carrying
cannot be consistently estimated.
To deal with this, Mykland and Zhang (2009)
propose to work with a new probability measure which is measure theoretically equivalent to
P
and
under which there is no drift (a statistical risk neutral measure). They pursue the analysis further and
4
propose an approximation measure
Qh,M
dened on the discretized observations
1
M h non overlapping blocks of size
1
of high frequency returns within a block, we have that M ≤ .
h
the volatility is constant on each of the
Specically, under the approximate measure
yi+(j−1)M h = C(j) ·
where
σt
ηi+(j−1)M ∼ i.i.d.N (0, Id ), Id
at the beginning of the
j -th
is a
√
d×d
Qh,M ,
in each block
hηi+(j−1)M ,
for
M.
Xih
M
Since
j = 1, . . . , M1h ,
only, for which
is the number
we have,
1 ≤ i ≤ M,
identity matrix and
C(j) = σ(j−1)M h ,
(4)
i.e., the value of
block (see Mykland and Zhang (2009), p.1417 for a formal denition of
Qh,M ).
The true distribution is
P,
but we may prefer to work with
simpler. Afterwards we adjust results back to
P
Qh,M
since then calculations are much
using the likelihood ratio (Radon-Nikodym derivative)
dQh,M /dP .
Remark 1. As pointed out in Mykland and Zhang's (2009) Theorem 3 and, in Mykland and Zhang's
(2011) Theorem 1, the measure P and its approximation Qh,M are contiguous on the observables. This
is to say that for any sequence Ah of sets, P (Ah ) → 0 if and only if Qh,M (Ah ) → 0. In particular, if
an estimator is consistent under Qh,M , it is also consistent under P . Rates of convergence (typically
h−1/2 ) are also preserved, but the asymptotic distribution may change. More specically, when adjusting
from Qh,M to P , the asymptotic variance of the estimator is unchanged (due to the preservation of
quadratic variation under limit operations), while the asymptotic bias may change. It appears that
a given sequence Zh of martingales will have exactly the same asymptotic distribution under Qh,M
and P , when the Qh,M martingale part of the log likelihood ratio log(dP /dQh,M ) has zero asymptotic
covariation with Zh . In this case, we do not need to adjust the distributional result from Qh,M to P .
Two important examples where this is true are the realized volatility and realized beta which we will
study in details in Sections 3 and 4.
Remark 2. In the particular case where the window length M increases with the sample size h−1
such that M = o(h−1/2 ), there is also no contiguity adjustment (see Remark 2 of Mykland and Zhang
(2011)). However, it is important to highlight that all results using contiguity arguments in Mykland
and Zhang (2009) apply only to the case with a bounded M. For the case M → ∞, Mykland and Zhang
(2011) use a dierent representation than (4).
It should be noted that throughout this paper, we will focus only on the representation given in
(4). As a consequence, we would not assume that the approximate measure Qh,M is contiguous to the
measure P when M grows to innity. This means that, we would neither use results in Mykland and
Zhang (2009), nor those in Mykland and Zhang (2011) when M → ∞.
Next we introduce a new bootstrap method that exploits the structure of (4). In particular, we
C(j) by
0
i=1 yi+(j−1)M yi+(j−1)M . That is, we follow
mimic the original observed vector of returns, and we use the normality of the data and replace
0
its estimate Ĉ(j) , where Ĉ(j) is such that Ĉ(j) Ĉ
(j)
=
5
1
Mh
PM
the main idea of Mykland and Zhang (2009), and assume constant volatility within blocks.
inside each block
where
j
M (j = 1, . . . , M1h ), we generate the M vector of returns
√
∗
∗
= Ĉ(j) ·
hηi+(j−1)M
, for i = 1, . . . , M,
yi+(j−1)M
of size
∗
ηi+(j−1)M
∼ i.i.d.N (0, Id )
across
(i, j),
and
Id
is a
d×d
Then,
as follows,
(5)
identity matrix.
Although we motivate the local Gaussian bootstrap data-generating process (DGP) by relying on
Qh,M
the approximate measure
with representation given by (4), note that to generate the bootstrap
vector of returns, we only need an estimator of the spot covariance matrix and a normal external
random vector. The existence of the approximate measure
validity of the local Gaussian bootstrap approach.
bootstrap remains asymptotically valid when
the approximate measure
Qh,M
Qh,M
is not a necessary condition to the
As we show in this paper, the local Gaussian
M → ∞,
i.e., a setting where we do not know whether
(with representation given by (4)) is contiguous to
P. In such situation,
for the proof of the validity of the bootstrap the key aspect is that we proceed directly under the true
P,
probability measure
without using any contiguity arguments.
In this paper, and as usual in the bootstrap literature,
P ∗ (E ∗
and
V ar∗ )
denotes the proba-
bility measure (expected value and variance) induced by the bootstrap resampling, conditional on
a realization of the original time series.
probability distribution
statistics
∗ ,
Zh,M
ability under
P
we write
υ,
In the following additional notations,
or the approximate probability measure
For a sequence of bootstrap
∗
in probability under
h
∗
= OP ∗ (1) as h → 0, in probability under υ
Zh,M
i
h ∗ limh→0 υ P ∗ Zh,M
> Mε > ε = 0. Finally, we
υ,
denotes the true
∗
→P 0, as h → 0, in
υ, or Zh,M
i
∗ ε > 0, δ > 0, limh→0 υ P ∗ Zh,M
> δ > ε = 0. Similarly, we
∗
Zh,M
= oP ∗ (1)
if for any
Qh,M .
υ
if conditional on the sample,
∗
Zh,M
∗
write Zh,M
weakly converges to
set with probability converging to one under
ε>0
if for all
Z
→d∗
under
there exists a
Z
as
P ∗,
h → 0,
Mε < ∞
probwrite
such that
in probability under
for all samples contained in a
υ.
The following result is crucial in obtaining our bootstrap results only in the context where those
Qh,M .
results are derived under the approximate measure
∗
Theorem 2.1. Let Zh,M
be a sequence of bootstrap statistics. Given the probability measure P and its
approximation Qh,M , we have that
∗
∗
∗
∗
Zh,M
→P 0, as h → 0, in probability under P , if and only if Zh,M
→P 0, as h → 0, in probability
under Qh,M .
Proof of Theorem 2.1
∗
Zh,M
∗
→P
0,
as
For any
h → 0,
ε > 0, δ > 0,
in probability under
is equivalent to
limh→0 Qh,M (Ah,M ) = 0,
∗
then that Zh,M
∗
→P
0,
as
letting
h → 0,
since
P,
P
o
n ∗ Ah,M ≡ P ∗ Zh,M
>δ >ε ,
if for any
and
in probability under
Qh,M
we have that
ε > 0, δ > 0, limh→0 P (Ah,M ) = 0.
This
are contiguous (see Remark 1). It follows
Qh,M .
The inverse follows similarly.
Theorem 2.1 provides a theoretical justication to derive bootstrap consistency results under the
approximate measure
Qh,M
as well as under
P.
This may simplify the bootstrap inference. We will
6
subsequently rely on this theorem to establish the bootstrap consistency results, when necessary (see
e.g., Theorem 4.2 and Theorem 4.3 in Section 4.2).
3
Results for realized volatility
3.1 Existing asymptotic theory
To describe the asymptotic properties of realized volatility, we need to introduce some notation. In
the following we assume that
1/M h
is an integer. For any
q > 0,
dene the realized
q -th
order power
variation as
1/M h Rq ≡ M h
X
j=1
RVj,M
Mh
q/2
.
(6)
PM
2
i=1 yi+(j−1)M is the realized volatility over the period
where
RVj,M =
1, . . . ,
1
M h . Note that when
q = 2, R2 = RV
[(j − 1) M h, jM h]
(realized volatility). Similarly, for any
q > 0,
for
j =
dene the
integrated power variation by
σq ≡
Z1
σuq du.
0
Mykland and Zhang (2009) show that
standard
χ2
distribution with
M
1
cM,q Rq
cM,q =
Γ is the Gamma function.
(CLT) result for
Rq
with
M
→
σ q , where
cM,q ≡ E
with
χ2M
the
(7)
Similarly, Mykland and Zhang (2009) provide a central limit theorem
bounded, whereas Mykland and Zhang (2011) based also on a contiguity
−1 , provided
the sample size h
Rq ,
q/2 q/2 Γ q+M
2
2
,
M
M
Γ 2
argument, but using a dierent representation than (4) allow the block size
M h → 0,
χ2M
M
degrees of freedom and
where
P
M
−1/2 ). Note that when
is of order O(h
without using any contiguity argument between
Qh,M
and
M
to go to innity with
M →∞
P,
as
h→0
such that
a CLT result still holds for
although this may involves the presence of de-biasing term, see e.g., Jacod and Protter (2012),
Jacod and Rosembaum (2013) (cf. equations (3.8) and (3.11)) and the recent work of Li, Todorov and
Tauchen (2016) (cf. Theorems 2 and 3).
We would like to highlight that for the special case
M → ∞, Rq
q = 2 (i.e.,
the case of realized volatility), when
enjoys a CLT without any bias correction term. Specically, for
q = 2,
as the number of
intraday observations increases to innity
√
h−1 R2 − σ 2 d
p
→ N (0, 1),
Vσ¯2
7
(8)
where
1
M cM,4 − c2M,2 Z
Vσ¯2 =
Note that for all
that for xed
c2M,2
0
M (cM,4 −c2M,2 )
M +2
, it follows that
M
c2M,2
M > 0, cM,2 = 1, cM,4 =
M,
σu4 du.
(8) holds true under both
P
and
Qh,M
= 2.
In addition, note
(see Mykland and Zhang (2009)). Whereas
M → ∞, (8) holds directly under the true measure P, as soon as h → 0 such that M 3 h → ∞ and
when
M 2h → 0
(see e.g., Jacod and Rosembaum (2013) (cf. equations (3.6), (3.8) and (3.11))). In practice,
result in (8) is infeasible since the asymptotic variance
integrated quarticity
R1
σu4 du.
For, xed
M,
Vσ¯2 ,M
depends on an unobserved quantity, the
Mykland and Zhang (2009) (cf.
Remark 8) propose a
0
consistent estimator of
have
P
Rq → σ q ,
M (cM,4 −c2M,2 ) 1
cM,4 R4 ). When
c2M,2
Vσ¯2 (V̂σ¯2 ,h,M =
M → ∞, cM,q → 1
and we still
see e.g., Jacod and Rosembaum (2013) (cf. equation (3.13) and Corollary 3.7), and
Theorem 3 of Li, Todorov and Tauchen (2016), see also Theorem 9.4.1 of Jacod and Protter (2012)
for similar result.
M → ∞.
Hence together with (8), we have the feasible CLT, for both cases: xed
It follows that as
M
and
h→0
√
TσMZ
¯2 ,h,M ≡
h−1 R2 − σ 2 d
q
→ N (0, 1),
V̂σ¯2 ,h,M
(9)
where we added the superscripts MZ to precise that the statistic is related to Mykland and Zhang
(2009) blocking method to avoid confusion when necessary.
When the block size
M
is xed such that
M = 1,
this result is equivalent to the CLT for realized
volatility derived by Barndor-Nielsen and Shephard (2002).
c1,4 = E
2
χ21
= 3.
We dene for
In particular,
c1,2 = E χ21 = 1,
and
M = 1, Tσ¯2 ,h = Tσ¯2 ,h,1 . Here, when M > 1, the realized volatility RV
using the blocking approach is the same realized volatility studied by Barndor-Nielsen and Shephard
(2002), but the
t-statistic
is dierent because
V̂σ¯2 ,h,M
changes with
M.
One advantage of the block-
based estimator is to improve eciency by varying the size of the block (see for e.g. Mykland, Shephard
and Sheppard (2012)).
3.2 Bootstrap consistency
Here we show that the new bootstrap method we proposed in Section 2 is consistent when applied to
realized volatility. Specically, given (5) with
r
∗
yi+(j−1)M
=
where
∗
ηi+(j−1)M
∼
i.i.d.N (0, 1) across
d = 1,
for
j = 1, . . . , 1/M h,
we let
RVj,M ∗
ηi+(j−1)M , i = 1, . . . , M,
M
(i, j).
Note that this bootstrap method is related to the wild
bootstrap approach proposed by Gonçalves and Meddahi (2009). In particular, when
RVj,M
M
= yj2
(10)
M = 1 and d = 1,
this amounts to the wild bootstrap based on a standard normal external random variable.
8
As the results of Gonçalves and Meddahi (2009) show, this choice of
on
σ2
because it does not replicate the higher order cumulants of
RV
η∗
is not optimal for inference
up to
o h1/2
propose a two-point distribution that matches the rst three cumulants up to
provides a second-order renement. We show that by replacing
RV
an asymptotically valid bootstrap method for
through order
o (h)
h1/2
. Instead, they
with the local average
and therefore
RVj,M
M , yields
and can matches the rst four cumulants of
even when volatility is stochastic.
RV
To see the gain of using the local Gaussian
dXt = σdWt
bootstrap method. Let consider a simplied model
without drift), where
yj2
o
d
(i.e. the constant volatility model
d
yi = σh1/2 νi , with νi ∼ i.i.d. N (0, 1) , and `=' denotes equivalence in distribution.
We have that
2
RVj,M = σ h
M
X
!
2
νi+(j−1)M
≡ σ 2 h · χ2M,j ,
i=1
where
σ 2 h,
χ2M,j
is i.i.d.∼
χ2M .
Thus,
E (RVj,M ) = σ 2 hE χ2M,j = σ 2 hM ,
the integrated volatility over the
2
of σ h, it is clear that choosing
RVj,M
is
variance of
M
2σ 4 h2 /M ,
h-horizon.
M > 1
Although
yi2
implying that
RVj,M
estimates
M
can also be viewed as an estimator
improves the eciency of the estimator (for given
which decreases with increasing
M ).
h,
the
The local Gaussian bootstrap
exploited this source of eciency while at the same time preserving the local Gaussianity inherent in
the semimartingale continuous time model driving
Xt .
Although (10) is motivated by this simplied
model, as we will prove below, this does not prevent the local Gaussian bootstrap method to be valid
more generally.
Its rst-order validity extends to the case where volatility is stochastic, there is a
leverage eect and the drift is non-zero.
In particular, it remains asymptotically valid for general
stochastic volatility models described by (1) and (2).
We dene the bootstrap realized volatility estimator as follows
R2∗ =
1/h
X
1/M h
yi∗2 =
i=1
where
∗ =
RVj,M
X
∗
RVj,M
,
j=1
PM
∗2
i=1 yi+(j−1)M . Letting
M
χ2j,M
1 X ∗2
ηi+(j−1)M ≡
,
M
M
i=1
it follows that
∗ =
RVj,M
χ2j,M
M
RVj,M .
E ∗ (R2∗ ) = cM,2 R2 ,
We can easily show that
and
Vσ∗¯2 ,h,M ≡ V ar∗ h−1/2 R2∗ = M cM,4 −µ22 R4 .
Hence, we propose the following consistent estimator of
V̂σ∗¯2MZ
=M
,h,M
Vσ∗¯2 ,h,M :
cM,4 − c2M,2
cM,4
9
R4∗ .
The bootstrap analogue of
TσMZ
¯2 ,h,M
is given by
√
Tσ∗¯2MZ
,h,M
≡
h−1 (R2∗ − cM,2 R2 )
q
.
V̂σ∗¯2MZ
,h,M
(11)
Theorem 3.1. Suppose (1), (2) and (5) hold. We have that
−c2
M,2 ) 1
a) For xed M, Vσ∗¯2 ,h,M = MM+2 V̂σ¯2 ,h,M , where V̂σ¯2 ,h,M = ( M,4
cM,4 R4 is a consistent estimator
c2M,2
of Vσ¯2 . If M is xed or M → ∞ as h → 0 such that M = o h−1/2 , then as h → 0,
M c
M
V ∗¯
− Vσ¯2 = oP (1) ,
M + 2 σ2 ,h,M
MZ
and sup P ∗ T̃σ∗¯2MZ
≤
x
−
P
T̃
≤
x
→ 0,
¯
2
,h,M
σ ,h,M
in probability under P , where T̃σ∗¯2MZ
≡
,h,M
b) If M is xed or M
→∞
x∈<
q
M
M +2
√
h−1 (R2∗ − cM,2 R2 ) and T̃σMZ
¯2 ,h,M ≡
√
h−1 R2 − σ 2 .
as h → 0 such that M = o h−1/2 , then as h → 0,
MZ
sup P ∗ Tσ∗¯2MZ
≤
x
−
P
T
≤
x
→ 0,
,h,M
σ¯2 ,h,M
x∈<
in probability under P.
Part a) of Theorem 3.1 justies using the local Gaussian bootstrap to estimate the entire distribution
of
T̃σ¯2 ,h,M ,
thus to construct bootstrap unstudentized (percentile) intervals for integrated volatility.
Whereas part b) of Theorem 3.1 provides a theoretical justication for using the bootstrap distribution
of
Tσ∗¯2 ,h,M
to estimate the entire distribution of
Tσ¯2 ,h,M . This result also justies the use of the bootstrap
for constructing the studentized bootstrap (percentile-t) intervals. Once again, it is worth emphasising
that these results hold under the general context studied by Mykland and Zhang (2009) which allow
for the presence of drifts and leverage eects under
P.
As the proof of Theorem 3.1 shows, here we
derive directly bootstrap results under the true probability measure
measure
Qh,M ;
P , without using any approximate
and /or the machinery of contiguity. Then we do not need to rely on Theorem 2.1 to
establish the bootstrap consistency results. The asymptotic validity of the local Gaussian bootstrap
depends on the availability of a CLT result of
R2
and a law of large numbers for
under the assumptions of Mykland and Zhang (2009) when
Jacod and Rosembaum (2013) when
M
Rq , which hold directly
is xed, and under the assumptions of
M → ∞.
To see the gain from the new local Gaussian bootstrap procedure, one should compare these results
with those of Gonçalves and Meddahi (2009). Results in part a) of Theorem 3.1 is in contrast to the
bootstrap methods studied by Gonçalves and Meddahi (2009), which are only valid for percentile-t
intervals i.e., when the statistic is normalized by its (bootstrap) standard deviation.
In particular,
the i.i.d. bootstrap variance estimator for the asymptotic variance of the realized volatility studied by
Gonçalves and Meddahi (2009) (cf. page 287) is given by
h−1
1/h
X
i=1

2
Z
1/h
X
yi4 − 
yi2  →P 3
0
i=1
10
1
σs4 ds −
Z
0
1
2
σs2 ds ,
which is equal to
2
R1
σs4 ds
(i.e.
the asymptotic conditional variance of the realized volatility) only
when the volatility is constant.
Similarly, the wild bootstrap variance estimator of Gonçalves and
0
Meddahi (2009) (cf. page 288) for the asymptotic variance of the realized volatility is given by

µ∗4
−
µ∗2
2
h−1
1/h
X

yi4 
→
µ∗4
P
i=1
where
µ∗q ≡ E ∗ |ν ∗ |q ,
with
ν∗
−
µ∗2
2
|
Z
·3
{z
≡VGM
0
1
σs4 ds,
}
(12)
the external random variable used in Gonçalves and Meddahi (2009) to
generate the wild bootstrap observations. In particular, using the best existing choice of
ν∗
(i.e. the
optimal two-point distribution) suggested in Proposition 4.5 of Gonçalves and Meddahi (2009) yield:
µ∗q
where
p=
1
2
−
q
q
q
q
√
√
1
1
31 + 186 · p + −
31 − 186 · (1 − p)
=
5
5
√ 3 . Thus, we have
186
µ∗2 ≈ 0.991,
It follows that
VGM
and
µ∗4 ≈ 1.217.
becomes
1
Z
σs4 ds
VGM ≈ 0.679 ·
Z
1
σs4 ds.
6= 2
0
0
This means that we would neither use the i.i.d.
bootstrap nor the best existing bootstrap method
of realized volatility (i.e. the optimal two-point wild bootstrap of Gonçalves and Meddahi (2009)) to
estimate
Vσ¯2 .
However, note that although the bootstrap methods in Gonçalves and Meddahi (2009)
do not consistently estimate
(percentile-t) intervals.
Vσ¯2 , their bootstrap methods are still asymptotically valid for studentized
It is also possible to consistently estimate the asymptotic variance of
R2 ,
by using the wild bootstrap of Gonçalves and Meddahi (2009) with certain external random variable.
Specically, given (12), we can show that a necessary condition for the consistency of the wild bootstrap
variance in Gonçalves and Meddahi (2009) is that
this choice of
ν∗
µ∗4 − µ∗2
2 =
2
3 . Unfortunately, it is easy to see that
does not deliver an asymptotic renement.
According to part a) of Theorem 3.1, the local Gaussian bootstrap variance estimator
out adjustment is not consistent for
(small) as
h → 0, Vσ∗¯2 ,h,M →P
Vσ¯2
M +2
M Vσ¯2
when the block size
6= Vσ¯2 ,
for large
mately). This explains the presence of the scale factor
of the local Gaussian bootstrap variance even when
this section, the estimator of
Vσ¯2
q
M
M
M
with-
is nite. In particular, for xed
(but bounded)
M
M +2 in
Vσ∗¯2 ,h,M
Vσ∗¯2 ,h,M → Vσ¯2
M
(approxi-
T̃σMZ
¯2 ,h,M , in order to restore consistency
is nite. In the univariate context considered in
is rather simple (it is given by a (scaled) version of
R4 ), but this is not
necessarily the case for other applications. For instance, for realized regression coecients dened by
the Mykland and Zhang's (2009) blocking approach the bootstrap percentile method could be useful
in that context.
As for all blocking methods, for the local Gaussian bootstrap there is an inherent trade-o at stake
11
by increasing the block size
M.
To gain further insight, let's consider once again the simplied model
h
Vσ¯2 = 2σ 4 , Vσ∗¯2 ,h,M = 2σ 4 M
dXt = σdWt .
It is easy to see that for this model
χ2M,j is i.i.d.∼
χ2M . A straightforward calculation shows that
Bias
Vσ∗¯2 ,h,M
4σ 2
=O
=
M
1
M
and
V ar
Vσ∗¯2 ,h,M
whereas the variance increases with increasing
j=1
χ2M,j
2
where
(M + 2) M 2 + 9M + 24 4
σ = O (M h) ,
= 4M h
M3
see Lemma B.3 in the Appendix for further details. Then the bias of
M,
P1/M h M.
Vσ∗¯2 ,h,M
decreases with increasing
Hence, in order to pick up any biases due to
the blocking in our local Gaussian bootstrap procedure, one should choose
M
very large. Of course,
∗
∗
this comes with a cost: large variance of V ¯2
. It follows that the mean squared error of V ¯2
σ ,h,M
σ ,h,M
2 + O (M h) . To minimize it, one should pick M proportional to h−1/3 , in which
is of order O 1/M
case
M SE Vσ∗¯2 ,h,M = O h−2/3 .
However, the optimal block size for mean squared error may be
suboptimal for the purpose of distribution estimation; see Hall, Horowitz and Jing (1995). We will not
pursue this approach further here. We leave this analysis to future work.
To summarize, Theorem 3.1 provides a theoretical justication for using the local Gaussian bootstrap approach as alternative to the existing bootstrap methods proposed by Gonçalves and Meddahi
(2009). Compared with the latter, there is no gain to use the local Gaussian bootstrap method when
M = 1.
However, when
M > 1,
M
with
large, in addition to consistently estimate the (conditional)
asymptotic variance of realized volatility, as we will see shortly (see Section 5.1 below), one can also
reduce the local Gaussian bootstrap error in estimating
cally, if the block length
M → ∞,
as
h→0
such that
P Tσ¯2 ,h ≤ x
M h → 0,
up to
oP (h).
More speci-
then the local Gaussian bootstrap is
an improvement of the best existing bootstrap methods of Gonçalves and Meddahi (2009). However,
in contrast to Gonçalves and Meddahi (2009) our new bootstrap method requires the choice of an
additional tuning parameter i.e., the block size
4
M.
Results for realized regression
Our aim in this section is to study the rst-order asymptotic properties of the local Gaussian bootstrap
method in multivariate setting. In particular, we investigate the validity of our approach for two dierent estimators of the slope coecient in a linear-time regression model for the continuous martingale
parts of two semimartingales observed over the xed time interval
[0, 1].
The rst is the estimator in
Barndor-Nielsen and Shephard (2004). The second is an estimator based on the results in Mykland
and Zhang (2009). To simplify the discussion, in the following, we consider the bivariate case where
d=2
and
and look at results for assets
yli ,
Let
k
and
l,
whose
ith
high frequency returns will be written as
yki
respectively.
Ĉ(j) ≡
Ĉkk(j)
0
Ĉlk(j) Ĉll(j)
!
q
Γ̂kk(j)
0
r

=  Γ̂kl(j)
q
Γ̂ll(j) −
Γ̂kk(j)
12

Γ̂2kl(j)
Γ̂kk(j)


be the Cholesky decomposition of
1
Mh
PM
0
i=1 yi+(j−1)M yi+(j−1)M . Given (5) with
∗
yk,i+(j−1)M
∗
yl,i+(j−1)M
!
= Ĉ(j) ·
√ 
h
for
j = 1, . . . , 1/M h,
!
and
i = 1, . . . , M,
∗
ηk,i+(j−1)M
,
∗
ηl,i+(j−1)M
q
∗
Γ̂kk(j) ηk,i+(j−1)M
r
√
h

=
d = 2,
Γ̂
q kl(j) η ∗
k,i+(j−1)M
Γ̂kk(j)
Γ̂ll(j) −
+

Γ̂2kl(j)
η∗
Γ̂kk(j) l,i+(j−1)M

,
(13)
where
Γ̂kk(j) =
and
M
M
M
1 X 2
1 X
1 X 2
yk,i+(j−1)M , Γ̂ll(j) =
yl,i+(j−1)M , Γ̂kl(j) =
yk,i+(j−1)M yl,i+(j−1)M ,
Mh
Mh
Mh
i=1
i=1
i=1
!
∗
ηk,i+(j−1)M
∗
ηl,i+(j−1)M
∼ i.i.d.N (0, I2 ),
I2
with
a
2×2
identity matrix.
4.1 Barndor-Nielsen and Shephard's (2004) type estimator
Under (1) and given our assumptions on
µ and σ
(Barndor-Nielsen and Shephard (2004) cf. equations
(25) and (26) in conjunction with, e.g., Jacod and Protter (2012)), we have that

1/h
P
2
yk,i

R1
− 0 Σkk,s ds




 i=1
1/h
 1/h
X
√
√
R1
 P
Sh = h−1 vech 
yi yi0  − vech (Γ) = h−1 
yk,i yl,i − 0 Σkl,s ds
 i=1
i=1
 1/h
 P
R
2 − 1Σ
yl,i
ll,s ds
0




 st
 −→ N (0, Π) ,



(14)
i=1
where
st
−→
denotes stable convergence and

R1
R1
R1
2 0 Σ2kl,s ds
2 0 Σkk,s Σkl,s ds
2 0 Σ2kk,s ds
R
R
R

 R
Π =  2 01 Σkk,s Σkl,s ds 01 Σkk,s Σll,s ds + 01 Σ2kl,s ds 2 01 Σll,s Σkl,s ds  .
R1
R1
R1
2 0 Σll,s Σkl,s ds
2 0 Σ2ll,s ds
2 0 Σ2kl,s ds

Barndor-Nielsen and Shephard (2004) propose the following consistent estimator of
−1
Π̂h = h
1/h
X
xi x0i
i=1
(15)
Π:
1/h−1
1 −1 X
− h
xi x0i+1 + xi+1 x0i ,
2
i=1
where
xi =
2
2
yk,i
yk,i yl,i yl,i
0
.
Thus
−1/2
Th = Π̂h
where here
d(d+1)
2
= 3,
since we let
d = 2.
st
Sh −→ N 0, I d(d+1) ,
(16)
2
We consider a general class of nonlinear transformations
that satisfy the following assumption. Throughout we let
13
∇f, (d × 1
vector-valued function) denote
the gradient of
f.
Assumption F:
The function
f : R
d(d+1)
2
→ R
is continuously dierentiable with
∇f (vech (Γ))
is
Γ.
non-zero for any sample path of
The corresponding statistics are dened as
√
Sf,h =
 

h−1 f vech 
1/h
X


yi yi0  − f (vech (Γ))
and
−1/2
Tf,h = Π̂f,h
Sf,h ,
i=1
where


Π̂f,h ≡ ∇0 f vech 
1/h
X


1/h
X
P
yi yi0  −→ Πf .
yi yi0  Π̂h ∇f vech 


(17)
i=1
i=1
In particular, we consider the question of estimating the integrated beta, i.e., the parameter
R1
βlk = R 01
0
Σkl,s ds
Σkk,s ds
.
(18)
Given (16) and by using the delta method, the asymptotic distribution of the realized regression
coecient

β̂lk = 
1/h
X
−1
2 
yk,i
i=1
obtained from regressing
yl,i
on
yk,i
1/h
X
yk,i yl,i ,
(19)
i=1
is
Sβ,h st
Tβ,h = q −→ N (0, 1) ,
V̂β
(20)
where
√
Sβ,h =

h−1 β̂lk − βlk ,
V̂β = 
1/h
X
−2
h−1 ĝβlk
2 
yk,i
such that
i=1
ĝβlk =
1/h
X
1/h−1
x2βi
−
i=1
X
xβi xβ,i+1 ,
and
2
xβi = yl,i yk,i − β̂lk yk,i
.
i=1
Sf,h is given by





1/h
1/h
X
X
√
= h−1 f vech 
yi∗ yi∗0  − f vech 
yi yi0  .
The local Gaussian bootstrap version of
 
∗
Sf,h,M
i=1
i=1
For the raw statistic
∗
Sf,h,M
=
∗
Sh,M
√
=
h−1
1/h
X
(x∗i − xi ) ,
i=1
where
x∗i =
∗2 y ∗ y ∗
∗2
yk,i
k,i l,i yl,i
14
0
.
It is elementary (though somewhat tedious) to show that
∗
E ∗ Sh,M
= 0,
and

2Γ̂2kk(j)
2Γ̂kk(j) Γ̂kl(j)
2Γ̂2kl(j)
X 

∗
Sh,M
= Mh
 2Γ̂kk(j) Γ̂kl(j) Γ̂kk(j) Γ̂ll(j) + Γ̂2kl(j) 2Γ̂kl(j) Γ̂ll(j) 
j=1
2Γ̂2kl(j)
2Γ̂kl(j) Γ̂ll(j)
2Γ̂2ll(j)
1/M h
Π∗h,M ≡ V ar∗

(see Lemma B.6 in the Appendix). Similarly, we denote the bootstrap analogue of
∗
Tβ,h by Tf,h,M
,
∗,
β̂lk
∗
Sβ,h,M
, and
Tf,h , β̂lk , Sβ,h ,
(21)
and
∗
Tβ,h,M
, respectively. In particular, we dene
−1/2
∗
∗
Sf,h.M
,
Tf,h,M
≡ Π̂∗f,h,M
where





1/h
1/h
X
X
yi∗ yi∗0  ,
≡ ∇0 f vech 
yi∗ yi∗0  Π̂∗h,M ∇f vech 

Π̂∗f,h,M
i=1
i=1
with

2Γ̂∗2
2Γ̂∗kk(j) Γ̂∗kl(j)
2Γ̂∗2
kk(j)
kl(j)
X 

∗
∗
= Mh
 2Γ̂∗kk(j) Γ̂∗kl(j) Γ̂∗kk(j) Γ̂∗ll(j) + Γ̂∗2
kl(j) 2Γ̂kl(j) Γ̂ll(j) 
j=1
2Γ̂∗2
2Γ̂∗kl(j) Γ̂∗ll(j)
2Γ̂∗2
kl(j)
ll(j)

2
2
4
Γ̂
4
Γ̂
Γ̂
2
Γ̂
+
Γ̂
Γ̂
kk(j) kl(j)
kk(j) ll(j)
1/M h
kk(j)
kl(j)
X 
1
2

4
Γ̂
Γ̂
Γ̂
Γ̂
+
3
Γ̂
4
Γ̂
Γ̂
· Mh
+
kk(j) kl(j)
kk(j) ll(j)
kl(j) ll(j)
kl(j)
 M
j=1
4 2
2
2 Γ̂kl(j) + Γ̂kk(j) Γ̂ll(j)
4Γ̂kl(j) Γ̂ll(j)
M Γ̂ll(j)
1/M h
Π̂∗h,M



,

such that
Γ̂∗kk(j) =
M
M
M
i=1
i=1
i=1
1 X ∗2
1 X ∗2
1 X ∗
∗
yk,i+(j−1)M , Γ̂∗ll(j) =
yl,i+(j−1)M , Γ̂∗kl(j) =
yk,i+(j−1)M yl,i+(j−1)M
.
Mh
Mh
Mh
We also let

∗
β̂lk
=
1/h
X
−1
∗2 
yk,i
∗
Sβ,h,M
=
∗
h−1 β̂lk
− β̂lk ,
and
∗
Tβ,h,M
∗ ∗
yk,i
yl,i ,
i=1
i=1
√
1/h
X
∗
Sβ,h,M
=q
∗
V̂β,h,M
such that
∗
V̂β,h,M

−2
1/h
X
∗2 
=
yk,i
h−1 ĝβ∗lk ,
i=1
with
ĝβ∗lk =
1/h
X
i=1
1/h−1
x∗2
βi −
X
x∗βi x∗β,i+1 ,
and
∗ ∗
∗ ∗2
x∗βi = yl,i
yk,i − β̂lk
yk,i .
i=1
Theorem 4.1. Suppose (1), (2) and (5) hold. Under Assumption F and the true probability distribution
P , if M → ∞ as h → 0 such that M = o h−1/2 , then as h → 0, the following hold
15
a)


Π∗f,h,M ≡ ∇0 f vech 
1/h
X



yi yi0  Π∗h,M ∇f vech 
1/h
X

P
yi yi0  −→ Πf ,
i=1
i=1
where Πf = p lim V ar (Sf,h ) , in particular
h→0
where
P
∗
V ar∗ Sβ,h,M
−→ Vβ ,
R
−2
1
Vβ = p lim V ar (Sβ,h ) ≡ 0 Σkk,s ds
B, with
h→0
Z
B=
1
2 2
Σ2kl,s + Σll,s Σkk,s − 4βlk Σkl,s Σkk,s + 2βlk
Σkk,s ds.
0
b)
P
∗
sup P ∗ Sf,h,M
≤ x − P (Sf,h ≤ x) −→ 0,
x∈R
in particular
P
∗
≤ x − P (Sβ,h ≤ x) −→ 0.
sup P ∗ Sβ,h,M
x∈R
c)
P
∗
≤ x − P (Tf,h ≤ x) −→ 0,
sup P ∗ Tf,h,M
x∈R
in particular
P
∗
≤ x − P (Tβ,h ≤ x) −→ 0.
sup P ∗ Tβ,h,M
x∈R
Theorem 4.1 justies using the local Gaussian bootstrap to estimate the distribution (and functionals of it such as the variance) of smooth function of realized covariance matrix
P1/h
0
i=1 yi yi , in particular
the realized beta. This contrasts with results in Dovonon et al. (2013) (cf. Theorem 5.1) where the
traditional pairs bootstrap is not able to mimic the score heterogeneity.
4.2 New variance estimator of Mykland and Zhang's (2009) realized beta type
estimator and bootstrap consistency
The goal of this subsection is to describe the realized beta in the context of Mykland and Zhang's
(2009) blocking approach. In order to obtain a feasible CLT, we propose a consistent estimator of the
variance of the realized beta, which is a new estimator in this literature. To derive this result, we use
the approach of Dovonon et al. (2013) and suppose that
σ
is independent of
Dovonon et al. (2013), we do not need here to suppose that
directly proceed under
1
Qh,M
µt = 0.
W.1
Note that contrary to
The main reason is because we can
(we will see shortly how). This is in fact one gain to revisit Dovonon et al.
We make the assumption of no leverage for notational simplicity and because this allows us to easily compute the
moments of the intraday returns conditionally on the volatility path. It would cleary be desirable to have a formal proof
of this result by allowing for leverage eect.
16
(2013) analysis with Mykland and Zhang's (2009) techniques of contiguity. In Dovonon et al. (2013),
letting
µt = 0,
is mathematically convenient in order to easily compute the moments of cross product
of high frequency returns. One can see that when
µt = 0,
we can simply write a high frequency return
of a given asset as the product of its volatility (the spot volatility) multiplied by a normal standard
distribution under the true probability measure
dXt = σt dWt ,
suppose that
where
equation (5)), conditionally on
σ,
σ
P.
Specically, Dovonon et al. (2013) (cf. Session 5.1)
is independent of
W.
Then, following Dovonon et al. (2013) (cf.
for a given frequency of the observations, we can write
yli = βlki yki + ui ,
where independently across
with
Γlki =
R ih
(i−1)h Σlk
i = 1, . . . , 1/h, ui |yki ∼ N (0, Vi ) ,
(u) du.
(22)
with
Vi ≡ Γli −
Γ2lki
Γki , and
βlki ≡
Γlki
Γki ,
2
Notice that (22) only makes sense in discrete time.
As Dovonon et al. (2013) argue, the conditional mean parameters of realized regression models are
heterogeneous under stochastic volatility. This heterogeneity justies why the pairs bootstrap method
that they studied is not second-order accurate.
Here we consider the general stochastic volatility model described by (1), but we rule out leverage
eects.
Given (4) (without assume that
but under
√
Qh,M .
It follows that under
√
µt = 0)
high frequency returns have similar representation
√
Qh,M , yk,i+(j−1)M = Ckk(j) hηk,i+(j−1)M
and
yl,i+(j−1)M =
1
M h where
Clk(j) hηk,i+(j−1)M + Cll(j) hηl,i+(j−1)M , for i = 1, . . . , M and j = 1, . . . ,
Ckk(j)
0
ηk,i+(j−1)M
C(j) ≡
, ηi+(j−1)M ≡
∼ i.i.d.N (0, I2 ),
Clk(j) Cll(j)
ηl,i+(j−1)M
0 with C
I2 is a 2×2 identity matrix, C(j) is the Cholesky decomposition of Σ(j) ≡ C(j) C(j)
(j) = σ(j−1)M h ,
i.e., the value of
σt
observables in the
at the beginning of the
j th
block (j
j -th
= 1, . . . , M1h ),
block. Under the approximate measure
Qh,M
the regression (22) becomes
yl,i+(j−1)M = βlk(j) yk,i+(j−1)M + ui+(j−1)M ,
where
for the
2 , and β
ui+(j−1)M |yk,i+(j−1)M ∼i.i.d.N 0, V(j) , for i = 1, . . . , M , with V(j) = hCll(j)
lk(j) ≡
(23)
Clk(j)
Ckk(j) .
β̌lk(j) the ordinary least squares (OLS) estimator of βlk(j) . To estimate the parameter
R1
by βlk =
0 βlk,s ds, Mykland and Zhang (2009) proposed to use β̌lk dened as follows,
!−1 M
!
1/M h
1/M h
M
X
X X
X
2
β̌lk = M h
β̌lk(j) = M h
yk,i+(j−1)M
yk,i+(j−1)M yl,i+(j−1)M .
(24)
Let us denote by
given
j=1
j=1
i=1
i=1
Note that the realized beta estimator discussed in Section 4.1.1 (see equation (18) and (19)) and also
studied among others by Dovonon et al. (2013) is a dierent statistic than the term given by (24).
Here, the realized beta estimator
average of
is not directly a least squares estimator, but is the result of the
β̌ lk(j) , the OLS estimators for each block.
in each block
2
β̌lk
j,
Since under
Qh,M , the volatility matrix is constant
implying consequently that the score is not heterogeneous and has mean zero. This
The underlying data generating process is in continuous time while one observes a discrete time sample of the process.
17
simplies the asymptotic inference on
βlk(j) , and on βlk .
It is worth emphasising that contrary to what
3
M = 1,
we have observed in the case of realized volatility estimator in Section 3.1 , here when
the
realized beta estimator using the blocking approach becomes
1/h
X
yl,i
,
β̌lk = h
yk,i
i=1
which is a dierent statistic than the statistic studied by Barndor-Nielsen and Shephard (2004). But
when
M = h−1 ,
when
M → ∞ with the sample size h−1 ,
both estimators are equivalent. However, as Mykland and Zhang (2011) pointed out,
the local approximation is good only when
M = O(h−1/2 ).
It
follows then that we are not comfortable to contrast Mykland and Zhang (2009) block-based "realized
M =
beta" estimator asymptotic results with those of Barndor-Nielsen and Shephard (2004a) when
h−1 .
M = h−1 )
The rst reason is that in this case (i.e.
M → ∞,
enough. The second and main reason is that when
know whether the approximate measure
Qh,M
the local approximation is not accurate
to the best of our knowledge we do not
with representation given by (4) is contiguous to
Mykland and Zhang (2009) provide a CLT result for
βlk .
In particular, we have under
as the number of intraday observations increases to innity (i.e.
Mykland and Zhang (2009), as
if
h → 0),
h → 0, for any δ > 0 such that M > 2 (1+δ)
√
h−1 β̌lk − βlk d
q
→ N (0, 1),
Vβ̌
P
and
P.
Qh,M ,
by using Section 4.2 of
with
M = O(1),
(25)
where
M
Vβ̌ =
M −2
Z
1
0
Σll,s
2
− βlk,s ds.
Σkk,s
In practice, this result is infeasible since the asymptotic variance
Vβ̌
depends on unobserved quantities.
Mykland and Zhang (2009) did not provide any consistent estimator of
to propose a consistent estimator of
Vβ̌ .
Vβ̌ .
To this end, we exploit the special structure of the regression
model. To nd the asymptotic variance of realized regression estimator
√
One of our contributions is
β̌lk ,
we can write
Xh
√ 1/M
β̌lk(j) − βlk(j) .
h−1 β̌lk − βlk = M h
j=1
Since
β̌lk(j)
are independent across
Vβ̌,h,M ≡ V ar
√
j,
it follows that
1/M h
h−1
β̌lk − βlk
2
=M h
X
V ar β̌lk(j) − βlk(j) .
(26)
j=1
3
Where the realized volatility estimator
R2
using Mykland and Zhang's (2009) blocking approach with
same realized volatility studied by Barndor-Nielsen and Shephard (2002).
18
M = 1,
is the
To compute (26), note that from standard regression theory, we have that under

M
X
V ar β̌lk(j) − βlk(j) = E 
Qh,M ,
!−1 
2
yk,i+(j−1)M
 V(j) ,
i=1
which implies that

1/M h
X
Vβ̌,h,M = M 2 h
!−1 
M
X
E
j=1
2
yk,i+(j−1)M
 V(j) .
(27)
i=1
Vβ̌ with equation (72) of Mykland and Zhang (2009). In fact, we can
PM 2
d
d
2
2
2
2
i=1 yk,i+(j−1)M = hCkk(j)
i=1 vi+(j−1)M = hCkk(j) χj,M , where vi+(j−1)M ∼
2
follow the standard χ distribution with M degrees of freedom. Then for any
Note that we can contrast
write under
Qh,M ,
i.i.d.N (0, 1)
and
integer
M >2
PM
χ2j,M
and conditionally on the volatility path, by using the expectation of the inverse of a
Chi square distribution we have,

E
M
X
!−1 
2
yk,i+(j−1)M
=E
i=1
!
1
1
−2
h−1 Ckk(j)
.
M −2
−2
h−1 Ckk(j)
=
χ2j,M
(28)
It follows then that
Vβ̌,h,M
1/M h X Cll(j) 2
M
=
Mh
.
M −2
Ckk(j)
j=1
Vβ̌,h,M
By using the structure of (27), a natural consistent estimator of
1/M h
V̂β̌,h,M ≡ M 2 h
where
X
M
X
j=1
i=1
!−1
!
,
(see Lemma B.13 and Lemma B.14 in the Appendix).
P
and
Qh,M
the feasible result
h−1 (β̌lk − βlk ) d
q
→ N (0, 1) .
V̂β̌,h,M
Next we show that the local Gaussian bootstrap method is consistent when applied to
realized beta estimator given by (24). Recall (13), and let
∗
from the regression of y
l,i+(j−1)M on
∗
β̌lk(j)
j.
X
β̌ ∗lk(j) .
j=1
∗
β̌lk
converges in probability (under
1/M h
β̌lk = M h
X
j=1

E∗ 
M
X
!−1
2∗
yk,i+(j−1)M
i=1
P ∗)
M
X
i=1
19
i.e., the
The bootstrap realized beta
1/M h
∗
β̌lk
= Mh
β̌lk
denote the OLS bootstrap estimator
∗
yk,i+(j−1)M
inside the block
estimator is
It is easy to check that
(29)
i=1
Together with the CLT result (25), we have under
Tβ̌,h,M ≡
1 X 2
ûi+(j−1)M
M −1
2
yk,i+(j−1)M
ûi+(j−1)M = yl,i+(j−1)M − β̌lk(j) yk,i+(j−1)M
√
M
is
to
!
∗
∗
.
yk,i+(j−1)M
yl,i+(j−1)M
ui+(j−1)M
The bootstrap analogue of the regression error
∗
β̌lk(j) yk,i+(j−1)M
,
∗
∗
β̌lk(j)
yk,i+(j−1)M
.
in model (23) is thus
whereas the bootstrap OLS residuals are dened as
∗
û∗i+(j−1)M = yl,i+(j−1)M
−
Thus, conditionally on the observed vector of returns
∗
u∗i+(j−1)M |yk,i+(j−1)M
∼ i.i.d.N 0, V̂(j)
, for
i = 1, . . . , M ,
∗
u∗i+(j−1)M = yl,i+(j−1)M
−
yi+(j−1)M ,
it follows that
where
2
.
V̂(j) ≡ hĈll(j)
We can show that for xed
M,
V ar∗
Hence, for any xed
Vβ̌ .
M,
√
M −1
∗
h−1 (β̌lk
− β̌lk ) =
V̂
.
M − 2 β̌,h,M
we would not use the local Gaussian bootstrap variance estimator to estimate
However, since we know the scaled factor
M −1
M −2 , this does not create a problem if we adjust the
bootstrap statistic accordingly. Our next theorem summarizes these results.
Theorem 4.2. Consider DGP (1), (2) and suppose (5) holds. Then conditionally on σ, under Qh,M
and P for xed M , the following hold, as h → 0, for any δ > 0 such that M > 2 (1+δ) with M = O(1),
a)
∗
Vβ̌,h,M
≡ V ar∗
P
→
√
∗
− β̌lk )
h−1 (β̌lk
M −1
V ,
M − 2 β̌
b)
sup P ∗
x∈R r
M − 2 √ −1 ∗
h
β̌lk − β̌lk ≤ x
M −1
!
−P
√
h−1
P
β̌lk − βlk → 0.
Part (a) of Theorem 4.2 shows that the bootstrap variance estimator is not consistent for
the block size
any
M
is nite. Note that for large
δ > 0 such that M > 2 (1+δ),
M −1
M −2
−1
M,
∗
but bounded V
β̌,h,M
∗
Vβ̌,h,M
→ Vβ̌ .
→ Vβ̌ ,
Vβ̌
when
(approximately) and for
Results in part (b) imply that the bootstrap
realized beta estimator has a rst-order asymptotic normal distribution with mean zero and covariance
matrix
Vβ̌ .
This is in line with the existing results in the cross section regression context, where the
wild bootstrap and the pairs bootstrap variance estimator of the least squares estimator are robust to
heteroskedasticity in the error term.
Bootstrap percentile intervals do not promise asymptotic renements. Next, we propose a consistent
bootstrap variance estimator that allows us to form bootstrap percentile-t intervals. More specically,
we can show that the following bootstrap variance estimator consistently estimates
1/M h
∗
V̂β̌,h,M
2
≡M h
X
M
X
j=1
i=1
!−1
∗2
yk,i+(j−1)M
!
M
1 X ∗2
ûi+(j−1)M .
M −1
i=1
20
V ∗β̌,h,M :
(30)
Our proposal is to use this estimator to construct the bootstrap t-statistic, associated with the bootstrap
realized regression coecient
∗.
β̌lk
Let
√
∗
≡
Tβ̌,h,M
be the bootstrap analogue of
∗ − β̌
h−1 β̌lk
lk
q
,
∗
V̂β̌,h,M
(31)
Tβ̌,h,M .
Theorem 4.3. Consider DGP (1), (2) and suppose (5) holds. Let
that M is bounded, conditionally on σ, as h → 0, the following hold
∗
∗
Tβ̌,h,M
→d N (0, 1) ,
Note that when the block size
M
M > 4 (2+δ)
for any δ > 0 such
in probability, under Qh,M and P.
is nite the bootstrap is also rst-order asymptotically valid
∗
when applied to the t-statistic T
(dened in (31) without any scaled factor), as our Theorem 4.3
β̌,h,M
∗
proves. This rst-order asymptotic validity occurs despite the fact that Vβ,h,M does not consistently
estimate
∗
V̂β̌,h,M
Vβ̌
M
is xed. The key aspect is that we studentize the bootstrap OLS estimator with
(dened in (30)), a consistent estimator of
bootstrap
5
when
t-statistic
∗
Vβ̌,h,M
,
implying that the asymptotic variance of the
is one.
Higher-order properties
In this section, we investigate the asymptotic higher order properties of the bootstrap cumulants.
Section 5.1 considers the case of realized volatility whereas Section 5.2 considers realized beta estimator
as studied by Barndor-Nielsen and Shephard (2004). The ability of the bootstrap to accurately match
the cumulants of the statistic of interest is a rst step to showing that the bootstrap oers an asymptotic
renement.
The results in this section are derived under the assumption of zero drift and no leverage (i.e.
is assumed independent of
σ ).
As in Dovonon et al. (2013), a nonzero drift changes the expressions
of the cumulants derived here. For instance, for the realized volatility, the eect of the drift on
√ OP
h .
W
Th
is
While this eect is asymptotically negligible at rst-order, it is not at higher orders. See
also the recent work of Hounyo and Veliyev (2016) (cf. equation (3.8)) where an additional term shows
up in the Edgeworth expansions for realized volatility estimator when the drift
µt
is non zero. The no
leverage assumption is mathematically convenient as it allows us to condition on the path of volatility
when computing the cumulants of our statistics.
Allowing for leverage is a dicult but promising
extension of the results derived here.
We introduce some notation. For any statistics
Th
and
Th∗ ,
we write
κj (Th ) to denote the j th
∗
∗
cumulant of Th and κj (Th ) to denote the corresponding bootstrap cumulant. For
denotes the coecient of the terms of order
O
√
h
21
of the asymptotic expansion of
j = 1
and
κj (Th ),
order
3, κj
whereas
for
j = 2
κ∗j,h,M
and
4 , κj
O (h).
denotes the coecients of the terms of order
The bootstrap coecients
are dened similarly.
5.1 Higher order cumulants of realized volatility
Here, we focus on the
t-statistic Tσ¯2 ,h
Tσ¯2 ,h
where
V̂σ¯2 ,h =
(µ4 −µ22 )
µ4
h−1
t-statistic Tσ∗¯2 ,h,M
√
h−1 R2 − µ2 σ 2
q
,
=
V̂σ¯2 ,h
and the bootstrap
dened by
(32)
P1/h
4
i=1 yi , and
√
Tσ∗¯2 ,h,M
=
h−1 (R2∗ − µ2 R2 )
q
,
V̂σ∗¯2 ,h,M
(33)
where
V̂σ∗¯2 ,h,M
1/M h h−1
M X RVj,M 2 1 X
µ4 − µ22 −1 X ∗4
µ4 − µ22
∗4
=
h
Mh
ηi+(j−1)M
− µ4 ,
yi =
µ4
µ4
Mh
M
i=1
i=1
such that
Rq,p ≡
∗
ηi+(j−1)M
∼ i.i.d.N (0, 1), µq = E |η ∗ |q
for
i=1
q > 0.
Let
σq,p ≡
σq
q/p ,
σp
( )
for any
q, p > 0,
and
Rq
. We make the following assumption.
(Rp )q/p
Assumption H.
volatility
σ
The log price process follows (1) with
µt = 0 and σt
is independent of
Wt , where the
is a càdlàg process, bounded away from zero, and satises the following regularity
condition:
lim h
1/2
h→0
for some
r>0
and for any
ηi
and
ξi
1/h
X
r
ση − σ r = 0,
ξ
i
i
i=1
such that
0 ≤ ξ1 ≤ η1 ≤ h ≤ ξ2 ≤ η2 ≤ 2h ≤ . . . ≤ ξ1/h ≤
η1/h ≤ 1.
Assumption H is stronger than required to prove the CLT for
R2 ,
and the rst-order validity of
the local Gaussian bootstrap, but it is a convenient assumption to derive the cumulants expansions of
Tσ¯2 ,h
and
Tσ∗¯2 ,h,M .
that for any
q>
Specically, under Assumption H, Barndor-Nielsen and Shephard (2004b) show
0, σhq
− σ q = o(
√
h), where σhq
=
h1−q/2
1/h
P
Rsh
s=1
(s−1)h
subsequently rely on to establish the cumulants expansion of
22
Tσ¯2 ,h .
!q/2
σu2 du
, a result on which we
The following result gives the expressions of
κj
and
κ∗j,h,M
for
j = 1, 2, 3
and
4.
We need to
introduce some notation. Let
A1 =
B1 =
√
µ6 − µ2 µ4
=
2
2,
1/2
µ4 µ4 − µ22
√
µ6 − 3µ2 µ4 + 2µ32
=
2
2,
3/2
µ4 − µ22
A2 =
µ8 − µ24 − 2µ2 µ6 + 2µ22 µ4
= 12,
µ4 µ4 − µ22
B2 =
µ8 − 4µ2 µ6 + 12µ22 µ4 − 6µ42 − 3µ24
= 12,
2
µ4 − µ22
C1 =
µ8 − µ24
32
= .
2
3
µ4
Theorem 5.1. Suppose (1) and (2) hold with µ = 0 and W independent of σ. Furthermore assume
(5). Under Assumption H, conditionally on σ and under P, it follows that, as h → 0
a)
κ1 Tσ¯2 ,h
κ2 Tσ¯2 ,h
κ3 Tσ¯2 ,h
κ4 Tσ¯2 ,h
=
√
hκ1 + o (h) ,
with κ1 = −
A1
σ6,4 ,
2
7
2
= 1 + hκ2 + o (h) , with κ2 = (C1 − A2 ) σ8,4 + A21 σ6,4
,
4
√
hκ3 + o (h) , with κ3 = (B1 − 3A1 ) σ6,4 ,
=
= hκ4 + o (h) ,
2
with κ4 = (B2 + 3C 1 − 6A2 ) σ8,4 + 18A21 − 6A1 B1 σ6,4
.
b)
κ∗1 Tσ∗¯2 ,h,M
κ∗2 Tσ∗¯2 ,h,M
κ∗3 Tσ∗¯2 ,h,M
κ∗4 Tσ∗¯2 ,h,M
√
=
hκ∗1,h,M + oP (h) ,
with κ∗1,h,M = −
A1
R6,4 ,
2
7
2
= 1 + hκ∗2,h,M + oP (h) , with κ∗2,h,M = (C1 − A2 ) R8,4 + A21 R6,4
,
4
√ ∗
=
hκ3,h,M + oP (h) , with κ∗3,h,M = (B1 − 3A1 ) R6,4 ,
= hκ∗4,h,M + oP (h) ,
2
.
with κ∗4,h,M = (B2 + 3C 1 − 6A2 ) R8,4 + 18A21 − 6A1 B1 R6,4
c) For j = 1, 2, 3 and 4, p limκ∗j,h,M − κj is nonzero if M is nite and it is zero if M
such that M = o
h→0
h−1/2 .
→∞
as h → 0
Theorem 5.1 states our main ndings for realized volatility. Part (a) of Theorem 5.1 are well known
in the literature (see e.g., Gonçalves and Meddahi (2009) (cf. Theorem A.1) and Hounyo and Veliyev
(2016)). These results are only given here for completeness. The remaining results in parts (b) and (c)
are new. Part (b) gives the corresponding results for the local Gaussian bootstrap. Part (c) shows that
the cumulants of
Tσ¯2 ,h
and
Tσ∗¯2 ,h,M
do not agree up to
o h1/2
when the block size
M
is xed (although
they are consistent), implying that the bootstrap does not provide a higher-order asymptotic renement
23
for nite values of
up to
o h1/2
M.
when
M
is nite is that
as
h→0
p limRq,p − σq,p
does not always equal to zero. Nevertheless,
h→0
M h → 0, p limRq,p − σq,p = 0, then the bootstrap matches the
h→0 1/2 , which implies that it provides a second-order
rst and third order cumulants through order O h
∗ T∗
renement, i.e. the bootstrap distribution P
≤
x
consistently estimates P Tσ¯2 ,h ≤ x with
¯
2
σ ,h,M
1/2 (assuming the corresponding Edgeworth expansions exist).4 This is
an error that vanishes as o h
−1/2 .
in contrast with the rst-order asymptotic Gaussian distribution whose error converges as O h
when
M →∞
The main reason why the local Gaussian bootstrap is not able to match cumulants
such that
Note that Gonçalves and Meddahi (2009) also proposed a choice of the external random variable for
their wild bootstrap method which delivers second-order renements. Our results for the bootstrap
method based on the local Gaussianity are new. We will compare the two methods in the simulation
section.
Theorem 5.1 also shows that the new bootstrap method we propose is able to match the second and
Tσ¯2 ,h
imply that
through
order O (h) when M → ∞ as h → 0. These
results
∗
distribution of T ¯2
consistently estimate the distribution of Tσ¯2 ,h through order
σ ,h,M
fourth order cumulants of
the bootstrap
o (h),
in which case the bootstrap oers a third-order asymptotic renement (this again assumes that
the corresponding Edgeworth expansions of
Tσ∗¯2 ,h,M
exist, something we have not attempted to prove
in this paper). If this is the case, then the local Gaussian bootstrap will deliver symmetric percentile-t
intervals for integrated volatility with coverage probabilities that converge to zero at the rate
o (h) .
In contrast, the coverage probability implied by the asymptotic theory-based intervals converge to the
desired nominal level at the rate
O (h) .
Remark 3. We would like to highlight that results in Theorem 5.1 are stated for the t-statistics Tσ¯2 ,h
∗MZ
and Tσ∗¯2 ,h,M (see (32) and (33)) but not for TσMZ
¯2 ,h,M and Tσ¯2 ,h,M (see (9) and (11)). However, under the
same conditions as in Theorem 5.1 but strengthened by piecewise constant assumption on the volatility
∗MZ
process σt , it is easy to see that analogue results as in Theorem 5.1 hold true for TσMZ
¯2 ,h,M and Tσ¯2 ,h,M
directly under P (without assuming that Qh,M is contiguous to P ). More specically, if the volatility
process σt is such that for s = 1, . . . , M1h ,
σt = σ(s−1)M h > 0,
for t ∈ ((s − 1) M h, sM h],
(34)
for any sample path of σ, then we can deduce that, when M → ∞ as h → 0, we also have
∗
∗MZ
−1/2
P TσMZ
≤
x
−
P
T
≤
x
=
o
h
,
P
¯2 ,h,M
σ¯2 ,h,M
∗ ∗MZ P TσMZ
= oP (h)
Tσ¯2 ,h,M ≤ x
¯2 ,h,M ≤ x − P
4
Recently, Hounyo and Veliyev (2016) rigorously justify the Edgeworth expansions for realized volatility derived by
As a consequence the cumulants expansions of Tσ¯2 ,h exist under our assumptions.
Our focus in parts (b) and (c) of Theorem 5.1 is on using on formal expansions to explain the superior nite sample
Gonçalves and Meddahi (2009).
properties of the new local Gaussian bootstrap theoretically (see e.g., Mammen (1993), Davidson and Flachaire (2001)
and Gonçalves and Meddahi (2009) for a similar approach). This approach does not seek to determine the conditions
under which the relevant expansions are valid.
24
under P. The proof follows exactly the same line as the proof of Theorem 5.1. For reasons of space we
leave the details for the reader. When M → ∞ condition (34) amounts almost to say that the volatility
is constant (σt = σ).
The potential for the local Gaussian bootstrap intervals to yield third-order asymptotic renements is particularly interesting because Gonçalves and Meddahi (2009) show that their wild bootstrap
method is not able to deliver such renements. Thus, our method is an improvement not only of the
Gaussian asymptotic distribution but also of the best existing bootstrap methods for realized volatility
in the context of no microstructure eects (where prices are observed without any error).
5.2 Higher order cumulants of realized beta
In this section, we provide the rst and third order cumulants of realized beta estimator given in (19).
These cumulants enter the Edgeworth expansions of the one-sided distribution functions of
∗
Tβ,h,M
,
i.e.,
∗
P ∗ Tβ,h,M
≤x
and
P (Tβ,h ≤ x),
Tβ,h
and
respectively. To describe the Edgeworth expansions,
we need to introduce additional notation. Let
Z
1
Σkk,s Σkl,s + βlk Σ2kk,s ds,
Ã0 =
0
Z
1
3 3
2 2
Σkk,s ds,
Σkk,s Σkl,s − 8βlk
2Σ3kl,s + 6Σkk,s Σkl,s Σll,s − 18βlk Σkk,s Σ2kl,s − 6βlk Σ2kk,s Σll,s + 24βlk
Ã1 =
0
Z
B̃ =
1
2 2
Σkk,s ds,
Σ2kl,s + Σkk,s Σll,s − 4βlk Σkk,s Σkl,s + 2βlk
0
H̃1 =
4Ã0
p
Γkk B̃
and
H̃2 =
Ã1
.
B̃ 3/2
Similarly we let
1/M h
Ã∗0,h,M
= Mh
X
1/M h
Γ̂kk(j) Γ̂kl(j) − β̂lk M h
j=1
= Mh
2Γ̂3kl(j) + 6Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) − 18β̂lk Γ̂kk(j) Γ̂2kl(j)
2 Γ̂2
3 3
−6β̂lk Γ̂2kk(j) Γ̂ll(j) + 24β̂lk
kk(j) Γ̂kl(j) − 8β̂lk Γ̂kk(j)
X
j=1
∗
B̃h,M
= Mh
Γ̂2kk(j) ,
j=1
1/M h
Ã∗1,h,M
X
1/M h X
!
,
2 2
Γ̂2kl(j) + Γ̂kk(j) Γ̂ll(j) − 4β̂lk Γ̂kk(j) Γ̂kl(j) + 2β̂lk
Γ̂kk(j) ,
j=1

∗
R̃1,h,M
= −
∗
H̃1,h,M
=
1
∗3/2
B̃h,M

Γ̂
Γ̂
Γ̂
kk(j)
kl(j)
ll(j)
j=1
j=1


P1/M h
 −19β̂lk M h P1/M h Γ̂kk(j) Γ̂2
−
5
β̂lk M h j=1 Γ̂2kk(j) Γ̂ll(j) 
j=1
kl(j)

,
P
P
1/M
h
1/M
h
2 Mh
2
3
3
+24β̂lk
j=1 Γ̂kk(j) Γ̂kl(j) − 8β̂lk M h
j=1 Γ̂kl(j)
4Ã∗0,h,M
q
∗
Γ̂kk B̃h,M
3M h
and
P1/M h
H̃2,h,M =
Γ̂3kl(j) + 5M h
Ã∗1,h,M
∗3/2
B̃h,M
,
where
P1/M h
Γ̂kk =
1/h
X
2
yk,i
.
i=1
Theorem 5.2. Suppose (1) and (2) hold with µ = 0 and W independent of σ. Furthermore assume
(5). Under Assumption H, conditionally on σ and under P, it follows that, as h → 0
25
a)
√ h ,
√ √
κ3 (Tβ,h ) =
hκ3 + o
h ,
√
κ1 (Tβ,h ) =
hκ1 + o
with κ1 =
1
H̃1 − H̃2 ,
2
with κ3 = 3H̃1 − 2H̃2 ,
b)
√
∗
κ∗1 Tβ,h,M
=
∗
κ∗3 Tβ,h,M
=
√
√ h ,
√ + oP
h ,
R̃∗
1 ∗
1,h,M
∗
H̃1,h,M − H̃2,h,M
+
,
2
M 1
∗
∗
= 3H̃1,h,M
− 2H̃2,h,M
+ oP
,
M
hκ∗1,h,M + oP
with κ∗1,h,M =
hκ∗3,h,M
with κ∗3,h,M
c) For j = 1 and 3, p limκ∗j,h,M − κj is nonzero if M is nite and it is zero if M
that M = o
Theorem 5.2 shows that the cumulants of
Tβ,h,M
implies that the error of the bootstrap approximation
of order
√ o
h .
→∞
h→0
h−1/2 .
T∗
agree
β,h,M
∗
≤x
P ∗ Tβ,h,M
and
through order
as h → 0 such
O
√ h ,
to the distribution of
Since the normal approximation has an error of the order
√ O
h ,
which
Tβ,h,M
is
this implies that
the local Gaussian bootstrap is second-order correct. This result is an improvement over the bootstrap
results in Dovonon, Gonçalves and Meddahi (2013), who showed that the pairs bootstrap is not secondorder correct in the general case of stochastic volatility. Thus, for realized beta, the local Gaussian
bootstrap is able to replicate the rst and third cumulants of
β̂lk
through order
o
√ h when M → ∞.
We conjecture that similar analysis at higher-order (possibly third-order asymptotic renement) holds
for the realized beta estimator, but a full exploration of this is left for future research.
6
Monte Carlo results
In this section we assess by Monte Carlo simulation the accuracy of the feasible asymptotic theory
approach of Mykland and Zhang (2009).
We nd that this approach leads to important coverage
probability distortions when returns are not sampled too frequently. We also compare the nite sample
performance of the new local Gaussian bootstrap method with the existing bootstrap method for
realized volatility proposed by Gonçalves and Meddahi (2009).
For integrated volatility, we consider two data generating processes in our simulations.
First,
following Zhang, Mykland and Aït-Sahalia (2005), we use the one-factor stochastic volatility (SV1F)
model of Heston (1993) as our data-generating process, i.e.
dXt = (µ − νt /2) dt + σt dBt ,
and
dνt = κ (α − νt ) dt + γ (νt )1/2 dWt ,
26
where
νt = σt2 , B
and
W
are two Brownian motions, and we assume
values are all annualized. In particular, we let
Corr(B, W ) = ρ.
The parameter
µ = 0.05/252, κ = 5/252, α = 0.04/252, γ = 0.05/252,
ρ = −0.5.
We also consider the two-factor stochastic volatility (SV2F) model analyzed by Barndor-Nielsen
5
et al. (2008) and also by Gonçalves and Meddahi (2009), where
dXt = µdt + σt dBt ,
σt = s-exp (β0 , +β1 τ1t + β2 τ2t ) ,
dτ1t = α1 τ1t dt + dB1t ,
dτ2t = α2 τ2t dt + (1 + φτ2t ) dB2t ,
corr (dWt , dB1t ) = ϕ1 , corr (dWt , dB2t ) = ϕ2 .
We follow Huang and Tauchen (2005) and set
α2 = −1.386, φ = 0.25, ϕ1 = ϕ2 = −0.3.
µ = 0.03, β0 = −1.2, β1 = 0.04, β2 = 1.5, α1 = −0.00137,
We initialize the two factors at the start of each interval by
drawing the persistent factor from its unconditional distribution,
−1
and by starting the
τ10 ∼ N 0, 2α
1
strongly mean-reverting factor at zero.
For integrated beta, the design of our Monte Carlo study is roughly identical to that used by
Barndor-Nielsen and Shephard (2004a), and Dovonon Gonçalves and Meddahi (2013) with a minor
dierence. In particular, we add a constant drift component to the design of Barndor-Nielsen and
Shephard (2004a). Here we briey describe the Monte Carlo design we use. We assume that
σ (t) dW (t) +
Σ (t) =
and
= Σ (t), where
2
Σ11 (t) Σ12 (t)
σ1 (t) σ12 (t)
,
=
σ21 (t) σ22 (t)
Σ21 (t) Σ22 (t)
σ12 (t) = σ1 (t) σ2 (t) ρ (t) .
2(2)
σ1 (t), where for
and
µ=
µ1
µ2
As Barndor-Nielsen and Shephard (2004a), we let
√
2(s)
2(s)
(s)
s = 1, 2, dσ1 (t) = −λs (σ1 (t) − ξs )dt + ωs σ1 (t) λs dbs (t),
component of a vector of standard Brownian motions, independent from
ξ1 = 0.110,
dX (t) =
µd (t), with σ (t) σ 0 (t)
ω1 = 1.346,
λ2 = 3.74,
ξ2 = 0.398,
and
ω2 = 1.346.
GARCH(1,1) diusion studied by Andersen and Bollerslev (1998):
W.
2(1)
σ12 (t) = σ1
bi
where
We let
(t) +
is the
i-th
λ1 = 0.0429,
Our model for
σ22 (t)
is the
dσ22 (t) = −0.035(σ22 (t) − 0.636)dt +
0.236σ22 (t)db3 (t). We follow Barndor-Nielsen and Shephard (2004), and let ρ(t) = (e2x(t) − 1)/(e2x(t) +
1),
where
x
let
µ1 = µ2 = 0.03.
follows the GARCH diusion:
We simulate data for the unit interval
an Euler scheme. We then construct the
dx(t) = −0.03(x(t) − 0.64)dt + 0.118x(t)db4 (t).
[0, 1].
The observed log-price process
h-horizon
returns
yi ≡ Xih − X(i−1)h
X
Finally, we
is generated using
based on samples of size
1/h.
Tables 1 and 2 give the actual rates of 95% condence intervals of integrated volatility and integrated
5
The function s-exp is the usual exponential function with a linear growth function splined in at high values of its
s-exp(x) = exp(x) if x ≤ x0 and s-exp(x) = √ exp(x20 ) 2 if x > xo , with x0 = log(1.5).
x0 −x0 +x
argument:
27
beta, computed over 10,000 replications. Results are presented for six dierent samples sizes:
1/h 1152,
576, 288, 96, 48, and 12, corresponding to 1.25-minute, 2.5-minute, 5-minute, 15-minute, halfhour and 2-hour returns. In Table 1, for each sample size we have computed the coverage rate by
varying the block size, whereas in Table 2 we summarize results by selecting the optimal block size.
We also report results for condence intervals based on a logarithmic version of the statistic
Th,M
and
its bootstrap version.
In our simulations, bootstrap intervals use 999 bootstrap replications for each of the 10,000 Monte
Carlo replications. We consider the studentized (percentile-t) symmetric bootstrap condence interval
method computed at the 95% level.
As for all blocking methods, to implement our bootstrap methods, we need to choose the block
M.
size
We follow Politis and Romano (1999) and Hounyo, Gonçalves and Meddahi (2013) and use
the Minimum Volatility Method. Although the block length
M
that result for this approach may be
suboptimal for the purpose of distribution estimation (in the context of realized volatility) as we do in
our application, it is readily accessible to practitioners, typically perform well, and allows meaningful
comparisons of the local Gaussian bootstrap and asymptotic normal theory-based methods. Here we
describe the algorithm we employ for a two-sided condence interval.
Algorithm: Choice of the block size M by minimizing condence interval volatility
(i)
For
M = Msmall to M = Mbig compute a bootstrap interval for the parameter of interest (integrated
volatility or integrated beta) at the desired condence level, this resulting in endpoints
and
(ii)
ICM,low
ICM,up .
For each
M
compute the volatility index
in a neighborhood of
M.
V IM
as the standard deviation of the interval endpoints
More specically, for a smaller integer
dard deviation of the endpoints
{ICM −l,low , . . . , ICM +l,low }
l,
let
V IM
equal to the stan-
plus the standard deviation of the
{ICM −l,up , . . . , ICM +l,up }, i.e.
v
v
u
u
l
l
u 1 X
2 u
1 X
t
¯
¯ up 2 ,
V IM ≡
ICM +i,low − IC low + t
ICM +i,up − IC
2l + 1
2l + 1
endpoints
i=−l
where
(iii)
¯ low =
IC
Pick the value
1
2l+1
M∗
Pl
i=−l
i=−l
ICM +i,low
and
¯ up =
IC
1
2l+1
Pl
i=−l
ICM +i,up .
corresponding to the smallest volatility index and report
{ICM ∗ ,low , ICM ∗ ,up }
as the nal condence interval.
One might ask what is a selection of reasonable
size
1/h = 1152,
the choices
Msmall = 1
and
1152, 576, 288, 96, and 48, we have used
used
l = 2
in our simulations.
Msmall
Mbig = 12
Msmall = 1
and
Mbig ?
In our experience, for a sample
usually suce, for the samples sizes :
and
Mbig = 12.
1/h =
For results in Table 2, we
Some initial simulations (not recorded here) showed that the actual
28
coverage rate of the condence intervals using the bootstrap is not sensitive to reasonable choice of l,
in particular, for
l = 1, 2, 3.
Starting with integrated volatility, the Monte Carlo results in Tables 1 and 2 show that for both
models (SV1F and SV2F), the asymptotic intervals tend to undercover. The degree of undercoverage
is especially large, when sampling is not too frequent.
for the log-based statistics.
It is also larger for the raw statistics than
The SV2F model exhibits overall larger coverage distortions than the
SV1F model, for all sample sizes.
When
M = 1,
the Gaussian bootstrap method is equivalent to
the wild bootstrap of Gonçalves and Meddahi (2009) that uses the normal distribution as external
random variable. One can see that the bootstrap replicates their simulations results. In particular,
the Gaussian bootstrap intervals tend to overcover across all models. The actual coverage probabilities
of the condence intervals using the Gaussian bootstrap are typically monotonically decreasing in
and does not tend to decrease very fast in
M
M,
for larger values of sample size.
A comparison of the local Gaussian bootstrap with the best existing bootstrap methods for realized
6
volatility
shows that, for smaller samples sizes, the condence intervals based on Gaussian bootstrap
are conservative, yielding coverage rates larger than
95% for the SV1F model.
The condence intervals
tend to be closer to the desired nominal level for the SV2F than the best bootstrap proposed by
Gonçalves and Meddahi (2009). For instance, for SV1F model, the Gaussian bootstrap covers
of the time when
87.42%.
h−1 = 12
96.51%
whereas the best bootstrap of Gonçalves and Meddahi (2009) does only
These rates decrease to
93.21%
and
80.42%
for the SV2F model, respectively.
We also consider intervals based on the i.i.d. bootstrap studied by Gonçalves and Meddahi (2009).
Despite the fact that the i.i.d. bootstrap does not theoretically provide an asymptotic renement for
two-sided symmetric condence intervals, it performs well.
While none of the intervals discussed here (bootstrap or asymptotic theory-based) allow for
h−1 ,
we have also studied this setup which is nevertheless an obvious interest in practice.
M =
For the
SV1F model, results are not very sensitive to the choice of the block size, whereas for the SV2F model
coverage rates for intervals using a very large value of block size (M
lower than
95%
even for the largest sample sizes. When
M=
= h−1 )
are systematically much
h−1 , the realized volatility
R2
using the
blocking approach is the same realized volatility studied by Barndor-Nielsen and Shephard (2002),
but the estimator of integrated quarticity using the blocking approach is
h−1 +2 2
R2 . This means that
h−1
2
1 2
, which is only valid under constant volatility. By
σ
dt
0
0 t
R
2
R1
1 2
Cauchy-Schwarz inequality, we have
≤ 0 σt4 dt, it follows then that we underestimated
0 σt dt
asymptotically we replace
R1
σt4 dt
by
R
the asymptotic variance of the realized volatility estimator.
This explains the poor performance of
the theory based on the blocking approach when the block size is too large. This also conrms the
theoretical prediction, which require
measure
6
√
M = O( h−1 )
for a good approximation for the probability
P.
The wild bootstrap based on Proposition 4.5 of Gonçalves and Meddahi (2009).
29
For realized beta, we see that intervals based on the feasible asymptotic procedure using Mykland
and Zhang's (2009) blocking approach and the bootstrap tend to be similar for larger sample sizes
whereas, at the smaller sample sizes, intervals based on the asymptotic normal distribution are quite
severely distorted. For instance, the coverage rate for the feasible asymptotic theory of Mykland and
Zhang (2009) when
to
95.17% (94.84%),
h−1 = 12
(cf.
h−1 = 48)
is only equal to
88.49% (92.86%),
whereas it is equal
for the Gaussian bootstrap (the corresponding symmetric interval based on the
pairs bootstrap of Dovonon Gonçalves and Meddahi (2013) yields a coverage rate of
93.59% (93.96%),
better than Mykland and Zhang (2009) but worse than the Gaussian bootstrap interval). Our Monte
Carlo results also conrm that for a good approximation, a very large block size is not recommended.
Overall, all methods behave similarly for larger sample sizes, in particular the coverage rate tends
to be closer to the desired nominal level. The Gaussian bootstrap performance is quite remarkable and
−1
outperforms the existing methods, especially for smaller samples sizes (h
7
= 12
and
48).
Empirical results
As a brief illustration, in this section we implement the local Gaussian bootstrap method with real
high-frequency nancial intraday data, and compare it to the existing feasible asymptotic procedure of
Mykland and Zhang (2009). The data consists of transaction log prices of General Electric (GE) shares
carried out on the New York Stock Exchange (NYSE) in August 2011. Before analyzing the data we
have cleaned the data. For each day, we consider data from the regular exchange opening hours from
time stamped between 9:30 a.m. till 4 p.m. Our procedure for cleaning data is exactly identical to
that used by Barndor-Nielsen et al. (2008). We detail in Appendix A the cleaning we carried out on
the data.
We implemented the realized volatility estimator of Mykland and Zhang (2009) on returns recorded
every
S
transactions, where
S
is selected each day so that there are 96 observations a day. This means
that on average these returns are recorded roughly every 15 minutes. Table 3 in the Appendix provides
the number of transactions per day, and the sample size used. Typically each interval corresponds to
about 131 transactions.
This choice is motivated by the empirical study of Hansen and Lunde (2006), who investigate
30 stocks of the Dow Jones Industrial Average, in particular they have presented detailed work for
the GE shares. They suggest to use 10 to 15 minutes horizon for liquid assets to avoid the market
microstructure noise eect.
Hence the main assumptions underlying the validity of the Mykland and Zhang (2009) block-based
method and our new bootstrap method are roughly satised and we feel comfortable to implement
them on this data.
To implement the realized volatility estimator, we need to choose the block size
Minimum Volatility Method described above to choose
30
M.
M.
We use the
We consider bootstrap percentile-t intervals, computed at the 95% level. The results are displayed
in Figure 1 in the appendix in terms of daily 95% condence intervals (CIs) for integrated volatility.
Two types of intervals are presented: our proposed new local Gaussian bootstrap method , and the the
feasible asymptotic theory using Mykland and Zhang (2009) blocking approach. The realized volatility
estimate
R2
is in the center of both condence intervals by construction. A comparison of the local
Gaussian bootstrap intervals with the intervals based on the feasible asymptotic theory using Mykland
and Zhang (2009) block-based approach suggests that the both types of intervals tend to be similar.
The width of these intervals varies through time.
However there are instances where the bootstrap
intervals are wider than the asymptotic theory-based interval. These days often correspond to days
with large estimate of volatility.
We have asked whether it will be due to jumps.
At this end we
have implemented the jumps test using blocked bipower variation of Mykland, Shephard and Sheppard
(2012). We have found no evidence of jumps at 5% signicance level for these two days. The gures
also show a lot of variability in the daily estimate of integrated volatility.
8
Concluding remarks
This paper proposes a new bootstrap method for statistics that are smooth functions of the realized
multivariate volatility matrix based on Mykland and Zhang's (2009) blocking approach.
We show
how and to what extent the local Gaussianity assumption can be explored to generate a bootstrap
approximation.
We use Monte Carlo simulations and derive higher order expansions for cumulants
to compare the accuracy of the bootstrap and the normal approximations at estimating condence
intervals for integrated volatility and integrated beta. Based on these expansions, we show that the
local Gaussian bootstrap provides second-order renements for the realized beta, whereas it provides
a third-order asymptotic renement for realized volatility.
This is an improvement of the existing
bootstrap results. Our new bootstrap method also generalizes the wild bootstrap of Gonçalves and
Meddahi (2009). Monte Carlo simulations suggest that the Gaussian bootstrap improves upon the rstorder asymptotic theory in nite samples and outperform the existing bootstrap methods for realized
volatility and realized betas.
An important assumption we make throughout this paper is that asset prices are observed at regular
time intervals and without any error (so that markets are frictionless).
If prices are nonequidistant
but non noisy, results in Mykland and Zhang (2009) show that the contiguity argument holds and the
same variance estimators of the realized covariation measures remain consistent.
same
t-statistics
Consequently, the
as those considered here can be used for inference purposes. In this case, we can rely
on the same local Gaussian bootstrap
t-statistics
to approximate their distributions.
The case of noisy data is much more challenging because in this case the realized covariation
measures studied in this paper are not consistent estimators of their integrated volatility measures.
Dierent microstructure noise robust measures are required. We conjecture that for the noise robust
31
measures based on pre-averaging approach (a dierent kind of blocking method) of Podolskij and Vetter
(2009) and Jacod et al. (2009), the contiguity results can be found under common types of noise. Since,
in these papers, the latent semimartingale is itself given a locally constant approximation. However,
for the bootstrap, it is worth emphasising that we have to distinguish two cases. The non-overlapping
pre-averaging based estimators as in Podolskij and Vetter (2009), and the overlapping pre-averaging
estimators as in Jacod et al.
(2009).
For the latter, the local Gaussian bootstrap would not be
valid, since the pre-averaged returns are very strongly dependent because they rely on many common
high frequency returns.
A dierent bootstrap method such as the wild blocks of blocks bootstrap
studied by Hounyo, Gonçalves and Meddahi (2013) will be appropriate. But we conjecture that, when
the pre-averaging method is apply on non-overlapping intervals (as in Podolskij and Vetter (2009)
and Gonçalves, Hounyo and Meddahi (2014)), the local Gaussian bootstrap method remains valid.
However, it is important to highlight that in this case, we should apply the local Gaussian bootstrap
method to the pre-averaged returns, instead of the raw returns.
Since in the presence of noise, the
non-overlapping pre-averaged returns remains asymptotically Gaussian and conditionally independent,
which are not the case for noisy high frequency raw returns.
Another promising extension would be to allow for jumps in the log-price process. Following the
initial draft of this paper, Dovonon et al.
(2014) generalize the local Gaussian bootstrap method
and use it for jump tests. In particular, instead of the local Gaussian bootstrap observations within
a given block will have a normal distribution with variance equal to the realized volatility over the
corresponding block. They propose to generate the bootstrap observations with normal distribution,
but now with a variance given by a local jumps-robust realized measure of volatility, for instance the
local block multipower variation estimator, similar to that of Mykland, Shephard and Sheppard (2012).
We also have supposed in multivariate framework that prices on dierent assets are observed synchronously, when prices are non-synchronous, dierent robust measures (eg.
Hayashi and Yoshida
(2005) covariance estimator) are required. Another important extension is to prove the validity of the
Edgeworth expansions derived here and to provide a theoretical optimal choice of the block size
M
for
condence interval construction. The extension of the local Gaussian bootstrap to these alternative
estimators and test statistics is left for future research.
Appendix A
This appendix is organized as follows. First, we details the cleaning we carried out on the data. Second,
we report simulation results. Finally we report empirical results.
Data Cleaning
In line with Barndor-Nielsen et al. (2009) we perform the following data cleaning steps:
(i)
Delete entries outside the 9:30pm and 4pm time window.
32
(ii)
Delete entries with a quote or transaction price equal to be zero.
(iii)
Delete all entries with negative prices or quotes.
(iv)
Delete all entries with negative spreads.
(v)
Delete entries whenever the price is outside the interval [bid
(vi)
− 2 ∗ spread ; ask + 2 ∗ spread].
Delete all entries with the spread greater or equal than 50 times the median spread of that day.
(vii)
Delete all entries with the price greater or equal than 5 times the median mid-quote of that day.
(viii)
Delete all entries with the mid-quote greater or equal than 10 times the mean absolute deviation
from the local median mid-quote.
(ix)
Delete all entries with the price greater or equal than 10 times the mean absolute deviation from
the local median mid-quote.
We report in Table 1 below, the actual coverage rates for the feasible asymptotic theory approach
and for our bootstrap methods.
In Table 2 we summarize results using the optimal block size by
minimizing condence interval volatility.
Table 3 provides some statistics of GE shares in August
2011.
33
Table 1. Coverage rates of nominal 95% CI for integrated volatility and integrated beta
Integrated volatility
Integrated beta
SV1F
Raw
M
CLT
Boot
SV2F
Log
CLT
Boot
Raw
CLT
Boot
Log
Raw
CLT
Boot
M
CLT
Boot
1/h = 12
1
85.44
98.49
90.08
97.86
80.38
96.62
86.17
96.24
2
83.58
95.67
2
85.56
97.31
90.31
96.80
80.43
94.70
86.27
94.73
3
87.57
94.98
3
85.71
96.46
90.84
96.08
80.34
93.77
85.89
93.70
4
89.15
94.81
4
85.88
96.20
90.97
95.93
80.34
92.88
85.52
92.89
6
90.65
94.47
12
86.11
94.84
91.27
94.87
77.66
88.89
81.65
86.97
12
90.46
93.62
1/h = 48
1
92.04
98.55
93.51
97.71
88.28
97.09
90.93
96.67
3
92.36
95.65
2
92.10
97.28
93.59
96.50
88.13
95.63
91.08
95.48
4
92.69
95.29
4
92.20
96.40
93.80
95.80
88.16
94.55
91.10
94.53
8
92.91
94.71
8
92.33
95.60
93.88
95.18
87.89
93.32
90.33
93.20
12
92.62
93.75
48
92.74
95.06
94.22
95.04
81.83
86.63
82.92
84.57
48
91.61
92.45
1/h = 96
1
93.35
97.94
94.09
97.10
90.20
97.06
92.10
96.66
3
92.60
95.51
2
93.43
96.78
93.99
96.06
90.37
95.84
92.24
95.67
4
93.12
95.01
4
93.47
95.78
94.03
95.61
90.46
94.70
92.09
94.83
8
93.78
94.81
8
93.50
95.26
94.09
95.32
90.07
93.81
91.75
94.01
12
93.79
94.55
96
93.42
94.80
94.35
94.87
81.93
84.61
82.79
83.60
96
91.97
92.36
1/h = 288
1
94.57
97.09
94.61
96.25
93.39
97.44
93.96
96.76
3
93.81
95.72
2
94.56
96.00
94.61
95.67
93.51
96.35
93.95
95.95
4
94.73
95.61
4
94.62
95.48
94.67
95.36
93.50
95.57
93.98
95.28
8
94.92
95.45
8
94.55
95.26
94.81
95.19
93.43
95.06
93.82
94.75
12
94.65
94.99
288
94.46
94.78
94.84
94.99
82.43
83.86
83.34
83.53
288
90.08
90.33
1/h = 576
1
94.53
96.12
94.75
95.84
94.19
96.96
94.49
96.52
3
93.93
95.58
2
94.57
95.53
94.68
95.41
94.17
96.23
94.52
95.78
4
94.49
95.36
4
94.74
95.15
94.70
95.16
94.32
95.59
94.56
95.45
8
94.51
94.88
8
94.67
95.08
94.72
94.96
94.22
95.38
94.46
95.16
12
94.48
94.86
576
94.58
94.85
94.76
94.92
82.01
82.37
82.05
82.32
576
87.09
87.09
1/h = 1152
1
95.06
96.06
95.16
95.70
94.51
96.52
94.47
95.95
3
94.75
95.91
2
95.13
95.68
95.20
95.65
94.53
95.79
94.47
95.42
4
94.90
95.45
4
95.05
95.49
95.20
95.31
94.42
95.21
94.50
95.11
8
94.83
95.13
8
95.15
95.47
95.18
95.20
94.39
95.03
94.47
94.85
12
94.95
94.85
1152
94.86
94.97
94.83
94.91
82.60
82.73
82.85
82.89
1152
81.69
81.60
Notes: CLT-intervals based on the Normal; Boot-intervals based on our proposed new local Gaussian bootstrap;
M
is the block size used to compute condence intervals. 10,000 Monte Carlo trials with 999 bootstrap
replications each.
34
35
94.49
94.50
5.37
5.86
5.66
5.82
6.01
48
96
288
576
1152
94.55
94.25
4.96
5.60
5.76
5.79
5.94
48
96
288
576
1152
94.68
94.72
94.75
94.98
93.96
93.60
PairsB
94.87
94.54
94.70
94.67
94.62
93.66
iidB
94.76
94.64
94.98
94.62
94.85
95.16
Boot
95.13
94.51
94.98
94.38
93.98
87.42
WB
95.04
94.95
95.05
95.29
95.76
96.51
Boot
95.15
94.58
94.60
93.68
92.31
90.79
CLT
Log
94.85
94.65
94.86
95.48
95.45
96.11
iidB
95.14
94.66
94.81
94.85
94.71
88.35
WB
95.02
94.98
94.96
95.11
95.35
96.07
Boot
6.25
6.06
5.94
5.81
5.75
3.84
M∗
SV2F
94.41
94.08
93.32
88.94
85.40
80.42
CLT
Raw
94.83
94.89
94.85
94.05
92.77
90.82
iidB
94.56
94.47
94.18
92.01
89.98
79.41
WB
94.92
95.19
95.02
94.28
93.63
93.21
Boot
94.38
94.38
93.62
90.48
87.59
86.70
CLT
Log
94.92
94.90
94.99
94.28
94.15
93.43
iidB
94.76
94.81
94.36
93.11
90.82
80.41
WB
94.83
95.03
94.90
94.21
93.80
93.35
Boot
Monte Carlo trials with 999 bootstrap replications each.
the pairs bootstrap of Dovonon Gonçalves and Meddahi (2013);
M∗
is the optimal block size selected by using the Minimum Volatility method. 10,000
on Proposition 4.5 of Gonçalves and Meddahi (2009); Boot-intervals based on our proposed new local Gaussian bootstrap; PairsB-intervals based on
Notes: CLT-intervals based on the Normal; iidB-intervals based on the i.i.d. bootstrap of Gonçalves and Meddahi (2009); WB-wild bootstrap based
94.75
93.70
92.87
88.49
3.62
12
CLT
M∗
n
Integrated beta
95.05
93.01
90.89
86.44
3.75
12
CLT
M∗
Raw
n
SV1F
Integrated volatility
Table 2. Coverage rates of nominal 95% intervals for integrated volatility and integrated beta using the optimal block size
Table 3. Summary statistics
Days
Trans
n
S
1 Aug
11303
96
118
2 Aug
13873
96
145
3 Aug
13205
96
138
4 Aug
16443
96
172
5 Aug
16212
96
169
8 Aug
18107
96
189
9 Aug
18184
96
190
10 Aug
15826
96
165
11 Aug
15148
96
158
12 Aug
12432
96
130
15 Aug
12042
96
126
16 Aug
10128
96
106
17 Aug
9104
96
95
18 Aug
15102
96
158
19 Aug
11468
96
120
22 Aug
10236
96
107
23 Aug
11518
96
120
24 Aug
10429
96
109
25 Aug
9794
96
102
26 Aug
9007
96
94
29 Aug
10721
96
112
30 Aug
9131
96
96
31 Aug
10724
96
112
Trans denotes the number of transactions,
every
S 'th
n
the sample size used to compute the realized volatility, and sampling of
transaction price, so the period over which returns are calculated is roughly 15 minutes.
Figure 1: 95% Condence Intervals (CI's) for the daily
σ2 ,
for each regular exchange opening days in August 2011,
calculated using the asymptotic theory of Mykland and Zhang (CI's with bars), and the new wild bootstrap
method (CI's with lines). The realized volatility estimator is the middle of all CI's by construction. Days on
the
x-axis.
36
Appendix B
This appendix concerns only the case where
d = 1 (i.e.
when the parameter of interest is the integrated
volatility). We organized this appendix as follows. First, we introduce some notation. Second, we state
Lemmas B.1 and B.2 and their proofs useful for proofs of theorem Theorem 5.1 presented in the main
text. These results are used to obtain the formal Edgeworth expansions through order
O(h) for realized
volatility. Finally, we prove Theorem 3.1, Lemma B.3 and Theorem 5.1.
Notation
To make for greater comparability, and in order to use some existing results, we have kept the
notation from Gonçalves and Meddahi (2009) whenever possible. We introduce some notation, recall
that, for any
q > 0, σ q ≡
R1
σuq du,
and let
RVj,M
σi2
h
q/2
σq,p,h ,
and
i=1
0
σq,p ≡
1/h P
σ¯hq ≡ h
q
when σ q is replaced with σh
( )
M
P
2
=
yi+(j−1)M
. We also let Rq,p ≡
σq
q/p ,
σp
we write
Rq
. Let
(Rp )q/p
i=1
and note that
µ2 = 1, µ4 = 3, µ6 = 15,
and
Rq ≡
µq = E |η|q
µ8 = 105.
σi2 ≡
Rih
σu2 du. We let
(i−1)h
1/M
Ph RVj,M q/2
Mh
, where
Mh
j=1
, where
η ∼ N (0, 1), with q > 0,
χ2M q/2
cM,q ≡ E
with
M
where
Recall that
cM,2 = 1, cM,4 = MM+2 ,
It follows by using the denition of cM,q gives in
equation (7) and this property of the Gamma function, for all x > 0, Γ (x + 1) = xΓ (x). Recall the
−1
(µ4 −µ22 ) −1 hP 4
denition of Tσ¯2 ,h given in (32) and V̂σ¯2 ,h =
h
yi .
µ4
χ2M the standard χ2 distribution with M degrees
+4)
(M +2)(M +4)(M +6)
cM,6 = (M +2)(M
and cM,8 =
.
M2
M3
of freedom.
Note that
i=1
We follow Gonçalves and Meddahi (2009) and we write
√
Tσ∗¯2 ,h,M
=
h−1 (R2∗ − µ2 R2 )
q
,
V̂σ∗¯2 ,h
where
V̂σ∗¯2 ,h,M
1/M h h−1
M X RVj,M 2 1 X
µ4 − µ22 −1 X ∗4
µ4 − µ22
∗4
h
yi =
Mh
ηi+(j−1)M
− µ4 .
=
µ4
µ4
Mh
M
i=1
i=1
i=1
We can write

Tσ∗¯2 ,h,M = Sσ∗¯2 ,h,M 
V̂σ∗¯2 ,h,M
−1/2
Vσ∗¯2 ,h,M

−1/2
√
= Sσ∗¯2 ,h,M 1 + hUσ∗¯2 ,h,M
,
where
√
Sσ∗¯2 ,h,M =
and
√
h−1 (R2∗ − µ2 R2 )
q
,
Vσ∗¯2 ,h,M
Uσ∗¯2 ,h,M ≡
h−1 V̂σ∗¯2 ,h − Vσ∗¯2 ,h,M
1/M
Ph RVj,M 2
Vσ∗¯2 ,h,M = V ar∗ n1/2 R2∗ = µ4 − µ22 · M h
.
Mh
i=1
37
Vσ∗¯2 ,h,M
,
RV q/2
q > 0, Mj,M
h Note that for any
with zero mean since
hq/2
M
q
M P
∗
ηi+(j−1)M − µq
∗
ηi+(j−1)M
∼ i.i.d. N (0, 1).
R2∗
−
P
i6=j6=k
=
σ
independent
and
V̂σ∗¯2 ,h,M −Vσ∗¯2 ,h,M
as follows
1/M h M 2
X RVj,M 1 X
∗
− µ2 R2 = M h
ηi+(j−1)M − µ2 ,
Mh M
Vσ∗¯2 ,h,M
i=1
1/M h M 4
X RVj,M 2 1 X
µ4 − µ22
∗
=
−
µ
Mh
η
i+(j−1)M 4 .
Mh M
µ4
j=1
i=1
P
Finally, note that throughout we will use
example
R2∗ − µ2 R2
We rewrite
j=1
V̂σ∗¯2 ,h,M
are conditionally on
i=1
i6=j6=...6=k to denote a sum where all indices dier, for
P
i6=j,i6=k,j6=k .
Lemma B.1. Suppose (1), (2) and (5) hold. Let M
q > 0, we have
≥1
such that M ≈ ch−α with α ∈ [0, 1) for any
a2)
q ∗
RV q/2 q/2
E ∗ yi+(j−1)M
h , for i = 1, . . . , M
= µq Mj,M
h √
Vσ∗¯2 ,h,M ≡ V ar∗
h−1 R2∗ = µ4 − µ22 R4 ,
a3)
E∗
a4)
i
h
2
E ∗ (R2∗ − µ2 R2 )4 = 3h2 µ4 −µ22 (R4 )2 + h3 µ8 − 4µ2 µ6 + 12µ22 µ4 − 6µ42 − 3µ24 R8 ,
a5)
h
i µ −µ2
E ∗ (R2∗ − µ2 R2 ) V̂σ∗¯2 ,h,M − Vσ∗¯2 ,h,M = 4µ4 2 (µ6 − µ2 µ4 ) hR6 ,
a6)
i h
µ −µ2
E ∗ (R2∗ − µ2 R2 )2 V̂σ∗¯2 ,h,M − Vσ∗¯2 ,h,M = 4µ4 2 h2 µ8 − µ24 −2µ2 µ6 + µ22 µ4 R8 ,
a7)
i
h
2
(µ4 −µ22 ) (µ6 −µ2 µ4 )
E (R2∗ − µ2 R2 )3 V̂σ∗¯2 ,h,M − Vσ∗¯2 ,h,M = 3h2
R4 R6 + OP h3 as h → 0,
µ4
a8)
h
i
E ∗ (R2∗ − µ2 R2 )4 V̂σ∗¯2 ,h,M − Vσ∗¯2 ,h,M
"
#
2
3
2
4 µ6 − 3µ2 µ4 + 2µ2 (µ6 − µ2 µ4 ) R
µ −µ
6
= h3 4µ4 2
+ OP h4 as h → 0,
2
2
2
+6 µ8 − µ4 − 2µ2 µ6 + 2µ2 µ4 µ4 −µ2 R4 R8
2 ∗
∗
∗
∗
E (R2 − µ2 R2 ) V̂σ¯2 ,h,M − Vσ¯2 ,h,M
= OP h2 as h → 0,
a1)
a9)
a10)
h
(R2∗
− µ2 R2 )
3
i
and j = 1, . . . , M1h
= h2 µ6 − 3µ2 µ4 + 2µ32 R6 ,
2 2
∗
∗
∗
(R2 − µ2 R2 ) V̂σ¯2 ,h,M − Vσ¯2 ,h,M
2 (µ4 −µ2 )
= h2 µ2 2
µ4 − µ22 µ8 − µ24 R4 R8 + 2 (µ6 − µ2 µ4 )2 R62 + OP h3
E∗
4
2 3
∗
∗
∗
(R2 − µ2 R2 ) V̂σ¯2 ,h,M − Vσ¯2 ,h,M
= OP h3
as h → 0,
a11)
E∗
a12)
2 4
∗
∗
∗
(R2 − µ2 R2 ) V̂σ¯2 ,h,M − Vσ¯2 ,h,M
i
2 h
2
2
2
(µ4 −µ2 )
= h3 µ2 2
3 µ4 − µ22
µ8 − µ24 R4 R8 +12 µ4 − µ22 (µ6 − µ2 µ4 ) R6 R4 +OP h4 as h → 0.
E∗
4
38
as h → 0,
Lemma B.2. Suppose (1), (2) and (5) hold. Let M
q > 0, we have
E ∗ Sσ∗¯2 ,h,M
E ∗ Sσ∗2
¯2 ,h,M
E ∗ Sσ∗3
¯2 ,h,M
E ∗ Sσ∗4
¯2 ,h,M
E ∗ Sσ∗¯2 ,h,M Uσ∗¯2 ,h,M
∗
U
E ∗ Sσ∗2
¯2 ,h,M σ¯2 ,h,M
≥1
such that M ≈ ch−α with α ∈ [0, 1) for any
= 0,
= 1,
√
=
hB1 R6,4 ,
= 3 + hB2 R8,4 ,
= A1 R6,4 ,
√
hA2 R8,4 ,
=
and as h → 0 we have,
3
E ∗ Sh,M
Uσ∗¯2 ,h,M
4
E ∗ Sh,M
Uσ∗¯2 ,h,M
E ∗ Sσ∗¯2 ,h,M Uσ∗2
¯2 ,h,M
∗2
E ∗ Sσ∗3
¯2 ,h,M Uσ¯2 ,h,M
∗2
E ∗ Sσ∗2
¯2 ,h,M Uσ¯2 ,h,M
∗2
E ∗ Sσ∗4
¯2 ,h,M Uσ¯2 ,h,M
= A3 R6,4 + OP (h) ,
√
2
h D1 R8,4 + D2 R6,4
+ OP h3/2 ,
=
= OP h1/2 ,
= OP h1/2 ,
2
= C1 R8,4 + C2 R6,4
+ OP (h) ,
2
= E1 R8,4 + E2 R6,4
+ OP (h) .
The constant A1 , B1 , A2 , B2 and C1 are dened in the text, and we have A3 = 3A1 , C2 = 2A21 ,
D1 = 6A2 , D2 = 4A1 B1 , E1 = 3C1 and E2 = 12A21 .
Proof of Lemma B.1 (a1) follows from
q q
∗
RVj,M q/2 q/2 ∗
yi+(j−1)M = M h h ηi+(j−1)M ,
for
∗
ηi+(j−1)M
∼ i.i.d. N (0, 1). For (a2) recall the denition of
and j = 1, . . . ,
∗
∗
that given result in (a1) we have E (R2 ) = µ2 R2 , then note that we can write
1
M h , where
R2∗
i = 1, . . . , M
R2∗ and remark
1/M h M 2
X RVj,M 1 X
∗
− µ2 R2 = M h
ηi+(j−1)M − µ2 .
Mh M
j=1
i=1
It follows that,
V ar∗
√
h−1 R2∗
= h−1 E ∗ [R2∗ − µ2 R2 ]2
−1
= h
1/M h 2
M
2
X RVj,M 2 1 X
∗ ∗
(M h)
E ηi+(j−1)M − µ2
Mh M2
2
j=1
=
=
i=1
1/M h X RVj,M 2
2
µ4 −µ2 M h
Mh j=1
2
µ4 −µ2 R4 .
39
For (a3), we have

3
M 2
X RVj,M h X
∗
= E ∗ M
ηi+(j−1)M − µ2 
Mh M

E ∗ [R2∗ − µ2 R2 ]3
1/M h i=1
j=1
1/M h 1/M h 1/M h M X
M X
M
X X X RVj ,M RVj ,M RVj ,M X
1
2
3
= h
Ii∗1 ,j1 ,i2 ,j2 ,i3 ,j3 ,
Mh Mh Mh 3
j1 =1 j2 =1 j3 =1
where
i1 =1 i2 =1 i3 =1
2
2
2
∗
∗
∗
∗
= E ηi1 +(j1 −1)M − µ2
ηi2 +(j2 −1)M − µ2
ηi3 +(j3 −1)M − µ2 .
Ii∗1 ,j1 ,i2 ,j2 ,i3 ,j3
easy to see that the only nonzero contribution to
Ii∗1 ,j1 ,i2 ,j2 ,i3 ,j3
is when
It is
(i1 , j1 ) = (i2 , j2 ) = (i3 , j3 ) ,
in
which case we obtain
Ii∗1 ,j1 ,i2 ,j2 ,i3 ,j3
3
M 2
X
∗
=
ηi+(j−1)M − µ2 = M µ6 − 3µ2 µ4 + 2µ32 .
i=1
Hence,
E
∗
[R2∗
3
− µ2 R2 ]
1/M h X RVj,M 3
3
= h
M h M µ6 − 3µ2 µ4 + 2µ2
j=1
2
= h µ6 − 3µ2 µ4 + 2µ32 R6 .
3
To prove the remaining results we follow the same structure of proofs as Gonçalves and Meddahi
(2009). Here
RVj,M q/2
Mh hq/2
M
q
M P
∗
ηi+(j−1)M − µq
plays the role of
i=1
|ri∗ |q − µq |ri |q ,
where
ri∗
denotes
the wild bootstrap returns in Gonçalves and Meddahi (2009).
Proof of Lemma B.2 Results follow immediately by using Lemma B.1 given the denitions of Sσ∗¯2 ,h,M
and
Uσ∗¯2 ,h,M .
M (cM,4 −c2M,2 ) 1
∗
cM,4 R4 and Vσ¯2 ,h,M =
c2M,2
M cM,4 −µ22 R4 the rst part of the result follows immediately, i.e. for xed M, Vσ∗¯2 ,h,M = MM+2 V̂σ¯2 ,h,M .
∗
For the second part, given that V ¯2
= MM+2 V̂σ¯2 ,h,M , we can write
σ ,h,M
Proof of Theorem 3.1 part a).
Given the denitions of
M
V ∗¯
− Vσ¯2
M + 2 σ2 ,h,M
V̂σ¯2 ,h,M =
= V̂σ¯2 ,h,M − Vσ¯2
= oP (1) ,
V̂σ¯2 ,h,M is a consitent estimator of Vσ¯2 . Next
h → 0 such that M 2 h → 0, then as h → 0,
MZ
sup P ∗ T̃σ∗¯2MZ
≤
x
−
P
T̃
≤
x
→ 0.
¯
2
,h,M
σ ,h,M
where consistency follows since
xed or
M →∞
as
x∈<
in probability under
P.
To this end, let
zj∗
r
=
T̃σ∗¯2MZ
=
,h,M
1/M
Ph
j=1
zj∗ ,
where
M √ −1
∗
∗
h
RVj,M
− E ∗ RVj,M
.
M +2
40
we show that, if
M
is
1/M
Ph
∗
Note that E
j=1
!
zj∗
= 0,
and
1/M
Ph
ar∗
V
j=1
!
P
zj∗
= V̂σ¯2 ,h,M → Vσ¯2 .
Moreover, since
are conditionally independent, by the Berry-Esseen bound, for some small
C>0
∗
z1∗ , . . . , z1/M
h
δ > 0 and for some constant
(which changes from line to line),
1/M h
X
2+δ
p ∗
∗MZ
sup P T̃σ¯2 ,h,M ≤ x −Φ x/ Vσ¯2 ≤ C
,
E ∗ zj∗ x∈<
j=1
which converges to zero in probability under
as
h → 0.
X
1/M h
2+δ
=
E ∗ zj∗ X
j=1
j=1
≤ 2
= 2
M ≥1
such that
M ≈ Ch−α
with
α ∈ [0, 1/2),
r
2+δ
√
M
∗
∗
h−1 RVj,M
− E ∗ RVj,M
E∗ M +2
M
M +2
M
M +2
where the inequality follows from the
and
R2(2+δ) ,
X
2+δ
E ∗ zj∗ ≤ 2
j=1
2
h
−(2+δ)
2
X
≤ C
2+δ
2
2+δ
P
M
∗2
η
1/M h
(j−1)M +i X
−(2+δ)
i=1
∗
h 2 E |RVj,M |2+δ ,
M
j=1
and the Jensen inequalities. Then, given the denitions of
M
M +2
2+δ
2
M
M +2
2+δ
2
δ
cM,2(2+δ) M 1+δ h 2 R2(2+δ)
P
1
cM,2(2+δ) R2(2+δ) →
1
δ
c2M,2(2+δ) h 2 −α(1+δ)
δ > 0, M ≥ 1 such that M ≈ Ch−α
− α(1 + δ) > 0,
∗ 2+δ
E ∗ RVj,M
j=1
Cr
Note that for any
1/M h
2+δ
we can write
1/M h
δ
2
for any
Indeed, we have that
1/M h
cM,2(2+δ)
P
with
σ 2(2+δ) = OP (1),
cM,2(2+δ)
R2(2+δ) .
α ∈ [0, 1/2), as h → 0,
and
cM,2(2+δ) → 1.
M
M +2
2+δ
2
= O (1) ,
Thus, results follow, in
particular, we have
1/M h
X
2+δ
E ∗ zj∗ = OP h
δ
−α(1+δ)
2
c2M,2(2+δ)
j=1
= oP (1) .
It is worthwhile to precise that for a bounded
M
(i.e.,
α = 0), the consistency result
P
1
cM,2(2+δ) R2(2+δ) →
σ 2(2+δ) = OP (1) follows from Mykland and Zhang (2009). Whereas when M → ∞, in particular for
α ∈ (0, 1/2) , the consistency result still holds since as M → ∞, cM,2(2+δ) → 1 and using e.g., Jacod
and Rosembaum (2013) (cf. equations (3.8) and (3.11)) or the recent work of Li, Todorov and Tauchen
(2016) (cf.
Theorems 2 and 3), we have
P
R2(2+δ) → σ 2(2+δ) .
Protter (2012) for similar result.
41
See also Theorem 9.4.1 of Jacod and
Proof of Theorem 3.1 part b).
probability under
P.
Given that
d
TσMZ
¯2 ,h,M → N (0, 1),
it suces that
d∗
Tσ∗¯2MZ
→ N (0, 1)
,h,M
in
Let
∗MZ
T̃h,M
,
Hσ∗¯2 ,h,M = q
V̂σ¯2 ,h,M
and note that
v
u
u M + 2 V̂σ¯2 ,h,M
∗
= Hσ¯2 ,h,M t
,
M V̂ ∗¯2MZ
Tσ∗¯2MZ
,h,M
σ ,h,M
where
Tσ∗¯2 ,h,M , V̂σ¯2 ,h,M
and
V̂σ∗¯2MZ
,h,M
are dened in the main text. Part a) of Theorem 3.1 proved that
d∗
Hσ∗¯2 ,h,M → N (0, 1) in probability under P .
under
P.
In particular, we show that (1)
in probability under
Thus, it suces to show that
Bias∗
M
∗MZ
M +2 V̂σ¯2 ,h,M
= 0,
V̂σ¯2 ,h,M
M
V̂ ∗MZ
M +2 σ¯2 ,h,M
and (2)
P∗
→ 1 in probability
V ar∗ MM+2 V̂σ∗¯2MZ
→ 0,
,h,M
P.
We have that
Bias
∗
M
V̂ ∗¯MZ
M + 2 σ2 ,h,M
M
M
∗
= E
V̂σ¯2 ,h,M −
V ∗¯
M +2
M + 2 σ2 ,h,M
= E ∗ V̂σ¯2 ,h,M − V̂σ¯2 ,h,M
∗
= V̂σ¯2 ,h,M − V̂σ¯2 ,h,M = 0,
we also have
V ar∗
M
V̂ ∗¯MZ
M + 2 σ2 ,h,M
=
M
M +2
2
V ar∗ V̂σ∗¯2 ,h,M ,
(35)
where
2
2 = E ∗ V̂σ∗¯2 ,h,M −Vσ∗¯2 ,h,M − E ∗ V̂σ∗¯2 ,h,M −Vσ∗¯2 ,h,M
V ar∗ V̂σ∗¯2 ,h,M

2
!2
1/M h
2
X
c
−
c
M,4
M,2
2∗
2

= M2
(M h)−2 E ∗ 
RVj,M
− cM,4 RVj,M
cM,4
j=1

2
!2
!2
1/M h
2
2
X
c
−
c
χ
M,4
j,M
M,2
4
(M h)−2
RVj,M
E∗ 
= M2
− cM,4  ,
cM,4
M
j=1
then given (35), the denitions of
V ar
∗
M
V̂ ∗¯MZ
M + 2 σ2 ,h,M
=
=
cM,2 , cM,4 , cM,8
M
M +2
2
M
M +2
2
M
2
M2
and
=
we can write
cM,4 − c2M,2
!2
−2
(M h)
cM,4
cM,4 − c2M,2
cM,8 −
c2M,4
1/M h
X
4
RVj,M
j=1
!2
(M h) cM,8 − c2M,4 R8
cM,4
2 2
M
2M
(M + 2) (M + 4) (M + 6) − M (M + 2)2
h
R8
M +2
M +2
M2
OP (M h)
=
R8 ,
→ 0
42
in probability under
P,
as long as
Mh → 0
as
Lemma B.3. Consider the simplied model
drift), then we have
h → 0.
dXt = σdWt
a1)
Bias
a2)
(M +2)(M 2 +9M +24) 4
V ar Vσ∗¯2 ,h,M = 4M h
σ
M3
Vσ∗¯2 ,h,M
=
4σ 2
M ,
Proof of Lemma B.3 part a1).
we can write
Given that
yi = Xih − X(i−1)h =d σh1/2 νi ,
RVj,M
(i.e. the constant volatility model without
2
= σ h
M
X
dXt = σdWt , for a given frequency h of the observations,
νi ∼ i.i.d. N (0, 1) . It follows that
!
with
2
νi+(j−1)M
= σ 2 h · χ2M,j ,
i=1
1/M h
R4 = (M h)
−1
X
1/M h
2
RVj,M
−1
4 2
= σ h (M h)
j=1
Vσ∗¯2 ,h,M
= M cM,4 −
X
χ2M,j
2
,
j=1
c2M,2
1/M h
2
h X
R4 = 2σ
χ2M,j ,
M
4
and
Vσ¯2 = 2σ 4 .
j=1
where
χ2M,j
χ2M . Thus,
Bias Vσ∗¯2 ,h,M
= E Vσ∗¯2 ,h,M − Vσ¯2
is i.i.d.∼
= 2σ 4
1/M h
2
h X
E χ2M,j − 2σ 4
M
j=1
4σ 2
h
(M h)−1 2M + M 2 − 2σ 4 =
.
M
M
1/M
P h 2 2
∗
4 h
that V ¯2
=
2σ
χM,j we can
M
σ ,h,M
= 2σ 4
Proof of Lemma B.3 part a2).
V ar
Vσ∗¯2 ,h,M
=
=
=
=
=
=
Proof of Theorem 5.1
Given
write
j=1
1/M h
2
h2 X
V ar χ2M,j
4σ
2
M
j=1
2
4 2 2
−1
8 h
2
2
4σ
(M h)
E χM,j − E χM,j
M2

!4 
!2 2 
2
2
2
χ
χ
h
M,j
M,j
 
− E
4σ 8 2 (M h)−1 M 4 E
M
M
M
4σ 8 M h cM,8 − c2M,4
"
#
2
(M
+
2)
(M
+
4)
(M
+
6)
(M
+
2)
4σ 8 M h
−
M3
M2
(M + 2) M 2 + 9M + 24 4
4M h
σ .
M3
8
Part (a) follows under our assumptions by using Theorem A.1 of Gonçalves
and Meddahi (2009). The proof of part (b) use the same technique as in the proof of Theorem A.3 in
43
Tσ∗¯2 ,h,M
Gonçalves and Meddahi (2009). In particular, the rst four cumulants of
are given by (e.g.,
Hall, 1992, p.42):
κ∗1 Tσ∗¯2 ,h,M
κ∗2 Tσ∗¯2 ,h,M
κ∗3 Tσ∗¯2 ,h,M
κ∗4 Tσ∗¯2 ,h,M
= E ∗ Tσ∗¯2 ,h,M ,
2
∗
= E ∗ Tσ∗2
Tσ∗¯2 ,h,M
,
¯2 ,h,M − E
3
∗
∗2
∗
∗
∗
= E ∗ Tσ∗3
−
3E
T
E
T
+
2
E
T
,
¯2 ,h,M
σ¯2 ,h,M
σ¯2 ,h,M
σ¯2 ,h,M
2
∗
∗
= E ∗ Tσ∗4
Tσ∗3
Tσ∗¯2 ,h,M − 3 E ∗ Tσ∗2
¯2 ,h,M − 4E
¯2 ,h,M E
¯2 ,h,M
4
∗
+12E ∗ Tσ∗2
Tσ∗¯2 ,h,M )2 − 6 E ∗ Tσ∗¯2 ,h,M
.
¯2 ,h,M (E
Our goal is to identify the terms of order up to
OP (h)
in the asymptotic expansions of these four
cumulants. We will rst provide asymptotic expansions through order
of
Tσ∗¯2 ,h,M
by using a Taylor expansion.
f (x) = (1 + x)−k/2
integer k ,
k = 1, · · · , 4,
0
yields
k,
O(h)
for the rst four moments
a second-order Taylor expansion of
f (x) = 1 − k2 x + k4 ( k2 + 1)x2 + O(x3 ).
We have that for any xed
−k/2
√ ∗
= Sσ∗k
+ OP ∗ (h3/2 ),
¯2 ,h,M 1 + hUσ¯2 ,h,M
k √ ∗k
k k
3/2
= Sσ∗k
hSσ¯2 ,h,M Uσ∗¯2 ,h,M + ( + 1)hSσ∗¯2 ,h,M Uσ∗2
)
¯2 ,h,M −
¯2 ,h,M + OP ∗ (h
2
4 2
3/2
≡ T̂σ∗k
).
¯2 ,h,M + OP ∗ (h
Tσ∗k
¯2 ,h,M
For
around
For a xed value
the moments of
Tσ∗k
¯2 ,h,M
up to order
O(h3/2 )
7
are given by
√
3
h ∗ ∗
E Sσ¯2 ,h,M Uσ∗¯2 ,h,M + hE ∗ Sσ∗¯2 ,h,M Uσ∗2
¯2 ,h,M
2
8 √ ∗ ∗2
∗2
∗2
∗
∗
= 1 − hE Sσ¯2 ,h,M Uσ¯2 ,h,M + hE ∗ Sσ∗2
E Tσ¯2 ,h,M
¯2 ,h,M Uσ¯2 ,h,M
√ 3
15
∗
∗3
∗3
∗
∗3
∗2
∗
∗
E ∗ Tσ∗3
=
E
S
−
+
S
U
S
U
h
E
hE
¯2 ,h,M
¯
¯
σ¯2 ,h,M
σ¯2 ,h,M σ¯2 ,h,M
2 σ2 ,h,M σ2 ,h,M
8 √
∗
∗4
∗4
∗2
∗
∗
∗4
∗
+
3hE
S
U
.
U
E ∗ Tσ∗4
=
E
S
−
2
hE
S
¯2 ,h,M
σ¯2 ,h,M
σ¯2 ,h,M σ¯2 ,h,M
σ¯2 ,h,M σ¯2 ,h,M
E
∗
Tσ∗¯2 ,h,M
where we used
= 0−
E ∗ Sσ∗¯2 ,h,M = 0,
E ∗ Tσ∗¯2 ,h,M
E ∗ Tσ∗2
¯2 ,h,M
E ∗ Tσ∗3
¯2 ,h,M
E ∗ Tσ∗4
¯2 ,h,M
Thus
7
and
E ∗ Sσ∗2
¯2 ,h,M = 1.
By Lemma B.2 in Appendix B, we have that
√
A1
h − R6,4 + OP (h3/2 ),
2
√
2
= 1 + h (C1 − A2 ) R8,4 + C2 R6,4
+ OP (h2 )
√
3
h B1 − A3 R6,4 + OP (h3/2 )
=
2
=
2
= 3 + h (B2 − 2D1 + 3E1 ) R8,4 + (3E2 − 2D2 ) R6,4
+ OP (h2 ).
√
κ∗1 Tσ∗¯2 ,h,M = − h A21 R6,4 , this proves the rst result.
To be strictly rigorous, we should be taking expected value of
T̂σ∗k
¯2 ,h,M
The remaining results follow similarly.
rather than of
(cf. p. 72) or Gonçalves and Meddahi (2009) (cf. p. 302) for similar approach.
44
Tσ∗k
¯2 ,h,M .
See e.g. Hall (1992)
For part (c), results follows by noting that
1
cM,q Rq
→ σq
use results in Section 4.1 of Myklang and Zhang (2009),
P . For xed M,
M → ∞ use Jacod and
in probability under
whereas when
Rosembaum (2013) (cf. equations (3.8) and (3.11)).
Appendix C
This appendix concerns the multivariate case. First, we show results introduced in Sections 3.1 and
4.1. Then, we show results for Sections 3.2 and 4.2.
Proof of Bootstrap results in Section 3.1 and 4.1: Barndor-Nielsen and Shephard's
(2004) type estimator
Notation
We introduce some notation. For
∗
β̂lk yk,i+(j−1)M
,
and let
j = 1, . . . , 1/M h,
and
i = 1, . . . , M,
∗
∗ y∗
ˆ∗i+(j−1)M = yl,i+(j−1)M
− β̂lk
k,i+(j−1)M
let
∗
∗i+(j−1)M = yl,i+(j−1)M
−
be the bootstrap OLS residuals. We
can write
√
∗
Tβ,h,M
P
√
1/h ∗ ∗
∗ − β̂
−1
−1/2
h−1 β̂lk
h
y
√ ∗
lk
i=1 k,i i
∗
q
=
1
+
=
S
,
hU
≡ r
β,h,M
β,h,M
P1/h ∗2 −2 ∗
∗
B̂
h,M
B̂h,M
i=1 yk,i
where
√
∗
=
Sβ,h,M
h−1
P
1/h ∗ ∗
i=1 yk,i i
√
q
∗
Bh,M
and
∗
≡
Uβ,h,M
∗
∗
− Bh,M
h−1 B̂h,M
∗
Bh,M
(36)
,
such that

∗
Bh,M


1/M h 1/h
X
X
∗ ∗ 
Γ̂2
= V ar∗  h−1 
i
= Mh
yk,i
√
kl(j)
2 2
+ Γ̂kk(j) Γ̂ll(j) − 4β̂lk Γ̂kk(j) Γ̂kl(j) + 2β̂lk
Γ̂kk(j) ,
j=1
i=1
and
∗
B̂h,M
Recall that
Note that


1
∗2
∗
∗
2
Γ̂
+
Γ̂
Γ̂
+
Γ̂
Γ̂
+
3
Γ̂
kk(j) ll(j)
X 
M
kl(j)
kk(j) ll(j)
kl(j)

= Mh

.
Γ̂2kk(j)
2
∗2
∗2
∗
∗
∗
Γ̂
Γ̂
+
2
β̂
Γ̂
−4
β̂
Γ̂
+
2
Γ̂
+
kk(j)
kl(j)
j=1
lk
lk
M
M
kk(j) kl(j)
kk(j)
1/M h
cM,q ≡ E
χ2M
M
cM,2 = 1, cM,4 =
q/2 M +2
M ,
with
χ2M
cM,6 =
the standard
(M +2)(M +4)
and
M2
χ2
distribution with
cM,8 =
M
degrees of freedom.
(M +2)(M +4)(M +6)
.
M3
Lemma B.4. Suppose (1), (2) and (5) hold. We have that, for any q1 , q2 ≥ 0 such that q1 + q2 > 0,
and for any k, l, k0 , l0 = 1, . . . , d,
1/M h q1 q2
X Mh
Γ̂kl(j) Γ̂k0 l0 (j) = OP (1) ,
j=1
where Γ̂kl(j) =
1
Mh
PM
i=1 yk,i+(j−1)M yl,i+(j−1)M ,
Proof of Lemma B.4.
for j = 1, . . . , 1/M h.
Apply Theorem 3 of Li, Todorov and Tauchen (2016).
45
Lemma B.5. Suppose (1), (2) and (5) hold. We have that, for any q1 , q2 ≥ 0 such that q1 + q2 > 0,
and for any k, l, k0 , l0 = 1, . . . , d, as soon as M → ∞ such that M h → 0, as h → 0.
a1)
Mh
a2)
Mh
1/M
Ph
j=1
1/M
Ph
j=1
Γ̂∗kl(j) −M h
1/M
Ph
j=1
Γ̂∗kl(j) Γ̂∗k0 l0 (j) −M h
Proof of Lemma B.5.
1/M
Ph
in probability-P.
P∗
Γ̂kl(j) Γ̂k0 l0 (j) −→ 0,
j=1
in probability-P.
We show that the results hold in quadratic mean with respect to
probability approaching one.
under
P∗
Γ̂kl(j) −→ 0,
P ∗,
with
This ensures that the bootstrap convergence also holds in probability
P.
For part a1). Recall (13), and notice that we can write
x∗i+(j−1)M =
∗2
yk,i+(j−1)M
∗
∗
yk,i+(j−1)M
yl,i+(j−1)M
∗2
yl,i+(j−1)M
0
,
such that
∗2
yk,i+(j−1)M
∗2
= Γ̂kk(j) ηk,i+(j−1)M
,
∗
∗
yk,i+(j−1)M
yl,i+(j−1)M
∗2
yl,i+(j−1)M
Γ̂2kl(j)
=
Γ̂kk(j)
2∗
ηk,i+(j−1)M
+2
+ Γ̂ll(j) −
where
∗
ηk,i+(j−1)M
∗
ηl,i+(j−1)M

1/M h
E ∗ M h
X
q
∗
∗
Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
,
2∗
= Γ̂kl(j) ηk,i+(j−1)M
+
Γ̂2kl(j)
Γ̂kl(j) q
∗
∗
Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
Γ̂kk(j)
!
2∗
ηl,i+(j−1)M
,
Γ̂kk(j)
!
∼ i.i.d.N (0, I2 ).


It follows that
!
M
X
1
∗
∗

yk,i+(j−1)M
yl,i+(j−1)M
Mh
1/M h
X
Γ̂∗kl(j)  = E ∗ M h
j=1
j=1
=
1/M h M
X X
i=1
∗
∗
E ∗ yk,i+(j−1)M
yl,i+(j−1)M
j=1 i=1
= h
1/M h M
X X
E
∗
2∗
Γ̂kl(j) ηk,i+(j−1)M
j=1 i=1
1/M h
= Mh
X
j=1
P
Z
Γ̂kl(j) −→
1
Σkk,s ds.
0
46
q
∗
∗
2
+ Γ̂kk(j) Γ̂ll(j) − Γ̂kl(j) ηk,i+(j−1)M ηl,i+(j−1)M
Similarly, we have

V ar∗ M h

1/M h
X

1/M h M
X X
Γ̂∗kl(j)  = V ar∗ 

∗
∗

yl,i+(j−1)M
E ∗ yk,i+(j−1)M
j=1 i=1
j=1
= h2
1/M h M
X X
q
∗
2∗
∗
+ Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
V ar∗ Γ̂kl(j) ηk,i+(j−1)M
j=1 i=1
= h2
1/M h M
X X
q
2
2∗
∗
∗
+ Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
E ∗ Γ̂kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
j=1 i=1
1/M h M h
X X
q
i2
2∗
∗
∗
E ∗ Γ̂kl(j) ηk,i+(j−1)M
+ Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
−h2
j=1 i=1
= Mh
2
1/M h X
2Γ̂2kl(j)
+ Γ̂kk(j) Γ̂ll(j) − M h
2
1/M h M
X X
j=1

= hM h
Γ̂2kl(j)
j=1 i=1
1/M h X

Γ̂2kl(j) + Γ̂kk(j) Γ̂ll(j)  = OP (h) = oP (1) .
j=1
{z
|
}
=OP (1)
The proof of part a2) follows similarly and therefore we omit the details.
Lemma B.6. Suppose (1), (2) and (5) hold. We have that,
a1)
∗
= 0,
E ∗ Sh,M
a2)

2Γ̂2kk(j)
2Γ̂kk(j) Γ̂kl(j)
2Γ̂2kl(j)


 2Γ̂kk(j) Γ̂kl(j) Γ̂kk(j) Γ̂ll(j) + Γ̂2kl(j) 2Γ̂kl(j) Γ̂ll(j) .
2Γ̂2kl(j)
2Γ̂kl(j) Γ̂ll(j)
2Γ̂2ll(j)

∗
= Mh
Π∗h,M ≡ V ar∗ Sh,M
1/M
Ph
j=1
Proof of Lemma B.6 part a1).
∗
E ∗ Sh,M
√
=
h−1
Given the denitions of
1/M h M
X X
∗ , x∗ ,
Sh,M
i
and
xi
we have
E ∗ x∗i+(j−1)M − xi+(j−1)M
j=1 i=1
√
=
h−1
1/M h M
X X
E
∗
x∗i+(j−1)M
j=1 i=1
where results follow, since we have
1/M
M
Ph P
j=1 i=1
−
√
h−1
1/M h M
X X
xi+(j−1)M = 0,
j=1 i=1
E∗
∗
∗
yk,i+(j−1)M
yl,i+(j−1)M
47
=
−1
hP
i=1
yk,i yl,i .
Proof of Lemma B.6 part a2).
Given the denition of
Π∗M,h ,
we have


1/h
X
√
= V ar∗  h−1
(x∗i − xi )
Π∗h,M
i=1
1/h
= h−1
X
V ar∗ (x∗i )
i=1
= h
−1
1/M h M X X
E
∗
x∗i+(j−1)M x∗0
i+(j−1)M
−E
∗
x∗i+(j−1)M
E
∗
x∗i+(j−1)M
0 j=1 i=1
= h
1/M h M
X X
0 ∗
∗
∗
∗
h−2 E ∗ x∗i+(j−1)M x∗0
−
E
x
E
x
.
i+(j−1)M
i+(j−1)M
i+(j−1)M
j=1 i=1
Let
∗
h−2 x∗i+(j−1)M x∗0
i+(j−1)M ≡ ai1 ,i2
1≤i1 ,i2 ≤3
.
It follows that
∗4
a∗1,1 = Γ̂2kk(j) ηk,i+(j−1)M
,
∗4
a∗2,1 = Γ̂kk(j) Γ̂kl(j) ηk,i+(j−1)M
+
a∗3,1
q
∗3
∗
Γ̂3kk(j) Γ̂ll(j) − Γ̂2kk(j) Γ̂2kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
,
q
∗3
∗
4∗
ηl,i+(j−1)M
= Γ̂2kl(j) ηk,i+(j−1)M
+ 2Γ̂kl(j) Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
∗2
2∗
+ Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
,
a∗1,2 = a∗2,1 ,
q
h
i2
∗
∗
2∗
a∗2,2 = Γ̂kl(j) ηk,i+(j−1)M
+ Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
,
v
u
Γ̂2kl(j)
Γ̂2kl(j) u
Γ̂3kl(j)
t
4∗
3∗
∗
∗
ηk,i+(j−1)M + 2 q
Γ̂ll(j) −
ηk,i+(j−1)M
ηl,i+(j−1)M
a3,2 =
Γ̂kk(j)
Γ̂
kk(j)
Γ̂kk(j)
!
Γ̂3kl(j)
2∗
2∗
+ Γ̂kl(j) Γ̂ll(j) −
ηk,i+(j−1)M
ηl,i+(j−1)M
Γ̂kk(j)
Γ̂2kl(j) q
3∗
∗
+
Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
Γ̂kk(j)
!
Γ̂2kl(j)
2∗
2∗
+2Γ̂kl(j) Γ̂ll(j) −
ηk,i+(j−1)M
ηl,i+(j−1)M
Γ̂kk(j)
!
Γ̂2kl(j) q
∗
3∗
+ Γ̂ll(j) −
Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) ηk,i+(j−1)M
ηl,i+(j−1)M
,
Γ̂kk(j)
48
(37)
a∗1,3 = a∗3,1 ,
a∗2,3 = a∗3,2 ,

2
Γ̂kl(j)
 Γ̂kl(j) 2∗
∗
a3,3 = 
ηk,i+(j−1)M + 2 q
Γ̂kk(j)
Γ̂
kk(j)
v
u
u
Γ̂2kl(j)
tΓ̂
∗
∗
ηk,i+(j−1)M
ηl,i+(j−1)M
+
ll(j) −
Γ̂kk(j)
Γ̂ll(j) −
Γ̂2kl(j)
2
!

2∗
ηl,i+(j−1)M
 .
Γ̂kk(j)
Hence, we have

3Γ̂2kk(j)
3Γ̂kk(j) Γ̂kl(j)
2Γ̂2kl(j) + Γ̂kk(j) Γ̂ll(j)


3Γ̂kk(j) Γ̂kl(j)
2Γ̂2kl(j) + Γ̂kk(j) Γ̂ll(j)
3Γ̂kl(j) Γ̂ll(j)
=
.
2
2
2Γ̂kl(j) + Γ̂kk(j) Γ̂ll(j)
3Γ̂kl(j) Γ̂ll(j)
3Γ̂ll(j)

h−2 E ∗ x∗i+(j−1)M x∗0
i+(j−1)M
(38)
Next, remark that given the denition of
x∗i+(j−1)M ,
and by using the linearization property of
E ∗ (·) ,
we have
0
.
E ∗ x∗i+(j−1)M = h Γ̂kk(j) Γ̂kl(j) Γ̂ll(j)
Thus, we can write

Γ̂2kk(j)
Γ̂kk(j) Γ̂kl(j) Γ̂kk(j) Γ̂ll(j)


Γ̂2kl(j)
Γ̂kl(j) Γ̂ll(j)  .
=  Γ̂kk(j) Γ̂kl(j)
Γ̂kk(j) Γ̂ll(j) Γ̂kl(j) Γ̂ll(j)
Γ̂2ll(j)

h−2 E ∗ x∗i+(j−1)M E ∗ x∗i+(j−1)M
0
(39)
Given (38) and (39), we have
−2
h

E
∗
x∗i+(j−1)M x∗0
i+(j−1)M
−E
∗
x∗i+(j−1)M
∗
E

x∗i+(j−1)M
0 2Γ̂2kk(j)
2Γ̂kk(j) Γ̂kl(j)
2Γ̂2kl(j)


=  2Γ̂kk(j) Γ̂kl(j) Γ̂kk(j) Γ̂ll(j) + Γ̂2kl(j) 2Γ̂kl(j) Γ̂ll(j)  .
2Γ̂2kl(j)
2Γ̂kl(j) Γ̂ll(j)
2Γ̂2ll(j)
Result follows by using (37).
Proof of Theorem 4.1.
nonlinear
We prove results for
f (z) = z;
the delta method implies the result for
f.
Part (a). Result follows from Theorem 3 of Li, Todorov and Tauchen (2016) by noting the following:
First the elements of
Π∗h,M
are all of the form of
Mh
1/M
Ph
Γ̂kl(j) Γ̂k0 l0 (j) .
Second, since
X
is continuous
j=1
in our framework, the truncation in equation (3.1) of Li, Todorov and Tauchen (2016) is useless: one
may use a treshold
vn = ∞
which reduce to our denition of
Part (b). Given the denition of
∗
Sh,M
=
√
h−1
1/M h M X X
∗ ,
Sh,M
Γ̂kl(j) .
we have
x∗i+(j−1)M − xi+(j−1)M =
√
j=1 i=1
1/M h
h−1
X
M
X
j=1
i=1
"
x∗i+(j−1)M − E ∗
M
X
!#
x∗i+(j−1)M
.
i=1
d(d+1)
P1/M h
λ ∈ R 2 such that λ0 λ = 1, supx∈R |P ∗ ( j=1 x̃∗j ≤ x)−
M
M
P
√
P ∗
P ∗
P
1/M h ∗
0
∗
0
∗
−1
xi+(j−1)M − E
xi+(j−1)M
. Note that, E ∗
x̃
Φ(x/ (λ Πλ))| → 0, where x̃j = h λ
j =
j=1
i=1
i=1
P
P
1/M h ∗
∗
0
0 and V ar∗
j=1 x̃j = λΠh,M λ → λ Πλ by part a). Thus, by Katz's (1963) Berry-Essen Bound,
The proof follows from showing that for any
49
for some small
ε>0
C > 0 which changes from line to line,


1/M h
Xh
X
∗ 1/M
∗
0


x̃j ≤ x − Φ(x/ λ Πλ ) ≤ C
sup P
E ∗ |x̃∗j |2+ .
x∈R j=1
j=1
Next, we show that
and some constant
P1/M h
j=1
1/M h
X
E ∗ |x̃∗j |2+ε = op (1).
We have that
"M
!#2+ε
M
√
X
X
0
∗
∗
∗
∗
−1
xi+(j−1)M − E
xi+(j−1)M =
E h λ
i=1
i=1
j=1
2+ε
1/M h
M
X
−(2+ε) X
≤ 22+ h 2
E ∗ λ0
x∗i+(j−1)M j=1
i=1
2+ε
1/M h
M
X
−(2+ε) X
E∗ ≤ 22+ h 2
x∗i+(j−1)M 1/M h
E ∗ |x̃∗j |2+ε
j=1
X
j=1
≤ 22+ Ch
−(2+ε)
2
i=1
M 1+ε
1/M h M
X X
2+ε
E ∗ x∗i+(j−1)M ,
j=1 i=1
where the rst inequality follows from the
Cr
and the Jensen inequalities; the second inequality uses the
Cauchy-Schwarz inequality and the fact that
We let
2
|z| =
(z 0 z) for any vector
z.
E ∗ |x̃∗j |2+ε ≤ Ch
−(2+ε)
2
1/M h M
X X
M 1+ε
j=1 i=1
j=1
≤ Ch
and the third inequality uses the
Cr
inequality.
Then, we have
1/M h
X
λ0 λ = 1;
−(2+ε)
2
1/M h M
X X
M 1+ε
d X
d X
2 1+ε/2
∗
∗
E∗ yk,i+(j−1)M
yl,i+(j−1)M
k=1 l=1
2+ε
∗ ∗
∗
E yk,i+(j−1)M yl,i+(j−1)M j=1 i=1
≤ Ch
2+ε
2
∗ 2(2+ε)
E ∗ ηk,1
M 2+ε
{z
|
+Ch
2+ε
2
∗
∗ 2+ε
E ∗ ηk,1
ηl,1
M 2+ε
|
{z
1+ε
≤ Ch
| M
{z }M h
=o(1)
|
1/M h q
2+ε
X Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) j=1
}
=O(1)
1/M h
ε
2
2+ε
X Γ̂kl(j) j=1
}
=O(1)
1/M h "
X
2+ε
2+ε 2
+ Γ̂kk(j) Γ̂ll(j) Γ̂kl(j) {z
}
=OP (1)
∗
∗
yk,i+(j−1)M
yl,i+(j−1)M
.
Cr
inequality, the third inequality uses in
Finally, the last inequality and the consistency follow
from Minkowski
inequality, Lemma B.4 and the fact that for any
α ∈ 0,
ε
2(1+ε)
Part (c).
,h
ε
2
M 1+ε
Given that
→ 0,
= oP (1) .
j=1
where the second and third inequalities follow from the
addition the denition of
#
h → 0.
d
Th → N 0, I d(d+1) ,
M
such that
M ≈ Ch−α
with
as
it suces that
2
d∗
∗
Th,M
→ N 0, I d(d+1)
2
50
in probability
under
P.
Next, note that from Lemma B.5
P∗
Π̂∗h,M − Π∗h,M −→ 0,
P.
in probability under
In addition, both
probability approaching one, as
h → 0.
Π̂∗h,M
and
Π∗h,M
are non singular in large samples with
Then, using results in parts (a) and (b) of Theorem 4.1, it
follows that
−1/2
1/2 ∗ −1/2 ∗ d∗
∗
Π∗h,M
Th,M
= Π̂∗h,M
Πh,M
Sh,M → N 0, I d(d+1) ,
2
{z
}
|
{z
}|
d∗
P∗
→N 0,I d(d+1)
→I d(d+1)
2
2
in probability-P.
Lemma B.7. Suppose (1), (2) and (5) hold. We have that,
a1)
E ∗ Γ̂∗3
kl(j) =
a2)
E ∗ Γ̂∗kk(j) Γ̂∗kl(j) Γ̂∗ll(j) =
a3)
E ∗ Γ̂∗kk(j) Γ̂∗2
kl(j) =
(M +2)(M +3)
Γ̂kk(j) Γ̂2kl(j)
M2
a4)
∗
Γ̂
E ∗ Γ̂∗2
kk(j) ll(j) =
M (M +2) 2
Γ̂kk(j) Γ̂ll(j)
M2
a5)
∗
E ∗ Γ̂∗2
Γ̂
kk(j) kl(j) =
M (M +2)(M +4) 2
Γ̂kk(j) Γ̂kl(j) ;
M3
a6)
∗2
E ∗ Γ̂∗2
Γ̂
kk(j) kl(j) =
M 3 +11M 2 +42M +36 2
Γ̂kk(j) Γ̂2kl(j)
M3
a7)
∗
Γ̂
E ∗ Γ̂∗3
kk(j) kl(j) =
M (M +2)(M +4)(M +6) 3
Γ̂kk(j) Γ̂kl(j) ;
M3
a8)
E ∗ Γ̂∗kk(j) Γ̂∗kl(j) =
M (M +2)
Γ̂kk(j) Γ̂kl(j) ;
M2
a9)
E ∗ Γ̂∗2
kl(j) =
(M +2)(M +1) 3
Γ̂kl(j)
M2
M +1 2
M Γ̂kl(j)
a10)
E ∗ Γ̂∗2
ll(j) =
a11)
E ∗ Γ̂∗2
kk(j) =
+
3(M +2)
Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) ;
M2
(M +2)2
Γ̂kk(j) Γ̂kl(j) Γ̂ll(j)
M2
+
+
3
+ 2 (MM+2)
2 Γ̂kl(j) ;
(M +2) 2
Γ̂kk(j) Γ̂ll(j) ;
M2
+
4(M +2)
Γ̂kk(j) Γ̂2kl(j) ;
M2
+
M 2 +2M +12 3
Γ̂kk(j) Γ̂ll(j) ;
M3
1
M Γ̂kk(j) Γ̂ll(j) ;
M +2 2
M Γ̂ll(j) ;
M +2 2
M Γ̂kk(j) .
Proof of Lemma B.7.
Γ̂∗kk(j) , Γ̂∗kl(j) Γ̂∗ll(j) and
"
#
M
1 X ∗2
ηk,i+(j−1)M ,
= Γ̂kk(j)
M
Note that given the denitions of
M
Γ̂∗kk(j)
1 X ∗2
yk,i+(j−1)M
=
Mh
i=1
(13), we have
i=1
M
Γ̂∗kl(j)
1 X ∗
∗
yk,i+(j−1)M yl,i+(j−1)M
=
Mh
i=1
!
M
q
1 X 2∗
= Γ̂kl(j)
ηk,i+(j−1)M + Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
M
i=1
51
M
1 X ∗
∗
ηk,i+(j−1)M ηl,i+(j−1)M
M
i=1
!
and
M
Γ̂∗ll(j)
=
1 X ∗2
yl,i+(j−1)M
Mh
i=1
2
Γ̂kl(j)
!
!
M
M
Γ̂kl(j) q
1 X 2∗
1 X ∗
∗
2
=
ηk,i+(j−1)M + 2
Γ̂kk(j) Γ̂ll(j) − Γ̂kl(j)
ηk,i+(j−1)M ηl,i+(j−1)M
M
Γ̂kk(j) M i=1
Γ̂kk(j)
i=1
!
!
M
X
Γ̂2kl(j)
1
2∗
+ Γ̂ll(j) −
,
ηl,i+(j−1)M
M
Γ̂kk(j)
i=1
!
∗
ηk,i+(j−1)M
∼ i.i.d.N (0, I2 ). By tedious but simple algebra, all results follow from the norwhere
∗
ηl,i+(j−1)M
∗
∗
mality and i.i.d properties of η
k,i+(j−1)M , ηl,i+(j−1)M across (i, j) . For instance, for (a1)
E
∗
Γ̂∗3
kl(j)
= E
∗
= E∗
Γ̂kl(j)
M
1 X 2∗
ηk,i+(j−1)M
M
i=1
!
M
1 X 2∗
ηk,i+(j−1)M
M
!
q
+ Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
M
1 X ∗
∗
ηk,i+(j−1)M ηl,i+(j−1)M
M
!!3
i=1
Γ̂3kl(j)
i=1
M
1 X 2∗
ηk,i+(j−1)M
M
q
+3Γ̂2kl(j) Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) E ∗
!2
i=1
+3Γ̂kl(j) Γ̂kk(j) Γ̂ll(j) −
Γ̂2kl(j)
E
∗
!
M
1 X ∗
∗
ηk,i+(j−1)M ηl,i+(j−1)M
M
i=1
!3
!2
i=1
M
1 X 2∗
ηk,i+(j−1)M
M
!
i=1
3/2
+ Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
E∗
M
1 X ∗
∗
ηk,i+(j−1)M ηl,i+(j−1)M
M
M
1 X ∗
∗
ηk,i+(j−1)M ηl,i+(j−1)M
M
i=1
= cM,6 Γ̂3kl(j) + 0
3
+ 3 Γ̂kl(j) Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j) E ∗
M
=
=
M
X
i=1
!
2∗
ηk,i+(j−1)M
M
X
!2
∗
∗
ηk,i+(j−1)M
ηl,i+(j−1)M
i=1
M (M + 2) (M + 4) 3
3M (M + 2)
2
Γ̂
+
Γ̂
Γ̂
Γ̂
−
Γ̂
kl(j)
kk(j)
ll(j)
kl(j)
kl(j)
M3
M3
(M + 2) (M + 1) 3
3 (M + 2)
Γ̂kl(j) +
Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) .
M2
M2
52
+0
For (a2), note that
Γ̂∗kk(j) Γ̂∗kl(j) Γ̂∗ll(j)
= Γ̂3kl(j)
M
1 X 2∗
ηk,i+(j−1)M
M
!3
i=1
!2
!
M
M
1 X ∗
1 X 2∗
∗
ηk,i+(j−1)M
ηk,i+(j−1)M ηl,i+(j−1)M
M
M
i=1
i=1
!
!
2
M
M
1 X
X
1
2∗
2∗
+ Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) − Γ̂3kl(j)
ηk,i+(j−1)M
ηl,i+(j−1)M
M
M
i=1
i=1
!
!
2
M
M
q
X
X
1
1
∗
2∗
∗
+Γ̂2kl(j) Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
ηl,i+(j−1)M
ηk,i+(j−1)M
ηk,i+(j−1)M
M
M
i=1
i=1
!
!2
M
M
1 X
X
1
2∗
∗
∗
+2 Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) − Γ̂3kl(j)
ηk,i+(j−1)M
ηk,i+(j−1)M
ηl,i+(j−1)M
M
M
i=1
i=1
!
!
!
M
M
M
3/2 1 X
X
X
1
1
2∗
2∗
∗
∗
+ Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
.
ηk,i+(j−1)M
ηl,i+(j−1)M
ηk,i+(j−1)M
ηl,i+(j−1)M
M
M
M
q
+2Γ̂2kl(j) Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
i=1
i=1
i=1
Then, we have
M (M + 2) (M + 4) 3
M 2 (M + 2) 3
Γ̂
+
0
+
Γ̂
Γ̂
Γ̂
−
Γ̂
kk(j) kl(j) ll(j)
kl(j)
kl(j)
M3
M3
M (M + 2)
+0 + 2
Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) − Γ̂3kl(j) + 0
M3
(M + 2)2
(M + 2) 3
=
Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) + 2
Γ̂kl(j) .
2
M
M2
=
E ∗ Γ̂∗kk(j) Γ̂∗kl(j) Γ̂∗ll(j)
For (a3), note that
2
Γ̂∗kk(j) Γ̂∗2
kl(j) = Γ̂kk(j) Γ̂kl(j)
M
1 X 2∗
ηk,i+(j−1)M
M
!3
i=1
!2
!
M
M
1 X ∗
1 X 2∗
∗
ηk,i+(j−1)M
ηk,i+(j−1)M ηl,i+(j−1)M
M
M
i=1
i=1
!
!2
M
M
1 X 2∗
1 X ∗
∗
ηk,i+(j−1)M
ηk,i+(j−1)M ηl,i+(j−1)M
,
M
M
q
+2Γ̂kk(j) Γ̂kl(j) Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
+ Γ̂2kk(j) Γ̂ll(j) − Γ̂kk(j) Γ̂2kl(j)
i=1
i=1
then we have
E ∗ Γ̂∗kk(j) Γ̂∗2
=
kl(j)
=
M (M + 2) 2
M (M + 2) (M + 4)
2
2
Γ̂
Γ̂
Γ̂
+
0
+
Γ̂
−
Γ̂
Γ̂
kk(j)
ll(j)
kk(j)
kl(j)
kk(j)
kl(j)
M3
M3
(M + 2) 2
(M + 2) (M + 3)
Γ̂kk(j) Γ̂2kl(j) +
Γ̂kk(j) Γ̂ll(j) .
M2
M2
53
For (a4), note that
2
∗
Γ̂∗2
kk(j) Γ̂ll(j) = Γ̂kk(j) Γ̂kl(j)
M
1 X 2∗
ηk,i+(j−1)M
M
!3
i=1
!2
!
M
M
1 X ∗
1 X 2∗
∗
ηk,i+(j−1)M
ηk,i+(j−1)M ηl,i+(j−1)M
M
M
i=1
i=1
!
!2
M
M
1 X 2∗
1 X 2∗
,
ηl,i+(j−1)M
ηk,i+(j−1)M
M
M
Γ̂kl(j) q
+2Γ̂2kk(j)
Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
Γ̂kk(j)
+ Γ̂2kk(j) Γ̂ll(j) − Γ̂kk(j) Γ̂2kl(j)
i=1
i=1
then we can write
∗
Γ̂
=
E ∗ Γ̂∗2
kk(j) ll(j)
=
M (M + 2) (M + 4)
M 2 (M + 2) 2
2
2
+
0
+
Γ̂
−
Γ̂
Γ̂
Γ̂
Γ̂
Γ̂
kk(j) kl(j)
kk(j) kl(j)
kk(j) ll(j)
M3
M3
M (M + 2) 2
4 (M + 2)
Γ̂kk(j) Γ̂ll(j) +
Γ̂kk(j) Γ̂2kl(j) .
2
M
M2
For (a5), note that
∗2
Γ̂∗2
kk(j) Γ̂kl(j)
=
Γ̂2kk(j) Γ̂2kl(j)
M
1 X 2∗
ηk,i+(j−1)M
M
!4
i=1
!3
!
M
M
1 X ∗
1 X 2∗
∗
ηk,i+(j−1)M
ηk,i+(j−1)M ηl,i+(j−1)M
M
M
i=1
i=1
!2
!2
M
M
1 X 2∗
1 X ∗
∗
ηk,i+(j−1)M
ηk,i+(j−1)M ηl,i+(j−1)M
,
M
M
q
2
+2Γ̂kk(j) Γ̂kl(j) Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
+ Γ̂3kk(j) Γ̂ll(j) − Γ̂2kk(j) Γ̂2kl(j)
i=1
i=1
then we have
M (M + 2) (M + 4) (M + 6) 2
Γ̂kk(j) Γ̂2kl(j) + 0
M4
15M + 3 M 2 − M + M 3 − M 2 3
2
2
+
Γ̂
Γ̂
−
Γ̂
Γ̂
ll(j)
kk(j)
kk(j)
kl(j)
M4
3
2
M 2 + 2M + 12 3
M + 11M + 42M + 36 2
2
=
Γ̂
Γ̂
+
Γ̂kk(j) Γ̂ll(j) .
kk(j)
kl(j)
M3
M3
∗2
Γ̂
=
E ∗ Γ̂∗2
kk(j) kl(j)
Result follows for (a6), by noting that
∗
Γ̂∗3
kk(j) Γ̂kl(j)
=
Γ̂3kk(j) Γ̂kl(j)
M
1 X 2∗
ηk,i+(j−1)M
M
!4
i=1
+Γ̂3kk(j)
q
Γ̂kk(j) Γ̂ll(j) − Γ̂2kl(j)
M
1 X 2∗
ηk,i+(j−1)M
M
i=1
54
!3
M
1 X ∗
∗
ηk,i+(j−1)M ηl,i+(j−1)M
M
i=1
!
,
and

M
1 X 2∗
E∗ 
ηk,i+(j−1)M
M
!4 
 = cM,8 = M (M + 2) (M + 4) (M + 6)
M4
i=1
E∗
M
1 X 2∗
ηk,i+(j−1)M
M
!3
M
1 X ∗
∗
ηk,i+(j−1)M ηl,i+(j−1)M
M
i=1
!
= 0.
i=1
The remaining results follow similarly and therefore we omit the details.
Lemma B.8. Suppose (1), (2) and (10) hold. We have that,
a1)
E∗
P
a2)
E∗
P
a3)
E∗
P
a4)
E∗
hP
1/h
h
hP
h
=
−
a5)
E∗
=
1/h ∗ ∗
i=1 yk,i i
= 0;
1/h ∗ ∗ 2
y
i=1 k,i i
∗ ;
= hB̃h,M
1/h ∗ ∗ 3
y
i=1 k,i i
= h2 Ã∗1,h,M ;
ii
ii
1/h ∗ ∗ P1/M h
Γ̂∗kk(j) Γ̂∗kl(j) − E ∗ Γ̂∗kk(j) Γ̂∗kl(j)
j=1
i=1 yk,i i
h
P1/M h
P1/M h
(M +2)
3M h j=1 Γ̂kk(j) Γ̂2kl(j) + M h j=1 Γ̂2kk(j) Γ̂ll(j)
M2
a6)
E∗
hP
a7)
E∗
P
a8)


2
P
1/h ∗ ∗

E∗ 
i=1 yk,i i
1/h ∗ ∗ P1/M h
i=1 yk,i i
j=1
1/h ∗ ∗ 4
y
i=1 k,i i
h
ii
∗ Γ̂∗2
Γ̂∗2
−
E
=
kk(j)
kk(j)
∗2 + O (h) ,
= 3h2 B̃h,M
P
as h → 0;
a9)
P1/M h ∗2
∗
∗
∗ ∗
Γ̂kl(j) + Γ̂∗kk(j) Γ̂∗ll(j) − E ∗ Γ̂∗2
j=1
i=1 yk,i i
kl(j) + Γ̂kk(j) Γ̂ll(j)
P1/M h
P1/M h
2(M +3)
M h j=1 Γ̂3kl(j) + 6MM+10
M h j=1 Γ̂kk(j) Γ̂kl(j) Γ̂ll(j)
2
M2
P1/M h
P1/M h
+1)
(6M +14)
β̂lk M h j=1 Γ̂kk(j) Γ̂2kl(j) − 2(M
β̂lk M h j=1 Γ̂2kk(j) Γ̂ll(j) ;
M2
M2


P
3
1/h ∗ ∗

E∗ 
i=1 yk,i i
P1/M h
j=1
i
Γ̂2kk(j) Γ̂kl(j) ;
i
h
P1/M h
P1/M h
M h j=1 Γ̂2kk(j) Γ̂kl(j) − β̂lk M h j=1 Γ̂3kl(j) ;
as h → 0;

i
P1/M h h ∗2
∗
∗
∗ Γ̂∗2
∗
∗
Γ̂
+
Γ̂
Γ̂
−
E
+
Γ̂
Γ̂
j=1
kl(j)
kk(j) ll(j)
kk(j) ll(j)
i  = OP h2 ,
kl(j)
P1/M h h ∗
∗ Γ̂∗2
+ j=1 Γ̂kk(j) Γ̂∗kl(j) − E ∗ Γ̂∗kk(j) Γ̂∗kl(j) + Γ̂∗2
kk(j)
kk(j) − E
i

P1/M h h ∗2
∗
∗
∗ Γ̂∗2
∗
∗
Γ̂
+
Γ̂
Γ̂
−
E
+
Γ̂
Γ̂
j=1
kl(j)
kk(j) ll(j)
kk(j) ll(j)
kl(j)
i 
P1/M h h ∗
∗
∗
∗
∗
∗2
+ j=1 Γ̂kk(j) Γ̂kl(j) − E Γ̂kk(j) Γ̂kl(j) + Γ̂kk(j) − E ∗ Γ̂∗2
kk(j)
∗ Ã∗2
= 3h2 B̃h,M
1,h,M + OP (h),
Proof of Lemma B.8.
4(M +2)
M2
− 4β̂lk M h
as h → 0.
In the following, recall that by denition for
55
j = 1, . . . , 1/M h,
and
i =
∗
∗
1, . . . , M, ∗i+(j−1)M = yl,i+(j−1)M
− β̂lk yk,i+(j−1)M
,




1/h
1/h
X
X
∗ ∗
∗ ∗
E∗ 
yk,i
i
= E∗ 
yk,i
i
i=1
and (13). For (a1), we can write
i=1
1/M h M h
X X
=
i
∗
2∗
∗
yl,i+(j−1)M
− β̂lk E ∗ yk,i+(j−1)M
E ∗ yk,i+(j−1)M
j=1 i=1
1/M h
X
= Mh
1/M h
Γ̂kl(j) − β̂lk M h
j=1
Γ̂kk(j)
j=1
1/M h
X
= Mh
X
M
1 X
yk,i+(j−1)M yl,i+(j−1)M
Mh
i=1
j=1
!
1/M h
X
i=1 yk,i yl,i
− P1/h
Mh
2
j=1
i=1 yk,i
P1/h
M
1 X 2
yk,i+(j−1)M
Mh
i=1
= 0.
For (a2), note that by denition

E∗ 
1/h
X
2

∗ ∗
i
= V ar∗ 
yk,i
i=1
1/h
X
i=1



1/h
X
∗ ∗
∗ ∗
i + E ∗ 
i .
yk,i
yk,i
i=1
P
1/h ∗ ∗
∗ , and the fact that by (a1) E ∗
B̃h,M
y
i=1 k,i i = 0. For the
∗ y ∗ ∗ and note that by denition, the z ∗0 s are conditionally
∗
∗ ∗
remaining results, write zi = yk,i i − E
i
k,i i
P
P1/h ∗ P1/h ∗ ∗
1/h ∗ ∗
∗
∗
∗
,
since
E
independent with E (zi ) = 0. Next, note also that
z
=
y
y
i=1 i
i=1 k,i i
i=1 k,i i = 0.
Then, result follows given the denition of
For (a3), note that

E∗ 
1/h
X
3
1/h
1/h
1/M h M
X
X
X X
∗3
∗ ∗
.
i
= E ∗  zi∗  =
E ∗ zi∗3 =
E ∗ zi+(j−1)M
yk,i
i=1
3

i=1
i=1
j=1 i=1
Thus, we can write

E∗ 
1/h
X
3
∗ ∗
i
yk,i
i=1
=
1/M h M
X X
h
i3
∗
∗
∗i+(j−1)M
E ∗ yk,i+(j−1)M
∗i+(j−1)M − E ∗ yk,i+(j−1)M
j=1 i=1
=

∗3
∗ y ∗2
∗2
∗ y∗
∗
E ∗ yk,i+(j−1)M
∗3
−
3E
E
i+(j−1)M
k,i+(j−1)M i+(j−1)M
k,i+(j−1)M i+(j−1)M

.
h i3
∗
∗
∗
+2 E yk,i+(j−1)M i+(j−1)M
i=1
1/M h M
X X
j=1

56
!
Next, to arrive at the desired expression for
1/M h M
X X
E∗
1/h ∗ ∗ 3
y
, note that
i=1 k,i i
P
∗3
E ∗ yk,i+(j−1)M
∗3
i+(j−1)M
j=1 i=1
1/M h M
X X
=
3 ∗
∗3
∗
E yk,i+(j−1)M yl,i+(j−1)M − β̂lk yk,i+(j−1)M
∗
j=1 i=1
3
= h
1/M h M
X X
6Γ̂3kl(j) + 9Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) − 36Γ̂kk(j) Γ̂2kl(j) β̂lk
2 Γ̂2
3 3
−9Γ̂2kk(j) Γ̂ll(j) β̂lk + 45β̂lk
kk(j) Γ̂kl(j) − 15β̂lk Γ̂kk(j)
j=1 i=1
−3
1/M h M
X X
!
,
∗2
∗
∗
∗
E ∗ yk,i+(j−1)M
∗2
E
y
i+(j−1)M
k,i+(j−1)M i+(j−1)M
j=1 i=1
= −3h
1/M h M X X
2 2
2Γ̂2kl(j) + Γ̂kk(j) Γ̂ll(j) − 6β̂lk Γ̂kk(j) Γ̂kl(j) + 3β̂lk
Γ̂kk(j)
Γ̂kl(j) − β̂lk Γ̂kk(j)
j=1 i=1
= h
1/M h M
X X
2 Γ̂2
−6Γ̂3kl(j) − 3Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) + 18β̂lk Γ̂kk(j) Γ̂2kl(j) − 9β̂lk
kk(j) Γ̂kl(j)
2
3 Γ̂3
2
2
2
+6β̂lk Γ̂kk(j) Γ̂kl(j) + 3β̂lk Γ̂kk(j) Γ̂ll(j) − 186β̂lk Γ̂kk(j) Γ̂kl(j) + 9β̂lk
kk(j)
j=1 i=1
!
,
and
2
1/M h M h
X X
i3
∗
∗
∗
yl,i+(j−1)M
− β̂lk yk,i+(j−1)M
E ∗ yk,i+(j−1)M
j=1 i=1
3
= 2h
1/M h M X X
Γ̂kl(j) − β̂lk Γ̂kk(j)
3
j=1 i=1
= h3
1/M h M X X
3 3
2 2
Γ̂kk(j) .
2Γ̂3kl(j) − 6β̂lk Γ̂kk(j) Γ̂2kl(j) + 6β̂lk
Γ̂kk(j) Γ̂kl(j) − 2β̂lk
j=1 i=1
For (a4), note that

1/h
1/M h h
X
X
∗
∗ ∗
yk,i i
Γ̂∗2
E
kl(j)
i=1
∗
∗
+ Γ̂∗kk(j) Γ̂∗ll(j) − E ∗ Γ̂∗2
kl(j) + Γ̂kk(j) Γ̂ll(j)

i

j=1


1/h 1/M h h
i
X
X
∗
∗
∗
∗
∗

= E ∗  zi∗
Γ̂∗2
Γ̂∗2
kl(j) + Γ̂kk(j) Γ̂ll(j) − E
kl(j) + Γ̂kk(j) Γ̂ll(j)
i=1
j=1
∗
∗
E ∗ Γ̂∗kl(j) − β̂lk Γ̂∗kk(j) Γ̂∗2
+
Γ̂
Γ̂
kl(j)
kk(j)
 ll(j)
= Mh
1
2
2
−
Γ̂
−
β̂
Γ̂
Γ̂
+
Γ̂
Γ̂
+
Γ̂
Γ̂
+
3
Γ̂
lk kk(j)
kl(j)
kk(j) ll(j)
kk(j) ll(j)
j=1
M
kl(j)
kl(j)

1/M h
∗
∗
∗
∗
∗2
∗2 Γ̂∗
X
E ∗ Γ̂∗3
+
Γ̂
Γ̂
Γ̂
−
β̂
Γ̂
−
β̂
Γ̂
Γ̂
lk
lk
kl(j)
kk(j)
kk(j) kl(j)
kk(j) ll(j)
 kl(j) ll(j)
= Mh
1
2
2
−
Γ̂
+
Γ̂
Γ̂
+
Γ̂
Γ̂
+
3
Γ̂
Γ̂
−
β̂
Γ̂
lk
kl(j)
kk(j)
kk(j)
ll(j)
kk(j)
ll(j)
j=1
M
kl(j)
kl(j)
1/M h


X
57


.
Using Lemma B.7 we deduce that
∗
∗
∗
∗
∗2
∗2
∗
−
β̂
−
β̂
+
Γ̂
Γ̂
Γ̂
Γ̂
Γ̂
Γ̂
Γ̂
E ∗ Γ̂∗3
lk kk(j) kl(j)
lk kk(j) ll(j)
kk(j) kl(j) ll(j)
kl(j)
∗
∗2
∗2
∗
∗
∗
∗
∗
∗
∗
Γ̂
Γ̂
Γ̂
−
β̂
E
Γ̂
−
β̂
= E ∗ Γ̂∗3
+
E
Γ̂
Γ̂
Γ̂
E
lk
lk
kk(j) ll(j)
kk(j) kl(j)
kl(j)
kk(j) kl(j) ll(j)
(M + 2) (M + 1) 3
3 (M + 2)
Γ̂kl(j) +
Γ̂kk(j) Γ̂kl(j) Γ̂ll(j)
2
M
M2
(M + 2) 3
(M + 2)2
Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) + 2
Γ̂kl(j)
+
2
M
M2
(M + 2) 2
(M + 2) (M + 3)
2
−β̂lk
Γ̂kk(j) Γ̂kl(j) +
Γ̂kk(j) Γ̂ll(j)
M2
M2
M (M + 2) 2
4 (M + 2)
2
−β̂lk
Γ̂kk(j) Γ̂ll(j) +
Γ̂kk(j) Γ̂kl(j)
M2
M2
(M + 2) (M + 5)
(M + 2) (M + 3) 3
Γ̂kl(j) +
Γ̂kk(j) Γ̂kl(j) Γ̂ll(j)
=
2
M2
M
(M + 2) (M + 7)
(M + 2) (M + 1) 2
2
−β̂lk
Γ̂kk(j) Γ̂kl(j) +
Γ̂kk(j) Γ̂ll(j) .
M2
M2
=
Adding and substracting appropriately gives result for (a4).
For (a5), note that


1/M h h
1/h
X
X
∗ ∗
∗
Γ̂∗
yk,i i
E
∗
kk(j) Γ̂kl(j)
− E ∗ Γ̂∗kk(j) Γ̂∗kl(j)
i

j=1
i=1


1/h 1/M h h
i
X
X
Γ̂∗kk(j) Γ̂∗kl(j) − E ∗ Γ̂∗kk(j) Γ̂∗kl(j) 
= E ∗  zi∗
i=1
j=1
1/M h X
2
∗
∗2
∗2
∗
∗
Γ̂kk(j) Γ̂kl(j) +
= Mh
E Γ̂kk(j) Γ̂kl(j) − β̂lk Γ̂kk(j) Γ̂kl(j) − Γ̂kl(j) − β̂lk Γ̂kk(j)
Γ̂
Γ̂
,
M kk(j) kl(j)
j=1
then use Lemma B.7 and remark that
∗2
∗
Γ̂
Γ̂
−
β̂
E ∗ Γ̂∗kk(j) Γ̂∗2
lk
kl(j)
kk(j) kl(j)
=
(M + 2) (M + 3)
(M + 2) 2
(M + 2) (M + 4) 2
Γ̂kk(j) Γ̂2kl(j) +
Γ̂kk(j) Γ̂ll(j) − β̂lk
Γ̂kk(j) Γ̂kl(j) .
M2
M2
M2
Adding and substracting appropriately gives result for (a5).
For (a6), note that

1/h
1/M h h
X
X
∗
∗ ∗
E
yk,i i
Γ̂∗2
kk(j)
i=1
− E ∗ Γ̂∗2
kk(j)

i

j=1
1/M h X
2 2
∗
∗2
∗
∗3
2
= Mh
E Γ̂kk(j) Γ̂kl(j) − β̂lk Γ̂kk(j) − Γ̂kl(j) − β̂lk Γ̂kk(j)
Γ̂kk(j) +
Γ̂
.
M kk(j)
j=1
Next use Lemma B.7 and compute
M (M + 2) (M + 4) 2
3
∗
∗3
Γ̂
Γ̂
−
β̂
Γ̂
E ∗ Γ̂∗2
Γ̂
−
β̂
Γ̂
=
lk
lk
kl(j)
kk(j) kl(j)
kk(j)
kk(j)
kk(j) .
M3
58
Adding and substracting appropriately gives result for (a6). The remaining results follows similarly
and therefore we omit the details.
Lemma B.9. Suppose (1), (2) and (10) hold. We have that,
a1)
∗
E ∗ Sβ,h,M
= 0;
a2)
2
∗
E ∗ Sβ,h,M
= 1;
a3)
3 √ Ã∗
∗
E ∗ Sβ,h,M
= h 1,h,M
∗3/2 ;
a4)
a5)
B̃h,M
P1/M h
3
Γ̂
+
6M
h
j=1
j=1 Γ̂kk(j) Γ̂kl(j) Γ̂ll(j)
kl(j)
1
∗
∗
E ∗ Sβ,h,M
U1,β,h,M
= ∗3/2
P1/M h
P1/M h
B̃h,M
−6β̂lk M h j=1 Γ̂kk(j) Γ̂2kl(j) − 2β̂lk M h j=1 Γ̂2kk(j) Γ̂ll(j)
"
#
P1/M h
P1/M h 3
Γ̂
+
10M
h
Γ̂
6M
h
Γ̂
Γ̂
kk(j)
kl(j)
ll(j)
j=1
j=1
kl(j)
1
;
+ ∗3/2
P1/M h
P1/M h
B̃h,M M
−14β̂lk M h j=1 Γ̂kk(j) Γ̂2kl(j) − 2β̂lk M h j=1 Γ̂2kk(j) Γ̂ll(j)
"
2M h
P1/M h
#
∗
∗
U2,β,h,M
E ∗ Sβ,h,M
i
h
P1/M h
P1/M h
P1/M h
2
lk
3M h j=1 Γ̂kk(j) Γ̂2kl(j) + M h j=1 Γ̂2kk(j) Γ̂ll(j) − 4β̂lk M h j=1 Γ̂2kk(j) Γ̂kl(j) ;
= −4 β̂∗3/2
1+ M
B̃h,M
a6)
∗
∗
=8
U3,β,h,M
E ∗ Sβ,h,M
a7)
∗
∗
= −4
U4,β,h,M
E ∗ Sβ,h,M
a8)
4
∗
E ∗ Sβ,h,M
= 3 + OP (h) , as h → 0;
√ i
h
∗
∗2
= OP
Ǔβ,h,M
E ∗ Sβ,h,M
h , as h → 0;
a9)
a10)
2
β̂lk
∗3/2
B̃h,M
1+
2
M
Ã∗0,h,M
∗1/2 P1/h
B̃h,M i=1
h
2
yk,i
Mh
B̃h,M
We apply Lemma
∗
∗
∗
∗
, U4,β,h,M
, U3,β,h,M
U1,β,h,M
, U2,β,h,M
and
j=1
Γ̂2kk(j) Γ̂kl(j) − β̂lk M h
P1/M h
j=1
i
Γ̂3kl(j) ;
;
h
i
Ã∗
∗3
∗
E ∗ Sβ,h,M
Ǔβ,h,M
= 3 1,h,M
∗3/2 + OP (h) ,
Proof of Lemma B.9.
P1/M h
as h → 0.
∗
,
B.8, results follow directly given the denitions of Sβ,h,M
∗
Ǔβ,h,M
.
Lemma B.10. Suppose (1), (2) and (10) hold. We have that,
∗
Uβ,h,M
=
∗
Ǔβ,h,M
+
∗
U4,β,h,M
+O
P∗
√ h ,
in probability, where
∗
Ǔβ,h,M
=
√ 1/M h
i
M h X h ∗2
∗
∗
∗
∗2
∗
∗
Γ̂
+
Γ̂
Γ̂
−
E
Γ̂
+
Γ̂
Γ̂
kl(j)
kk(j)
ll(j)
kl(j)
kk(j)
ll(j)
∗
B̃h,M
j=1
i
Xh h
√ 1/M
β̂lk
Γ̂∗kk(j) Γ̂∗kl(j) − E ∗ Γ̂∗kk(j) Γ̂∗kl(j)
−4 ∗ M h
B̃h,M
j=1
+2
2
β̂lk
∗
B̃h,M
i
Xh h
√ 1/M
∗
∗2
M h
Γ̂∗2
−
E
Γ̂
kk(j)
kk(j)
j=1
∗
∗
∗
+ U2,β,h,M
+ U3,β,h,M
,
≡ U1,β,h,M
59
and
Ã∗0,h,M
∗
∗
U4,β,h,M
= −4 q
Sβ,h,M
.
P
1/h
∗
2
B̃h,M
y
i=1 k,i
Proof of Lemma B.10.
Using the denition of
∗ ,
B̂h,M
by adding and substracting appropriately, we
can write

∗
B̂h,M


X 

= Mh

j=1 


1/M h 
∗
Since β̂lk
Γ̂kk(j) Γ̂ll(j) + 3Γ̂2kl(j)
Γ̂2kk(j)
2
∗
∗
∗
∗2
∗
−4 β̂lk − β̂lk Γ̂kk(j) Γ̂kl(j) − M Γ̂kk(j) Γ̂kl(j) + 4β̂lk β̂lk − β̂lk
Γ̂kk(j) − 2 M
2
−4β̂lk Γ̂∗kk(j) Γ̂∗kl(j) − M
Γ̂kk(j) Γ̂kl(j)
2 Γ̂2kk(j)
Γ̂2kk(j)
∗
∗2
2
∗2
+ 2 β̂lk − β̂lk
Γ̂kk(j) − 2 M
+2β̂lk Γ̂kk(j) − 2 M
∗
∗
Γ̂∗2
kl(j) + Γ̂kk(j) Γ̂ll(j) −
h1/2
− β̂lk = OP ∗
∗
β̂lk
and
Mh
2
− β̂lk
P1/M h
j=1
Γ̂2kk(j)
∗2
Γ̂kk(j) − 2 M
= OP ∗ (1) ,
1/M h
Mh
1
M
X
Γ̂∗2
kk(j)
−2
Γ̂2kk(j)
j=1
M
in probability, then
!
= OP ∗ (h) ,
in probability. Next, note that
1/M h Mh
X
Γ̂∗kk(j) Γ̂∗kl(j)
j=1
2
Γ̂
Γ̂
−
M kk(j) kl(j)
in probability, which together with
∗
β̂lk
1/M h − β̂lk · M h
X
Γ̂∗kk(j) Γ̂∗kl(j)
j=1
1/M h
X
∗
β̂lk β̂lk
− β̂lk · M h
j=1
1/M h
∗ − β̂ = O ∗
β̂lk
P
lk
= Mh
X
Γ̂kk(j) Γ̂kl(j) + OP ∗ h1/2 ,
j=1
h1/2 ,
and
β̂lk = OP ∗ (1)
implies that
1/M h
X
2
∗
−
Γ̂kk(j) Γ̂kl(j)
Γ̂kk(j) Γ̂kl(j) + OP ∗ (h)
=
β̂lk − β̂lk · M h
M
j=1
!
1/M h
X
Γ̂2kk(j)
∗2
∗
Γ̂kk(j) − 2
Γ̂2kk(j) + OP ∗ (h) .
= β̂lk β̂lk
− β̂lk · M h
M
j=1
Note also that
1 2
∗
∗
2
Γ̂
Γ̂
+
3
Γ̂
E ∗ Γ̂∗2
+
Γ̂
Γ̂
=
Γ̂
+
Γ̂
Γ̂
+
kk(j) ll(j)
kk(j) ll(j)
kl(j) ,
kl(j)
kk(j) ll(j)
kl(j)
M
2
E ∗ Γ̂∗kk(j) Γ̂∗kl(j)
Γ̂
Γ̂
,
= Γ̂kk(j) Γ̂kl(j) +
M kk(j) kl(j)
Γ̂2kk(j)
∗
∗2
2
E Γ̂kk(j)
,
= Γ̂kk(j) + 2
M
and recall that by denition
∗
B̃h,M
= Mh
1/M h X
2 2
Γ̂2kl(j) + Γ̂kk(j) Γ̂ll(j) − 4β̂lk Γ̂kk(j) Γ̂kl(j) + 2β̂lk
Γ̂kk(j) .
j=1
60





.




Hence, we obtain
∗
B̂h,M
=
∗
B̃h,M
+ Mh
1/M h h
X
i
∗
∗
∗
∗
∗
Γ̂∗2
Γ̂∗2
kl(j) + Γ̂kk(j) Γ̂ll(j)
kl(j) + Γ̂kk(j) Γ̂ll(j) − E
j=1
−4β̂lk M h
1/M h h
X
Γ̂∗kk(j) Γ̂∗kl(j)
−E
∗
Γ̂∗kk(j) Γ̂∗kl(j)
i
+
2
2β̂lk
Mh
1/M h h
X
i
∗
∗2
−
E
Γ̂
Γ̂∗2
kk(j)
kk(j)
j=1
j=1
∗
−4Ã∗0,h,M β̂lk
− β̂lk + OP ∗ (h) .
∗
Next, we can use β̂lk
∗
Ã∗0,h,M β̂lk
− β̂lk
P1/h
− β̂lk =
∗ ∗
i=1 yk,i i
P1/h ∗2
i=1 yk,i
√
∗
and the denition of Sβ,h,M
=
h−1
P1/h
√
i=1
∗ ∗
yk,i
i
∗
Bh,M
to write
q
P
−1
1/h ∗2
∗
B̃h,M
√ ∗
y
i=1 k,i
∗
=
hÃ0,h,M Sβ,h,M
P1/h 2  P1/h 2 
i=1 yk,i
i=1 yk,i
q


P1/h ∗2 P1/h 2
∗
B̃h,M
√ ∗
y
−
y
i=1 k,i
k,i
∗
1 + i=1 P
hÃ0,h,M P1/h
=
Sβ,h,M
+ OP ∗ (h)
1/h 2
2
y
y
i=1 k,i
i=1 k,i
q
q
P
P1/h 2 
1/h ∗2
∗
∗
Ã∗0,h,M B̃h,M
√
√ Ã∗0,h,M B̃h,M
−
y
i=1 yk,i 
k,i
∗
∗
 i=1 P
Sβ,h,M
+ h P1/h
Sβ,h,M
h P1/h
+ OP ∗ (h) ,
=
1/h 2
2
2
i=1 yk,i
i=1 yk,i
i=1 yk,i
|
{z
}
=OP ∗ (h)
where we have used the fact that
P.
∗
= OP ∗ (1) and
Sβ,h,M
P1/h
∗2
i=1 yk,i −
P1/h
2
i=1 yk,i =OP ∗
√ h , in probability-
It follows that
∗
Uβ,h,M
√ 1/M h
i
M h X h ∗2
∗2
∗
∗
∗
∗
∗
Γ̂
+
Γ̂
Γ̂
Γ̂
+
Γ̂
Γ̂
−
E
kl(j)
kk(j) ll(j)
kl(j)
kk(j) ll(j)
∗
B̃h,M
j=1
=
i
Xh h
√ 1/M
β̂lk
−4 ∗ M h
Γ̂∗kk(j) Γ̂∗kl(j) − E ∗ Γ̂∗kk(j) Γ̂∗kl(j)
B̃h,M
j=1
i
Xh h
√ 1/M
Ã∗0,h,M
∗2
∗
∗
q
+2 ∗ M h
Γ̂∗2
−
E
Γ̂
−
4
kk(j)
kk(j)
P1/h 2 Sβ,h,M + OP ∗ (h)
∗
B̃h,M
B̃h,M i=1 yk,i
j=1
√ ∗
∗
∗
∗
≡ U1,β,h,M
+ U2,β,h,M
+ U3,β,h,M
+ U4,β,h,M
+ OP ∗
h .
|
{z
}
2
β̂lk
∗
≡Ǔβ,h,M
Proof of proposition 5.2.
Part (a) follows under our assume conditions by using Theorem 5.2 of
Dovonon et al. (2013). The proof of part (b) use the same technique as in the proof of Theorem A.3
in Gonçalves and Meddahi (2009). Given (36) and Lemma B.10, we may decompose
√ −1/2
√ ∗
∗
∗
∗
Tβ,h,M
= Sβ,h,M
1 + h Ǔβ,h,M
+ U4,β,h,M
+ OP ∗
h
.
k, we have that
√ k
∗k
∗
∗
∗
= Sβ,h,M 1 − h Ǔβ,h,M + U4,β,h,M
+ OP ∗ (h) ≡ T̃β,h,M
+ OP ∗ (h) .
2
Hence, for any xed integer
∗k
Tβ,h,M
61
For
k = 1, 2, 3,
the moments of
∗
T̃β,h,M
are given by
√ k
√ k
∗k
∗k
∗k
∗
∗k
∗
E ∗ T̃β,h,M
= E ∗ Sβ,h,M
− h E ∗ Sβ,h,M
Ǔβ,h,M
− h E ∗ Sβ,h,M
U4,β,h,M
.
2
2
∗
T̃β,h,M
The rst and third cumulants of
are given by
∗
∗
κ∗1 T̃β,h,M
= E ∗ T̃β,h,M
h i3
∗
∗3
∗2
∗
∗
.
κ∗3 T̃β,h,M
= E ∗ T̃β,h,M
− 3E ∗ T̃β,h,M
E ∗ T̃β,h,M
+ 2 E ∗ T̃β,h,M
By Lemma B.9, we deduce
√
∗
E ∗ T̃β,h,M
= −

h 
Mh
1/M h
1/M h
X
X
∗3/2
B̃h,M
√
Γ̂3kl(j) + 3M h

Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) 
j=1
j=1


1/M h
1/M h
X
X
h 
− ∗3/2 −3β̂lk M h
Γ̂kk(j) Γ̂2kl(j) − β̂lk M h
Γ̂2kk(j) Γ̂ll(j) 
B̃h,M
j=1
j=1

√ 
1/M h
1/M h
1/M h
X
X
X
h 
2
Γ̂2kk(j) Γ̂kl(j) 
Γ̂2kk(j) Γ̂ll(j) + 8β̂lk
Mh
Γ̂kk(j) Γ̂2kl(j) − 2β̂lk M h
− ∗3/2 −6β̂lk M h
B̃h,M
j=1
j=1
j=1


√
1/M h
1/M h
X
X
√
2Ã∗0,h,M
h  2
2
3
− ∗3/2 4β̂lk M h
Γ̂kk(j) Γ̂kl(j) − 4β̂lk M h
Γ̂3kl(j)  + h q
P1/h 2
∗
Bh,M
B̃h,M
j=1
j=1
i=1 yk,i


√
1/M h
1/M h
1/M h
X
X
X
h 
− ∗3/2
Γ̂kk(j) Γ̂2kl(j) 
Γ̂kk(j) Γ̂kl(j) Γ̂ll(j) − 19β̂lk M h
Γ̂3kl(j) + 5M h
3M h
B̃h,M M
j=1
j=1
j=1


√
1/M h
1/M h
1/M h
X
X
X
h 
3
2
− ∗3/2
Γ̂3kl(j)  .
Γ̂2kk(j) Γ̂kl(j) − 8β̂lk
Mh
Γ̂2kk(j) Γ̂ll(j) + 24β̂lk
Mh
−5β̂lk M h
B̃h,M M
j=1
j=1
j=1
Result follows since, we can write
∗
=
E ∗ T̃β,h,M
where
∗
Ã∗0,h,M , Ã∗1,h,M , B̃h,M
and
√


∗
2Ã∗0,h,M
Ã∗1,h,M
R̃1,h,M
,
h q
+
P1/h 2 −
∗3/2
M
∗
2
B̃
B̃h,M i=1 yk,i
h,M
∗
R̃1,h,M
are dened in the main text.
The remaining results follow
similarly. For part (c), apply Theorem 3 of Li, Todorov and Tauchen (2016).
62
Proof of Bootstrap results in Section 3.2 and 4.2: Mykland and Zhang's (2009)
type estimator
Notation
We introduce some notation. We let
V1,(j) =
V2,(j) =
We denote by
M 2h
M −1
M
X
M 2h
M −1
M
X
!−1
M
X
2
yk,i+(j−1)M
i=1
!
u2i+(j−1)M
,
i=1
!−2
M
X
2
yk,i+(j−1)M
i=1
!2
yk,i+(j−1)M ui+(j−1)M
.
i=1
yk(j) = yk,1+(j−1)M , · · · , yk,M j
0
, the
M
returns of asset
k
observed within the block
j.
Similarly for the bootstrap, we let
∗
V1,(j)
≡
∗
V2,(j)
and
M
X
M 2h
M −1
M
X
1/M
Ph j=1
M
M −1
q
2
Ĉkk(j)
0
Ĉlk(j) Ĉll(j)
1
bM,q cM −1,q
bll(j)
C
b
Ckk(j)
M
X
!
u∗2
i+(j−1)M
,
i=1
!−2
M
X
∗2
yk,i+(j−1)M
i=1
∗
∗
∗
yk(j)
= yk,1+(j−1)M
, · · · , yk,M
j
Ĉ(j) ≡
!−1
∗2
yk,i+(j−1)M
i=1
Recall that
Mh
≡
M 2h
M −1
!2
∗
yk,i+(j−1)M
u∗i+(j−1)M
,
i=1
0
.
q
!

=
Γ̂kk(j)
Γ̂
q kl(j)
Γ̂kk(j)

0
r
Γ̂ll(j) −
Γ̂2kl(j)

.
For any
q > M,
let
Rβ,q ≡
Γ̂kk(j)
q
, where the denition of
q cM,q
is given in equation (7), and for
2q Γ( M2 − 2q )
2
2
, where χM is the standard χ distribution
Γ( M
)
2
M
M2
M3
with M degrees of freedom. Note that bM,2 =
, bM,4 =
M −2
(M −2)(M −4) , and bM,6 = (M −2)(M −4)(M −6) .
It follows by using the denition of bM,q and this property of the Gamma function, for all x > 0,
Γ (x + 1) = xΓ (x).
any
q > M,
we have
bM,q ≡ E
M
χ2M
2
M
2
=
Auxiliary Lemmas
Lemma B.11. Suppose (1) and (2) hold. Then, we have that
1/M h
V̂β̌,h,M =
X
1/M h
V1,(j) −
j=1
X
V2,(j) .
j=1
Lemma B.12. Suppose (1) and (2) hold with W independent of σ. Assume that
conditionally on σ and under Qh,M , the following hold
a1)
E V1,(j) =
a2)
2
E V1,(j)
=
M 3h
(M −1)(M −2)
Cll(j)
Ckk(j)
2
M 5 (M +2)
h2
(M −1)2 (M −2)(M −4)
, for M > 2;
Cll(j)
Ckk(j)
4
, for M > 4;
63
M = O (1) ,
then
2
a3)
E V2,(j) =
a4)
2
E V2,(j)
=
a5)
E V1,(j) V2,(j) =
a6)
V ar V1,(j) =
4M 5
h2
(M −1)(M −2)2 (M −4)
Cll(j)
Ckk(j)
4
, for M > 4;
a7)
V ar V2,(j) =
2M 4
h2
(M −1)(M −2)2 (M −4)
Cll(j)
Ckk(j)
4
, for M > 4;
a8)
Cov V1,(j) , V2,(j) =
a9)
V ar V1,(j) − V2,(j) =
M 2h
(M −1)(M −2)
Cll(j)
Ckk(j)
3M 4
h2
(M −1) (M −2)(M −4)
, for M > 2;
2
Cll(j)
Ckk(j)
M 4 (M +2)
h2
(M −1)2 (M −2)(M −4)
4
, for M > 4;
Cll(j)
Ckk(j)
4M 4
h2
(M −1)(M −2)2 (M −4)
4
, for M > 4;
Cll(j)
Ckk(j)
2M 5 (2M −3)
h2
(M −1)(M −2)2 (M −4)
4
Cll(j)
Ckk(j)
, for M > 4;
4
, for M > 4.
Lemma B.13. Suppose (1) and (2) hold with W independent of σ. Assume that
conditionally on σ and under Qh,M , for any M > 4, the following hold
a1)
E V̂β̌,h,M = Vβ̌,h,M ;
a2)
V ar V̂β̌,h,M =
a3)
V̂β̌,h,M − Vβ̌,h,M → 0
a4)
Vβ̌,h,M → Vβ̌ .
2M 4 (2M −3)
h
(M −1)(M −2)2 (M −4)
· Mh
1/M
Ph j=1
Cll(j)
Ckk(j)
4
(Γ̂kk(j) )−1
2
i=1 ûi+(j−1)M =
q
E
a3)
Rβ,q − M h
M −1
M
1/M
Ph j=1
δ > 0;
a4)
=
2q
Cll(j)
Ckk(j)
V̂β̌,h,M − Vβ̌,h,M → 0
M = O (1) ,
then
in probability;
PM
a2)
bll(j)
C
b
Ckk(j)
then
;
Lemma B.14. Suppose (1) and (2) hold with W independent of σ. Assume that
conditionally on σ and under Qh,M , the following hold
a1)
M = O (1) ,
Cll(j)
Ckk(j)
bM,q cM −1,q
q
→0
2
;
Cll(j)
Ckk(j)
q
, for M > q;
in probability under Qh,M and P , for any M > q (1+δ) , for some
in probability under Qh,M and P , for any M > 2 (1+δ) , for some δ > 0.
Proof of Lemma B.11.
in the text (see Equation (29)), and the
denition of
can write
Given the denition of V̂β,h,M
ûi+(j−1)M = yl,i+(j−1)M − β̌lk(j) yk,i+(j−1)M , we
1/M h
V̂β̌,h,M
2
= M h
X
M
X
j=1
i=1
1/M h
=
M 2h X
M −1
j=1
!−1
2
yk,i+(j−1)M
M
X
i=1
M
2
1 X
yl,i+(j−1)M − β̌lk(j) yk,i+(j−1)M
M −1
!
i=1
!−1
2
yk,i+(j−1)M
M
X
i=1
64
ui+(j−1)M − β̌lk(j) − βlk(j) yk,i+(j−1)M
2
!
,
where we used the denition of
yl,i+(j−1)M
see equation (29). Adding and subtracting appropriately,
it follows that
V̂β̌,h,M
=
1/M h
M
M 2h X X 2
yk,i+(j−1)M
M −1
j=1
!−1
i=1
M
X
!
u2i+(j−1)M
+ β̌lk(j) − βlk(j)
2
!
i=1
!−1 M
!
1/M h
M
X
X
M 2h X
2
β̌lk(j) − βlk(j)
yk,i+(j−1)M
yk,i+(j−1)M ui+(j−1)M
−2
M −1
j=1
i=1
i=1
!−1 M
!
1/M h
M
X
M 2h X X 2
yk,i+(j−1)M
u2i+(j−1)M
=
M −1
j=1
i=1
i=1
!
!2
−2
1/M
h
M
M
X
M 2h X X 2
−
yk,i+(j−1)M
yk,i+(j−1)M ui+(j−1)M
M −1
j=1
1/M h
=
X
j=1
i=1
i=1
1/M h
V1,(j) −
X
V2,(j) ,
j=1
−1 P
PM 2
M
β̌lk(j) − βlk(j) =
y
y
u
i=1 k,i+(j−1)M
i=1 k,i+(j−1)M i+(j−1)M .
Proof of Lemma B.12 part a1). Given the denition of V1,(j) , the law of iterated expectations
the fact that ui+(j−1)M |yk(j) ∼ i.i.d.N 0, V(j) , we can write
E V1,(j) = E E V1,(j) |yk(j)
 

!−1 M
!
M
2
X
X
M h  
2
E E
yk,i+(j−1)M
u2i+(j−1)M  |yk(j) 
=
M −1
i=1
i=1

!
!
−1
M
M
2
X
X
M h 
2
E
yk,i+(j−1)M
E u2i+(j−1)M |yk(j) 
=
M −1
i=1
i=1


!
−1
M
3
X
M h
2
,
=
V E
yk,i+(j−1)M
M − 1 (j)
where we used
and
i=1
then given equation (28) in the text and by replacing
E V1,(j) =
Proof of Lemma B.12 part a2).
V(j)
by
M 3h
(M − 1) (M − 2)
Given the denition of
2 ,
hCll(j)
Cll(j)
Ckk(j)
V1,(j)
we deduce
2
.
and the law of iterated expectations,
we can write
2
2
E V1,(j)
= E E V1,(j)
|yk(j)
 

!−2 M
!2 
M
X
X
M 4 h2
2
E E 
yk,i+(j−1)M
u2i+(j−1)M  |yk(j) 
=
(M − 1)2
i=1
i=1


2 
!−2
!2
M
M
X
X
u
M 4h
i+(j−1)M
2
p
=
yk,i+(j−1)M
E
|yk(j)   .
V2 E
(M − 1)2 (j)
V(j)
i=1
i=1
65
Note that since
M (M + 2)
ui+(j−1)M |yk(j) ∼ i.i.d.N 0, V(j)
where
χ2j,M
follow the standard
2
E V1,(j)
=
then given the fact that
χ2
,
PM
E
√
i=1

+ 2) h 2 
V(j) E
(M − 1)2
M
X
M
!2
2
|yk(j)
V(j)
2
= E χ2j,M
=
degrees of freedom. Then we have
!−2 
2
yk,i+(j−1)M
,
i=1
d
2
i=1 yk,i+(j−1)M =
by using the second moment of an inverse
ui+(j−1)M
distribution with
M 5 (M
PM
d
2
hCkk(j)
χ2j,M , where `=' denotes equivalence in distribution,
2
1
1
2
= (M −2)(M
of χ distribution, we have E
−4) , and
χ2
j,M
by replacing
V(j)
by
2
hCll(j)
E
it follows that
2
V1,(j)
M 5 (M + 2)
=
h2
(M − 1)2 (M − 2) (M − 4)
Proof of Lemma B.12 part a3).
Given the denition of
V1,(j) ,
Cll(j)
Ckk(j)
4
.
the law of iterated expectations and
ui+(j−1)M |yk(j) ∼ i.i.d.N 0, V(j) , we can write
E V2,(j) = E E V2,(j) |yk(j)
 

!−2 M
!2 
M
2
X
X
M h  
2
E E
=
yk,i+(j−1)M
yk,i+(j−1)M ui+(j−1)M  |yk(j) 
M −1
i=1
i=1

!
!
−2
M
M
2
X
X
M h 
2
2
E
=
yk,i+(j−1)M
yk,i+(j−1)M
E u2i+(j−1)M |yk(j) 
M −1
i=1
i=1


!
−1
M
2
X
M h
2
,
V E
yk,i+(j−1)M
=
M − 1 (j)
the fact that
i=1
then using equation (28) in the text and replacing
E V2,(j) =
Proof of Lemma B.12 part a4).
V(j)
by
2
hCll(j)
M 2h
(M − 1) (M − 2)
Given the denition of
yields
Cll(j)
Ckk(j)
V2,(j)
2
.
and the law of iterated expectations,
we can write
2
2
= E E V2,(j)
|yk(j)
E V2,(j)


!−4
!4
M
M
4
2
X
X
M h
2
E
yk,i+(j−1)M
E
yk,i+(j−1)M ui+(j−1)M
|yk(j) 
=
(M − 1)2
i=1
i=1


!
−4
M
X
M 4 h2
2

≡
E
yk,i+(j−1)M
Ξ .
(M − 1)2
i=1
66
Then using the conditional independence and mean zero property of

Ξ ≡ E
M
X
!4
yk,i+(j−1)M ui+(j−1)M
we have that

|yk(j) 
yk,i+(j−1)M ui+(j−1)M
i=1
=
M
X
4
u4i+(j−1)M |yk(j)
E yk,i+(j−1)M
i=1
+3
X
2
2
E yk,i+(j−1)M
u2i+(j−1)M |yk(j) E yk,s+(j−1)M
u2s+(j−1)M |yk(j)
i6=s
2
= 3V(j)


M
X

y4
k,i+(j−1)M
i=1
+
X
2
2
 = 3V 2
yk,s+(j−1)M
yk,i+(j−1)M
(j)
M
X
!2
4
yk,i+(j−1)M
,
i=1
i6=s
thus we can write
2
E V2,(j)

!−2 
M
X
M 4 h2
2
,
=
3V 2 E 
yk,i+(j−1)M
(M − 1)2 (j)
i=1
result follows similarly where we use the same arguments as in the proof of Lemma B.12 part a2).
Proof of Lemma B.12 part a5).
The proof follows similarly as parts a2) and a4) of Lemma B.12
and therefore we omit the details.
Proof of Lemma B.13 part a1).
Given the denitions of
V̂β,h,M , V1,(j) , V2,(j)
and by using Lemma
B.11 and part 1 of Lemma B.12, we can write


1/M h
E V̂β̌,h,M
= E
X

V1,(j)  − E 
j=1
1/M h
=
X

1/M h
X
V2,(j) 
j=1
1/M h
X
E V1,(j) −
E V1,(j)
j=1
j=1
1/M h 2
1/M h X Cll(j) 2
M 2h
−
(M − 1) (M − 2)
Ckk(j)
=
X
M 3h
(M − 1) (M − 2)
=
1
M
Vβ̌,h,M −
V
= Vβ̌,h,M .
M −1
M − 1 β̌,h,M
j=1
Proof of Lemma B.13 part a2).
Cll(j)
Ckk(j)
j=1
Given the denitions of
V̂β̌,h,M , V1,(j) , V2,(j)
and Lemma B.11, we
can write

V ar V̂β̌,h,M = V ar 
1/M h
X
j=1

V1,(j)  + V ar 
1/M h
X
j=1


V2,(j)  − 2Cov 
1/M h
1/M h
X
X
j=1
V1,(j) ,

V2,(j)  ,
j=1
ui+(j−1)M |yk(j) ∼ i.i.d.N 0, V(j) , we have V1,(j) and V2,(j) are conditionally indeyk(j) , V1,(j) and V2,(t) are conditionally independent for all t 6= j given yk(j) . It follows
given the fact that
pendent given

67
that
V ar V̂β̌,h,M
1/M h X
=
2 2 2 2
E V1,(j)
− E V1,(j)
+ E V2,(j) − E V2,(j)
j=1
1/M h
−2
X
E V2,(j) V2,(j) − E V1,(j) E V2,(j) ,
j=1
nally results follow given Lemma B.12.
Proof
of Lemma B.13 part a3). Results follow directly
given Lemma B.12 parts a1) and a2) since
V ar V̂β̌,h,M − Vβ̌,h,M → 0 as h → 0.
Proof of Lemma B.13 part a4). This result follows from the boundedness of Σkk,s , Σll,s
Reimann integrable of Σkl,s for any k, l = 1, · · · , d.
Proof of Lemma B.14 part a1). Given the denition of ûi+(j−1)M , we can write
E V̂β̌,h,M − Vβ̌,h,M = 0
(Γ̂kk(j) )−1
M
X
û2i+(j−1)M
and
M
X
1
=
Γ̂kk(j)
i=1
Γ̂kk(j)
1
=
2
yl,i+(j−1)M
− 2β̌lk(j)
i=1
β̌lk(j) =
Proof of Lemma B.14 part a2).
2
2
2
yl,i+(j−1)M
− 2β̌lk(j) yl,i+(j−1)M yk,i+(j−1)M + β̌lk(j)
yk,i+(j−1)M
i=1
M
X
Γ̂kk(j)
thus results follow by replacing
2
i=1
M X
1
=
yl,i+(j−1)M − β̌lk(j) yk,i+(j−1)M
and the
M
X
2
yl,i+(j−1)M yk,i+(j−1)M + β̌lk(j)
i=1
Γ̂lk(j)
Γ̂kk(j)
M
X
!
2
yk,i+(j−1)M
i=1
.
Given the denitions of
Γ̂ll(j) , Γ̂kl(j)
and
Γ̂lk(j)
and using part a1)
of Lemma B.14, we can write
E
Cll(j)
Ckk(j)
PM
2
i=1 ûi+(j−1)M
q
= E
! 2q
Γ̂kk(j)

= E
M
X
!− 2q
2
yk,i+(j−1)M
M
X
E
i=1

= E
M
X
! 2q
û2i+(j−1)M
2
yk,i+(j−1)M

V(j) E 
E
cM,q ,
V(j)
! 2q

|yk(j)  = E
χ2j,M
q
q 2
= (M − 1) 2 cM −1,q ,
68
! 2q

|yk(j)  ,
ui+(j−1)M |yk(j) ∼ i.i.d.N 0, V(j)
we can write
M û2
X
i+(j−1)M
i=1
V(j)
i=1
where we use the law of iterated expectations and the fact that

|yk(j) 
M û2
X
i+(j−1)M
q
2
i=1
Then given the denition of

i=1
!− 2q
.
,
it follows that
E
bll(j)
C
bkk(j)
C
!q
PM
2
i=1 ûi+(j−1)M
= E
! 2q
Γ̂k(j)

M
X
q
2
q
= (M − 1) 2 cM −1,q V(j) E 
!− 2q 
2
yk,i+(j−1)M

i=1
!q 
2
M


= (M − 1) cM −1,q V(j) Γk(j) E
χ2j,M

q
2
q
2
=
where
bM,q = E
M
χ2j,M
M −1
M
− 2q
q
2
bM,q cM −1,q
q
;
q !
2
, for
M > q.
Proof of Lemma B.14 part a3).
We verify the moments conditions of the Weak Law of Large
q
Numbers for independent and nonidentically distributed on
1, . . . , M1h .
Cll(j)
Ckk(j)
By using part a2) of Lemma B.14, for any
1+δ
E |zj |
=
M −1
M
δq
2
δ > 0,
zj ≡
M2
q
(M −1) 2 bM,q cM −1,q
and conditionally on
bM,(1+δ)q cM −1,(1+δ)q
bM,q
cM −1,q
Cll(j)
Ckk(j)
σ,
bll(j)
C
b
Ckk(j)
q
,
j =
we can write
(1+δ)q
<∞
Σ is an adapted càdlàg spot covolatility matrix and locally bounded and invertible (in particular,
2
Ckk(j)
> 0).
since
Proof of Lemma B.14 part a4).
Result follows directly given the denition of
part a3) of Lemma B.14, where we let
q = 2.
V̂β̌,h,M , Vβ̌,h,M
and
∗ , V
∗
Remark 4. The bootstrap analogue of Lemma B.11 and B.12 replace V1(j) with V1(j)
2(j) with V2(j) .
ll(j)
∗
∗
The bootstrap analogue of Lemma B.13 replaces V̂β̌,h,M with V̂β̌,h,M
, Vβ̌,h,M with Vβ̌,h,M
and CCkk(j)
with
bll(j)
C
b
Ckk(j)
.
Lemma B.15. Suppose (1) and (2) hold with W independent of σ. Assume that
conditionally on σ and under Qh,M , for some small δ > 0, the following hold
−2(2+δ)
a1)
E∗
P
a2)
E∗
2(2+δ) PM ∗
2+δ
∗
≤ µ22(2+δ) M 2+δ Γ̂2+δ
i=1 yk,i+(j−1)M ui+(j−1)M k(j) V̂(j) ;
M
2∗
i=1 yk,i+(j−1)M
Proof of Lemma B.15 part a1).
−2(2+δ)
= bM,4(2+δ) Γ̂k(j)
M = O (1) ,
then
, for M > 4 (2 + δ);
Given the denition of
∗
yk,i+(j−1)M
, we can write
d
2∗
i=1 yk,i+(j−1)M =
PM
PM 2
2
2
b2
b2
hC
i=1 vi+(j−1)M = hCkk(j) χj,M , where vi+(j−1)M ∼ i.i.d.N (0, 1), and χj,M follow the standard
kk(j)
χ2 distribution with M degrees of freedom. Then for any integer M > 4 (2 + δ), we have that,
!−2(2+δ)
!2(2+δ)
M
X
M
−2(2+δ)
−2(2+δ)
2
E
yk,i+(j−1)M
=E
Γ̂kk(j) = bM,4(2+δ) Γ̂kk(j) .
2
χ
j,M
i=1
69
Proof of Lemma B.15 part a2).
and the fact that
Cr
Indeed by using the
∗
u∗i+(j−1)M |yk(j)
∼ i.i.d.N 0, V̂(j)
inequality, the law of iterated expectations
, we can write for any
δ > 0,

2(2+δ) 
M
M
2(2+δ)
X
X
∗
∗
 ≤ M 3+2δ
u∗i+(j−1)M E ∗ 
u∗i+(j−1)M yk,i+(j−1)M
E ∗ yk,i+(j−1)M
i=1
i=1
= M
=
where the last equality follows since
and
2(2+δ)
µ2(2+δ) = E |v|
3+2δ
M
X
∗2(2+δ)
∗2(2+δ)
∗
E ∗ yk,i+(j−1)M E ∗ ui+(j−1)M |yk(j)
i=1
2+δ
2
µ2(2+δ) M 2+δ Γ̂2+δ
kk(j) V̂(j) ,
d
2∗
b2 v2
yk,i+(j−1)M
= hC
kk(j) i+(j−1)M ,
where
vi+(j−1)M ∼ i.i.d.N (0, 1)
.
Proof of Theorem 4.2
For part a), the proof follows the same steps as the proof of
Vβ̌,h,M
which
∗
we explain in the main text, in particular, given the denition of β̂lk , we have that
∗
Vβ̌,h,M
= V ar∗
√
∗
− β̌lk )
h−1 (β̌lk
1/M h
X
2
= M h
∗
V ar∗ β̌lk(j)
− β̌lk(j)
j=1

1/M h
X
= M 2h
M
X
E∗ 
j=1
=
M 2h
M −2
!−1 
2∗
yk,i+(j−1)M
 V̂(j)
i=1
1/M h X
j=1
Cll(j)
Ckk(j)
2
=
M −1
,
V̂
M − 2 β̌,h,M
.
then results follows, given Lemma B.13 or part a4) of Lemma B.14
For part b), we have
zj,∗ β̌ =
r
q
M −2
M −1
√
∗ − β̌
h−1 β̌lk
lk =
1/M
Ph
j=1
M
X
M −2 √
∗2
M h
yk,i+(j−1)M
M −1
zj,∗ β̌ ,
!−1
i=1
Note that
E ∗ zj,∗ β̌ = 0,
where
M
X
!
∗
yk,i+(j−1)M
u∗i+(j−1)M
.
i=1
and that

1/M h

X
zj∗  =
V ar∗ 
j=1
by part a) moreover, since
∗ , . . . , z∗
z1,
β̌
1/M h,β̌
M −2 ∗
P
V
= V̂β̌,h,M → Vβ̌ ,
M − 1 β̌,h,M
are conditionally independent, by the Berry-Esseen bound,
δ > 0 and for some constant C > 0,
!
r
M − 2 √ −1 ∗
∗
sup P
h (β̌lk − β̌lk ) ≤ x − Φ
M −1
x∈< for some small
70
x
Vβ̌
!
1/M h
2+δ
X
E ∗ zj,∗ β̌ ,
≤C
j=1
Next, we show that
1/M
Ph
j=1
1/M h
X
j=1
2+δ
E ∗ zj,∗ β̌ = op (1).
We have that

2+δ 
!−(2+δ) M
2+δ 1/M h
M
2+δ
X
X
X
√
2
2+δ
M −2

∗
∗2
∗
=
u
M h
E∗ 
yk,i+(j−1)M
y
E ∗ zj,∗ β̌ k,i+(j−1)M i+(j−1)M M −1
j=1
≡
i=1
i=1
2+δ
1/M h
M − 2 2 √ 2+δ X ∗ ∗ ∗ M h
E Aj B j ,
M −1
j=1
it follows then by using Cauchy-Schwarz inequality that
E∗
v
u
u
A∗j Bj∗ ≤ tE ∗
M
X
i=1
v 
2(2+δ) 
M
!−2(2+δ) u
u
X
u ∗ 
∗2
∗

tE
yk,i+(j−1)M
u∗i+(j−1)M yk,i+(j−1)M
i=1
1
δ
− 2+δ
2
2+δ
2
2
≤ µ2(2+δ) bM,4(2+δ)
M 1+ 2 Γ̂k(j) V̂(j)

!2  2+δ
2
1
Γ̂
Γ̂
l(j)
lk(j)
2

= µ2(2+δ) bM,4(2+δ) 
,
−
Γ̂k(j)
Γ̂k(j)
where the second inequatily used part a1) and a2) of Lemma B.15 and
v ∼ N (0, 1).
Finally, given the denition of
1/M h
X
2+δ
E ∗ zj,∗ β̌ ≤
j=1
=
M −2
M −1
2+δ
M −2
M −1
2+δ
= OP
δ
2
Rβ,2+δ ,
1/M h
µ2(2+δ) bM,4(2+δ) M
2+δ 1+ 2δ
h
X
j=1
1
2
2
µ2(2+δ) bM,4(2+δ)
1
2
h bM,4(2+δ)
M −2
M
with
we can write
1
2
2
µ2(2+δ) = E |v|2(2+δ)
M −1
M
2+δ
2
2+δ
2
bll(j)
C
bkk(j)
C
!2+δ
δ
bM,2+δ cM −1,2+δ M 1+δ h 2 Rβ,2+δ
!
bM,2+δ cM −1,2+δ
= oP (1) .
δ > 0, µ2(2+δ) = E |v|2(2+δ) ≤ ∆ < ∞ where v ∼ N (0, 1), moreover as h → 0, cM −1,2+δ =
O (1), bM,4(2+δ) = O (1), bM,2+δ = O (1) and by using Lemma B.14 we have Rβ,2+δ = OP (1).
Since for any
Proof of Theorem 4.3 Let
q
∗
Hβ̌,h,M
=
M −2
M −1
√
∗ − β̌ )
h−1 (β̌lk
lk
q
=
V̂β̌,h,M
√
∗ − β̌ )
h−1 (β̌lk
lk
q
,
∗
Vβ̌,h,M
and note that
∗
Tβ̌,h,M
where
∗
V̂β̌,h,M
is dened in the main text. Theorem 4.2 proved that
Bias∗
P∗
d∗
∗
Hβ̌,h,M
→ N (0, 1)
in probability.
∗
V̂ ∗
− Vβ̌,h,M
→ 0 in probability under Qh,M and P . In particular, we
β̌,h,M
P
∗
∗
V̂β̌,h,M
= 0, and (2) V ar∗ V̂β̌,h,M
→ 0. Results follows directly by using the
Thus, it suces to show that
show that (1)
v
uV ∗
u β̌,h,M
∗
= Hβ̌,h,M t ∗
,
V̂β̌,h,M
71
bootstrap analogue of parts a1), a2) and a3) of Lemma B.13.
References
[1] Barndor-Nielsen, O., S. E. Graversen, J. Jacod, M. Podolskij, and N. Shephard, (2006). A central limit
theorem for realised power and bipower variations of continuous semimartingales. In Y. Kabanov, R.
Lipster, and J. Stoyanov (Eds.), From Stochastic Analysis to Mathematical Finance, Festschrift for Albert
Shiryaev, 33 - 68. Springer.
[2] Barndor-Nielsen, O. and N. Shephard, (2002). Econometric analysis of realized volatility and its use in
estimating stochastic volatility models, Journal of the Royal Statistical Society, Series B, 64, 253-280.
[3] Barndor-Nielsen, O. and N. Shephard, (2004a). Econometric analysis of realised covariation: high frequency based covariance, regression and correlation in nancial economics, Econometrica, 72, 885-925.
[4] Barndor-Nielsen, O. and N. Shephard, (2004b). Power and bipower variation with stochastic volatility
and jumps,
Journal of Financial Econometrics, 2, 1-48.
[5] Barndor-Nielsen, O., P. Hansen, A. Lunde, and N. Shephard, (2008). Designing realised kernels to
measure the ex-post variation of equity prices in the presence of noise, Econometrica, 76, 1481-1536.
[6] Barndor-Nielsen, O.E., Hansen, P.R., Lunde, A., Shephard, N., (2009). Realised kernels in practice:
trades and quotes, Econometrics Journal 12, C1-C32.
[7] Davidson, R., and E. Flachaire, (2001). The Wild Bootstrap, Tamed at Last, Working Paper Darp58,
STICERD, London School of Economics.
[8] Dovonon, P., Gonçalves, S. and N. Meddahi, (2013). Bootstrapping realized multivariate volatility measures, Journal of Econometrics 172,49-65.
[9] Dovonon, P., Gonçalves, S., Hounyo, U. and N. Meddahi (2014). Bootstrapping high-frequency jump
tests, manuscript.
[10] Gonçalves, S., Hounyo, U. and N. Meddahi (2014). Bootstrap inference for pre-averaged realized volatility
based on non-overlapping returns, Journal of Financial Econometrics 12 (4), 679-707.
[11] Gonçalves, S. and N. Meddahi, (2009). Bootstrapping realized volatility, Econometrica, 77(1), 283-306.
[12] Hall, P., Horowitz, J. L. and Jing, B.-Y. (1995). On blocking rules for the bootstrap with dependent data,
Biometrika 82, 561-574.
[13] Hansen, P.R. and A. Lunde, (2006). Realized variance and market microstructure noise, Journal of Business and Economics Statistics, 24, 127-161.
[14] Heston, S. (1993). Closed-form solution for options with stochastic volatility with applications to bonds
and currency options, Review of Financial Studies, 6, 327-343.
[15] Hounyo, U. , Gonçalves, S., and N. Meddahi (2013). Bootstrapping pre-averaged realized volatility under
market microstructure noise, Research paper 2013-28, CREATES, Aarhus University.
[16] Hounyo, U., and B. Veliyev, (2016). Validity of Edgeworth expansions for realized volatility estimators,
forthcoming in the Econometrics Journal.
72
[17] Jacod, J. and Protter, P. (2012). Discretization of Processes, Springer-Verlag.
[18] Jacod, J., and M. Rosenbaum, (2013). Quarticity and other functionals of volatility ecient estimation,
The Annals of Statistics, 41, 1462-1484.
[19] Li, J., V. Todorov and G. Tauchen, (2016). Adaptive estimation of continuous-time regression models
using high-frequency data, Working paper, Duke University
[20] Politis, D.N., Romano, J.P., Wolf, M., (1999). Subsampling, Springer-Verlag, New York.
[21] Mammen, E., (1993). Bootstrap and wild bootstrap for high dimensional linear models, Annals of Statistics 21, 255-285.
[22] Meddahi, N., (2002). A theoretical comparison between integrated and realized volatility, Journal of
Applied Econometrics 17, 475-508.
[23] Mykland, P.A. and L. Zhang, (2009). Inference for continous semimartingales observed at high frequency,
Econometrica 77, 1403-1455.
[24] Mykland, P.A. and L. Zhang, (2011). The double Gaussian approximation for high frequency data,
Scandinavian Journal of Statistics 38, 215-236.
[25] Mykland, P.A., and L. Zhang, (2012). The Econometrics of High Frequency Data, Statistical Methods
for Stochastic Dierential Equations, M. Kessler, A. Lindner, and M. Sørensen, eds. p. 109-190.
[26] Mykland, P.A. N. Shephard, and K. Shepphard, (2012). Ecient and feasible inference for the components
of nancial variation using blocked multipower variation, Working paper, Oxford University.
[27] Zhang, L, P.A. Mykland, and Y. Aït-Sahalia, (2005). A tale of two time-scales: determining integrated
volatility with noisy high frequency data, Journal of the American Statistical Association, 100, 1394-1411.
73