Adaptive state space models with applications to

Adaptive state space models with
applications to the business cycle and
financial stress∗
Davide Delle Monache†
Ivan Petrella‡
Fabrizio Venditti§
Banca d’Italia
Bank of England
Banca d’Italia
February 5, 2016
Abstract
The estimation of state-space models with time-varying parameters typically implies
the use of computationally intensive methods. Moreover, when volatility evolves stochastically the model ceases to be conditionally Gaussian and requires nonlinear filtering techniques. In this paper, we model parameters’ variation in a Gaussian state-space model by
letting their dynamics to be driven by the score of the predictive likelihood. In this setup,
conditionally on past data, the model remains Gaussian and the likelihood function can
be evaluated using the Kalman filter. We derive a new set of recursions running in parallel with the standard Kalman filter recursions that allows us estimate simultaneously
the unobserved state vector and the time-varying parameters by maximum likelihood.
Given that a variety of time series models have a state space representation, the proposed
methodology is of wide interest in econometrics and applied macroeconomics. Specifically, we use it for improving GDP measurement based on alternative noisy measures and
to construct an index of financial conditions that can be used to nowcast GDP in real
time.
JEL codes: C22, C32, C51, C53, E31.
Keywords: time-varying parameters, score driven models, state space models, dynamics factor models, financial indicator.
∗
The views expressed in this paper are those of the authors and do not necessarily reflect those of the
Bank of England or the Banca d’Italia. While assuming the scientific responsibility for any error in the paper,
the authors would like to thank. We thank the participants at the workshop on “Dynamic Models driven by
the Score of Predictive Likelihoods-Tenerife 2014”, the “9th CFE Conference-London 2015”, and the seminar
participants at the University of Glasgow.
†
Monetay Policy and Economic Outlook Directorate. Email: [email protected]
‡
Bank of England, Birkbeck University London, and CEPR. Email: [email protected]
§
Financial Stability Directorate. Email: [email protected]
1
1
Introduction
After two decades of Great Moderation, a period of low macroeconomic volatility and stable
correlations across macroeconomic time series, the Great Recession has brought the issue of
structural breaks back in the spotlight, spurring a wealth of new research on modeling and
forecasting economic time series in the presence of parameter instability. Recent work in the
field builds on two well established research agendas. The former maintains that parameter
instability manifests itself through large, infrequent breaks, and it is typically associated with
institutional changes (shifts in the monetary policy strategy or changes in the wage bargaining
process). A second view is based on the idea that the collection of information is costly and
sticky, and that the learning process through which economic agents adapt to shocks results
in a slow and continuous adjustment of their behaviour, which is reflected in a gradual change
of the parameters of empirical models.1 Born and bred in different contexts, the former being
more oriented to solving forecast failure, the latter more often used for structural modeling, the
two modeling approaches have recently crossed their paths. Alessandri and Mumtaz (2014),
for example, use threshold VARs for analyzing the response of the macroeconomy to a financial
shock, while D’Agostino, Gambetti and Giannone (2013) document the forecasting properties
of models with gradual time variation in coefficients.
However, the quest for flexibility that underscores these econometric tools, does not come
without costs, since the non-linearities introduced in the models often require the use of computationally intensive estimation methods and sophisticated filtering techniques. These complications can be particularly burdensome for unobserved component models, such as the dynamic
factor models, where the hurdle implied by time varying coefficients is compounded by the
presence of latent states, which need to be inferred on the basis of the observed data.
Motivated by these challenges, in this paper we develop a theoretical framework for the
estimation of state space models with time-varying parameters and since a variety of univariate
and multivariate time-series models can be cast in state space, the proposed approach is of
interest for a wide spectrum of empirical applications. The estimator, which retains most of
the desirable properties of its linear counterpart, generalizes existing algorithms in the adaptive
models literature, like the one proposed by Koop and Korobilis (2013), justifying the definition
of Adaptive State Space models adopted in the title of the paper. The building block of the
method consists in positing a law of motion for the parameters that is a (linear) function of
the score of the predictive likelihood, in line with Creal at al. (2013) and Harvey (2013). In
this setup, conditionally on past data, we show that the model remains Gaussian and that
the likelihood function can be evaluated using the Kalman filter (KF). Still, the nonlinear
setup poses an important challenge. In the linear case the accrual of new information (i.e.
a data release) allows to compute a prediction error that, through the KF recursions, is the
basis for updating and forecasting the unobserved components. In practice, conditional on the
1
Unfortunately the question of which view of the world is more appropriate can not really be settled with an
econometric test, as tests designed for detecting large breaks typically perform poorly when the data generating
process features slow, continuous changes in the model parameters, see Benati (2007).
2
model parameters, new information allows us to refine our view on the most likely location
of the unobserved states and to form the best possible guess of where the data will be in the
future. Time variation in the parameters introduces an additional margin of adjustment to this
process. Since past forecast errors could be due to both a poor estimate of the unobserved states
and to an estimation error of the model parameters, new information calls for a simultaneous
update of both these ingredients. A crucial contribution of our paper consists in deriving the
analytical expressions for a new set of recursions that, running in parallel with the standard
KF recursions, allow to update at each point in time both the unobserved states as well as the
time varying parameters. Therefore, by means of the proposed augmented KF, the unobserved
state vector and the time-varying parameters can be estimated simultaneously by maximum
likelihood.
After presenting the main theorems we discuss two extensions of the theoretical framework
that could be particularly valuable in applied work. First, since in some applications the
researcher would like to impose parameters restrictions, such as stationarity and non-negative
variances, we show how to incorporate general restrictions on the model parameters. Second,
we illustrate how the model can be extended to deal with data at mixed frequencies and
missing observations. We conclude the theoretical part of the paper by briefly discussing how
our estimation strategy is sufficiently general to nest existing adaptive algorithms used in the
literature as a special case.
Given that a wide range of time series models have a state space representation, the estimation method we propose is of interest for a large number of applications in econometrics
and applied macroeconmics. Specifically, we illustrate the potential of the method through two
empirical applications.
In the first exercise we focus on business cycle measurement. In particular, we take as a
starting point the model that Aruoba et al. (2016) have developed to estimate GDP on the
basis of underlying (noisy) measures and extend it to account for time variation in the model
parameters. The measure of GDP developed by Aruoba et al. (2016), labelled GDPplus, is
produced by combining expenditure- and income-side measures of GDP (GDPE and GDPI
respectively) through optimal filtering methods in a constant parameter framework. We show
that most of the parameters of the state space model used by Aruoba et al. (2016) are subject
to frequent breaks, which are well captured by our score driven model. Our approach can
then retrieve some well known stylized facts of the U.S. business cycle that are overlooked
in a constant parameter framework. From our estimates, for instance, four distinct phases
of U.S. long-run growth (measured by the long-run forecast of the estimated model) emerge:
the first (in the 1950s and 1960s) characterized by a steady high level of growth, the second
(from the mid-1970s) marked by a rapid deceleration, a third period of resurgence starting
from mid-1990s and, finally, a sharp decline since the burst of the dotcom bubble in the early
2000s, a development which has been further reinforced by the Great Recession. As a result, we
currently estimate long-run growth in the U.S. to stand at levels between 1.5 and 2 percent, in
3
line with current Congressional Budget Office estimates2 Our model also succesfully captures
the decline of GDP growth volatility during the Great Moderation, as first documented by
Perez-Quiros and McConnell (2000), and the subsequent leap in volatility due to the Great
Recession.
In a second application we derive from a panel of business cycle and financial variables
an indicator of financial stress for the euro area. In this respect, we follow a literature that
has grown tremendously in the recent past, as the financial crisis has shown that the financial
sector can be the origin of powerful shock that have disruptive repercussions on the business
cycle. As a result a large number of Financial Condition Indexes (FCIs) have been developed,
aimed at summarising in a single indicator the (sometimes conflicting) signals coming from
different segments of the financial system. A non exhaustive list of contributions on the topic
includes Illing and Liu (2006), Hakkio and Keeton (2009), Hatzius et al. (2010), Matheson
(2011), Brave and Butters (2012), Hollo’ et al (2012) and Koop and Korobilis (2014).3 Our
modeling approach allows us to address three related challenges in the construction of measures
of financial stress. The first relates to model instability. Since the financial/real economy nexus
is not necessarily stable over time, being shaped by the evolution of technological progress
and by financial innovation, our time varying parameters model seems well suited to capture
shifts in the correlations between the involved variables. The second issue that we address is
endogeneity. Indeed, the relevant information that one would like to extract from financial
variables regards primitive shocks that are coincident or leading with respect to the business
cycle, rather than an endogenous response of the financial sector to past shocks originated in the
real economy. We show how to place appropriate restrictions in the State-Space representation
of our model that allow us to extract a financial component that Granger-causes the business
cycle. Third, since our estimation method handles both missing data and mixed frequencies
we can (i) deal with the fact that some financial indicators are very short and have irregular
patterns (ii) model montly data together with quarterly GDP. We assess the performance
of our model in a nowcasting exercise in which financial variables are used to complement
standard business cycle indicators in monitoring GDP developments in real time. The model
detects significant shifts in the parameters and delivers GDP forecasts that are on average more
accurate than those obtained on the basis of simple univariate benchmarks.
The rest of the paper is structured as follows. Section 2 constitutes the theoretical body of
the paper, where we present the main theorems as well as the extensions discussed above. In
Section 3 we apply the model to the problem of measuring GDP in the U.S. In Section 4 we
use it to measure financial stress in a hte euroa area. In Section 5 we conclude. A technical
Appendix collects the proofs.
2
Up-to-date estimates of potential output for the U.S. are available here https://research.stlouisfed.
org/fred2/series/GDPPOT
3
Most of the available FCI papers borrow their methodological setup from the factor model literature that
was developed in the 2000s to forecast GDP and inflation in a data-rich environment, see Stock and Watson
(2002), Forni, Hallin, Lippi and Reichlin (2000), Doz, Giannone and Reichlin (2012) and the review article by
Stock and Watson (2010).
4
2
The score driven state space model
Let’s assume that a given time series model has the following state space representation
yt = Zt αt + εt , εt ∼ N (0, Ht ),
αt+1 = Tt αt + ηt , ηt ∼ N (0, Qt ),
t = 1, ..., n,
(1)
where yt is the N × 1 vectors of observed variables, εt is the N × 1 vectors of measurement
errors that are Gaussian distributed with zero mean and covariance matrix Ht , αt is the m × 1
vector of state variables and ηt is the corresponding m × 1 vector of Gaussian disturbances with
zero mean and covariance matrix Qt . Thus, the system matrices are Zt that is N × m matrix,
Ht that is N × N , and Tt and Qt that are m × m.
The observation and the state vector are conditionally Gaussian given the information
set Yt−1 = {yt−1 , ..., y1 } and the vector of parameters θ; namely yt |Yt−1 θ ∼ N (Zt at , Ft ) and
αt |Yt−1 θ ∼ N (at , Pt ), and the conditional log-likelihood function is equal to
`t (yt |Yt−1 ; θ) = −
N
log 2π + log |Ft | + vt0 Ft−1 vt .
2
(2)
The prediction error, vt , its covariance matrix Ft , the conditional mean of the state vector
at and its covariance matrix Pt , are estimated optimally (in the sense of the minimum mean
square error) using the KF recursions
vt = yt − Zt at , Ft = Zt Pt Zt0 + Ht , Kt = Tt Pt Zt0 Ft−1 ,
at|t = at + Pt Zt0 Ft−1 vt , Pt|t = Pt − Pt Zt0 Ft−1 Zt Pt ,
at+1 = at + Kt vt , Pt+1 = Tt Pt Tt0 − Kt Ft Kt0 + Qt , t = 1, ..., n,
(3)
where at = E(αt |Yt−1 ) and at|t = E(αt |Yt ) are the predictive filter and real-time filter respectively together with their MSEs Pt = E[(at −αt )(at −αt )0 |Yt−1 ] and Pt|t = E[(at −αt )(at −αt )0 |Yt ].
For the Gaussian model, the matrices Zt , Ht , Tt and Qt can be time-varying but are assumed
to be given (or predetermined); see details in Harvey (1989). In case those matrices contain
time-varying parameters driven by a additional stochastic processes the KF looses its optimality
and Bayesian simulation techniques are usually needed; see Durbin and Koopman (2001).
In this paper we propose a new approach in which the parameters variation is driven by past
observations only. In particular, we exploit the score driven models recently proposed by Creal
et al (2013) and Harvey (2013). Specifically, first we assume a given conditional distribution
of the observations yt ∼ p(yt |ft , Yt−1 , θ), where we condition to past informations Yt−1 , to the
vector of static parameters θ and to the vector ft containing the last estimated time-varying
parameters. Secondly, we posit a law of motion for the the vector of time-varying parameters
ft+1 = ω + Φft + Ωst ,
5
st = St ∇t ,
(4)
where the driving mechanism st is equal to the scaled-score of the conditional likelihood:
∂`t
∇t =
,
∂ft
St = −Et
∂`2t
∂ft ∂ft0
−i
,
with `t = log p(yt |ft , Yt−1 , θ).
Therefore, for i = 0 we have a gradient based algorithm in which st = ∇t is zero mean and
variance equal to It which is the Fisher information matrix. Alternatively, for i = 1 we have
that St = It−1 and st has variance equal to It−1 , and for i = −1/2 we have that st has variance
equal to identity matrix I. If it is not differently stated we consider the case of i = 1. Note
that in order to avoid numerical instability it is often desirable to replace St with a smoothed
estimator S̃t = (1 − λ)St + λS̃t−1 .
In this setup the system matrices are function of the past estimated parameters, ft , and
the static parameters θ; namely, Zt = Z(ft , θ), Tt = T (ft , θ), Ht = H(ft , θ) and Qt = Q(ft , θ).
The the conditional Gaussianity is preserved and the predictive likelihood (2) is amended to
account for the parameters’ variation.
Let’s focus briefly on the updating rule (4); at each point in time the vector of time-varying
parameters is updated so as to maximise the local fit of the model. Specifically, the size
of the update depends on the slope and curvature of the likelihood function. As such, the
updating law of motion (4) can be rationalised as a stochastic analog of the Gauss–Newton
search direction for the time-varying parameters. Blasques et al. (2014) show that updating
the parameters using the score is optimal, as it locally reduces the Kullback-Leibler divergence
between the true conditional density and the one implied by the model.
Creal et al (2008) propose the use the score-driven approach to model the parameters’
variation for univariate unobserved component model and we generalised their result for any
Gaussian state space model with time-varying parameters.4 Specifically, the two theorems
below describe how the standard KF is augmented by a new set of recursions that compute
the scaled-score st driving the parameters’ variation.
Let first introduce some notation. Given a N × m matrix X, vec(X) is the vector stacking
the columns of X one underneath the other, thus vec(X) is a N m×1 vector. Moreover, vech(X)
eliminates all supradiagonal elements of X from vec(X), thus vech(X) is a N (N + 1)/2 × 1
vector. Following Abadir and Magnus (2005, ch 11), we define the following matrices: the N 2 ×p
duplication matrix D and the p × N 2 cancellation matrix D+ , such that Dvech(X) = vec(X)
and D+ vec(X) = vech(X). It is worth stressing that the cancellation matrix is equal to
the left inverse of duplication matrix, namely D+ = (D0 D)−1 D0 . Moreover, the N m × N m
commutation matrix CN,m is such that CN,m vec(X) = vec(X 0 ), and for N = m the m2 × m2
commutation matrix is denoted by Cm . The identity matrix of order N is denoted by IN . We
also use the symemtrizer matrix Nn = 21 (In2 + Cn ), such that Nn vec(A) = vec 12 (A + A0 ), and
for A symmetric Nn vec(A) = vec(A). Finally, “⊗” is the Kronecker product.
4
Creal et al (2008, sec. 4.4) focus on the local level model with stochastic volatility proposed by Stock and
Watson (2007).
6
Theorem 1 Given the Gaussian model (1) with the conditional log-likelihood (2), the gradient
and information matrix are obtained as follows
•0
1 • 0 −1
−1
−1
F t (Ft ⊗ Ft )[vt ⊗ vt − vec(Ft )] − 2V t Ft vt ,
=
2
•
•0
•
1 • 0 −1
−1
−1
=
F t (Ft ⊗ Ft )F t + 2V t Ft V t , t = 1, ..., n,
2
∇t
It
•
where V t =
∂vt
∂ft0
•
and F t =
∂vec(Ft )
.
∂ft0
(5)
(6)
Proofs in the Appendix A.1.
The previous theorem show how to compute analytically the score and the information
matrix, which depend on the prediction error vt and its covariance matrix Ft that are recursively
•
•
computed using the KF (3). Furtheremore, we have two new elements V t and F t and the
following theorem shows the analytical expressions to recursively compute them.
Theorem 2 In order to compute the gradient and the information matrix we need the following
set of additional recursions:
•
•
•
V t = −[(a0t ⊗ IN )Z t + Zt At ],
•
t = 1, ..., n,
•
•
(7)
•
F t = 2NN (Zt Pt ⊗ IN )Z t + (Zt ⊗ Zt )P t + H t ,
•
•
(8)
•
K t = (Ft−1 Zt Pt ⊗ Im )T t + (Ft−1 Zt ⊗ Tt )P t
•
•
+(Ft−1 ⊗ Tt Pt )CN m Z t − (Ft−1 ⊗ Kt )F t ,
•
•
•
•
•
•
(9)
•
At+1 = (a0t ⊗ Im )T t + Tt At + (vt0 ⊗ Im )K t + Kt V t ,
•
(10)
•
P t+1 = (Tt ⊗ Tt )P t − (Kt ⊗ Kt )F t + Qt
•
•
+2Nm [(Tt Pt ⊗ Im )T t − (Kt Ft ⊗ Im )K t ],
•
where Z t =
∂vec(Zt )
,
∂ft0
•
Ht =
∂vec(Ht )
,
∂ft0
•
Tt =
∂vec(Tt )
∂ft0
•
and Qt =
∂vec(Qt )
.
∂ft0
(11)
Proofs in the Appendix A.2.
Remark 3 The expressions (5)-(11) generalize the results in Harvey (1989, pp. 140-3) for the
case of time-varying parameters.
•
•
•
•
It is worth to stress that the Jacobians matrices Z t , H t , T t and Qt are model specific,
i.e. they depend on the specific time-variation of the model under consideration. In the
simplest case they result in selection matrices, but in general they incorporate Jacobians due
to parameters restrictions. This point will be clarified in later on.
Putting together Theorem 1 and 2, we obtain the new set of recursions (5)-(11) running in
parallel with the KF (3) to compute the scaled score st and thus the time-varying parameters
in (4). To summarise, the augment algorithm runs as described below.
•
•
Algorithm 4 Given the initial values a1 , P1 , A1 , P 1 and f1 , for t = 1, ..., n:
• evaluate Zt , Ht , Tt , Qt ;
7
•
•
•
•
• evaluate Z t , H t , T t , Qt ;
• compute vt , Ft , `t , Kt ;
•
•
•
• compute V t , F t , K t , ∇t , It , st ;
•
•
• update at+1 , Pt+1 , At+1 , P t+1 , ft+1 .
All the static parameters in the law of motion (4), and possibly any static element in Zt ,
Tt , Ht and Qt , are collected in the vector θ, which is then estimated by maximum likelihood
P
(ML); θb = arg max nt=1 `t (θ). The evaluation of the log-likelihood is straightforward and the
maximisation can be obtained numerically. Following Harvey (1989, p. 128) we have that
√ b
n(θ − θ) → N (0, Ξ), where is the asymptotic variance Ξ is evaluated by numerical derivative
at the optimum as discussed in Creal et al (2013, sec. 2.3).
2.1
State space models with mean adjustment
It is sometimes convenient to include mean adjustment terms in (1) resulting in
yt = Zt αt + dt + εt , εt ∼ N (0, Ht ),
αt+1 = Tt αt + ct + ηt , ηt ∼ N (0, Qt ),
(12)
where dt and ct are known vector possibly time-varying. While the specification (1) is adequate
for most purposes, the specification (12) may be suitable for occasional use and the filtering
(3) is amended only in the following elements
vt = yt − Zt at − dt ,
at+1 = Tt at + ct + Kt vt ,
(13)
thus expressions (7) and (10) are modified as follows
•
•
•
•
V t = −[(a0t ⊗ IN )Z t +dt + Zt At ],
•
•
•
•
•
•
At+1 = (a0t ⊗ Im )T t +ct + Tt At + (vt0 ⊗ Im )K t + Kt V t ,
•
where dt =
2.2
∂dt
∂ft0
•
and ct =
∂ct
∂ft0
(14)
will be model specific.
Missing observations
Assume to have a data set containing missing observations, namely the observed vector
is Wt yt , where Wt is Nt × N selection matrix with 1 ≤ Nt ≤ N, meaning that at least one
observation is available, Wt is obtained by eliminating the i − th row from IN if the i − th
variable is missing at time t and if there is no missing observations Wt = IN . If no data is
available Nt = 0, the KF computes the prediction step only and the likelihood is equal to a
constant so that st = 0, and the time-varying parameters are not updated.
8
The measurement equation (12) is therefore modified taking into account the missing observations as following
Wt yt = Wt Zt αt + Wt dt + Wt εt , Wt εt ∼ N (0, Wt Ht Wt0 ).
(15)
Thus, at each point in time the likelihood `t is computed using Nt observations, the KF
recursions (3) are modified as follows
vt = Wt (yt − Zt at − dt ),
Ft = Wt (Zt Pt Zt0 + Ht )Wt0 ,
Kt = Tt Pt Zt0 Wt0 Ft−1 ,
(16)
while the expression for at+1 and Pt+1 are still the same. Therefore, the formulae (7), (8) and
(9) need to be accommodated as following
•
•
•
•
V t = −[(a0t ⊗ Wt )Z t + Wt dt + Wt Zt At ],
•
•
•
•
F t = 2NNt (Wt Zt Pt ⊗ Wt )Z t + (Wt Zt ⊗ Wt Zt )P t + (Wt ⊗ Wt )H t ,
•
•
•
K t = (Ft−1 Wt Zt Pt ⊗ Im )T t + (Ft−1 Wt Zt ⊗ Tt )P t
•
(17)
•
+(Ft−1 Wt ⊗ Tt Pt )CN m Z t − (Ft−1 ⊗ Kt )F t
•
•
while recursion (10) and (11) for At+1 and P t+1 are still the same. While the formulae in (16)
can be find in standard textbook such as Durbin and Koopman (2001), the formulae (17) can
be obtained following similar steps as the Appendix by taking into account the selection matrix
Wt .
2.3
Mixed frequencies and temporal aggregation
This section deals with the relation between the high frequency variable xt , which for key
economic concepts are unobserved, and the corresponding observed low frequency series, xkt
with k > 1. The relation between the observed low frequency variable and the corresponding
indicator depends on whether the variable is a flow or a stock variable and on how the variable is
transformed before entering the model. In all cases the variable can be rewritten as a weighted
average of the unobserved high frequency indicator, specifically
xt =
2k−2
X
ωjk xkt−j
(18)
j=0
Here a summary of the implied weights (see e.g. Banbura and Modugno 2014):
• If the variable enters in level and is a stock variable then ωjk = 1 for j = 0 and ωjk = 0
for j > 0.
• If the variable enters in level and is a flow then ωjk = 1 for j = 0, ..., k − 1 and ωjk = 0 for
9
j ≥ k.
• If the variable enters in first difference and is a stock ωjk = 1 for j = 0, ..., k − 1 and
ωjk = 0 for j ≥ k.
• If the variable enters in first difference and is a flow ωjk = j + 1 for j = 0, ..., k − 1 and
ωjk = 2k − j − 1 for j ≥ k.
Assume that the unobserved high frequency variable follows the state space model
2
yi,t = Zi,t αt + εi,t , εi,t ∼ N 0; σi,t
,
(19)
we consider a single indicator (e.g., the GDP) to be aggregated from quarterly to monthly (i.e.
k = 3). Therefore, at quarterly frequency we observe
q
yi,t
=
4
X
ωj3 Zi,t−j αt−j +
j=0
4
X
ωj4 ε4i,t−j ,
(20)
j=0
and the state space needs to be accommodated taking into account the aggregation (20) and
the implied missing observations.5 In empirical section we will describe in detail the state space
specification.
2.4
Relationship with the existing literature
Koopman et al (2010) introduce time-varying parameters in the dynamic Nelson–Siegel
yield curve model of interest rates of different maturities. This results in a state space model
in which the factor loading are time-varying and the volatility follows a GARCH process. The
state vector is augmented in order to include the time-varying loading and as such they obtain
a non linearity masurement equation which is solved by means of the extended KF. Moreover,
they allow for time-varying volatility. In the spirit of Harvey, Ruiz, and Sentana (1992) they use
the common GARCH specification where the time-varying volatility is driven by the squared
error estimated within the KF as additional element of the state vector.
SI POTREBBE CERCARE DI VEDERE SE APPROCCIO DI Harvey, Ruiz,
and Sentana (1992) SI PUO RAZIONALIZZARE NEL NOSTRO FRAMEWORK.
Koop and Korobilis (2013) propose the use of the adaptive algorithm, specifically forgetting
factor algorithm, to deal with time-varying parameters and stochastic volatility in large VARs
models. In a different paper, Koop and Korobilis (2014) exploit the adaptive algorithm to model
the parameters variation in a state space framework using a two-step procedure, in which the
unobserved state vector is extract by means of KF. Here, we show that this approach is a
special case of our general framework. Specifically, if the state vector αt is observable (and
5
Note that although the aggregation implies that the measurement error now follows a moving average
process of order related to the dimension of the high frequency, it nevertheless remains white noise when
observed at the low frequency frequency (see Auroba et al., 2011). Hence we treat the measurement error at
the low frequency as white noise in what follows.
10
denoted with at ) we have that both the state space model reduces to two regression models
with rime-varying parameters:
yt = XZt βZt + εt , εt ∼ N (0, Ht ),
at+1 = XT t βT t + ηt , ηt ∼ N (0, Qt ),
(21)
where βZt = vec(Zt ), XZt = (a0t ⊗ IN ), βT t = vec(Tt ) and XT t = (a0t ⊗ Im ). For simplicity, we
consider only the first equation of (21), while the derivation for the second equation follows
accordingly. The conditional log-likelihood for yt is
`t = −
N
log (2π) + log |Ht | + ε0t Ht−1 εt ,
2
0
and the vector of time-varying parameters is ft = (βZt
, vec(Ht )0 )0 . Given the formulae (5)-(6),
the scaled score is qeual to st = (s0βt , s0Ht )0 with:
0
−1 0
•
•
•
−1
sβt = − B t Ht B t
B t Ht−1 εt ,
0
−1 0
•
•
•
−1
−1
−1
−1
sHt = H t (Ht ⊗ Ht )H t
H t (Ht ⊗ Ht )[εt ⊗ εt − vec(Ht )] .
•
•
•
∂εt
= −XZt , H=D if we model the vech(Ht ), or H t = IN 2 if we model vec(Ht ).
with B t = ∂β
0
Zt
Finally, the two scaled score driving the drifing coefficients and volatilities are:
sβt = Xt0 Ht−1 Xt
−1
Xt0 Ht−1 εt ,
sHt = [εt ⊗ εt − vec(Ht )].
Assuming a random walk law of motion for ft we obtain the following filter to6
βZt+1 = βZt +κβ Xt0 Ht−1 Xt
−1
Xt0 Ht−1 εt ,
vec(Ht+1 ) = vec(Ht )+κH (εt ⊗εt −vec(Ht )). (22)
If we replace the scaling matrix Xt0 Ht−1 Xt with its smoothed estimator
Sβt = (1 − κβ )Sβt−1 + κβ Xt0 Ht−1 Xt
we have that the filter used in Koop and Korobilis (2013, 2014) boils down to the filter in (22).
For details on the exact equivalence see Delle Monache and Petrella (2014).
2.5
Parameters restrictions
Applications of time-varying parameters models often require imposing restrictions on the
parameters space. For instance, stationarity restrictions may be imposed or positivity of the
volatilities. When restrictions are implemented within a score-driven setup, the resulting model
can still be estimated by MLE without the need of computational demanding simulation meth6
Starting from the law of motion (4), we assume that ω = 0, Φ = I and Ω is block diagonal that depends
on two scalar only; κβ and κH regulating the time-variation of the coefficients and volatility, respectively.
11
ods. This requires the re-parameterization of the vector of time-varying parameters as follows
f˜t = ψ(ft ),
(23)
where ft is the unrestricted vector of parameters we model, and f˜t is the vector of interest
restricted through the function ψ(·). The latter is a time invariant, continuous and twice
differentiable function, often called link function (Creal et al., 2013, and Harvey, 2013). The
vector ft continues to follow the updating rule (4), but the score needs to be amended taking
into account the Jacobian of ψ(.). In practice, we model ft = h(f˜t ), where h(·) is the inverse
function of ψ(·). Given a continuous and differentiable function ψ(·), Ψt is a deterministic
function given past information, whose role is to re-weight the original score such that the
restrictions are satisfied at each point in time.
•
•
•
In our setup this transformation will have an effect on the Jacobian matrices Z t , H t , T t and
•
Qt . Specifically, we propose the general form for a generic time-varying matrix of the system:
vec (Mt ) = cM + SM 1 ψ (SM 2 ft ) ,
(24)
where Mt denotes a nr × nc matrix containing both constant and time-varying parameters, cM
is a nr nc × 1 vector with constant elements, and SM 1 and SM 2 are selection matrices. SM 1
is obtained as follows: take Inr nc , retain the columns associated to the time varying elements
of vec (Mt ), while SM 2 selects the sub-vector of ft belonging to Mt . Finally, ψ (.) denotes a
(time-invariant) link function and its Jacobian is Ψt = ∂ψ(ft )/∂ft0 . The representation (24)
helps to compute
•
M t = SM 1 Ψt SM 2 .
In the application, we will impose restrictions on the time-varying autoregressive coefficients
such that they have stable roots; this is implemented by re-parameterizing the autoregressive
model with respect to the partial autocorrelations as in Delle Monache and Petrella (2014). A
second set of restrictions concern a generic time-varying covariance matrix Σt to be positive
defined and this achived by means of log-Cholesky transformation. Specifically, we have Σt =
Lt L0t , with Lt being a lower triangular matrix, and we model vech(L̃t ), where L̃ij,t = Lij,t for
i 6= j, and L̃ii,t = log Lii,t . Thus, the Jacobian is equal to
Ψt =
∂vec(Σt )
= (IN 2 + KN )(Lt ⊗ IN )SL Slog t ,
∂vech(L̃t )0
where KN commutation matrix, SL is a selection matrix such that SL vech(Lt ) = vec(Lt ) and
it is obtained form diag(vec(Lt )) by dropping the columns containing only 0s and replacing
∂vech(Lt )
the no-zeros elements with 1s. Finally, Slog t = ∂vech(
= diag(exp(L̃ii,t )). Note that in case
L̃t )0
2
Σt = σ is the time-varying variance, ψ(˜lt ) = exp(2˜lt ), where ˜lt = log σt , and Ψt = 2 exp(2˜lt ).
t
12
3
Empirical application 1: GDP plus revisited
Our first empirical application consists of extending the model proposed by Aruoba et al.
(2016) in the context of GDP measurement. Because the source of all incomes lies in the
flow of value added generated by production and all production must be either consumed at
home or abroad or else invested, Gross Domestic Product (GDP) can be measured in different
ways. In the U.S. two GDP estimates exist, a widely-used expenditure-side version, GDPE ,
and a much less widely-used income-side version, GDPI . These two related concepts can,
at times, diverge quite significantly. In the top panel of Figure (1) we show the annualized
percentage quarterly rate of growth of both GDPE (green line) and GDPI (blue line) as well
as their difference (bottom panel). The discrepancy between the two concepts is generally non
negligible and it can be as large as five percentage points on an annualized basis. The question
of which GDP measure is more reliable is of particular interest to policy makers who need to
monitor business cycle developments in real time, since some researchers argue that GDPI ,
which has traditionally received much less attention than GDPE , has actually done a better
job at recognizing the start of recessions, see Nalewaik (2012).
Aruoba et al. (2016) suggest to use the two existing measure of GDP computed from the
expenditure side and from the income side to extract, through optimal filtering techniques, a
measure of underlying GDP that is relatively cleaner of measurement error. To this end they
work with a dynamic factor model where both statistical measures of output are assumed to
be driven by a common latent component (which they label GDP plus). An optimal estimate
of GDP plus is then delivered by the Kalman filter conditional on the parameters of the model,
which are estimated via maximum likelihood. In our empirical exercise we extend their framework by allowing for the parameters of the dynamic factor model to change over time, driven
by the score of the conditional likelihood.
Formally, the baseline constant parameter specification from which we start is described in
the following State-Space system:7
"
GDPE,t
GDPI,t
#
"
=
1
1
#
"
αt +
εE,t
εI,t
#
"
,
αt+1 = ρ0 + ρ1 αt + ρ2 αt−1 + et ,
εE,t
εI,t
#
∼ N (0, H)
(25)
2
et ∼ N (0, σ )
where H is a full 2 × 2 matrix.8 It can be readily recognized that the model is a traditional
small-scale dynamic factor model in the spirit of Sargent and Sims (1977). We extend the
7
Aruoba et al. (2016) consider up to five alternative model specifications, which differ from each other with
respect to the correlation structure of the measurement errors. Given that the focus of our paper is mainly
methodological, in our application we consider only one of their specifications.
8
In the original specification, Aruoba et al. (2016) work with an AR(1) transition equation. We prefer an
AR(2) specification that is more in line with traditional business cycle models. Notice, in particular, that the
spectrum of an AR(1) model has its peak at zero and most of its mass on the left hand side of the frequency
band, while in an AR(2) model the peak of the spectrum can occur at cyclical frequencies.
13
model to allow for all the seven parameters involved to change over time:
"
GDPE,t
GDPI,t
#
"
=
1
1
#
"
αt +
εE,t
εI,t
#
"
,
αt+1 = ρ0,t + ρ1,t αt + ρ2,t αt−1 + et ,
εE,t
εI,t
#
∼ N (0, Ht )
et ∼ N
(26)
(0, σt2 )
In the vector ft we collect the three group of parameters, namely the intercept and autoregressive elements of the transition equation,
the variance of the erorr
the elements of Ht and
0 0
term in the transition equation, i.e. ft = ρ0,t , ρ1,t , ρ2,t , σt2 , vech L̃t
, where L̃t is a lower
tringular matrix containing the elements of the log-Cholesky transformation of Ht as described
earlier. We assume that
ft+1 = ft + Ωst
where Ω is a diagonal matrix function of only 3 static coefficients: κρ , κσ , κl , one for each group
of parameters. These parameters are then estimated by maximum likelihood.
The results of the analysis are shown in Figures 2 to 9.
• In Figure 2 we report the estimated time varying intercept ρ0,t (blue line), together
with the estimate obtained on the basis of the constant parameters model. Two points
are worth highlighting. First, there is a secular downward trend that drives this term,
indicating that the mean growth rate of the U.S. economy has progressively fallen over
time. Second, there are marked falls corresponding to some recessions, namely those of
the early 70s and early 2000s. As we will see below, this has important implications
for the estimated long-run growth of the economy. Third the constant parameter model
severely overestimates this term at the end of the sample.
• The evolution of the autoregressive parameters is shown in Figure 3. Notice that the second autoregressive term (red line) fluctuates markedly over time, although it is reasonably
centered around zero, the value to which is constrained in the original specification by
Aruoba et al. (2016). The first AR term (blue line), also displays non-negligible time
variation, especially at the end of the sample. Aruoba et al. (2016) place a lot of emphasis on this parameter, which they estimate at around 0.5 on the whole sample. This
value is higher than the one estimated for either GDPE or GDPI , which leads them to
conclude that the filtered measure is more predictable than its components. Our findings
cast some doubt over the possibility of exploiting this auto-correlation in real time, since
it seems to be the result of the prolonged fall in macroeconomic activity during the Great
Recession rather than a robust, full-sample, feature of GDP plus. Notice that both the
autoregressive terms in Figure 2 and the time-varying interecept in Figure 1 contribute
ρ0,t
. This implies
to defining the long-run forecast of GDP, which is given by µt = (1−ρ1,t
−ρ1,t )
that underestimating or overestimating any of this terms can lead to a poor estimate
of the long-run GDP growth, so that letting both vary over time can have important
implications for the real-time assessment of long-run growth.
14
• Next, in Figure 4 we report the estimated long-run growth path implied by our model,
ρ0,t
. It is worth stressing that this is a key variable for both
computed as µt = (1−ρ1,t
−ρ1,t )
monetary and fiscal policy. Orphanides (2003), for instance, stresses how poor real-time
estimates of potential growth can generate sizeable monetary policy mistakes. According
to his analysis, for instance, part of the monetary policy mistake that led to the Great
Inflation of the 70s can be attributed to a failure in identifying a permanent slowdonw in
productivity growht in the first part of the 70s. Long-run projections are also at the center
of the fiscal policy debate as they determine the solvency of pension systems and, more
generally, the sustainability of public debt. According to our estimates, four distinct
phases of U.S. long-run growth (measured by the long-run forecast of the estimated
model) emerge: the first (in the 1950s and 1960s) characterized by a steady high level of
growth, the second (from the mid-1970s) marked by a rapid deceleration, a third period
of resurgence starting from mid-1990s and, finally, a sharp decline since the burst of the
dotcom bubble in the early 2000s, a development which has been further reinforced by
the Great Recession. As a result, we currently estimate long-run growth in the U.S. to
stand at levels between 1.5 and 2 percent, in line with current Congressional Budget
Office estimates. These results are strikingly in line with both the analysis in Benati
(2007) and with the conclusions reached by Antolin-Diaz et al. (2015).
• Figure 5 shows the variance of the error of the common component, a measure of macroeconomic volatility. It is interesting to notice that, after a prolonged period of decrease
in the 50s and 60s, volatility reaches a plateau where it stays up to the mid-80s. From
1984 onwards there is a second marked fall of volatility, corresponding to the Great
Moderation, which is interrupted, and partially reversed by the Great Recession in 2008.
• Next, in Figure 6 we report the Kalman filter gain of GDPE relative to that of GDPI over
the whole sample, which can be interpreted as the relative weights that the two observed
measures receive in the construction of GDP plus. Given how this is computed, we have
that values below 1 imply that a relatively larger weight is assigned to GDPI . From
the picture it emerges that, indeed, as claimed by part of the literature, GDPI better
captured the behaviour of GDP plus in the middle of the sample. However, in recent
years, an equal weighting scheme seems warranted. This result squares well with the fact
that from July 2015 the BEA started publishing the mean of GDPE and GDPI as an
indicator of the business cycle. Our results corroborate the choice of an equal weighting
scheme at the current juncture.
• In Figure 7 we show the resulting GDP plus estimate (blue line) together with the two
underlying components. From this picture it is clear that the estimated latent factor is
considerably smoother than the two observed series, so that in many periods GDPI and
GDPE fall outside of the confidence intervals. In Figures 8 and 9 we zoom these comparisons in two recession periods, namely the 1970s recessions and the Great Recession
of 2008. It is evident that during these turbulent periods GDP plus gives less volatile
15
indications on output dynamics than the other two measures.
Concluding, when applied to the problem of extracting a cleaner measure of output from
observed series of Gross Domestic Product, our model reveals a number of interesting aspects
that are overlooked in the constant parameters framework proposed by Aruoba et al. (2016). In
particular, our model delivers a real time/time-varying assessment (i) of the long-run economic
growth (ii) of macroeconomic volatility and (iii) of the relative importance of GDPI and GDPE .
All these three ingredients are of great interest to policy makers and business cycle analysts.
4
Empirical application 2: measuring financial stress in
the euro area
We focus our second empirical application on a measure of financial stress for the euro
area. Since the focus is on the methodological aspects of our model, we do not engage in a
search of new financial indicators but rely on the 15 sub-indexes that form the ECB-CISS as
the representative indicators of financial stress for the euro area.9 and model these 15 variables
together with 4 business cycle indicators, namely GDP, industrial production and two survey
indicators, the composite Purchasing Manager Index (PMI) and its orders subcomponent. The
financial stress variables are:
• Realised volatility of the 3-month Euribor rate, Interest rate spread between 3-month
Euribor and 3-month French T-bills, Monetary Financial Institution’s (MFI) emergency
lending at Eurosystem central banks
• Realised volatility of the German 10-year benchmark government bond index, Yield
spread between A-rated non-financial corporations and government bonds (7-year maturity bracket), 10-year interest rate swap spread: weekly average of daily data
• Realised volatility of the Datastram non-financial sector stock market index, CMAX
for the Datastream non-financial sector stock market index (Maximum cumulated index
losses over a moving 2-year window), Stock-bond correlation
• Realised volatility of the idiosyncratic equity return of the Datastream bank sector stock
market index over the total market index, Yield spread between A-rated financial and
non-financial corporations (7-year maturity), CMAX as defined above interacted with the
inverse price-book ratio (book-price ratio) for the financial sector equity market index
• Realised volatility of the euro exchange rate vis-à-vis the US, dollar, Realised volatility of
the euro exchange rate vis-à-vis the Japanese Yen, Realised volatility of the euro exchange
rate vis-à-vis the British Pound
9
We thank the authors of the CISS for sharing their data with us.
16
See Hollo’ et al. (2012) for an accurate description of their construction. In the next two
subsections we discuss more in details the dynamic factor model specification and how this can
be nested in the more general State Space framework discussed in Section 2. Next we present
the results of full sample estimation and of a real time forecast exercise.
4.1
Financial stress and economic conditions
The vector of observable is yt = (ytr0 , ytx0 )0 , in which ytr0 in the vector macroeconomic variable
including also the quarter on quarter growth rate of GDP and other cyclical indicators, and
ytx0 , the chosen financial indicators. Although financial variables are potentially available at
frequencies higher than monthly, we restrict ourselves to the use of monthly averages in order to
limit the dimension of the state vector and also because a monthly update of such an indicator
seems sufficiently informative for policy makers.
Hatzius et al. (2012) define the financial condition as ”the current state of financial variables
that influence economic behaviour and (thereby) the future state of the economy”. Furthermore
they add: ”ideally, a FCI should measure financial shocks – exogenous shifts in financial conditions that influence or otherwise predict future economic activity. True financial shocks should
be distinguished from the endogenous reflection or embodiment in financial variables of past
economic activity that itself predicts future activity”. The definition above is implicitly a statement about Granger-causality; i.e. we look for an indicator of financial activity that predicts
output beyond the information in the past levels of economic activity. However, in Hatzius et
al. (2010) the financial assets beside responding to financial shocks also react to real shocks as
well as the predictable component of real activity. The latter though contradicts the idea that
all available information is already priced, which implies that asset prices should only respond
to the unpredictable component, i.e. the news, in real activity. Therefore, our proposed FCI is
defined as the summary of all the relevant information in the asset prices that helps to predict
the future behaviour of the coincident indicator of economic activity; it is a leading indicator
of future activity.
Formally, we can cast the model in the following state space form. The measurement
equation is block diagonal and can be written as
"
ytr
ytx
#
"
=
λrt 0
0 λxt
#"
αtr
αtx
#
"
+
εrt
εxt
#
"
,
εrt
εxt
#
∼ N (0, Ht ).
(27)
Time variation in the loading vector allows for the relevance of the variables included in the
system for the factors to vary over time. For instance the impact of financial variables that
are related to the real estate market is likely to have increased during the early part of the
latest recession. More generally, the information content of different asset classes could change
in different stages of the business cycle.
17
The transition equation imposes the Granger-causality restrictions, specifically
"
r
αt+1
x
αt+1
#
"
= Φ1t
r
αt−1
x
αt−1
#
"
+ Φ2t
r
αt−2
x
αt−2
#
+ ut ,
ut ∼ N (0, Σt )
(28)
where Φjt are upper triangular matrices. The time variation in the feedback matrices in the
transition equations is meant to capture possible time variation in the leading properties of
the financial indicator. Furthermore, time varying volatilities are allowed for both in the
measurement error and to the innovations of the transition equations. The former are included
so as to pick up changes in the volatility that are specific to the variables included in the
system, whereas the latter capture common changes in volatilities (and the associated time
varying correlation), i.e. the great moderation (see e.g. Perez-Quiros and McConnell, 2000) or
the business cycle related fluctuations in uncertainty documented by Jurado et al. (2015) as
well as the broader increase of the volatility of financial markets.
Note that the contemporaneous relation between the factors is fully captured by the covariance matrix Σt on which we have made no restriction. Therefore, current development in
real activity, i.e. in the coincident indicator αxt , affect contemporaneously asset prices in this
context. However, it is important to stress that it is only the unpredictable component, i.e. the
shock to real activity, that affects the asset prices (therefore implicitly incorporating the idea
that financial market are efficient or that informational frictions are not relevant at monthly
frequency). To illustrate this point let us consider the identification that financial shocks affect
real activity with a (1 month) delay, financial market react instantaneously to any development
in the real economy. This ordering restriction allows to identify real shocks, urt , from financial
shocks, uxt ,
#
#
"
"
#"
ert
h11t 0
ert
∼ N (0, I).
(29)
,
ut =
ext
ext
h21t h22t
Substituting this into the measurement equation we have that the unpredictable component of
the real activity, urt , has a direct impact on asset prices. Therefore assuming a block diagonal
measurement equation ensures that is only the shocks, i.e. the news, in real economic activity
that move the financial markets (and not their predictable component). Note that even though
the measurement equation is block diagonal the fact that the covariance matrix is unrestricted
implies that financial variables help to nowcast current economic activity. In fact, it is easy
to show that the Kalman gain is not diagonal in this setting. Intuitively, if financial markets
react systematically to macroeconomic developments, their reaction helps to pin down the real
shocks and therefore to improve the nowcast of the coincident indicator and its components.
4.2
The state space specification
Here we discuss the details of the state space representation of the factor model (27)-(28)
taking into account the mixed frequency and aggregation. For the macro variables ytr , we have
three types of indicators: one quarterly (i.e. GDP, ytq ), three monthly business cycle indicators
18
ytm (industrial production and two surveys) and 15 monthly financial indicators, ytx . Therefore,
the vector of observables yt = (ytq , ytm0 , ytx0 )0 have dimension 19. The state vector is
0
y
y
y
y
x
αt = αty , αtx , αt−1
, αt−1
, αt−2
, αt−3
, αt−4
, εqt , εqt−1 , εqt−2 , εqt−3 , εqt−4
where αty is the real factor αtx is the financial factor and εqt is the measurement error of the
quarterly macro variable (i.e., GDP) and due to the time aggregation we have the moving
average terms. The corresponding time-varying matrix of factor loading, consistent with the
time aggregation, is equal to



Zt = 


1 q
λ
3 t
1 s
λ
3 t
λm
t
0
0
0
0
λxt
2 q
λ
3 t−1
1 s
λ
3 t−1
0
0
0
0
0
0
λqt−2
1 s
λ
3 t−2
0
0
2 q
λ
3 t−3
1 q
λ
3 t−4
0
0
0
0
0
0
1
3
2
3
1 23 31
0 0 0 0 0
0 0 0 0 0
0 0 0 0 0






x
where λqt is scalar, λst is ns × 1 vector, λm
t is nm × 1 vector and λt is nx × 1 vector. Note that we
impose restrictions on the factor loadings such that the one relative to the first macro variables
(i.e. the GDP) and the one relative to the first financial variables are both normalize to one.
m0 x0 0
The measurement errors is εt = (0, εs0
t , εt , εt ) ∼ N (0, Ht ), where Ht is diagonal and thus
we impose positive volatility on the elements of Ht by log transformation element by element.
Consistently with the previous discussion on the identification of real and financial factors,
the transition equation is specified as follows

Tt
Qt
φ11,t φ12,t φ13,t φ14,t
"
#

 0 φ22,t
T11,t T12
0 φ24,t
=
, T11,t = 

T21 T22
0
0
0
 1
0
1
0
0
h
i
= blockdiag Σt 05×5 σε2q t 04×4






while T12 , T21 and T22 are sub-matrices of 0s and 1s. In order to impose stable roots in the
transition matrix Tt , we use the transformation described early on for the two pairs
φr,t = (φ11,t , φ13,t )0 ∈ S 2 and φx,t = (φ22,t , φ24,t )0 ∈ S 2
where S 2 is the space with stable roots, while φ12,t and φ14,t are left unrestricted. Finally, we
restrict the matrix Σt to be positive definite using the log-Cholesky transformation. Collecting
all the time varying parameters in the vector ft we specialize their score driven law of motion
as
ft+1 = ft + Ωst
We restrict the matrix Ω to be diagonal and to depend on two constants, κs and κv , where
the former drives the amount of time variation in the factor loadings and in the AR coefficients,
19
while κv is a smoothing parameter that governs the time variation of the model volatilities.
Given that the model dynamics depend only on these constants and that irregular behaviour of
the financial variables over the considered sample makes likelihood estimation sensitive to the
starting parameter values, we estimate the model via dynamic model selection and dynamic
model averaging, see the Appendix for a discussion.
We start from a description of the full sample parameter estimates and of the estimated
business cycle and financial factors.
In Figures 10 and 11 we report a time series plot of the weights assigned by the Dynamic
Model Averaging algorithm and the indicator of the best model as selected by Dynamic Model
Selection. It can be noticed that the model achieves model sparsity, in the sense that, after a
necessary period of learning, it assigns nontrivial weights only to a limited number of models.
Figure 11, on the other hand, reveals that the highest weights are typically assigned to models
in the upper part of grid, where time variation is relatively more pronounced.
Figure 12 shows the estimated factors. For comparison, we also report the ECB-CISS,
rescaled to have zero mean and unit variance. Two observations are in order. First, the
business cycle factor captures the main features of the euro-area business cycle, namely the
two expansion phases before the Great Recession, as well as the double dip downturn generated
by the Great Recession first and by the Sovereign Debt Crisis. The estimated financial factor,
our measure of financial stress, also captures correctly the two period of financial stress after
the Lehman and the Greek crises. Unlike the CISS, however, it also signals the build up of
financial vulnerabilities in the early part of the sample.
Next, in Figures 13 to 14 we show the time varying elements of the covariance matrix of
the unobserved states, i.e. the volatilities of the business cycle and of the financial factor, as
well as their time varying covariance. The model detects significant shifts in the volatilities in
correspondence with the two recent recessions, although, financial stress volatility is quite high
already at the beginning of the 2000s. The covariance between the disturbances of these two
factors, rather stable for most of the sample, turns significantly negative after 2007, a period
of both severe financial stress and real economy recession. The volatilities of the measurement
equations, shown in Figure 16, similarly display significant spikes towards the end of the sample.
Finally, in Figures 17 and 18 we show the time varying factor loadings. Although some variation
can be noticed in these coefficients, especially when DMS is used, it seems the most of the time
variation in the whole model originates from the variances rather than from the factor loadings.
Finally, in Figure 19 we report the results of a forecast exercise in which we predict GDP
1 to 18 months ahead and contrast the results of our model with that of two simple univariate
benchmarks, namely an AR and a random walk model. The figure shows the evolution of the
RMSE relative to that of the AR model for the various forecast horizons of the model selected
with DMA, DMS or the simple median model. Notice that the RMSE of our factor models
decline as the forecast horizon shortens, meaning that they use new information efficiently.
Also the factor model always outperforms the AR model at short-medium horizons (between
1 and 10 months ahead).
20
5
Conclusions
In this paper we develop a score-driven approach to estimate state-space models with timevarying parameters. By letting the dynamics of the parameters be driven by the score of
the predictive likelihood we show that the model retains many desirable properties, namely it
can be analyzed using an extended KF procedure. We derive the new set of recursions that,
running in parallel with the standard KF , allows us estimate simultaneously the unobserved
state vector and the time-varying parameters. Given that a variety of time series models
have a state space representation, the proposed methodology has to be considered
of wide interest in econometrics and applied macroeconomics.
After presenting the main theorems we discuss two extensions of the theoretical framework
that could be particularly valuable in applied work. First, since in some applications the
researcher would like to impose parameters restrictions, such as stationarity and non-negative
variances, we show how to incorporate general restrictions on the model parameters. Second,
we illustrate how the model can be extended to deal with data at mixed frequencies and
missing observations. We conclude the theoretical part of the paper by briefly discussing how
our estimation strategy is sufficiently general to nest existing adaptive algorithms used in the
literature as a special case. Given that a wide range of time series models have a state
space representation, the estimation method we propose is of interest for a large
number of applications in econometrics and applied macroeconmics.
Specifically, we illustrate the potential of the method through two empirical applications.
In the first exercise we focus on business cycle measurement. In particular, we take as a
starting point the model that Aruoba et al. (2016) have developed to estimate GDP on the
basis of underlying (noisy) measures and extend it to account for time variation in the model
parameters. Our model delivers a real time and time-varying assessment (i) of the long-run
economic growth (ii) of macroeconomic volatility and (iii) of the relative importance of GDPI
and GDPE . In a second application we extract from a panel of business cycle and financial
variables an indicator of financial stress for the euro area. Our modeling approach allows us
to address three related challenges in the construction of measures of financial stress. The
first relates to model instability. The second issue that we address is endogeneity and we
place appropriate restrictions in the State-Space representation of our model that allow us
to extract a financial component that Granger-causes the business cycle. Third, since our
estimation method handles both missing data and mixed frequencies we can (i) deal with the
fact that some financial indicators are very short and have irregular patterns (ii) model montly
data together with quarterly GDP. We assess the performance of our model in a nowcasting
exercise in which financial variables are used to complement standard business cycle indicators
in monitoring GDP developments in real time. The model detects significant shifts in the
parameters and delivers GDP forecasts that are on average more accurate than those obtained
on the basis of simple univariate benchmarks.
21
Figure 1: GDP(E) vs. GDP(I)
22
Figure 2: Time varying intercept
Figure 3: Time varying AR coefficients
23
Figure 4: Long-run growth
Figure 5: Variance of the error of the common component
24
Figure 6: Relative Kalman Filter Gain
Figure 7: Estimated Common Factor, GDPE and GDPI
25
Figure 8: Zooming in the 1970s recession
Figure 9: Zooming in the Great Recession
26
Figure 10: Model weights (DMA)
Model weights
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
1986
1989
1992
1995
1998
2001
2004
2007
2010
2013
2010
2013
Figure 11: Best Model (DMS)
Best model (Dynamic Model Selection)
70
60
50
40
30
20
10
0
1986
1989
1992
1995
1998
2001
27
2004
2007
Figure 12: Estimated Factors
Estimated Business Cycle and Financial factors
4
3
2
1
0
−1
−2
−3
−4
−5
0
DMS−FinCycle
DMA−FinCycle
median model−FinCycle
DMS−BusCycle
DMA−BuCycle
median model−BusCycle
CISS
GDP
20
40
60
80
100
120
140
160
180
200
Figure 13: Business Cycle factor volatility
Bus−cycle volatility
0.4
DMS
median model
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
1986
1989
1992
1995
1998
2001
28
2004
2007
2010
2013
Figure 14: Financial factor volatility
Fin−factor volatility
0.7
DMS
median model
0.6
0.5
0.4
0.3
0.2
0.1
0
1986
1989
1992
1995
1998
2001
2004
2007
2010
2013
Figure 15: Covariance between financial and business cycle factors
Covariance between financial and business cycle factors
0.05
DMS
median model
0
−0.05
−0.1
−0.15
−0.2
−0.25
−0.3
1986
1989
1992
1995
1998
2001
29
2004
2007
2010
2013
Figure 16: Measurement equation volatilities
0.8
0.08
0.6
0.4
0.06
2.5
3
0.3
2
2
1.5
0.4
0.04
0
0.02
0
1996 2006
2.5
0.1
1996 2006
3
2
0
1996 2006
1.5
1
1
0
1.5
1996 2006
1.4
1.2
1
0
1996 2006
2
0.8
1.5
0.6
1
0.4
0.5
0.2
1996 2006
0
1996 2006
0
2
4
1
1.5
3
0.8
1
2
0.6
0.5
1
0.4
1996 2006
1996 2006
0.5
0.6
0.4
0
1
0.8
1996 2006
1996 2006
1.5
1
0.5
0
0.5
0.5
1996 2006
0
1
1.5
0.5
0
1996 2006
0.5
2
1
0
0
1
1
0.5
2.5
2
1.5
1.5
0.2
1
0.2
2.5
2
1996 2006
0
1996 2006
0
1996 2006
0.2
1996 2006
0
1996 2006
Figure 17: Financial Cycle factor loadings
Real−loadings
4
2
0
1986
DMS
median model
1989
1992
1995
1998
2001
2004
2007
2010
2013
2007
2010
2013
2007
2010
2013
Real−loadings
4
2
0
1986
DMS
median model
1989
1992
1995
1998
2001
2004
Real−loadings
4
2
0
1986
DMS
median model
1989
1992
1995
1998
2001
30
2004
Figure 18: Financial Cycle factor loadings
2
2
2
4
1
2
0
0
0
−2
−1
1994 2004 2014
0
1994 2004 2014
−2
5
2
2
0
0
0
1994 2004 2014
−2
1994 2004 2014
4
2
0
−5
−2
1994 2004 2014
2
1994 2004 2014
2
−2
1994 2004 2014
−2
2
2
0
0
1994 2004 2014
0
0
−2
−2
−4
1994 2004 2014
5
2
0
0
−5
−2
1994 2004 2014
1994 2004 2014
−2
1994 2004 2014
−2
1994 2004 2014
1994 2004 2014
Figure 19: Root Mean Square Forecast Errors
RMSE relative to AR model
1
0.9
0.8
0.7
0.6
0.5
0.4
19
AR
RW
DMS model
DMA model
Median model
17
15
13
11
9
Months to GDP release
31
7
5
3
1
A
Appendix
We follow the notation and the main results on the matrix differential calculus in Abadir
and Magnus (2005, ch 13)
A.1
Proof of theorem 1
The gradient vector is
0
0
∂`t
1 ∂ log |Ft | ∂vt0 Ft−1 vt
=−
+
∂ft0
2
∂ft0
∂ft0
0
1 1
∂|Ft | ∂vec(Ft ) ∂vt0 Ft−1 vt ∂vt
∂vt0 Ft−1 vt ∂vec(Ft−1 ) ∂vec(Ft )
−
+
+
2 |Ft | ∂vec(Ft )0 ∂ft0
∂vt0
∂ft0 ∂vec(Ft−1 )0 ∂vec(Ft )0
∂ft0
•
•
•
1
− [vec(Ft−1 )0 F t + 2vt0 Ft−1 V t − (vt0 ⊗ vt0 )(Ft−1 ⊗ Ft−1 )F t ]0
2
•0
•0
1 •0
−1
−1
−1
−1
− F t vec(Ft ) − F t (Ft ⊗ Ft )(vt ⊗ vt ) + 2V t Ft vt
2
0
•0
•0
1 •
−1
−1
−1
−1
−1
F t (Ft ⊗ Ft )(vt ⊗ vt ) − F t vec(Ft Ft Ft ) − 2V t Ft vt
2
0
•0
•0
1 •
−1
−1
−1
−1
−1
F t (Ft ⊗ Ft )(vt ⊗ vt ) − F t (Ft ⊗ Ft )vec(Ft ) − 2V t Ft vt
2
•0
1 • 0 −1
−1
−1
F t (Ft ⊗ Ft )[vt ⊗ vt − vec(Ft )] − 2V t Ft vt .
(30)
2
∇t =
=
=
=
=
=
=
We compute the information matrix as the expected value of the Hessian
It = −Et
∂ 2 `t
.
∂ft ∂ft0
(31)
Let re-write the Gradient (30) as follows
∇t =
=
=
=
=
=
•0
1 • 0 −1
−1
0
−1
− F t (Ft ⊗ Ft )[vec(Ft ) − vec(vt vt )] + 2V t Ft vt
2
•0
1 •0
−1
−1
−1
−1
0
−1
− F t [(Ft ⊗ Ft )vec(Ft ) − (Ft ⊗ Ft )vec(vt vt )] − 2V t Ft vt
2
•0
1 •0
0 −1
−1
−1
−1
− F t [vec(Ft ) − vec(Ft vt vt Ft )] + 2V t Ft vt
2
•0
1 •0
−1
−1
−1
0 −1
− F t vec(Ft − Ft vt vt Ft ) + 2V t Ft vt
2
•0
1 •0
−1
0 −1
−1
− F t vec{Ft (I − vt vt Ft )} + 2V t Ft vt
2
•0
1 •0
−1
−1
0 −1
− F t (I ⊗ Ft )vec(I − vt vt Ft ) + 2V t Ft vt .
2
32
(32)
The negative Hessian is equal to
∂∇t
1
∂ 2 `t
=− 0 =
−
0
∂ft ∂ft
∂ft
2
|
•0
∂ F t (I ⊗
Ft−1 )vec(I
−
vt vt0 Ft−1 )
0
•
−1
∂ V t F t vt
+
∂f 0
{z t
}
Φ1t
.
∂ft0
{z
|
(33)
}
Φ2t
We start by computing the first term of (33)
0
•
∂ F t (I⊗Ft−1 )vec(I−vt vt0 Ft−1 )
•0
∂vec(F t )
•0
∂ft0
0
∂vec(F t )
0
•
∂ F t (I⊗Ft−1 )vec(I−vt vt0 Ft−1 )
∂vec(I⊗Ft−1 ) ∂vec(Ft−1 ) ∂vec(Ft )
1
+2
−1 0
−1 0
0
∂ft0
∂vec(I⊗F
)
t
∂vec(Ft ) ∂vec(Ft )
1
2
Φ1t =
•0
∂ F t (I⊗Ft−1 )vec(I−vt vt0 Ft−1 )
∂vec(I−v v 0 F −1 )0
t t t
+ 21
∂ft0
∂vec(I−vt vt0 Ft−1 )0
#
−1
0 −1 0
−
v
= 12 vec(I
t vt Ft ) (I ⊗ Ft ) ⊗ I F t
•0
•
∂vec(I⊗F −1 )
−1
− 12 vec(I − vt vt0 Ft )0 ⊗ F t ∂vec(F −1t )0 (Ft−1 ⊗ Ft−1 )F t
(34)
t
∂vec(vt vt0 Ft−1 )
− 21 F t (I ⊗ Ft−1 )
∂ft0
|
{z
}
•0
Γt
•0
#
where F t =
∂vec(F t )
∂ft0
Γt =
and
∂vec(vt vt0 Ft−1 ) ∂vec(vt vt0 ) ∂vt
∂vec(vt vt0 )0
∂vt0
∂ft0
Ft−1 ⊗ I (vt ⊗ I + I ⊗
=
∂vec(vt vt0 Ft−1 ) ∂vec(Ft−1 ) ∂vec(Ft )
∂ft0
∂vec(Ft−1 )0 ∂vec(Ft )0
•
•
vt ) V t − (I ⊗ vt vt0 ) (Ft−1 ⊗ Ft−1 )F t .
+
(35)
Putting together (34) and (35)we obtain
#
−1
0 −1 0
vec(I
−
v
t vt Ft ) (I ⊗ Ft ) ⊗ I F t
•0
•
∂vec(I⊗F −1 )
1
0 −1 0
= − 2 vec(I − vt vt Ft ) ⊗ F t ∂vec(F −1t )0 (Ft−1 ⊗ Ft−1 )F t
Φ1t =
1
2
(36)
t
•0
− 21 F t (Ft−1 ⊗ Ft−1 ) (vt ⊗ I + I ⊗
•0
•
+ 12 F t Ft−1 ⊗ Ft−1 vt vt0 Ft−1 F t .
•
vt ) V t
The second term of (33) is equal to
Φ2t =
0
•
∂ V t Ft−1 vt
•0
∂vec(V t )0
•0
∂vec(V t )
∂ft0
#
+
0
•
∂ V t Ft−1 vt
∂vec(Ft−1 )
•
∂vec(Ft−1 ) ∂vec(Ft )
∂vec(Ft )0
∂ft0
•
+
•0
0
•
∂ V t Ft−1 vt
∂vt0
•
= (vt0 Ft−1 ⊗ I)V t − (vt ⊗ V t )0 (Ft−1 ⊗ Ft−1 )F t + V t Ft−1 V t ,
33
∂vt
∂ft0
(37)
•0
#
where V t =
∂vec(V t )
.
∂ft0
Putting together (36) and (37) we obtain the following expression for (33)
2
∂ `t
=
− ∂f
0
tf
t
#
vec(I − vt vt0 Ft−1 )0 (I ⊗ Ft−1 ) ⊗ I F t
•0
•
∂vec(I⊗F −1 )
1
0 −1 0
− 2 vec(I − vt vt Ft ) ⊗ F t ∂vec(F −1t )0 (Ft−1 ⊗ Ft−1 )F t
1
2
t
•0
•
− 21 F t (Ft−1 ⊗ Ft−1 ) (vt ⊗ I + I ⊗ vt ) V t
•0
•
+ 12 F t Ft−1 ⊗ Ft−1 vt vt0 Ft−1 F t
(38)
#
+(vt0 Ft−1 ⊗ I)V t
•
•
−(vt ⊗ V t )0 (Ft−1 ⊗ Ft−1 )F t
•0
•
+V t Ft−1 V t .
Following Harvey (1989, p.141), taking the conditional expectation of (??)the only random
#
•
is the prediction error vt and its first and second derivatives are V t and V t . However, such
derivatives are fixed given the conditional expectation at time t − 1. Moreover, we have that
Et (vt ) = 0, Et (vt vt0 ) = Ft and Et [vec(I − vt vt0 Ft−1 )] = 0. Therefore, applying the expectations
Et (.), the fourth and the seventh term in (38) are the only nonzero elements and the information
matrix is equal to
•0
•
•
1 •0
It = F t Ft−1 ⊗ Ft−1 F t + V t Ft−1 V t .
(39)
2
A.2
Proof of theorem 2
Given the model-specific Jacobian matrices
•
Zt =
N m×k
∂vec(Zt )
,
∂ft0
•
Ht =
N 2 ×k
•
∂vec(Ht ) •
∂vec(Tt )
∂vec(Qt )
, Tt =
and Qt =
.
0
0
∂ft
∂ft
∂ft0
m2 ×k
m2 ×k
We show how compute the following Jacobian matrices
•
Vt
=
∂vt
∂ft0
=
∂Zt at
∂fht0
N ×k
∂Zt at ∂vec(Zt )
t at ∂at
+ ∂Z
∂vec(Zt )0
∂ft0
∂a0t ∂ft0
•
•
−[(a0t ⊗ IN )Z t + Zt At ],
= −
=
•
Ft
=
∂vec(Ft )
∂ft0
=
∂vec(Ft ) ∂vec(Zt Pt Zt0 ) ∂vec(Zt )
∂ft0
∂vec(Zt )0
∂ft0
N 2 ×k
+
•
∂vec(Zt Pt Zt0 ) ∂vec(Pt )
+
∂vec(Pt )0
∂ft0
•
•
= (IN 2 + CN 2 )(Zt Pt ⊗ IN )Z t + (Zt ⊗ Zt )P t + H t
•
•
•
= 2NN (Zt Pt ⊗ IN )Z t + (Zt ⊗ Zt )P t + H t
34
(40)
i
•
Ht
(41)
•
Kt
=
mN ×k
=
=
=
∂vec(Kt )
∂ft0
∂vec(Tt Pt Zt0 Ft−1 ) ∂vec(Tt )
∂vec(Tt Pt Zt0 Ft−1 ) ∂vec(Pt )
+
∂vec(Tt )0
∂ft0
∂vec(Pt )0
∂ft0
∂vec(Tt Pt Zt0 Ft−1 ) ∂vec(Zt0 ) ∂vec(Zt )
∂vec(Tt Pt Zt0 Ft−1 ) ∂vec(Ft−1 ) ∂vec(Ft )
+ ∂vec(Z 0 )0
+ ∂vec(F −1 )0 ∂vec(Ft )0 ∂f 0
∂vec(Zt )0
∂ft0
t
t
t
•
•
•
−1
−1
−1
(Ft Zt Pt ⊗ Im )T t + (Ft Zt ⊗ Tt )P t + (Ft ⊗ Tt Pt )CN m Z t −
•
•
•
(Ft−1 Zt Pt ⊗ Im )T t + (Ft−1 Zt ⊗ Tt )P t + (Ft−1 ⊗ Tt Pt )CN m Z t −
•
(IN ⊗ Tt Pt Zt0 )(Ft−1 ⊗ Ft−1 )F t
•
(Ft−1 ⊗ Kt )F t .
(42)
•
At
m×k
=
∂at
∂ft0
=
∂Tt at ∂vec(Tt )
∂vec(Tt )0
∂ft0
•
(a0t ⊗ Im )T t +
=
∂Tt at ∂at
∂a0t ∂ft0
•
Tt At + (vt0
+
+
∂Kt vt ∂vec(Kt )
∂vec(Kt )0
∂ft0
•
•
+
(43)
∂Kt vt ∂vt
∂vt0 ∂ft0
⊗ Im )K t + Kt V t
,
•
Pt
=
∂vec(Pt )
∂ft0
=
∂vec(Tt Pt Tt0 ) ∂vec(Tt )
∂vec(Tt )0
∂ft0
m2 ×k
+
∂vec(Tt Pt Tt0 ) ∂vec(Pt )
∂vec(Kt Ft Kt0 ) ∂vec(Kt )
− ∂vec(K
0
∂vec(Pt )0
∂ft0
∂ft0
t)
•
•
−
∂vec(Kt Ft Kt0 ) ∂vec(Ft )
∂vec(Ft )0
∂ft0
•
•
+ Qt
•
•
= (Im2 + Cm )(Tt Pt ⊗ Im )T t + (Tt ⊗ Tt )P t − (Im2 + Cm )(Kt Ft ⊗ Im )K t − (Kt ⊗ Kt )F t + Qt
•
•
•
•
•
= 2Nm [(Tt Pt ⊗ Im )T t − (Kt Ft ⊗ Im )K t ] + (Tt ⊗ Tt )P t − (Kt ⊗ Kt )F t + Qt
(44)
A.3
Dynamic model averaging and selection
Estimation via model averaging and selection proceeds as follows. We specify a grid of
values for the parameters λs and λv . Each point in this grid defines a new model. Weights for
each model j (defined πt|t−1,j ) are obtained as a function of the predictive density at time t − 1
through the following recursions:
α
πt−1|t−1,j
πt|t−1,j = PJ
α
l=1 πt−1|t−1,l
πt|t−1,j pj (yt |y t−1 )
πt|t,j = PJ
l=1
πt|t−1,l pl (yt |y t−1 )
(45)
(46)
where pj (yt |y t−1 ) is the predictive likelihood. Since this is a function of the prediction errors
and of the prediction errors variance, which are part of the output of the KF, the model weights
can be computed at no cost along with the model parameters estimation. Note that here a new
forgetting factor appears, α, which discounts past predictive likelihoods. We set this parameter
to 0.9. At each point in time, forecast are obtained on the basis of the model with the highest
weight πt|t−1,j or by averaging on the basis of the model weights, see also Koop and Korobilis
(2013).
35
References
Abadir K. and J.R. Magnus (2005). Matrix Albegra. Cambridge University Press.
Alessandri P, and H. Mumtaz, (2014). Financial indicators and density forecasts for US output
and inflation. Temi di discussione (Economic working papers) 977, Bank of Italy.
Antolin-Diaz, J, T Drechsel, I and Petrella (2015). Following the Trend: Tracking GDP when
long-run growth is uncertainaity, CEPR Discussion Papers 10272, C.E.P.R. Discussion
Papers.
Aruoba, S.B., Diebold, F.X., Nalewaik, J. Schorfheide, F. and Song, D. (2016), ”Improving GDP Measurement: A Measurement-Error Perspective. Journal of Econometrics,
forthcoming.
Bańbura M. and M. Modugno (2014). Maximum Likelihood Estimation Of Factor Models
On Datasets With Arbitrary Pattern Of Missing Data. Journal of Applied Econometrics,
29(1),133-160.
Benati L (2007). Drift and breaks in labor productivity. Journal of Economic Dynamics and
Control, Elsevier, 31(8), 2847-2877.
Brave, S. and A. Butters, (2012). Diagnosing the Financial System: Financial Conditions and
Financial Stress. International Journal of Central Banking, vol. 8(2), 191-239.
Chan, J.C.C., G. Koop and S.M. Potter (2013). A New Model of Trend Inflation, Journal of
Business and Economic Statistics, 31(1), 94-106.
Cogley, T. and T.J. Sargent (2005).Drifts and Volatilities: Monetary Policies and Outcomes
in the Post WWII US. Review of Economic Dynamics, 8(2), 262-302.
Cogley, T., G. Primiceri and T.J. Sargent (2010). Inflation-Gap persistence in the US. American Economic Journal: Macroeconomics, 2(1), 43-69
Creal D., Kopman S.J. and A. Lucas (2008). A General Framework for Observation Driven
Time-Varying Parameter Models. Tinbergen Institute Discussion Paper. 108/4
Creal D., Kopman S.J. and A. Lucas (2011). A dynamic Multivariate Heavy-tail Model for
Time-Varying Volatilities and Correlations. Journal of Business and Economic Statistics,
29(4), 552-563.
Creal D, S.J. Kopman and A. Lucas (2013). Generalised Autoregressive Score Models with
Applications. Journal of Applied Econometrics, 28(5), 777–795.
D’Agostino, A., L. Gambetti and D. Giannone (2013). Macroeconomic forecasting and structural change. Journal of Applied Econometrics, 28(1), 82-101.
36
Delle Monache, Davide and Ivan Petrella (2014). Adaptive Models and Heavy Tails. Birkbeck
working paper, BWPEF 1409.
Doz, C., Giannone D. and L. Reichlin, (2012). A Quasi–Maximum Likelihood Approach for
Large, Approximate Dynamic Factor Models. The Review of Economics and Statistics,
94(4), 1014-1024.
Durbin, J. and S.J. Koopman (2001). Time Series Analysis by State Space Methods. Oxford:
Oxford University Press.
Forni, M. Hallin, M, Lippi, M, and L. Reichlin, (2000). The Generalized Dynamic-Factor
Model: Identification And Estimation. The Review of Economics and Statistics, 82(4),
540-554.
Jurado, K., S.C. Ludvigson, and S. Ng. (2015). Measuring Uncertainty. American Economic
Review, 105(3), 1177-1216.
Hakkio, C. S., and W. R. Keeton. 2009. Financial Stress: What Is It, How Can It Be
Measured, and Why Does It Matter? Economic Review (Federal Reserve Bank of Kansas
City) (Second Quarter): 5–50.
Harvey, A.C. (1989). Forecasting, structural time series models and Kalman filter. Cambridge
University Press.
Harvey, A. C., Ruiz, E., and Sentana, E. (1992). Unobserved Component Time Series Models
With ARCH Disturbances. Journal of Econometrics, 52, 129-158.
Harvey, A.C. (2013). Dynamic Models for Volatility and Heavy Tails. With Applications to
Financial and Economic Time Series. Cambridge University Press.
Hatzius J., P. Hooper, F.S. Mishkin, L.K. Schoenholtz, and M.W. Watson (2010). Financial
Conditions Indexes: A Fresh Look after the Financial Crisis. NBER Working Papers
16150.
Holló, D., Kremer, M. and M. Lo Duca, (2012). CISS - a composite indicator of systemic
stress in the financial system. Working Paper Series 1426, European Central Bank.
Koop, G. and D. Korobilis (2010). Bayesian Multivariate Time Series Methods for Empirical
Macroeconomics. Foundations and Trends in Econometrics, 3 (4), 267–358.
Koop, G. and D. Korobilis (2013). Large Time-Varying Parameter VARs. Journal of Econometrics, 177, 185-198.
Koop, G. and D. Korobilis (2014). A New Index of Financial Conditions. European Economic
Review, 71, 101-116.
37
Koopman, S.J, M.I.P. Mallee and M. Van der Wel (2010). Analyzing the Term Structure of
Interest Rates Using the Dynamic Nelson–Siegel Model With Time-Varying Parameters.
Journal of Business & Economic Statistics, 28(3), 329-343.
Illing, M., and Y. Liu (2006). Measuring Financial Stress in a Developed Country: An
Application to Canada. Journal of Financial Stability 2 (3): 243–65.
Matheson, T. D. (2012). Financial conditions indexes for the United States and euro area.
Economics Letters, 115(3), 441-446.
Nalewaik, J. J. (2012), Estimating Probabilities of Recession in Real Time Using GDP and
GDI. Journal of Money, Credit and Banking, 44, 235-253.
Perez-Quiros, G. and McConnell, M. 2000. ”Output Fluctuations in the United States: What
Has Changed since the Early 1980’s?, American Economic Review, American Economic
Association, 90(5),1464-1476.
Perez-Quiros G. and M.M. McConnell (2000). Output Fluctuations in the United States:
What Has Changed since the Early 1980’s? American Economic Review, 90(5), 14641476..
Primiceri, G. (2005). Time Varying Structural Vector Autoregressions and Monetary Policy,
The Review of Economic Studies, 72, 821-852.
Proietti T. (1998). Characterizing Asymmetries in Business Cycle Using Smooth-Transition
Structural Time-Series Models. Studies in Nonlinear Dynamics and Econometrics, 3(3),
141-156.
Sargent, T.J. and C.A. Sims (1977), Business Cycle Modeling Without Pretending to Have
too Much a Priori Theory,” In C.A. Sims (ed.), New Methods in Business Cycle Research:
Proceedings from a Conference, Federal Reserve Bank of Minneapolis, 45-109.
Stock J.H. and M.W. Watson (2002). Forecasting Using Principal Components From a Large
Number of Predictors. Journal of the American Statistical Association, American Statistical Association, 97,1167-1179.
Stock, J. and M.W. Watson (2007). Why has U.S. inflation become harder to forecast?
Journal of Money, Credit and Banking 39, supplement F, 3-33.
Stock J.H. and M.W. Watson (2010). Dynamic Factor Models. Handbook of Economic Forecasting, M. Clements and D. Hendry (eds.), Oxford University Press.
38