Issues in Connectedness Measurement for Large Networks Francis X. Diebold April 4, 2014 1/9 Diebold-Yilmaz Connectedness via Variance Decomposition Connectedness Table x2 ... xN x1 x1 x2 .. . H d11 H d21 .. . H d12 H d22 .. . ··· ··· .. . H d1N H d2N .. . xN H dN1 H dN2 ··· H dNN H i6=2 di2 ··· To Others H i6=1 di1 P H = Ci←• P PN H = C•←j CH = 1 N PN j=1 j6=i i6=j i6=N H diN P i6=j dijH dijH , are the “from degrees” PN i,j=1 P From Others P dH Pj6=1 1j H j6=2 d2j .. P . H j6=N dNj i=1 i6=j dijH , are the “to degrees” dijH , is the mean degree (to or from) 2/9 Static Pairwise Connectedness Country Banking Systems, 2004-14, Adaptive Elastic Net 3/9 Dynamic Total Connectedness Country Banking Systems, 2004-14, Adaptive Elastic Net 100 90 80 70 60 50 04 05 06 07 08 09 10 11 12 13 4/9 Network Analysis with a VAR Approximating Model yt = Ayt−1 + εt εt ∼ iid(0, Σ) I Diebold-Yilmaz variance decomposition-based (V ). Based on variance decomposition matrix D = f (A). What fraction of the H-step-ahead prediction-error variance of variable i is due to shocks in variable j? I Brownlees et al. coefficient-based (C ). Based directly on A. How is yi,t connected to yj,t−1 ? 5/9 V vs. C Issues I V is readily calculated for any approximating model, not just AR(1). Not obvious how to do C . I V requires an identifying assumption, whereas C does not. But C just skirts the issue by ignoring Σ. I V facilitates examination of multi-step connectedness via H, whereas C is tied to 1-step. (We could also do IRF’s.) I V cleanly breaks down connectedness from granular to aggregate, with pieces adding to 100%. Not clear for C . I Sparse A estimation methods do not force a sparse D, since D = f (A) for nonlinear f (·). Sparse A estimation methods do force a sparse A (by construction). Hence V is more appealing than C . 6/9 V and C Issues I Recovering degrees of freedom. Selection (e.g., forward stepwise), shrinkage (e.g., ridge), or both (e.g., lasso). I System estimation methods vs. equation-by-equation I Time-variation: Smooth evolution (e.g. rolling) vs. sharp breaks (e.g. fused ridge, fused lasso). Bandwidth selection. I Many flavors of lasso (e.g., lasso, adaptive lasso, elastic net, adaptive elastic net) 7/9 Lasso I: Penalized Estimation β̂ = arg min β T X !2 yt − t=1 X βi xit s.t. i K X (|βi |q )1/q ≤ c i=1 Equivalently, β̂ = arg min β T X t=1 !2 yt − X i βi xit K X +λ (|βi |q )1/q i=1 Concave penalty functions non-differentiable at the origin produce selection. Smooth convex penalties produce shrinkage. q → 0 produces selection, q = 2 produces ridge, q = 1 produces lasso. 8/9 Lasso II: Flavors of Lasso β̂Lass !2 T K X X X = arg min yt − βi xit +λ |βi | β t=1 β̂ALass = arg min β β̂Enet i i=1 !2 T X yt − X t=1 βi xit +λ i K X wi |βi | i=1 where wi = 1/β̂iν , β̂i is OLS or ridge, and ν > 0 !2 T K X X X = arg min yt − βi xit +λ α|βi | + (1 − α)βi2 β β̂AEnet = arg min β t=1 i T X X t=1 i=1 !2 yt − i βi xit +λ K X 2 αwi |βi | + (1 − α)βi i=1 where wi = 1/β̂iν , β̂i is OLS or ridge, and ν > 0. 9/9
© Copyright 2026 Paperzz