A Complete Characterization of a Class of Robust Linear Average

2010 American Control Conference
Marriott Waterfront, Baltimore, MD, USA
June 30-July 02, 2010
ThB02.3
A Complete Characterization of a Class of
Robust Linear Average Consensus Protocols
Randy A. Freeman, Thomas R. Nelson, and Kevin M. Lynch
Abstract— We provide a set of verifiable necessary and sufficient conditions for a member of a broad class of linear discretetime consensus protocols to guarantee the robust convergence
of the outputs of the networked dynamics to the correct average
consensus value.
I. I NTRODUCTION
We consider the average consensus problem: each member
of a network of agents has some assigned input value, and it
must calculate the average of the inputs of all agents. We seek
solutions which are decentralized (agents can only communicate with their immediate neighbors), scalable (each agent’s
memory, computation, and communication requirements are
independent of the total number of agents in the network),
asymptotically correct (if the inputs and the network are both
constant, then each agent’s estimate of the group average
converges to the right value with zero steady-state error), and
robust (temporary faults in communication or computation
have no bearing on the final steady-state result).
There has been much work on this type of problem in
the past several years, and we refer the reader to [1], [2],
[3], [4], [5] for details. In many cases, the agent inputs are
assigned as initial states in a networked dynamic system,
and as time progresses these states converge to the average
consensus value. However, such solutions are inherently nonrobust: a fault which causes a single incorrect state update
at a single time instant can cause the network to converge to
the incorrect consensus value. To avoid this problem, we
consider dynamic consensus schemes in which the agent
inputs are assigned not as initial states but as actual inputs
into the networked dynamic system [6], [7]. For robustness
in this case, the final value of the dynamic state should be the
correct average consensus value (at least for constant inputs
and constant networks), regardless of the value of the initial
state. In this paper we provide a complete characterization
of a class of linear discrete-time protocols which possess
this robustness property. The characterization of nonlinear
schemes, such as the consensus propagation scheme of [8],
is a topic of future research.
We begin in Section II with a review of relevant results
from graph theory, we present some results on a particularly
useful decentralized graph weighting scheme in Section III,
and we present our main result as Theorem 5 in Section IV.
This work was supported in part by NSF grant ECS-0601661 and by
the Office of Naval Research. The authors are with the Department of
Electrical Engineering and Computer Science (Freeman and Nelson) and the
Department of Mechanical Engineering (Lynch), Northwestern University,
Evanston, IL 60208, USA, [email protected],
[email protected],
[email protected].
978-1-4244-7425-7/10/$26.00 ©2010 AACC
II. L APLACIANS OF WEIGHTED DIGRAPHS
For each finite set S, we let S S denote the product S×S
minus its diagonal, so that (s,t) ∈ S S when s,t ∈ S and
s 6= t. For K = R or K = C, we let K S denote the free Kvector space over S, with the canonical basis given by the
indicator functions es of elements s ∈ S (where es (t) = 1
when s = t and es (t) = 0 when s 6= t). We endow K S with
the inner product hx, yi , ∑s∈S x(s) y∗(s) (where ∗ denotes
the conjugate transpose) so that the canonical basis is orthonormal. We let 0, 1 ∈ K S denote the constant functions
on S with values 0 and 1, respectively. We regard RS as
an ordered vector space by adopting the product order: for
x, y ∈ RS , we say x 6 y when x(s) 6 y(s) for all s ∈ S. Thus
the positive cone of RS is the set of mappings from S to the
nonnegative reals. We let L (K S ) denote the space of linear
maps from K S to K S , and we let Mx = M(x) ∈ K S denote the
value of a mapping M ∈ L (K S ) at a point x ∈ K S . We let I
denote both the identity map I ∈ L (K S ) and the identity
matrix, and we adopt the convention that M 0 = I for any
mapping M ∈ L (K S ) (or any square matrix M), including
the zero mapping. Finally, we let K ∅ denote the singleton
vector space of dimension zero.
A vertex set V is any nonempty finite set, and we let U(V)
and W(V) denote the sets of unweighted and positively
weighted simple digraphs on V, respectively. Specifically, we
identify U(V) with the power set of V V, and we say that
an unweighted digraph G ∈ U(V) has an arc from vertex a to
vertex b when (a, b) ∈ G. Also, we identify W(V) with the
positive cone of the vector space RVV , and the arc set of
a weighted digraph G ∈ W(V) is the underlying unweighted
digraph Arc(G) ∈ U(V) given by the support of G,
namely,
Arc(G) , (a, b) ∈ V V : G(a, b) > 0 .
(1)
Here G(a, b) represents the weight on an arc (a, b) ∈ Arc(G).
We regard the mapping Arc(·) as a surjection from W(V)
to U(V), and we refer to any right inverse of this surjection
as a weighting scheme on V.
in (a) and
Given a digraph G ∈ U(V) or G ∈ W(V), we let NG
out
NG (a) respectively denote the sets of in- and out-neighbors
of a vertex a ∈V (not including a itself). The unweighted inout
and out-degrees dGin (a) and
dG (a)outare
out in
in
dG (a) , NG (a) ,
dG (a) , NG
(a) .
(2)
We let Src(G) ⊂ V and Sink(G) ⊂ V denote the sets of
sources and sinks of G, respectively: a ∈ Src(G) when
dGin (a) = 0 and a ∈ Sink(G) when dGout (a) = 0. A weighted
digraph G ∈ W(V) has weighted in- and out-degrees win
G (a)
and wout
G (a) given by
win
wout
(3)
G (a) , ∑ G(b, a) ,
G (a) , ∑ G(a, b) .
b6=a
b6=a
Using terminology from [4], [9], we say that G ∈ W(V) is
3198
out
balanced when win
G (a) = wG (a) for all a ∈V and symmetric
when G(a, b) = G(b, a) for all (a, b) ∈ V V. All symmetric
weighted digraphs are balanced, and if G is balanced then
Src(G) = Sink(G). In what follows, the term “digraph” will
refer to a weighted digraph unless otherwise indicated.
We define the Laplacian of a digraph G ∈ W(V) to be the
V
linear map
∆G ∈ L (R ) given by
∆Gx (a) = (∆G)(x) (a) , ∑ G(a, b) x(a) − x(b) (4)
b6=a
for each x ∈ RV and a ∈ V. The map ∆ : W(V) → L (RV )
is injective, which means a digraph is characterized by its
Laplacian. We regard ∆G as a member of L (CV ) when
discussing its eigenvalues and eigenvectors, and we let
σ (G) ⊂ C denote the spectrum of ∆G. One can show that
V
the adjoint ∆∗ G
∈ L (R ) of the Laplacian is given by
∆∗Gx (a) = ∑ G(a, b)x(a) − G(b, a)x(b)
(5)
b6=a
for each x ∈ RV and a ∈ V, and it is clear that a digraph
is symmetric if and only if its Laplacian is self-adjoint.
We say that a digraph G is diagonalizable when ∆G is
diagonalizable, that is, when there is a basis of CV consisting
of eigenvectors of ∆G. Clearly all symmetric digraphs are
diagonalizable.
It follows from (4) and (5) that ∆G1 = 0 for every digraph
G ∈ W(V), and moreover ∆∗ G1 = 0 if and only if G is
balanced. Therefore rank(∆G) 6 |V| − 1 for every digraph
G ∈ W(V), and we say that G has maximal rank when
rank(∆G) = |V| − 1. To characterize the rank of ∆G, we need
the concept of a quotient digraph. Every partition P of the
vertex set V induces a quotient digraph GP ∈ W(P), where
GP (A, B) , ∑ G(a, b)
(6)
(a,b) ∈ A×B
for all A, B ∈ P P. It is clear from flow conservation that
any quotient of a balanced digraph is itself balanced. Recall
that a digraph is weakly connected when any distinct pair
of its vertices lie on some undirected path and strongly
connected when any distinct pair of its vertices lie on some
directed cycle. The acyclic quotient of G is the quotient Gaq
induced by the partition of V into its strongly connected
components. The acyclic quotient of a balanced digraph is
totally disconnected (every vertex is both a source and a
sink). Hence a balanced digraph is strongly connected if
and only if it is weakly connected, which means the phrase
“balanced and connected digraph” is unambiguous. It is
shown in [10] that
rank(∆G) + Sink(Gaq ) = |V| .
(7)
Equation (7) is a direct consequence of an oft-rediscovered
theorem on determinants [11]. Thus a digraph has maximal
rank if and only if its acyclic quotient has a single sink.
In particular, every strongly connected digraph has maximal
rank, which was proved independently as Theorem 1 in [4].
Also, it is clear that if a digraph is balanced, then it has
maximal rank if and only if it is connected.
For each vertex set V, the set of balanced digraphs
in W(V), the set of symmetric digraphs in W(V), and the set
of maximal-rank digraphs in W(V) are all convex. However,
if |V| > 3, then the set of balanced, connected, diagonalizable
digraphs in W(V) is not convex.
If we order the elements of a vertex set V, then we
can label the vertices as V = {1, . . . , n} for n = |V| and
regard the canonical basis for RV as an ordered basis. The
Laplacian of each digraph G ∈ W(V) has the following
matrix representation L ∈ Rn×n with respect to this ordered
canonical basis:
(
wout
if i = j
G (i)
Li j =
(8)
−G(i, j) if i 6= j .
We will use this representation (8) in some of our proofs.
III. I NVERSE OUT- DEGREE WEIGHTS
For each x ∈ C, we let Dx ⊂ C denote the closed unit disc
centered at x, and we note that rDx for r ∈ C is the closed
disc centered at rx with radius |r|.
Consider a complete digraph G ∈ W(V) with equal weights
(there is a constant c > 0 such that G(a, b) = c for all a 6= b).
Then the Laplacian of G has a single eigenvalue at zero, and
all other eigenvalues lie at c|V|. Thus without knowledge
of an upper bound on the digraph order |V|, there is no
way to choose c so that all digraphs whose weights are
all equal to c have Laplacians with eigenvalues guaranteed
to lie in some fixed disc rDx . Because the stability region
for discrete-time, linear, time-invariant systems is bounded
(being the interior of D0 ), it will be difficult to design
discrete-time algorithms involving equal-weight Laplacians
which are stable for digraphs of arbitrary order. Instead, we
seek weighting schemes which guarantee that all Laplacian
eigenvalues lie in a known bounded region, regardless of
the digraph order. Furthermore, to be useful for decentralized algorithms, the resulting weights should be computable
using information only from immediate neighbors. We now
propose such a weighting scheme which is similar to the
local-degree weighting scheme discussed in [3].
Given an unweighted digraph G ∈ U(V) on a vertex set V,
the associated inverse out-degree digraph IOD(G) ∈ W(V) is
given by the weights
(
1
when (a, b) ∈ G
out
out
(9)
IOD (G)(a, b) = dG (a) + dG (b)
0
when (a, b) 6∈ G .
This mapping IOD(·) is a right inverse of Arc(·) and is thus
a weighting scheme. Its image IOD(U(V)) is the set of all
digraphs on V having IOD weights.
Theorem 1: If a digraph G ∈ W(V) belongs to the convex
hull of IOD(U(V)), then σ (G) ⊂ D1 . If G itself belongs to
IOD ( U (V)), then σ (G) ⊂ D0 ∩ D1 .
The proof (omitted) follows the main idea in [12]. Note
that there exist digraphs, even symmetric ones, which belong
to the convex hull of IOD(U(V)) but whose Laplacians have
eigenvalues outside of D0 . Also, we conjecture that the
eigenvalue estimates in Theorem 1 can be improved:
Conjecture 2: If G ∈ IOD(U(V)) then σ (G) ⊂ 21 D1 .
If this conjecture is true, then 12 D1 is the smallest convex
region in the complex plane containing all eigenvalues of all
digraphs having IOD weights. Indeed, let G ∈ U(V) consist
of a single directed cycle with n = |V| vertices. Then it is
straightforward to show that the Laplacian of IOD(G) has
eigenvalues at 21 − 12 e2π jk/n for 1 6 k 6 n, which are equally
3199
spaced around the boundary of 12 D1 .
We next show that IOD weights are particularly useful
when digraphs are drawn from a “symmetric” probability
distribution. Let V be a vertex set, and let p : U(V) → [0, 1]
be a probability distribution (mass function) on the set U(V)
of unweighted digraphs. The probability digraph Π p ∈ W(V)
associated with p is given by the weights
Π p (a, b) = Prob (a, b) ∈ G = ∑ χG (a, b) p(G) , (10)
G∈ U(V)
where χG denotes the indicator function of G. We say
that
p is an
arc-independent distribution when the events
(a, b) ∈ G in (10) are mutually independent for distinct
ordered pairs of vertices (a, b); in this case the probability
digraph Π p completely characterizes the distribution p.
For a pair of unweighted digraphs G1 , G2 ∈ U(V), we write
G1 ↔ G2 whenthere exist
vertices a, b ∈ V such that
G1 \ G2 = (a, b)
and G2 \ G1 = (b, a) .
(11)
Thus G1 ↔ G2 when G1 and G2 have the same arcs except
for a single pair of vertices between which G1 and G2 have
arcs in opposite directions. We say that p is an arc-symmetric
distribution when G1 ↔ G2 implies p(G1 ) = p(G2 ). If p is
arc-symmetric, then the associated probability digraph Π p
will be a symmetric digraph, but the converse is not true
in general. However, if p is arc-independent, then the arcsymmetry of p is equivalent to the symmetry of Π p .
Let W : U(V) → W(V) be a weighting scheme on V, and
let p be a probability distribution over U(V). The expected
digraph E p [W] ∈ W(V) for W and p is given
by the weights
E p [W](a, b) = ∑ W(G)(a, b) p(G) .
(12)
G∈ U(V)
We note that for the unity equal weighting scheme in which
W(G)(·, ·) = χG (·, ·) for each G, the expected digraph E p [W]
coincides with the probability digraph Π p .
Theorem 3: If the distribution p is arc-symmetric, then
the expected digraph E p [IOD] is symmetric.
The proof of Theorem 3 (omitted) relies on the fact
that the IOD weights in (9) are a function of the sum of
the neighboring out-degrees. Indeed, if we were instead to
use some other weighting scheme, such as the local-degree
weighting scheme
( in [3] given by
1
when (a, b) ∈ G
out
out
(13)
W(G)(a, b) = max{dG (a), dG (b)}
0
when (a, b) 6∈ G ,
then the arc-symmetry of the distribution p need not guarantee that the expected digraph is symmetric or even balanced,
even when p is arc-independent.
IV. ROBUST AVERAGE CONSENSUS
A. Convergent matrices
A square complex matrix M is convergent when its power
sequence {M k }∞
k=1 converges to a finite constant matrix as
k → ∞.1 It is clear from the Jordan decomposition that M is
convergent if and only if all Jordan blocks associated with an
eigenvalue at λ = 1 are of size one, and all other eigenvalues
lie in the interior of D0 . Equivalently, M is convergent if and
only if it is similar to a block diagonal matrix diag(N, I) for
some square matrix N with ρ(N) < 1, where ρ(·) denotes
the spectral radius (neither N nor I need be present).
Lemma 4: Suppose M ∈ Cn×n and b ∈ Cn are given, and
consider the discrete-time dynamics
x(k + 1) = Mx(k) + b
(14)
with x(k) ∈ Cn . If M is convergent and there exists an equilibrium state x̄ = M x̄ + b, then for every initial state x(0), the
trajectory x(k) converges exponentially to some equilibrium
as k → ∞. If M is convergent and there does not exist an
equilibrium state, then the trajectory diverges from every
initial state, that is, |x(k)| → ∞ as k → ∞. Finally, if M is
not convergent, then there is some initial state from which
the trajectory does not converge.
B. Average consensus
Let G ∈ W(V) be a digraph on a vertex set V, and suppose
each vertex a ∈ V is an agent which receives a scalar input
signal ua (k) ∈ R at each discrete time step k and passes it
through a filter to produce a scalar output signal ya (k) ∈ R.
In computing its output ya (k), agent a also uses information
it receives from its out-neighbors at the current time step (so
that information flows against the arcs of G). The goal is
for each ya (k) to be an estimate of the average of all agent
inputs at time step k. We let u(k), y(k) ∈ RV be the vectors
of these inputs and outputs so that u(k)(a) = ua (k) and
y(k)(a) = ya (k) for each a ∈ V, and we write the estimation
error e(k) ∈ RV as
1 u(k), 1 · 1 .
(15)
e(k) = y(k) −
|V|
If the inputs ua (k) are changing rapidly with k and if the
diameter of G is large, then we cannot hope to keep the
error small by communicating only with neighbors (by the
time information about a particular input reaches a distant
agent, such information is no longer relevant). Thus we can
achieve our goal of keeping the error small only when the
inputs are varying slowly relative to the diameter of G.
We seek scalable filters for calculating the local estimates ya (k): the memory, computation, and communication
requirements for each agent should depend only on its outdegree, not on the total number of agents in the network. A
well-studied example of such a filter is given by the linear
update rule
ya (k + 1) = ya (k) − ∑ G(a, b) ya (k) − yb (k) .
(16)
b6=a
It is well known that if G is balanced, connected, and such
that I − ∆G is convergent, then y(k) converges to the vector
1 y(0), 1 · 1
(17)
ȳ =
|V|
as k → ∞ [4]. Thus if we have a constant input vector
u(k) ≡ ū and we choose the initial estimate y(0) = ū, then the
estimation error (15) will converge to zero as k → ∞. Unfortunately, this filter is not robust to errors in the calculation
of the update rule (16): if just one of the components ya (k)
is updated incorrectly at just one time instant, then the entire
estimate y(k) will converge to the wrong value. We can
characterize this lack of robustness in a different way by
examining a state-space version of the
filter (16):
xa (k + 1) = xa (k) − ∑ G(a, b) ya (k) − yb (k)
(18)
also refer to linear mappings in L (·) as convergent when their matrix
representations are convergent.
1 We
3200
b6=a
ya (k) = xa (k) + ua (k) ,
(19)
where xa (k) ∈ R is a scalar filter state variable for agent a.
One can show that, under a constant input u(k) ≡ ū and
with G as above, the output y(k) converges to
1 ȳ =
x(0) + ū, 1 · 1
(20)
|V|
as k → ∞. Here
the dependence
of ȳ on the initial state x(0)
is apparent: if x(0), 1 6= 0, then the estimate y(k) will converge to the wrong value. Such incorrect state initializations
might come from errors in inter-agent communication, the
addition or removal of agents, or incorrect local calculations.
In this paper, we seek alternative linear filters for which
the estimation error (15) converges to zero regardless of the
value of the initial filter state x(0).
C. Polynomial linear protocols
If we let x(k) ∈ RV denote the vector of filter states in the
update rule (18), so that x(k)(a) = xa (k) for each a ∈ V, then
we can combine the filter equations (18)–(19) for each agent
into the global rule
x(k + 1) = [I − ∆G]x(k) − ∆Gu(k)
(21)
y(k) = x(k) + u(k) .
(22)
We can generalize the structure of this filter by allowing the
internal state xa (k) of each agent to be a vector rather than a
scalar, and by allowing the coefficient matrices in the statespace description to be general polynomial functions of the
digraph Laplacian. Specifically, a polynomial linear protocol
is a collection Σ = [A(X), B(X),C(X), D(X)], where
`
`
A(X) , ∑ Ai X i
B(X) , ∑ Bi X i
(23)
C(X) , ∑ Ci X i
D(X) , ∑ Di X i
(24)
i=0
`
i=0
i=0
`
i=0
are polynomials in the formal symbol X with matrix coefficients Ai ∈ R p×p , Bi ∈ R p×q , Ci ∈ Rm×p , and Di ∈ Rm×q . Here
` > 0 denotes the degree of Σ, and p > 1, q > 1, and m > 1
denote its state, input, and output dimensions, respectively
(and we refer to p as the dimension of the protocol). We
regard such a system Σ as a protocol or a template as it does
not by itself constitute a dynamic system. Rather, given a
linear map M ∈ L (RS ) for some finite set S, the protocol Σ
generates the discrete-time linear system Σ(M) given by
x(k + 1) = A(M)x(k) + B(M)u(k)
(25)
y(k) = C(M)x(k) + D(M)u(k) ,
(26)
with state x(k) ∈ R p ⊗ RS , input u(k) ∈ Rq ⊗ RS , and output
y(k) ∈ Rm ⊗ RS , where
`
`
A(M) , ∑ Ai ⊗ M i
B(M) , ∑ Bi ⊗ M i
(27)
C(M) , ∑ Ci ⊗ M i
D(M) , ∑ Di ⊗ M i
(28)
i=0
`
i=0
i=0
`
i=0
(here ⊗ denotes the tensor product). In this manner, a
protocol Σ will generate the dynamics Σ(∆G) for any digraph
G ∈ W(V) on a vertex set V. For example, the dynamics
in (21)–(22) are generated by the SISO protocol of degree
` = 1 and dimension p = 1 given by the matrix coefficients
A0 = C0 = D0 = 1, A1 = B1 = −1, and B0 = C1 = D1 = 0.
Given a digraph G ∈ W(V), we can implement the dynamics Σ(∆G) using a local filter on each agent in which the
agent receives information from its out-neighbors. Indeed,
we can write x(k), u(k), and y(k) uniquely as
x(k) = ∑ xa (k) ⊗ ea ,
xa (k) ∈ R p ,
(29)
a∈V
u(k) =
∑ ua (k) ⊗ ea ,
ua (k) ∈ Rq ,
(30)
ya (k) ∈ Rm ,
(31)
a∈V
y(k) =
∑ ya (k) ⊗ ea ,
a∈V
where {ea }a∈V are the canonical basis vectors for RV . Then
each agent a ∈ V implements
the filter
xa (k)
a
z0 (k) =
∈ R p+q
(32)
ua (k)
zai (k) = ∑ G(a, b) zai−1 (k) − zbi−1 (k) for 1 6 i 6 ` (33)
b6=a
` xa (k + 1)
A Bi a
=∑ i
z (k) .
(34)
ya (k)
Ci Di i
i=0
It is straightforward to show that the combination of these
local filters (32)–(34) results in the global dynamics (25)–
(26) with M = ∆G.
We see from (33) that each time step k requires ` stages
of inter-agent communication: first, the za0 variables are
communicated to obtain the za1 variables, which are then
communicated to obtain the za2 variables, and so on. Thus at
each time step, each agent transmits at most `(p + q) scalar
values to its in-neighbors (the actual number could be smaller
if any of the Ai , Bi , Ci , or Di are zero). The same information
goes to each in-neighbor, which means the transmission is
possible using a simple local broadcast.
D. Robust average consensus protocols
Let Σ = [A(X), B(X),C(X), D(X)] be a SISO polynomial
linear protocol with dimension p > 1, and let G ∈ W(V) be
a digraph on a vertex set V. Suppose ū ∈ RV is a constant
vector of agent inputs, and suppose x̄ ∈ R p ⊗ RV and ȳ ∈ RV
are an equilibrium state and output (respectively) for the
dynamics Σ(∆G) under
this constant
input:
I − A(∆G) x̄ = B(∆G)ū
(35)
ȳ = C(∆G)x̄ + D(∆G)ū .
(36)
We seek protocols for which, under appropriate assumptions
on the digraph G, the state always converges to such an
equilibrium value x̄ and the only possible corresponding
value for ȳ is the average consensus value
1 ū, 1 · 1 .
(37)
ȳ =
|V|
Definition 1: Let G ⊂ W(V) be a family of digraphs on a
vertex set V. A SISO polynomial linear protocol Σ achieves
robust average consensus over G when for every G ∈ G, the
dynamics Σ(∆G) are such that for any initial state x(0) and
any constant input u(k) ≡ ū, the state x(k) converges to a
constant and the output y(k) converges to ȳ in (37) as k → ∞.
Before we state our main result on polynomial linear
protocols, we first introduce some notation. If we evaluate
the formal polynomials in (23)–(24) at a complex scalar
X = µ ∈ C, then A(µ), B(µ), C(µ), and D(µ) become
complex matrices of appropriate dimensions. Given these
complex matrices, we define the complex
scalar
+
H(µ) , C(µ) I − A(µ) B(µ) + D(µ) ,
(38)
where (·)+ denotes the Moore-Penrose pseudoinverse.
3201
Theorem 5: Let Σ = [A(X), B(X),C(X), D(X)] be a SISO
polynomial linear protocol, and let G ⊂ W(V) be a nonempty
collection of diagonalizable digraphs on a vertex set V with
|V| > 2. Then Σ achieves robust average consensus over G if
and only if all of the following are true:
(i) A(µ) is convergent for all µ ∈ σ (G).
(ii) B(µ) ∈ Col I − A(µ) for all µ ∈ σ (G),
(iii) C∗(µ) ∈ Col I − A∗(µ) for all µ ∈ σ (G),
(iv) H(0) = 1, where H(·) is from (38),
(v) H(µ) = 0 for all nonzero µ ∈ σ (G), and
(vi) each digraph in G is balanced and connected.
Using this theorem as a sufficient condition on a protocol Σ requires knowledge of a region in the complex plane
containing the spectrum σ (G) of each G ∈ G. For example,
suppose G is the collection of all balanced, connected, and
diagonalizable digraphs in IOD(U(V)); then we can conclude
from Theorem 1 that Σ achieves robust average consensus
over G if conditions (i)–(v) hold for all µ ∈ D0 ∩ D1 .
As we will see in the proof of Theorem 5, condition (i)
is equivalent to A(∆G) being convergent for all G ∈ G,
which from Lemma 4 is necessary and sufficient for the
dynamics Σ(∆G) to converge to an equilibrium from any
initial state and for any constant input vector (provided an
equilibrium exists). Conditions (ii) and (iii) are equivalent
to the statement that, for each µ ∈ σ (G), any eigenvalue
of A(µ) at λ = 1 is both uncontrollable through B(µ) and
unobservable through C(µ).
+ Conditions (iv) and (v) imply
that the matrix I − A(µ) is discontinuous as a function
of µ at µ = 0, which can only happen when I − A0 is
singular, namely, when A0 has an eigenvalue at λ = 1;
hence this eigenvalue must be uncontrollable through B0
and unobservable through C0 . This is clearly violated by the
protocol (21)–(22) for which A0 = C0 = 1. In fact, there is no
linear consensus protocol having degree ` = 1 and dimension
p = 1 which can achieve robust average consensus. Indeed,
if there were such a protocol, then necessarily A0 = D0 = 1
and B0 = C0 = 0. If also A1 = 0, then A(µ) = 1 for all µ,
and we conclude from (v) that D0 + µD1 = 0 for µ > 0,
which implies D0 = 0, a contradiction. Therefore A1 6= 0,
+
which means 1 − A(µ) = −(µA1 )−1 for µ > 0 and thus
H(µ) = D0 + µ(D1 −C1 B1 /A1 ) for µ > 0. But now (v) again
implies D0 = 0, a contradiction.
Fortunately, there do exist linear consensus protocols of
higher degree or higher dimension that achieve robust average consensus. Consider the protocol of degree ` = 1 and
dimension
p = 2 given by 1−γ 0
−k p ki
γ
A0 =
, A1 =
, B0 =
,
(39)
0
1
−ki 0
0
C0 = 1 0 ,
B1 = C1T = 0 ,
D0 = D1 = 0 , (40)
where γ, k p , and ki are positive gains. This is a discrete-time
version of the PI estimator studied in [7]. If we choose the
gains so that the matrix A(µ) = A0 + µA1 is convergent for
all µ ∈ σ (G), then it is straightforward to verify that this
protocol satisfies conditions (i)–(v) in Theorem 5. Another
example is the linear consensus protocol of degree ` = 2 and
dimension p = 1 given by
A0 = C1 = D0 = 1 ,
A2 = B1 = −1 ,
(41)
A1 = B0 = B2 = C0 = C2 = D1 = D2 = 0 ,
(42)
which yields A(µ) = 1 − µ 2 , B(µ) = −µ, C(µ) = µ, and
D(µ) = 1. If G is such that µ 2 ∈ D1 for all µ ∈ σ (G), then
this protocol satisfies conditions (i)–(v) in Theorem 5.
Although we state and prove Theorem 5 only for the case
of diagonalizable digraphs, we conjecture that the results
hold also for the non-diagonalizable case.
E. Proof of Theorem 5
We begin with two lemmas about Kronecker products;
their straightforward proofs are omitted.
Lemma 6: Given n matrices A1 , . . . , An ∈ C pi ×qi and n
vectors b1 , . . . , bn ∈ C pi , the equation
A1 ⊗ · · · ⊗ An x = b1 ⊗ · · · ⊗ bn
(43)
admits a solution x ∈ Cq , where q = ∏i qi , if and only if
either (i) bi = 0 for some i, or (ii) bi ∈ Col(Ai ) for each i.
Lemma 7: Let b1 , . . . , bn ∈ Cm be an independent collection of vectors, and let A1 , . . . , An ∈ C p×q each have rank q.
Then the matrix
A , A1 ⊗ b1 . . . An ⊗ bn ∈ Cmp×nq
(44)
has rank nq.
Let n = |V| > 2, suppose G ∈ G, and let L ∈ Rn×n denote
the matrix representation (8) of the Laplacian ∆G with respect to some order on V. Because L is diagonalizable (by assumption), it has n independent eigenvectors w1 , . . . , wn ∈ Cn
with corresponding eigenvalues λ1 , . . . , λn ∈ C, and we define
the nonsingular matrix W , [w1 . . . wn ]. For j ∈ {1, . . . , n},
let J j ∈ C p×p be a Jordan form of A(λ j ) with A(λ j ) =
Pj J j Pj−1 for some invertible matrix Pj ∈ C p×p . The matrix
J , diag(J1 , . . . , Jn ) is a Jordan form of A(L); indeed, the
np × np complex matrix
P , P1 ⊗ w1 . . . Pn ⊗ wn ,
(45)
which is invertible by Lemma 7, is such that A(L) = PJP−1 .
We conclude that A(L) is convergent if and only if A(λ j )
is convergent for each j, and it follows from Lemma 4
that condition (i) of Theorem 5 is necessary and sufficient
for the dynamics Σ(∆G) to converge under constant inputs
from every initial state and for every G ∈ G (provided these
dynamics admit an equilibrium state).
We define the polynomial F(X) , I − A(X), and for each
j ∈ {1, . . . , n} we let F(λ j ) = Φ j S j Ψ∗j be a singular value decomposition so that S j ∈ R p×p is diagonal with nonnegative
real entries listed in descending order and Φ∗j Φj = Ψ∗j Ψj = I.
We define the np ×np complex matrices Φ , Φ1 ⊗ w1 . . . Φn ⊗ wn
(46)
Ψ , Ψ1 ⊗ w1 . . . Ψn ⊗ wn ,
(47)
which are both invertible by Lemma 7. Next we define the
matrix Γ , [WW ∗ ]−1 so that W ∗ ΓW = I, and we observe that
∗
Φ∗ (I ⊗ Γ)Φ =
Ψ (I ⊗ Γ)Ψ = I. We next computeF(L)Ψ:
F(L)Ψ = F(λ1 )Ψ1 ⊗ w1 . . . F(λn )Ψn ⊗ wn
= Φ1 S1 ⊗ w1 . . . Φn Sn ⊗ wn
= (Φ1 ⊗ w1 )S1 . . . (Φn ⊗ wn )Sn = ΦS , (48)
3202
where S is the block diagonal matrix S , diag(S1 , . . . , Sn ).
We introduce the inner product h·, ·iΓ on Cnp by defining
hx, yiΓ , y∗ (I ⊗ Γ)x
(49)
np
for all x, y ∈ C . Thus Φ and Ψ have orthonormal columns
with respect to the inner product (49), and from (48) we have
F(L) = ΦSΨ−1 = ΦSΨ∗(I ⊗ Γ) .
(50)
For each j ∈ {1, . . . , n}, we label the columns of Φj and Ψj
as φ1 j , . . . , φ p j ∈ C p and ψ1 j , . . . , ψ p j ∈ C p (respectively), and
we write S j = diag(s1 j , . . . , s p j ). If we let r j , rank F(λ j )
for each j, then (50) implies
rj
n
F(L)x = ∑ ∑ sm j (φm j ⊗ w j ) x, (ψm j ⊗ w j ) Γ
(51)
j=1 m=1
for all x ∈Cnp . It is clear from (51) that
Col F(L) = span φm j ⊗ w j : 1 6 j 6 n, 1 6 m 6 r j (52)
Null F(L) = span ψm j ⊗ w j : 1 6 j 6 n, r j < m 6 p . (53)
In particular we note that
B(L) = B(L)WW ∗ Γ
`
= ∑ (Bi ⊗ Li )w1
...
(Bi ⊗ Li )wn
∗
W Γ
i=0
= B(λ1 ) ⊗ w1 . . . B(λn ) ⊗ wn W ∗ Γ ,
and we conclude
from (52) that
Col B(L) ⊂ Col F(L)
(54)
m
B(λ j ) ∈ Col F(λ j ) for all j ∈ {1,. . . , n} .
(55)
Next, we note that Null F(L) ⊂ Null C(L) if and only if
each basis vector ψm j ⊗ w j in the span (53) satisfies
`
0 = C(L)(ψm j ⊗ w j ) = ∑ (Ci ⊗ Li )(ψm j ⊗ w j )
i=0
= C(λ j )ψm j ⊗ w j = C(λ j )ψm j w j ,
and we conclude
from (53) that
Null F(L) ⊂ Null C(L)
(56)
m
∗
C (λ j ) ∈ Col F ∗(λ j ) for all j ∈ {1, . . . , n} . (57)
It follows from (55) and (57) that the dynamics Σ(∆G) admit
an equilibrium state x̄ with a unique corresponding equilibrium output ȳ for all constant inputs ū ∈ RV and all G ∈ G if
and only if conditions (ii) and (iii) of Theorem 5 hold. We
thus proceed assuming (ii) and (iii) hold, in which case the
vectors v j , F +(λ j )B(λ j ) are such that F(λ j )v j = B(λ j ) for
each j. It follows
from (54) that
B(L) = F(λ1 )v1 ⊗ w1 . . . F(λn )vn ⊗ wn W ∗ Γ
= F(L) v1 ⊗ w1 . . . vn ⊗ wn W ∗ Γ .
(58)
Thus we can
calculate
the
unique
output
equilibrium
ȳ
as
ȳ = C(L) v1 ⊗ w1 . . . vn ⊗ wn W ∗ Γū + D(L)ū
= C(λ1 )v1 w1 . . . C(λn )vn wn W ∗ Γū + D(L)WW ∗ Γū
= H(λ1 )w1 . . . H(λn )wn W ∗ Γū
n
= ∑ H(λ j ) w∗j Γū wj .
(59)
j=1
Because L1 = 0, we can assume without loss of generality
that w1 = 1 and λ1 = 0, and we can write (59) as
n
ȳ = H(0) 1T Γū 1 + ∑ H(λ j ) w∗j Γū wj .
(60)
j=2
We have left to show that ȳ in (60) has the value in (37) for
all inputs ū and all digraphs G ∈ G if and only if conditions
(iv)–(vi) hold. To this end, we first compute
n
n
j=1
j=2
1T Γ−1 = 1T WW ∗ = ∑ hwj , 1i w∗j = n1T + ∑ hwj , 1i w∗j , (61)
and we multiply both sides from the right by Γ/n to obtain
1 n
1 T
(62)
1 = 1T Γ + ∑ hwj , 1i w∗j Γ .
n
n j=2
In particular, we can multiply both sides of (62) from the
right by w1 = 1 to obtain
1 = 1T Γ 1 .
(63)
We also note that
G is balanced ⇔ LT1 = 0
⇔ 0 = 1TLW = λ1 hw1 , 1i . . . λn hwn , 1i
⇔ λ j hwj , 1i = 0 ∀ j ∈ {2, . . . , n} .
(64)
Now suppose conditions (iv)–(vi) hold; then G has maximal
rank and thus λ j 6= 0 for all j > 2, and it follows from (62)
and (64) that ȳ in (60) has the value in (37). Conversely,
suppose that ȳ in (60) has the value in (37) for every input ū.
It follows from (60) that H(0) 6= 0 and H(λ j ) = 0 for all
j > 2. Therefore λ j 6= 0 for all j > 2, which means G has
maximal rank and condition (v) holds. For the particular case
in which ū = 1, we see from (60) and (63) that ȳ = H(0)1,
and it follows that condition
(iv)
holds. At this point we
know that ȳ = 1T Γū 1 = 1Tū/n 1 for all ū, which implies
1T Γ = 1T/n. We conclude from (62) that hwj , 1i = 0 for all
j > 2, and it follows from (64) that G is balanced. A balanced
digraph with maximal rank is connected, which implies (vi).
R EFERENCES
[1] J. N. Tsitsiklis, Problems in Decentalized Decision-Making and Computation. PhD thesis, MIT, Cambridge, MA, 1984.
[2] A. Jadbabaie, J. Lin, and A. S. Morse, “Coordination of groups of
mobile autonomous agents using nearest neighbor rules,” IEEE Trans.
Automat. Contr., vol. 48, pp. 988–1001, June 2003.
[3] L. Xiao and S. Boyd, “Fast linear iterations for distributed averaging,”
Syst. Contr. Lett., vol. 53, pp. 65–78, Sept. 2004.
[4] R. Olfati-Saber and R. M. Murray, “Consensus problems in networks
of agents with switching topology and time-delays,” IEEE Trans.
Automat. Contr., vol. 49, pp. 1520–1533, Sept. 2004.
[5] W. Ren, R. W. Beard, and E. M. Atkins, “A survey of consensus
problems in multi-agent coordination,” in Proceedings of the 2005
American Control Conference, (Portland, Oregon), pp. 1859–1864,
June 2005.
[6] D. P. Spanos, R. Olfati-Saber, and R. M. Murray, “Dynamic consensus for mobile networks,” in Proceedings of the 2005 IFAC World
Congress, (Prague), 2005.
[7] R. A. Freeman, P. Yang, and K. M. Lynch, “Stability and convergence
properties of dynamic average consensus estimators,” in Proceedings
of the 45th IEEE Conference on Decision and Control, (San Diego,
CA), pp. 398–403, Dec. 2006.
[8] C. C. Moallemi and B. Van Roy, “Consensus propagation,” IEEE
Transactions on Information Theory, vol. 52, pp. 4753–4766, Nov.
2006.
[9] C. W. Wu, “Algebraic connectivity of directed graphs,” Linear and
Multilinear Algebra, vol. 53, no. 3, pp. 203–223, 2005.
[10] C. W. Wu, “On bounds of extremal eigenvalues of irreducible and
m-reducible matrices,” Linear Algebra and its Applications, vol. 402,
pp. 29–45, 2005.
[11] O. Taussky, “A recurring theorem on determinants,” The American
Mathematical Monthly, vol. 56, pp. 672–676, Dec. 1949.
[12] R. Merris, “A note on Laplacian graph eigenvalues,” Linear Algebra
and its Applications, vol. 285, pp. 33–35, 1998.
3203