Tracking and Formation Control of Multiple Autonomous Agents: A Two-Level Consensus Approach Maurizio Porfiri a , D. Gray Roberson b , Daniel J. Stilwell a b,1 Department of Mechanical, Aerospace and Manufacturing Engineering, Polytechnic University, Six MetroTech Center, Brooklyn, NY 11201 b The Bradley Department of Electrical and Computer Engineering, Virginia Polytechnic Institute and State University, Blacksburg, VA 24061 Abstract Simultaneous tracking and formation control is addressed for a team of autonomous agents that evolve dynamically in a space containing a measurable vector field. Each agent measures the local value of the field along its trajectory and occasionally shares relevant information with other agents, in order to estimate the spatial average obtained from averaging measurements across all agents. Using shared information, agents control their trajectories in a cooperative manner, with the dual goals of driving the average field measurement to a specified value and maintaining a desired formation about the average. Two approaches to virtual leader estimation are considered. The first involves the synthesis of a common virtual leader state, whereas the second involves decentralized estimation of the virtual leader by individual agents. Under the second approach, control is posed as a two-level consensus problem, where agents reach agreement on the virtual leader state at one level and reach formation about the virtual leader at the other level. The decentralized approach is effective even when communication among agents is limited, in the sense that the associated network graph can be disconnected in frozen time. Key words: Consensus problem; Formation flying; Switched systems; Stability; Tracking; Multi-agent systems 1 Introduction Widespread interest in multi-agent systems has been fueled by the breadth of natural and technological systems that exhibit coordinated activity of a group of autonomous agents. In nature, the flocking of birds [32], schooling of fish [12], and foraging of insects [30] provide insight into the emergence of group behaviors from the coordinated actions of individuals. In the field of robotics, autonomous vehicle platoons are modeled and studied as multi-agent systems. Applications have been identified in practically every environment accessible to vehicles, from mobile sensing in oceans [1], to navigation through unknown terrestrial environments [4], to surveillance by autonomous aircraft [2], to control of satellite formations in space [20]. Investigations into fundamental aspects of multi-agent system behavior have been widely reported, e.g. [14,19,21,23,26]. For example, 1 Corresponding author. Tel.: +1 540 231 3204. Fax: +1 540 231 3362. E-mail addresses: [email protected] (M. Porfiri), [email protected] (D. G. Roberson), [email protected] (D. J. Stilwell) Preprint submitted to Automatica [23] develops a comprehensive framework of flocking that captures the major features of natural systems, such as animal herds and schools, and that can be applied to vehicle platoons as well. It provides mathematical understanding of collective behaviors such as agent cohesion, separation, and alignment that have been identified as fundamental aspects of flocking behavior in [27]. In this paper, we study combined tracking and formation control of multiple autonomous agents. The agents evolve in a space that possesses a measurable vector field. For example, the agents may represent vehicles traveling in an environment that possesses a temperature, chemical, or magnetic field, e.g. [16,22]. The agents measure the value of the field locally and communicate their measurements with other agents. Using shared information, agents control their trajectories cooperatively, with the dual goals of driving the spatial average of the field measurement to a desired value and maintaining a desired formation about the average. To compute a spatial average of field measurements among agents would seem imply that all agents can communicate with all other agents or to an exogenous 12 January 2007 system. Although we briefly consider the case of an exogenous virtual leader to compute the spatial average, we are most concerned with the problem of potentially limited or sparse communication. To address limited communication, we consider the case that each agent is endowed with a local observer that estimates the spatial average given local measurements and information that is occasionally communicated from other agents. only tracking stability. The formation stability problem can be addressed by analyzing a collection of timeinvariant problems, while the tracking problem is inherently time-varying. In the decentralized virtual leader estimation case, the ability to decouple tracking and formation control is dependent on the communication network topology. Although it is not possible to decouple the two controller components for a general network, it is possible to do so for specific network topologies. We present a technique for decoupling tracking and formation control under a specific but practically useful class of communication network topologies, and for subsequently designing stable controllers. In this case, vector field properties and the graph spectrum simultaneously affect both tracking and formation stability. The decoupling between formation and tracking problems results in a collection of time-varying subsystems. Our approach to tracking and formation control utilizes the concept of a virtual leader. In contrast to [8,22], which use artificial potentials to synthesize a virtual leader trajectory, we adopt a graph theoretic approach. The use of virtual leaders within a graph framework has also been considered in [7,23]. In these works, the virtual leader acts as an external reference signal that represents global group objectives. The role of the virtual leader is different in the present work, in that the virtual leader synthesis is tied to a measured vector field in the underlying ambient environment, and is not defined by an absolute trajectory. We retain the mathematical structure for multi-agent formation problems developed in [7], and based on consensus theory, e.g. [24,25]. However, to address the dependence of the tracking problem on the ambient vector field, we extend the mathematical model and control design by connecting the virtual leader synthesis to the agent field measurements. Furthermore, we consider the case of limited inter-agent communication by allowing for a time-varying communication network. Our approach results in a coupled formation and tracking problem where system behavior is simultaneously affected by the measured field, individual agent dynamics, and dynamic communication topology. The decentralized estimation strategy is developed first for systems whose communication network is static, and is then extended to systems whose communication is dynamic. Control of spatially distributed systems under dynamic communication is an important research area, e.g. [11,13,18,33]. For example, in [11] simultaneous optimization of communication and control objectives is studied. That work considers discrete-time systems with a central decision maker that communicates with subsystems independently across limited capacity channels. We instead focus on the decentralized case. We also limit our attention to intermittent communication between subsystems, and not to channel bandwidth constraints. We model the system as a switched system, consisting of a set of time-varying subsystems and a protocol for transitioning between them, e.g. [17,31]. We take advantage of results from fast switching theory to derive new stability conditions for systems with limited communication, e.g. [29]. We start by presenting the common virtual leader approach for systems whose communication network is connected and static. The approach is practically suitable when an exogenous system is present, and is theoretically useful to present the basic tools for solving the formation tracking problem under severely limited communication. We then build on the single virtual leader strategy to develop a novel technique that involves the independent estimate of the virtual leader trajectory by each agent, based on local measurements and shared information. Tracking and formation control takes place in a two-level consensus framework, where agents reach agreement on their local virtual leader estimates at one level and achieve formation at a second level. The theoretical results of this paper may be applied to the problem of tracking and formation control with a platoon of autonomous underwater vehicles (AUVs) for environmental monitoring applications, e.g. [5,15]. Equipped with appropriate sensors, AUVs can measure underwater fields such as scalar temperature, salinity, or pollutant concentration fields, or vector magnetic fields such as those produced by submarines. Furthermore, communication within AUV platoons may be severely bandwidth-limited, e.g. [8]. For the common virtual leader case, we present a design technique for decoupling the centralized virtual leader computation from the decentralized formation control, and we give stability conditions for the system. We show that formation stability may be studied by analyzing reduced-order system models whose dynamics are related to the spectral properties of the communication graph and to the vector field properties. In particular, stability of the formation is only affected by the graph spectral properties, while field characteristics influence 2 Model of Multi-Agent System We consider a platoon of N agents with identical dynamics. Each agent is modeled as the linear system ẋi = PA xi + PB ui , yi = PC xi (1) where i = 1, . . . , N identifies the agent, xi ∈ Rn represents the agent’s state, ui ∈ Rm is an input generated 2 by a control law, and yi ∈ Rk is an internal state measurement. Denoting x = (x1 , . . . , xN ), u = (u1 , . . . , uN ), and y = (y1 , . . . , yN ) as the concatenation of the vectors xi , ui , and yi , the state equation for the entire system is ẋ = (I ⊗ PA )x + (I ⊗ PB )u , Because α and β are unknown parameters, the virtual leader cannot be viewed as an external reference signal. The formation is specified by a possibly time dependent vector h(t) = (h1 , . . . , hN ) ∈ RN k , whose entries are the desired internal state measurement offsets from the virtual leader state for the corresponding agents, so that the vector of desired internal state measurements is y = (I ⊗ PC )x where ⊗ represents the matrix Kronecker product and all identity matrices are N × N . ỹ(t) = 1N ⊗ q̃(t) + h(t) In addition, we consider a set of measurements φi ∈ Rk for each agent corresponding to an external vector field. These measurements are dependent on the internal state measurement through φi = ϕ(yi , t). When the distance between agents is small relative to spatial variation in the field, it is reasonable to assume an affine description of the field. We therefore assume that locally in the immediate area in which the agents are operating, the field ϕ can be represented adequately as an affine function of the internal state measurements. The vector field measurement of agent i is then φi = αyi +β, where α ∈ Rk×k and β ∈ Rk are time-varying parameters. The vector of field measurements Φ = (φ1 , . . . , φN ) for all agents is Φ(y, t) = IN ×N ⊗ α(t)y + 1N ⊗ β(t) For consistency, the relationships hT (1N ⊗ ei ) = 0 must hold for i = 1, . . . , k, where ei ∈ Rk is a vector whose elements are 0, except for a 1 in the ith location. These relationships indicate that the virtual leader should occupy the geometric center of the formation. In the sequel, we simplify notation by omitting explicit time dependence, except when doing otherwise clarifies the presentation. The multi-agent system and vector field model is used in Sections 3, 4, and 5 to study tracking and formation control design. The following cases are considered: • A common virtual leader is computed by an exogenous system under a static communication topology. • A separate virtual leader estimate is computed by each agent, and the network topology is static. • A separate virtual leader estimate is computed by each agent, and the network topology switches periodically among a set of possibly unconnected topologies. (2) where 1N ∈ RN is a vector all of whose elements equal 1. Following [7], graph theory is used to model the platoon communication network, with nodes representing individual agents and edges representing active communication links. All communication is bidirectional, so that the corresponding graph is undirected. A connected network is one in which a direct or indirect path exists between each pair of nodes in the associated graph. The Laplacian matrix is used extensively in characterizing the effects of the communication network. The Laplacian is defined as L = D − A, where A is the graph adjacency matrix, and D is the diagonal degree matrix whose diagonal entries equal the number of edges incident with the corresponding nodes. Defined in this manner, L is symmetric and positive semidefinite. In addition, the vector 1N is in the kernel of L. The total number of zero eigenvalues of L equals the number of connected components of the graph. The second smallest eigenvalue λ2 is termed the algebraic connectivity of the graph [9]. 3 Tracking and Formation Control: Common Virtual Leader and Static Communication To introduce the basic concepts of our approach, we begin with the case of a single exogenous virtual leader. The virtual leader is modeled as a set of integrators, q̇ = w, where w ∈ Rk is an input generated by a control law. The information sensed by the ith agent is zij = (yi − yj ) − (hi − hj ) , zii = yi − hi − q j ∈ Ji (5a) (5b) where the set Ji ⊂ {1, . . . , N } represents the set of agents that the ith agent can sense, i.e. the graph nodes connected to the ith node. The quantity zij represents the error committed by the j th agent in maintaining the desired offset with respect to the ith agent, while zii represents the error committed by the ith agent in maintaining the desired offset with respect to the virtual leader. These errors P are synthesized in a unique error signal zi = zii + j∈Ji zij . The vector z = (z1 , . . . , zN ) is given by z = ((L + IN ×N ) ⊗ Ik×k )(y − h − 1N ⊗ q). We also define a decentralized control law K which maps zi to ui and has internal states vi ∈ Rs , represented by Agents are tasked with a twofold mission: tracking a desired vector field value φ̃, and maintaining a desired formation about the desired vector. The control strategies involve the synthesis of a virtual leader q ∈ Rk , whose purpose is to track the desired field vector and to provide a reference about which the agents form. Under the affine field assumption, the desired virtual leader state q̃(t) = α−1 (t)(φ̃(t) − β(t)) (4) (3) corresponds to the desired location in the vector field. v̇i = KA vi + KB zi , 3 ui = KC vi + KD zi (6) The information sensed by the virtual leader τ= where i = 2, ..., N , and if K and V stabilize the system 1 T (1 ⊗ Ik×k )Φ − φ̃ N N q̇ = w τ = α(t)y1 (7) represents the error committed by the set of agents in keeping their average measurement to the desired value φ̃. We define a control law V which maps τ to w and has internal state r ∈ Rl , represented by ṙ = VA r + VB τ , w = VC r + VD τ Proof. Since L is a symmetric matrix, it is diagonalizable and has eigenvectors ϕ1 , ϕ2 , . . . , ϕN that are orthogonal and may be chosen to be orthonormal, with ϕ1 = 1N N −1/2 . Define the block diagonal transformation matrix Q = diag(Ik+l×k+l , T ⊗ In×n , T ⊗ Is×s ), where T = [ϕ1 ϕ2 · · · ϕN ]. Because the column vectors ϕi are orthonormal, T is an orthogonal matrix. Consider the similarity transformation η̃ = Q−1 η = (q̃, r̃, x̃, ṽ), which is the state of a system whose state equation is Specification of (6) and (8) as dynamic control laws is not meant to imply the superiority of dynamic control for the tracking and formation problem. The intent is merely to present the control design in a general framework. We now consider the complete set of agents together with the virtual leader and use Kronecker product to assemble the matrices governing the formation and tracking behavior. Let v indicate the concatenation of the vectors v1 , . . . , vN . The closed loop system dynamics becomes η̃˙ = Q−1 AQη̃ + Q−1 Bp (9) 0 VC 0 VA A= −1N ⊗ PB KD 0 h iT p = β T φ̃T hT 1 T N VD 1N 1 T N VB 1N ⊗ αPC Υ (10) 0 0 Λi = −KB 0 KB P C KA PA + PB KD PC (λi + 1) PB KC q̃˙ 0 r̃˙ 0 = x̃˙ 1 −PB KD ṽ˙ 1 −KB (11) Analysis of (9) begins by presenting a necessary and sufficient condition for uniform asymptotic stability. KB PC (λi + 1) # KA VC VD α(t)PC VA VB α(t)PC 0 PA + PB KD PC 0 KB P C 0 q̃ r̃ PB KC x̃1 KA ṽ1 0 which correspond to the state coefficient matrix Λ1 (t). The closed loop dynamics obtained by applying K to (12) are Proposition 1 The controllers K and V stabilize the dynamics of the system (9) if and only if K simultaneously stabilizes the set of N − 1 systems zi = (λi + 1)PC xi 0 where all identity matrices are N × N and ẋi = PA xi + PB ui , VD α(t)PC for i = 2, . . . , N . Stability of the original system (9) may be determined from stability of (14). Stability may therefore be determined from the time-varying homogeneous system whose state coefficient matrix is Λ1 (t), and the N − 1 time-invariant homogeneous systems whose state coefficient matrices are Λi . The closed loop dynamics obtained by applying the controllers K and V to (13) are −(L + I) ⊗ KB Υ = I ⊗ PA + (L + I) ⊗ PB KD PC VC " I ⊗ P B KC I ⊗ KA 0 −1N ⊗ KB 0 (L + I) ⊗ KB PC V −VD 0 D V −V 0 B B B= 0 0 −(L + I) ⊗ PB KD 0 0 0 VA VB α(t)PC 0 Λ1 (t) = (15) −PB KD 0 PA + PB KD PC PB KC ⊗ αPC (14) The state matrix Q−1 AQ = diag Λi of (14) is block diagonal, with with h iT η = q T rT xT v T , (13) where λi are the eigenvalues of L, ordered so that 0 = λ1 ≤ · · · ≤ λN . (8) η̇ = Aη + Bp ẋ1 = PA x1 + PB u1 z1 = PC x1 − q " # " #" # x̃˙ i PA + PB KD PC (λi + 1) PB KC x̃i = ṽ˙ i KB PC (λi + 1) KA ṽi (12) 4 which represents the error committed by the ith agent in maintaining the desired offset with respect to its local virtual leader estimate qi . By synthesizing the error signals in a unique error signal zi defined as the sum of the errors sensed by the ith agent, the error vector z becomes which correspond to the state coefficient matrices Λi . Therefore the stability of these closed loop systems is equivalent to stability of the original system (9). 2 Tracking control, achieved through the virtual leader, is associated with Λ1 in (15). Indeed, since the ambient vector field parameter α appears in Λ1 only, tracking is decoupled from the platoon formation control problem, which is associated with the Λi for i ≥ 2. Because the Λi are constant for i ≥ 2, stability of the formation problem reduces to the requirement of Hurwitz Λi . Time-varying methods, such as those presented in [28], are required to determine the stability of the tracking problem, due to the time-varying nature of Λ1 . The case where α(t) varies slowly may be particularly relevant. Additional investigations on the formation stability may be undertaken by using Nyquist methods, as done in [7] for formation control about static targets. We emphasize that the offset function h, the desired measurement φ̃ and the parameter β act only as inputs to the dynamical system, while α(t) enters directly in the state matrix and influences the system stability. z = ((L + IN ×N ) ⊗ Ik×k )(y − h − q) (16) The sensed information is used by the decentralized controllers (6). On the other hand, we replace the measurement error in (7) by the vector τ = (τ1 , . . . , τN ) comprised of the local measurement errors computed by each agent τ 6= (F ⊗ Ik×k )(Φ − 1N ⊗ φ̃) (17) where F ∈ RN ×N computes a local weighted average and satisfies Fij 6= 0, if Aij 6= 0, i, j = 1, . . . N . Thus, each agent computes its own average estimate by using the measurement from the communicating agents. Consistent with its averaging function, F has 1N as an eigenvector corresponding to the eigenvalue 1. Due to the incomplete averaging inherent in F , the control law does not aim at exactly regulating τ to zero. Indeed, the control objective is not to drive each entry of τ to zero, but to ensure that the average of the entries in τ , which corresponds to (17), is zero. Computation of the virtual leader trajectory from agent measurements is a unique feature of the controller. Other works, such as [7,14], also consider formation about a leader or target. In such works, though, the leader or target is modeled as an additional agent that communicates with or is sensed by the agents. In our system, the leader trajectory cannot be sensed independently by the agents, due to the nature of the environmental field. Construction of the virtual leader trajectory from measurements is therefore a distinguishing characteristic of our controller. We emphasize that our virtual leader trajectory is not known a priori, as in [23], but is rather constructed in real time as field measurements are taken. A possible choice for F is F = (IN ×N + D)−1 (IN ×N + A) (18) which provides unweighted local average measurements. When using (18) in (17), the elements of τ become τi = ³ X ´ 1 Φi + Φj − φ̃ 1 + |Ji | j∈Ji 4 Tracking and Formation Control: Decentralized Virtual Leader Estimation and Static Communication where |Ji | is the number of agents communicating with the ith agent. When each agent communicates with ρ other agents, F may be written as A second tracking and formation control strategy involves independent estimation of the virtual leader trajectory by the agents. Each agent estimates a spatial average of the field measurements based on its local measurements and information from the agents with which it communicates. We model the virtual leader estimators as single integrators by q̇ = w, where q = (q1 , . . . , qN ) contains the virtual leader position estimates and w ∈ RN k are inputs generated by decentralized control laws. F = IN ×N − (1 + ρ)−1 L In addition to the error signal zij representing the errors committed by the ith agent in keeping the required offsets with respect to j th one or to its virtual leader estimate, we define the error signal eij = qi − qj , where j ∈ Ji , to represent the disagreement error between the virtual leader estimate of the ith and j th agents. For simplicity, these errors are synthesized in a unique error signal ei defined as the sum of the errors sensed by the ith agent. Thus, the vector e = (e1 , . . . , eN ) is given by e = (L ⊗ Ik×k )q. In order to drive the virtual leader estimates to the desired common value, we define a control law V which maps τi and ei to wi and has internal state Since we assume that each agent does not sense the actual virtual leader position but only its proper estimate, we modify (5a) and (5b) as zij = (yi − yj ) − (hi − hj ) − (qi − qj ) , zii = yi − hi − qi (19) j ∈ Ji 5 ri ∈ Rl , represented in state-form by column vectors ϕi are orthonormal. Applying the similarity transformation Q to A in (21) yields ṙi = VA ri + VB1 τi + VB2 ei wi = VC ri + VD1 τi + VD2 ei Q−1 AQ (22) I ⊗ VC F̃ ⊗ VD1 αPC L̃ ⊗ VD2 0 L̃ ⊗ V 0 I ⊗ VA F̃ ⊗ VB1 αPC B2 = −M̃ ⊗ PB KD 0 Υ̃ I ⊗ P K B C −M̃ ⊗ KB 0 M̃ ⊗ KB PC I ⊗ KA Let r indicate the concatenation of the vectors r1 , . . . , rN . The closed-loop system dynamics becomes η̇ = Aη + Bp (20) with η and p defined as in (10) and where L̃ and F̃ are diagonal matrices whose diagonal elements are the eigenvalues λi and µi of L and F , respectively, ordered by the eigenvectors ϕi , where all identity matrices are N × N , where M̃ = L̃ + I, and where Υ̃ = I ⊗ PA + M̃ ⊗ PB KD PC . By permuting rows and columns, the matrix (22) may be written in block diagonal form, with the generic 2 × 2 blocks L ⊗ VD2 I ⊗ VC F ⊗ (VD1 αPC ) 0 L⊗V I ⊗ VA F ⊗ (VB1 αPC ) 0 B2 A = −M ⊗ PB KD 0 Υ I ⊗ P B KC −M ⊗ KB 0 M ⊗ KB PC I ⊗ KA (21) 1 ⊗ VD1 −1N ⊗ VD1 0 N 1 ⊗ V −1 ⊗ V 0 B1 N B1 N B= 0 0 −(L + I) ⊗ PB KD 0 0 −(L + I) ⊗ KB λi VD2 VC µi VD1 αPC −(λi + 1)KB 0 (λi + 1)KB PC 0 λi VB 2 VA µi VB1 αPC 0 −(λi + 1)PB KD 0 PA + (λi + 1)PB KD PC PB KC KA (23) where all identity matrices are N ×N , where M = L+I, and where Υ is defined in (11). which yields the claim. In general, a decomposition of the system such as the one performed in Section 3 does not simplify stability analysis. However, it is possible to work with (20) directly and to specify properties of F and L that simplify analysis when they are present. The following proposition provides an example. Each subsystem (23) is time-varying, and stability must be evaluated using time-varying system analysis methods. The commutativity of L and F in Proposition 2 is motivated by a platoon network in which each agent communicates with a fixed set of relative neighboring agents, and computes an unweighted average as in (19), or in which the communication graph is circulant [9]. In the latter case, F can be defined as a circulant matrix. Since L is also circulant under the considered topology, L and F commute, according to a standard property of circulant matrices [6]. When the graph is complete (fully connected) and F is chosen as in (19), the tracking error is zero for any constant value of h. Proposition 2 Suppose that L and F commute. The controllers K and V stabilize the dynamics of the system (20) if and only if K and V simultaneously stabilize the set of N systems ẋi = PA xi + PB ui q̇i = wi zi = (λi + 1)(PC xi − qi ) τi = µi αPC xi e i = λi q i 5 for i = 1, . . . , N , where λi are the eigenvalues of L, ordered such that 0 = λ1 ≤ · · · ≤ λN , and µi are the eigenvalues of F , 1 = µ1 ≥ · · · ≥ µN ordered according to the common eigenvectors. 2 Tracking and Formation Control: Periodic Switched Communication We now investigate tracking and formation control for a platoon with switched communication. Our main result is that the average network determines stability of the system if the network switches at a sufficiently fast rate. Moreover, the average system provides a useful approximation of the switched system, as the simulations in Section 6 illustrate. Communication may be severely limited, so that at any point in time there may be isolated agents that do not share information. Hence the approach of estimating the virtual leader trajectory on each agent, as described in Section 4, is adopted. Proof. Since L and F commute, they share the set of orthonormal eigenvectors ϕ1 , . . . , ϕN . These eigenvectors are real due to the symmetry of L. It follows that the eigenvalues of F are real, and thus its eigenvalues 1 = µ1 ≥ · · · ≥ µN are real. Define the block diagonal matrix Q = diag(T ⊗ Ik×k , T ⊗ Il×l , T ⊗ In×n , T ⊗ Is×s ), where T = [ϕ1 ϕ2 · · · ϕN ] is orthogonal because the 6 The platoon formation error z continues to be defined by (16) with the qualification that L is time-varying. The estimation error (17) is modified so that τ = (F ⊗ In×n )(Φ − 1N ⊗ φ̃) to account for the fact that isolated agents rely on the virtual leader estimate computed during their most recent communication event to drive their navigation. The time-varying matrix F is the same as F in (17) with the exception that rows corresponding to isolated agents vanish. The matrices L and F switch among sets of constant matrices {Lj } and {Fj }. At every time instant, L = Lj and F = Fj for some j ∈ {1, . . . , J}, where J is the total number of constant matrices. To simplify the analysis, we assume that the platoon communication network switches periodically, i.e. that there is a T ∈ R+ such that F(t) = F(t + kT ) and L(t) = L(t+kT ) for all t ∈ R+ and k ∈ N. Each pair Lj , Fj is active for a set percentage δj of the switching PJ period T , with i=1 δi = 1, so that L(t) = Lj , F(t) = Fj Lemma 3 Suppose the matrix functions S1 (·), S2 (·) : R+ → Rn×n are right-continuous, bounded and discontinuous in a finite number of instants over every finite interval, and the matrix function G(·) : R+ → Rn×n is differentiable and bounded, and has a bounded derivative for any t ∈ R+ . Also suppose there exists τ > 0 such that S1 and S2 haveRuniform average over any interval of dut+τ ration τ , i.e. τ1 t Si (t) = S̄i is constant for all t ∈ R+ and i = 1, 2. If the linear time-varying system ψ̇(t) = (S̄1 + G(t)S̄2 )ψ(t) , ẏ(t) = (S1 (t/ε) + G(t)S2 (t/ε))y(t) , Using Lemma 3, we seek to characterize stability of (25) in terms of the uniform average values L̄ = J X δj Lj , j=1 (25) I ⊗ VC F ⊗ VD1 αPC 0 L⊗V I ⊗ VA F ⊗ VB1 αPC 0 B2 A= −M ⊗ PB KD 0 Υ I ⊗ PB KC −M ⊗ KB 0 F̄ = J X δ j Fj (29) j=1 of L(t) and F(t) over the period T and the corresponding average state coefficient matrix L̄ ⊗ VD2 I ⊗ VC F̄ ⊗ VD1 αPC 0 L̄ ⊗ V I ⊗ VA F̄ ⊗ VB1 αPC 0 B2 Ā = −M̄ ⊗ PB KD 0 Ῡ I ⊗ PB KC −M̄ ⊗ KB 0 M̄ ⊗ KB PC I ⊗ KA (30) where all identity matrices are N × N , M̄ = L̄ + I, and Ῡ = I ⊗ PA + M̄ ⊗ PB KD PC . Although Ā incorporates the constants L̄ and F̄, it is nevertheless time-varying due to the parameter α. The following proposition provides a useful control design tool for T -periodic systems. is obtained from (21) upon substituting F with F, i.e. L ⊗ VD 2 y(0) = y0 (28) is uniformly exponentially stable. ´ h Pj Pj−1 for t ∈ kT + T i=1 δi , kT + T i=1 δi . Therefore, the state matrix of the switched linear system (27) is uniformly exponentially stable, then there exists ε∗ > 0 such that for all fixed ε ∈ (0, ε∗ ), (24) ẏ(t) = A(t)y(t) + B(t)p(t) ψ(0) = ψ0 M ⊗ KB PC I ⊗ KA (26) where L and F are defined in (24), where all identity matrices are N × N , where M = L + I, and where Υ is defined as in (11). Note that A is time-varying, due to the time-varying nature of L, F, M , Υ, and α. The time-varying input vector is Proposition 4 Let α be a bounded differentiable matrix function with a bounded derivative on the entire time axis. If the time-varying average system F1N ⊗ VD1 −F1N ⊗ VD1 0 F1 ⊗ V −F1 ⊗ V 0 B1 N B1 N B= 0 0 −M ⊗ P K B D 0 0 −M ⊗ KB ψ̇(t) = Ā(t)ψ(t) , ψ(0) = ψ0 since in general F1N 6= 1N . To obtain stability results, we study the response of the time-averaged system and relate it to the response of a platoon whose communication network switches at a sufficiently fast rate. with Ā(t) defined in (30), is uniformly asymptotically stable, then there exists a T ∗ > 0 such that the switched system ẏ(t) = A(t)y(t) , y(0) = y0 with A(t) given in (26), is uniformly asymptotically stable for any period T < T ∗ . A new result for analyzing the stability of time-varying switched systems is required, the proof of which is found in the Appendix. Proof. Define auxiliary functions L∗ (t/T ) = L(t) and F ∗ (t/T ) = F(t). We note that the period of L∗ and F ∗ 7 is τ = 1. Using L∗ and F ∗ , the homogenous linear differential equation associated with (25) has state matrix A(t) = S1 (t/T ) + G(t)S2 (t/T ), where An example platoon with switched communication whose average system meets the conditions of Proposition 5 is one in which switching occurs in a circulant manner. That is, the commutativity of L̄ and F̄ are satisfied if, over the course of the switching period, each agent communicates with a common set of relative neighbors for a common amount of time. S1 (t/T ) L(t) ⊗ VD2 I ⊗ VC 0 0 L(t) ⊗ V I ⊗ VA 0 0 B2 = −M (t) ⊗ PB KD 0 Υ(t) I ⊗ P K B C −M (t) ⊗ KB 0 M (t) ⊗ KB PC I ⊗ KA 6 In this section we illustrate tracking and formation control, by considering a platoon of agents moving in a twodimensional space possessing a two-dimensional affine and time-invariant vector field. Each agent is modeled as a point mass with viscous-like damping. The internal measurement corresponds to a position-like variable, and the control input corresponds to a force-like variable. Therefore the agent model in (1) is given by S2 (t/T ) = F(t) ⊗ I(s+l+n+k)×(s+l+n+k) 0 0 VD1 α(t)PC 0 0 0 V α(t)P 0 B1 C G(t) = I ⊗ 0 0 0 0 00 0 0 " where all unlabeled identity matrices are N × N , M = L + I, and Υ is defined as in (11). By applying Lemma 3 with ε = T , the claim follows. 2 PA = 0 1 # , 0a PB = " # 0 1 , h i PC = 1 0 (31) We apply the control design procedures presented in Sections 3 and 5 and simulate the controlled motion of a platoon of four agents. As example control laws, we use purely static feedback. For the common and virtual leader approach we use scalar K and V of the type ui = KD zi and w = VD τ . For the decentralized virtual leader approach, we use ui = KD zi and wi = VD1 τi + VD2 ei . Proposition 4 guarantees that nonzero stabilizing switching periods exist under the given hypotheses and provides an upper bound estimate for the stabilizing periods. The estimate is typically conservative, and for certain systems it may even be the case that stability is obtained for all switching periods. Other switching techniques involving dwell time or state feedback, e.g. [31], may be effective under appropriate conditions. For the simulation, the configurable parameter in the agent model (31) is set to a = −3. The vector field is constant with parameters In general, L̄ in (29) is not a graph Laplacian matrix, i.e. there is no graph whose Laplacian is L̄. Nevertheless, it is easy to show that it is symmetric and has 1N in its kernel. By application of the Geršgorin disc theorem [10], it is straightforward to demonstrate that L̄ is positive semidefinite. By combining Propositions 2 and 4 the following result holds. " α= 0.1 0.02 0.02 0.2 # " , β= # 0 0.1 The control objectives are given by h iT φ̃ = −0.2 −0.4 , Proposition 5 Let α be a bounded and differentiable matrix function with a bounded derivative for t ∈ R+ , and suppose that L̄ and F̄ commute. Suppose that the controllers K and V stabilize the dynamics of the systems ẋi = PA xi + PB ui q̇i = wi zi = (λ̄i + 1)(PC xi − qi ) τi = µ̄i αPC xi Tracking and Formation Control Simulation h iT h = 1 1 −1 1 −1 −1 1 −1 The control law is defined as KD = −0.5, VD = −1, VD1 = −1, and VD2 = −1. ei = λ̄i qi The first simulation illustrates the common virtual leader approach with a static and fully connected communication network, where the communication graph is complete. Figure 1 shows the agent trajectories. where i = 1, . . . , N , the eigenvalues λ̄i of L are ordered such that 0 = λ̄1 ≤ · · · ≤ λ̄N , and the eigenvalues µ̄i of F̄ are ordered such that µ̄1 ≥ · · · ≥ µ̄N according to the common eigenvectors. Then there exists T ∗ > 0, such that for any T < T ∗ the controllers K and V stabilize the dynamics of the switched system (25). The second simulation illustrates the switched communication case. Most of the time all agents do not share information. At periodic time intervals, three agents communicate, and the communication pattern 8 -4 -3 -2 4 4 3 3 2 2 1 1 1 -1 2 3 4 -4 0 1 0 , 1 0 0 0 0 0 1 -1 -1 -2 -2 -3 -3 -4 -4 1/4 3/8 F2 = 3/8 0 2 3 4 Fig. 2. Agent trajectories, decentralized virtual leader estimation with switched communication. repeats in a circulant manner. Over the course of a switching period, eight communication events occur. The corresponding adjacency matrices for the timevarying network are such that A1 , A3 , A5 , A7 are 4 × 4 zero matrices, and A2 , A4 , A6 , A8 represent all possible three-vehicle communication patterns. The averaging matrices are defined so that F1 , F3 , F5 , F7 are 4 × 4 zero matrices, and F2 , F4 , F6 , F8 reflect the active communication network. So for example 1 1 0 -2 -1 Fig. 1. Agent trajectories, common virtual leader with static and fully connected communication. 0 1 A2 = 1 0 -3 4 3 2 1 -4 3/8 3/8 0 1/4 3/8 0 3/8 1/4 0 0 0 0 -3 -2 1 -1 2 3 4 -1 -2 -3 -4 Fig. 3. Agent trajectories, average model of switched communication. This choice of F matrices does not correspond to (18), but rather yields an unweighted F̄ = 0.075(1N 1T N ). The overall switching period is T = 5, and by setting δ1 = δ3 = δ5 = δ7 = 0.15 and δ2 = δ4 = δ6 = δ8 = 0.1, each agent is isolated for 80% of the time. That is, each agent communicates with no other agent 80% of the time. This communication pattern yields the average graph Laplacian L̄ = 0.2Lf , where Lf = N IN ×N − 1N 1T N is the Laplacian of a fully connected communication network. ory of averaging, see e.g. [3], the error for the switched communication case is bounded and may be constrained to be arbitrary small by decreasing the switching period. With the choice of controller gains, the formation error is regulated faster than the tracking error. Roughly speaking, agents reach formation quickly, then move as a rigid group to satisfy the tracking objective. Figure 2 shows the agent trajectories under switched communication. Figure 3 displays the trajectories computed using the average model. From a comparison between the switched and average cases, the average model provides a reasonable approximation of the switched communication model. Indeed, the average and switched system trajectories in Figures 2 and 3 are nearly aligned. The simulations illustrate the effectiveness of the theoretical results and the control design procedure. In the switched communication case, the agent trajectories approximately follow those achieved by the average system. Acceptable performance is obtained in the switched communication case with reasonable communication network switching rates. This means that stability can be assessed even if the network is disconnected in frozen time. The simulations represent an idealized system. For practical applications, other factors such as obstacle and interagent collision avoidance must be considered. The convergence rate of the decentralized systems is slower than the rate of the system possessing a common virtual leader. This slower rate is directly attributable to infrequent communication. As one expects from the the- 9 7 Conclusion Φ̄, we obtain Z There are several notable features of the control problem considered in this paper and the strategies used to accomplish control objectives. In contrast to typical tracking problems, e.g. [7,34], the entity being tracked is not a moving target or a path defined in spatial coordinates. The location of the target evolves in time with the agent motions and is tied to an unknown process external to the agents, introducing additional complexity. Tracking and formation control are coupled and influenced by time-varying, state-dependent measurements. Second, the control and estimation architecture is placed in a consensus framework. When a common virtual leader is estimated, formation about the virtual leader is posed as a consensus problem. In the decentralized virtual leader case, the control and estimation is addressed in a twolevel consensus framework, with agents reaching consensus on the virtual leader at one level, and forming about the virtual leader at the other level. Finally, the possibility of simplifying control design by decoupling the tracking and formation control has been explored, and design criteria that guarantee system stability have been identified in cases where decoupling is possible. t+ετ ∆Φ(t, t + ετ ) = I + + ∞ Z X i=2 (S1 (ξ/ε) + G(ξ)S2 (ξ/ε))dξ t Z t+ετ ξ1 (S1 (ξ1 /ε) + G(ξ1 )S2 (ξ1 /ε)) Z ··· t t ξi−1 (S1 (ξi /ε) + G(ξ1 )S2 (ξi /ε))dξi · · · dξ1 t t+ετ Z −I − − (S̄1 + G(ξ)S̄2 )dξ ∞ Z X i=2 t Z t+ετ ξ1 (S̄1 + G(ξ1 )S̄2 ) t ··· t Z ξi−1 (S̄1 + G(ξi )S̄2 )dξi · · · dξ1 (34) t where I is the identity matrix. Consider the term Z t+ετ (S1 (ξ/ε) + G(ξ)S2 (ξ/ε) − S̄1 − G(ξ)S̄2 )dξ (35) t By definition of S̄1 , (35) may be rewritten as 8 Appendix Z t+ετ t £ ¤ S1 (ξ) + α(ξ)S2 (ξ) − S̄1 − α(ξ)S̄2 dξ Z t+ετ £ ¤ = G(ξ) S2 (ξ/ε) − S̄2 dξ t Proof of Lemma 3. Since the average system (27) is uniformly asymptotically stable, there is a matrix function Q(t) and positive scalars η, ρ, and µ such that the Lyapunov function V (ψ(t), t) = ψ T (t)Q(t)ψ(t) satisfies ηkψ(t)k2 ≤ V (ψ(t), t) ≤ ρkψ(t)k2 and V̇ (ψ(t), t) ≤ −µkψ(t)k2 for all t in the Euclidean norm. To establish uniform asymptotic stability of (28) we show that V (y(t), t) is also a Lyapunov function for (28) if ε is sufficiently small, by showing that for small ετ Define r(ξ, σ) = G(ξ) − G(σ). Since the first derivative of G(t) is bounded, there exists a positive constant m such that kr(ξ, σ)k ≤ |ξ − σ|m in the Euclidean norm for any ξ and σ. Therefore, (35) becomes Z t+ετ (S1 (ξ/ε) + G(ξ)S2 (ξ/ε) − S̄1 − G(ξ)S̄2 )dξ Z t+ετ £ ¤ = (G(t) + r(ξ, t)) S2 (ξ/ε) − S̄2 dξ t ∆V (y, t + ετ, t) = V (y(t + ετ ), t + ετ ) − V (y(t), t) (32) t £ ¤ R t+ετ which equals t r(ξ, t) S2 (ξ) − S̄2 dξ by definition of S̄2 . Therefore, the following bound holds is negative definite for all t, see e.g. [29]. Let Φ and Φ̄ be the transition matrices of (28) and (27), respectively, and let °Z ° ° ° t ∆Φ(s, t) = Φ(s, t) − Φ̄(s, t) (33) t+ετ ° ° (S1 (ξ/ε) + G(ξ)S2 (ξ/ε) − S̄1 − G(ξ)S̄2 )dξ ° ° Z t+ετ ° ° ≤ kr(ξ, t)k °S2 (ξ/ε) − S̄2 ° dξ t ≤ M (ετ )2 By using the Peano-Baker series representation of Φ and 10 (36) ° ° where M = m supt≥0 °S2 (t) − S̄2 °. Letting [2] R. W. Beard and T. W. McLain. Multiple UAV cooperative search under collision avoidance and limited range communication constraints. In Proceedings of the 42nd IEEE Conference on Decision and Control, pages 25–30, Maui, HI, December 2003. ¶ µ Ω = max sup kS1 (t)k, kS̄1 k t≥0 µ ¶ + sup kG(t)k max sup kS2 (t)k, kS̄2 k t≥0 [3] N. N. Bogoliubov and Y. A. Mitropolsky. Asymptotic Methods in the Theory of Non-Linear Oscillations. Gordon and Breach Science Publishers, New York, 1961. t≥0 [4] W. Burgard, M. Moors, C. Stachniss, and F. E. Schneider. Coordinated multi-robot exploration. IEEE Transactions on Robotics, 21(3):376–386, June 2005. and accounting for (36), (34) may be bounded by k∆Φ(t, t + ετ )k ≤ f (ε) [5] J. Clark and R. Fierro. Cooperative hybrid control of robotic sensors for perimeter detection and tracking. In Proceedings of the American Control Conference, pages 3500– 3505, Portland, OR, June 2005. (37) where f (ε) = 2(exp(Ωετ ) − 1 − Ωετ ) + M (ετ )2 . Note df that f (0) = 0 and dε (0) = 0. By using (33), (32) may be expressed as [6] P. J. Davis. Circulant Matrices. John Wiley & Sons, New York, 1979. [7] J. A. Fax and R. M. Murray. Information flow and cooperative control of vehicle formations. IEEE Transactions on Automatic Control, 49(9):1465–1476, September 2004. ∆V (y, ετ, t) T = [Φ(t, t + ετ )y(t)] Q(t + ετ ) [Φ(t, t + ετ )y(t)] [8] E. Fiorelli, P. Bhatta, N. E. Leonard, and I. Shulman. Adaptive sampling using feedback control of an autonomous underwater glider fleet. In Proceedings of the 13th International Unmanned Untethered Submersible Technology Symposium, pages 1–16, Durham, NH, August 2003. − y T (t)Q(t)y(t) £ ¤ = y T (t) Φ̄T (t, t + ετ )Q(t + ετ )Φ̄(t, t + ετ )−Q(t) y(t) £ + y T (t) Φ̄T (t, t + ετ )Q(t + ετ )∆Φ(t, t + ετ ) [9] C. Godsil and G. Royle. Algebraic Graph Theory. SpringerVerlag, New York, NY, 2001. + ∆Φ(t, t + ετ )Q(t + ετ )Φ̄T (t, t + ετ ) ¤ + ∆ΦT (t, t + ετ )Q(t + ετ )∆Φ(t, t + ετ ) y(t) [10] R. A. Horn and C. R. Johnson. Matrix Analysis. Cambridge University Press, Cambridge, UK, 1985. Following the arguments in [29] for bounding each summand, and using (37), we obtain [11] D. Hristu and K. Morgansen. Limited communication control. Systems & Control Letters, 37:193–205, 1999. [12] S. Hubbard, P. Babak, S. T. Sigurdsson, and K. G. Magnússon. A model of the formation of fish schools and migrations of fish. Ecological Modelling, 174(4):359–374, June 2004. µ ¶ ¶ h µ µετ ∆V (y, t + ετ, t) ≤ ρ exp − −1 (38) ρ µr µ ¶¶ i ρ µετ +2 exp − ρf (ε) + ρ2 f 2 (ε) ky(t)k2 η 2ρ [13] H. Ishii and B. A. Francis. Stabilization with control networks. Automatica, 38(10):1745–1751, 2002. [14] A. Jadbabaie, J. Lin, and A. S. Morse. Coordination of groups of mobile autonomous agents using nearest neighbor rules. IEEE Transactions on Automatic Control, 48(6):988– 1001, June 2003. Defining the function g(ε, z) to be the right hand side of (38), it can be shown that g(0, z) = 0 and ∂g ∂ε |ε=0 = −µτ ky(t)k2 . Thus, since g(ε, z) → +∞ as ε → ∞ and g(ε, z) is a continuous function of ε, there exists a ε∗ such that g(ε∗ , z) = 0 and g(ε, z) < 0 for any ε < ε∗ . Thus ∆V (y, t + ετ, t) is negative definite for any ε < ε∗ . 2 [15] M. Kumagai, T. Ura, Y. Kuroda, and R. Walker. A new autonomous underwater vehicle designed for lake environment monitoring. Advanced Robotics, 16(1):17–26, January 2002. [16] W. Li, J. A. Farrell, S. Pang, and R. M. Arrieta. Moth-inspired chemical plume tracing on an autonomous underwater vehicle. IEEE Transactions on Robotics, 22(2):292–307, April 2006. Acknowledgment [17] D. Liberzon. Switching in Systems and Control. Birkhauser, Boston, MA, 2003. The authors gratefully acknowledge the support of the National Science Foundation via grant IIS-0238092, and the Office of Naval Research via grant N00014-03-1-0444. [18] S. A. Matveev AS. Optimal design of networked control systems: computer control via asynchronous communication channels. International Journal of Control, 77(16):1426– 1437, 2004. References [19] M. Mesbahi. On state-dependent dynamic graphs and their controllability properties. IEEE Transactions on Automatic Control, 50(3):387–392, March 2005. [1] R. Bachmayer, N. E. Leonard, J. Graver, E. Fiorelli, P. Bhatta, and D. Paley. Underwater gliders: Recent developments and future applications. In International Symposium on Underwater Technology, pages 195–200, 2004. [20] M. Mesbahi and F. Y. Hadaegh. Formation flying control of multiple spacecraft via graphs, matrix inequalities, and switching. Journal of Guidance, Control, and Dynamics, 24(2):369–377, March-April 2001. 11 [21] L. Moreau. Stability of multiagent systems with timedependent communication links. IEEE Transactions on Automatic Control, 50(2):169–182, February 2005. [22] P. Ögren, E. Fiorelli, and N. E. Leonard. Cooperative control of mobile sensor networks: Adaptive gradient climbing in a distributed environment. IEEE Transactions on Automatic Control, 49(8):1292–1302, August 2004. [23] R. Olfati-Saber. Flocking for multi-agent dynamic systems: Algorithms and theory. IEEE Transactions on Automatic Control, 51(3):401–420, march 2006. [24] R. Olfati-Saber and R. M. Murray. Consensus problems in networks of agents with switching topology and time-delays. IEEE Transactions on Automatic Control, 49(9):1520–1533, September 2004. [25] W. Ren and E. Atkins. Second-order consensus protocols in multiple vehicle systems with local interactions. In AIAA Guidance, Navigation, and Control Conference, San Francisco, CA, 2005. [26] W. Ren and R. W. Beard. Consensus seeking in multiagent systems under dynamically changing interaction topologies. IEEE Transactions on Automatic Control, 50(5):655–661, May 2005. [27] C. W. Reynolds. Flocks, herds, and schools: A distributed behavioral model. Computer Graphics, 21:35–34, 1987. [28] W. J. Rugh. Linear System Theory. Prentice Hall, London, 2nd edition, 1996. [29] D. J. Stilwell, E. M. Bollt, and D. G. Roberson. Sufficient conditions for fast switching synchronization in time varying network topologies. SIAM Journal on Applied Dynamical Systems, 5(1):140–156, 2006. [30] D. J. T. Sumpter and S. C. Pratt. A modelling framework for understanding social insect foraging. Behavioral Ecology and Sociobiology, 53(3):131–144, February 2003. [31] Z. Sun and S. S. Ge. Switched Linear Systems: Control and Design. Springer, London, 2005. [32] J. Toner and Y. Tu. Flocks, herds, and schools: A quantitative theory of flocking. Physical Review E, 58(4):4828–4858, 1998. [33] W. S. Wong and R. W. Brockett. Systems with finite communication bandwidth constraints .1. state estimation problems. IEEE Transactions on Automatic Control, 42(9):1294–1299, 1997. [34] H. Yamaguchi. A distributed motion coordination strategy for multiple nonholonomic mobile robots in cooperative hunting operations. Robotics and Autonomous Systems, 43:257–282, 2003. 12
© Copyright 2025 Paperzz