1534 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 58, NO. 6, JUNE 2013 [12] M. Fliess, C. Join, and M. Mboup, “Algebraic change-point detection,” Applicable Algebra Eng., Commun., Comp., vol. 21, no. 2, pp. 131–143, 2010. [13] E. Fridman, F. Gouaisbaut, M. Dambrine, and J.-P. Richard, “A descriptor approach to sliding mode control of systems with time-varying delays,” Int. J. Syst. Sci., vol. 34, no. 8–9, pp. 553–559, 2003. [14] E. Fridman, “Stability of systems with uncertain delays: A new ‘complete’ Lyapunov-Krasovskii functional,” IEEE Trans. Autom. Control, vol. 51, no. 5, pp. 885–890, May 2006. [15] J.-E. Fonde, “Delay Differential Equation Models in Mathematical Biology,” Ph.D. dissertation, Univ. Michigan, Dearborn, 2005. [16] J. P. Gauthier, H. Hammouri, and S. Othman, “A simple observer for nonlinear systems. Applications to bioreactors,” IEEE Trans. Autom. Control, vol. 37, no. 6, pp. 875–880, Jun. 1992. [17] M. Ghanes, J. B. Barbot, J. DeLeon, and A. Glumineau, “A robust output feedback controller of the induction motor drives: New design and experimental validation,” Int. J. Control, vol. 83, no. 3, pp. 484–497, 2010. [18] A. Germani, C. Manes, and P. Pepe, “An asymptotic state observer for a class of nonlinear delay systems,” Kybernetika, vol. 37, no. 4, pp. 459–478, 2001. [19] M. Hou and R. T. Patton, “An observer design for linear time-delay systems,” IEEE Trans. Autom. Control, vol. 47, no. 1, pp. 121–125, Jan. 2002. [20] S. Ibrir, “Adaptive observers for time delay nonlinear systems in triangular form,” Automatica, vol. 45, no. 10, pp. 2392–2399, 2009. [21] V. L. Kharitonov and D. Hinrichsen, “Exponential estimates for time delay systems,” Syst. Control Lett., vol. 53, no. 5, pp. 395–405, 2004. [22] N. N. Krasovskii, “On the analytical construction of an optimal control in a system with time lags,” J. Appl. Math. Mech., vol. 26, no. 1, pp. 50–67, 1962. [23] V. Laskhmikanthan, S. Leela, and A. Martynyuk, “Practical stability of nonlinear systems,” in Proc. Word Scientific, 1990, [CD ROM]. [24] N. MacDonald, “Time lags in biological models,” in Lecture Notes in Biomath. New York: Springer, 1978. [25] L. A. Marquez-Martinez, C. H. Moog, and V. V. Martin, “Observability and observers for nonlinear systems with time delays,” Kybernetika, vol. 38, no. 4, pp. 445–456, 2002. [26] H. Mounier and J. Rudolph, “Flatness based control of nonlinear delay systems: A chemical reactor example,” Int. J. Control, vol. 71, no. 5, pp. 871–890, 1998. [27] K. Natori and K. Ohnishi, “A design method of communication disturbance observer for time-delay compensation, taking the dynamic property of network disturbance into account,” IEEE T. Indust. Electron., vol. 55, no. 5, pp. 2152–2168, 2008. [28] S.-I. Niculescu, C.-E. de Souza, L. Dugard, and J.-M. Dion, “Robust exponential Stability of uncertain systems with time-varying delays,” IEEE Trans. Autom. Control, vol. 43, no. 5, pp. 743–748, May 1998. [29] S.-I. Niculescu, Delay Effects on Stability: A Robust Control Approach. New York: Springer LNCIS, 2001. [30] P. Picard, O. Sename, and J. F. Lafay, “Observers and observability indices for linear systems with delays,” in Proc. IEEE Conf. Computat. Eng. Syst. Appl. (CESA 96) , 1996, vol. 1, pp. 81–86. [31] J.-P. Richard, “Time-delay systems: An overview of some recent advances and open problems,” Automatica, vol. 39, no. 10, pp. 1667–1694, 2003. [32] J.-P. Richard, F. Gouaisbaut, and W. Perruquetti, “Sliding mode control in the presence of delay,” Kybernetica, vol. 37, no. 4, pp. 277–294, 2001. [33] O. Sename, “New trends in design of observers for time-delay systems,” Kybernetica, vol. 37, no. 4, pp. 427–458, 2001. [34] O. Sename and C. Briat, “H1 observer design for uncertain time-delay systems,” in Proc. IEEE ECC’07, 2007, pp. 5123–4130. [35] A. Seuret, T. Floquet, J.-P. Richard, and S. K. Spurgeon, “A sliding mode observer for linear systems with unknown time-varying delay,” in Proc. IEEE ACC’07, 2007, pp. 4558–4563. [36] A. Seuret, T. Floquet, J.-P. Richard, and S. K. Spurgeon, Topics in Time-Delay Systems: Analysis, Algorithms and Control. Berlin, Germany: Springer Verlag, 2008. [37] E. Shustin, L. Fridman, E. Fridman, and F. Castaos, “Robust semiglobal stabilization of the second order system by relay feedback with an uncertain variable time delay,” SIAM J. Control Optim., vol. 47, no. 1, pp. 196–217, 2008. [38] R. Villafuerte, S. Mondie, and Z. Poznyak, “Practical stability of time delay systems: LMI’s approach,” in Proc. IEEE CDC, 2008, pp. 4807–4812. [39] Z. Wang, B. Huang, and H. Unbehausen, “Robust H1 observer design for uncertain time-delay systems: (I) the continuous case,” in Proc. IFAC 14th World Congress, Beijing, China, 1999, pp. 231–236. [40] J. Zhang, X. Xia, and C. H. Moog, “Parameter identifiability of nonlinear systems with time-delay,” IEEE Trans. Autom. Control, vol. 47, no. 2, pp. 371–375, Feb. 2006. [41] G. Zheng, J. P. Barbot, D. Boutat, T. Floquet, and J. P. Richard, “On obserability of nonlinear time-delay systems with unknown inputs,” IEEE Trans. Autom. Control, vol. 56, no. 8, pp. 1973–1978, Aug. 2011. An Approximate Dual Subgradient Algorithm for Multi-Agent Non-Convex Optimization Minghui Zhu and Sonia Martínez Abstract—We consider a multi-agent optimization problem where agents subject to local, intermittent interactions aim to minimize a sum of local objective functions subject to a global inequality constraint and a global state constraint set. In contrast to previous work, we do not require that the objective, constraint functions, and state constraint sets are convex. In order to deal with time-varying network topologies satisfying a standard connectivity assumption, we resort to consensus algorithm techniques and the Lagrangian duality method. We slightly relax the requirement of exact consensus, and propose a distributed approximate dual subgradient algorithm to enable agents to asymptotically converge to a pair of primal-dual solutions to an approximate problem. To guarantee convergence, we assume that the Slater’s condition is satisfied and the optimal solution set of the dual limit is singleton. We implement our algorithm over a source localization problem and compare the performance with existing algorithms. Index Terms—Dual subgradient algorithm, Lagrangian duality. I. INTRODUCTION Recent advances in computation, communication, sensing and actuation have stimulated an intensive research in networked multi-agent systems. In the systems and controls community, this has translated into how to solve global control problems, expressed by global objective functions, by means of local agent actions. Problems considered include multi-agent consensus or agreement [11], [17], coverage control [4], formation control [7], [21] and sensor fusion [24]. The seminal work [2] provides a framework to tackle optimizing a global objective function among different processors where each processor knows the global objective function. In multi-agent environments, a problem of focus is to minimize a sum of local objective functions by a group of agents, where each function depends on a common global decision vector and is only known to a specific agent. This problem is motivated by others in distributed estimation [16], [23], distributed source localization [20], and network utility maximization [12]. More recently, consensus techniques have been proposed to address the issues of switching topologies, asynchronous computation and coupling in objective functions; see for instance [14], [15], [27]. More specifically, the paper [14] presents the first Manuscript received October 14, 2010; revised January 25, 2012; accepted October 08, 2012. Date of publication November 16, 2012; date of current version May 20, 2013. This work was supported by the NSF CAREER Award CMMI-0643673. Recommended by Associate Editor E. K. P. Chong. M. Zhu is with the Laboratory for Information and Decision Systems, Massachusetts Institute of Technology, Cambridge MA 02139 USA (e-mail: [email protected]). S. Martínez is with the Department of Mechanical and Aerospace Engineering, University of California, San Diego, La Jolla, CA 92093 USA (e-mail: [email protected]). Digital Object Identifier 10.1109/TAC.2012.2228038 0018-9286/$31.00 © 2012 IEEE IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 58, NO. 6, JUNE 2013 analysis of an algorithm that combines average consensus schemes with subgradient methods. Using projection in the algorithm of [14], the authors in [15] further address a more general scenario that takes local state constraint sets into account. Further, in [27] we develop two distributed primal-dual subgradient algorithms, which are based on saddle-point theorems, to analyze a more general situation that incorporates global inequality and equality constraints. The aforementioned algorithms are extensions of classic (primal or primal-dual) subgradient methods which generalize gradient-based methods to minimize non-smooth functions. This requires the optimization problems of interest to be convex in order to determine a global optimum. The focus of the current technical note is to relax the convexity assumption in [27]. In order to deal with all aspects of our multi-agent setting, our method integrates Lagrangian dualization, subgradient schemes, and average consensus algorithms. Distributed function computation by a group of anonymous agents interacting intermittently can be done via agreement algorithms [4]. However, agreement algorithms are essentially convex, and so we are led to the investigation of nonconvex optimization solutions via dualization. The techniques of dualization and subgradient schemes have been popular and efficient approaches to solve both convex programs (e.g., in [3]) and nonconvex programs (e.g., in [5], [6]). Statement of Contributions: Here, we investigate a multi-agent optimization problem where agents desire to agree upon a global decision vector minimizing the sum of local objective functions in the presence of a global inequality constraint and a global state constraint set. Agent interactions are changing with time. The objective, constraint functions, as well as the state-constraint set, can be nonconvex. To deal with both nonconvexity and time-varying interactions, we first define an approximate problem where the exact consensus is slightly relaxed. We then propose a distributed dual subgradient algorithm to solve it, where the update rule for local dual estimates combines a dual subgradient scheme with average consensus algorithms, and local primal estimates are generated from local dual optimal solution sets. This algorithm is shown to asymptotically converge to a pair of primal-dual solutions to the approximate problem under the following assumptions: firstly, the Slater’s condition is satisfied; secondly, the optimal solution set of the dual limit is singleton; thirdly, dynamically changing network topologies satisfy some standard connectivity condition. A conference version of this manuscript was published in [26], and an enlarged archived version of this paper is [25]. Main differences are the following: i) by assuming that the optimal solution set of the dual limit is a singleton, and changing the update rule in the dual estimates, we are able to determine a global solution in contrast to an approximate solution in [26]; ii) we present a simple criterion to check the new sufficient condition for nonconvex quadratic programs; iii) new simulation results of our algorithm on a source localization example and a comparison of its performance with existing algorithms are performed. Due to space limitations, details of the technical proofs and simulations can be found in [25]. II. PROBLEM FORMULATION AND PRELIMINARIES Consider a networked multi-agent system where agents are labeled . The multi-agent system operates in a synby , and its topology will be chronous way at time instants , represented by a directed weighted graph . Here, is the adjacency matrix, for is the weight assigned to the edge where the scalar pointing from agent to agent , and is the set of edges with non-zero weights. The set of in-neighbors of agent at time is denoted by . Similarly, we define the set of out-neighbors of agent at time as 1535 . We here make the following assumptions on communication graphs: Assumption 2.1 (Non-Degeneracy): There exists a constant such that , and , for , satisfies , for all . Assumption 2.2 (Balanced Communication): It holds that for all and , and for and . all Assumption 2.3 (Periodical Strong Connectivity): There is a such that, for all , the directed graph positive integer is strongly connected. The above network model is standard to characterize a networked multi-agent system, and has been widely used in the analysis of average consensus algorithms; e.g. see [17], and distributed optimization in [15], [27]. Recently, an algorithm is given in [9] which allows agents to construct a balanced graph out of a non-balanced one under certain assumptions. The objective of the agents is to cooperatively solve the : following primal problem (1) where is the global decision vector. The function is only known to agent , continuous, and referred to as the objec, the state constraint set, is tive function of agent . The set are continuous, and the incompact. The function is understood component-wise; i.e., , for equality , and represents a global inequality constraint. We all and . We will denote . will assume that the set of feasible points is non-empty; i.e., is Since is compact and is closed, then we can deduce that compact. The continuity of follows from that of . In this way, the is finite and , the set of primal optimal value of the problem optimal solutionss, is non-empty. We will also assume the following Slater’s condition holds: Assumption 2.4 (Slater’s Condition): There exists a vector such that . Such is referred to as a Slater vector of the . problem Remark 2.1: All the agents can agree upon a common Slater vector through a maximum-consensus scheme. This can be easily implemented as part of an initialization step, and thus the assumption that the Slater vector is known to all agents does not limit the applicability of our algorithm; see [25] for an algorithm solving this problem. In [27], in order to solvethe convex caseof the problem (i.e.; and are convex functions and is a convex set), we propose two distributed primal-dual subgradient algorithms where primal (resp. dual) estimates move along subgradients (resp. supergradients) and are projected onto convex sets. The absence of convexity impedes the use of the algorithms in [27] since, on the one hand, (primal) gradient-based algorithms are easily trapped in local minima; on the other hand, projection maps may not be well-defined when (primal) state constraint sets are nonconvex. In the sequel, we will employ Lagrangian dualization, subgradient methods and average consensus schemes to design a distributed algorithm which . can find an approximate solution to the problem Towards this end, we construct a directed cyclic graph where . We assume that each agent has a unique in-neighbor (and out-neighbor). The out-neighbor (resp. in-neighbor) (resp. ). With the graph , we will of agent is denoted by : study the following approximate problem of problem (2) 1536 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 58, NO. 6, JUNE 2013 where , with a small positive scalar, and is the column , and vector of ones. Problem (2) provides an approximation to . In particular, the approximate will be referred to as problem when . Its optimal value problem (2) reduces to the problem and , respecand the set of optimal solutions will be denoted by , is finite and . tively. Similarly to the problem can be replaced by any strongly Remark 2.2: The cyclic graph connected graph . Given , each agent is endowed with two inand , for equality constraints: each out-neighbor . This set of inequalities implies that any feasible of problem satisfies the approximate consolution . For simplicity, we will use sensus; i.e., , with a minimum number of constraints, as the the cyclic graph initial graph. A. Dual Problems Before introducing dual problems, let us denote by , , and . The dual problem associated with is given by , , , where and is given as , is not separable since . depends on B. Dual Solution Sets The Slater’s condition ensures the boundedness of dual solution sets for convex optimization; e.g., [10], [13]. We will shortly see that the Slater’s condition plays the same role in nonconvex optimization. To as achieve this, we define the function follows: Let be a Slater vector for problem . Then with is a Slater vector of the problem . Similarly to (3) and (4) in [27], which employ Lemma 3.2 of [27], we have that for any , it holds that (5) (3) where Here, the dual function Note that neighbor’s multipliers where in (5). This leads to the upper bound on . Let be zero . (6) is the Lagrangian function where note can be computed locally. We de- (7) We denote the dual optimal value of the problem by and . We endow each agent with the set of dual optimal solutions by and the local dual the local Lagrangian function defined by function In the approximate problem , the introduction of , renders the and separable. As a result, the global dual function can be decomposed into a simple sum of the local dual functions . More precisely, the following holds: Since and are continuous and is compact, then is continuous; see Theorem 1.4.16 in [1]. Similarly, is continuous. Since is bounded, then . in the Remark 2.3: The requirement of exact agreement on is slightly relaxed in the problem by introducing a problem . In this way, the global dual function is a sum of the small , as in (4); is non-empty and uniformly local dual functions bounded. These two properties play important roles in the devise of our subsequent algorithm. C. Other Notation Define the set-valued map ; i.e., given , the set lutions to the following local optimization problem: as are the so(8) Notice that in the sum of , each for appears in two terms: one is , and the any . With this observation, we regroup the other is terms in the summation in terms of , and have the following: (4) Here, is referred to as the marginal map of agent . Since is comare continuous, then in (8) for any . In pact and the algorithm we will develop in next section, each agent is required to obtain one (globally) optimal solution and the optimal value the local optimization problem (8) at each iterate. We assume that this can be , or and easily solved, and this is the case for problems of being smooth (the extremum candidates are the critical points of the objective function and isolated corners of the boundaries of the constraint regions) or having some specific structure which allows the use of global optimization methods such as branch and bound algorithms. , we define the distance between a point In the space to a set as , as and the Hausdorff distance between two sets . IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 58, NO. 6, JUNE 2013 We denote by and where . III. DADS ALGORITHM In this section, we devise a distributed approximate dual subgradient algorithm which aims to find a pair of primal-dual solutions to the ap. proximate problem For each agent , let be the estimate of the primal soluat time tion to the approximate problem be the estimate of the multiplier on the inequality constraint (resp. )1 be the estimate of the multiplier associated with the collection of the local (resp. ), inequality constraints . We let , for all to be the collection of dual estimates of agent . And for where denote and are convex combinations of dual estimates of agent and its neighbors at time . At time , we associate each agent a supergradient vector defined as , where , has components , , and for , while the components of are given by: , , and , for . For each agent , we define the set for some where . Let is to be the projection onto the set . It is easy to check that is well-defined. closed and convex, and thus the projection map The Distributed Approximate Dual Subgradient (DADS) Algorithm is described in Table I. Algorithm 1 The DADS Algorithm Initialization: Initially, all the agents agree upon some in the . Each agent chooses a common Slater approximate problem and obtains through a vector , computes is given in (7). After that, each max-consensus algorithm where and . agent chooses initial states Iteration: Every , each agent executes the following steps: , given , solve the local optimization 1) For each and the dual problem (8), obtain a solution . optimal value , generate the dual estimate according 2) For each to the following rule: (9) where the scalar 3) Repeat for is a step-size. . Remark 3.1: The DADS algorithm is an extension of the classical dual algorithm, e.g., in [19] and [3] to the multi-agent setting and nonconvex case. In the initialization of the DADS algorithm, the value serves as an upper bound on . In Step 1, one solution in is needed, and it is unnecessary to compute the whole set . To assure primal convergence, we assume that dual estimates converge to the set where each has a single optimal solution. 1We will use the superscript to indicate that of some global variables. and are estimates 1537 Definition 3.1 (Singleton Optimal Dual Solution Set): The set of is the set of such that the set is a singleton, for each . where The primal and dual estimates in the DADS algorithm converge to . We a pair of primal-dual solutions to the approximate problem formally state this in the following theorem: Theorem 3.1: (Convergence Properties of the DADS Algorithm): and the corresponding approximate problem Consider the problem with some . We let the non-degeneracy assumption 2.1, the balanced communication assumption 2.2 and the periodic strong connectivity assumption 2.3 hold. In addition, suppose the Slater’s con. Consider the dual sequences of dition 2.4 holds for the problem , , and the primal sequence of of the satdistributed approximate dual subgradient algorithm with , , isfying . 1) (Dual estimate convergence) There exists a dual solution where and such that the following holds for all 2) (Primal estimate convergence) If the dual solution satisfies , i.e. is a singleton for all , then there is such that, for all IV. DISCUSSION Before outlining the technical proofs for Theorem 3.1, we would like to make the following observations. First, our methodology is motivated by the need of solving a nonconvex problem in a distributed way by a group of agents whose interactions change with time. This places a number of restrictions on the solutions that one can find. Timevarying interactions of anonymous agents can be currently solved via agreement algorithms; however these are inherently convex operations, which does not work well in nonconvex settings. To overcome this, one can resort to dualization. Admittedly, zero duality gap does not hold in general for nonconvex problems. A possibility would be to resort to nonlinear augmented Lagrangians, for which strong duality holds in a broad class of programs [5], [6], [22]. However, we find here another problem, as a distributed solution using agreement requires separability, as the one ensured by the linear Lagrangians we use here. Thus, we have looked for alternative assumptions that can be easier to check and allow the dualization approach to work. More precisely, Theorem 3.1 shows that dual estimates always converge to a dual optimal solution. The convergence of primal estimates requires an additional assumption that the dual limit has a single optimal solution. We refer to this assumption as the singleton dual optimal solution set (SD for short). The assumption may not be easy to check a priori, however it is of similar nature as existing algorithms for nonconvex optimization. In [5] and [6], subgradient methods are defined in terms of (nonlinear) augmented Lagrangians, and it is shown that every accumulation point of the primal sequence is a primal solution provided that the dual function is required to be differentiable at the dual limit. An open question is how to resolve the above issues imposed by the multi-agent setting with less stringent conditions on the nature of the nonconvex optimization problem. In the following, we study a class of nonconvex quadratic programs for which a sufficient condition guarantees that the SD assumption holds. Nonconvex quadratic programs hold great importance from both 1538 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 58, NO. 6, JUNE 2013 theoretic and practical aspects. In general, nonconvex quadratic programs are NP-hard, and please refer to [18] for detailed discussion. The aforementioned sufficient condition only requires checking the positive definiteness of a matrix. Consider the following nonconvex quadratic program: (10) and are real and symmetric matrices. where is given by The approximate problem of Lemma 5.2 (Lipschitz Continuity of ): There is a constant such that for any , it holds that . In the DADS algorithm, the error induced by the projection map is given by A basic iterate relation of dual estimates in the DADS algorithm is the following. Lemma 5.3 (Basic Iterate Relation): Under the assumptions in Thewith for all , orem 3.1, for any we have for all (11) as before. The local LaWe introduce the dual multipliers can be written as follows: grangian function where the term independent of is dropped and is a linear function . The dual function and dual problem can be defined of : as before. Consider any dual optimal solution . If for all is positive definite; (P1) (P2) ; then the SD assumption holds. The properties (P1) and (P2) are easy is obtained. We to verify in a distributed way once a dual solution would like to remark that (P1) is used in [8] to determine the unique global optimal solution via canonical duality when is absent. V. CONVERGENCE ANALYSIS This section outlines the analysis steps to prove Theorem 3.1. Please refer to [25] for more details. Recall that is continuous and is comsuch that and pact. Then there are for all . We start our analysis from the computation of supergradients of . , then Lemma 5.1 (Supergradient Computation): If is a supergradient of at . That is, for any (12) A direct result of Lemma 5.1 is that the vector is a supergradient of ; i.e., the following supergradient inequality holds for any : at (14) Asymptotic convergence of dual estimates is shown next. Lemma 5.4 (Dual Estimate Convergence): Under the assumptions in Theorem 3.1, there exists such that , , and . The remainder of the section is dedicated to characterizing the convergence properties of primal estimates. Lemma 5.5 (Properties of Marginal Maps): The set-valued marginal is closed. In addition, it is upper semicontinuous at ; map , there is such that for any , i.e., for any . it holds that Lemma 5.6 (Primal Estimate Convergence): Under the assumptions , it holds that in Theorem 3.1, for each where . The main result of this technical note, Theorem 3.1, can be shown next. In particular, we will show complementary slackness, primal feasibility of , and its primal optimality, respectively. Proof for Theorem 3.1: , Claim 1: and . Proof: Rearranging the terms related to in (14) leads to the folwith lowing inequality holding for any for all : (13) It can be seen that (9) of dual estimates in the DADS algorithm is a combination of a dual subgradient scheme and average consensus is Lipschitz continuous algorithms. The following establishes that with some Lipschitz constant . (15) IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 58, NO. 6, JUNE 2013 Summing (15) over [0, ], and dividing by (16) We now proceed to show for each . , for and , in Let is not summable but square summable, and (16). Recall that is uniformly bounded. Take , and then it follows from Lemma 5.1 in [27] that: (17) On the other hand, since , we have given the . Let where fact that is an upper bound of . Then we could choose a sufficiently small and in (16) such that where is given in the definition of and is given by: , for . Following the same lines toward (17), it gives that . Hence, it holds that . The rest of the proof is analogous and thus omitted. The proofs of the following two claims can be found in [25]. Claim 2: is a primal feasible solution to the approximate problem . . Claim 3: is a primal solution to the problem VI. CONCLUSION We have studied a distributed dual algorithm for a class of multiagent nonconvex optimization problems. The convergence of the algorithm has been proven under the assumptions that i) the Slater’s condition holds; ii) the optimal solution set of the dual limit is singleton; iii) the network topologies are strongly connected over any given bounded period. An open question is how to address the shortcomings imposed by nonconvexity and multi-agent interactions settings. REFERENCES [1] J. P. Aubin and H. Frankowska, Set-Valued Analysis. Boston, MA: Birkhäuser, 1990. [2] D. P. Bertsekas and J. N. Tsitsiklis, Parallel and Distributed Computation: Numerical Methods. Boston, MA: Athena Scientific, 1997. [3] D. P. Bertsekas, Convex Optimization Theory. Boston, MA: Anthena Scietific, 2009. [4] F. Bullo, J. Cortés, and S. Martínez, Distributed Control of Robotic Networks, ser. Applied Mathematics Series. Princeton, NJ: Princeton Univ. Press, 2009 [Online]. Available: http://www.coordinationbook. info [5] R. S. Burachik, “On primal convergence for augmented Lagrangian duality,” Optimization, vol. 60, no. 8, pp. 979–990, 2011. [6] R. S. Burachik and C. Y. Kaya, “An update rule and a convergence result for a penalty function method,” J. Ind. Manag. Optim., vol. 3, no. 2, pp. 381–398, 2007. 1539 [7] J. A. Fax and R. M. Murray, “Information flow and cooperative control of vehicle formations,” IEEE Trans. Autom. Control, vol. 49, no. 9, pp. 1465–1476, Sep. 2004. [8] D. Y. Gao, N. Ruan, and H. Sherali, “Solutions and optimality criteria for nonconvex constrained global optimization problems with connections between canonical and Lagrangian duality,” J. Global Optim., vol. 45, no. 3, pp. 474–497, 2009. [9] B. Gharesifard and J. Cortés, “Distributed strategies for generating weight-balanced and doubly stochastic digraphs,” Eur. J. Control, to be published. [10] J.-B. Hiriart-Urruty and C. Lemaréchal, Convex Analysis and Minimization Algorithms: Part 1: Fundamentals. New York: Springer, 1996. [11] A. Jadbabaie, J. Lin, and A. S. Morse, “Coordination of groups of mobile autonomous agents using nearest neighbor rules,” IEEE Trans. Autom. Control, vol. 48, no. 6, pp. 988–1001, Jun. 2003. [12] F. P. Kelly, A. Maulloo, and D. Tan, “Rate control in communication networks: Shadow prices, proportional fairness and stability,” J. Oper. Res. Soc., vol. 49, no. 3, pp. 237–252, 1998. [13] A. Nedic and A. Ozdaglar, “Approximate primal solutions and rate analysis for dual subgradient methods,” SIAM J. Optim., vol. 19, no. 4, pp. 1757–1780, 2009. [14] A. Nedic and A. Ozdaglar, “Distributed subgradient methods for multiagent optimization,” IEEE Trans. Autom. Control, vol. 54, no. 1, pp. 48–61, 2009. [15] A. Nedic, A. Ozdaglar, and P. A. Parrilo, “Constrained consensus and optimization in multi-agent networks,” IEEE Trans. Autom. Control, vol. 55, no. 4, pp. 922–938, Apr. 2010. [16] R. D. Nowak, “Distributed EM algorithms for density estimation and clustering in sensor networks,” IEEE Trans. Signal Processing, vol. 51, pp. 2245–2253, 2003. [17] R. Olfati-Saber and R. M. Murray, “Consensus problems in networks of agents with switching topology and time-delays,” IEEE Trans. Autom. Control, vol. 49, no. 9, pp. 1520–1533, Sep. 2004. [18] P. M. Pardalos and S. A. Vavasis, “Quadratic programming with one negative eigenvalue is NP-hard,” J. Global Optim., vol. 1, no. 1, pp. 15–22, 1991. [19] B. T. Polyak, “A general method for solving extremum problems,” Soviet Math. Doklady, vol. 3, no. 8, pp. 593–597, Aug. 1967. [20] M. G. Rabbat and R. D. Nowak, “Decentralized source localization and tracking,” in Proc. IEEE Int. Conf. Acoust., Speech, Signal Processing, May 2004, pp. 921–924. [21] W. Ren and R. W. Beard, Distributed Consensus in Multi-vehicle Cooperative Control, ser. Communications and Control Engineering. New York: Springer, 2008. [22] R. T. Rockafellar and R. J.-B. Wets, Variational Analysis. New York: Springer, 1998. [23] S. Sundhar Ram, A. Nedic, and V. V. Veeravalli, “Distributed and recursive parameter estimation in parametrized linear state-space models,” IEEE Trans. Autom. Control, vol. 55, no. 2, pp. 488–492, Feb. 2010. [24] L. Xiao, S. Boyd, and S. Lall, “A scheme for robust distributed sensor fusion based on average consensus,” in Proc. Symp. Inform. Processing Sensor Networks, Los Angeles, CA, Apr. 2005, pp. 63–70. [25] M. Zhu and Martínez, “An approximate dual subgradient algorithm for multi-agent non-convex optimization,” IEEE Trans. Autom. Control 2013 [Online]. Available: http://arxiv.org/abs/1010.2732 [26] M. Zhu and S. Martínez, “An approximate dual subgradient algorithm for multi-agent non-convex optimization,” in Proc. IEEE Int. Conf. Decision Control, Atlanta, GA, Dec. 2010, pp. 7487–7492. [27] M. Zhu and S. Martínez, “On distributed convex optimization under inequality and equality constraints,” IEEE Trans. Autom. Control, vol. 57, pp. 151–164, 2012.
© Copyright 2026 Paperzz