Multiagent Teamwork: Analyzing the Optimality and Complexity of Key Theories and Models David V. Pynadath and Milind Tambe Information Sciences Institute and Department of Computer Science University of Southern California 1 Agent Teamwork Agents, robots, sensors, spacecraft, etc. Performing a common task Operating in an uncertain environment Distinct, uncertain observations Distinct actions with uncertain effects Limited, costly communication Battlefield Simulation Satellite Clusters Disaster Rescue 2 Motivation Performance ? Optimal New algorithm Optimal ? Practical Systems ? Theoretical Approaches ? No communication Outline of Results 1) Unified teamwork framework 2) Complexity of optimal teamwork 3) New coordination algorithm 4) Optimality-Complexity evaluation of existing methods Complexity 3 Example Domain: Helicopter Team Did they see that? I destroyed the enemy radar. Enemy Radar Goal 4 Communicative Multiagent Team Decision Problem (COM-MTDP) S: states of the world e.g., position of helicopters, position of the enemy A: domain-level actions e.g., fly below radar, fly normal altitude P: transition probability function e.g., world dynamics, effects of actions S: communication capabilities, possible “speech acts” e.g., “I have destroyed enemy radar.” 5 COM-MTDPs (cont’d) W: observations e.g., enemy radar, position of other helicopter O: probability (for each agent) of observation Maps state and actions into distribution over observations (e.g., sensor noise model) R: reward (over states, actions, messages) e.g., good if we reach destination, better if we reach it earlier e.g., saying, “I have destroyed enemy,” has a cost Teamwork Definition: All members share same preferences (i.e., R) 6 Problem Complexity No communication Individually Observable Collectively Observable Free communication COM-MTDPs 7 To Communicate or Not To Communicate Local decision of one agent at a single point in time: “I have achieved a joint goal.” “Should I tell my teammate?” Joint intentions theory: “I must attain mutual belief.” Always communicate [Jennings] STEAM: “I must communicate if the expected cost of miscoordination outweighs the cost of communication.” [Tambe] Each cost is a fixed parameter specified by designer 8 Locally Optimal Criterion for Communication Communicate if and only if: E[R | communicate] E[R | do Expectation over possible histories of and beliefs not states communicate] up to current time Expected reward over future trajectories of states and beliefs WITH communication Expected reward over future trajectories of states and beliefs WITHOUT communication Expected cost of communicating 9 Empirical Results V_opt-V Communication Cost Observability 10 Empirical Results 11 Empirical Results 12 Optimality vs. Complexity Optimality E[R] 1.46 Locally Optimal Globally Optimal STEAM Silent 1.43 1.40 Jennings 0.1 1.0 Observability = 0.2 Comm. Cost = 0.7 Complexity 10,000 seconds (log) 13 Optimality vs. Complexity Optimality E[R] STEAM 1.80 Jennings Locally Optimal 1.43 Observability = 0.2 Comm. Cost = 0.3 Silent 0.1 Globally Optimal 1.0 Complexity 10,000 seconds (log) 14 Summary COM-MTDPs provide a unified framework for agent teamwork Representation subsumes many existing agent models Policy space subsumes many existing prescriptive theories This framework supports deeper analyses of teamwork problems Quantitative characterization of optimality-efficiency tradeoff , for different policies, in different domains Derivation of novel coordination algorithms http://www.isi.edu/teamcore/Teamwork Detailed proofs Source code JAIR article 15
© Copyright 2026 Paperzz