A C o l l a b o r a t i o n M e c h a n i s m on Positive I n t e r a c t i o n s in Multi-agent Environments K e i M a t s u b a y a s h i * and M a r i o T o k o r o + Department of Computer Science, Keio University. 3-14-1 Hiyoshi, Yokohama 223 Japan Abstract In m u l t i - a g e n t e n v i r o n m e n t s where agents independently generate and execute plans to satisfy their goals, the resulting plans may sometimes overlap. In this paper, we propose a c o l l a b o r a t i o n mechanism using social law, t h r o u g h w h i c h r a t i o n a l agents can s m o o t h l y delegate and receive the execution of the overl a p p i n g parts of plans in order to reduce the cost of plan execution. Also, we consider col l a b o r a t i o n w i t h agents t h a t do not abide by social law, t h a t is, self-centered agents. Simu l a t i o n results show t h a t our mechanism also has the p r o p e r t y of balancing t he cost of plan execution and shows f l e x i b i l i t y towards selfcentered agents. 1 Introduction In m u l t i - a g e n t e n v i r o n m e n t s where agents generate and execute plans w h i c h satisfy their o w n goals, plans are generated independently, thus two relationships [ M a r t i a l , 1992] exist among plans. T h e first is called a negative relationship and exists between plans w h i c h a t t e m p t to use the same resource at the same t i m e . T h i s rel a t i o n s h i p w i l l cause conflicts. Such conflicts must be avoided or resolved because they prevent actions f r o m being p r o p e r l y executed. T h e other relation is called a positive relationship and exists between plans w h i c h i n clude o v e r l a p p i n g subplans, subplans consisting of b o t h c o m m o n actions and i n d i v i d u a l actions. W h e n there is a positive relationship and one agent delegates the exec u t i o n of the o v e r l a p p i n g subplan to the other agent, it can reduce the cost of execution. A lot of research has been done on negative relationships, such as conflict avoidance, by using a social law [Shoham and T e n n e n h o l t z , 1992] and conflict resolution based on game theory [ Z l o t k i n and Rosenschein, 1990]. However, despite the i m p o r t a n c e of a positive relationship for t h e agents' effective execution of plans, not much has been made [Foulser et al., 1991] to utilize this, except to reduce costs by deleting the overlapping actions 346 Distributed AI in a d o m a i n consisting of only one agent [Hayes, 1989] and for multi-agent systems where i r r a t i o n a l agents delegate and receive plans [ M a r t i a l , 1992] and for m u l t i - a g e n t systems where agents m u t u a l l y construct a c o l l a b o r a t i v e plan [Osawa and T o k o r o , 199l] . In this paper we propose a c o l l a b o r a t i o n mechanism by which r a t i o n a l agents can delegate and receive subplans in a positive relationship so t h a t the cost of ex ecution is reduced. If an agent delegates its actions to the other agent, it can reduce the execution cost greatly and benefit f r o m i t . On the other h a n d , when an agent receives actions, the cost increases a l i t t l e even t h o u g h the reduction in the cost of the sender is larger t h a n the gain in its o w n cost. As r a t i o n a l agents can delegate actions but may not be w i l l i n g to receive, it is very h a r d to collaborate in a positive relationship. However, if an agent w h i c h receives actions is guaranteed to be able to delegate some actions in the f u t u r e , t h e n even a r a t i o n a l agent can collaborate in a positive r e l a t i o n s h i p . We propose a mechanism based on a social law (the guarantee) in w h i c h agents must balance the costs of received actions. W i t h this mechanism, r a t i o n a l agents can collab orate t h r o u g h the expected cost based on game theory. Moreover, there are t w o other benefits arising f r o m this law: (1) j u s t as agents can expect another agent's p r o p o s i t i o n (strategy) at negotiation on delegation, they can avoid conflicts of strategies and can collaborate s m o o t h l y ; and (2) the macroscopic p r o b l e m t h a t arises when the load of the whole m u l t i - a g e n t system is concentrated on an i r r a t i o n a l agent can be solved w i t h this microscopic cost-balancing law. A l t h o u g h the image related to the benefit of such a law is i m p l i c i t in m a n y AI works, the social law was only recently formalized and discussed [Shoham and T e n nenholtz, 1992]. However, this w o r k does not consider agents w h i c h do not abide by t h e law. W h e n t r e a t i n g a multi-agent e n v i r o n m e n t as an open d i s t r i b u t e d envir o n m e n t , it is q u i t e n a t u r a l to t h i n k t h a t such agents exist. Therefore, our mechanism considers c o l l a b o r a t i o n w i t h self-centered agents. It is a very simple s o l u t i o n for agents to not collaborate w i t h self-centered agents. Nevertheless, it is problematic t h a t an agent w h i c h except i o n a l l y d i d not abide by the law for some reason, cannot collaborate any more, and t h a t agents could benefit by c o l l a b o r a t i n g w i t h the self-centered agents. T h e r e f o r e , agents should collaborate in a more r a t i o n a l and fiexi- ble way: they collaborate if they can benefit and do not collaborate when they cannot. O u r mechanism allows agents to collaborate in this way. S i m u l a t i o n results show t h a t w i t h our mechanism, r a t i o n a l agents can reduce their cost by delegati n g / r e c e i v i n g actions. T h e results also show t h a t the mechanism has the i m p o r t a n t properties of flexibility and guarantee cost-balancing. T h e o u t l i n e of the rest of the paper is as follows. In the next section, some of the assumptions are presented before discussion. O u r mechanism is proposed in Sections 3 and 4. In Section 3, t w o evaluation functions are defined using t h e game theory. Also, the social law and the m o d e l of agents are described. In Section 4, the other p a r t of our m e c h a n i s m is described. In Section 5, some examples are presented. E x p e r i m e n t a l results are shown and discussed in sections 6 and 7, respectively. 2 3 Definitions 3.1 C h a n g e o f cost b y d e l e g a t i o n T h e change in agent A's subplan, w h i c h has a favor relationship to B's subplan and the associated cost t h r o u g h d e l e g a t i o n / r e c e p t i o n , is shown in Figure 1 where some actions are indicated as circles. For instance, if a subplan of agent A (a) consists of three actions: "go to store" then "buy f o o d " and finally "come back," and t h a t of agent B consists of: "go to store" then " b u y j u i c e " and finally "come back," t h e n the subplans are in a favor relationship ( p a r t l y overlapp i n g ) . If A receives the execution of B's s u b p l a n , the replanned subplan (b) t h a t A m u s t execute is: "go to store" then " b u y f o o d " and " b u y j u i c e " ( B ' s action) and finally "come back" and " h a n d B j u i c e " (action to adjust A's initial plan). Assumptions A s s u m p t i o n s on the agents' plans and actions are as follows: 1. Each agent has a plan to achieve its o w n goal, and doesn't know t h e plans of other agents'. 2. Subplans (plans w h i c h achieve sub goals) consist of sequences of actions. In order to express costs, we use the f o l l o w i n g n o t a tion: 3. T h e meanings of actions and associated costs are predefined and c o m m o n to all agents. • : A ' s cost for e x e c u t i n g a subplan w h i c h has a favor relationship to B's s u b p l a n . In order to concentrate on the a l g o r i t h m of delegat i o n / r e c e p t i o n decision, we assume the f o l l o w i n g : • : T h e cost of A ' s actions when executed by B when delegated (e.g., " b u y f o o d " ) . 4. A g e n t s can find w h i c h parts (subplans) of their plan can be delegated. • : T h e cost of A ' s a d d i t i o n a l actions in the case of d e l e g a t i o n / r e c e p t i o n (e.g., " h a n d B j u i c e " ) . Positive relationship t h a t is i m p l i c i t l y contained between plans is d i v i d e d i n t o t w o relationships 1 [ M a r t i a l , 1992]: a s u b s u m p t i o n relationship and a favor relation If A receives the execution of B's actions, A ' s overall cost is the cost of its o w n actions plus the cost of the delegated actions f r o m B and the cost of a d j u s t i n g a plan to a new p l a n On the other h a n d , if A delegates the execution of its subplan, the subplan w i l l become ( c ) . In this case, A ' s overall cost is only the cost of a d j u s t i n g a p l a n ship. 3.2 5. Agents can discover other agents w i t h which to neg o t i a t e delegation. 6. N e g o t i a t i o n is carried out between t w o agents only. If the actions of an agent's plan are the same or subsume another agent's actions, the plans are in a subs u m p t i o n relationship and one agent does not need to execute its plan if the other agent executes its o w n p l a n . If an agent's plan p a r t i a l l y overlaps w i t h another agent's p l a n and is n o t in a s u b s u m p t i o n relationship, the plans are in a favor r e l a t i o n s h i p . In this case, an agent does not need to execute its plan if another agent is w i l l i n g to receive the plan and to execute its own plan together w i t h t h e n o n - o v e r l a p p i n g p a r t o f the received plan. N o w , we assume the f o l l o w i n g 2 : 7. A g e n t s collaborate when their plans are in a favor relationship. 1 Martial divided the relationships in more detail. I t is not necessary for agents to collaborate in a subsumption relationship because it is enough for an agent to check whether its actions were already done after execution of the other agent's plan. 2 Social law for delegation Social law is formalized in [Shoham and T e n n e n h o l t z , 1992] and is treated as the pair of actions and c o n d i t i o n s about states t h a t p r o h i b i t t a k i n g a c t i o n . I n t u i t i v e l y , social law / is the law by which agents must balance the costs of the individual received actions. Definition 1 (Social law for delegation) l is defined to obey the next constraints. (a) Social law If an agent receives more actions than the other agent, it cannot receive actions at the next negotiation. (b) If an agent delegates more actions than the other agent, it cannot delegate actions at the next negotiation. W h e n agents abide by gate actions and benefits even if it receives actions even a r a t i o n a l agent can this law, each agent can delein the f u t u r e according to (a), and loses by t h a t . Therefore, propose the strategy "receive." Matsubayashi and Tokoro 347 Though this law may not be the only one that makes agents collaborate on positive relationships, the search for such a law is beyond the scope of this paper. However, this law has the following two advantages: (1) as (a) and (b) are exclusive, agents can avoid the conflicts of strategies proposed in the negotiation, or collaborate smoothly, and (2) the microscopic law defined between two agents can balance the agents' costs of executing received actions in a whole system. 3.3 Expected payoff We use payoff in game theory as the increase in cost arising when an agent proposes its strategy at negotiation and defines the expected payoff. The strategy consists of two propositions: "delegate" and "receive" 3 Therefore, agent A's strategy p is a mixed strategy (p 1 , P2) where p1 is A's probability of proposing strategy "delegate" and is that of the reception. Also, the other agent B proposes a mixed strategy expressing the same idea. Please note that the expected payoff is not the expected cost but the expected increase in cost. Because of the existence of social law /, the strategy that is selected by A not only influences the strategy of B in the future, but also influences the cost of A in the future. Hence, the expected payoff is the sum of the expected payoff of current negotiation, present expected payoff (Ep(p, /)) and the negotiation in the future, future expected payoff (EF(P, /)). However, other agents (say, agent B) may not abide by the law. W i t h the reliability of B (Pi(B)), i.e., the probability of B abiding by the law, the expected payoff, is defined as follows: D e f i n i t i o n 2 ( E x p e c t e d p a y o f f of agent A) 3.3.1 W h e n t h e o t h e r agent abides b y t h e law If the other agent, B, abides by the law, A's payoff of current negotiation is expressed in the next payoff matrix. Entries '—' (undefined) represent cases that cannot happen because both agents abide by the law. Table 1: A's payoff (increase in cost) Table 2: A's payoff in the future 3 There is also one other strategy: "do nothing," in which agents must execute the subplan for themselves. This strategy is taken when the expected payoff is not less than 0. 348 Distributed Al Matsubayashi and Tokoro 349 T h e costs for agent B are taken for the same table w i t h B replacing A. However, t h e cost of the subplan (25-30) and the cost of the delegated p a r t of the plan (5-12) was r a n d o m l y selected and was not necessarily common to b o t h agents. T h e results are t h e average of 5000 s i m u lations. 6.1 negotiation is n o t cost-consuming (Cneg = 2), t h e agent continues to collaborate w i t h t h e self-centered agent, because it can reduce its cost more t h a n w h e n d o i n g it alone (i) b u t less t h a n when in s i t u a t i o n ( i i ) . The cost of B Table 3: Cost of actions for agent A R e d u c t i o n o f cost Figure 3 shows the relation between the cost for executing a subplan and the frequency of negotiation for delegation. The cost of A Figure 4: Sum of the cost of the plan received Curves in the figures are the result of the next conditions except for (iii b)). (i) B does not exist (agent A has to execute the actions alone), (ii) B uses our mechanism, (iii a) B is self-centered (B's strategy is only "dele(iii b) (iv) B is self-centered B is the same as (iii a) till the 30-th negotiation, after then, it is the same as (ii). Although the cost doesn't decrease when agent A executes the actions alone (i), A can reduce the cost when the other agent uses our mechanism (ii). When the other agent is self-centered (iii), the cost is at first a little bit more than when the other agent uses our mechanism from the start. This is caused by a decision made by both flipping a coin when strategies collide and usage of the "do nothing" strategy when A cannot benefit from it. However, when the delegation precedes with (a cost-consuming negotiation), as the agent using our mechanism does not come to collaborate with the selfcentered agent, the cost approaches the cost of (i). When 350 Distributed Al In the case where the other agent changes its mind and decides to use our mechanism rather than being self-centered (iv), the same result obtained for the selfcentered agent is obtained at first. If the other agent has decided to use our mechanism, it returns the benefit obtained up till then in order to balance the individual received cost. Therefore, the cost is extremely reduced at that time, and approaches the cost of (ii). Furthermore, we simulated the case when and are different from 15 and 0.07. W i t h a less than 15, result (iii) presents the largest change among all simulations. The agent decides not to collaborate with the selfcentered agent earlier. If is greater than 0.07, result (iv) represents the largest change. The cost when the other agent has come to use our mechanism is smaller. The reason for this is that as larger, larger so that the agent comes to collaborate earlier on the social law. But if is smaller, the agent will propose the "do nothing" strategy so that the "delegate" strategy would be less used and the cost is not reduced. 6.2 Cost b a l a n c i n g of r e c e i v e d a c t i o n s Figure 4 shows the received cost of both agents. Curves (ii)-(iv) in the figure are the same conditions as before. When both agents use our mechanism (ii), the costs received are balanced. When the other agent is selfcentered (iii), the agent receives too many actions because of the flexibility of the mechanism. However, this loss cannot continue to increase because A does not get to participate in the negotiation on delegation if A thinks that it cannot benefit from it. Also, in the case of the self-centered agent which changes its mind and decides to use our mechanism in the course of negotiation, the received costs are finally balanced (iv). 7 Discussion We discuss some issues related to our mechanism. T h e social law w h i c h guarantees t h a t an agent can delegate its actions in the f u t u r e even if it receives the actions of another agent, n o t o n l y enables r a t i o n a l agents to reduce t h e i r cost by c o l l a b o r a t i n g , b u t also enables our m e c h a n i s m to have t w o i m p o r t a n t properties: a g u a r a n tee of balanced costs for i n d i v i d u a l l y received actions a n d an avoidance of conflicts in strategies. As o u r microscopic law, defined between t w o agents, enables t h e agents to balance t h e received costs between t h e m so t h a t t h e costs are macroscopically balanced in a w h o l e m u l t i - a g e n t system. W i t h t h i s law, a n i r r a t i o n a l agent t h a t m a y receive t o o m a n y actions w o u l d not be a b o t t l e n e c k of the whole system if other agents thatdelegated tasks to it w a i t e d u n t i l they finished. W i t h agents A and B a b i d i n g to this law, if A receives more than A proposes "deleg a t e " and B proposes "receive." Therefore, the strategies do not conflict, so t h a t agents can easily reach an agreement. In [Shoham and T e n n e n h o l t z , 1992], social law is used to constrain actions to avoid conflicts. It became unnecessary for agents to negotiate for conflict r e s o l u t i o n . However, agents m u s t negotiate to exchange i n f o r m a t i o n a b o u t plans when they collaborate in a posi t i v e r e l a t i o n s h i p . O u r mechanism used the law to con s t r a i n t h e strategy of n e g o t i a t i o n so as to avoid conflict. Moreover, we considered self-centered agents, or agents t h a t do not abide by the law. T h i s is a case w h i c h is not considered in [Shoharn and T e n n e n h o l t z , 1992]. It is very simple for agents not to collaborate w i t h self centered agents after they realize w h a t k i n d of agents they are dealing w i t h . However, it is p r o b l e m a t i cal t h a t an agent w h i c h exceptionally does not abide by the law for some reason, cannot collaborate any more. A l s o , p r o b l e m a t i c a l is the fact t h a t agents do not col l a b o r a t e w i t h self-centered agents even when benefit can be o b t a i n e d . O u r mechanism is flexible: r e l i a b i l i t y P 1 is g r a d u a l l y decreased every t i m e the strategies conflict, so t h a t t h e chance to collaborate w i t h a real self-centered agent is decreased and agents collaborate only if they can benefit f r o m i t . However, agents w i t h our mechan i s m m a y receive a l i t t l e more action t h a n a self-centered agent because of this flexibility. Here, we consider the s i t u a t i o n where the bank and post office are in the same direction and b o t h agents p l a n to go to b o t h places. Because their plans are "go to b a n k " and "go to post office" and " r e t u r n , " the res u l t i n g plan of our mechanism is t h a t one agent goes to b o t h places. Nevertheless, if the places are in different directions, the r e s u l t i n g plans are that each agent goes to a different place. T h e l a t t e r result is obtained t h r o u g h t w o n e g o t i a t i o n s because their plans are "go to b a n k " and " r e t u r n " and "go t o post office" and " r e t u r n . " O u r m e c h a n i s m also provides such an o p t i m a l s o l u t i o n . In [ Z l o t k i n and Rosenschein, 1990], flipping a coin is used to resolve conflicts, b u t the purpose for our work is to avoid conflict. Since t h e m e n t i o n e d work is designed for c o l l a b o r a t i o n in negative relationships, the conflict in goals m u s t be resolved. On the other h a n d , as our work is directed to p o s i t i v e relationships, the conflicts are not necessarily resolved (in this case, agents cannot reduce costs). In a related piece of research [Martial, 1992], the proposed algorithm focuses on the optimization of plans of agents. Therefore, agents are willing to receive the actions even if they cannot benefit from them, and the received costs would not be balanced. Our mechanism does not require such irrational agents; each agent can rationally and flexibly select its own strategies. 8 Conclusion The utilization of positive relationships has not received much attention in D A I . We proposed a mechanism using a social law by which rational agents can reduce the cost of their execution by delegating and receiving actions. W i t h this law, agents can avoid conflicts of strategies in a negotiation and reach an agreement easily. Furthermore, we considered the case in which agents do not abide by the law (self-centered agents) and the case in which agents that are self-centered at first come to abide by the law eventually. Simulation results show that by using our mechanism, rational and flexible collaboration is possible with these kinds of agents. We intend to use a more complex model of agents, and consider communication costs and collaboration with multiple agents by introducing market mechanisms. Also, we will introduce a mechanism based on incomplete information for more detail, as stated in [Kakehi and Tokoro, 1993]. Acknowledgments We would like to thank T. Hirotsu, R. Kakehi, V. Vasconcelos and I. Satoh for their helpful comments. References [Foulser et al., 1991] D. E. Foulser, M. L i , and Q. Yang. A Quantitative Theory for Plan Merging. In AAAI91, 1991. [Hayes, 1989] C. C. Hayes. A Model of Planning for Plan Efficiency: Taking Advantage of Operator Overlap. In IJCAI-89, 1989. [Kakehi and Tokoro, 1993] R. Kakehi and M. Tokoro. A Negotiation Protocol for Conflict Resolution in Multiagent Environments. In fCICIS '93, 1993. [Martial, 1992] F . V . M a r t i a l . Coordinating Plans of Autonomous Agents. Lecture Notes in Artificial Intelligence, Subseries of LNCS. Springer-Verlag, 1992. [Osawa and Tokoro, 1991] E. Osawa and M. Tokoro. Collaborative Plan Construction for Multiagent Mutual Planning. In MAAMAW '91, 1991. [Shoham an d Tennenholtz, 1992] Y.Shoham and M. Tennenholtz. On the synthesis of useful social laws for artificial agent societies (preliminary report). In AAAI-92, 1992. [Zlotkin and Rosenschein, 1990] G. Zlotkin and J. S. Rosenschein. Negotiation and Conflict Resolution in Non-Cooperative Domains. In AAAI-90, 1990. Matsubayashi and Tokoro 351
© Copyright 2026 Paperzz