1 / 46

Multi-Agent Resource allocation

Multi-Agent Resource allocation. Notes prepared by Ulle Endriss Used with permission. What is MARA?. A tentative definition would be the following: Multiagent Resource Allocation (MARA) is the process of distributing a number of items amongst a number of agents.

neola
Télécharger la présentation

Multi-Agent Resource allocation

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Multi-Agent Resource allocation Notes prepared by Ulle Endriss Used with permission

  2. What is MARA? • A tentative definition would be the following: • Multiagent Resource Allocation (MARA) is the process of distributing a number of items amongst a number of agents. • What kind of items (resources) are being distributed? • How are they being distributed? • And finally, why are they being distributed?

  3. Outline • Concerning the specification of MARA problems: • – Overview of different types of resources • – Representation of the preferences of individual agents (done) • – Notions of social welfare to specify the quality of an • allocation (partly done already) • • Concerning methods for solving MARA problems: • – Discussion of allocation procedures • – Some complexity results concerning allocation procedures • Y. Chevaleyre, P.E. Dunne, U. Endriss, J. Lang, M. Lemaˆıtre, N. Maudet, • J. Padget, S. Phelps, J.A. Rodr´ıguez-Aguilar and P. Sousa. Issues in Multiagent • Resource Allocation. Informatica, 30:3–31, 2006.

  4. Types of Resources • • A central parameter in any resource allocation problem is the nature of the resources themselves. • • There is a whole range of different types of resources, and each of them may require different techniques . . . • • Distinguish properties of the resources themselves and characteristics of the chosen allocation mechanism. Examples: • – Resource-inherent property: Is the resource perishable? • – Characteristic of the allocation mechanism: Can the • resource be shared amongst several agents?

  5. Continuous vs. Discrete Resources • Resources may be continuous (e.g. energy) or discrete • (e.g. fruit). • • Discrete resources are indivisible; continuous resources may be • treated either as (infinitely) divisible or as indivisible (e.g. only • sell orange juice in units of 50 liters ; discretization). • • Representation of a single bundle: – Several continuous resources: vector over non-negative reals – Several discrete resources: vector over non-negative integers – Several distinguishable discrete resources: vector over {0, 1} • • Classical literature in economics mostly concentrates on a • single continuous resource; recent work in AI and Computer • Science focuses on discrete resources.

  6. Divisible or not • Resources may be treated as being divisible or indivisible. • Continuous/discrete: physical property of resources • Divisible/indivisible: feature of the allocation mechanism

  7. Sharable or not • • A sharable resource can be allocated to a number of different agents at the same time. Examples: • – a photo taken by an earth observation satellite • – path in a network (network routing) • • More often though, resources are assumed to be non-sharable and can only have a single owner at a time. Examples: • – energy to power a specific device • – fruit to be eaten by the agent obtaining it

  8. Static or not • Resources that do not change their properties during a negotiation process are called static resources. There are at least two types of resources that are not static: • consumable goods such as fuel • perishable goods such as food • In general, resources cannot be assumed to be static. However, in many cases it is reasonable to assume that they are as far as the negotiation process at hand is concerned.

  9. Single-unit vs. Multi-unit • • In single-unit settings there is exactly one copy of each type of • good; all items are distinguishable (e.g. several houses). • • In multi-unit settings there may be several copies of the same • type of good (e.g. 10 bottles of milk). • • Note that this distinction is only a matter or representation: • Every multi-unit problem can be translated into a single-unit problem by introducing new names for the items • (inefficient, but possible). • Every single-unit problem is in fact also a (degenerate) multi-unit problem. • • Multi-unit problems allow for a more compact representation of allocations and preferences, but also require a richer language (variables ranging over integers, not just binary values).

  10. Resources vs. Tasks • • Tasks may be considered resources with negative utility. • • Hence, task allocation may be regarded a MARA problem. • • However, tasks are often coupled with constraints regarding their timing. • Remark: From now on, we are going to deal with the allocation of static indivisible resources that are available in single units and that cannot be shared . . .

  11. Setting • Set of agents A = {1..n} and finite set of indivisible resources R. • An allocation A is a partitioning of R amongst the agents in A. • Each agent iA has got a utility function ui: 2RR. • We usually write ui(A) as a shorthand for ui(A(i)). • We shall sometimes refer to the preference relation i induced by the utility function ui: R ≤i R’ iff ui(R) ≤ ui(R’). • Remark: MARA with indivisible resources is a prime example for a combinatorial domain.

  12. Social Welfare • An important parameter in the specification of a MARA problem • concerns our goals: what kind of allocation do we want to achieve? • • Success may depend on a single factor (e.g. revenue of an • auctioneer), but more often on an aggregation of preferences • of the individual agents in the system. • • Concepts from social choice theory and welfare economics can • be useful here (“multiagent systems as societies of agents”). • • Here we use the term social welfare in a broad sense, to • describe the quality of an allocation in view of a suitable • aggregation of the individual agent preferences. • Pareto optimality is probably the most basic criterion for social • optimality, but there are many others . . .

  13. Social Welfare Orderings • A collective utility function is a function W : RnR mapping utility vectors to the reals. Here we define them over allocations A (inducing utility vectors): • • The utilitarian social welfare is defined as the sum of utilities: • • The egalitarian social welfare is given by the utility of the agent that is currently worst off: • • The Nash product is the product of the individual utilities:

  14. Social Welfare Orderings (cont.) • • The elitist social welfare is given by the utility of the agent that is currently best off: • • Let uAbe the ordered utility vector induced by allocation A (from lowest to highest). Then the k-rank dictator CUF (collective utility function) swk is defined as the kth element of the vector. When k=1, it is egalitarian. • • The leximin-ordering is a social welfare ordering that may be regarded as a refinement of the egalitarian collective utility function: a is less than b if a lexically precedes b in the uA order

  15. Normalized Utility • It can be useful to normalize utility functions before aggregation: • • If A0 is the initial allocation, then we may restrict attention to • allocations A that Pareto-dominate A0 and use utility gains • ui(A) − ui(A0) rather than ui(A) as problem input. • • We could evaluate an agent’s utility gains relative to the gains • it could expect in the best possible case. Define an agent’s • maximum utility with respect to a set Adm of admissible allocations: • bui= best of possible allocations • Then define the normalized individual utility of agent i: • u’i(A) = ui(A)/bui • The optimum of the leximin-ordering wrt. normalized utilities • is known as the Kalai-Smorodinsky solution.

  16. Ordered Weighted Averaging • We can build families of parameterized collective utility functions that induce several social welfare orders. An example are the ordered weighted averaging operators. • Let w = <w1,w2, . . . ,wn> be a vector of real numbers. Define: This generalizes several other social welfare orders: • If w is such that wi= 0 for all i≠k and wk = 1, then we have exactly the k-rank dictator collective utility function. • If wi = 1 for all i, then we obtain the utilitarian collective utility function. • If wi=αi−1, with α > 0, then the leximin-ordering is the limit of the social welfare order induced by sww as goes to 0.

  17. Envy-Freeness • An allocation is called envy-free iff no agent would rather have one of the bundles allocated to any of the other agents: A(i)iA(j) • Recall that A(i) is the bundle allocated to agent i in allocation A. Note that envy-free allocations do not always exist (at least not if we require either complete or Pareto optimal allocations).

  18. Example • Consider the following example with two agents and three goods: • A = {1, 2} and R = {a, b, c}. Suppose utility functions are additive: • u1({a}) = 18u1({b}) = 12 u1({c}) = 8 • u2({a}) = 15 u2({b}) = 8 u2({c}) = 12 • Let T be the allocation giving {a} to agent 1 and {b,c} to agent 2. • • T has maximal egalitarian social welfare (18); utilitarian social • welfare is not maximal (38 rather than 42); and neither is • elitist social welfare (20 rather than 38). • • T is Pareto optimal and leximin-optimal, but not envy-free. • • There is no allocation that would be both Pareto optimal and • envy-free. But if we change u1({a}) = 20 (from 18), then T • becomes Pareto optimal and envy free.

  19. Degrees of Envy • As we cannot always ensure envy-free allocations, another approach would be to try to reduce envy as much as possible. • But what does that actually mean? • A possible approach to systematically defining different ways of • measuring the degree of envy of an allocation: • • Envy between two agents: • max{ui(A(j)) − ui(A(i)), 0} [or even without max] • • Degree of envy of a single agent: • 0-1, max, sum • • Degree of envy of a society: • max, sum [or indeed any social welfare order/collective utility function]

  20. Allocation Procedures and Complexity • • We have now seen the various components that are needed to • specify a MARA problem (type of resource, agent preferences, • optimality criterion). • • In the future, we are going to see how agents can negotiate optimal allocations in a distributed manner and could consider a centralized allocation procedure (combinatorial auctions). • • Now we are going to look into the computational complexity of the problem of finding an optimal allocation, independently from any specific allocation procedure.

  21. Resource Allocation Problems • For the purpose of formally stating the resource allocation • problems for which we want to analyze the complexity, let a • resource allocation setting <A,R, U> be given by: • • A = {1, 2 . . . , n} is a set of n agents; • • R = {r1, r2, . . . , rm} is a collection of m resources; and • • U = {u1, u2, . . . , un} describes the utility function ui : 2RQ for the agent iin A. • The set of allocations A is the set of partitionings of R amongst A (or equivalently, the set of total functions from R to A).

  22. Welfare Optimization • How hard is it to find an allocation with maximal social welfare? • Rephrase this optimization problem as a decision problem: • Welfare Optimization (WO) • Instance: <A,R, U>; K Reals • Question: Is there an allocation A such that swu(A) > K? • Unfortunately, the problem is intractable:

  23. Envy-Freeness • Checking whether a given setting admits an envy-free allocation (if all goods need to be allocated) is again NP-complete:

  24. Allocation Procedures • To solve a MARA problem, we firstly need to decide on an • allocation procedure. This is a complex issue, involving at least: • • Protocols: What types of deals are possible? What messages do agents have to exchange to agree on one such deal? • • Strategies: What strategies can agents use for a given protocol? How can we incentivize agents to behave in a certain way? • • Algorithms: How do we solve the computational problems • faced by agents when engaged in negotiation?

  25. Centralized vs. Distributed Negotiation • An allocation procedure to determine a suitable allocation of • resources may be either centralized or distributed: • • In the centralized case, a single entity decides on the final • allocation, possibly after having elicited the preferences of the • other agents. Example: combinatorial auctions • • In the distributed case, allocations emerge as the result of a • sequence of local negotiation steps. Such local steps may or • may not be subject to structural restrictions (for instance, the • protocol may only allow for bilateral deals – mutual obligations). • Which approach is appropriate under what circumstances?

  26. Advantages of the Centralized Approach • Much recent work on negotiation and resource allocation • (particularly in the MAS community) has concentrated on • centralized approaches, in particular on combinatorial auctions. • There are several reasons for this: • • The communication protocols required are relatively simple. • • Many results from economics and game theory, in particular on mechanism design, can be exploited. • • Recent advances in the design of powerful algorithms for • winner determination in combinatorial auctions.

  27. Disadvantages of the Centralized Approach • But there are also some disadvantages to the centralized approach: • • Can we trust the centre (the auctioneer)? • • Does the centre have the computational resources required? • (but beware: distributing it doesn’t dissolve NP-hardness) • • Less natural to take an initial allocation into account (in an • auction, usually the auctioneer owns everything to begin with). • • Less natural to model step-wise improvements. • • Arguably, only the distributed approach is a serious • implementation of the MAS paradigm (that is, while • admittedly being difficult, we would really like to understand • how to make distributed decision making work . . . ).

  28. Auction Protocols • Auctions are centralized mechanisms for the allocation of goods amongst several agents. Agents report their preferences (bidding) and the auctioneer decides on the final allocation (and on prices).

  29. The Contract Net Protocol • Originally developed for task decomposition and allocation, but • also applicable to distributed negotiation over resources. Each agent • may assume the roles of manager and bidder. The Contract Net • protocol is a one-to-many protocol matching an offer by a manager • to one of potentially many bidders. There are four phases: • • Announcement phase: The manager advertises a deal to a • number of partner agents (the bidders). • • Bidding phase: The bidders send proposals to the manager. • • Assignment phase: The manager elects the best bid and • assigns the resource(s) accordingly. • • Confirmation phase: The elected bidder sends a confirmation. • R.G. Smith. The Contract Net Protocol: High-level Communication and Control • in a Distributed Problem Solver. IEEE Trans. Comp., 29:1104–1113, 1980.

  30. Extensions • The immediate adaptation of the original Contract Net protocol • only allows managers to advertise a single resource at a time, and a bidder can only offer money in return for that resource (not other items). • Possible extensions: • • Allow for negotiation over the exchanges of bundles of items. • • Allow for deals without explicit utility transfers (monetary • payments). The announcement phase remains the same, but • bids are now about offering resources in exchange, not money. • • Allow agents to negotiate several deals concurrently and to • de-commit from deals within a certain period. • • In leveled-commitment contracts, agents are also allowed to • de-commit, but have to pay a pre-defined penalty.

  31. Properties of Allocation Procedures • We may study different properties of allocation procedures: • • Termination: Is the procedure guaranteed to terminate • eventually? • • Convergence: Will the final allocation be optimal according to • our chosen social welfare measure? • • Incentive-compatibility: Do agents have an incentive to report • their valuations truthfully? (mechanism design) • • Complexity results: What is the computational complexity of • finding a socially optimal allocation of resources? • Next, we are going to see an example for a convergence property . . .

  32. Negotiating Socially Optimal Allocations • We are now going to analyze a specific model of distributed • negotiation (defined on the next slide). • We are not going to talk about designing a concrete negotiation protocol, but rather study the framework from an abstract point of view. The main question concerns the relationship between • • the local view: what deals will agents make in response to their individual preferences?; and • • the global view: how will the overall allocation of resources evolve in terms of social welfare?

  33. An Abstract Negotiation Framework • • Finite set of agents A and finite set of indivisible resources R. • • An allocation A is a partitioning of R amongst the agents in A. • Example: A(i) = {r5, r7} — agent i owns resources r5 and r7 • • Every agent iA has got a utility function ui : 2RR. • Example: ui(A) = ui(A(i)) = 577.8 — agent i is pretty happy • • Agents may engage in negotiation to exchange resources in • order to benefit either themselves or society as a whole. • • A deal = (A,A’) is a pair of allocations (before/after). • • A deal may come with a number of side payments to • compensate some of the agents for a loss in utility. • A payment function is a function p : A R with • Example: p(i) = 5 and p(j) = −5 means that agent i pays $5, • while agent j receives $5. • What are the benefits of side-payments?

  34. The Global/Social Perspective • As emphasized in earlier lectures, there are many different (fairness • or efficiency) criteria that we could use to define our goals. • For now, suppose that as system designers we are interested in • maximizing utilitarian social welfare: • swu(A) =  ui(A) • Observe that there is no need to include the agents’ side-payments into this definition, because they’d always add up to 0. • While the local perspective is driving the negotiation process, we • use the global perspective to assess how well we are doing. • Note that trying to improve social welfare is at odds with being self-interested.

  35. Example • Let A = {ann, bob} and R = {chair, table} and suppose our agents • use the following utility functions: • uann({ }) = 0 ubob({ }) = 0 • uann({chair}) = 2 ubob({chair}) = 3 • uann({table}) = 3 ubob({table}) = 3 • uann({chair, table}) = 7 ubob({chair, table}) = 8 • Furthermore, suppose the initial allocation of resources is A0 with • A0(ann) = {chair, table} and A0(bob) = { }. • Social welfare for allocation A0 is 7, but it could be 8. By moving • only a single resource from agent ann to agent bob, the former • would lose more than the latter would gain (not IR). The only • possible deal would be to move the whole set {chair, table}. • What would the appropriate side payment be?

  36. Linking the Local and the Global Perspectives • It turns out that individually rational deals are exactly those deals • that increase social welfare: • Lemma 1 (Rationality and social welfare) A deal = (A,A1’ • with side payments is individually rational iffswu(A) < swu(A’). • Proof: “ →”: Rationality means that overall utility gains outweigh • overall payments (which are = 0). • “←”: The social surplus can be divided amongst all deal • participants by using, say, equal division. • Discussion: The lemma confirms that individually rational • behavior is “appropriate” in utilitarian societies.

  37. Convergence • It is now easy to prove the following convergence result (originally stated by Sandholm in the context of distributed task allocation): • Theorem 1 (Sandholm, 1998) Any sequence of IR deals will • eventually result in an allocation with maximal social welfare. • Proof: Termination follows from our lemma and the fact that the • number of allocations is finite. So let A be the terminal allocation. • Assume A is not optimal, i.e. there exists an allocation A’ with • swu(A) < swu(A’). Then, by our lemma, deal = (A,A’) is individually • rational → contradiction. • Discussion: Agents can act locally and need not be aware of the • global picture (convergence is guaranteed by the theorem).

  38. Multilateral Negotiation • On the downside, outcomes that maximize social welfare can only • be guaranteed if the negotiation protocol allows for deals involving • any number of agents and resources: • Theorem 2 (Necessity of complex deals) Any deal δ= (A,A’) • may be necessary, i.e. there are utility functions and an initial • allocation such that any sequence of individually rational deals • leading to an allocation with maximal social welfare would have to • include (unless is δ“independently decomposable”). • The proof involves the systematic definition of utility functions • such that A’ is optimal and A is the second best allocation. • Independently decomposable deals (to which the result does not • apply) are deals that can be split into two sub-deals concerning • distinct sets of agents.

  39. Negotiation in Restricted Domains • Most work on negotiation in MAS is concerned with bilateral • negotiation or auctions. THUS Multilateral negotiation is difficult! • Maybe we can guarantee convergence to a socially optimal • allocation for structurally simpler types of deals if we restrict the • range of utility functions that agents can use? • First, two negative results: • • Theorem 2 continues to hold even when all agents have to use • monotonic utility functions. [R1 R2 ) ui(R1) ≤ ui(R2)] • • Theorem 2 continues to hold even when all agents have to use • dichotomous utility functions. [ui(R) = 0 or ui(R) = 1] • NOTE: that often progress is made by limiting the cases.

  40. Modular Domains • A utility function ui is called modular iff it satisfies the following • condition for all bundles R1,R2 R: • ui(R1R2) = ui(R1) + ui(R2) − ui(R1R2) • That is, in a modular domain there are no synergies between items; you • can get the utility of a bundle by adding up the utilities of its elements. • →Negotiation in modular domains is feasible: • Theorem 3 (Modular domains) If all utility functions are modular, • then individually rational 1-deals (each involving just one resource) • suffice to guarantee outcomes with maximal social welfare. • We also know that the class of modular utility functions is maximal : no • strictly larger class could still guarantee the same convergence property.

  41. Simulation and Experiments • While we know from Theorem 3 that 1-deals (blue) guarantee an optimal result, an experiment (20 agents, 200 resources, modular utilities) suggests that general bilateral deals (red) achieve the same goal faster. The graph shows how utilitarian social welfare (y-axis) develops as agents attempt to contract more an more deals (x-axis) amongst themselves.

  42. Communication Complexity • We should also consider the communication complexity of negotiating socially optimal allocations: focus on the length of negotiation processes and the amount of information exchanged, rather than just on computational aspects. • U. Endriss and N. Maudet. On the Communication Complexity of Multilateral Trading. Journal of Autonomous Agents and MAS, 11(1):91–107, 2005.

  43. Aspects of Complexity • (1) How many deals are required to reach an optimal allocation? • – communication complexity as number of individual deals • (2) How many dialogue moves are required to make one such deal? • – affects communication complexity as number of moves • (3) How expressive a communication language do we require? • – Minimum requirements: propose, accept, reject • + content language to specify multilateral deals • – affects communication complexity as number of bits • exchanged • (4) How complex is the reasoning task faced by an agent when • deciding on its next dialogue move? • – computational complexity (local rather than global view)

  44. Number of Deals • There are two results on upper bounds pertaining to the first variant of our negotiation framework (with side payments, general utility functions, and aiming at maximizing utilitarian social welfare): • Theorem 4 (Shortest path) A single rational deal is sufficient • to reach an allocation with maximal social welfare. • Proof: Use Lemma 1δ= (A,A0) is IR iffswu(A) < swu(A’)]. • Theorem 5 (Longest path) A sequence of rational deals can • consist of up to |A||R| − 1 deals, but not more. • Proof: No allocation can be visited twice (same lemma) and there • are |A||R| distinct allocations ) upper bound follows. • To show that the upper bound is tight, we need to show that it is • possible that all allocations have distinct social welfare . . .

  45. Path Length in Modular Domains • If all agents are using modular utility functions and only negotiate • 1-deals, then we obtain the following bounds: • • Shortest path: ≤ |R| • • Longest path: ≤|R| · (|A| − 1) • There are similar results for a framework without monetary side • payments (where the goal is to reach a Pareto optimal allocation). • Dunne (2005) has also worked on the topic of communication complexity in distributed negotiation, but generally this is still very much an under-explored area . . . • P.E. Dunne. ExtremalBehaviour in Multiagent Contract Negotiation. Journal • of Artificial Intelligence Research, 23:41–78, 2005.

  46. More on Convergence • Generally, it is interesting to see for what kind of combination of deals and optimality criteria we can get convergence results: • • Deals: structural constraints and rationality criteria • • Optimality criteria: various social welfare orders, degrees of envy, . . . • For example, the result (Theorem 1) shows that by using the rationality criterion given by our definition of individual rationality and by not imposing any structural constraints, we can guarantee convergence with respect to the optimality criterion given by the notion of utilitarian social welfare.

More Related