ics 806-Week 4 & 5

download ics 806-Week 4 & 5

of 61

Transcript of ics 806-Week 4 & 5

  • 8/14/2019 ics 806-Week 4 & 5

    1/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    MSC COMPUTER SCIENCE

    Distributed rational decision-making for multi-agent systems

    Session Topics

    1. Definitions

    2. How humans make decisions

    3. SMART technique

    4. Decision making in group setting5. Game Theoretic Perspective

    6. Voting

    7. Auctions8. Bargaining

    9. Negotiation

    10.Coalition Formation

  • 8/14/2019 ics 806-Week 4 & 5

    2/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Distributed rational decision-making for multi-

    agent systemsSources: Sandholm(1999). Distributed rational decision making; Paul Goodwin, George

    Wright (2005). Decision Analysis for Management Judgment. John Wiley & Sons Ltd,

    Third Edition; Michael Wooldridge (2002). Introduction to multi-agent systems.

    Decision- is a choice from various

    alternatives.

    Rationality optimization from point of view

    of goalsDistribution- involvement of several

    participants with no central of control

  • 8/14/2019 ics 806-Week 4 & 5

    3/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    ICS 806 - MULTI-AGENT SYSTEMS

    Distributed rational decision-making for multi-agentsystems

    How Human beings make decision

    1. They approximate or apply heuristics-rules of

    thumb usually when they lack tools for decision

    making

    2. They do so because they have limited in abilities to compute.

    This is referred to as bounded rationality (Paul Goodwin,

    George Wright (2005)).

  • 8/14/2019 ics 806-Week 4 & 5

    4/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    ICS 806 - MULTI-AGENT SYSTEMS

    Types of Heuristics used by Human beings in decision making

    Minimalist strategy: recognize all the options and simply select one atrandom.

    The last strategy: use the choice that worked the last time a decision hadto be made in a similar situation.

    Lexicographic strategy: decide on the best attribute and use it to make a

    selection such as the price property in buying a car.

    Semi lexicographic decision strategy: consider several attributes and

    apply some rules to choose. For example one may consider cost,

    engine capacity and some rules for buying a car.

  • 8/14/2019 ics 806-Week 4 & 5

    5/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    ICS 806 - MULTI-AGENT SYSTEMS

    Types of Heuristics used by Human beingsin decision making (cont)

    Elimination by aspect strategy: select an attributeand apply a cut off point. People also make

    decisions using satisficing strategy in which they

    wait for options that appear sequentially until theymeet their expectations.

    Reason based choice strategy: options aredescribed using various attributes and the optionwith the preferred attributes is selected. The

    attributes are used as a ground for selection.

  • 8/14/2019 ics 806-Week 4 & 5

    6/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Factors that affect how people make choices

    timeavailable to make the decision

    effortthat a given strategy will involve

    knowledgeof the decision maker about the environment,

    accurate decisionconstraints

    Need for justificationof choice to others

    desire to minimize conflict(for example, the conflict between thepros and cons of moving to another job).

    - decision makers may choose their strategies so as to

    balance the effort involved in making the decision

    against the accuracy that they wish to achieve.

  • 8/14/2019 ics 806-Week 4 & 5

    7/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Some terminologyObjective- an indication of the preferred direction of movement.

    Example of objectives are statements such as minimize marketlosses or maximize market share.

    Attribute -a feature that is used to measure performance in relation

    to an objective. For example, if the objective is :-maximize the

    exposure of a television advertisement then the attribute may benumber of people surveyed who recall seeing the advertisement in

    order to measure the degree to which the objective was achieved.

    Proxy attribute- an attribute which is not directly related to the

    objective. For example, a company may use the proxy attribute staff

    turnover to measure how well they are achieving their objective of

    maximizing job satisfaction for their staff.

    Value-a numerical score to measure attractiveness an action to adecision maker where there is no element of risk nor uncertainty.

    Utility- a numerical score to measure attractiveness of an action to adecision maker where there is element of risk or uncertainty.

  • 8/14/2019 ics 806-Week 4 & 5

    8/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    SMART decision making technique- multiple objectivesSimple Multi-attribute Rating Technique (SMART):

    Stage 1: Identify the decision maker (or decision makers).Stage 2: Identify the alternative courses of action. These can be,represented by the different offices the owner can choose.

    Stage 3: Identify the attributes which are relevant to the decision problem.The attributes which distinguish the different offices can be factors such

    as rent, size and quality of working conditions.

    Stage 4: For each attribute, assign values to measure the performance ofthe alternatives on that attribute. For example, how well do theConstructing a value tree offices compare when considering the quality

    of the working conditions they offer?Stage 5: Determine a weight for each attribute. This may reflect how

    important the attribute is to the decision maker

    Stage 6: For each alternative, take a weighted average of the values

    assigned to that alternative. This will give us a measure of how well an anitem such as officeperforms over all the attributes.

    Stage 7: Make a provisional decision.Stage 8: Perform sensitivity analysisto see how robust the decision is tochanges in the figures supplied by the decision maker.

  • 8/14/2019 ics 806-Week 4 & 5

    9/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    SMART example ( where should an office be located?)

  • 8/14/2019 ics 806-Week 4 & 5

    10/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Smart examples: Values

    Aggregate benefits = [(weighti * valuei)] / [(weighti)]

  • 8/14/2019 ics 806-Week 4 & 5

    11/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Basic axioms related to SMART(1) Decidability: The decision maker should be able to decide which of the

    options s/he prefers.

    (2) Transitivity: If the decision maker prefers A to B; and s/he also prefers Bto C; then s/he must therefore also prefer A to C.

    (3) Summation: This implies that if the decision maker prefers A to B and Bto C, then the strength of preference of A over C must be greater than the

    strength of preference of A over B (or B over C).

    (4) Solvability: This is necessary for the method of obtaining a valuefunction. For example it can be the distance from the center of town

    which had a value halfway between the worst and best distances.

    (5) Finite upper and lower bounds for value: The values that turn out to be

    the best option are not necessarily so wonderful and the worst option is

    not so awful.

  • 8/14/2019 ics 806-Week 4 & 5

    12/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Decision making in group settingGroups- managers or stakeholders

    There are essentially two approaches to the problem:

    1. mathematical and2. behavioral aggregation

    (although the approaches can be combined).

    Mathematical aggregation rely on techniques such as thecalculation of a simple average of the judgments of the

    individual group members.

    Behavioral aggregation relies on reaching a decision by membersof the group communicating with each other either in open

    discussion or via a more structured communication process.

  • 8/14/2019 ics 806-Week 4 & 5

    13/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Examples: mathematical aggregation

  • 8/14/2019 ics 806-Week 4 & 5

    14/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Examples: Aggregating probabilities (p.330)

    Note potential problems with using probabilities

  • 8/14/2019 ics 806-Week 4 & 5

    15/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Aggregating preferences

    Member Preferences

    Jana A > B > C

    Jing B > C > AJumbo C> A > B

    Taking votes for A>B: A gets two and B gets 1.

    Comparing B and C (B > C): B gets 2 and C gets 1.

    Comparing C and A (C>A): C gets 2 and A gets 1. So thepreference is not transitive. This is known as

    Condorcets paradox.

  • 8/14/2019 ics 806-Week 4 & 5

    16/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Arrowshowed in his Impossibility TheoremThere is no aggregation procedure that can produce

    preferences with the following properties:1. A transitive group preference order for the options

    under consideration;

    2. If every member of the group prefers one option toanother then so must the group.

    3. The group choice between two options, A and B,

    depends only upon the preferences of members

    between these options and not on preferences for

    any other option. This discourages lying.

    4. There is no dictator. No individual is able to impose

    his or her preferences on the group.

  • 8/14/2019 ics 806-Week 4 & 5

    17/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Values and utilities may be aggregated based on

    simple average.

    As given in the example that follows:

    Destination Member 1 Member 2 AverageNanyuki 150 100 125

    Usenge 80 90 85

    Witu 100 86 93

    Decision: members choose to go to Nanyuki

  • 8/14/2019 ics 806-Week 4 & 5

    18/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Structured group decision making process- minimize negative group effects such as domineering by single

    individuals;- realize this by controlling inter-personal interaction and information

    flow such as using the Delphi method;

    Delphi method:

    (1) Panelists provide opinions about the likelihood of futureevents, or when those events will occur, or what the impact of

    such event(s) will be. These opinions are often given as

    responses to questionnaires which are completed individually by

    members of the panel.

    (2) The results of this polling of panelists are then tallied and statisticalfeedback of the whole panels opinions (e.g. range or medians) are

    provided to individual panelists before a re-polling takes place. At this

    stage, anonymous discussion (often in written form) may occur so thatdissenting opinion is aired.

    (3) The output of the Delphi technique is a quantified group consensus,

    which is usually expressed as the median response of the group of

    panelists. (Paul Goodwin, George Wright (2005) p.337)

  • 8/14/2019 ics 806-Week 4 & 5

    19/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    The game theoretic perspective (Wooldridge)

    Utilities and Preferences

    Consider only two agents: Ag = {i, j} that are self-interested

    Let = {1, 2, .. n} be a set of the preferred outcomes by agentsWe capture preferences by utility functions:

    ui : Ru

    j: R

    Utility functions lead to preference orderingsover outcomes:

    i means ui() ui()j means ui() > uj()Environment behaviour given by state transformer function:

    : Ac x Ac agent is action (C or D) agentjs action (C or D)

    Here is a state transformer function:

    (D,D) =1, (D,C) =2, (C,D) =3, (C,C) =4(This environment is sensitive to actions of both agents.)Here is another:

    (D,D) =1, (D,C) =1, (C,D) =1, (C,C) =1(Neither agent has any influence in this environment.)

  • 8/14/2019 ics 806-Week 4 & 5

    20/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Rational ActionSuppose we have the case where bothagents can

    influence the outcome, and they have utility functions asfollows:

    ui(1)=1, ui(2)=2, ui(3)=3, ui(4)=4uj(

    1)=1, uj(

    2)=2, uj(

    3)=3, uj(

    4)=4

    With a bit of mixing of notation:

    ui

    (D,D)=1, ui

    (D,C)=2, ui

    (C,D)=3, ui

    (C,C)=4uj(D,D)=1, uj(D,C)=2, uj(C,D)=3, uj((C,C)=4

    Then agent is preferences are:

    C, CiC, DiD,CiD,DC is the rational choicefori.(Because iprefers all outcomes that arise through Cover

    all outcomes that arise through D.)

  • 8/14/2019 ics 806-Week 4 & 5

    21/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Payoff MatricesWe can characterize the previous scenario in a payoff matrix whereAgent iis

    the column playerand Agentjis the row player.

  • 8/14/2019 ics 806-Week 4 & 5

    22/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Dominant StrategiesGiven any particular strategy s(eitherCorD)agent i, there will be a number of possible

    outcomes.

    We say s1 dominates s2 if every outcome

    possible by iplaying s1, is preferred overevery outcome possible by iplaying s2.A rational agent will never play a dominated

    strategy. So in deciding what to do, we candelete dominated strategies. Unfortunately,there isnt always a unique undominated

    strategy.

  • 8/14/2019 ics 806-Week 4 & 5

    23/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Nash Equilibrium

    In general, we will say that two strategies s1and s2are in Nash equilibrium if:

    Under the assumption that agent iplays s1,

    agentjcan do no better than play s2; andUnder the assumption that agentjplays s2,

    agent ican do no better than play s1.

    Neither agent has any incentive to deviatefrom a Nash equilibrium.

    Unfortunately: Not every interaction scenario

    has a Nash equilibrium.

    Some interaction scenarios have more than

    one Nash equilibrium.

    C titi d Z S I t ti

  • 8/14/2019 ics 806-Week 4 & 5

    24/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Competitive and Zero-Sum InteractionsWhere preferences of agents are diametrically

    opposed we have strictly competitivescenarios.

    Zero-sum encounters are those where utilities

    sum to zero:

    ui( ) + uj() = 0for all , Zero sum impliesstrictly competitive.

    Zero sum encounters in real life are very rare . .. but people tend to act in many scenarios as

    if they were zero sum.

  • 8/14/2019 ics 806-Week 4 & 5

    25/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    The Prisoners Dilemma

    Two men are collectively charged with a crimeand held in separate cells, with no way of

    meeting or communicating. They are told

    that:1. if one confesses and the other does not, the

    confessor will be freed, and the other will be

    jailed for three years;

    2. if both confess, then each will be jailed for

    two years.Both prisoners know that if neither confesses,

    then they will each be jailed for one year.

    Payoff matrix for prisoners dilemma:

  • 8/14/2019 ics 806-Week 4 & 5

    26/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Payoff matrix for prisoners dilemma:

    Top left: If both defect, then both get punishment for mutual defection.

    Top right: Ificooperates andjdefects, igets suckers payoff of 1, whilejgets 4.Bottom left: Ifjcooperates and idefects,jgets suckers payoff of 1, while igets 4.

    Bottom right: Reward for mutual cooperation.

  • 8/14/2019 ics 806-Week 4 & 5

    27/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Prisoners Dilemma cont.The individual rationalaction is defect. This guarantees a payoff of no worse than

    2, whereas cooperating guarantees a payoff of at most 1.

    So defection is the best response to all possible strategies: both agents defect,

    and get payoff = 2.

    But intuitionsays this is notthe best outcome: Surely they should both cooperateand each get payoff of 3!

    This apparent paradox is the fundamental problem of multi-agent interactions. It

    appears to imply that cooperation will not occur in societies of self-interestedagents.

    Real world examples

    nuclear arms reduction (why dont I keep mine. . . )free rider systems public transport;

    in Kenya- television licenses.

    The prisoners dilemma is ubiquitous. Can we recover cooperation?

  • 8/14/2019 ics 806-Week 4 & 5

    28/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Distributed rational decision making continues(Tuomas W. Sandholm(1999). Distributed rational decision making.)

    Automated negotiation with self interested agents is

    important because:

    Technology push that has led to standardized

    communication infrastructure such as : Intenet, NII,EDI, FIPA, Concordia, Voyager, Odyssey, Telescript,

    Java. Agents can easily communicate even if they are

    heterogeneous.There is a move to computerize operational decisions

    such as Internet transactions for purchase of goods,

    information or bandwidth.There is a trend towards virtual organizations in

    which dynamic alliances of small enterprises occur and

    take advantage of economics of scale.

    M lti A t t i t i t ti

  • 8/14/2019 ics 806-Week 4 & 5

    29/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Multi-Agents systems can assist in automating

    negotiation at operative decision level.

    The agents are self interested and may pursue different goals.

    Usually protocols and a strategy (way to use the protocol) are imposed

    by the designer. Agents have interaction protocol but each agent

    chooses their own strategy.

    For self interested agents strategies are chosen so that individual goals

    are best met.

    The protocols should therefore be designed based on non-cooperative

    perspective.

    The Protocols considered are: voting, auctions, contracting, bargaining,

    and coalition formation. The protocols however affect social outcomes.Note that self interested agents are concerned with their own local goals

    and not the global(social) good. This is why protocols are important for

    them.

  • 8/14/2019 ics 806-Week 4 & 5

    30/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Criteria for evaluating protocolsSocial welfare sum of all agents utilities; can be used to compare protocols;

    Pareto efficiency a solution X is Pareto efficient if there is no other solution Xin which an agent is better off than in X; and there is no agent worse off in X

    than in X.

    Individual rationality- occurs when negotiated payoff is better than the payoff

    without negotiation. The mechanism should be individual rational for everyagent.

    Stability the mechanism or protocol should not be manipulated. Agents shouldbe motivated to behave in some desired ways. Some mechanisms may have

    dominant strategies. Sometimes Nash equilibrium can be used where

    appropriate.

    Computational efficiency agents use mechanisms and computationaloverheads should be as minimal as possible; protocols with minimum

    computational overheads are preferred.

    Distribution and communication efficiency- distributed protocols are preferred tocentralized ones to avoid single point of failures; however, communication

    overheads should be minimized.

    V ti

  • 8/14/2019 ics 806-Week 4 & 5

    31/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Voting

    Voting occurs on some social choice such asa swimming pool, public health centre, etc.

    All agents give inputs to a mechanism.

    The mechanism then makes a choice that willapply to all agents. Voters can be truthful or

    non truthful (insincere).

    Truthful voters f f f

  • 8/14/2019 ics 806-Week 4 & 5

    32/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Truthful voters- will rank feasible social outcomes faithfully.Suppose A ={agents or voters} and O ={feasible outcomes}

    Let agent i A have a strict transitive preference relation ion social choices in O. Suppose > is a social choice rule

    that takes as inputs agent preferences and produces as

    outputs ordered social choice preferences. Then thefollowing features are desirable of the mechanism:

    1. should exist for all agents;

    2. should be defined for every pair o, w O;

    3. should be asymmetric and transitive over O;

    4. Outcome should be Pareto efficient i.e. o io, o, o O,then o > o

    5. Irrelevant alternatives should have no effect; if all for agents ranks o io iff o i o then > should also rank similarly;

    6. No agent should be a dictator in that o i o then o > o for all other

    agents;

    Arrows impossibility theorem

  • 8/14/2019 ics 806-Week 4 & 5

    33/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Arrows impossibility theorem

    There is no social choice rule that satisfies all the six

    conditions above.

    Note that all agents preferences are known.

    Strategic (insincere) voters

  • 8/14/2019 ics 806-Week 4 & 5

    34/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Strategic (insincere) voters-occur in situations where agents reveal their preferences.-assumption: preferences known are those that are truthfully revealed.

    - However, the agents that can benefit from lying about their preferences will

    do so.

    So design protocols, with stability items included such as dominant strategy

    or Nash equilibrium.

    This will make agents behave in some ways that leads to some social good.

    The overall outcome should be the same as if the agents revealed theirpreferences or types truthfully.

    Theorem: Revelation Principle:suppose some protocol implements socialchoice function f(.) in Nash (or dominant strategy) equilibrium where agents

    are not necessarily truthful. Then the social choice function f(.) is

    implementable in Nash (or dominant strategy) via a single step where agents

    reveal their types truthfully.

    Theorem: Gibbard-Sattethwait

  • 8/14/2019 ics 806-Week 4 & 5

    35/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Theorem: Gibbard Sattethwaitimpossibility theoremlet agent types i consist of preferences orderi on O. Let there beno restrictions on i so that agents rank their preferences in anyorder.

    Let |O| >= 3. If social choice function f(.) is truthfully implementable

    in a dominant strategy equilibrium then f(.) is dictatorial in thatsome agent gets most preferred outcome irrespective of the types

    others reveal.

    Restricting preference and using Groves-Clarke Tax mechanism

    enables constructing protocols that cannot be manipulated. The

    voters are taxed according to how their votes may affect the

    outcome. Gibbard-Sattethwait impossibility theorem is thereforecircumvented. There are other ways of circumventing the theorem.

    Auctions

  • 8/14/2019 ics 806-Week 4 & 5

    36/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    AuctionsAn auction takes place between an agent known as the auctioneerand a

    collection of agents known as the bidders.

    The goal of the auction is for the auctioneer to allocate the goodto one of thebidders. In most settings the auctioneer desires to maximize the price; bidders

    desire to minimize price.

    Auction settings

    Private value auctionsThe value of goods depends only on the agents. Only the winning agents utilize

    the goods and do not re-sell them. Agents are assumed to know the valuesexactly.

    Common value auctionsThe value of the goods depends entirely on others value of the good. For

    example, where the goods are re-sold such as in treasury bills etc, tea or coffee

    auctions.

    Correlated value auctionsValue of goods depend partly on agents own preferences and partly on value of

    others such as in negotiation within a contract.

    Auction parameters

  • 8/14/2019 ics 806-Week 4 & 5

    37/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Auction parameters

    Valueof goods

    Bid types(English, sealed first price, Dutch, Vickrey),and bidding process.

    Goods value type:private value public/common value;correlated value.

    Winner determinationmay be first price; second price.

    Bidding process:open cry; sealed bid.

    Bidding type: one shot; ascending; descending.

    Auction protocols

  • 8/14/2019 ics 806-Week 4 & 5

    38/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Auction protocols

    English AuctionsAre: first-price, open cry, ascending.

    The dominant strategy in an English Auction is

    for agent to successively bid a small amount

    more than the current highest bid until it reachestheir valuation, then withdraw.

    The English Auction is susceptible to: winnerscurse where the good worn may of lower value;and shills in which time is wasted in making

    decisions.

    Dutch Auctions

  • 8/14/2019 ics 806-Week 4 & 5

    39/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Dutch Auctions

    Are: open-cry; descending

    The auctioneer starts by good at artificially highvalue;

    auctioneer lowers the offer price until some agentmakes a bid equal to the current offer price;

    the good is then allocated to the agent that made

    the offer.

    First-Price Sealed-Bid Auctions

  • 8/14/2019 ics 806-Week 4 & 5

    40/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    First-Price Sealed-Bid Auctions

    Are: one-shot

    There is a single round;

    bidders submit a sealed bid for the good;

    the good is allocated to the agent that madehighest bid;

    and the winner pays price of highest bid.

    Best strategy is to bid less than true valuation.

    Vickrey Auctions

  • 8/14/2019 ics 806-Week 4 & 5

    41/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Vickrey Auctions

    Are: second-price; sealed-bid auctions.

    Good is awarded to the agent that made the

    highest bid; at the price of the second highestbid.

    Bidding to your true valuation is dominantstrategy in Vickrey auctions.

    Vickrey auctions susceptible to antisocialbehavior.

    Some issues in Auctions

  • 8/14/2019 ics 806-Week 4 & 5

    42/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Some issues in Auctions

    All auction mechanisms are not collusion proof

    Bidders can collude and lower final awards

    Most vulnerable: English and Vickrey auctionsThe sealed first price, Dutch and Vickrey

    auctions collusions are only possible if the

    bidders know each other ahead of bidding.

    Vickrey Auction: auctioneer can lie by

    overstating the value of the second highest bid.

    B i i

  • 8/14/2019 ics 806-Week 4 & 5

    43/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Bargaining

    Agent agree on outcomes that are mutuallybeneficial.

    Monopoly: one agent gets all the benefits of theinteraction.

    Perfect competition: no agent gets the benefits

    of the interaction.

    Types: axiomatic; strategic.

    Axiomatic bargaining theory

  • 8/14/2019 ics 806-Week 4 & 5

    44/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    g g yConsists of postulates and solutions that satisfy the

    axioms are sought.An example is the Nash bargaining axioms (for 2 agents)

    that follow:

    Let agent i {1, 2}, ui be the utility function for eachagent, ui: O R, and the set of feasible utility vectors beconvex ({(u1(o), u2(o))| o O)}). Let u* = (u1(o*), u2(o*)).Then the axioms are:

    Invariance- numeric utility functions only representordinal preferences among outcomes.Anonymity agents are anonymous in that switching labels of

    players does not affect the outcome.Independence of irrelevant alternatives- removing o and not o* does

    not affect the solution o*.

    Pareto efficiency- both agents cannot get utility than under u* =

    (u1(o*), u2(o*)).

    Strategic bargaining theory

  • 8/14/2019 ics 806-Week 4 & 5

    45/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    g g g yAgents make offers to each other in pre-specified order.

    The bargaining process may take a finite number ofsteps.

    One approach: use discounts on every step where one

    agent concedes some value.

    Another approach: agents incur bargaining costs every

    step of the process and the one whose costs are higher

    concedes.

    Computations:

    intra-agent deliberativesearch: agents locally generate

    alternatives, evaluates them, counter-speculate and lookahead in negotiation process

    Inter-agent committal search:agents make binding

    agreements with each other concerning the outcome.

    Negotiation (Wooldridge)

  • 8/14/2019 ics 806-Week 4 & 5

    46/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    - process of reaching agreements on matters of

    common interestComponents of a negotiation setting:

    1.A negotiation set: possible proposals that

    agents can make.

    2.A protocol. Strategies, one for each agent,

    which are private.

    3.A rule that determines when a deal has been

    struck and what the agreement deal is.

    4.Process: Negotiation usually proceeds in aseries of rounds, with every agent making a

    proposal at every round.

    Negotiation in Task-Oriented Domains

  • 8/14/2019 ics 806-Week 4 & 5

    47/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Negotiation in Task Oriented Domains

    (TOD)A TOD is a triple where:

    Tis the (finite) set of all possible tasks;

    Ag ={ 1, .., n} is set of participant agents;

    c=(T) R+ defines costof executing each subset oftasks;

    An encounteris a collection of tasks where

    Ti Tfor each i Ag.

    Deals in TODs

  • 8/14/2019 ics 806-Week 4 & 5

    48/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Deals in TOD s

    Given encounter a dealwill be an allocation ofthe tasks T1 U T2to the agents 1 and 2.

    The costto iof deal is c(Di

    ), and will bedenoted costi().

    The utilityof deal to agent iis:utilityi() = c(Ti) costi()

    The conflict deal, , is the deal consisting ofthe tasks originally allocated.

    Deal is individual rationalif it weakly dominates theconflict deal.

  • 8/14/2019 ics 806-Week 4 & 5

    49/61

    The Monotonic Concession Protocol

  • 8/14/2019 ics 806-Week 4 & 5

    50/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Rules of this protocol are as follows. . .

    Negotiation proceeds in rounds.1. On round 1, agents simultaneously propose a deal from

    the negotiation set. Agreement is reached if one agent

    finds that the deal proposed by the other is at least asgood or better than its proposal.

    2. If no agreement is reached, then negotiation proceeds to

    another round of simultaneous proposals.

    3. In round u+ 1, no agent is allowed to make a proposalthat is less preferred by the other agent than the deal it

    proposed at time u.

    4. If neither agent makes a concession in some round u >0, then negotiation terminates, with the conflict deal.

    The Zeuthen Strategy

  • 8/14/2019 ics 806-Week 4 & 5

    51/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    The Zeuthen Strategy

    Three problems:

    What should an agents first proposal be? Its mostpreferred deal

    On any given round, who should concede? The agentleast willing to risk conflict.

    If an agent concedes, then how muchshould itconcede? Just enough to change the balance of risk.

    Willingness to Risk Conflict

  • 8/14/2019 ics 806-Week 4 & 5

    52/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Suppose you have conceded a lot.

    Then:1. Your proposal is now near to conflict deal.

    2. In case conflict occurs, you are not much worse off.

    3. You are more willingto risk conflict.

    An agent will be more willingto risk conflict if thedifference in utility between its current proposal and

    the conflict deal is low.

    The Contract NetIs a well known task sharing protocol for task allocation:

  • 8/14/2019 ics 806-Week 4 & 5

    53/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Is a well known task-sharing protocol fortask allocation:1. Recognition; 2. Announcement; 3. Bidding; 4. Awarding; 5. Expediting

    Coalition formation (Sandholm, Brad Spangler(2003). Coalition Building.

    http://www.beyondintractability.org/action/author.jsp?id=24548http://www.beyondintractability.org/action/author.jsp?id=24548
  • 8/14/2019 ics 806-Week 4 & 5

    54/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    (http://www.beyondintractability.org/essay/coalition_building/) June 2003. )

    A coalition is a temporary alliance or partnering

    of agents in order to achieve a common purpose

    or to engage in joint activity.

    Coalition building is the process by which groups

    (agents, individuals, organizations, or nations)

    come together to form a coalition.

    Building a successful coalition

  • 8/14/2019 ics 806-Week 4 & 5

    55/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    g

    involves a series of steps.

    Potential coalition members must be persuadedthat forming a coalition would be to their

    benefit. This is done by demonstrating:

    1. that the goals are similar and compatible,2. that working together will enhance agents'

    abilities to reach their goals,

    3. that the benefits of coalescing are greater thanthe costs.

    Benefits of a coalition

  • 8/14/2019 ics 806-Week 4 & 5

    56/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    1.A coalition can bring more expertise and

    resources to bear on complex issues, where the

    technical or other resources of any one agent orother agent groups would not be sufficient.

    2.A coalition will increase the impact of each agent's

    effort.

    3.A coalition will increase available resources. Eachagent will gain access to the contacts,

    connections, and relationships established by

    other agents.

    Problems with working in a Coalition

  • 8/14/2019 ics 806-Week 4 & 5

    57/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    1. Member agents can get distracted.

    2. A coalition may only be as strong as its weakest link. Eachmember agent will have different levels of resources

    different internal limitations.

    3. To keep a coalition together, it is often necessary to caterto one side more than another, especially when

    negotiating tactics.

    4. The democratic principle of one agent-one vote may notalways be acceptable to agents with a lot of resources.

    The coalition must carefully define the relationships

    between powerful and less-powerful agents.

    5. Individual organizations may not get credit for their

    contributions to a coalition. Members that contribute a lot

    may think they did not receive enough credit.

    Implementing multi-agent coalitions

  • 8/14/2019 ics 806-Week 4 & 5

    58/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    Implemented by: looking for some equilibrium

    solution; or using characteristic function game.

    Equilibrium solution: formulating the coalitionformation protocols and examining the way the

    coalition evolves.

    The characteristic function game (CFG): coalition

    S has characteristic function sv .

  • 8/14/2019 ics 806-Week 4 & 5

    59/61

    Example of coalition structure for 4 agents (Sandholm, p.44)

  • 8/14/2019 ics 806-Week 4 & 5

    60/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    End of Session (Week 5) Exercises

  • 8/14/2019 ics 806-Week 4 & 5

    61/61

    September/October

    2009

    School of Computing and

    Informatics, University of Nairobi

    1. Describe how human beings make decisions.

    2. What is the SMART decision making technique ?

    3. Construct an agent model for decision making in a group.

    4. What is game theory ?

    5. Construct an agent model for implementing voting.

    6. Construct an agent model that implements auctions.

    7. Construct an agent model that implements bargaining.

    8. Construct an agent model that implements negotiation.

    9. Construct an agent model that implements coalitions.