LECTURE 4: MULTIAGENT INTERACTIONS

Similar documents
In reality; some cases of prisoner s dilemma end in cooperation. Game Theory Dr. F. Fatemi Page 219

In the Name of God. Sharif University of Technology. Graduate School of Management and Economics

Introduction to Industrial Organization Professor: Caixia Shen Fall 2014 Lecture Note 5 Games and Strategy (Ch. 4)

Game Theory - Lecture #8

SI Game Theory, Fall 2008

Week 8: Basic concepts in game theory

Introduction to Multi-Agent Programming

Early PD experiments

CMPSCI 240: Reasoning about Uncertainty

CS 331: Artificial Intelligence Game Theory I. Prisoner s Dilemma

Regret Minimization and Security Strategies

preferences of the individual players over these possible outcomes, typically measured by a utility or payoff function.

Prisoner s dilemma with T = 1

Introduction to Game Theory

Week 8: Basic concepts in game theory

Lecture 5 Leadership and Reputation

Static Games and Cournot. Competition

Problem 3 Solutions. l 3 r, 1

S 2,2-1, x c C x r, 1 0,0

Game theory and applications: Lecture 1

In the Name of God. Sharif University of Technology. Microeconomics 2. Graduate School of Management and Economics. Dr. S.

Game Theory: Minimax, Maximin, and Iterated Removal Naima Hammoud

Noncooperative Oligopoly

CMPSCI 240: Reasoning about Uncertainty

FDPE Microeconomics 3 Spring 2017 Pauli Murto TA: Tsz-Ning Wong (These solution hints are based on Julia Salmi s solution hints for Spring 2015.

An introduction on game theory for wireless networking [1]

Game Theory. VK Room: M1.30 Last updated: October 22, 2012.

m 11 m 12 Non-Zero Sum Games Matrix Form of Zero-Sum Games R&N Section 17.6

Managerial Economics ECO404 OLIGOPOLY: GAME THEORETIC APPROACH

Iterated Dominance and Nash Equilibrium

Prisoner s Dilemma. CS 331: Artificial Intelligence Game Theory I. Prisoner s Dilemma. Prisoner s Dilemma. Prisoner s Dilemma.

Warm Up Finitely Repeated Games Infinitely Repeated Games Bayesian Games. Repeated Games

ECON/MGEC 333 Game Theory And Strategy Problem Set 9 Solutions. Levent Koçkesen January 6, 2011

Stochastic Games and Bayesian Games

Economics and Computation

The Nash equilibrium of the stage game is (D, R), giving payoffs (0, 0). Consider the trigger strategies:

Introduction to Game Theory Lecture Note 5: Repeated Games

CS711 Game Theory and Mechanism Design

HE+ Economics Nash Equilibrium

Elements of Economic Analysis II Lecture X: Introduction to Game Theory

Chapter 8. Repeated Games. Strategies and payoffs for games played twice

Exercises Solutions: Game Theory

Repeated, Stochastic and Bayesian Games

Economics 171: Final Exam

Repeated Games. Econ 400. University of Notre Dame. Econ 400 (ND) Repeated Games 1 / 48

Introductory Microeconomics

Player 2 L R M H a,a 7,1 5,0 T 0,5 5,3 6,6

Game Theory. Wolfgang Frimmel. Repeated Games

6.207/14.15: Networks Lecture 9: Introduction to Game Theory 1

Economics and Computation

Chapter 2 Strategic Dominance

Stochastic Games and Bayesian Games

ECONS 424 STRATEGY AND GAME THEORY MIDTERM EXAM #2 ANSWER KEY

Name. FINAL EXAM, Econ 171, March, 2015

Repeated Games. September 3, Definitions: Discounting, Individual Rationality. Finitely Repeated Games. Infinitely Repeated Games

G5212: Game Theory. Mark Dean. Spring 2017

Multiagent Systems. Multiagent Systems General setting Division of Resources Task Allocation Resource Allocation. 13.

FDPE Microeconomics 3 Spring 2017 Pauli Murto TA: Tsz-Ning Wong (These solution hints are based on Julia Salmi s solution hints for Spring 2015.

Game Theory I 1 / 38

Repeated games. Felix Munoz-Garcia. Strategy and Game Theory - Washington State University

January 26,

Economics 431 Infinitely repeated games

SI 563 Homework 3 Oct 5, Determine the set of rationalizable strategies for each of the following games. a) X Y X Y Z

Game Theory I 1 / 38

CS 798: Homework Assignment 4 (Game Theory)

The Ohio State University Department of Economics Econ 601 Prof. James Peck Extra Practice Problems Answers (for final)

Answer Key: Problem Set 4

Game Theory: Normal Form Games

CHAPTER 14: REPEATED PRISONER S DILEMMA

6.254 : Game Theory with Engineering Applications Lecture 3: Strategic Form Games - Solution Concepts

Algorithms and Networking for Computer Games

Economics 109 Practice Problems 1, Vincent Crawford, Spring 2002

Duopoly models Multistage games with observed actions Subgame perfect equilibrium Extensive form of a game Two-stage prisoner s dilemma

LECTURE NOTES ON GAME THEORY. Player 2 Cooperate Defect Cooperate (10,10) (-1,11) Defect (11,-1) (0,0)

PAULI MURTO, ANDREY ZHUKOV. If any mistakes or typos are spotted, kindly communicate them to

6.207/14.15: Networks Lecture 9: Introduction to Game Theory 1

ECO303: Intermediate Microeconomic Theory Benjamin Balak, Spring 2008

March 30, Why do economists (and increasingly, engineers and computer scientists) study auctions?

Review Best Response Mixed Strategy NE Summary. Syllabus

M.Phil. Game theory: Problem set II. These problems are designed for discussions in the classes of Week 8 of Michaelmas term. 1

CUR 412: Game Theory and its Applications Final Exam Ronaldo Carpio Jan. 13, 2015

Econ 323 Microeconomic Theory. Practice Exam 2 with Solutions

is the best response of firm 1 to the quantity chosen by firm 2. Firm 2 s problem: Max Π 2 = q 2 (a b(q 1 + q 2 )) cq 2

A brief introduction to evolutionary game theory

Name. Answers Discussion Final Exam, Econ 171, March, 2012

Econ 323 Microeconomic Theory. Chapter 10, Question 1

UC Berkeley Haas School of Business Game Theory (EMBA 296 & EWMBA 211) Summer 2016

Agenda. Game Theory Matrix Form of a Game Dominant Strategy and Dominated Strategy Nash Equilibrium Game Trees Subgame Perfection

Game Theory: Additional Exercises

Sequential-move games with Nature s moves.

ECON 803: MICROECONOMIC THEORY II Arthur J. Robson Fall 2016 Assignment 9 (due in class on November 22)

CUR 412: Game Theory and its Applications, Lecture 12

Strategies and Nash Equilibrium. A Whirlwind Tour of Game Theory

Introduction to Game Theory

Chapter 2 Discrete Static Games

Test 1. ECON3161, Game Theory. Tuesday, September 25 th

Rationalizable Strategies

Computational Examination of Strategies for Play in IDS Games

PAULI MURTO, ANDREY ZHUKOV

Microeconomics of Banking: Lecture 5

Transcription:

What are Multiagent Systems? LECTURE 4: MULTIAGENT INTERACTIONS Source: An Introduction to MultiAgent Systems Michael Wooldridge 10/4/2005 Multi-Agent_Interactions 2 MultiAgent Systems Thus a multiagent system contains a number of agents which interact through communication are able to act in an environment have different spheres of influence (which may coincide) will be linked by other (organizational) relationships Utilities and Preferences Assume we have just two agents: Ag = {i, j} Agents are assumed to be self-interested: they have preferences over how the environment is Assume Ω = {ω 1, ω 2, }is the set of outcomes that agents have preferences over We capture preferences by utility functions: u i = Ω R u j = Ω R Utility functions lead to preference orderings over outcomes: ω i ω means u i (ω) u i (ω ) ω i ω means u i (ω) > u i (ω ) 10/4/2005 Multi-Agent_Interactions 3 10/4/2005 Multi-Agent_Interactions 4 1

What is Utility? Utility is not money (but it is a useful analogy) Typical relationship between utility & money: Multiagent Encounters We need a model of the environment in which these agents will act agents simultaneously choose an action to perform, and as a result of the actions they select, an outcome in Ω will result the actual outcome depends on the combination of actions assume each agent has just two possible actions that it can perform, C ( cooperate ) and D ( defect ) Environment behavior given by state transformer function: 10/4/2005 Multi-Agent_Interactions 5 10/4/2005 Multi-Agent_Interactions 6 Multiagent Encounters Here is a state transformer function: (This environment is sensitive to actions of both agents.) Here is another: Rational Action Suppose we have the case where both agents can influence the outcome, and they have utility functions as follows: With a bit of abuse of notation: (Neither agent has any influence in this environment.) And here is another: (This environment is controlled by j.) 10/4/2005 Multi-Agent_Interactions 7 Then agent i s preferences are: C is the rational choice for i. (Because i prefers all outcomes that arise through C over all outcomes that arise through D.) 10/4/2005 Multi-Agent_Interactions 8 2

Payoff Matrices We can characterize the previous scenario in a payoff matrix: Agent i is the column player Agent j is the row player Dominant Strategies Given any particular strategy (either C or D) of agent i, there will be a number of possible outcomes We say s 1 dominates s 2 if every outcome possible by i playing s 1 is preferred over every outcome possible by i playing s 2 A rational agent will never play a dominated strategy So in deciding what to do, we can delete dominated strategies Unfortunately, there isn t always a unique undominated strategy 10/4/2005 Multi-Agent_Interactions 9 10/4/2005 Multi-Agent_Interactions 10 Nash Equilibrium In general, we will say that two strategies s 1 and s 2 are in Nash equilibrium if: 1. under the assumption that agent i plays s 1, agent j can do no better than play s 2 ; and 2. under the assumption that agent j plays s 2, agent i can do no better than play s 1. Neither agent has any incentive to deviate from a Nash equilibrium Unfortunately: 1. Not every interaction scenario has a Nash equilibrium 2. Some interaction scenarios have more than one Nash equilibrium Competitive and Zero-Sum Interactions Where preferences of agents are diametrically opposed we have strictly competitive scenarios Zero-sum encounters are those where utilities sum to zero: u i (ω) + u j (ω) = 0 for all ω Ω Zero sum implies strictly competitive Zero sum encounters in real life are very rare but people tend to act in many scenarios as if they were zero sum 10/4/2005 Multi-Agent_Interactions 11 10/4/2005 Multi-Agent_Interactions 12 3

Two men are collectively charged with a crime and held in separate cells, with no way of meeting or communicating. They are told that: if one confesses and the other does not, the confessor will be freed, and the other will be jailed for three years if both confess, then each will be jailed for two years Both prisoners know that if neither confesses, then they will each be jailed for one year Payoff matrix for prisoner s dilemma: Top left: If both defect, then both get punishment for mutual defection Top right: If i cooperates and j defects, i gets sucker s payoff of 1, while j gets 4 Bottom left: If j cooperates and i defects, j gets sucker s payoff of 1, while i gets 4 Bottom right: Reward for mutual cooperation 10/4/2005 Multi-Agent_Interactions 13 10/4/2005 Multi-Agent_Interactions 14 The individual rational action is defect This guarantees a payoff of no worse than 2, whereas cooperating guarantees a payoff of at most 1 So defection is the best response to all possible strategies: both agents defect, and get payoff = 2 But intuition says this is not the best outcome: Surely they should both cooperate and each get payoff of 3! This apparent paradox is the fundamental problem of multi-agent interactions. It appears to imply that cooperation will not occur in societies of self-interested agents. Real world examples: nuclear arms reduction ( why don t I keep mine... ) free rider systems public transport; in the UK television licenses. The prisoner s dilemma is ubiquitous. Can we recover cooperation? 10/4/2005 Multi-Agent_Interactions 15 10/4/2005 Multi-Agent_Interactions 16 4

Arguments for Recovering Cooperation Conclusions that some have drawn from this analysis: the game theory notion of rational action is wrong! somehow the dilemma is being formulated wrongly Arguments to recover cooperation: We are not all Machiavelli! The other prisoner is my twin! The shadow of the future The Iterated Prisoner s Dilemma One answer: play the game more than once If you know you will be meeting your opponent again, then the incentive to defect appears to evaporate Cooperation is the rational choice in the infinititely repeated prisoner s dilemma (Hurrah!) 10/4/2005 Multi-Agent_Interactions 17 10/4/2005 Multi-Agent_Interactions 18 Backwards Induction But suppose you both know that you will play the game exactly n times On round n - 1, you have an incentive to defect, to gain that extra bit of payoff But this makes round n 2 the last real, and so you have an incentive to defect there, too. This is the backwards induction problem. Playing the prisoner s dilemma with a fixed, finite, pre-determined, commonly known number of rounds, defection is the best strategy Axelrod s Tournament Suppose you play iterated prisoner s dilemma against a range of opponents What strategy should you choose, so as to maximize your overall payoff? Axelrod (1984) investigated this problem, with a computer tournament for programs playing the prisoner s dilemma 10/4/2005 Multi-Agent_Interactions 19 10/4/2005 Multi-Agent_Interactions 20 5

Strategies in Axelrod s Tournament ALLD: Always defect the hawk strategy; TIT-FOR-TAT: 1. On round u = 0, cooperate 2. On round u > 0, do what your opponent did on round u 1 TESTER: On 1st round, defect. If the opponent retaliated, then play TIT-FOR-TAT. Otherwise intersperse cooperation and defection. JOSS: As TIT-FOR-TAT, except periodically defect Recipes for Success in Axelrod s Tournament Axelrod suggests the following rules for succeeding in his tournament: Don t be envious: Don t play as if it were zero sum! Be nice: Start by cooperating, and reciprocate cooperation Retaliate appropriately: Always punish defection immediately, but use measured force don t overdo it Don t hold grudges: Always reciprocate cooperation immediately 10/4/2005 Multi-Agent_Interactions 21 10/4/2005 Multi-Agent_Interactions 22 Game of Chicken Consider another type of encounter the game of chicken: (Think of James Dean in Rebel without a Cause: swerving = coop, driving straight = defect.) Difference to prisoner s dilemma: Mutual defection is most feared outcome. (Whereas sucker s payoff is most feared in prisoner s dilemma.) Strategies (c,d) and (d,c) are in Nash equilibrium 10/4/2005 Multi-Agent_Interactions 23 Other Symmetric 2 x 2 Games Given the 4 possible outcomes of (symmetric) cooperate/defect games, there are 24 possible orderings on outcomes CC i CD i DC i DD Cooperation dominates DC i DD i CC i CD Deadlock. You will always do best by defecting DC i CC i DD i CD Prisoner s dilemma DC i CC i CD i DD Chicken 10/4/2005 Multi-Agent_Interactions 24 6