Stochastic Games and Bayesian Games

Similar documents
Stochastic Games and Bayesian Games

Repeated, Stochastic and Bayesian Games

G5212: Game Theory. Mark Dean. Spring 2017

M.Phil. Game theory: Problem set II. These problems are designed for discussions in the classes of Week 8 of Michaelmas term. 1

Lecture 5 Leadership and Reputation

FDPE Microeconomics 3 Spring 2017 Pauli Murto TA: Tsz-Ning Wong (These solution hints are based on Julia Salmi s solution hints for Spring 2015.

February 23, An Application in Industrial Organization

ECE 586BH: Problem Set 5: Problems and Solutions Multistage games, including repeated games, with observed moves

CHAPTER 14: REPEATED PRISONER S DILEMMA

Game Theory. Wolfgang Frimmel. Repeated Games

An introduction on game theory for wireless networking [1]

Introduction to Game Theory Lecture Note 5: Repeated Games

Economics 171: Final Exam

Game Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India October 2012

Repeated Games. EC202 Lectures IX & X. Francesco Nava. January London School of Economics. Nava (LSE) EC202 Lectures IX & X Jan / 16

Game Theory. Important Instructions

Duopoly models Multistage games with observed actions Subgame perfect equilibrium Extensive form of a game Two-stage prisoner s dilemma

Advanced Microeconomics

Microeconomics II. CIDE, MsC Economics. List of Problems

Microeconomic Theory II Preliminary Examination Solutions

Preliminary Notions in Game Theory

Warm Up Finitely Repeated Games Infinitely Repeated Games Bayesian Games. Repeated Games

Games of Incomplete Information

CMSC 474, Introduction to Game Theory 16. Behavioral vs. Mixed Strategies

Infinitely Repeated Games

CUR 412: Game Theory and its Applications, Lecture 12

Game Theory for Wireless Engineers Chapter 3, 4

Microeconomics of Banking: Lecture 5

HW Consider the following game:

GAME THEORY. Department of Economics, MIT, Follow Muhamet s slides. We need the following result for future reference.

MA300.2 Game Theory 2005, LSE

Chapter 8. Repeated Games. Strategies and payoffs for games played twice

Extensive-Form Games with Imperfect Information

FDPE Microeconomics 3 Spring 2017 Pauli Murto TA: Tsz-Ning Wong (These solution hints are based on Julia Salmi s solution hints for Spring 2015.

Repeated Games with Perfect Monitoring

Beliefs and Sequential Rationality

Economics 209A Theory and Application of Non-Cooperative Games (Fall 2013) Repeated games OR 8 and 9, and FT 5

BAYESIAN GAMES: GAMES OF INCOMPLETE INFORMATION

Not 0,4 2,1. i. Show there is a perfect Bayesian equilibrium where player A chooses to play, player A chooses L, and player B chooses L.

Sequential Rationality and Weak Perfect Bayesian Equilibrium

m 11 m 12 Non-Zero Sum Games Matrix Form of Zero-Sum Games R&N Section 17.6

Repeated Games. Econ 400. University of Notre Dame. Econ 400 (ND) Repeated Games 1 / 48

Decision Theory: Value Iteration

Spring 2017 Final Exam

Simon Fraser University Fall Econ 302 D200 Final Exam Solution Instructor: Songzi Du Wednesday December 16, 2015, 8:30 11:30 AM

Problem 3 Solutions. l 3 r, 1

Repeated Games. September 3, Definitions: Discounting, Individual Rationality. Finitely Repeated Games. Infinitely Repeated Games

Lecture Note Set 3 3 N-PERSON GAMES. IE675 Game Theory. Wayne F. Bialas 1 Monday, March 10, N-Person Games in Strategic Form

Introduction to Industrial Organization Professor: Caixia Shen Fall 2014 Lecture Note 5 Games and Strategy (Ch. 4)

General Examination in Microeconomic Theory SPRING 2014

Simon Fraser University Spring 2014

Finitely repeated simultaneous move game.

Answers to Problem Set 4

CMPSCI 240: Reasoning about Uncertainty

CUR 412: Game Theory and its Applications Final Exam Ronaldo Carpio Jan. 13, 2015

CS 331: Artificial Intelligence Game Theory I. Prisoner s Dilemma

The Nash equilibrium of the stage game is (D, R), giving payoffs (0, 0). Consider the trigger strategies:

Strategies and Nash Equilibrium. A Whirlwind Tour of Game Theory

6.254 : Game Theory with Engineering Applications Lecture 3: Strategic Form Games - Solution Concepts

6 Dynamic Games with Incomplete Information

Iterated Dominance and Nash Equilibrium

Introduction to Game Theory

CMPSCI 240: Reasoning about Uncertainty

CHAPTER 15 Sequential rationality 1-1

Ph.D. Preliminary Examination MICROECONOMIC THEORY Applied Economics Graduate Program June 2017

Introductory Microeconomics

Signaling Games. Farhad Ghassemi

6.207/14.15: Networks Lecture 9: Introduction to Game Theory 1

1 Solutions to Homework 4

Game Theory: Normal Form Games

Subgame Perfect Cooperation in an Extensive Game

SI Game Theory, Fall 2008

Outline Introduction Game Representations Reductions Solution Concepts. Game Theory. Enrico Franchi. May 19, 2010

January 26,

Economics 502 April 3, 2008

Player 2 L R M H a,a 7,1 5,0 T 0,5 5,3 6,6

Prisoner s dilemma with T = 1

PAULI MURTO, ANDREY ZHUKOV. If any mistakes or typos are spotted, kindly communicate them to

Early PD experiments

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2017

Notes for Section: Week 7

6.207/14.15: Networks Lecture 9: Introduction to Game Theory 1

TR : Knowledge-Based Rational Decisions

Answer Key: Problem Set 4

Finite Memory and Imperfect Monitoring

Introduction to Political Economy Problem Set 3

Rationalizable Strategies

High Frequency Repeated Games with Costly Monitoring

Regret Minimization and Security Strategies

arxiv: v1 [cs.gt] 12 Jul 2007

PAULI MURTO, ANDREY ZHUKOV

Outline for today. Stat155 Game Theory Lecture 19: Price of anarchy. Cooperative games. Price of anarchy. Price of anarchy

In the Name of God. Sharif University of Technology. Microeconomics 2. Graduate School of Management and Economics. Dr. S.

ECONS 424 STRATEGY AND GAME THEORY MIDTERM EXAM #2 ANSWER KEY

The folk theorem revisited

October 9. The problem of ties (i.e., = ) will not matter here because it will occur with probability

Discounted Stochastic Games with Voluntary Transfers

Outline for today. Stat155 Game Theory Lecture 13: General-Sum Games. General-sum games. General-sum games. Dominated pure strategies

Maintaining a Reputation Against a Patient Opponent 1

4. Beliefs at all info sets off the equilibrium path are determined by Bayes' Rule & the players' equilibrium strategies where possible.

Transcription:

Stochastic Games and Bayesian Games CPSC 532l Lecture 10 Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 1

Lecture Overview 1 Recap 2 Stochastic Games 3 Bayesian Games 4 Analyzing Bayesian games Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 2

Finitely Repeated Games Everything is straightforward if we repeat a game a finite number of times we can write the whole thing as an extensive-form game with imperfect information at each round players don t know what the others have done; afterwards they do overall payoff function is additive: sum of payoffs in stage games Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 3

Infinitely Repeated Games Consider an infinitely repeated game in extensive form: an infinite tree! Thus, payoffs cannot be attached to terminal nodes, nor can they be defined as the sum of the payoffs in the stage games (which in general will be infinite). Definition Given an infinite sequence of payoffs r 1, r 2,... for player i, the average reward of i is lim k k j=1 r j k. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 4

Nash Equilibria With an infinite number of equilibria, what can we say about Nash equilibria? we won t be able to construct an induced normal form and then appeal to Nash s theorem to say that an equilibrium exists Nash s theorem only applies to finite games Furthermore, with an infinite number of strategies, there could be an infinite number of pure-strategy equilibria! It turns out we can characterize a set of payoffs that are achievable under equilibrium, without having to enumerate the equilibria. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 5

Definitions Consider any n-player game G = (N, A, u) and any payoff vector r = (r 1, r 2,..., r n ). Let v i = min max u i (s i, s i ). s i S i s i S i i s minmax value: the amount of utility i can get when i play a minmax strategy against him Definition A payoff profile r is enforceable if r i v i. Definition A payoff profile r is feasible if there exist rational, non-negative values α a such that for all i, we can express r i as a A αu i(a), with a A α a = 1. a payoff profile is feasible if it is a convex, rational combination of the outcomes in G. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 6

Folk Theorem Theorem (Folk Theorem) Consider any n-player game G and any payoff vector (r 1, r 2,..., r n ). 1 If r is the payoff in any Nash equilibrium of the infinitely repeated G with average rewards, then for each player i, r i is enforceable. 2 If r is both feasible and enforceable, then r is the payoff in some Nash equilibrium of the infinitely repeated G with average rewards. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 7

Folk Theorem (Part 1) Payoff in Nash enforceable Part 1: Suppose r is not enforceable, i.e. r i < v i for some i. Then consider a deviation of this player i to b i (s i (h)) for any history h of the repeated game, where b i is any best-response action in the stage game and s i (h) is the equilibrium strategy of other players given the current history h. By definition of a minmax strategy, player i will receive a payoff of at least v i in every stage game if he adopts this strategy, and so i s average reward is also at least v i. Thus i cannot receive the payoff r i < v i in any Nash equilibrium. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 8

Folk Theorem (Part 2) Feasible and enforceable Nash Part 2: Since r is a feasible payoff profile, we can write it as r i = ( βa a A u i (a), where β a and γ are non-negative integers. 1 γ ) Since the combination was convex, we have γ = a A β a. We re going to construct a strategy profile that will cycle through all outcomes a A of G with cycles of length γ, each cycle repeating action a exactly β a times. Let (a t ) be such a sequence of outcomes. Let s define a strategy s i of player i to be a trigger version of playing (a t ): if nobody deviates, then s i plays a t i in period t. However, if there was a period t in which some player j i deviated, then s i will play (p j ) i, where (p j ) is a solution to the minimization problem in the definition of v j. 1 Recall that α a were required to be rational. So we can take γ to be their common denominator. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 9

Folk Theorem (Part 2) Feasible and enforceable Nash First observe that if everybody plays according to s i, then, by construction, player i receives average payoff of r i (look at averages over periods of length γ). Second, this strategy profile is a Nash equilibrium. Suppose everybody plays according to s i, and player j deviates at some point. Then, forever after, player j will receive his min max payoff v j r j, rendering the deviation unprofitable. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 9

Lecture Overview 1 Recap 2 Stochastic Games 3 Bayesian Games 4 Analyzing Bayesian games Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 10

Introduction What if we didn t always repeat back to the same stage game? A stochastic game is a generalization of repeated games agents repeatedly play games from a set of normal-form games the game played at any iteration depends on the previous game played and on the actions taken by all agents in that game A stochastic game is a generalized Markov decision process there are multiple players one reward function for each agent the state transition function and reward functions depend on the action choices of both players Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 11

Formal Definition Definition A stochastic game is a tuple (Q, N, A, P, R), where Q is a finite set of states, N is a finite set of n players, A = A 1 A n, where A i is a finite set of actions available to player i, P : Q A Q [0, 1] is the transition probability function; P (q, a, ˆq) is the probability of transitioning from state s to state ˆq after joint action a, and R = r 1,..., r n, where r i : Q A R is a real-valued payoff function for player i. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 12

Remarks This assumes strategy space is the same in all games otherwise just more notation Again we can have average or discounted payoffs. Interesting special cases: zero-sum stochastic game single-controller stochastic game transitions (but not payoffs) depend on only one agent Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 13

Strategies What is a pure strategy? Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 14

Strategies What is a pure strategy? pick an action conditional on every possible history of course, mixtures over these pure strategies are possible too! Some interesting restricted classes of strategies: behavioral strategy: s i (h t, a ij ) returns the probability of playing action a ij for history h t. the substantive assumption here is that mixing takes place at each history independently, not once at the beginning of the game Markov strategy: s i is a behavioral strategy in which s i (h t, a ij ) = s i (h t, a ij ) if q t = q t, where q t and q t are the final states of h t and h t, respectively. for a given time t, the distribution over actions only depends on the current state stationary strategy: s i is a Markov strategy in which s i (h t1, a ij ) = s i (h t 2, a ij ) if q t1 = q t 2, where q t1 and q t 2 are the final states of h t1 and h t 2, respectively. no dependence even on t Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 14

Equilibrium (discounted rewards) Markov perfect equilibrium: a strategy profile consisting of only Markov strategies that is a Nash equilibrium regardless of the starting state analogous to subgame-perfect equilibrium Theorem Every n-player, general sum, discounted reward stochastic game has a Markov perfect equilibrium. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 15

Equilibrium (average rewards) Irreducible stochastic game: every strategy profile gives rise to an irreducible Markov chain over the set of games irreducible Markov chain: possible to get from every state to every other state during the (infinite) execution of the stochastic game, each stage game is guaranteed to be played infinitely often for any strategy profile without this condition, limit of the mean payoffs may not be defined Theorem For every 2-player, general sum, average reward, irreducible stochastic game has a Nash equilibrium. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 16

A folk theorem Theorem For every 2-player, general sum, irreducible stochastic game, and every feasible outcome with a payoff vector r that provides to each player at least his minmax value, there exists a Nash equilibrium with a payoff vector r. This is true for games with average rewards, as well as games with large enough discount factors (i.e. with players that are sufficiently patient). Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 17

Lecture Overview 1 Recap 2 Stochastic Games 3 Bayesian Games 4 Analyzing Bayesian games Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 18

Fun Game Choose a phone number none of your neighbours knows; consider it to be ABC-DEFG Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 19

Fun Game Choose a phone number none of your neighbours knows; consider it to be ABC-DEFG take DE as your valuation play a first-price auction with three neighbours, where your utility is your valuation minus the amount you pay Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 19

Fun Game Choose a phone number none of your neighbours knows; consider it to be ABC-DEFG take DE as your valuation play a first-price auction with three neighbours, where your utility is your valuation minus the amount you pay now play the auction again, same neighbours, same valuation Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 19

Fun Game Choose a phone number none of your neighbours knows; consider it to be ABC-DEFG take DE as your valuation play a first-price auction with three neighbours, where your utility is your valuation minus the amount you pay now play the auction again, same neighbours, same valuation now play again, with FG as your valuation Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 19

Fun Game Choose a phone number none of your neighbours knows; consider it to be ABC-DEFG take DE as your valuation play a first-price auction with three neighbours, where your utility is your valuation minus the amount you pay now play the auction again, same neighbours, same valuation now play again, with FG as your valuation Questions: what is the role of uncertainty here? can we model this uncertainty using an imperfect information extensive form game? Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 19

Fun Game Choose a phone number none of your neighbours knows; consider it to be ABC-DEFG take DE as your valuation play a first-price auction with three neighbours, where your utility is your valuation minus the amount you pay now play the auction again, same neighbours, same valuation now play again, with FG as your valuation Questions: what is the role of uncertainty here? can we model this uncertainty using an imperfect information extensive form game? imperfect info means not knowing what node you re in in the info set here we re not sure what game is being played (though if we allow a move by nature, we can do it) Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 19

Introduction So far, we ve assumed that all players know what game is being played. Everyone knows: the number of players the actions available to each player the payoff associated with each action vector Why is this true in imperfect information games? We ll assume: 1 All possible games have the same number of agents and the same strategy space for each agent; they differ only in their payoffs. 2 The beliefs of the different agents are posteriors, obtained by conditioning a common prior on individual private signals. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 20

Definition 1: Information Sets Bayesian game: a set of games that differ only in their payoffs, a common prior defined over them, and a partition structure over the games for each agent. Definition (Bayesian Game: Information Sets) A Bayesian game is a tuple (N, G, P, I) where N is a set of agents, G is a set of games with N agents each such that if g, g G then for each agent i N the strategy space in g is identical to the strategy space in g, P Π(G) is a common prior over games, where Π(G) is the set of all probability distributions over G, and I = (I 1,..., I N ) is a set of partitions of G, one for each agent. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 21

Definition 1: Example 6 Richer Representations: Beyond the Normal and Extensive Forms I 2,1 I 2,2 I 1,1 A 1 0 0 3 p = 0.2 B 2 1 1 4 p = 0.1 C 1 3 2 1 p = 0.1 D 1 0 4 0 p = 0 I 1,2 E 3 2 0 3 p = 0.1 F 2 4 1 1 p = 0.1 G 0 1 3 1 p = 0.25 H 2 0 3 4 p = 0.15 Figure 6.7 A Bayesian game Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 22

Definition 2: Extensive Form with Chance Moves Add an agent, Nature, who follows a commonly known mixed strategy. Thus, reduce Bayesian games to extensive form games of imperfect information. This definition is cumbersome for the same reason that IIEF is a cumbersome way of representing matrix games like Prisoner s dilemma however, it makes sense when the agents really do move sequentially, and at least occasionally observe each other s actions. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 23

the Recap choices of others. StochasticThus, Games we have reduced Bayesian Games games of incomplete Analyzing information Bayesian gamesto games of imperfect information, albeit ones with chance moves. These chance moves of Nature require minor adjustments of existing definitions, replacing payoffs by their expectations, given Nature s moves. 5 For example, the Bayesian game of Figure 6.7 can be represented in extensive form as depicted in Figure 6.8. Definition 2: Example Nature A, B C, D E, F G, H 1 1 U D U D U D U D 2 2 L R L R L R L R L R L R L R L R 4 1 1 10 5 1 3 5 17 3 3 3 3 1 3 2 1 2 3 2 2 4 8 3 8 Figure 6.8 The Bayesian game from Figure 6.7 in extensive form Although this second definition of Bayesian games can be more initially intuitive than our first definition, it can also be more cumbersome to work with. This is be- Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 24

Definition 3: Epistemic Types Directly represent uncertainty over utility function using the notion of epistemic type. Definition A Bayesian game is a tuple (N, A, Θ, p, u) where N is a set of agents, A = (A 1,..., A n ), where A i is the set of actions available to player i, Θ = (Θ 1,..., Θ n ), where Θ i is the type space of player i, p : Θ [0, 1] is the common prior over types, u = (u 1,..., u n ), where u i : A Θ R is the utility function for player i. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 25

Definition 3: Example 150 6 Richer Representations: Beyond the Normal and Extensive Forms I 2,1 I 2,2 I 1,1 A 1 0 0 3 p = 0.2 B 2 1 1 4 p = 0.1 C 1 3 2 1 p = 0.1 D 1 0 4 0 p = 0 I 1,2 E 3 2 0 3 p = 0.1 F 2 4 1 1 p = 0.1 G 0 1 3 1 p = 0.25 H 2 0 3 4 p = 0.15 6 Richer Representations: Beyond the Normal and Extensive Forms a 1 a 2 θ 1 θ 2 u 1 U L θ 1,1 θ 2,1 4/3 U L θ 1,1 θ 2,2 1 Figure 6.7 A Bayesian game receive individual signals about Nature s choice, and these are captured by their information sets, in a standard way. The agents have no additional information; in particular, the information sets capture the fact that agents make their choices without knowing the choices of others. Thus, we have reduced games of incomplete information to games of imperfect information, albeit ones with chance moves. These chance moves of Nature require minor adjustments of existing definitions, replacing payoffs by their expectations, given Nature s moves. 5 For example, the Bayesian game of Figure 6.7 can be represented in extensive form as depicted in Figure 6.8. U L θ 1,2 θ 2,1 5/2 U L θ 1,2 θ 2,2 3/4 U R θ 1,1 θ 2,1 1/3 U R θ 1,1 θ 2,2 3 U R θ 1,2 θ 2,1 3 U R θ 1,2 θ 2,2 5/8 a 1 a 2 θ 1 θ 2 u 1 D L θ 1,1 θ 2,1 1/3 D L θ 1,1 θ 2,2 2 D L θ 1,2 θ 2,1 1/2 D L θ 1,2 θ 2,2 3 D R θ 1,1 θ 2,1 10/3 D R θ 1,1 θ 2,2 1 D R θ 1,2 θ 2,1 2 D R θ 1,2 θ 2,2 17/8 Figure 6.9 Utility function u 1 fornature the Bayesian game from Figure 6.7. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 26

Lecture Overview 1 Recap 2 Stochastic Games 3 Bayesian Games 4 Analyzing Bayesian games Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 27

Strategies Pure strategy: s i : Θ i A i a mapping from every type agent i could have to the action he would play if he had that type. Mixed strategy: s i : Θ i Π(A i ) a mapping from i s type to a probability distribution over his action choices. s j (a j θ j ) the probability under mixed strategy s j that agent j plays action a j, given that j s type is θ j. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 28

Expected Utility Three meaningful notions of expected utility: ex-ante the agent knows nothing about anyone s actual type; ex-interim an agent knows his own type but not the types of the other agents; ex-post the agent knows all agents types. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 29

Ex-interim expected utility Definition (Ex-interim expected utility) Agent i s ex-interim expected utility in a Bayesian game (N, A, Θ, p, u), where i s type is θ i and where the agents strategies are given by the mixed strategy profile s, is defined as EU i (s θ i ) = p(θ i θ i ) s j (a j θ j ) u i (a, θ i, θ i ). θ i Θ i a A j N i must consider every θ i and every a in order to evaluate u i (a, θ i, θ i ). i must weight this utility value by: the probability that a would be realized given all players mixed strategies and types; the probability that the other players types would be θ i given that his own type is θ i. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 30

Ex-ante expected utility Definition (Ex-ante expected utility) Agent i s ex-ante expected utility in a Bayesian game (N, A, Θ, p, u), where the agents strategies are given by the mixed strategy profile s, is defined as EU i (s) = θ i Θ i p(θ i )EU i (s θ i ) or equivalently as EU i (s) = p(θ) s j (a j θ j ) u i (a, θ). θ Θ a A j N Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 31

Ex-post expected utility Definition (Ex-post expected utility) Agent i s ex-post expected utility in a Bayesian game (N, A, Θ, p, u), where the agents strategies are given by s and the agent types are given by θ, is defined as EU i (s, θ) = a A j N s j (a j θ j ) u i (a, θ). The only uncertainty here concerns the other agents mixed strategies, since i knows everyone s type. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 32

Best response Definition (Best response in a Bayesian game) The set of agent i s best responses to mixed strategy profile s i are given by BR i (s i ) = arg max EU i (s i, s i ). s i S i it may seem odd that BR is calculated based on i s ex-ante expected utility. However, write EU i (s) as θ i Θ i p(θ i )EU i (s θ i ) and observe that EU i (s i, s i θ i ) does not depend on strategies that i would play if his type were not θ i. Thus, we are in fact performing independent maximization of i s ex-interim expected utility conditioned on each type that he could have. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 33

Nash equilibrium Definition (Bayes-Nash equilibrium) A Bayes-Nash equilibrium is a mixed strategy profile s that satisfies i s i BR i (s i ). we can also construct an induced normal form for Bayesian games the numbers in the cells will correspond to ex-ante expected utilities however as argued above, as long as the strategy space is unchanged, best responses don t change between the ex-ante and ex-interim cases. Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 34

ex-post Equilibrium Definition (ex-post equilibrium) A ex-post Bayes-Nash equilibrium is a mixed strategy profile s that satisfies θ, i, s i arg max s i S i EU i (s i, s i, θ). somewhat similar to dominant strategy, but not quite EP: agents do not need to have accurate beliefs about the type distribution DS: agents do not need to have accurate beliefs about others strategies Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 35