Mixed Strategies. Samuel Alizon and Daniel Cownden February 4, 2009

Similar documents
Chapter 10: Mixed strategies Nash equilibria, reaction curves and the equality of payoffs theorem

Problem Set 2 - SOLUTIONS

SI 563 Homework 3 Oct 5, Determine the set of rationalizable strategies for each of the following games. a) X Y X Y Z

6.254 : Game Theory with Engineering Applications Lecture 3: Strategic Form Games - Solution Concepts

Best-Reply Sets. Jonathan Weinstein Washington University in St. Louis. This version: May 2015

PAULI MURTO, ANDREY ZHUKOV

MS&E 246: Lecture 5 Efficiency and fairness. Ramesh Johari

Version 01 as of May 12, 2014 Author: Steven Sagona To be submitted to PRL (not really)

6.207/14.15: Networks Lecture 10: Introduction to Game Theory 2

Player 2 H T T -1,1 1, -1

ECE 586BH: Problem Set 5: Problems and Solutions Multistage games, including repeated games, with observed moves

Almost essential MICROECONOMICS

Game theory and applications: Lecture 1

Yao s Minimax Principle

Solution to Tutorial 1

Solution to Tutorial /2013 Semester I MA4264 Game Theory

Not 0,4 2,1. i. Show there is a perfect Bayesian equilibrium where player A chooses to play, player A chooses L, and player B chooses L.

Economics 171: Final Exam

Can we have no Nash Equilibria? Can you have more than one Nash Equilibrium? CS 430: Artificial Intelligence Game Theory II (Nash Equilibria)

10.1 Elimination of strictly dominated strategies

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2017

preferences of the individual players over these possible outcomes, typically measured by a utility or payoff function.

ANASH EQUILIBRIUM of a strategic game is an action profile in which every. Strategy Equilibrium

Elements of Economic Analysis II Lecture X: Introduction to Game Theory

Introduction to Industrial Organization Professor: Caixia Shen Fall 2014 Lecture Note 5 Games and Strategy (Ch. 4)

FDPE Microeconomics 3 Spring 2017 Pauli Murto TA: Tsz-Ning Wong (These solution hints are based on Julia Salmi s solution hints for Spring 2015.

Game Theory Fall 2003

Iterated Dominance and Nash Equilibrium

Notes for Section: Week 4

CMSC 474, Introduction to Game Theory 16. Behavioral vs. Mixed Strategies

Elements of Economic Analysis II Lecture XI: Oligopoly: Cournot and Bertrand Competition

Outline for today. Stat155 Game Theory Lecture 13: General-Sum Games. General-sum games. General-sum games. Dominated pure strategies

Exercises Solutions: Oligopoly

Game Theory. VK Room: M1.30 Last updated: October 22, 2012.

Mixed Strategies. In the previous chapters we restricted players to using pure strategies and we

Game Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India October 2012

Regret Minimization and Security Strategies

March 30, Why do economists (and increasingly, engineers and computer scientists) study auctions?

Microeconomics III Final Exam SOLUTIONS 3/17/11. Muhamet Yildiz

Introduction to Game Theory

ECE 586GT: Problem Set 1: Problems and Solutions Analysis of static games

Microeconomics of Banking: Lecture 5

January 26,

FDPE Microeconomics 3 Spring 2017 Pauli Murto TA: Tsz-Ning Wong (These solution hints are based on Julia Salmi s solution hints for Spring 2015.

CS 331: Artificial Intelligence Game Theory I. Prisoner s Dilemma

Ph.D. Preliminary Examination MICROECONOMIC THEORY Applied Economics Graduate Program June 2017

Introduction to Game Theory Evolution Games Theory: Replicator Dynamics

Problem 3 Solutions. l 3 r, 1

Game Theory: Normal Form Games

Strategies and Nash Equilibrium. A Whirlwind Tour of Game Theory

m 11 m 12 Non-Zero Sum Games Matrix Form of Zero-Sum Games R&N Section 17.6

Outline Introduction Game Representations Reductions Solution Concepts. Game Theory. Enrico Franchi. May 19, 2010

Econ 101A Final exam May 14, 2013.

Game Theory and Economics Prof. Dr. Debarshi Das Department of Humanities and Social Sciences Indian Institute of Technology, Guwahati

Game Theory with Applications to Finance and Marketing, I

Microeconomic Theory II Preliminary Examination Solutions Exam date: June 5, 2017

Best Reply Behavior. Michael Peters. December 27, 2013

PAULI MURTO, ANDREY ZHUKOV. If any mistakes or typos are spotted, kindly communicate them to

CMPSCI 240: Reasoning about Uncertainty

Unraveling versus Unraveling: A Memo on Competitive Equilibriums and Trade in Insurance Markets

EC487 Advanced Microeconomics, Part I: Lecture 9

TEACHING STICKY PRICES TO UNDERGRADUATES

1 Two Period Exchange Economy

Game Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India August 2012

Math 135: Answers to Practice Problems

Advanced Microeconomic Theory EC104

Exercises Solutions: Game Theory

Jianfei Shen. School of Economics, The University of New South Wales, Sydney 2052, Australia

Solution Problem Set 2

S 2,2-1, x c C x r, 1 0,0

Game Theory: Global Games. Christoph Schottmüller

Rationalizable Strategies

Game Theory Notes: Examples of Games with Dominant Strategy Equilibrium or Nash Equilibrium

Econ 101A Final exam May 14, 2013.

The Ohio State University Department of Economics Second Midterm Examination Answers

Chapter 19 Optimal Fiscal Policy

Econ 711 Homework 1 Solutions

ECO410H: Practice Questions 2 SOLUTIONS

3 Arbitrage pricing theory in discrete time.

Continuing game theory: mixed strategy equilibrium (Ch ), optimality (6.9), start on extensive form games (6.10, Sec. C)!

An Introduction to the Mathematics of Finance. Basu, Goodman, Stampfli

Finding Equilibria in Games of No Chance

ECONS 424 STRATEGY AND GAME THEORY HANDOUT ON PERFECT BAYESIAN EQUILIBRIUM- III Semi-Separating equilibrium

Game Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India August 2012

KIER DISCUSSION PAPER SERIES

On Forchheimer s Model of Dominant Firm Price Leadership

Answers to Problem Set 4

Name: Midterm #1 EconS 425 (February 20 th, 2015)

Notes for Section: Week 7

REPEATED GAMES. MICROECONOMICS Principles and Analysis Frank Cowell. Frank Cowell: Repeated Games. Almost essential Game Theory: Dynamic.

CS364A: Algorithmic Game Theory Lecture #14: Robust Price-of-Anarchy Bounds in Smooth Games

MA200.2 Game Theory II, LSE

A Decentralized Learning Equilibrium

Game Theory Problem Set 4 Solutions

Now we return to simultaneous-move games. We resolve the issue of non-existence of Nash equilibrium. in pure strategies through intentional mixing.

Université du Maine Théorie des Jeux Yves Zenou Correction de l examen du 16 décembre 2013 (1 heure 30)

All Equilibrium Revenues in Buy Price Auctions

The Ohio State University Department of Economics Econ 601 Prof. James Peck Extra Practice Problems Answers (for final)

ELEMENTS OF MATRIX MATHEMATICS

HE+ Economics Nash Equilibrium

Transcription:

Mixed Strategies Samuel Alizon and Daniel Cownden February 4, 009 1 What are Mixed Strategies In the previous sections we have looked at games where players face uncertainty, and concluded that they choose strategies to maximize their expected payoff. We now extend our notion of strategy to what is called a mixed strategy, by allowing players to incorporate probability into their strategies. The strategies that we have been referring to thus far will now be called pure strategies, and a mixed strategy will refer to a set of pure strategies for a player, coupled with a probability distribution which tells the player which of the pure strategies to play with which probability. definition 6. Let s i, i 1,,..., n, be the pure strategies available to a player in a game. A mixed strategy, denoted σ, for this player is a probability distribution over s 1, s,..., s n ; i.e. σ = p 1 s 1, p s,... p n s n where each p 1, p,..., p n are all nonnegative and n i=1 p i = 1, and each p i is the probabilities with which the player will play the corresponding pure strategy s i. Note that pure strategies can be represented as mixed strategies. Consider the case where all but one of the p i s are zero, in a mixed strategy σ, then the one p i which is not zero must be one, and so a single pure strategy is played with certainty. In this case the mixed strategy is equivalent to the pure strategy. Also a finicky note on notation: Previously s i has denoted a strategy of player i, where i indexes the players. In this definition s i denotes the ith strategy available to a player, where i indexes the strategies. It should be clear from context what i is indexing, but to keep things clear try to index strategies with a super script and players with a sub script. Now that we know what a mixed strategy is, we would like to know how to compute the payoffs when mixed strategies are involved. This is a simple application of the expected utility principle. Describing precisely how to do this would be tedious and uninstructive, so instead here is an example. Consider a two player simultaneous game where player one has pure strategies s 1,..., s n and player two has pure strategies t 1,..., t m. We have payoff functions π 1 (s i, t j ) for player one and π (s i, t j ) for player two, i [1,,..., n] and j [1,,..., m]. Now suppose that player one is playing mixed strategy σ = p 1 s 1, p s,..., p n s n and that player two is playing mixed strategy τ = q 1 t 1, q t,..., q m t m. We calculate π 1 (σ, τ) as follows. We first apply the expected utility principle and decompose π 1 (σ, τ) in terms of σ yielding π 1 (σ, τ) = n i=1 p iπ 1 (s i, τ). We then apply the expected utility principle again and decompose each on of the π 1 (s i, τ) in terms of τ. This yields π 1 (s i, τ) = m j=1 p iπ 1 (s i, t j ). Putting it all together 1

we get π 1 (σ, τ) = n m i=1 j=1 p iq j π 1 (s i, t j ). We do the same expansion for player two and get π (σ, τ) = n m i=1 j=1 p iq j π (s i, t j ). Notice that the order in which we decompose does not matter, so we could have decomposed π 1 (σ, τ) in terms of τ first and then in terms of σ and arrived at the same answer. The definition of a Nash equilibrium is still the same as before only now we allow the set of all possible strategies of a player to include mixed strategies. One of the reasons we are interested i n mixed strategies is that there is a powerful theorem about mixed strategies and Nash equilibria. The theorem can be stated as follows. Theorem 1. If each player in an N-player game has a finite number of pure strategies, then the game has a, not necessarily unique, Nash equilibrium, provided mixed strategies are allowed. Note that this does not say there will always be a mixed strategy Nash equilibrium, it is quite possible that there will only be pure strategy equilibria. We do not give a proof of this theorem here, as it is beyond the scope of this course, but if you were interested you could look up John Nashs 1950 Equilibrium Points in N-person Games in the Proceedings of the National Academy of Science. Before when we encountered games with no Nash equilibrium it was difficult to predict which strategies the players would choose, with the introduction of mixed strategies we are now able to make probabilistic predictions about the actions of rational players in a game. Mixed strategies will also become an important tool for thinking about games played within an evolving population, but we will leave this for later sections. Here is a simple example to where you can try out your ideas of what a mixed strategy is. At a market, two farmers (1 and ) have a choice between selling vegetables, V, or meat, M,. If they choose to sell different things then they each are able to make an equal amount of profit. If they both sell the same thing they are unable to make a profit. The payoff matrix is simple. Note that each player has only two pure strategies (M or V). M V M 0,0 1,1 V 1,1 0,0 There are two obvious Nash equilibria in this game, (M, V ) and (V, M) however, unless the farmers have a means of coordinating their actions (e.g. phoning each other before going over to the market, deciding what to do at a previous market), there is no reason why these Nash equilibria should be attained. Attempting to maximize his expected payoff, farmer 1 might use the following reasoning: I will bring vegetables half of the time and meat half of the time. This way, I should be able to sell things half of the time. This plan of action is precisely a mixed strategy because each pure strategy with a given probability. Lets see if farmer 1s reasoning is rational, that is does it correspond to a Nash equilibrium. Let σ be the mixed strategy of farmer 1, and τ the mixed strategy of farmer. Let p be the probability with which farmer 1 sells meat and let q be the probability with which farmer sells meat. Clearly then the probabilities of selling vegetables are (1 p) and (1 q) for farmer one and two respectively. Then we can write: σ = pm, (1 p)v and τ = qm, (1 q)v. Now let s look at

the payoff functions π 1 (σ, τ) = pq π 1 (M, M)+p(1 q) π 1 (M, V )+(1 p)q π 1 (V, M)+(1 p)(1 q) π 1 (V, V ) (1) Plugging in the values from the payoff matrix we get: π 1 (σ, τ) = pq 0 + p(1 q) 1 + (1 p)q 1 + (1 p)(1 q) 0 () = p(1 q) + (1 p)q (3) = p(1 q) + q (4) = π 1 (p, q) (5) To find the best response of farmer 1 to q we would usually take the partial derivative of π 1 (p, q) with respect to p and set it equal to zero and then solve for p, but π 1 (p, q) is linear in p and so upon inspection of 4 we see that if q < 1 then p should be as big as possible and if q > 1 then p should be as small as possible, and if q = 1 then it doesn t matter what p is, so: 1 q < 1 B 1 (q) = 0 q > 1 (6) anything q = 1 Note that we used the expected utility principle twice to simplify the expression, expanding first the mixed-strategy of farmer 1 then that of farmer. This game is completely symmetric because you cannot make any distinction between the players: they play at the same time, have the same strategy sets and the same payoff functions. So we know that player two s best response function looks like 1 p < 1 B (p) = 0 p > 1 (7) anything p = 1 Looking at these best response functions we see that they intersect at exactly three point (0, 1), (1, 0) and ( 1, 1 ). The first two Nash equilibria correspond to the pure strategy equilibria we were able to identify earlier. The third equilibrium point refers to the when both farmers bring Meat or Vegetables with equal probability. The expected value when both play this mixed strategy Nash equilibrium is π 1 (p, q) = p(1 q) + q (8) = 1 (1 1 ) + 1 = 1 (9) (10) So the expected payoff when playing the mixed strategy Nash equilibrium is half that of playing at the pure strategy equilibrium, but as we mentioned before since the farmers have no way of coordinating the mixed strategy Nash equilibria is their best option. You may have noticed that the way we found the Nash equilibria in the above game was identical to how we found pure strategy Nash equilibria in symmetric games with continuous strategies. That is because in 3

the above example allowing the farmers to play mixed strategies made the game equivalent to one where each player chose a number between zero and one, p and q respectively, and then received a payoff of π 1 (p, q) = p(1 q)+q and π (p, q) = q(1 p) + p respectively. This is precisely the sort of symmetric game that we saw in the previous section on continuous strategies. You can imagine that while this continuous strategies technique for finding mixed strategy Nash equilibria works well for simple games, a similar game where players had even slightly larger pure strategy sets would quickly become a computational nightmare. Fortunately there is a useful theorem which will help us find mixed strategy Nash equilibria, in games where previous techniques would have failed. Finding Mixed Strategy Nash Equilibria We now present the this useful theorem called the Fundamental Theorem of Mixed Strategy Nash Equilibria. Theorem. Let σ = (σ 1, σ,..., σ n ) be a mixed strategy profile for an n-player game. For any player i = 1,,..., n let σ i represent the mixed strategies used by all the players other than player i, let S i be the finite set of pure strategies available to player i, and let π i (s, σ i ) s S i be the payoff to player i when playing s against σ i. Note that i is indexing players here and not strategies, hence it is a subscript. Then σ is a Nash equilibrium the following two conditions hold: 1. If s, s S i are two strategies that occur with positive proability in σ i, then π i (s, σ i ) = π i (s, σ i ). If s, s S i where s occurs with positive probability in σ i and s occurs with zero probability in σ i, then π i (s, σ i ) > π i (s, σ i ) Why this theorem is true, and how it will be useful may not be immediately clear. We will give an intuitive argument for why the theorem is true, and then we will look at some examples of how the theorem is used. In words the theorem is saying that at a Nash equilibrium, σ, in mixed strategies, each player i receives the same expected payoff for playing each of the pure strategies that occur with positive probability within his equilibrium mixed strategy σ i when playing against σ i, the Nash equilibrium mixed strategies of all the other players. To see why this must be, suppose that the payoffs for some of the strategies in σ i were different. Then one of the strategies is worse and one is better in terms of expected payoff, so player i would choose to play the strategy with lower payoff with probability zero, since player i wants to maximize her expected utility. This implies that σ was not a Nash equilibrium mixed strategy, since player i had incentive to change their strategy. The theorem also says that at a Nash equilibrium, σ, in mixed strategies, all the pure strategies played with positive probability in σ i, the mixed strategy of player i, give a higher expected payoff to player i than those pure strategies which are played with probability zero. To see why this must be so suppose that one of the pure strategies played with probability zero actually gives player i a higher expected payoff than one of the strategies played with positive probability. If this were the case player i would play this better pure strategy preferably 4

over any of the strategies which occur with positive proability in σ i. This implies σ is not a Nash equilibrium since player i had incentive to change their strategy. These two arguments show that, σ is a Nash equilibrium implies conditions 1 and, because we showed that if 1 and did not hold than σ was not a Nash equilibrium. An argument for the other direction of implication is as follows. Suppose that if s, s S i are two strategies that occur with positive proability in σ i, then π i (s, σ i ) = π i (s, σ i ), and that if s, s S i where s occurs with positive probability in σ i and s occurs with zero probability in σ i, then π i (s, σ i ) > π i (s, σ i ). Then clearly player i has no incentive to change his strategy, since increasing the probability of playing a pure strategy that is currently being played with probability zero will decrease his expected payoff, and rearranging the probability with which the pure strategies which are played with positive probability are played will have no effect on the expected payoff. Since player i has no incentive to change his strategy we can say that in this case σ i is a best response to σ i Since this is true for every player, we have that σ is a Nash equilibrium since each player is playing a best response. So we have that conditions 1 and imply that σ is a Nash equilibrium. Now that you know why the theorem is true, we show how it can be used. First we will apply the theorem to solving the simple farmer problem of previous section. Then we will use the insights gained from that to write up a recipe and finally we will apply that recipe to a significantly more difficult problem. Recall the farmer problem where two farmers had to decide whether to sell Meat or Vegetables at a market. The payoff matrix looked like: M V M 0,0 1,1 V 1,1 0,0 We start by supposing that there is a mixed strategy Nash equilibrium (σ, τ ). Where σ = p M, (1 p )V and τ = q M, (1 q )V are the mixed strategies of farmer one and two respectively. Applying the theorem the first condition of the theorem we have that if (σ, τ ) is indeed a Nash equilibrium then the following must hold: π 1 (M, τ ) = π 1 (V, τ ) (11) π (σ, M) = π σ, V ) (1) Working with equation 11, we use the expected utility principle to expand both sides of the equation and we get: π 1 (M, τ ) = q π 1 (M, M) + (1 q )π 1 (M, V ) = 0q + 1(1 q ) = 1 q = (13) π 1 (V, τ ) = q π 1 (V, M) + (1 q )π 1 (V, V ) = 1q + 0(1 q ) = q ) (14) So for (σ, τ ) to be a Nash equilibrium q = 1 q must be satisfied, which it is when q = 1. We can expand equation 1 and using an identical procedure deduce that for (σ, τ ) to be a Nash equilibrium then p = 1. So we can conclude that the mixed strategy Nash equilibrium for this simple farmer game is (σ, τ ) where σ = τ = 1 M, 1 V, which agrees with our earlier analysis of the game. A generalization of what we just did is given in the following recipe. 5

.1 Recipe:How to Find Mixed Strategy Nash Equilibria Consider an n-player, simultaneous game, where the players are indexed by i [1,,..., n]. Let player i s strategy set be denoted by S i = [s 1 i, s i,..., sm(i) i ] where m(i) is the number of pure strategies in player i s strategy set. Let σ i = p 1 i s1 i, p i s i,..., pm(i) i s m(i) i denote a mixed strategy of player i. We first assume that there is a mixed strategy Nash equilibrium denoted (σ1, σ,..., σn). If this assumption is false then we will either find a contradiction or a mixed strategy Nash equilibrium that is equivalent to a pure strategy Nash equilibrium. Applying condition 1 of the theorem a whole lot of times we generate lots of equations, each of the form: π i (s i, σ i ) = π i (s i, σ i) where s i, s i S i. We will then use the expected utility principle to expand these equations until they are expressed in terms of the payoff functions of pure strategies. We will then solve this system of equation for the p j i s and this will give us the Nash equilibrium strategies of the players. 6