COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture #24 Scribe: Jordan Ash May 1, 2014
|
|
- Silas Singleton
- 5 years ago
- Views:
Transcription
1 COS 5: heoretical Machine Learning Lecturer: Rob Schapire Lecture #24 Scribe: Jordan Ash May, 204 Review of Game heory: Let M be a matrix with all elements in [0, ]. Mindy (called the row player) chooses row i while Max (called the column player) chooses column j. In this case, from Mindy s expected loss is: Loss = M(i, j) Alternatively, Mindy could select a move randomly from a distribution over the rows and Max could select a move randomly from a distribution over the columns. Here, the expected loss for Mindy is: Loss = X i,j (i)m(i, j)(j) = M = M(, ) and are called mixed strategies, while i and j are called pure strategies. 2 Minimax heorem: In some games, such as Rock, aper, Scissors, players move at exactly the same time. In this way, both players have the same information available to them at the time of moving. Now we suppose that Mindy plays first, followed by Max. Max knows the that Mindy chose, and further knows M(, ) for any he chooses. Consequently, he chooses a that maximizes M(, ). Because Mindy knows that Max will choose = arg max M(, ) for any she chooses, she selects a that minimizes max M(, ). hus, if Mindy goes first, she could expect to su er a loss of min max M(, ). Overall, it may initially seem like the player to go second has an advantage because she has more information available to her. From Mindy s perspective again, this leads to: max min M(, ) apple min max M(, ) So Mindy playing after Max seems to be better than if the two play in reverse order. However, John von Neumann showed that the expected outcome of a game is always the same, regardless of the order in which players move. v = max min M(, ) =min max M(, )
2 Here, v denotes the value of the game. his may seem counterintuitive, because the player that goes second has more information available to her at the time of choosing a move. We will prove the above statement using an online learning algorithm. Let: hen, = arg min = arg max max M(, ) min M(, ) 8 : M(,) apple v () 8 : M(, ) v (2) In other words, for some optimal, the maximum loss that Max could cause is bounded by v and Mindy s loss is at least v, regardless of the particular strategies they choose. If we had knowledge of M, we might be able to find by employing techniques from linear programming. However, we don t necessarily have this knowledge, and even if we did, M could be massively large. Further, applies here only for opponents that are perfectly adversarial, so it doesn t account for an opponent that might make mistakes. hus, it makes sense to try to learn M and iteratively. We do this with the following formulation: for t =,..., Mindy chooses t Max chooses t (with knowledge of t ) Mindy observes M(i, t )8i Loss = M( t, t ) end Clearly, the total loss of this algorithm is simply M( t, t ). We want to be able to compare this loss to the best possible loss that could have been achieved by fixing any single strategy for all iterations. In other words, we want to show: M( t, t ) apple min M(, t) + [Small Regret erm] 2. Multiplicative Updates Suppose we use a multiplicative weight algorithm that updates weights in the following way, where n is the number of rows in matrix M: 2
3 2 [0, ) (3) (i) = 8i (4) n t+ (i) = t (i) M(i,t) Normalizing Constant (5) Our algorithm is similar to the weighted majority algorithm. he idea is decrease the probability of choosing a particular row proportionally to the loss su ered by selecting that row. After making an argument using potentials, we could use this algorithm to obtain the following bound: where = ln( ) and c =. 2.2 Corollary M( t, t ) apple min M(, t )+c ln(n) (6) We can choose such that: M(, ) apple min M(, t )+ (7) q ln(n) where = O( ), which goes to zero for large. In other words, the loss su ered by Mindy per round approaches the optimal average loss per round. We ll use this result to prove the Minimax theorem. 2.3 roof Suppose that Mindy uses the above algorithm to choose t, and that Max chooses t such that t = arg max M(,), maximizing Mindy s loss. Also, let: = = t (8) t (9) We also know intuitively, as mentioned before, that max min M(, ) apple min max M(, ), because as stated earlier, the player that goes second has more information available to her. 3
4 o show equality, which would prove the Minimax theorem stated earlier, it s enough to show that max min M(, ) min max M(, ) also. min max M apple max M By definition of : = max t M By convexity: apple max t M By definition of t : = t M t By corollary 2.2: apple min M t + he proof is finished because that: By definition of : =min M + apple max min M+ goes to zero as gets large. his proof also shows max M apple v + where v = max min M. If we take the average of the t terms computed at each round of the algorithm, we get something within of the optimal value. Because goes to zero for large values of, we can get closer to the optimal strategy by simply increasing. In other words, this strategy becomes more and more optimal as the number of rounds increases. For this reason, is called an approximate min max strategy. A similar argument could be made to show that is an approximate max min strategy. 3 Relation to Online Learning In order to project our analysis into an online learning framework, consider the following problem setting: 4
5 for t =,..., Observe x t from X redict ŷ t 2{0, } Observe true label c(x t ) end Here we consider each hypothesis h as being an expert from the set of all hypotheses H. We want to show that: number of mistakes apple number of mistakes of best h + [Small Regret erm] We set up a game matrix M where M(i, j) =M(h, x) =ifh(x) 6= c(x) and 0 otherwise. hus, the size of this matrix is H X. Given an x t, the algorithm must choose some t, a distribution used to predict x t s label. h is chosen according to the distribution t, and then ŷ t is chosen as h(x t ). t in this context is the distribution concentrated on x t (is at x t and 0 at all other x 2 X). Consequently: M( t,x t ) = E[number of mistakes] apple min M(h, x t ) + [Small Regret erm] h Notice that min h M(h, x t) is equal to the number of mistakes made by the best hypothesis h. IfwesubstituteM( t,x t )with h t(h) {h(x t ) 6= c(x t )} = r[h(x) 6= c(x)] above, we obtain the same bound as was found in the previous section. 4 Relation to Boosting We could think of boosting as a game between the boosting algorithm and the weak learner it calls. Consider the following problem: for t =,..., he boosting algorithm selects a distribution D t over the training set samples X he weak learner chooses a hypothesis h t end Here we assume that all weak learners h t obey the weak learning assumption, i.e. that r (x,y) Dt [h t (x) 6= y)] apple 2 and >0. We could define the game matrix M 0 in terms of the matrix M used in the last section. However, here we want a distribution over the X samples rather than over the hypotheses, so we need to transpose and normalize M. M 0 = M 5
6 In other words, M 0 (i, j) =M 0 (x, h) =ifh(x) =c(x) and 0 otherwise. Here, t = D t, and t is a distribution fully concentrated on the particular h t chosen by the weak learner. We could apply our same analysis from the multiplicative weights algorithm: M 0 ( t,h t ) apple min x M 0 (x, h t )+ Also, M 0 ( t,h t )= x (x) {h(x) =c(x)} = r[h t (x) =c(x)] 2 + Combining these facts: 2 + apple M 0 (x, h t ) apple min x M 0 (x, h t )+ 8x : Rearranging, X M 0 (x, h t ) 2 + > 2 which is again true because approaches 0 as gets large. In other words, we have found that over 2 of the weak hypotheses correctly classify any x when gets su ciently large. Because the final hypothesis is just a majority vote of these weak learners, we have proven that the boosting algorithm drives training error to zero when enough weak learners are employed. 6
TTIC An Introduction to the Theory of Machine Learning. Learning and Game Theory. Avrim Blum 5/7/18, 5/9/18
TTIC 31250 An Introduction to the Theory of Machine Learning Learning and Game Theory Avrim Blum 5/7/18, 5/9/18 Zero-sum games, Minimax Optimality & Minimax Thm; Connection to Boosting & Regret Minimization
More information15.053/8 February 28, person 0-sum (or constant sum) game theory
15.053/8 February 28, 2013 2-person 0-sum (or constant sum) game theory 1 Quotes of the Day My work is a game, a very serious game. -- M. C. Escher (1898-1972) Conceal a flaw, and the world will imagine
More informationMath 167: Mathematical Game Theory Instructor: Alpár R. Mészáros
Math 167: Mathematical Game Theory Instructor: Alpár R. Mészáros Midterm #1, February 3, 2017 Name (use a pen): Student ID (use a pen): Signature (use a pen): Rules: Duration of the exam: 50 minutes. By
More informationRegret Minimization and Correlated Equilibria
Algorithmic Game heory Summer 2017, Week 4 EH Zürich Overview Regret Minimization and Correlated Equilibria Paolo Penna We have seen different type of equilibria and also considered the corresponding price
More information6.896 Topics in Algorithmic Game Theory February 10, Lecture 3
6.896 Topics in Algorithmic Game Theory February 0, 200 Lecture 3 Lecturer: Constantinos Daskalakis Scribe: Pablo Azar, Anthony Kim In the previous lecture we saw that there always exists a Nash equilibrium
More informationTUFTS UNIVERSITY DEPARTMENT OF CIVIL AND ENVIRONMENTAL ENGINEERING ES 152 ENGINEERING SYSTEMS Spring Lesson 16 Introduction to Game Theory
TUFTS UNIVERSITY DEPARTMENT OF CIVIL AND ENVIRONMENTAL ENGINEERING ES 52 ENGINEERING SYSTEMS Spring 20 Introduction: Lesson 6 Introduction to Game Theory We will look at the basic ideas of game theory.
More informationThursday, March 3
5.53 Thursday, March 3 -person -sum (or constant sum) game theory -dimensional multi-dimensional Comments on first midterm: practice test will be on line coverage: every lecture prior to game theory quiz
More informationBest counterstrategy for C
Best counterstrategy for C In the previous lecture we saw that if R plays a particular mixed strategy and shows no intention of changing it, the expected payoff for R (and hence C) varies as C varies her
More informationPAULI MURTO, ANDREY ZHUKOV. If any mistakes or typos are spotted, kindly communicate them to
GAME THEORY PROBLEM SET 1 WINTER 2018 PAULI MURTO, ANDREY ZHUKOV Introduction If any mistakes or typos are spotted, kindly communicate them to andrey.zhukov@aalto.fi. Materials from Osborne and Rubinstein
More informationIEOR E4004: Introduction to OR: Deterministic Models
IEOR E4004: Introduction to OR: Deterministic Models 1 Dynamic Programming Following is a summary of the problems we discussed in class. (We do not include the discussion on the container problem or the
More informationNear-Optimal No-Regret Algorithms for Zero-Sum Games
Near-Optimal No-Regret Algorithms for Zero-Sum Games Constantinos Daskalakis, Alan Deckelbaum 2, Anthony Kim 3 Abstract We propose a new no-regret learning algorithm. When used ( ) against an adversary,
More informationUsing the Maximin Principle
Using the Maximin Principle Under the maximin principle, it is easy to see that Rose should choose a, making her worst-case payoff 0. Colin s similar rationality as a player induces him to play (under
More informationLecture 5. 1 Online Learning. 1.1 Learning Setup (Perspective of Universe) CSCI699: Topics in Learning & Game Theory
CSCI699: Topics in Learning & Game Theory Lecturer: Shaddin Dughmi Lecture 5 Scribes: Umang Gupta & Anastasia Voloshinov In this lecture, we will give a brief introduction to online learning and then go
More informationYao s Minimax Principle
Complexity of algorithms The complexity of an algorithm is usually measured with respect to the size of the input, where size may for example refer to the length of a binary word describing the input,
More informationMATH 210, PROBLEM SET 1 DUE IN LECTURE ON WEDNESDAY, JAN. 28
MATH 210, PROBLEM SET 1 DUE IN LECTURE ON WEDNESDAY, JAN. 28 1. Frankfurt s theory of lying and bullshit. Read Frankfurt s book On Bullshit. In particular, see the description of the distinction he makes
More informationMAT 4250: Lecture 1 Eric Chung
1 MAT 4250: Lecture 1 Eric Chung 2Chapter 1: Impartial Combinatorial Games 3 Combinatorial games Combinatorial games are two-person games with perfect information and no chance moves, and with a win-or-lose
More informationCS711 Game Theory and Mechanism Design
CS711 Game Theory and Mechanism Design Problem Set 1 August 13, 2018 Que 1. [Easy] William and Henry are participants in a televised game show, seated in separate booths with no possibility of communicating
More informationChapter 10: Mixed strategies Nash equilibria, reaction curves and the equality of payoffs theorem
Chapter 10: Mixed strategies Nash equilibria reaction curves and the equality of payoffs theorem Nash equilibrium: The concept of Nash equilibrium can be extended in a natural manner to the mixed strategies
More informationORF 307: Lecture 12. Linear Programming: Chapter 11: Game Theory
ORF 307: Lecture 12 Linear Programming: Chapter 11: Game Theory Robert J. Vanderbei April 3, 2018 Slides last edited on April 3, 2018 http://www.princeton.edu/ rvdb Game Theory John Nash = A Beautiful
More informationGame theory for. Leonardo Badia.
Game theory for information engineering Leonardo Badia leonardo.badia@gmail.com Zero-sum games A special class of games, easier to solve Zero-sum We speak of zero-sum game if u i (s) = -u -i (s). player
More informationIterated Dominance and Nash Equilibrium
Chapter 11 Iterated Dominance and Nash Equilibrium In the previous chapter we examined simultaneous move games in which each player had a dominant strategy; the Prisoner s Dilemma game was one example.
More informationLecture l(x) 1. (1) x X
Lecture 14 Agenda for the lecture Kraft s inequality Shannon codes The relation H(X) L u (X) = L p (X) H(X) + 1 14.1 Kraft s inequality While the definition of prefix-free codes is intuitively clear, we
More informationStrategies and Nash Equilibrium. A Whirlwind Tour of Game Theory
Strategies and Nash Equilibrium A Whirlwind Tour of Game Theory (Mostly from Fudenberg & Tirole) Players choose actions, receive rewards based on their own actions and those of the other players. Example,
More informationGame Theory: Minimax, Maximin, and Iterated Removal Naima Hammoud
Game Theory: Minimax, Maximin, and Iterated Removal Naima Hammoud March 14, 17 Last Lecture: expected value principle Colin A B Rose A - - B - Suppose that Rose knows Colin will play ½ A + ½ B Rose s Expectations
More informationOutline for today. Stat155 Game Theory Lecture 13: General-Sum Games. General-sum games. General-sum games. Dominated pure strategies
Outline for today Stat155 Game Theory Lecture 13: General-Sum Games Peter Bartlett October 11, 2016 Two-player general-sum games Definitions: payoff matrices, dominant strategies, safety strategies, Nash
More informationRegret Minimization and Security Strategies
Chapter 5 Regret Minimization and Security Strategies Until now we implicitly adopted a view that a Nash equilibrium is a desirable outcome of a strategic game. In this chapter we consider two alternative
More informationGame theory and applications: Lecture 1
Game theory and applications: Lecture 1 Adam Szeidl September 20, 2018 Outline for today 1 Some applications of game theory 2 Games in strategic form 3 Dominance 4 Nash equilibrium 1 / 8 1. Some applications
More informationCOS 445 Final. Due online Monday, May 21st at 11:59 pm. Please upload each problem as a separate file via MTA.
COS 445 Final Due online Monday, May 21st at 11:59 pm All problems on this final are no collaboration problems. You may not discuss any aspect of any problems with anyone except for the course staff. You
More informationStrategy Lines and Optimal Mixed Strategy for R
Strategy Lines and Optimal Mixed Strategy for R Best counterstrategy for C for given mixed strategy by R In the previous lecture we saw that if R plays a particular mixed strategy, [p, p, and shows no
More informationEcon 172A, W2002: Final Examination, Solutions
Econ 172A, W2002: Final Examination, Solutions Comments. Naturally, the answers to the first question were perfect. I was impressed. On the second question, people did well on the first part, but had trouble
More informationBest-Reply Sets. Jonathan Weinstein Washington University in St. Louis. This version: May 2015
Best-Reply Sets Jonathan Weinstein Washington University in St. Louis This version: May 2015 Introduction The best-reply correspondence of a game the mapping from beliefs over one s opponents actions to
More informationPAULI MURTO, ANDREY ZHUKOV
GAME THEORY SOLUTION SET 1 WINTER 018 PAULI MURTO, ANDREY ZHUKOV Introduction For suggested solution to problem 4, last year s suggested solutions by Tsz-Ning Wong were used who I think used suggested
More informationMaximizing Winnings on Final Jeopardy!
Maximizing Winnings on Final Jeopardy! Jessica Abramson, Natalie Collina, and William Gasarch August 2017 1 Abstract Alice and Betty are going into the final round of Jeopardy. Alice knows how much money
More informationLecture 7: Bayesian approach to MAB - Gittins index
Advanced Topics in Machine Learning and Algorithmic Game Theory Lecture 7: Bayesian approach to MAB - Gittins index Lecturer: Yishay Mansour Scribe: Mariano Schain 7.1 Introduction In the Bayesian approach
More informationDefinition 4.1. In a stochastic process T is called a stopping time if you can tell when it happens.
102 OPTIMAL STOPPING TIME 4. Optimal Stopping Time 4.1. Definitions. On the first day I explained the basic problem using one example in the book. On the second day I explained how the solution to the
More informationLINEAR PROGRAMMING. Homework 7
LINEAR PROGRAMMING Homework 7 Fall 2014 Csci 628 Megan Rose Bryant 1. Your friend is taking a Linear Programming course at another university and for homework she is asked to solve the following LP: Primal:
More informationNear-Optimal No-Regret Algorithms for Zero-Sum Games
Near-Optimal No-Regret Algorithms for Zero-Sum Games Constantinos Daskalakis Alan Deckelbaum Anthony Kim Abstract We propose a new no-regret learning algorithm. When used against an adversary, ( ) our
More informationMATH 121 GAME THEORY REVIEW
MATH 121 GAME THEORY REVIEW ERIN PEARSE Contents 1. Definitions 2 1.1. Non-cooperative Games 2 1.2. Cooperative 2-person Games 4 1.3. Cooperative n-person Games (in coalitional form) 6 2. Theorems and
More informationSubject : Computer Science. Paper: Machine Learning. Module: Decision Theory and Bayesian Decision Theory. Module No: CS/ML/10.
e-pg Pathshala Subject : Computer Science Paper: Machine Learning Module: Decision Theory and Bayesian Decision Theory Module No: CS/ML/0 Quadrant I e-text Welcome to the e-pg Pathshala Lecture Series
More informationMixed Strategies. Samuel Alizon and Daniel Cownden February 4, 2009
Mixed Strategies Samuel Alizon and Daniel Cownden February 4, 009 1 What are Mixed Strategies In the previous sections we have looked at games where players face uncertainty, and concluded that they choose
More information6.254 : Game Theory with Engineering Applications Lecture 3: Strategic Form Games - Solution Concepts
6.254 : Game Theory with Engineering Applications Lecture 3: Strategic Form Games - Solution Concepts Asu Ozdaglar MIT February 9, 2010 1 Introduction Outline Review Examples of Pure Strategy Nash Equilibria
More informationLecture 6. 1 Polynomial-time algorithms for the global min-cut problem
ORIE 633 Network Flows September 20, 2007 Lecturer: David P. Williamson Lecture 6 Scribe: Animashree Anandkumar 1 Polynomial-time algorithms for the global min-cut problem 1.1 The global min-cut problem
More informationComparative Study between Linear and Graphical Methods in Solving Optimization Problems
Comparative Study between Linear and Graphical Methods in Solving Optimization Problems Mona M Abd El-Kareem Abstract The main target of this paper is to establish a comparative study between the performance
More informationm 11 m 12 Non-Zero Sum Games Matrix Form of Zero-Sum Games R&N Section 17.6
Non-Zero Sum Games R&N Section 17.6 Matrix Form of Zero-Sum Games m 11 m 12 m 21 m 22 m ij = Player A s payoff if Player A follows pure strategy i and Player B follows pure strategy j 1 Results so far
More informationGame Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India October 2012
Game Theory Lecture Notes By Y. Narahari Department of Computer Science and Automation Indian Institute of Science Bangalore, India October 22 COOPERATIVE GAME THEORY Correlated Strategies and Correlated
More informationMaximum Contiguous Subsequences
Chapter 8 Maximum Contiguous Subsequences In this chapter, we consider a well-know problem and apply the algorithm-design techniques that we have learned thus far to this problem. While applying these
More informationOutline for today. Stat155 Game Theory Lecture 19: Price of anarchy. Cooperative games. Price of anarchy. Price of anarchy
Outline for today Stat155 Game Theory Lecture 19:.. Peter Bartlett Recall: Linear and affine latencies Classes of latencies Pigou networks Transferable versus nontransferable utility November 1, 2016 1
More informationIntroduction to game theory LECTURE 2
Introduction to game theory LECTURE 2 Jörgen Weibull February 4, 2010 Two topics today: 1. Existence of Nash equilibria (Lecture notes Chapter 10 and Appendix A) 2. Relations between equilibrium and rationality
More informationTheir opponent will play intelligently and wishes to maximize their own payoff.
Two Person Games (Strictly Determined Games) We have already considered how probability and expected value can be used as decision making tools for choosing a strategy. We include two examples below for
More informationOutline. 1 Introduction. 2 Algorithms. 3 Examples. Algorithm 1 General coordinate minimization framework. 1: Choose x 0 R n and set k 0.
Outline Coordinate Minimization Daniel P. Robinson Department of Applied Mathematics and Statistics Johns Hopkins University November 27, 208 Introduction 2 Algorithms Cyclic order with exact minimization
More information6.207/14.15: Networks Lecture 10: Introduction to Game Theory 2
6.207/14.15: Networks Lecture 10: Introduction to Game Theory 2 Daron Acemoglu and Asu Ozdaglar MIT October 14, 2009 1 Introduction Outline Review Examples of Pure Strategy Nash Equilibria Mixed Strategies
More informationMonte-Carlo Planning: Introduction and Bandit Basics. Alan Fern
Monte-Carlo Planning: Introduction and Bandit Basics Alan Fern 1 Large Worlds We have considered basic model-based planning algorithms Model-based planning: assumes MDP model is available Methods we learned
More informationMaximizing Winnings on Final Jeopardy!
Maximizing Winnings on Final Jeopardy! Jessica Abramson, Natalie Collina, and William Gasarch August 2017 1 Introduction Consider a final round of Jeopardy! with players Alice and Betty 1. We assume that
More informationThe Loans_processed.csv file is the dataset we obtained after the pre-processing part where the clean-up python code was used.
Machine Learning Group Homework 3 MSc Business Analytics Team 9 Alexander Romanenko, Artemis Tomadaki, Justin Leiendecker, Zijun Wei, Reza Brianca Widodo The Loans_processed.csv file is the dataset we
More informationGame Theory I. Author: Neil Bendle Marketing Metrics Reference: Chapter Neil Bendle and Management by the Numbers, Inc.
Game Theory I This module provides an introduction to game theory for managers and includes the following topics: matrix basics, zero and non-zero sum games, and dominant strategies. Author: Neil Bendle
More informationComplexity of Iterated Dominance and a New Definition of Eliminability
Complexity of Iterated Dominance and a New Definition of Eliminability Vincent Conitzer and Tuomas Sandholm Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA 15213 {conitzer, sandholm}@cs.cmu.edu
More informationGame Theory Notes: Examples of Games with Dominant Strategy Equilibrium or Nash Equilibrium
Game Theory Notes: Examples of Games with Dominant Strategy Equilibrium or Nash Equilibrium Below are two different games. The first game has a dominant strategy equilibrium. The second game has two Nash
More informationLecture 10: The knapsack problem
Optimization Methods in Finance (EPFL, Fall 2010) Lecture 10: The knapsack problem 24.11.2010 Lecturer: Prof. Friedrich Eisenbrand Scribe: Anu Harjula The knapsack problem The Knapsack problem is a problem
More information1. better to stick. 2. better to switch. 3. or does your second choice make no difference?
The Monty Hall game Game show host Monty Hall asks you to choose one of three doors. Behind one of the doors is a new Porsche. Behind the other two doors there are goats. Monty knows what is behind each
More informationBest Reply Behavior. Michael Peters. December 27, 2013
Best Reply Behavior Michael Peters December 27, 2013 1 Introduction So far, we have concentrated on individual optimization. This unified way of thinking about individual behavior makes it possible to
More informationMA300.2 Game Theory 2005, LSE
MA300.2 Game Theory 2005, LSE Answers to Problem Set 2 [1] (a) This is standard (we have even done it in class). The one-shot Cournot outputs can be computed to be A/3, while the payoff to each firm can
More informationWeek 8: Basic concepts in game theory
Week 8: Basic concepts in game theory Part 1: Examples of games We introduce here the basic objects involved in game theory. To specify a game ones gives The players. The set of all possible strategies
More informationQ1. [?? pts] Search Traces
CS 188 Spring 2010 Introduction to Artificial Intelligence Midterm Exam Solutions Q1. [?? pts] Search Traces Each of the trees (G1 through G5) was generated by searching the graph (below, left) with a
More information56:171 Operations Research Midterm Examination Solutions PART ONE
56:171 Operations Research Midterm Examination Solutions Fall 1997 Write your name on the first page, and initial the other pages. Answer both questions of Part One, and 4 (out of 5) problems from Part
More informationTTIC An Introduction to the Theory of Machine Learning. The Adversarial Multi-armed Bandit Problem Avrim Blum.
TTIC 31250 An Introduction to the Theory of Machine Learning The Adversarial Multi-armed Bandit Problem Avrim Blum Start with recap 1 Algorithm Consider the following setting Each morning, you need to
More informationCS 331: Artificial Intelligence Game Theory I. Prisoner s Dilemma
CS 331: Artificial Intelligence Game Theory I 1 Prisoner s Dilemma You and your partner have both been caught red handed near the scene of a burglary. Both of you have been brought to the police station,
More informationRationalizable Strategies
Rationalizable Strategies Carlos Hurtado Department of Economics University of Illinois at Urbana-Champaign hrtdmrt2@illinois.edu Jun 1st, 2015 C. Hurtado (UIUC - Economics) Game Theory On the Agenda 1
More informationMath 135: Answers to Practice Problems
Math 35: Answers to Practice Problems Answers to problems from the textbook: Many of the problems from the textbook have answers in the back of the book. Here are the answers to the problems that don t
More informationTug of War Game. William Gasarch and Nick Sovich and Paul Zimand. October 6, Abstract
Tug of War Game William Gasarch and ick Sovich and Paul Zimand October 6, 2009 To be written later Abstract Introduction Combinatorial games under auction play, introduced by Lazarus, Loeb, Propp, Stromquist,
More information1 Shapley-Shubik Model
1 Shapley-Shubik Model There is a set of buyers B and a set of sellers S each selling one unit of a good (could be divisible or not). Let v ij 0 be the monetary value that buyer j B assigns to seller i
More informationMonte-Carlo Planning: Introduction and Bandit Basics. Alan Fern
Monte-Carlo Planning: Introduction and Bandit Basics Alan Fern 1 Large Worlds We have considered basic model-based planning algorithms Model-based planning: assumes MDP model is available Methods we learned
More informationRecap First-Price Revenue Equivalence Optimal Auctions. Auction Theory II. Lecture 19. Auction Theory II Lecture 19, Slide 1
Auction Theory II Lecture 19 Auction Theory II Lecture 19, Slide 1 Lecture Overview 1 Recap 2 First-Price Auctions 3 Revenue Equivalence 4 Optimal Auctions Auction Theory II Lecture 19, Slide 2 Motivation
More informationCS364A: Algorithmic Game Theory Lecture #3: Myerson s Lemma
CS364A: Algorithmic Game Theory Lecture #3: Myerson s Lemma Tim Roughgarden September 3, 23 The Story So Far Last time, we introduced the Vickrey auction and proved that it enjoys three desirable and different
More informationCS360 Homework 14 Solution
CS360 Homework 14 Solution Markov Decision Processes 1) Invent a simple Markov decision process (MDP) with the following properties: a) it has a goal state, b) its immediate action costs are all positive,
More informationCS221 / Spring 2018 / Sadigh. Lecture 9: Games I
CS221 / Spring 2018 / Sadigh Lecture 9: Games I Course plan Search problems Markov decision processes Adversarial games Constraint satisfaction problems Bayesian networks Reflex States Variables Logic
More informationProblem Set 2 - SOLUTIONS
Problem Set - SOLUTONS 1. Consider the following two-player game: L R T 4, 4 1, 1 B, 3, 3 (a) What is the maxmin strategy profile? What is the value of this game? Note, the question could be solved like
More informationLecture Notes 1
4.45 Lecture Notes Guido Lorenzoni Fall 2009 A portfolio problem To set the stage, consider a simple nite horizon problem. A risk averse agent can invest in two assets: riskless asset (bond) pays gross
More informationOutline Introduction Game Representations Reductions Solution Concepts. Game Theory. Enrico Franchi. May 19, 2010
May 19, 2010 1 Introduction Scope of Agent preferences Utility Functions 2 Game Representations Example: Game-1 Extended Form Strategic Form Equivalences 3 Reductions Best Response Domination 4 Solution
More information343H: Honors AI. Lecture 7: Expectimax Search 2/6/2014. Kristen Grauman UT-Austin. Slides courtesy of Dan Klein, UC-Berkeley Unless otherwise noted
343H: Honors AI Lecture 7: Expectimax Search 2/6/2014 Kristen Grauman UT-Austin Slides courtesy of Dan Klein, UC-Berkeley Unless otherwise noted 1 Announcements PS1 is out, due in 2 weeks Last time Adversarial
More information1 Overview. 2 The Gradient Descent Algorithm. AM 221: Advanced Optimization Spring 2016
AM 22: Advanced Optimization Spring 206 Prof. Yaron Singer Lecture 9 February 24th Overview In the previous lecture we reviewed results from multivariate calculus in preparation for our journey into convex
More informationECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2017
ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2017 These notes have been used and commented on before. If you can still spot any errors or have any suggestions for improvement, please
More informationECS171: Machine Learning
ECS171: Machine Learning Lecture 15: Tree-based Algorithms Cho-Jui Hsieh UC Davis March 7, 2018 Outline Decision Tree Random Forest Gradient Boosted Decision Tree (GBDT) Decision Tree Each node checks
More informationEfficiency and Herd Behavior in a Signalling Market. Jeffrey Gao
Efficiency and Herd Behavior in a Signalling Market Jeffrey Gao ABSTRACT This paper extends a model of herd behavior developed by Bikhchandani and Sharma (000) to establish conditions for varying levels
More informationGame Theory - Lecture #8
Game Theory - Lecture #8 Outline: Randomized actions vnm & Bernoulli payoff functions Mixed strategies & Nash equilibrium Hawk/Dove & Mixed strategies Random models Goal: Would like a formulation in which
More informationECE 586GT: Problem Set 1: Problems and Solutions Analysis of static games
University of Illinois Fall 2018 ECE 586GT: Problem Set 1: Problems and Solutions Analysis of static games Due: Tuesday, Sept. 11, at beginning of class Reading: Course notes, Sections 1.1-1.4 1. [A random
More informationLecture 8: Introduction to asset pricing
THE UNIVERSITY OF SOUTHAMPTON Paul Klein Office: Murray Building, 3005 Email: p.klein@soton.ac.uk URL: http://paulklein.se Economics 3010 Topics in Macroeconomics 3 Autumn 2010 Lecture 8: Introduction
More informationVersion 01 as of May 12, 2014 Author: Steven Sagona To be submitted to PRL (not really)
Version 01 as of May 12, 2014 Author: Steven Sagona To be submitted to PRL (not really) The Game Theory of Pokemon: AI implementing Nash Equilibrium (Dated: May 12, 2014) This paper is written to propose
More informationEpistemic Game Theory
Epistemic Game Theory Lecture 1 ESSLLI 12, Opole Eric Pacuit Olivier Roy TiLPS, Tilburg University MCMP, LMU Munich ai.stanford.edu/~epacuit http://olivier.amonbofis.net August 6, 2012 Eric Pacuit and
More information2. This algorithm does not solve the problem of finding a maximum cardinality set of non-overlapping intervals. Consider the following intervals:
1. No solution. 2. This algorithm does not solve the problem of finding a maximum cardinality set of non-overlapping intervals. Consider the following intervals: E A B C D Obviously, the optimal solution
More informationGAME THEORY. Game theory. The odds and evens game. Two person, zero sum game. Prototype example
Game theory GAME THEORY (Hillier & Lieberman Introduction to Operations Research, 8 th edition) Mathematical theory that deals, in an formal, abstract way, with the general features of competitive situations
More informationChapter 2 Strategic Dominance
Chapter 2 Strategic Dominance 2.1 Prisoner s Dilemma Let us start with perhaps the most famous example in Game Theory, the Prisoner s Dilemma. 1 This is a two-player normal-form (simultaneous move) game.
More informationMicroeconomics II. CIDE, MsC Economics. List of Problems
Microeconomics II CIDE, MsC Economics List of Problems 1. There are three people, Amy (A), Bart (B) and Chris (C): A and B have hats. These three people are arranged in a room so that B can see everything
More informationJanuary 26,
January 26, 2015 Exercise 9 7.c.1, 7.d.1, 7.d.2, 8.b.1, 8.b.2, 8.b.3, 8.b.4,8.b.5, 8.d.1, 8.d.2 Example 10 There are two divisions of a firm (1 and 2) that would benefit from a research project conducted
More informationApplications of Linear Programming
Applications of Linear Programming lecturer: András London University of Szeged Institute of Informatics Department of Computational Optimization Lecture 8 The portfolio selection problem The portfolio
More informationLecture 5: Iterative Combinatorial Auctions
COMS 6998-3: Algorithmic Game Theory October 6, 2008 Lecture 5: Iterative Combinatorial Auctions Lecturer: Sébastien Lahaie Scribe: Sébastien Lahaie In this lecture we examine a procedure that generalizes
More informationGame Theory: introduction and applications to computer networks
Game Theory: introduction and applications to computer networks Zero-Sum Games (follow-up) Giovanni Neglia INRIA EPI Maestro 20 January 2014 Part of the slides are based on a previous course with D. Figueiredo
More informationHomework 2: Solutions Sid Banerjee Problem 1: Practice with Dynamic Programming Formulation
Problem 1: Practice with Dynamic Programming Formulation A product manager has to order stock daily. Each unit cost is c, there is a fixed cost of K for placing an order. If you order on day t, the items
More informationElements of Economic Analysis II Lecture XI: Oligopoly: Cournot and Bertrand Competition
Elements of Economic Analysis II Lecture XI: Oligopoly: Cournot and Bertrand Competition Kai Hao Yang /2/207 In this lecture, we will apply the concepts in game theory to study oligopoly. In short, unlike
More informationLecture 9: Games I. Course plan. A simple game. Roadmap. Machine learning. Example: game 1
Lecture 9: Games I Course plan Search problems Markov decision processes Adversarial games Constraint satisfaction problems Bayesian networks Reflex States Variables Logic Low-level intelligence Machine
More informationLecture 5 January 30
EE 223: Stochastic Estimation and Control Spring 2007 Lecture 5 January 30 Lecturer: Venkat Anantharam Scribe: aryam Kamgarpour 5.1 Secretary Problem The problem set-up is explained in Lecture 4. We review
More information