The Game-Theoretic Framework for Probability

Similar documents
Defensive Forecasting

Game-Theoretic Probability and Defensive Forecasting

Comparison of proof techniques in game-theoretic probability and measure-theoretic probability

Laws of probabilities in efficient markets

Probability without Measure!

Introduction to Game-Theoretic Probability

Probability, Price, and the Central Limit Theorem. Glenn Shafer. Rutgers Business School February 18, 2002

An introduction to game-theoretic probability from statistical viewpoint

Outline of Lecture 1. Martin-Löf tests and martingales

Time Resolution of the St. Petersburg Paradox: A Rebuttal

X i = 124 MARTINGALES

On Existence of Equilibria. Bayesian Allocation-Mechanisms

Remarks on Probability

The Accrual Anomaly in the Game-Theoretic Setting

arxiv: v1 [cs.lg] 21 May 2011

6. Martingales. = Zn. Think of Z n+1 as being a gambler s earnings after n+1 games. If the game if fair, then E [ Z n+1 Z n

The Subjective and Personalistic Interpretations

MA300.2 Game Theory 2005, LSE

N(A) P (A) = lim. N(A) =N, we have P (A) = 1.

The Kelly Criterion. How To Manage Your Money When You Have an Edge

Martingales. by D. Cox December 2, 2009

Rational Behaviour and Strategy Construction in Infinite Multiplayer Games

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models

Black-Scholes and Game Theory. Tushar Vaidya ESD

Best-Reply Sets. Jonathan Weinstein Washington University in St. Louis. This version: May 2015

How quantitative methods influence and shape finance industry

A GENERALIZED MARTINGALE BETTING STRATEGY

What is accomplished by successful non stationary stochastic prediction?

Stochastic Games and Bayesian Games

Central Limit Theorem 11/08/2005

Game Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India October 2012

Lecture 5. 1 Online Learning. 1.1 Learning Setup (Perspective of Universe) CSCI699: Topics in Learning & Game Theory

The Capital Asset Pricing Model as a corollary of the Black Scholes model

A new formulation of asset trading games in continuous time with essential forcing of variation exponent

Finite Additivity in Dubins-Savage Gambling and Stochastic Games. Bill Sudderth University of Minnesota

Lecture Notes for Chapter 6. 1 Prototype model: a one-step binomial tree

GEK1544 The Mathematics of Games Suggested Solutions to Tutorial 3

Martingale Measure TA

Math-Stat-491-Fall2014-Notes-V

Variation Spectrum Suppose ffl S(t) is a continuous function on [0;T], ffl N is a large integer. For n = 1;:::;N, set For p > 0, set vars;n(p) := S n

Internet Trading Mechanisms and Rational Expectations

9 Expectation and Variance

Problem 3 Solutions. l 3 r, 1

Lecture 9 Game theoretic finance

Mixed Strategies. In the previous chapters we restricted players to using pure strategies and we

MATHEMATICAL ENGINEERING TECHNICAL REPORTS. A new formulation of asset trading games in continuous time with essential forcing of variation exponent

CONSISTENCY AMONG TRADING DESKS

Game Theory: Normal Form Games

Finite Memory and Imperfect Monitoring

18.440: Lecture 32 Strong law of large numbers and Jensen s inequality

Building Infinite Processes from Regular Conditional Probability Distributions

ECON FINANCIAL ECONOMICS

Efficiency in Decentralized Markets with Aggregate Uncertainty

Lecture 7: Bayesian approach to MAB - Gittins index

INTERIM CORRELATED RATIONALIZABILITY IN INFINITE GAMES

Game Theory Fall 2006

Tug of War Game. William Gasarch and Nick Sovich and Paul Zimand. October 6, Abstract

February 23, An Application in Industrial Organization

Math 180A. Lecture 5 Wednesday April 7 th. Geometric distribution. The geometric distribution function is

Characterization of the Optimum

3 Arbitrage pricing theory in discrete time.

MA200.2 Game Theory II, LSE

Iterated Dominance and Nash Equilibrium

Chapter 5. Sampling Distributions

Stochastic Games and Bayesian Games

Copyright (C) 2001 David K. Levine This document is an open textbook; you can redistribute it and/or modify it under the terms of version 1 of the

Expected Utility Theory

G5212: Game Theory. Mark Dean. Spring 2017

Computational Independence

What do Coin Tosses and Decision Making under Uncertainty, have in common?

Sampling; Random Walk

Microeconomics II. CIDE, MsC Economics. List of Problems

Game Theory and Economics Prof. Dr. Debarshi Das Department of Humanities and Social Sciences Indian Institute of Technology, Guwahati.

An Introduction to Stochastic Calculus

Math 489/Math 889 Stochastic Processes and Advanced Mathematical Finance Dunbar, Fall 2007

4 Martingales in Discrete-Time

How speculation can explain the equity premium

Economics 431 Infinitely repeated games

The Game-Theoretic Capital Asset Pricing Model

SYLLABUS AND SAMPLE QUESTIONS FOR MSQE (Program Code: MQEK and MQED) Syllabus for PEA (Mathematics), 2013

Statistical Methods in Practice STAT/MATH 3379

Casino gambling problem under probability weighting

Discrete Mathematics for CS Spring 2008 David Wagner Final Exam

Stochastic Processes and Financial Mathematics (part one) Dr Nic Freeman

Sidney I. Resnick. A Probability Path. Birkhauser Boston Basel Berlin

Convergence. Any submartingale or supermartingale (Y, F) converges almost surely if it satisfies E Y n <. STAT2004 Martingale Convergence

Introduction to Probability Theory and Stochastic Processes for Finance Lecture Notes

LECTURE 2: MULTIPERIOD MODELS AND TREES

Comparing Allocations under Asymmetric Information: Coase Theorem Revisited

MATH20180: Foundations of Financial Mathematics

MATH3075/3975 FINANCIAL MATHEMATICS TUTORIAL PROBLEMS

Standard Decision Theory Corrected:

The Yield Envelope: Price Ranges for Fixed Income Products

FE 5204 Stochastic Differential Equations

Follow the Leader I has three pure strategy Nash equilibria of which only one is reasonable.

Why Bankers Should Learn Convex Analysis

Answers to Odd-Numbered Problems, 4th Edition of Games and Information, Rasmusen

GAME THEORY. Department of Economics, MIT, Follow Muhamet s slides. We need the following result for future reference.

Strategy -1- Strategy

Chapter 14. From Randomness to Probability. Copyright 2010 Pearson Education, Inc.

Transcription:

11th IPMU International Conference The Game-Theoretic Framework for Probability Glenn Shafer July 5, 2006 Part I. A new mathematical foundation for probability theory. Game theory replaces measure theory. Part II. Application to statistics: Defensive forecasting. Good probability forecasting is possible. 1

Part I. A new mathematical foundation for probability theory. Game theory replaces measure theory. Mathematics: Classical probability theorems become theorems in game theory (someone has a winning strategy). Philosophy: Cournot s principle (an event of small probability does not happen) becomes game-theoretic (you do not get rich without risking bankruptcy). 2

Part II. Application to statistics: Defensive forecasting. Good probability forecasting is possible. We call it defensive forecasting because it defends against a portmanteau (quasi-universal) test. Your probability forecasts will pass this portmanteau test even if reality plays against you. Defensive forecasting is a radically new method, not encountered in classical or measure-theoretic probability. 3

Part I. Basics of Game-Theoretic Probability 1. Pascal & Ville. Pascal assumed no arbitrage (you cannot make money for sure) in a sequential game. Ville added Cournot s principle (you will not get rich without risking bankruptcy). 2. The strong law of large numbers 3. The weak law of large numbers 4

Pascal: Fair division Peter and Paul play for $100. Paul is behind. Paul needs 2 points to win, and Peter needs only 1. Peter $0 $? Paul Peter $0 Paul $100 Blaise Pascal (1623 1662), as imagined in the 19th century by Hippolyte Flandrin. If the game must be broken off, how much of the $100 should Paul get? 5

It is fair for Paul to pay $a in order to get $2a if he defeats Peter and $0 if he loses to Peter. $25 So Paul should get $25. Peter $0 Peter Paul $50 Paul $0 $100 $a $0 $2a Modern formulation: If the game on the left is available, the prices above are forced by the principle of no arbitrage. 6

Binary probability game. (Here K n is Skeptic s capital and s n is the total stakes.) K 0 := 1. FOR n = 1, 2,... : Forecaster announces p n [0, 1]. Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n (y n p n ). No Arbitrage: If Forecaster announces a strategy in advance, the strategy must obey the rules of probability to keep Skeptic from making money for sure. In other words, the p n should be conditional probabilities from some probability distribution for y 1, y 2,.... 7

Blaise Pascal Probability is about fair prices in a sequential game. Pascal s concept of fairness: no arbitrage. Jean Ville A second concept of fairness: you will not get rich without risking bankruptcy. 8

In 1939, Ville showed that the laws of probability can be derived from a principle of market efficiency: If you never bet more than you have, you will not get infinitely rich. Jean Ville, 1910 1988, on entering the École Normale Supérieure. As Ville showed, this is equivalent to the principle that events of small probability will not happen. We call both principles Cournot s principle. 9

Binary probability game when Forecaster uses the strategy given by a probability distribution P. K 0 := 1. FOR n = 1, 2,... : Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n (y n P{Y n = 1 Y 1 = y 1,..., Y n 1 = y n 1 }). Restriction on Skeptic: Skeptic must choose the s n so that K n 0 for all n no matter how Reality moves. 10

Two sides of fairness in game-theoretic probability. Pascal Constraint on Forecaster: Don t let Skeptic make money for sure. (No arbitrage.) Ville Constraint on Skeptic: Do not risk bankruptcy. (Cournot s principle say s he will then not make a lot of money.) 11

Part I. Basics of Game-Theoretic Probability 1. Pascal & Ville 2. The strong law of large numbers (Borel). The classic version says the proportion of heads converges to 1 2 except on a set of measure zero. The game-theoretic version says it converges to 2 1 unless you get infinitely rich. 3. The weak law of large numbers 12

Fair-coin game. (Skeptic announces the amount M n he risks losing rather than the total stakes s n.) K 0 = 1. FOR n = 1, 2,... : Skeptic announces M n R. Reality announces y n { 1, 1}. K n := K n 1 + M n y n. Skeptic wins if (1) K n is never negative and (2) either lim n 1 n ni=1 y i = 0 or lim n K n =. Otherwise Reality wins. Theorem Skeptic has a winning strategy. 13

Who wins? Skeptic wins if (1) K n is never negative and (2) either lim n 1 n n i=1 y i = 0 or lim n K n =. So the theorem says that Skeptic has a strategy that (1) does not risk bankruptcy and (2) guarantees that either the average of the y i converges to 0 or else Skeptic becomes infinitely rich. Loosely: The average of the y i converges to 0 unless Skeptic becomes infinitely rich. 14

The Idea of the Proof Idea 1 Establish an account for betting on heads. On each round, bet ɛ of the account on heads. Then Reality can keep the account from getting indefinitely large only by eventually holding the cumulative proportion of heads at or below 2 1 (1 + ɛ). It does not matter how little money the account starts with. Idea 2 Establish infinitely many accounts. Use the kth account to bet on heads with ɛ = 1/k. This forces the cumulative proportion of heads to stay at 1/2 or below. Idea 3 Set up similar accounts for betting on tails. This forces Reality to make the proportion converge exactly to one-half. 15

Definitions A path is an infinite sequence y 1 y 2... of moves for Reality. An event is a set of paths. A situation is a finite initial sequence of moves for Reality, say y 1 y 2... y n. is the initial situation, a sequence of length zero. When ξ is a path, say ξ = y 1 y 2..., write ξ n for the situation y 1 y 2... y n. 16

Game-theoretic processes and martingales A real-valued function on the situations is a process. A process P can be used as a strategy for Skeptic: Skeptic buys P(y 1... y n 1 ) of y n Skeptic in situation y 1... y n 1. A strategy for Skeptic, together with a particular initial capital for Skeptic, also defines a process: Skeptic s capital process K(y 1... y n ). We also call a capital process for Skeptic a martingale. 17

Notation for Martingales Skeptic begins with capital 1 in our game, but we can change the rules so he begins with α. Write K P for his capital process when he begins with zero and follows strategy P: K P ( ) = 0 and K P (y 1 y 2... y n ) := K P (y 1 y 2... y n 1 ) + P(y 1 y 2... y n 1 )y n. When he starts with α, his capital process is α + K P. The capital processes that begin with zero form a linear space, for βk P = K βp and K P 1 + K P 2 = K P 1+P 2. So the martingales also form a linear space. 18

Convex Combinations of Martingales If P 1 and P 2 are strategies, and α 1 + α 2 = 1, then α 1 (1 + K P 1) + α 2 (1 + K P 2) = 1 + K α 1P 1 +α 2 P 2. LHS is the convex combination of two martingales that each begin with capital 1. RHS is the martingale produced by the same convex combination of strategies, also beginning with capital 1. Conclusion: In the game where we begin with capital 1, we can obtain a convex combination of 1 + K P 1 and 1 + K P 2 by splitting our capital into two accounts, one with initial capital α 1 and one with initial capital α 2. Apply α 1 P 1 to the first account and α 2 P 2 to the second. 19

Infinite Convex Combinations: Suppose P 1, P 2,... are strategies and α 1, α 2,... are nonnegative real numbers adding to one. If k=1 α k P k converges, then k=1 α k K P k also converges. k=1 α k K P k is the capital process from k=1 α k P k. You can prove this by induction on K P (y 1 y 2... y n ) := K P (y 1 y 2... y n 1 ) + P(y 1 y 2... y n 1 )y n. In game-theoretic probability, you can usually get an infinite convex combination of martingales, but you have to check on the convergence of the infinite convex combination of strategies. In a sense, this explains the historical confusion about countable additivity in measure-theoretic probability (see Working Paper #4). 20

The greater power of game-theoretic probability Instead of a probability distribution for y 1, y 2,..., maybe you have only a few prices. Instead of giving them at the outset, maybe your make them up as you go along. Instead of use or Skeptic announces M n R. Reality announces y n { 1, 1}. K n := K n 1 + M n y n. Skeptic announces M n R. Reality announces y n [ 1, 1]. K n := K n 1 + M n y n. Forecaster announces m n R. Skeptic announces M n R. Reality announces y n [m n 1, m n + 1]. K n := K n 1 + M n (y n m n ). 21

Part I. Basics of Game-Theoretic Probability 1. Pascal & Ville 2. The strong law of large numbers. Infinite and impractical: You will not get infinitely rich in an infinite number of trials. 3. The weak law of large numbers. Finite and practical: You will not multiply your capital by a large factor in N trials. 22

The weak law of large numbers (Bernoulli) K 0 := 1. FOR n = 1,..., N: Skeptic announces M n R. Reality announces y n { 1, 1}. K n := K n 1 + M n y n. Winning: Skeptic wins if K n is never negative and either K N C or N n=1 y n /N < ɛ. Theorem. Skeptic has a winning strategy if N C/ɛ 2. 23

Part II. Defensive Forecasting 1. Thesis. Good probability forecasting is possible. 2. Theorem. Forecaster can beat any test. 3. Research agenda. Use proof to translate tests of Forecaster into forecasting strategies. 4. Example. Forecasting using LLN (law of large numbers). 24

THESIS Good probability forecasting is possible. We can always give probabilities with good calibration and resolution. PERFECT INFORMATION PROTOCOL FOR n = 1, 2,... Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. There exists a strategy for Forecaster that gives p n with good calibration and resolution. 25

FOR n = 1, 2,... Reality announces x n X. Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. 1. Fix p [0, 1]. Look at n for which p n p. If the frequency of y n = 1 always approximates p, Forecaster is properly calibrated. 2. Fix x X and p [0, 1]. Look at n for which x n x and p n p. If the frequency of y n = 1 always approximates p, Forecaster is properly calibrated and has good resolution. 26

FOR n = 1, 2,... Reality announces x n X. Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. Forecaster can give ps with good calibration and resolution no matter what Reality does. Philosophical implications: To a good approximation, everything is stochastic. Getting the probabilities right means describing the past well, not having insight into the future. 27

THEOREM. Forecaster can beat any test. FOR n = 1, 2,... Reality announces x n X. Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. Theorem. Given a test, Forecaster has a strategy guaranteed to pass it. Thesis. There is a test of Forecaster universal enough that passing it implies the ps have good calibration and resolution. (Not a theorem, because good calibration and resolution is fuzzy.) 28

The probabilities are tested by another player, Skeptic. FOR n = 1, 2,... Reality announces x n X. Forecaster announces p n [0, 1]. Skeptic announces s n R. Reality announces y n {0, 1}. Skeptic s profit := s n (y n p n ). A test of Forecaster is a strategy for Skeptic that is continuous in the ps. If Skeptic does not make too much money, the ps pass the test. Theorem If Skeptic plays a known continuous strategy, Forecaster has a strategy guaranteeing that Skeptic never makes money. 29

This concept of test generalizes the standard stochastic concept. Stochastic setting: There is a probability distribution P for the xs and ys. Forecaster uses P s conditional probabilities as his ps. Reality chooses her xs and ys from P. Standard concept of statistical test: Choose an event A whose probability under P is small. Reject P if A happens. In 1939, Jean Ville showed that in the stochastic setting, the standard concept is equivalent to a strategy for Skeptic. 30

Why insist on continuity? Why count only strategies for Skeptic that are continuous in the ps as tests of Forecaster? 1. Brouwer s thesis: A computable function of a real argument is continuous. 2. Classical statistical tests (e.g., reject if LLN fails) correspond to continuous strategies. 31

Skeptic adopts a continuous strategy S. FOR n = 1, 2,... Reality announces x n X. Forecaster announces p n [0, 1]. Skeptic makes the move s n specified by S. Reality announces y n {0, 1}. Skeptic s profit := s n (y n p n ). Theorem Forecaster can guarantee that Skeptic never makes money. We actually prove a stronger theorem. Instead of making Skeptic announce his entire strategy in advance, only make him reveal his strategy for each round in advance of Forecaster s move. FOR n = 1, 2,... Reality announces x n X. Skeptic announces continuous S n : [0, 1] R. Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. Skeptic s profit := S n (p n )(y n p n ). Theorem. Forecaster can guarantee that Skeptic never makes money. 32

FOR n = 1, 2,... Reality announces x n X. Skeptic announces continuous S n : [0, 1] R. Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. Skeptic s profit := S n (p n )(y n p n ). Theorem Forecaster can guarantee that Skeptic never makes money. Proof: If S n (p) > 0 for all p, take p n := 1. If S n (p) < 0 for all p, take p n := 0. Otherwise, choose p n so that S n (p n ) = 0. 33

Research agenda. Use proof to translate tests of Forecaster into forecasting strategies. Example 1: Use a strategy for Sceptic that makes money if Reality does not obey the LLN (frequency of y n = 1 overall approximates average of p n ). The derived strategy for Forecaster guarantees the LLN i.e., its probabilities are calibrated in the large. Example 2: Use a strategy for Skeptic that makes money if Reality does not obey the LLN for rounds where p n is close to p. The derived strategy for Forecaster guarantees calibration for p n close to p. Example 3: Average the preceding strategies for Skeptic for a grid of values of p. The derived strategy for Forecaster guarantees good calibration everywhere. Example 4: Average over a grid of values of p and x. Then you get good resolution too. 34

Example 3: Average strategies for Skeptic for a grid of values of p. (The p -strategy makes money if calibration fails for p n close to p.) The derived strategy for Forecaster guarantees good calibration everywhere. Example of a resulting strategy for Skeptic: S n (p) := n 1 i=1 e C(p p i) 2 (y i p i ) Any kernel K(p, p i ) can be used in place of e C(p p i) 2. 35

Skeptic s strategy: S n (p) := n 1 i=1 e C(p p i) 2 (y i p i ) Forecaster s strategy: Choose p n so that n 1 i=1 e C(p n p i ) 2 (y i p i ) = 0. The main contribution to the sum comes from i for which p i is close to p n. So Forecaster chooses p n in the region where the y i p i average close to zero. On each round, choose as p n the probability value where calibration is the best so far. 36

Example 4: Average over a grid of values of p and x. (The (p, x )-strategy makes money if calibration fails for n where (p n, x n ) is close to (p, x ).) Then you get good calibration and good resolution. Define a metric for [0, 1] X by specifying an inner product space H and a mapping continuous in its first argument. Φ : [0, 1] X H Define a kernel K : ([0, 1] X) 2 R by K((p, x)(p, x )) := Φ(p, x) Φ(p, x ). The strategy for Skeptic: S n (p) := n 1 i=1 K((p, x n )(p i, x i ))(y i p i ). 37

Skeptic s strategy: S n (p) := n 1 i=1 K((p, x n )(p i, x i ))(y i p i ). Forecaster s strategy: Choose p n so that n 1 i=1 K((p n, x n )(p i, x i ))(y i p i ) = 0. The main contribution to the sum comes from i for which (p i, x i ) is close to (p n, x n ). So we need to choose p n to make (p n, x n ) close (p i, x i ) for which y i p i average close to zero. Choose p n to make (p n, x n ) look like (p i, x i ) for which we already have good calibration/resolution. 38

References Probability and Finance: It s Only a Game! Glenn Shafer and Vladimir Vovk, Wiley, 2001. www.probabilityandfinance.com: Chapters from book, reviews, many working papers. www.glennshafer.com: Most of my published articles, including the two following. Statistical Science, 21 70 98, 2006: The sources of Kolmogorov s Grundebegriffe. Journal of the Royal Statistical Society, Series B 67 747 764, 2005: Good randomized sequential probability forecasting is always possible. 39

Standard stochastic concept of statistical test: Choose an event A whose probability under P is small. Reject P if A happens. Ville s Theorem: In the stochastic setting... Given an event of probability less than 1/C, there is a strategy for Skeptic that turns $1 into $C without risking bankruptcy. Given a strategy for Skeptic that starts with $1 and does not risk bankruptcy, the probability that it turns $1 into $C or more is no more than 1/C. So the concept of a strategy for Skeptic generalizes the concept of testing with events of small probability. 40

Continuity rules out Dawid s counterexample FOR n = 1, 2,... Skeptic announces continuous S n : [0, 1] R. Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. Skeptic s profit := S n (p n )(y n p n ). Reality can make Forecaster uncalibrated by setting y n := { 1 if pn < 0.5 0 if p n 0.5, Skeptic can then make steady money with S n (p) := { 1 if p < 0.5 1 if p 0.5, But if Skeptic is forced to approximate S n by a continuous function of p, then the continuous function will have a zero close to p = 0.5, and so Forecaster will set p n 0.5. 41

THREE APPROACHES TO FORECASTING FOR n = 1, 2,... Forecaster announces p n [0, 1]. Skeptic announces s n R. Reality announces y n {0, 1}. 1. Start with strategies for Forecaster. Improve by averaging (prediction with expert advice). 2. Start with strategies for Skeptic. Improve by averaging (approach of this talk). 3. Start with strategies for Reality (probability disributions). Improve by averaging (Bayesian theory). 42