Blackwell Optimality in Markov Decision Processes with Partial Observation

Similar documents
arxiv: v1 [math.oc] 23 Dec 2010

An Application of Ramsey Theorem to Stopping Games

Yao s Minimax Principle

Commutative Stochastic Games

Persuasion in Global Games with Application to Stress Testing. Supplement

Randomization and Simplification. Ehud Kalai 1 and Eilon Solan 2,3. Abstract

Best response cycles in perfect information games

On Existence of Equilibria. Bayesian Allocation-Mechanisms

Game Theory: Normal Form Games

Equivalence between Semimartingales and Itô Processes

Deterministic Multi-Player Dynkin Games

Protocols with No Acknowledgment

INTERIM CORRELATED RATIONALIZABILITY IN INFINITE GAMES

Forecast Horizons for Production Planning with Stochastic Demand

In Discrete Time a Local Martingale is a Martingale under an Equivalent Probability Measure

Log-linear Dynamics and Local Potential

Total Reward Stochastic Games and Sensitive Average Reward Strategies

Equilibrium payoffs in finite games

Lecture 7: Bayesian approach to MAB - Gittins index

GAME THEORY. Department of Economics, MIT, Follow Muhamet s slides. We need the following result for future reference.

INTERIM CORRELATED RATIONALIZABILITY IN INFINITE GAMES

A reinforcement learning process in extensive form games

4: SINGLE-PERIOD MARKET MODELS

Stochastic Games with 2 Non-Absorbing States

Kutay Cingiz, János Flesch, P. Jean-Jacques Herings, Arkadi Predtetchinski. Doing It Now, Later, or Never RM/15/022

Discounted Stochastic Games

6.207/14.15: Networks Lecture 10: Introduction to Game Theory 2

On the Lower Arbitrage Bound of American Contingent Claims

Introduction to Probability Theory and Stochastic Processes for Finance Lecture Notes

Martingales. by D. Cox December 2, 2009

Microeconomic Theory II Preliminary Examination Solutions

Finite Memory and Imperfect Monitoring

Economics 209A Theory and Application of Non-Cooperative Games (Fall 2013) Repeated games OR 8 and 9, and FT 5

On Complexity of Multistage Stochastic Programs

Building Infinite Processes from Regular Conditional Probability Distributions

Long run equilibria in an asymmetric oligopoly

Functional vs Banach space stochastic calculus & strong-viscosity solutions to semilinear parabolic path-dependent PDEs.

Econometrica Supplementary Material

MATH3075/3975 FINANCIAL MATHEMATICS TUTORIAL PROBLEMS

4 Martingales in Discrete-Time

Best-Reply Sets. Jonathan Weinstein Washington University in St. Louis. This version: May 2015

Sy D. Friedman. August 28, 2001

EXTENSIVE AND NORMAL FORM GAMES

Efficiency in Decentralized Markets with Aggregate Uncertainty

Outline Introduction Game Representations Reductions Solution Concepts. Game Theory. Enrico Franchi. May 19, 2010

6: MULTI-PERIOD MARKET MODELS

Finding Equilibria in Games of No Chance

BAYESIAN GAMES: GAMES OF INCOMPLETE INFORMATION

CONVERGENCE OF OPTION REWARDS FOR MARKOV TYPE PRICE PROCESSES MODULATED BY STOCHASTIC INDICES

A No-Arbitrage Theorem for Uncertain Stock Model

Asymptotic results discrete time martingales and stochastic algorithms

BROWNIAN MOTION Antonella Basso, Martina Nardon

The folk theorem revisited

Information Aggregation in Dynamic Markets with Strategic Traders. Michael Ostrovsky

COMBINATORICS OF REDUCTIONS BETWEEN EQUIVALENCE RELATIONS

Mathematical Finance in discrete time

Strategies and Nash Equilibrium. A Whirlwind Tour of Game Theory

Finite Population Dynamics and Mixed Equilibria *

Markov Decision Processes II

Dynamic Admission and Service Rate Control of a Queue

Handout 4: Deterministic Systems and the Shortest Path Problem

4 Reinforcement Learning Basic Algorithms

Stochastic Games and Bayesian Games

Non replication of options

Sequential Decision Making

Arbitrage Theory without a Reference Probability: challenges of the model independent approach

On the Number of Permutations Avoiding a Given Pattern

Goal Problems in Gambling Theory*

Competitive Outcomes, Endogenous Firm Formation and the Aspiration Core

Bilateral trading with incomplete information and Price convergence in a Small Market: The continuous support case

Mixed Strategies. In the previous chapters we restricted players to using pure strategies and we

The Real Numbers. Here we show one way to explicitly construct the real numbers R. First we need a definition.

Essays on Some Combinatorial Optimization Problems with Interval Data

Two-Dimensional Bayesian Persuasion

Stochastic Games and Bayesian Games

Solutions of Bimatrix Coalitional Games

Approximate Revenue Maximization with Multiple Items

SHORT-TERM RELATIVE ARBITRAGE IN VOLATILITY-STABILIZED MARKETS

Math-Stat-491-Fall2014-Notes-V

Credibilistic Equilibria in Extensive Game with Fuzzy Payoffs

An Adaptive Learning Model in Coordination Games

Outline of Lecture 1. Martin-Löf tests and martingales

A Decentralized Learning Equilibrium

UNIVERSITY OF VIENNA

Laurence Boxer and Ismet KARACA

Rational Behaviour and Strategy Construction in Infinite Multiplayer Games

The ruin probabilities of a multidimensional perturbed risk model

Outline. 1 Introduction. 2 Algorithms. 3 Examples. Algorithm 1 General coordinate minimization framework. 1: Choose x 0 R n and set k 0.

A relation on 132-avoiding permutation patterns

Comparison of proof techniques in game-theoretic probability and measure-theoretic probability

Optimal stopping problems for a Brownian motion with a disorder on a finite interval

Game Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India August 2012

Convergence of trust-region methods based on probabilistic models

arxiv: v1 [math.pr] 6 Apr 2015

Information Acquisition under Persuasive Precedent versus Binding Precedent (Preliminary and Incomplete)

Introduction to game theory LECTURE 2

THE NUMBER OF UNARY CLONES CONTAINING THE PERMUTATIONS ON AN INFINITE SET

Regression estimation in continuous time with a view towards pricing Bermudan options

3 The Model Existence Theorem

Option Pricing under Delay Geometric Brownian Motion with Regime Switching

Transcription:

Blackwell Optimality in Markov Decision Processes with Partial Observation Dinah Rosenberg and Eilon Solan and Nicolas Vieille April 6, 2000 Abstract We prove the existence of Blackwell ε-optimal strategies in finite Markov Decision Processes with partial observation. Laboratoire d Analyse Geometrie et Applications Institut Galilée, Université Paris Nord, avenue Jean Baptiste Clément, 93430 Villetaneuse, France. e-mail: dinah@math.univ-paris13.fr Department of Managerial Economics and Decision Sciences, Kellogg School of Management, Northwestern University, Evanston IL 60208. e-mail: e-solan@nwu.edu GRAPE, Université Montesquieu-Bordeaux 4, and Laboratoire d Econométrie de l Ecole Polytechnique, 1 rue Descartes, 75 005 Paris, France. e-mail: vieille@poly.polytechnique.fr 1

1 Introduction A well-known result by Blackwell [1] states that, in any Markov Decision Process (MDP thereafter) with finitely many states and finitely many actions, there is a pure stationary strategy that is optimal, for every discount factor close enough to one. This strong optimality property is now referred to as Blackwell optimality. In this paper, we address the problem of existence of Blackwell optimal strategies for finite MDP with partial observation; that is, for finite MDP s in which at the end of every stage, the decision maker receives a signal that depends randomly on the current state and on the action that has been chosen. We prove that, in any such MDP, there is a strategy that is Blackwell ε-optimal; that is, ε-optimal for every discount factor close enough to one. The strategy we construct is moreover ε-optimal in the n-stage MDP, for every n large enough. The standard approach to MDP s with partial observation is to convert it into an auxiliary MDP with full observation and Borel state space. The conditional distribution over the state space Ω given the available information (sequence of past signals and past actions) plays the role of the state variable in the auxiliary MDP. This approach has been developed for instance in [7], [8] and [9]. One then looks for optimal stationary strategies (strategies such that the action chosen in any given stage is only a function of the belief held on the underlying state in Ω). A commonly used criterion is the long-run average cost criterion, see, e.g., [2], [3]. It is well-known that optimal strategies for this criterion do not exist in general MDP s with Borel state space. Hence one imposes assumptions which guarantee the existence of optimal strategies. These assumptions usually have the flavor of an irreducibility condition that one imposes on the transition function of the MDP. For MDP s that arise from a MDP with partial observation, these conditions may be difficult to interpret in terms of the underlying data; see for instance Assumption 7.2, p. 329 in [6]. In the present paper we do not follow this approach but rather use the structure on the auxiliary MDP that is derived from the underlying MDP. Specifically, using a sequence of optimal strategies in the n-stage MDP, and using the compactness of the state space of the auxiliary MDP and the continuity of the payoff on this space, we construct a Blackwell ε-optimal strategy. In Section 2, we present the model and the main results. In section 3, we show on an example that the result is in some respect tight. In Section 6, we construct a Blackwell ε-optimal strategy. This strategy is neither pure nor stationary. In the case of degenerate observation (the decision maker receives no information whatsoever), we construct a pure, stationary Blackwell ε- 2

optimal strategy. Part of this proof serves as an introduction for the general case. It is therefore presented in Section 5. Section 4 contains a number of preliminary results that are used in both proofs. 2 The Model and the Main Results Given a set M, we denote by (M) the set of probability distributions over M, and we identify M with the set of extreme points of (M). A Markov decision process with partial observation is given by: (i) A state space Ω, (ii) an action set A, (iii) a signal set S, (iv) a transition rule q : Ω A (S Ω), (v) a payoff function r : Ω A R, (vi) A probability distribution x 1 (Ω). We assume that Ω, A and S are finite sets. Extensions to more general cases are discussed below. W.l.o.g., we assume that 0 r(ω, a) 1 for every (ω, a) Ω A. An initial state ω 1 is drawn according to x 1. At every stage n the decision maker chooses an action a A, and a pair (s n, ω n+1 ) S Ω of a signal and a new state is drawn according to q(ω n, a n ). The decision maker is informed of the signal s n, but not of the new state ω n+1. Thus, the information available to the decision maker at stage n is the finite sequence a 1, s 1, a 2, s 2,..., a n 1, s n 1 and a behavior strategy for the decision maker is a function that assigns for every such sequence a probability distribution over (A). We set H n = (A S) n 1, and we denote respectively by H = n 1 H n and H = (A Ω S) N the set of finite histories and infinite plays. We denote by H n the algebra over H induced by H n. Each strategy σ and every initial distribution x 1 induce a probability distribution P x1,σ over (H, H ), where H = σ(h n, n 1). Expectations under P x1,σ are denoted by E x1,σ. All norms in the paper are supremum norms. We let γ n (x 1, σ) = E x1,σ[(r(ω 1, a 1 ) + + r(ω n, a n ))/n] denote the expected average payoff in the first n stages. We denote by v n (x 1 ) = sup σ γ n (x 1, σ) the value of the n-stage process. We simply write v n where there is no risk of confusion about the initial distribution. For every λ (0, 1) and every strategy σ we define the λ-discounted payoff as [ ] γ λ (σ) = γ λ (x 1, σ) = E x1,σ (1 λ) λ m 1 r(ω m, a m ), 3 m=1

and the discounted value by v λ = sup γ λ (σ). σ Definition 1 v R is the (uniform) value of the MDP with p.o. (with initial probability distribution x 1 ) if v = lim n v n = lim λ 1 v λ and, for every ε > 0, there exists a strategy σ, a positive integer N 0 N, and λ 0 (0, 1) such that: γ n (x 1, σ) v n ε, n N 0 (1) γ λ (x 1, σ) v λ ε, λ λ 0 (2) Our first main result is that the value always exists. Theorem 2 If Ω, A and S are finite, then v exists. In the case where S = 1, that is, the decision maker receives no informative signal, we get a stronger result. To state this result we need additional notions. For n 1, we denote by y n the conditional law of ω n given H n : for each ω Ω, y n [ω] is the posterior probability in stage n that the process is at state ω given the information available to the decision maker (we do not assume here that S = 1.) Thus, y 1 = x 1. Observe that the value y n (h n ) (Ω) of y n after a given history h n may be computed without knowledge of the strategy. y n is therefore a function H n (Ω) or, equivalently, a random variable (H, H n ) (Ω). Clearly, the law of y n is influenced by the strategy that is followed. A pure strategy is a strategy σ : H (A), such that σ(h) A for each h H. A strategy is stationary if σ(h n ) depends only on the belief y n (h n ) held at stage n. If S = 1, the ε-optimal strategies can be chosen to be pure and stationary. Theorem 3 If Ω and A are finite, and S = 1, then for every ε > 0 there exists a pure stationary ε-optimal strategy. Comment: It might seem that stationarity is an extremely desirable requirement. However, it may well be the case that the decision maker cannot hold twice the same belief over time. In such a case, the stationarity requirement is empty. Comment: It is not clear that the existence of a pure ε-optimal strategy follows from the existence of ε-optimal strategies (i.e., from Theorem 2). The 4

reason is the following. By Kuhn s theorem [4], given x 1 and a strategy σ, there exists a mixed strategy π, i.e., a probability distribution over pure strategies, such that the probability distribution over H obtained by first choosing a pure strategy f according to π, and then following f, coincides with P x1,σ. In particular, given n 1, there exists a strategy f n in the support of π, such that γ n (x 1, f n ) γ n (x 1, σ). However, it is not clear at all that f n can be chosen independently of n. 3 An example Define a MDP with no signals as follows. Set Ω = {ω, ω}, and A = {a, a}. The transition rule q is given by The payoff function r is given by q( ω ω, a) = 1 for each a q(ω ω, ā) = 1, q(ω ω, a) = 1 2. r( ω, ā) = 1, and r(ω, a) = 0 otherwise. The MDP starts from state ω. We identify a probability distribution over Ω with the probability assigned to ω. Observe that the state ω is absorbing. Observe also that: whenever the player chooses ā, the current state does not change, hence the belief remains the same; whenever the player choose a, the current belief (i.e., the probability of being in ω) is divided by two. The uniform value of this MDP is equal to one. Indeed, given ε > 0, let σ be the (stationary) strategy that plays a in the first N = log 2 ε + 2 stages, and plays ā afterwards. Given σ, one has y N+1 < ε. Therefore, E x1,σ [r(ω n, a n )] = 1 y N+1 > 1 ε for each n > N. In particular, lim inf n γ n (σ) = lim inf λ 1 γ λ (σ) > 1 ε. Since v n 1, and v λ 1, the uniform value is indeed equal to 1. This implies lim λ 1 v λ = lim n v n = 1. We now claim that there is no Blackwell optimal strategy. By Kuhn s theorem, it is enough to prove that there is no pure Blackwell optimal strategy. Let σ = (a n ) n N be a pure strategy. We distinguish three (non-exclusive) cases. Case 1: There exists N N, such that a n = ā for every n N. 5

In that case, the sequence (y n ) is constant from stage N on. Therefore, lim n γ n (σ) = lim λ 1 γ λ (σ) = 1 y N < 1. In particular, γ λ (σ) < v λ for λ close to one. Case 2: There exists N N, such that a n = a for every n N. In that case, E σ [r(ω n, a n )] = 0 for each n N. Therefore, lim n γ n (σ) = lim λ 1 γ λ (σ) = 0. Case 3: There exists n 0 N, such that a n0 = ā and a n0 +1 = a. Denote by τ the strategy obtained from σ by permutation of a n0 and a n0 +1. Observe that E τ [r(ω n, a n )] = E σ [r(ω n, a n )] for each n N\ {n 0, n 0 + 1}, E τ [r(ω n0, a n0 )] = E σ [r(ω n0 +1, a n0 +1)] = 0, E τ [r(ω n0 +1, a n0 +1)] > E σ [r(ω n0, a n0 )]. Therefore, γ λ (τ) > γ λ (σ) for λ close to one. In particular, σ is not optimal for λ close to one. A natural question arises. Does there exist a strategy that is Blackwell ε-optimal for each ε > 0? We claim that there is such a pure strategy, but no stationary one. Indeed, let σ = (a n ) n N be a pure stationary strategy. Since y n+1 = y n whenever a n = ā, the stationarity of σ implies that a n+1 = ā as soon as a n = ā. This implies that the sequence (a n ) is eventually constant, i.e., it must be that either case 1 or case 2 above holds. In both cases, σ fails to be ε-optimal, provided ε is small enough. Let now σ = (a n ) be any sequence such that the subset A = {n N,a n = a} of N is infinite and has density zero. Since A is infinite, the sequence (y n ) converges to zero under σ. Therefore, lim E σ [r(ω n, a n )] = 1. (3) n,n/ A Since A has density zero, (3) yields lim n γ n (σ) = lim λ 1 γ λ (σ) = 1. 4 Preliminaries The purpose of this section is to introduce several general results. The first result is standard. It asserts that, given N N, there exists a pure optimal strategy in the N-stage MDP such that the action played in stage n depends only on n and y n. 6

Lemma 4 For each N 1, there exists a pure strategy σ N such that γ N (x 1, σ N ) = v N (x 1 ) and σ N (h n ) is only a function of n and y n (h n ). Proof. Let a strategy σ be given, and define a strategy ˆσ as follows. In stage n 1, it plays a A with the probability P x,σ (a n = a y 1,..., y n ). Since y n is a sufficient statistic about ω n, it is easy to check that γ N (x 1, σ) = γ N (x 1, ˆσ). Observe that ˆσ(h n ) depends only on n and y n (h n ). Using Kuhn s Theorem, there exists a pure strategy σ N such that γ N (x 1, σ N ) γ N (x 1, ˆσ). The result follows. Whenever in the sequel we refer to optimal strategies in the n-stage process, we mean a pure strategy that satisfies the two conditions in Lemma 4. Given m < n, we denote by [ ] 1 γ m,n (x 1, σ) = E x1,σ n m + 1 (r(ω m, a m ) + + r(ω n, a n )), the expected average payoff from stage m up to stage n. Thus, γ n (x 1, σ) = γ 1,n (x 1, σ). Proposition 5 Let x, x (Ω). For every strategy σ and every m < n, γ m,n (x, σ) γ m,n (x, σ) x x. Proof. Let n 1 and h n H n be given. Observe that, for every x (Ω) and for every strategy σ, one has P x,σ (h n = h n ) = ω Ω x(ω)p ω,σ (h n = h n ). In particular, E x,σ [r(s n, a n )] = ω Ω x(ω)e ω,σ [r(s n, a n )]. The result follows. For simplicity, we write γ n (σ) and γ m,n (σ) instead of γ n (x 1, σ) and γ m,n (x 1, σ) whenever there is no possible confusion about x 1. Comment: We claim here that to prove that v is the value, it is enough to prove that v = lim n v n and (1) holds. Since Ω is finite, Proposition 5 implies that (v n ) converges to v uniformly over (Ω). Hence, by Lehrer and Sorin [5], (v λ ) converges uniformly to v. Moreover, one can show that lim inf λ 1 γ λ (x 1, σ) lim inf n γ n (x 1, σ). Hence (2) holds as well. 7

Proposition 6 Let σ, ε > 0 and n N be given, and set N = inf {k N, such that γ m (σ) γ n (σ) ε for every k m n}. (4) Then N 1 + (1 ε)n. Moreover, γ N,m (σ) γ n (σ) ε for every N m n. (5) Given ε > 0 and σ, let N n denote the integer associated with n in (4). Observe that lim n (n N n ) = +. This Proposition has the same flavor as Proposition 2 in [5]. Proof. Clearly, N n. Note that if N > 1 then γ N 1 (σ) < γ n (σ) ε. We first show that N 1 + (1 ε)n. Indeed, otherwise, N > 1, hence γ N 1 (σ) < γ n (σ) ε. Since payoffs are bounded by 1, γ n (σ) N 1 n γ N 1(σ) + n N + 1 n < γ n (σ) ε + ε = γ n (σ) a contradiction. Next we show that (5) holds. Fix an integer m such that N m n. If N = 1, one has γ N,m (σ) = γ m (σ) γ n (σ) ε. If N > 1, γ N 1 (σ) < γ n (σ) ε, while γ m (σ) γ n (σ) ε. It follows that γ N,m (σ) γ n (σ) ε. 5 The Case of No Signals This section is devoted to the proof of Theorem 3. Thus, we assume that no signal is available. The initial distribution x 1 is fixed throughout the section. A pure strategy is reduced to a sequence of actions: the action that is played at each stage. Moreover, if σ is pure, the posterior distribution at stage n depends deterministically on σ. We write y n (σ) for the posterior distribution at stage n: y n (σ)[ω] = P x1,σ(ω n = ω). If σ = (a 1, a 2,... ) A N is a strategy, we define for every positive integer m N the truncated strategy σ m = (a m, a m+1,... ) and the prefix m σ = (a 1,..., a m ). Define w = lim sup n v n, and fix ε > 0. Let (n i ) i N be a subsequence such that lim i v ni = w, and v ni w < ε/2 for every i N. Let σ i be a pure optimal strategy in the n i -stage problem (that satisfies the two conditions of Lemma 4.) Thus, γ ni (σ i ) = v ni. 8

Given i N, we let N i 1 + (1 ε)n i be the integer obtained by applying Proposition 6 to n i. Possibly by taking a subsequence, we may assume w.l.o.g. that N 1 N i for each i. We let y i = y Ni (σ i ) be the posterior distribution over states induced by σ i at stage N i. Since Ω is finite, (Ω) is compact, hence there exists y (Ω) and a subsequence of {y i }, still denoted by {y i }, such that y i y < ε, for each i N. For each i N define π i as: follow σ 1 up to N 1, switch to σ N i i N 1. Formally, at stage π i (n) = { σ1 (n) for 1 n N 1 1 σ i (N i + n N 1 ) for N 1 n Set m i = N 1 + n i N i. Since N 1 N i, one has m i n i. Note that lim inf i m i = +. Proposition 7 If m satisfies (N 1 1)/ε < m m i then γ m (π i ) w 4ε. Proposition 7 asserts that each π i gives high payoff in all m-stage problems, provided m is sufficiently large (but smaller than m i ). Moreover, the lower bound on m is independent of i. Proof. Fix an integer m such that (N 1 1)/ε < m m i. By construction, y N1 (π i ) = y 1, hence γ m (x 1, π i ) = N 1 1 m γ N 1 1(x 1, π i ) + m N 1 + 1 γ N1,m(x 1, π i ) m = N 1 1 m γ N 1 1(x 1, π i ) + m N 1 + 1 γ m N1 +1(y 1, π N 1 i ) m By the assumption on m, (m N 1 + 1)/m 1 ε. Since y 1 y i < ε, we get by Proposition 5, and since γ N1 1(π i ) 0, γ m (x 1, π i ) (1 ε) (γ Ni,m N 1 +N i (x 1, π i ) ε). Since N 1 N i, m N 1 + N i n i, hence γ Ni,m N 1 +N i (y i, π i ) w 2ε. The result follows. Proposition 8 In the case S = 1, the uniform value exists. 9

Proof. Since A is finite, by a diagonal extraction argument there exists a pure strategy π such that every prefix of π is a prefix of infinitely many π i s: for each m, m π = m π i for infinitely many i. In particular, for every m > (N 1 1)/ε, γ m (π) w 4ε. In particular, v m w 4ε. Since ε > 0 is arbitrary, one has w = lim n v n and π is a 4ε-optimal strategy. Proof of Theorem 3. Let π = (a 1, a 2,... ) be a pure ε-optimal strategy; that is, for some n 0 N, γ n (π) w ε for every n n 0. Let y n = y n (π) be the posterior distribution at stage n. Case 1: (y n ) n N is eventually periodic; that is, there exists n 1 N and d N such that y n = y n + d for every n n 1. Since π is ε-optimal, it follows that the expected average payoff along the period is at least w ε: γ n1,n 1 +d 1(π) w ε. It follows that there exist n 2 n 3 such that (i) y n2 = y n3, (ii) y i y j for every n 2 i < j < n 3, and (iii) γ n2,n 3 1(π) w ε. Let Y = {y n, n = 1,..., n 3 } be the set of all posterior distributions in the first n 3 stages. Consider the directed graph whose vertices are the elements in Y, and which contains the edge (y, y ) Y Y if and only if (y, y ) = (y n, y n+1 ) for some n {1,..., n 3 1}. Thus we connect with an edge any two consecutive elements in the finite sequence (y n ) n 3 n=1. Clearly there is a path from y 1 to any y Y. Let y 1 = y i1, y i2,..., y ik be a shortest path that connects y 1 to the set {y n2, y n2 +1,..., y n3 }. In particular, y ij y ij for every 1 j < j k. Assume w.l.o.g. that y ik = y n2. Define π = (a i1, a i2,..., a ik 1, a n2, a n2 +1,..., a n3 1, a n2, a n2 +1,..., a n3 1,... ). By construction, y n (π ) = y in (π) for each n < k, y k (π ) = y n2 (π), and the sequence (y n (π)) n k coincides with the periodic sequence (y n2 (π),...y n3 1(π), y n2 (π),..., y n3 1(π),...). Each of the posteriors y n (π ), n < k appears only once, hence π is stationary. Since γ n2,n 3 1(π) w ε, we have γ n (π ) w 2ε for every n k(n 3 n 2 )/ε. Case 2: There are two integers 0 < n 1 < n 2 such that y n1 = y n2, and γ n1,n 2 1(π) w ε. Define the strategy π = (a 1, a 2,..., a n1, a n1 +1,..., a n2 1, a n1,..., a n2 1,... ). Then π is 2ε-optimal, and (y n (π )) is eventually periodic. We can then apply Case 1 to π. 10

Case 3: There is some y (Ω) that appears infinitely often in the sequence (y n ) n N. Since for every n sufficiently large, γ n (π) w ε, it follows that there exist n 1 < n 2 such that y n1 = y n2 = y and γ n1,n 2 1(π) w ε. Apply now Case 2. Case 4: None of the above hold. Since Case 3 does not hold, every y (Ω) that appears in the sequence (y n ) n N, does so only finitely many times. Since Case 2 does not hold, the expected average payoff between two appearances of any y (Ω) in (y n ) is below w ε. Define a sequence (i k ) k N as follows: and i 1 = max {n 1, y n = y 1 }, i k+1 = max{n 1, y n = y ik +1}. (6) In words, i 1 is the last occurrence of the initial belief, i 2 is the last occurrence of the belief held in stage i 1 + 1, and so on. Since y ik appears only finitely many times in the sequence (y n ), the maximum in (6) is finite. Clearly i k+1 > i k. Note that y ik+1 = y ik +1, for each k. Define now a strategy π = (a i1, a i2, a i3,... ). Since y ik+1 = y ik +1, it follows by induction that y ik+1 = y(a i1, a i2,..., a ik ), where y(a i1, a i2,..., a ik ) is the posterior probability held after playing actions a i1, a i2,..., a ik. It also follows that no element in the sequence (y ik ) appears twice. In particular, the strategy π is stationary. We now argue that for every k 0 n 0, γ k0 (π ) w ε. Set n = i k0 and i 0 = 0. Note that Clearly, k 0 n = (i k i k 1 ) = k 0 + k=1 nγ n (π) = k 0 γ k0 (π ) + k k 0 i k+1 >i k +1 k k 0 i k+1 >i k +1 (i k i k 1 1). (i k+1 i k 1)γ ik +1,i k+1 1(π). Since Case 2 does not hold, γ ik +1,i k+1 1(π) < w ε, whenever i k+1 > i k + 1, Since n k 0 n 0, γ n (π) w ε. It follows that γ k0 (π ) w ε, as desired. 11

Comment. The fact that the action set A is finite was used in the diagonal extraction argument in the proof of Proposition 8. However, the proof can be extended to compact metric action spaces provided the functions a r(ω, a) and a q(ω, a) are continuous in a, for each ω Ω. To see why the diagonal extraction argument works in that case, take for every n N a finite subset A n A such that for each a A there is some ā(a) A n with sup ω r(ω, a) r(ω, ā(a)) < ε and sup q(ω, a) q(ω, ā(a)) < ε/2 n. (7) ω Define for every i N the strategy π i by π i(n) = ā(π i (n)). By (7), γ n (π i ) γ n (π i) < 2ε. Since for each fixed n, {π i(n)} i N is finite, one can apply the diagonal extraction argument to {π i} i N, and get a strategy π such that every prefix of π is a prefix of infinitely many π i s. Then π i is 3ε-optimal. 6 The General Case This section is devoted to the proof of Theorem 2. At first we follow the same path as in the proof of Theorem 3. However, since now the signal set is not degenerate, the posterior distribution at stage N i depends on the signals the decision maker received. Hence, before the process starts, the decision maker who follows some strategy has a probability distribution over the possible posteriors he may have at stage N i. We are thus forced to work with the space ( (Ω)), which is no longer finite dimensional. The proof will be amended to deal with this difficulty. Fix ε > 0 once and for all. Denote w = lim sup n v n, and let (n i ) be a subsequence such that lim i v ni = w and w v ni < ε for every i N. For each i N, let σ i be an optimal strategy in the n i -stage MDP (that satisfies the two conditions of Lemma 4), and let N i 1 + (1 ε)n i be the integer obtained by applying Proposition 6 to n i. We assume w.l.o.g. that N 1 N i for each i. Recall that y Ni is the posterior distribution over Ω at stage N i, given the history up to that stage. Since A and S are finite, y Ni may take only finitely many values. We denote by p i the law of y Ni when the strategy σ i is followed (under P σi ): p i has finite support supp(p i ) and p i (y) = P σi (y Ni = y) for each y (Ω). Comment. A natural idea is to repeat the proof of the previous section, by using the law of the belief as relevant state variable, i.e. by dealing with 12

the auxiliary state space ( (Ω)). Observe that ( (Ω)) is no longer finitedimensional but is compact in the w -topology, which is a metric topology. Let d be a corresponding metric. The proof of the previous section would go through if one was able to prove the following Lipschitz property: for every p, p ( (Ω)), σ and n N, γ n (p, σ) γ n (p, σ) d(p, p ), where γ n (p, σ) denotes the expectation of γ n (x, σ) under p. However, it is not clear that this condition holds. We therefore choose a different route, which involves a discretization of (Ω), and uses the Lipschitz condition expressed in Lemma 5. Let T be a finite partition of (Ω) into sets of diameter smaller than ε. By Lemma 5, given T T, x, x T, a strategy σ and every n N, one has γ n (x, σ) γ n (x, σ) < ε. (8) Given p ( (Ω)) with finite support, we denote by ˆp the probability induced by p on T : ˆp[T ] = p[x] T T. x supp(p) T Since T is a finite partition, there is a subsequence of (ˆp i ) i N that converges to a limit ˆp. We still denote this subsequence by (ˆp i ) i N. We assume moreover that for every i N, ˆp i ˆp < ε/2. In particular, ˆp i ˆp 1 < ε for every i N. In the case of no signals, we defined a strategy π i as: follow σ 1 up to stage N 1, then switch to the sequence of actions prescribed by σ i after stage N i. We will proceed in a similar way here. There is however a small difficulty. The action that σ i plays in stage N i depends on the belief y Ni. Therefore, one needs to define a map that associates to the true belief y N1 held at stage N 1 a fictitious value for y Ni. Indeed, the possible beliefs in stage N 1 need not be the same as the possible beliefs in stage N i. The solution is simply to select a fictitious belief x according to the conditional distribution p i [ T (y N1 )], where, given y (Ω), T (y) is the element of T that contains y. We need an additional notation. For each x (Ω), we define the strategy σ N i i [x] induced by σ i after stage N i, given the belief x, as follows. For each history (a 1, s 1,..., a m, s m), we set σ N i i [x] (a 1, s 1,..., a m, s m) = σ i (a 1, s 1,..., a Ni 1, s Ni 1, a 1, s 1,..., a m, s m), 13

where (a 1, s 1,..., a Ni 1, s Ni 1) is any sequence in H Ni such that y Ni (a 1, s 1,..., a Ni 1, s Ni 1) = x. Since σ i is stationary, this is independent of the particular sequence (a 1, s 1,..., a Ni 1, s Ni 1). (If no such sequence exists, the definition of σ N i i [x] is irrelevant). We now define, for every i N, a strategy π i as follows: Follow σ 1 up to stage N 1 1. If p i [T (y N1 )] = 0, continue in an arbitrary way. Otherwise, choose x according to p i [ T (y N1 )], and continue with σ N i i [x]. Observe that the definition of π i involves choosing at stage N 1 a pure strategy at random. Such a strategy is called a mixed strategy. By Kuhn s theorem [4], there is a behavior strategy that induces the same probability distribution over H as π i. We may therefore view π i as a behavior strategy. Proposition 9 For any m such that N 1 /ε m N 1 + n i N i, one has γ m (π i ) w 5ε. Proof. By the definition of π i, and since payoffs are bounded by 1: γ m (π i ) = N 1 1 m γ N 1 1(σ 1 ) + m N 1 + 1 m y (Ω) x T (y) p 1 (y)p i (x T (y))γ m N1 +1(y, σ N i i If x, y (Ω) belong to the same element of T, one has γ m N1 +1(y, σ N i i [x]) γ m N1 +1(x, σ N i i [x]) ε. Therefore [x]). γ m (π i ) N 1 1 m γ N 1 1(σ 1 ) (9) + m N 1 + 1 ˆp 1 (T ) p i (x T )γ m N1 +1(x, σ N i i [x]) ε. (10) m T T x T 14

Since ˆp i ˆp 1 < ε, ˆp 1 (T ) p i (x T )γ m N1 +1(x, σ N i i [x]) T T x T Since m N 1 /ε, substituting into (9) yields x (Ω) p i (x)γ m N1 +1(x, σ N i i γ Ni,m N 1 +N i (σ i ) ε γ m (π i ) (1 ε)γ Ni,m N 1 +N i (σ i ) 2ε w 5ε. [x]) ε The last step is to construct from the sequence (π i ) i N, using a diagonal extraction argument, a strategy π that is 6ε-optimal. Let n 1 be given. Since H n is finite, there exists a sequence (i n (j)) j N such that lim j π in (j)(h) exists for every h H n. We denote by π(h) the limit. W.l.o.g., we may assume that (i n+1 (j)) j is a subsequence of (i n (j)) j for each n. Clearly, for each n, γ n (π) = lim j γ n (π in (j)). By Proposition 9, γ n (π) w 5ε, for every n N 1 /ε. Hence Theorem 2 is proved. We conclude by discussing several extensions. Comment. The extension to a compact set of actions also holds in the general case, under the same conditions as in the case of no signals, as discussed above. Comment. The extension to MDP with finite Ω, A and countable set of signals S is straightforward. Indeed, given ε > 0, there exist finite subsets S n of S such that, given any strategy σ and any initial distribution x 1 (Ω), P x1,σ(s n / S n for some n) ε. The proof then essentially reduces to the case of a finite set of signals. Comment. The extension to MDP with finite A, countable Ω does not hold, even when S is a singleton. Indeed, there are examples, see [5] for instance, of MDP with finite A, countable Ω and deterministic transitions, that have no value. For such MDP, the sequence of past actions enables the decision maker to recover the current state of the MDP. Hence the assumption of partial observation is irrelevant. Comment. Our proof works in the case of MDPs with a compact metric space Ω, and finite action set A and signal set S, as long as (8) holds. 15

References [1] D. Blackwell. Discrete dynamic programming. Annals of Mathematical Statistics, 33:719 726, 1962. [2] V.S. Borkar. Control of markov chains with long-run average cost criterion. In W. Fleming and P.L. Lions, editors, Stochastic Differential Systems, Stochastic Control Theory and Applications, The IMA Volumes in Mathematics and Its Applications, Vol. 10, pages 57 77. Springer-Verlag, Berlin, 1988. [3] E. Fernandez-Gaucherand, A. Araposthatis, and S.I. Marcus. On partially observable markov decision processes with an average cost criterion. In Proceedings of the 28th IEEE Conference on Decision and Control, pages 1267 1272, Tampa, FL., 1989. [4] H.W. Kuhn. Extensive games and the problem of information. In H.W. Kuhn and A.W. Tucker, editors, Contributions to the Theory of Games. Annals of Mathematics Study 28, Princeton University Press, 1953. [5] E. Lehrer and S. Sorin. A uniform tauberian theorem in dynamic prgramming. Mathematics of Operations Research, 17:303 307, 1992. [6] M.L. Puterman. Markov Decision Processes: Discrete Stochastic Dynamic Programming. Wiley, New York, 1994. [7] D. Rhenius. Incomplete information in markovian decision models. Annals of Statistics, 2:1327 1334, 1974. [8] S. Sawaragi and T. Yoshikawa. Discrete time markovian decision processes with incomplete state observation. Annals of Mathematical Statistics, 41:78 86, 1970. [9] A.A. Yushkevich. Reduction of a controlled markov model with incomplete date to a problem with complete information in the cas of borel state and control spaces. Theory of Probability and its Applications, 21:153 158, 1976. 16