Stochastic Dual Dynamic Programming Algorithm for Multistage Stochastic Programming

Size: px
Start display at page:

Download "Stochastic Dual Dynamic Programming Algorithm for Multistage Stochastic Programming"

Transcription

1 Stochastic Dual Dynamic Programg Algorithm for Multistage Stochastic Programg Final presentation ISyE 8813 Fall 2011 Guido Lagos Wajdi Tekaya Georgia Institute of Technology November 30, 2011

2 Multistage Stochastic Linear Programg A 1 x 1 =b 1 x 1 0 c1 T x 1 +E Analysis of Stochastic Dual Dynamic Programg Method Alex Shapiro (2010) B 2 x 1 +A 2 x 2 =b 2 x 2 0 c2 T x 2 + E... + E B T x T 1 +A T x T =b T x T 0 Applications: planning problems in ing, energy, forestry, etc. ct T x T... Challenges: Tractability of E Stagewise dependence of data process {ξ t := (c t, B t, A t, b t)} t=1,...,t Curse of dimensionality

3 Summary 1 Two-stage stochastic programg Cutting-plane method SDDP algorithm for two-stage SP 2 Convergence and main contributions

4 Two-stage stochastic programg 1 Two-stage stochastic programg Cutting-plane method SDDP algorithm for two-stage SP 2 Convergence and main contributions

5 Two-stage stochastic programg True problem: c T x + E Ax=b x 0 T x+w y=h y 0 q T y Take random sample ξ 1,..., ξ N and approximate E 1 N N j=1 = Sample Average Approximation (SAA) problem: c T x + 1 N N q jt y Ax=b x 0 j=1 T j x+w j y=h j y 0 } {{ } } Q(x,ξ j ):= {{ } Q(x):=

6 Two-stage stochastic programg Cutting-plane method Basic Ingredients c T x + 1 N Ax=b x 0 N Q(x, ξ j ) j=1 } {{ } Q(x):= where Q(x, ξ) := max Assume: relatively complete recourse, i.e. feasible x, Q(x, ξ) < a.s. 1 N N j=1 Q(, ξj ) convex piecewise-linear, and problem is T x+w y=h y 0 q T y!!! For f convex: f (x 0) := { d : x f (x) f (x 0) + d T (x x } 0) Q(, [ ξ)(x 0) = T T {π : π opt. sol. of dual of Q(x 0, ξ)} ] 1 N N j=1 Q(, ξj ) (x 0) = 1 N N j=1 T j T {π :. π opt. sol. of dual of Q(x 0, ξ j )}

7 Two-stage stochastic programg Cutting-plane method Basic Ingredients c T x + 1 N Ax=b x 0 N Q(x, ξ j ) j=1 } {{ } Q(x):= where Q(x, ξ) := max Assume: relatively complete recourse, i.e. feasible x, Q(x, ξ) < a.s. Conclusion: c T x + 1 N Q(x, ξ j ) is: N j=1 easy to compute on given x difficult to optimize easy to compute subgradient on given x T x+w y=h y 0 q T y

8 Two-stage stochastic programg Cutting-plane method Cutting Plane Algorithm Given sample { ξ j = (q j, T j, W j, h j ) } j=1,...,n, 0. k 1; LB 1 ; UB 1 ; Q 1 (x) LB 1 x 1. ( Forward Step ) Let x k be the solution of: LB k c T x + Q k (x) Ax=b x 0 2. ( Backward Step ) Compute: Q(x k ) 1 N N j=1 Q(x k, ξ j ) g k 1 N N j=1 T jt π j,k (subgradient) Let UB k c T x k + Q(x k ). If UB k LB k ɛ, END. Else (LB k < UB k ), add plane Q(x k ) + g kt (x x k ) to Q k : Q k+1 (x) max{q k (x), Q(x k ) + g kt (x x k )} 3. k k + 1, iterate from 1.

9 Two-stage stochastic programg SDDP algorithm for two-stage SP SDDP algorithm I Given sample { ξ j = (q j, T j, W j, h j ) } j=1,...,n, 0. k 1; UB 1 ; LB 1 ; Q 1 (x) LB 1 x 1. Forward Step 1.1 Let x k be the solution of: LB k c T x + Q k (x) Ax=b x Take subsample {ξ (j) } M j=1 of {ξj } N j=1 (N >> M), and with values {ϑ j := c T x k + Q(x k, ξ (i) )} M j=1 compute (1 α) confidence upper bound of true problem opt. value ϑ : UB k ϑ + z α/2 σ ϑ / M where ϑ := 1 M M j=1 ϑ j and σ 2 ϑ := 1 M 1 M j=1 (ϑ j ϑ) If UB k LB k ɛ, END.

10 Two-stage stochastic programg SDDP algorithm for two-stage SP SDDP algorithm II 2. Backward Step 2.1 Compute: Q(x k ) 1 N N j=1 Q(x k, ξ j ) = 1 N N j=1 g k 1 N N j=1 T jt π j,k (subgradient) 2.2 Add plane Q(x k ) + g kt ( x k ) to Q k ( ): max T j x k +W j y=h j y 0 Q k+1 (x) max{q k (x), Q(x k ) + g kt (x x k )} 3. k k + 1, iterate from 1. q jt y

11 1 Two-stage stochastic programg Cutting-plane method SDDP algorithm for two-stage SP 2 Convergence and main contributions

12 True problem A 1 x 1 =b 1 x 1 0 c1 T x 1 +E B 2 x 1 +A 2 x 2 =b 2 x 2 0 c2 T x 2 + E... + E Equivalently (Dynamic Programg equations): where A 1 x 1 =b 1 x 1 0 c T 1 x 1 + E [Q 2(x 1, ξ 2)] }{{} Q 2 (x 1 ):= B T x T 1 +A T x T =b T x T 0 ct T x T Q t(x t 1, ξ t) := inf B t x t 1 +A t x t=b t x t 0 Q T (x T 1, ξ T ) := inf B T x T 1 +A T x T =b T x T 0 ct T x t + E [Q t+1(x t, ξ t+1)] }{{} Q t+1 (x t) ct T x T t = 2,..., T 1 Assumptions 1 Process is stagewise independent, i.e. ξ t+1 indep. of ξ 1,..., ξ t. 2 Problem has relatively complete recourse

13 SAA problem Take random sample { ξj t = ( c tj, Ãtj, B tj, b } tj) for each stage j=1,...,n t t = 2,..., T. SAA problem is: A 1 x 1 =b 1 x 1 0 c1 T x 1+ 1 N 2 N 2 j=1 B j2 x 1 +Ãj2x 2 = b j2 x 2 0 N T c j2 T x N 3 N N j=1 T j=1 Equivalently (Dynamic Programg equations): where A 1 x 1 =b 1 x 1 0 c1 T x N 2 Q 2,j(x 1) N 2 j=1 } {{ } Q 2 (x 1 ) B jt x T 1 +ÃjT x T = b jt x T 0 c jt T x T Q t,j(x t 1) := B tj x t 1 +Ãtj x t= b tj x t 0 Q T,j(x T 1) := B Tj x T 1 +ÃTj x T = b Tj x T 0 c T tj x t + 1 c T Tjx T N t+1 N t+1 j=1 Q t+1,j(x t) } {{ } Q t+1 (x t) t = 2,..., T 1

14 SDDP method: the idea!!! Cost-to-go functions Q t (x t 1 ) = 1 N t N t j=1 Q t,j (x t 1 ) t = 2,..., T are convex piecewise-linear Approximate A 1x 1 =b 1 x 1 0 B tj x t 1 +Ãtj xt= b tj x t 0 c1 Tx 1 + Q 2 (x 1 ) by A 1x 1 =b 1 x 1 0 c tj Tx t + Q t+1 (x t ) by B tj x t 1 +Ãtj xt= b tj x t 0 c T 1 x 1 + Q 2 (x 1 ) c T tj x t + Q t+1 (x t ) where Q 2 ( ),..., Q T ( ) are convex, piecewise-linear and Q t ( ) Q t ( ) t = 2,..., T In successive iterations, refine lower approximations Q t ( ) using subgradient of Q t ( ): Q k t ( ) Q k+1 t ( ) Q k+2 t ( )... Q t ( )

15 SDDP method: Forward step At iteration k 1, we have lower approximations Q 2,..., Q T Take subsample {( ξ (j) (j) 2,..., ξ T )}M j=1 of original sample For j = 1,..., M, take sampled process ( ξ (j) (j) 2,..., ξ ) and solve T A1x1=b1 x1 0 ξ (j) t, x t 1,j B t(j) x t 1,j +Ãt(j) xt = b t(j) xt 0 ξ (j) T, x T 1,j B T (j) x T 1,j +ÃT (j) x T = b T (j) x T 0 c1 T x1 + Q2(x1) c t(j) T xt + Qt+1(xt) c T T (j) x T x 1j x tj ( ξ (j) [t] ) t = 2,..., T 1 x T j ( ξ (j) [T ] ) obtaining candidate policy values x 1j, x 2j,..., x T j with cost T ϑ j ct(j) T x tj t=1 It s a (1 α) confidence upper bound of true opt. value ϑ : UB k ϑ + z α σ ϑ / M where ϑ := 1 M M j=1 ϑ j and σ ϑ 2 := 1 M M 1 j=1 (ϑ j ϑ) 2.

16 SDDP method: Backward step We have candidate values x 1, x 2,..., x T. At stage t = T : For x T 1 = x T 1 and for j = 1,..., N T, solve: Q T,j (x T 1 ) := B T,j x T 1 +ÃT,j x T = b T,j x T 0 and let π T,j be opt. dual solution. Let 1 NT Q T (x T 1 ) := N T Add cut g T := 1 N T c T T,jx T Q j=1 T,j (x T 1 ) NT B j=1 T T,j π T,j. L T (x T 1 ) := Q T (x T 1 ) + g T T (x T 1 x T 1 ) to lower approx. Q T used in stage t = T 1: Q T ( ) := max{q T ( ), L T ( )}.

17 SDDP method: Backward step At stage t = T 1: For x T 2 = x T 2 and for j = 1,..., N T 1, solve: Q T 1,j (x T 2 ) := B T 1,j x T 2 +ÃT 1,j x T 1 = b T 1,j x T 1 0 and let π T 1,j be opt. dual solution. Let Add cut c T T 1,jx T 1 +Q T (x T 1 ) 1 NT Q T 1 (x T 2 ) := 1 N T 1 j=1 Q T 1,j (x T 2 ) g T 1 := 1 NT 1 N T 1 j=1 B T T 1,j π T 1,j. L T 1 (x T 2 ) := Q T 1 (x T 2 ) + g T T 1(x T 2 x T 2 ) to lower approx. Q T 1 used in stage t = T 2: Q T 1 ( ) := max{q T 1 ( ), L T 1 ( )}.

18 SDDP method: Backward step At stage t = T 2: At stage t = 1: solve LB k A 1 x 1 =b 1 x 1 0 c T 1 x 1 + Q 2 (x 1 ) LB k is, on average, lower bound to ϑ true optimal value

19 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: True problem

20 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SAA problem

21 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Forward step

22 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Forward step

23 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Forward step

24 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Forward step

25 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Backward step

26 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Backward step

27 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Backward step

28 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Forward step

29 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Forward step

30 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Forward step

31 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Backward step

32 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Backward step

33 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Backward step...

34 Summary: KEY IDEAS for SDDP algorithm Forward step: sample process ξ 1,..., ξ T implementable policy x 1,..., x T Repetitions Upper bound on true optimal value Backward step: Cost-to-go functions Q t (x t 1 ) = 1 N t N t max c t,jx T t + Q t+1 (x t ) j=1 B t,j x t 1+Ãt,j x t= b t,j x t 0 are convex piecewise-linear functions of x t 1 Refine lower approximations Q t ( ) using subgradient [ ] Qt ( ) (x t 1 ) = 1 N t N t j=1 Lower bound on true optimal value B T t,j {π t,j : opt. sol. of dual... }

35 Convergence and main contributions Proposition (Convergence) Assume i. At the forward step, process subsamples are taken independently of each other ii. At all iterations, approximated problems A 1 x 1 =b 1 x 1 0 c1 T x 1 + Q 2 (x 1 ) and B t,j x t 1 +Ãt,j x t = b t,j x t 0 have finite optimal value iii. In the backward step, basical solutions are used c T t,jx t + Q t+1 (x t ) Then, after a sufficiently but finite large amount of iterations of the SDDP algorithm, the forward procedure defines an optimal policy for the SAA problem.

36 Convergence and main contributions Tractability? Issue: total number of scenarios Π T t=2 N t Cutting plane for two-stage: bad SDDP algorithm: generalization of cutting plane method even worse? One run of Backward step: solve 1 + N N T LP s One run of Forward step: solve 1 + M(T 1) LP s Tractability : SDDP method construct feasible policy

37 Convergence and main contributions THE END

Worst-case-expectation approach to optimization under uncertainty

Worst-case-expectation approach to optimization under uncertainty Worst-case-expectation approach to optimization under uncertainty Wajdi Tekaya Joint research with Alexander Shapiro, Murilo Pereira Soares and Joari Paulo da Costa : Cambridge Systems Associates; : Georgia

More information

Robust Dual Dynamic Programming

Robust Dual Dynamic Programming 1 / 18 Robust Dual Dynamic Programming Angelos Georghiou, Angelos Tsoukalas, Wolfram Wiesemann American University of Beirut Olayan School of Business 31 May 217 2 / 18 Inspired by SDDP Stochastic optimization

More information

DASC: A DECOMPOSITION ALGORITHM FOR MULTISTAGE STOCHASTIC PROGRAMS WITH STRONGLY CONVEX COST FUNCTIONS

DASC: A DECOMPOSITION ALGORITHM FOR MULTISTAGE STOCHASTIC PROGRAMS WITH STRONGLY CONVEX COST FUNCTIONS DASC: A DECOMPOSITION ALGORITHM FOR MULTISTAGE STOCHASTIC PROGRAMS WITH STRONGLY CONVEX COST FUNCTIONS Vincent Guigues School of Applied Mathematics, FGV Praia de Botafogo, Rio de Janeiro, Brazil vguigues@fgv.br

More information

Report for technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional do Sistema Elétrico Risk Averse Approach

Report for technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional do Sistema Elétrico Risk Averse Approach Report for technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional do Sistema Elétrico Risk Averse Approach Alexander Shapiro and Wajdi Tekaya School of Industrial and

More information

On solving multistage stochastic programs with coherent risk measures

On solving multistage stochastic programs with coherent risk measures On solving multistage stochastic programs with coherent risk measures Andy Philpott Vitor de Matos y Erlon Finardi z August 13, 2012 Abstract We consider a class of multistage stochastic linear programs

More information

Multistage risk-averse asset allocation with transaction costs

Multistage risk-averse asset allocation with transaction costs Multistage risk-averse asset allocation with transaction costs 1 Introduction Václav Kozmík 1 Abstract. This paper deals with asset allocation problems formulated as multistage stochastic programming models.

More information

Stochastic Dual Dynamic Programming

Stochastic Dual Dynamic Programming 1 / 43 Stochastic Dual Dynamic Programming Operations Research Anthony Papavasiliou 2 / 43 Contents [ 10.4 of BL], [Pereira, 1991] 1 Recalling the Nested L-Shaped Decomposition 2 Drawbacks of Nested Decomposition

More information

SOLVING ROBUST SUPPLY CHAIN PROBLEMS

SOLVING ROBUST SUPPLY CHAIN PROBLEMS SOLVING ROBUST SUPPLY CHAIN PROBLEMS Daniel Bienstock Nuri Sercan Özbay Columbia University, New York November 13, 2005 Project with Lucent Technologies Optimize the inventory buffer levels in a complicated

More information

Stochastic Dual Dynamic integer Programming

Stochastic Dual Dynamic integer Programming Stochastic Dual Dynamic integer Programming Shabbir Ahmed Georgia Tech Jikai Zou Andy Sun Multistage IP Canonical deterministic formulation ( X T ) f t (x t,y t ):(x t 1,x t,y t ) 2 X t 8 t x t min x,y

More information

Financial Optimization ISE 347/447. Lecture 15. Dr. Ted Ralphs

Financial Optimization ISE 347/447. Lecture 15. Dr. Ted Ralphs Financial Optimization ISE 347/447 Lecture 15 Dr. Ted Ralphs ISE 347/447 Lecture 15 1 Reading for This Lecture C&T Chapter 12 ISE 347/447 Lecture 15 2 Stock Market Indices A stock market index is a statistic

More information

IE 495 Lecture 11. The LShaped Method. Prof. Jeff Linderoth. February 19, February 19, 2003 Stochastic Programming Lecture 11 Slide 1

IE 495 Lecture 11. The LShaped Method. Prof. Jeff Linderoth. February 19, February 19, 2003 Stochastic Programming Lecture 11 Slide 1 IE 495 Lecture 11 The LShaped Method Prof. Jeff Linderoth February 19, 2003 February 19, 2003 Stochastic Programming Lecture 11 Slide 1 Before We Begin HW#2 $300 $0 http://www.unizh.ch/ior/pages/deutsch/mitglieder/kall/bib/ka-wal-94.pdf

More information

Approximation of Continuous-State Scenario Processes in Multi-Stage Stochastic Optimization and its Applications

Approximation of Continuous-State Scenario Processes in Multi-Stage Stochastic Optimization and its Applications Approximation of Continuous-State Scenario Processes in Multi-Stage Stochastic Optimization and its Applications Anna Timonina University of Vienna, Abraham Wald PhD Program in Statistics and Operations

More information

A Robust Option Pricing Problem

A Robust Option Pricing Problem IMA 2003 Workshop, March 12-19, 2003 A Robust Option Pricing Problem Laurent El Ghaoui Department of EECS, UC Berkeley 3 Robust optimization standard form: min x sup u U f 0 (x, u) : u U, f i (x, u) 0,

More information

Scenario Generation and Sampling Methods

Scenario Generation and Sampling Methods Scenario Generation and Sampling Methods Güzin Bayraksan Tito Homem-de-Mello SVAN 2016 IMPA May 9th, 2016 Bayraksan (OSU) & Homem-de-Mello (UAI) Scenario Generation and Sampling SVAN IMPA May 9 1 / 30

More information

MULTISTAGE STOCHASTIC PROGRAMS WITH A RANDOM NUMBER OF STAGES: DYNAMIC PROGRAMMING EQUATIONS, SOLUTION METHODS, AND APPLICATION TO PORTFOLIO SELECTION

MULTISTAGE STOCHASTIC PROGRAMS WITH A RANDOM NUMBER OF STAGES: DYNAMIC PROGRAMMING EQUATIONS, SOLUTION METHODS, AND APPLICATION TO PORTFOLIO SELECTION MULTISTAGE STOCHASTIC PROGRAMS WITH A RANDOM NUMBER OF STAGES: DYNAMIC PROGRAMMING EQUATIONS, SOLUTION METHODS, AND APPLICATION TO PORTFOLIO SELECTION Vincent Guigues School of Applied Mathematics, FGV

More information

On Complexity of Multistage Stochastic Programs

On Complexity of Multistage Stochastic Programs On Complexity of Multistage Stochastic Programs Alexander Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205, USA e-mail: ashapiro@isye.gatech.edu

More information

Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion

Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion A.B. Philpott y and V.L. de Matos z October 7, 2011 Abstract We consider the incorporation of a time-consistent coherent

More information

Scenario tree generation for stochastic programming models using GAMS/SCENRED

Scenario tree generation for stochastic programming models using GAMS/SCENRED Scenario tree generation for stochastic programming models using GAMS/SCENRED Holger Heitsch 1 and Steven Dirkse 2 1 Humboldt-University Berlin, Department of Mathematics, Germany 2 GAMS Development Corp.,

More information

Investigation of the and minimum storage energy target levels approach. Final Report

Investigation of the and minimum storage energy target levels approach. Final Report Investigation of the AV@R and minimum storage energy target levels approach Final Report First activity of the technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional

More information

Contents Critique 26. portfolio optimization 32

Contents Critique 26. portfolio optimization 32 Contents Preface vii 1 Financial problems and numerical methods 3 1.1 MATLAB environment 4 1.1.1 Why MATLAB? 5 1.2 Fixed-income securities: analysis and portfolio immunization 6 1.2.1 Basic valuation of

More information

Quasi-Convex Stochastic Dynamic Programming

Quasi-Convex Stochastic Dynamic Programming Quasi-Convex Stochastic Dynamic Programming John R. Birge University of Chicago Booth School of Business JRBirge SIAM FM12, MSP, 10 July 2012 1 General Theme Many dynamic optimization problems dealing

More information

Approximations of Stochastic Programs. Scenario Tree Reduction and Construction

Approximations of Stochastic Programs. Scenario Tree Reduction and Construction Approximations of Stochastic Programs. Scenario Tree Reduction and Construction W. Römisch Humboldt-University Berlin Institute of Mathematics 10099 Berlin, Germany www.mathematik.hu-berlin.de/~romisch

More information

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models IEOR E4707: Foundations of Financial Engineering c 206 by Martin Haugh Martingale Pricing Theory in Discrete-Time and Discrete-Space Models These notes develop the theory of martingale pricing in a discrete-time,

More information

The Irrevocable Multi-Armed Bandit Problem

The Irrevocable Multi-Armed Bandit Problem The Irrevocable Multi-Armed Bandit Problem Ritesh Madan Qualcomm-Flarion Technologies May 27, 2009 Joint work with Vivek Farias (MIT) 2 Multi-Armed Bandit Problem n arms, where each arm i is a Markov Decision

More information

Optimal Security Liquidation Algorithms

Optimal Security Liquidation Algorithms Optimal Security Liquidation Algorithms Sergiy Butenko Department of Industrial Engineering, Texas A&M University, College Station, TX 77843-3131, USA Alexander Golodnikov Glushkov Institute of Cybernetics,

More information

Dynamic Programming: An overview. 1 Preliminaries: The basic principle underlying dynamic programming

Dynamic Programming: An overview. 1 Preliminaries: The basic principle underlying dynamic programming Dynamic Programming: An overview These notes summarize some key properties of the Dynamic Programming principle to optimize a function or cost that depends on an interval or stages. This plays a key role

More information

Asset-Liability Management

Asset-Liability Management Asset-Liability Management John Birge University of Chicago Booth School of Business JRBirge INFORMS San Francisco, Nov. 2014 1 Overview Portfolio optimization involves: Modeling Optimization Estimation

More information

Assessing Policy Quality in Multi-stage Stochastic Programming

Assessing Policy Quality in Multi-stage Stochastic Programming Assessing Policy Quality in Multi-stage Stochastic Programming Anukal Chiralaksanakul and David P. Morton Graduate Program in Operations Research The University of Texas at Austin Austin, TX 78712 January

More information

6.231 DYNAMIC PROGRAMMING LECTURE 3 LECTURE OUTLINE

6.231 DYNAMIC PROGRAMMING LECTURE 3 LECTURE OUTLINE 6.21 DYNAMIC PROGRAMMING LECTURE LECTURE OUTLINE Deterministic finite-state DP problems Backward shortest path algorithm Forward shortest path algorithm Shortest path examples Alternative shortest path

More information

Risk aversion in multi-stage stochastic programming: a modeling and algorithmic perspective

Risk aversion in multi-stage stochastic programming: a modeling and algorithmic perspective Risk aversion in multi-stage stochastic programming: a modeling and algorithmic perspective Tito Homem-de-Mello School of Business Universidad Adolfo Ibañez, Santiago, Chile Joint work with Bernardo Pagnoncelli

More information

Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion

Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion A.B. Philpott y and V.L. de Matos z March 28, 2011 Abstract We consider the incorporation of a time-consistent coherent

More information

Risk Management for Chemical Supply Chain Planning under Uncertainty

Risk Management for Chemical Supply Chain Planning under Uncertainty for Chemical Supply Chain Planning under Uncertainty Fengqi You and Ignacio E. Grossmann Dept. of Chemical Engineering, Carnegie Mellon University John M. Wassick The Dow Chemical Company Introduction

More information

Multistage Stochastic Demand-side Management for Price-Making Major Consumers of Electricity in a Co-optimized Energy and Reserve Market

Multistage Stochastic Demand-side Management for Price-Making Major Consumers of Electricity in a Co-optimized Energy and Reserve Market Multistage Stochastic Demand-side Management for Price-Making Major Consumers of Electricity in a Co-optimized Energy and Reserve Market Mahbubeh Habibian Anthony Downward Golbon Zakeri Abstract In this

More information

Multistage Stochastic Mixed-Integer Programs for Optimizing Gas Contract and Scheduling Maintenance

Multistage Stochastic Mixed-Integer Programs for Optimizing Gas Contract and Scheduling Maintenance Multistage Stochastic Mixed-Integer Programs for Optimizing Gas Contract and Scheduling Maintenance Zhe Liu Siqian Shen September 2, 2012 Abstract In this paper, we present multistage stochastic mixed-integer

More information

Decomposition Methods

Decomposition Methods Decomposition Methods separable problems, complicating variables primal decomposition dual decomposition complicating constraints general decomposition structures Prof. S. Boyd, EE364b, Stanford University

More information

Markov Decision Processes II

Markov Decision Processes II Markov Decision Processes II Daisuke Oyama Topics in Economic Theory December 17, 2014 Review Finite state space S, finite action space A. The value of a policy σ A S : v σ = β t Q t σr σ, t=0 which satisfies

More information

Optimal energy management and stochastic decomposition

Optimal energy management and stochastic decomposition Optimal energy management and stochastic decomposition F. Pacaud P. Carpentier J.P. Chancelier M. De Lara JuMP-dev workshop, 2018 ENPC ParisTech ENSTA ParisTech Efficacity 1/23 Motivation We consider a

More information

Column generation to solve planning problems

Column generation to solve planning problems Column generation to solve planning problems ALGORITMe Han Hoogeveen 1 Continuous Knapsack problem We are given n items with integral weight a j ; integral value c j. B is a given integer. Goal: Find a

More information

High-dimensional risk-constrained dynamic asset allocation via Markov stochastic dual dynamic programming

High-dimensional risk-constrained dynamic asset allocation via Markov stochastic dual dynamic programming High-dimensional risk-constrained dynamic asset allocation via Markov stochastic dual dynamic programming Davi Valladão Industrial Engineering Department, PUC-Rio, davimv@puc-rio.br Thuener Silva Industrial

More information

Lecture Notes 1

Lecture Notes 1 4.45 Lecture Notes Guido Lorenzoni Fall 2009 A portfolio problem To set the stage, consider a simple nite horizon problem. A risk averse agent can invest in two assets: riskless asset (bond) pays gross

More information

Computer Vision Group Prof. Daniel Cremers. 7. Sequential Data

Computer Vision Group Prof. Daniel Cremers. 7. Sequential Data Group Prof. Daniel Cremers 7. Sequential Data Bayes Filter (Rep.) We can describe the overall process using a Dynamic Bayes Network: This incorporates the following Markov assumptions: (measurement) (state)!2

More information

Ellipsoid Method. ellipsoid method. convergence proof. inequality constraints. feasibility problems. Prof. S. Boyd, EE364b, Stanford University

Ellipsoid Method. ellipsoid method. convergence proof. inequality constraints. feasibility problems. Prof. S. Boyd, EE364b, Stanford University Ellipsoid Method ellipsoid method convergence proof inequality constraints feasibility problems Prof. S. Boyd, EE364b, Stanford University Ellipsoid method developed by Shor, Nemirovsky, Yudin in 1970s

More information

Optimal monetary policy when asset markets are incomplete

Optimal monetary policy when asset markets are incomplete Optimal monetary policy when asset markets are incomplete R. Anton Braun Tomoyuki Nakajima 2 University of Tokyo, and CREI 2 Kyoto University, and RIETI December 9, 28 Outline Introduction 2 Model Individuals

More information

Policy iterated lower bounds and linear MC upper bounds for Bermudan style derivatives

Policy iterated lower bounds and linear MC upper bounds for Bermudan style derivatives Finance Winterschool 2007, Lunteren NL Policy iterated lower bounds and linear MC upper bounds for Bermudan style derivatives Pricing complex structured products Mohrenstr 39 10117 Berlin schoenma@wias-berlin.de

More information

Modeling Time-dependent Randomness in Stochastic Dual Dynamic Programming

Modeling Time-dependent Randomness in Stochastic Dual Dynamic Programming Modeling Time-dependent Randomness in Stochastic Dual Dynamic Programming Nils Löhndorf Department of Information Systems and Operations Vienna University of Economics and Business Vienna, Austria nils.loehndorf@wu.ac.at

More information

A Stochastic Approximation Algorithm for Making Pricing Decisions in Network Revenue Management Problems

A Stochastic Approximation Algorithm for Making Pricing Decisions in Network Revenue Management Problems A Stochastic Approximation Algorithm for Making ricing Decisions in Network Revenue Management roblems Sumit Kunnumkal Indian School of Business, Gachibowli, Hyderabad, 500032, India sumit kunnumkal@isb.edu

More information

An Empirical Study of Optimization for Maximizing Diffusion in Networks

An Empirical Study of Optimization for Maximizing Diffusion in Networks An Empirical Study of Optimization for Maximizing Diffusion in Networks Kiyan Ahmadizadeh Bistra Dilkina, Carla P. Gomes, Ashish Sabharwal Cornell University Institute for Computational Sustainability

More information

Stochastic Optimization Methods in Scheduling. Rolf H. Möhring Technische Universität Berlin Combinatorial Optimization and Graph Algorithms

Stochastic Optimization Methods in Scheduling. Rolf H. Möhring Technische Universität Berlin Combinatorial Optimization and Graph Algorithms Stochastic Optimization Methods in Scheduling Rolf H. Möhring Technische Universität Berlin Combinatorial Optimization and Graph Algorithms More expensive and longer... Eurotunnel Unexpected loss of 400,000,000

More information

The Values of Information and Solution in Stochastic Programming

The Values of Information and Solution in Stochastic Programming The Values of Information and Solution in Stochastic Programming John R. Birge The University of Chicago Booth School of Business JRBirge ICSP, Bergamo, July 2013 1 Themes The values of information and

More information

Support Vector Machines: Training with Stochastic Gradient Descent

Support Vector Machines: Training with Stochastic Gradient Descent Support Vector Machines: Training with Stochastic Gradient Descent Machine Learning Spring 2018 The slides are mainly from Vivek Srikumar 1 Support vector machines Training by maximizing margin The SVM

More information

Approximate Composite Minimization: Convergence Rates and Examples

Approximate Composite Minimization: Convergence Rates and Examples ISMP 2018 - Bordeaux Approximate Composite Minimization: Convergence Rates and S. Praneeth Karimireddy, Sebastian U. Stich, Martin Jaggi MLO Lab, EPFL, Switzerland sebastian.stich@epfl.ch July 4, 2018

More information

New Policies for Stochastic Inventory Control Models: Theoretical and Computational Results

New Policies for Stochastic Inventory Control Models: Theoretical and Computational Results OPERATIONS RESEARCH Vol. 00, No. 0, Xxxxx 0000, pp. 000 000 issn 0030-364X eissn 1526-5463 00 0000 0001 INFORMS doi 10.1287/xxxx.0000.0000 c 0000 INFORMS New Policies for Stochastic Inventory Control Models:

More information

From Discrete Time to Continuous Time Modeling

From Discrete Time to Continuous Time Modeling From Discrete Time to Continuous Time Modeling Prof. S. Jaimungal, Department of Statistics, University of Toronto 2004 Arrow-Debreu Securities 2004 Prof. S. Jaimungal 2 Consider a simple one-period economy

More information

Multistage Stochastic Programming

Multistage Stochastic Programming Multistage Stochastic Programming John R. Birge University of Michigan Models - Long and short term - Risk inclusion Approximations - stages and scenarios Computation Slide Number 1 OUTLINE Motivation

More information

Forecast Horizons for Production Planning with Stochastic Demand

Forecast Horizons for Production Planning with Stochastic Demand Forecast Horizons for Production Planning with Stochastic Demand Alfredo Garcia and Robert L. Smith Department of Industrial and Operations Engineering Universityof Michigan, Ann Arbor MI 48109 December

More information

Multistage Stochastic Programming

Multistage Stochastic Programming IE 495 Lecture 21 Multistage Stochastic Programming Prof. Jeff Linderoth April 16, 2003 April 16, 2002 Stochastic Programming Lecture 21 Slide 1 Outline HW Fixes Multistage Stochastic Programming Modeling

More information

Sequential Decision Making

Sequential Decision Making Sequential Decision Making Dynamic programming Christos Dimitrakakis Intelligent Autonomous Systems, IvI, University of Amsterdam, The Netherlands March 18, 2008 Introduction Some examples Dynamic programming

More information

Information Relaxations and Duality in Stochastic Dynamic Programs

Information Relaxations and Duality in Stochastic Dynamic Programs Information Relaxations and Duality in Stochastic Dynamic Programs David Brown, Jim Smith, and Peng Sun Fuqua School of Business Duke University February 28 1/39 Dynamic programming is widely applicable

More information

CPS 270: Artificial Intelligence Markov decision processes, POMDPs

CPS 270: Artificial Intelligence  Markov decision processes, POMDPs CPS 270: Artificial Intelligence http://www.cs.duke.edu/courses/fall08/cps270/ Markov decision processes, POMDPs Instructor: Vincent Conitzer Warmup: a Markov process with rewards We derive some reward

More information

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation IEOR E4703: Monte-Carlo Simulation Other Miscellaneous Topics and Applications of Monte-Carlo Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Allocation of Risk Capital via Intra-Firm Trading

Allocation of Risk Capital via Intra-Firm Trading Allocation of Risk Capital via Intra-Firm Trading Sean Hilden Department of Mathematical Sciences Carnegie Mellon University December 5, 2005 References 1. Artzner, Delbaen, Eber, Heath: Coherent Measures

More information

Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration

Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration Piyush Rai CS5350/6350: Machine Learning November 29, 2011 Reinforcement Learning Supervised Learning: Uses explicit supervision

More information

Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration

Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration Piyush Rai CS5350/6350: Machine Learning November 29, 2011 Reinforcement Learning Supervised Learning: Uses explicit supervision

More information

Shape-Preserving Dynamic Programming

Shape-Preserving Dynamic Programming Shape-Preserving Dynamic Programming Kenneth Judd and Yongyang Cai July 20, 2011 1 Introduction The multi-stage decision-making problems are numerically challenging. When the problems are time-separable,

More information

THE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE

THE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE THE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE GÜNTER ROTE Abstract. A salesperson wants to visit each of n objects that move on a line at given constant speeds in the shortest possible time,

More information

Markov Decision Processes

Markov Decision Processes Markov Decision Processes Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. AIMA 3. Chris Amato Stochastic domains So far, we have studied search Can use

More information

Some simple Bitcoin Economics

Some simple Bitcoin Economics Some simple Bitcoin Economics Linda Schilling 1 and Harald Uhlig 2 1 École Polytechnique - CREST Department of Economics lin.schilling@gmail.com 2 University of Chicago Department of Economics huhlig@uchicago.edu

More information

DM559/DM545 Linear and integer programming

DM559/DM545 Linear and integer programming Department of Mathematics and Computer Science University of Southern Denmark, Odense May 22, 2018 Marco Chiarandini DM559/DM55 Linear and integer programming Sheet, Spring 2018 [pdf format] Contains Solutions!

More information

Stochastic Optimal Control With Dynamic, Time-Consistent Risk Constraints

Stochastic Optimal Control With Dynamic, Time-Consistent Risk Constraints Stochastic Optimal Control With Dynamic, Time-Consistent Risk Constraints Yin-Lam Chow, Marco Pavone risk metric is applied to the future stream of costs; typical examples include variance-constrained

More information

COMP331/557. Chapter 6: Optimisation in Finance: Cash-Flow. (Cornuejols & Tütüncü, Chapter 3)

COMP331/557. Chapter 6: Optimisation in Finance: Cash-Flow. (Cornuejols & Tütüncü, Chapter 3) COMP331/557 Chapter 6: Optimisation in Finance: Cash-Flow (Cornuejols & Tütüncü, Chapter 3) 159 Cash-Flow Management Problem A company has the following net cash flow requirements (in 1000 s of ): Month

More information

Stochastic Programming and Financial Analysis IE447. Midterm Review. Dr. Ted Ralphs

Stochastic Programming and Financial Analysis IE447. Midterm Review. Dr. Ted Ralphs Stochastic Programming and Financial Analysis IE447 Midterm Review Dr. Ted Ralphs IE447 Midterm Review 1 Forming a Mathematical Programming Model The general form of a mathematical programming model is:

More information

Comparison of Static and Dynamic Asset Allocation Models

Comparison of Static and Dynamic Asset Allocation Models Comparison of Static and Dynamic Asset Allocation Models John R. Birge University of Michigan University of Michigan 1 Outline Basic Models Static Markowitz mean-variance Dynamic stochastic programming

More information

Lecture 7: Bayesian approach to MAB - Gittins index

Lecture 7: Bayesian approach to MAB - Gittins index Advanced Topics in Machine Learning and Algorithmic Game Theory Lecture 7: Bayesian approach to MAB - Gittins index Lecturer: Yishay Mansour Scribe: Mariano Schain 7.1 Introduction In the Bayesian approach

More information

Online Appendix: Extensions

Online Appendix: Extensions B Online Appendix: Extensions In this online appendix we demonstrate that many important variations of the exact cost-basis LUL framework remain tractable. In particular, dual problem instances corresponding

More information

6.231 DYNAMIC PROGRAMMING LECTURE 3 LECTURE OUTLINE

6.231 DYNAMIC PROGRAMMING LECTURE 3 LECTURE OUTLINE 6.21 DYNAMIC PROGRAMMING LECTURE LECTURE OUTLINE Deterministic finite-state DP problems Backward shortest path algorithm Forward shortest path algorithm Shortest path examples Alternative shortest path

More information

Interpolation. 1 What is interpolation? 2 Why are we interested in this?

Interpolation. 1 What is interpolation? 2 Why are we interested in this? Interpolation 1 What is interpolation? For a certain function f (x we know only the values y 1 = f (x 1,,y n = f (x n For a point x different from x 1,,x n we would then like to approximate f ( x using

More information

Online Appendix Optimal Time-Consistent Government Debt Maturity D. Debortoli, R. Nunes, P. Yared. A. Proofs

Online Appendix Optimal Time-Consistent Government Debt Maturity D. Debortoli, R. Nunes, P. Yared. A. Proofs Online Appendi Optimal Time-Consistent Government Debt Maturity D. Debortoli, R. Nunes, P. Yared A. Proofs Proof of Proposition 1 The necessity of these conditions is proved in the tet. To prove sufficiency,

More information

CSCI 1951-G Optimization Methods in Finance Part 00: Course Logistics Introduction to Finance Optimization Problems

CSCI 1951-G Optimization Methods in Finance Part 00: Course Logistics Introduction to Finance Optimization Problems CSCI 1951-G Optimization Methods in Finance Part 00: Course Logistics Introduction to Finance Optimization Problems January 26, 2018 1 / 24 Basic information All information is available in the syllabus

More information

17 MAKING COMPLEX DECISIONS

17 MAKING COMPLEX DECISIONS 267 17 MAKING COMPLEX DECISIONS The agent s utility now depends on a sequence of decisions In the following 4 3grid environment the agent makes a decision to move (U, R, D, L) at each time step When the

More information

Markov Decision Processes (MDPs) CS 486/686 Introduction to AI University of Waterloo

Markov Decision Processes (MDPs) CS 486/686 Introduction to AI University of Waterloo Markov Decision Processes (MDPs) CS 486/686 Introduction to AI University of Waterloo Outline Sequential Decision Processes Markov chains Highlight Markov property Discounted rewards Value iteration Markov

More information

Ellipsoid Method. ellipsoid method. convergence proof. inequality constraints. feasibility problems. Prof. S. Boyd, EE392o, Stanford University

Ellipsoid Method. ellipsoid method. convergence proof. inequality constraints. feasibility problems. Prof. S. Boyd, EE392o, Stanford University Ellipsoid Method ellipsoid method convergence proof inequality constraints feasibility problems Prof. S. Boyd, EE392o, Stanford University Challenges in cutting-plane methods can be difficult to compute

More information

Strategies and Nash Equilibrium. A Whirlwind Tour of Game Theory

Strategies and Nash Equilibrium. A Whirlwind Tour of Game Theory Strategies and Nash Equilibrium A Whirlwind Tour of Game Theory (Mostly from Fudenberg & Tirole) Players choose actions, receive rewards based on their own actions and those of the other players. Example,

More information

European Contingent Claims

European Contingent Claims European Contingent Claims Seminar: Financial Modelling in Life Insurance organized by Dr. Nikolic and Dr. Meyhöfer Zhiwen Ning 13.05.2016 Zhiwen Ning European Contingent Claims 13.05.2016 1 / 23 outline

More information

Large-Scale SVM Optimization: Taking a Machine Learning Perspective

Large-Scale SVM Optimization: Taking a Machine Learning Perspective Large-Scale SVM Optimization: Taking a Machine Learning Perspective Shai Shalev-Shwartz Toyota Technological Institute at Chicago Joint work with Nati Srebro Talk at NEC Labs, Princeton, August, 2008 Shai

More information

Optimal investments under dynamic performance critria. Lecture IV

Optimal investments under dynamic performance critria. Lecture IV Optimal investments under dynamic performance critria Lecture IV 1 Utility-based measurement of performance 2 Deterministic environment Utility traits u(x, t) : x wealth and t time Monotonicity u x (x,

More information

Dynamic Replication of Non-Maturing Assets and Liabilities

Dynamic Replication of Non-Maturing Assets and Liabilities Dynamic Replication of Non-Maturing Assets and Liabilities Michael Schürle Institute for Operations Research and Computational Finance, University of St. Gallen, Bodanstr. 6, CH-9000 St. Gallen, Switzerland

More information

Lecture 5 January 30

Lecture 5 January 30 EE 223: Stochastic Estimation and Control Spring 2007 Lecture 5 January 30 Lecturer: Venkat Anantharam Scribe: aryam Kamgarpour 5.1 Secretary Problem The problem set-up is explained in Lecture 4. We review

More information

Note on Oblivious equilibrium

Note on Oblivious equilibrium Note on Oblivious equilibrium November 29, 2016 1 Introduction Ericson and Pakes(1995) introduced a framework for dynamic oligopoly equilibrium with entry and exit, which can be used for policy analysis.

More information

An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity

An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity Coralia Cartis, Nick Gould and Philippe Toint Department of Mathematics,

More information

The Uncertain Volatility Model

The Uncertain Volatility Model The Uncertain Volatility Model Claude Martini, Antoine Jacquier July 14, 008 1 Black-Scholes and realised volatility What happens when a trader uses the Black-Scholes (BS in the sequel) formula to sell

More information

HIGHER ORDER BINARY OPTIONS AND MULTIPLE-EXPIRY EXOTICS

HIGHER ORDER BINARY OPTIONS AND MULTIPLE-EXPIRY EXOTICS Electronic Journal of Mathematical Analysis and Applications Vol. (2) July 203, pp. 247-259. ISSN: 2090-792X (online) http://ejmaa.6te.net/ HIGHER ORDER BINARY OPTIONS AND MULTIPLE-EXPIRY EXOTICS HYONG-CHOL

More information

ROBUST OPTIMIZATION OF MULTI-PERIOD PRODUCTION PLANNING UNDER DEMAND UNCERTAINTY. A. Ben-Tal, B. Golany and M. Rozenblit

ROBUST OPTIMIZATION OF MULTI-PERIOD PRODUCTION PLANNING UNDER DEMAND UNCERTAINTY. A. Ben-Tal, B. Golany and M. Rozenblit ROBUST OPTIMIZATION OF MULTI-PERIOD PRODUCTION PLANNING UNDER DEMAND UNCERTAINTY A. Ben-Tal, B. Golany and M. Rozenblit Faculty of Industrial Engineering and Management, Technion, Haifa 32000, Israel ABSTRACT

More information

Tangent Lévy Models. Sergey Nadtochiy (joint work with René Carmona) Oxford-Man Institute of Quantitative Finance University of Oxford.

Tangent Lévy Models. Sergey Nadtochiy (joint work with René Carmona) Oxford-Man Institute of Quantitative Finance University of Oxford. Tangent Lévy Models Sergey Nadtochiy (joint work with René Carmona) Oxford-Man Institute of Quantitative Finance University of Oxford June 24, 2010 6th World Congress of the Bachelier Finance Society Sergey

More information

ADVANCED MACROECONOMIC TECHNIQUES NOTE 7b

ADVANCED MACROECONOMIC TECHNIQUES NOTE 7b 316-406 ADVANCED MACROECONOMIC TECHNIQUES NOTE 7b Chris Edmond hcpedmond@unimelb.edu.aui Aiyagari s model Arguably the most popular example of a simple incomplete markets model is due to Rao Aiyagari (1994,

More information

Optimally Thresholded Realized Power Variations for Lévy Jump Diffusion Models

Optimally Thresholded Realized Power Variations for Lévy Jump Diffusion Models Optimally Thresholded Realized Power Variations for Lévy Jump Diffusion Models José E. Figueroa-López 1 1 Department of Statistics Purdue University University of Missouri-Kansas City Department of Mathematics

More information

EARLY EXERCISE OPTIONS: UPPER BOUNDS

EARLY EXERCISE OPTIONS: UPPER BOUNDS EARLY EXERCISE OPTIONS: UPPER BOUNDS LEIF B.G. ANDERSEN AND MARK BROADIE Abstract. In this article, we discuss how to generate upper bounds for American or Bermudan securities by Monte Carlo methods. These

More information

Chapter 3: Black-Scholes Equation and Its Numerical Evaluation

Chapter 3: Black-Scholes Equation and Its Numerical Evaluation Chapter 3: Black-Scholes Equation and Its Numerical Evaluation 3.1 Itô Integral 3.1.1 Convergence in the Mean and Stieltjes Integral Definition 3.1 (Convergence in the Mean) A sequence {X n } n ln of random

More information

Optimal Switching Games for Emissions Trading

Optimal Switching Games for Emissions Trading Optimal Switching Games for Emissions Trading Mike Department of Statistics & Applied Probability University of California Santa Barbara MSRI, May 4, 2009 1 / 29 Outline Cap-and-Trade: Producer Perspective

More information

Dynamic Portfolio Choice II

Dynamic Portfolio Choice II Dynamic Portfolio Choice II Dynamic Programming Leonid Kogan MIT, Sloan 15.450, Fall 2010 c Leonid Kogan ( MIT, Sloan ) Dynamic Portfolio Choice II 15.450, Fall 2010 1 / 35 Outline 1 Introduction to Dynamic

More information

POMDPs: Partially Observable Markov Decision Processes Advanced AI

POMDPs: Partially Observable Markov Decision Processes Advanced AI POMDPs: Partially Observable Markov Decision Processes Advanced AI Wolfram Burgard Types of Planning Problems Classical Planning State observable Action Model Deterministic, accurate MDPs observable stochastic

More information