Robust Dual Dynamic Programming

Size: px
Start display at page:

Download "Robust Dual Dynamic Programming"

Transcription

1 1 / 18 Robust Dual Dynamic Programming Angelos Georghiou, Angelos Tsoukalas, Wolfram Wiesemann American University of Beirut Olayan School of Business 31 May 217

2 2 / 18 Inspired by SDDP Stochastic optimization - Optimizes expected value - Needs to know distribution min E P [f (x, ξ)] x Robust optimization - Optimizes for the worst case - Works with uncertainty sets Nested Benders min x max f (x, ξ) ξ Ξ SDDP RDDP

3 3 / 18 minimize max ξ Ξ Multistage Robust Optimization T qt x t (ξ t ) t=1 subject to T t x t 1 (ξ t 1 ) + W t x t (ξ t ) H t ξ t ξ Ξ, t x t (ξ t ) R nt, ξ t Ξ t, ξ t = (ξ 1,, ξ t ) Optimize over decision policies x t ( ). Ξ t polyhedral uncertainty sets. Constraints entangle consecutive stages. Infinite number of variables and constraints.

4 3 / 18 minimize max ξ Ξ Multistage Robust Optimization T qt x t (ξ t ) t=1 ( [ T ]) E qt x t (ξ t ) subject to T t x t 1 (ξ t 1 ) + W t x t (ξ t ) H t ξ t ξ Ξ, t t=1 x t (ξ t ) R nt, ξ t Ξ t, ξ t = (ξ 1,, ξ t ) Optimize over decision policies x t ( ). Ξ t polyhedral uncertainty sets. Constraints entangle consecutive stages. Infinite number of variables and constraints.

5 4 / 18 Relatively complete recourse Assumptions - any partial feasible solution x 1,..., x t can be extended to a complete solution x 1,..., x T - Can be addressed using feasibility cuts Right hand side uncertainty only - T t x t 1 (ξ t 1 ) + W t x t (ξ t ) H t ξ t - Ensures finite convergence - Makes problem easier - Relationship with Nested Benders/SDDP easier to see Both assumptions can be lifted

6 Nested Formulation The multistage problem can be expressed through a nested formulation [ ] min q1 x 1 + x 1 X 1 max min ξ 2 Ξ 2 x 2 X 2 (x 1,ξ 2 ) q 2 x max min ξ T Ξ T x T X T (x T 1,ξ T ) q T x T }{{} First stage problem t stage problem Q 3 (x 2 ) } {{ } Q 2 (x 1 ) min x 1 R n 1 Q t (x t 1 ) = max min ξ t Ξ t x t R n t q 1 x 1 + Q 2 (x 1 ) W 1 x 1 h 1 q t x t + Q t+1 (x t ) T t x t 1 + W t x t H t ξ t 5 / 18

7 6 / 18 Nested Formulation Q 2 Q 3 Q T Q T +1 min x 1 R n 1 q 1 x 1 + Q 2 (x 1 ) W 1 x 1 h 1 Q t (x t 1 ) = max min ξ t Ξ t x t R n t q t x t + Q t+1 (x t ) T t x t 1 + W t x t H t ξ t Optimal value of inner problem convex in ξ t.

8 6 / 18 Nested Formulation Q 2 Q 3 Q T Q T +1 min x 1 R n 1 q 1 x 1 + Q 2 (x 1 ) W 1 x 1 h 1 Q t (x t 1 ) = max min ξ t ext Ξ t x t R n t q t x t + Q t+1 (x t ) T t x t 1 + W t x t H t ξ t Optimal value of inner problem convex in ξ t. We can replace Ξ t with ext Ξ t. Problem decomposes. If only we knew the value functions...

9 Nested Benders Decomposition FP ξ2 ξ3 x3 min x2 R n 2 subject to q 2 x2 + Q 3 (x2) Tt x1 + W2x2 H2ξ2 x2 x3 x1 x2 x3 max ξ2 extξ2 [ min x2 R n 2 subject to q 2 x2 + Q 3 (x2) Tt x1 + W2x2 H2ξ2 ] x3 BP Maintain one (outer approximation of a ) value function per node. Traverse the scenario tree forwards and backwards. FP: At every node, solve and decide where to refine. Move x t forward. We refine at all nodes, i.e., for all scenarios. BP: introduce Benders cuts to refine outer approximations. 7 / 18

10 Nested Benders Decomposition FP ξ2 ξ3 x3 min x2 R n 2 subject to q 2 x2 + Q 3 (x2) Tt x1 + W2x2 H2ξ2 x2 x3 x1 X2(ξ2, x1) x2 x3 max ξ2 extξ2 [ min x2 R n 2 subject to q 2 x2 + Q 3 (x2) Tt x1 + W2x2 H2ξ2 ] x3 BP Maintain one (outer approximation of a ) value function per node. Traverse the scenario tree forwards and backwards. FP: At every node, solve and decide where to refine. Move x t forward. We refine at all nodes, i.e., for all scenarios. BP: introduce Benders cuts to refine outer approximations. But cuts are valid for all nodes of a stage. 7 / 18

11 8 / 18 Towards SDDP: Cut Sharing ξ2 ξ3 FP min x2 R n 2 subject to q 2 x2 + Q 3 (x2) Tt x1 + W2x2 H2 ξ2 x1 x2 x3 X2(ξ2, x1) max ξ2 extξ2 [ min x2 R n 2 subject to q 2 x2 + Q 3 (x2) Tt x1 + W2x2 H2ξ2 ] BP Maintain one approximation per stage. But which scenario to propagate forwards? Where to refine the approximation? Exponential number of end-to-end choices.

12 Towards SDDP: Cut Sharing ξ2 ξ3 FP min x2 R n 2 subject to q 2 x2 + Q 3 (x2) Tt x1 + W2x2 H2 ξ2 x1 x2 x3 X2(ξ2, x1) max ξ2 extξ2 [ min x2 R n 2 subject to q 2 x2 + Q 3 (x2) Tt x1 + W2x2 H2ξ2 ] BP Maintain one approximation per stage. But which scenario to propagate forwards? Where to refine the approximation? Exponential number of end-to-end choices. SDDP Solution (for stochastic programming): pick at random! Small number of refinements. Good performance in practice. No deterministic upper bound/termination criterion. Stochastic convergence. 8 / 18

13 9 / 18 Robust Dual Dynamic Programming Not all scenarios are important. Pick worst case scenarios. Maintain both inner and outer approximations. In the FP: - use inner approximations to choose scenarios. - use outer approximations to choose decisions(points of refinement). In the BP refine both inner and outer approximations.

14 1 / 18 Where to Refine? Forward pass. Minimizing a convex function Maximizing a convex function ξt f = arg max min x f t = ξ t Ξ t x t R n t arg min x t R n t q t x t + Q t+1 (x t ) T t x f t 1 + W tx t H t ξ t q t x t + Q t+1 (x t ) T t x f t 1 + W tx t H t ξ f t

15 1 / 18 Where to Refine? Forward pass. Minimizing a convex function Maximizing a convex function ξt f = arg max min x f t = ξ t Ξ t x t R n t arg min x t R n t q t x t + Q t+1 (x t ) T t x f t 1 + W tx t H t ξ t q t x t + Q t+1 (x t ) T t x f t 1 + W tx t H t ξ f t

16 11 / 18 ξ b t = How to Refine? Backward Pass. arg max min ξ t Ξ t x t R n t qt x t + Q t+1 (x t ) T t xt 1 f + W tx t H t ξ t with corresponding inner solution x b t. Add (x f t 1, q t x b t + Q t+1 (x b t )) to the description of Q t By solving (the dual of) Q t (xt 1 f ) = min x t R n t q t x t + Q t+1 (x t ) T t x f t 1 + W tx t H t ξ b t Q t with a hyperplane at x f t 1. - Subgradients of perturbation functions Lagrange multipliers. refine

17 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x1 f = / 43

18 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 1 = 5.9 Stage 2. Update Region. [ 15., 5.] 17 / 43

19 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 1 = 5.9 Stage 2. Update Region. [ 15., 5.] x f 2 = / 43

20 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 1 = 5.9 Stage 2. Update Region. [ 15., 5.] Stage 3. Update Region. [12., 15.] x f 2 = / 43

21 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 1 = 5.9 Stage 2. Update Region. [ 15., 5.] x f 2 = 5. x f 3 = 12. Stage 3. Update Region. [12., 15.] 2 / 43

22 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 2. Update Region. [ 15., 5.] x f 2 = 5. Stage 3. Update Region. [12., 15.] Stage 2. Update Q 3 (-5.,32.) x f 3 = / 43

23 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 2 = 5. Stage 3. Update Region. [12., 15.] x f 3 = 12. Stage 2. Update Q 3 Stage 2. Update Q 3 4. x x 2 (-5.,32.) 22 / 43

24 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 3. Update Region. [12., 15.] x f 3 = 12. Stage 2. Update Q 3 (-5.,32.) Stage 2. Update Q 3 Stage 1. Update Q 2 (5.9,2.2) 4. x x 2 23 / 43

25 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 3 = 12. Stage 2. Update Q 3 (-5.,32.) Stage 2. Update Q 3 4. x x 2 Stage 1. Update Q 2 Stage 1. Update Q 2 2. x x 1 (5.9,2.2) 24 / 43

26 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 2. Update Q 3 (-5.,32.) Stage 2. Update Q 3 4. x x 2 Stage 1. Update Q2 (5.9,2.2) x f 1 = 1. Stage 1. Update Q 2 2. x x 1 25 / 43

27 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 2. Update Q 3 4. x x 2 Stage 1. Update Q2 (5.9,2.2) Stage 1. Update Q 2 2. x x 1 Stage 2. Update Region. [ 15., 1.9] x f 1 = / 43

28 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 1. Update Q 2 (5.9,2.2) Stage 1. Update Q 2 2. x x 1 x f 1 = 1. x f 2 = 1.9 Stage 2. Update Region. [ 15., 1.9] 27 / 43

29 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 1. Update Q 2 2. x x 1 x f 1 = 1. Stage 2. Update Region. [ 15., 1.9] Stage 3. Update Region. [59., 15.] x f 2 = / 43

30 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 1 = 1. Stage 2. Update Region. [ 15., 1.9] x f 2 = 1.9 x f 3 = 59. Stage 3. Update Region. [59., 15.] 29 / 43

31 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 2. Update Region. [ 15., 1.9] x f 2 = 1.9 Stage 3. Update Region. [59., 15.] Stage 2. Update Q 3 (1.9,15.4) x f 3 = / 43

32 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 2 = 1.9 Stage 3. Update Region. [59., 15.] x f 3 = 59. Stage 2. Update Q 3 Stage 2. Update Q 3 4. x x 2 (1.9,15.4) 31 / 43

33 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 3. Update Region. [59., 15.] x f 3 = 59. Stage 2. Update Q 3 (1.9,15.4) Stage 2. Update Q 3 Stage 1. Update Q 2 (-1.,35.4) 4. x x 2 32 / 43

34 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 3 = 59. Stage 2. Update Q 3 (1.9,15.4) Stage 2. Update Q 3 4. x x 2 Stage 1. Update Q 2 Stage 1. Update Q 2 2. x x 1 (-1.,35.4) 33 / 43

35 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 2. Update Q 3 (1.9,15.4) Stage 2. Update Q 3 4. x x 2 Stage 1. Update Q2 (-1.,35.4) x f 1 = 1.1 Stage 1. Update Q 2 2. x x 1 34 / 43

36 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 2. Update Q 3 4. x x 2 Stage 1. Update Q2 (-1.,35.4) Stage 1. Update Q 2 2. x x 1 Stage 2. Update Region. [ 15., 2.] x f 1 = / 43

37 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 1. Update Q 2 (-1.,35.4) Stage 1. Update Q 2 2. x x 1 x f 1 = 1.1 x f 2 = 2. Stage 2. Update Region. [ 15., 2.] 36 / 43

38 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 1. Update Q 2 2. x x 1 x f 1 = 1.1 Stage 2. Update Region. [ 15., 2.] Stage 3. Update Region. [2., 15.] x f 2 = / 43

39 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 1 = 1.1 Stage 2. Update Region. [ 15., 2.] x f 2 = 2. x f 3 = 2. Stage 3. Update Region. [2., 15.] 38 / 43

40 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 2. Update Region. [ 15., 2.] x f 2 = 2. Stage 3. Update Region. [2., 15.] Stage 2. Update Q 3 (2.,-6.) x f 3 = / 43

41 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 2 = 2. Stage 3. Update Region. [2., 15.] x f 3 = 2. Stage 2. Update Q 3 Stage 2. Update Q 3 4. x x 2 (2.,-6.) 4 / 43

42 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x Stage 3. Update Region. [2., 15.] x f 3 = 2. Stage 2. Update Q 3 (2.,-6.) Stage 2. Update Q 3 Stage 1. Update Q 2 (-1.1,-3.8) 4. x x 2 41 / 43

43 2. x 1 + Q 2 (x 1 ) 4. x 2 + Q 3 (x 2 ) 1. x x f 3 = 2. Stage 2. Update Q 3 (2.,-6.) Stage 2. Update Q 3 4. x x 2 Stage 1. Update Q 2 Stage 1. Update Q 2 2. x x 1 (-1.1,-3.8) 42 / 43

44 Numerical Results: Inventory Control 13 / 18

45 Numerical Results Scalability w.r.t. horizon T={1, 5, 1} - 5 products - 4 random variables per stage (24 = 16 scenarios) optimization time (secs) optimization time (secs) relative distance % relative distance % relative distance % , 1,5 2, optimization time (secs) RDDP scales better than LDR w.r.t. the horizon 14 / 18

46 15 / 18 Numerical Results Scalability w.r.t. products ={1, 15, 2} - horizon T=1-4 random variables per stage (2 4 = 16 scenarios) relative distance % 5 5 relative distance % 5 5 relative distance % optimization time (secs) , optimization time (secs) 1 2, 4, 6, 8, optimization time (secs) RDDP does not solve the curse of dimensionality But, can address problems of practical importance

47 16 / 18 Scalability w.r.t. random variables ={5, 7, 9} - i.e., scenarios per stage= {32, 128, 512} - products= {6, 8, 1} - horizon T=1 Numerical Results relative distance % 5 5 relative distance % 5 5 relative distance % optimization time (secs) optimization time (secs) 1 4, 8, 12, optimization time (secs)

48 17 / 18 Current Work Extension to Stochastic Programming - Same inner approximation, different algorithm - Same deterministic convergence guarantees - Preliminary results indicate comparable complexity Robust Optimization Stochastic Optimization

49 18 / 18 RDDP Summary Converges to optimal solution. Implementable strategy at every iteration(ub). Lower bound available at every iteration. Finite convergence for RHS/ technology matrix uncertainty Deterministic asymptotic convergence for Recourse matrix/objective uncertainty. 2-Stage sub-problems hard. State of the art multistage problems small 2-Stage problems. Stochastic optimization/ distributionally robust optimization

Stochastic Dual Dynamic Programming Algorithm for Multistage Stochastic Programming

Stochastic Dual Dynamic Programming Algorithm for Multistage Stochastic Programming Stochastic Dual Dynamic Programg Algorithm for Multistage Stochastic Programg Final presentation ISyE 8813 Fall 2011 Guido Lagos Wajdi Tekaya Georgia Institute of Technology November 30, 2011 Multistage

More information

Worst-case-expectation approach to optimization under uncertainty

Worst-case-expectation approach to optimization under uncertainty Worst-case-expectation approach to optimization under uncertainty Wajdi Tekaya Joint research with Alexander Shapiro, Murilo Pereira Soares and Joari Paulo da Costa : Cambridge Systems Associates; : Georgia

More information

Multistage risk-averse asset allocation with transaction costs

Multistage risk-averse asset allocation with transaction costs Multistage risk-averse asset allocation with transaction costs 1 Introduction Václav Kozmík 1 Abstract. This paper deals with asset allocation problems formulated as multistage stochastic programming models.

More information

SOLVING ROBUST SUPPLY CHAIN PROBLEMS

SOLVING ROBUST SUPPLY CHAIN PROBLEMS SOLVING ROBUST SUPPLY CHAIN PROBLEMS Daniel Bienstock Nuri Sercan Özbay Columbia University, New York November 13, 2005 Project with Lucent Technologies Optimize the inventory buffer levels in a complicated

More information

On solving multistage stochastic programs with coherent risk measures

On solving multistage stochastic programs with coherent risk measures On solving multistage stochastic programs with coherent risk measures Andy Philpott Vitor de Matos y Erlon Finardi z August 13, 2012 Abstract We consider a class of multistage stochastic linear programs

More information

Asset-Liability Management

Asset-Liability Management Asset-Liability Management John Birge University of Chicago Booth School of Business JRBirge INFORMS San Francisco, Nov. 2014 1 Overview Portfolio optimization involves: Modeling Optimization Estimation

More information

Report for technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional do Sistema Elétrico Risk Averse Approach

Report for technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional do Sistema Elétrico Risk Averse Approach Report for technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional do Sistema Elétrico Risk Averse Approach Alexander Shapiro and Wajdi Tekaya School of Industrial and

More information

Risk aversion in multi-stage stochastic programming: a modeling and algorithmic perspective

Risk aversion in multi-stage stochastic programming: a modeling and algorithmic perspective Risk aversion in multi-stage stochastic programming: a modeling and algorithmic perspective Tito Homem-de-Mello School of Business Universidad Adolfo Ibañez, Santiago, Chile Joint work with Bernardo Pagnoncelli

More information

Stochastic Dual Dynamic Programming

Stochastic Dual Dynamic Programming 1 / 43 Stochastic Dual Dynamic Programming Operations Research Anthony Papavasiliou 2 / 43 Contents [ 10.4 of BL], [Pereira, 1991] 1 Recalling the Nested L-Shaped Decomposition 2 Drawbacks of Nested Decomposition

More information

Stochastic Dual Dynamic integer Programming

Stochastic Dual Dynamic integer Programming Stochastic Dual Dynamic integer Programming Shabbir Ahmed Georgia Tech Jikai Zou Andy Sun Multistage IP Canonical deterministic formulation ( X T ) f t (x t,y t ):(x t 1,x t,y t ) 2 X t 8 t x t min x,y

More information

DASC: A DECOMPOSITION ALGORITHM FOR MULTISTAGE STOCHASTIC PROGRAMS WITH STRONGLY CONVEX COST FUNCTIONS

DASC: A DECOMPOSITION ALGORITHM FOR MULTISTAGE STOCHASTIC PROGRAMS WITH STRONGLY CONVEX COST FUNCTIONS DASC: A DECOMPOSITION ALGORITHM FOR MULTISTAGE STOCHASTIC PROGRAMS WITH STRONGLY CONVEX COST FUNCTIONS Vincent Guigues School of Applied Mathematics, FGV Praia de Botafogo, Rio de Janeiro, Brazil vguigues@fgv.br

More information

Optimal energy management and stochastic decomposition

Optimal energy management and stochastic decomposition Optimal energy management and stochastic decomposition F. Pacaud P. Carpentier J.P. Chancelier M. De Lara JuMP-dev workshop, 2018 ENPC ParisTech ENSTA ParisTech Efficacity 1/23 Motivation We consider a

More information

Multistage Stochastic Programming

Multistage Stochastic Programming IE 495 Lecture 21 Multistage Stochastic Programming Prof. Jeff Linderoth April 16, 2003 April 16, 2002 Stochastic Programming Lecture 21 Slide 1 Outline HW Fixes Multistage Stochastic Programming Modeling

More information

Approximations of Stochastic Programs. Scenario Tree Reduction and Construction

Approximations of Stochastic Programs. Scenario Tree Reduction and Construction Approximations of Stochastic Programs. Scenario Tree Reduction and Construction W. Römisch Humboldt-University Berlin Institute of Mathematics 10099 Berlin, Germany www.mathematik.hu-berlin.de/~romisch

More information

Multistage Stochastic Programming

Multistage Stochastic Programming Multistage Stochastic Programming John R. Birge University of Michigan Models - Long and short term - Risk inclusion Approximations - stages and scenarios Computation Slide Number 1 OUTLINE Motivation

More information

Scenario reduction and scenario tree construction for power management problems

Scenario reduction and scenario tree construction for power management problems Scenario reduction and scenario tree construction for power management problems N. Gröwe-Kuska, H. Heitsch and W. Römisch Humboldt-University Berlin Institute of Mathematics Page 1 of 20 IEEE Bologna POWER

More information

Scenario tree generation for stochastic programming models using GAMS/SCENRED

Scenario tree generation for stochastic programming models using GAMS/SCENRED Scenario tree generation for stochastic programming models using GAMS/SCENRED Holger Heitsch 1 and Steven Dirkse 2 1 Humboldt-University Berlin, Department of Mathematics, Germany 2 GAMS Development Corp.,

More information

Multistage Stochastic Mixed-Integer Programs for Optimizing Gas Contract and Scheduling Maintenance

Multistage Stochastic Mixed-Integer Programs for Optimizing Gas Contract and Scheduling Maintenance Multistage Stochastic Mixed-Integer Programs for Optimizing Gas Contract and Scheduling Maintenance Zhe Liu Siqian Shen September 2, 2012 Abstract In this paper, we present multistage stochastic mixed-integer

More information

Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion

Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion A.B. Philpott y and V.L. de Matos z October 7, 2011 Abstract We consider the incorporation of a time-consistent coherent

More information

A Multi-Stage Stochastic Programming Model for Managing Risk-Optimal Electricity Portfolios. Stochastic Programming and Electricity Risk Management

A Multi-Stage Stochastic Programming Model for Managing Risk-Optimal Electricity Portfolios. Stochastic Programming and Electricity Risk Management A Multi-Stage Stochastic Programming Model for Managing Risk-Optimal Electricity Portfolios SLIDE 1 Outline Multi-stage stochastic programming modeling Setting - Electricity portfolio management Electricity

More information

Multistage Stochastic Demand-side Management for Price-Making Major Consumers of Electricity in a Co-optimized Energy and Reserve Market

Multistage Stochastic Demand-side Management for Price-Making Major Consumers of Electricity in a Co-optimized Energy and Reserve Market Multistage Stochastic Demand-side Management for Price-Making Major Consumers of Electricity in a Co-optimized Energy and Reserve Market Mahbubeh Habibian Anthony Downward Golbon Zakeri Abstract In this

More information

Essays on Some Combinatorial Optimization Problems with Interval Data

Essays on Some Combinatorial Optimization Problems with Interval Data Essays on Some Combinatorial Optimization Problems with Interval Data a thesis submitted to the department of industrial engineering and the institute of engineering and sciences of bilkent university

More information

Approximation of Continuous-State Scenario Processes in Multi-Stage Stochastic Optimization and its Applications

Approximation of Continuous-State Scenario Processes in Multi-Stage Stochastic Optimization and its Applications Approximation of Continuous-State Scenario Processes in Multi-Stage Stochastic Optimization and its Applications Anna Timonina University of Vienna, Abraham Wald PhD Program in Statistics and Operations

More information

Dynamic Programming: An overview. 1 Preliminaries: The basic principle underlying dynamic programming

Dynamic Programming: An overview. 1 Preliminaries: The basic principle underlying dynamic programming Dynamic Programming: An overview These notes summarize some key properties of the Dynamic Programming principle to optimize a function or cost that depends on an interval or stages. This plays a key role

More information

Integer Programming Models

Integer Programming Models Integer Programming Models Fabio Furini December 10, 2014 Integer Programming Models 1 Outline 1 Combinatorial Auctions 2 The Lockbox Problem 3 Constructing an Index Fund Integer Programming Models 2 Integer

More information

Continuous-time Stochastic Control and Optimization with Financial Applications

Continuous-time Stochastic Control and Optimization with Financial Applications Huyen Pham Continuous-time Stochastic Control and Optimization with Financial Applications 4y Springer Some elements of stochastic analysis 1 1.1 Stochastic processes 1 1.1.1 Filtration and processes 1

More information

Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion

Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion A.B. Philpott y and V.L. de Matos z March 28, 2011 Abstract We consider the incorporation of a time-consistent coherent

More information

Scenario Generation and Sampling Methods

Scenario Generation and Sampling Methods Scenario Generation and Sampling Methods Güzin Bayraksan Tito Homem-de-Mello SVAN 2016 IMPA May 9th, 2016 Bayraksan (OSU) & Homem-de-Mello (UAI) Scenario Generation and Sampling SVAN IMPA May 9 1 / 30

More information

Lecture outline W.B.Powell 1

Lecture outline W.B.Powell 1 Lecture outline What is a policy? Policy function approximations (PFAs) Cost function approximations (CFAs) alue function approximations (FAs) Lookahead policies Finding good policies Optimizing continuous

More information

IE 495 Lecture 11. The LShaped Method. Prof. Jeff Linderoth. February 19, February 19, 2003 Stochastic Programming Lecture 11 Slide 1

IE 495 Lecture 11. The LShaped Method. Prof. Jeff Linderoth. February 19, February 19, 2003 Stochastic Programming Lecture 11 Slide 1 IE 495 Lecture 11 The LShaped Method Prof. Jeff Linderoth February 19, 2003 February 19, 2003 Stochastic Programming Lecture 11 Slide 1 Before We Begin HW#2 $300 $0 http://www.unizh.ch/ior/pages/deutsch/mitglieder/kall/bib/ka-wal-94.pdf

More information

Investigation of the and minimum storage energy target levels approach. Final Report

Investigation of the and minimum storage energy target levels approach. Final Report Investigation of the AV@R and minimum storage energy target levels approach Final Report First activity of the technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional

More information

MULTISTAGE STOCHASTIC PROGRAMS WITH A RANDOM NUMBER OF STAGES: DYNAMIC PROGRAMMING EQUATIONS, SOLUTION METHODS, AND APPLICATION TO PORTFOLIO SELECTION

MULTISTAGE STOCHASTIC PROGRAMS WITH A RANDOM NUMBER OF STAGES: DYNAMIC PROGRAMMING EQUATIONS, SOLUTION METHODS, AND APPLICATION TO PORTFOLIO SELECTION MULTISTAGE STOCHASTIC PROGRAMS WITH A RANDOM NUMBER OF STAGES: DYNAMIC PROGRAMMING EQUATIONS, SOLUTION METHODS, AND APPLICATION TO PORTFOLIO SELECTION Vincent Guigues School of Applied Mathematics, FGV

More information

CS 188: Artificial Intelligence

CS 188: Artificial Intelligence CS 188: Artificial Intelligence Markov Decision Processes Dan Klein, Pieter Abbeel University of California, Berkeley Non Deterministic Search Example: Grid World A maze like problem The agent lives in

More information

Support Vector Machines: Training with Stochastic Gradient Descent

Support Vector Machines: Training with Stochastic Gradient Descent Support Vector Machines: Training with Stochastic Gradient Descent Machine Learning Spring 2018 The slides are mainly from Vivek Srikumar 1 Support vector machines Training by maximizing margin The SVM

More information

A Robust Option Pricing Problem

A Robust Option Pricing Problem IMA 2003 Workshop, March 12-19, 2003 A Robust Option Pricing Problem Laurent El Ghaoui Department of EECS, UC Berkeley 3 Robust optimization standard form: min x sup u U f 0 (x, u) : u U, f i (x, u) 0,

More information

Financial Optimization ISE 347/447. Lecture 15. Dr. Ted Ralphs

Financial Optimization ISE 347/447. Lecture 15. Dr. Ted Ralphs Financial Optimization ISE 347/447 Lecture 15 Dr. Ted Ralphs ISE 347/447 Lecture 15 1 Reading for This Lecture C&T Chapter 12 ISE 347/447 Lecture 15 2 Stock Market Indices A stock market index is a statistic

More information

Optimal liquidation with market parameter shift: a forward approach

Optimal liquidation with market parameter shift: a forward approach Optimal liquidation with market parameter shift: a forward approach (with S. Nadtochiy and T. Zariphopoulou) Haoran Wang Ph.D. candidate University of Texas at Austin ICERM June, 2017 Problem Setup and

More information

Stochastic Optimization Methods in Scheduling. Rolf H. Möhring Technische Universität Berlin Combinatorial Optimization and Graph Algorithms

Stochastic Optimization Methods in Scheduling. Rolf H. Möhring Technische Universität Berlin Combinatorial Optimization and Graph Algorithms Stochastic Optimization Methods in Scheduling Rolf H. Möhring Technische Universität Berlin Combinatorial Optimization and Graph Algorithms More expensive and longer... Eurotunnel Unexpected loss of 400,000,000

More information

An Empirical Study of Optimization for Maximizing Diffusion in Networks

An Empirical Study of Optimization for Maximizing Diffusion in Networks An Empirical Study of Optimization for Maximizing Diffusion in Networks Kiyan Ahmadizadeh Bistra Dilkina, Carla P. Gomes, Ashish Sabharwal Cornell University Institute for Computational Sustainability

More information

Making Decisions. CS 3793 Artificial Intelligence Making Decisions 1

Making Decisions. CS 3793 Artificial Intelligence Making Decisions 1 Making Decisions CS 3793 Artificial Intelligence Making Decisions 1 Planning under uncertainty should address: The world is nondeterministic. Actions are not certain to succeed. Many events are outside

More information

Building Consistent Risk Measures into Stochastic Optimization Models

Building Consistent Risk Measures into Stochastic Optimization Models Building Consistent Risk Measures into Stochastic Optimization Models John R. Birge The University of Chicago Graduate School of Business www.chicagogsb.edu/fac/john.birge JRBirge Fuqua School, Duke University

More information

Dynamic Replication of Non-Maturing Assets and Liabilities

Dynamic Replication of Non-Maturing Assets and Liabilities Dynamic Replication of Non-Maturing Assets and Liabilities Michael Schürle Institute for Operations Research and Computational Finance, University of St. Gallen, Bodanstr. 6, CH-9000 St. Gallen, Switzerland

More information

Robust Optimization Applied to a Currency Portfolio

Robust Optimization Applied to a Currency Portfolio Robust Optimization Applied to a Currency Portfolio R. Fonseca, S. Zymler, W. Wiesemann, B. Rustem Workshop on Numerical Methods and Optimization in Finance June, 2009 OUTLINE Introduction Motivation &

More information

6.231 DYNAMIC PROGRAMMING LECTURE 5 LECTURE OUTLINE

6.231 DYNAMIC PROGRAMMING LECTURE 5 LECTURE OUTLINE 6.231 DYNAMIC PROGRAMMING LECTURE 5 LECTURE OUTLINE Stopping problems Scheduling problems Minimax Control 1 PURE STOPPING PROBLEMS Two possible controls: Stop (incur a one-time stopping cost, and move

More information

Dynamic Portfolio Choice II

Dynamic Portfolio Choice II Dynamic Portfolio Choice II Dynamic Programming Leonid Kogan MIT, Sloan 15.450, Fall 2010 c Leonid Kogan ( MIT, Sloan ) Dynamic Portfolio Choice II 15.450, Fall 2010 1 / 35 Outline 1 Introduction to Dynamic

More information

Non-Deterministic Search

Non-Deterministic Search Non-Deterministic Search MDP s 1 Non-Deterministic Search How do you plan (search) when your actions might fail? In general case, how do you plan, when the actions have multiple possible outcomes? 2 Example:

More information

Journal of Computational and Applied Mathematics. The mean-absolute deviation portfolio selection problem with interval-valued returns

Journal of Computational and Applied Mathematics. The mean-absolute deviation portfolio selection problem with interval-valued returns Journal of Computational and Applied Mathematics 235 (2011) 4149 4157 Contents lists available at ScienceDirect Journal of Computational and Applied Mathematics journal homepage: www.elsevier.com/locate/cam

More information

Decomposition Methods

Decomposition Methods Decomposition Methods separable problems, complicating variables primal decomposition dual decomposition complicating constraints general decomposition structures Prof. S. Boyd, EE364b, Stanford University

More information

Contents Critique 26. portfolio optimization 32

Contents Critique 26. portfolio optimization 32 Contents Preface vii 1 Financial problems and numerical methods 3 1.1 MATLAB environment 4 1.1.1 Why MATLAB? 5 1.2 Fixed-income securities: analysis and portfolio immunization 6 1.2.1 Basic valuation of

More information

CSEP 573: Artificial Intelligence

CSEP 573: Artificial Intelligence CSEP 573: Artificial Intelligence Markov Decision Processes (MDP)! Ali Farhadi Many slides over the course adapted from Luke Zettlemoyer, Dan Klein, Pieter Abbeel, Stuart Russell or Andrew Moore 1 Outline

More information

Action Selection for MDPs: Anytime AO* vs. UCT

Action Selection for MDPs: Anytime AO* vs. UCT Action Selection for MDPs: Anytime AO* vs. UCT Blai Bonet 1 and Hector Geffner 2 1 Universidad Simón Boĺıvar 2 ICREA & Universitat Pompeu Fabra AAAI, Toronto, Canada, July 2012 Online MDP Planning and

More information

Approximate Composite Minimization: Convergence Rates and Examples

Approximate Composite Minimization: Convergence Rates and Examples ISMP 2018 - Bordeaux Approximate Composite Minimization: Convergence Rates and S. Praneeth Karimireddy, Sebastian U. Stich, Martin Jaggi MLO Lab, EPFL, Switzerland sebastian.stich@epfl.ch July 4, 2018

More information

Assessing Policy Quality in Multi-stage Stochastic Programming

Assessing Policy Quality in Multi-stage Stochastic Programming Assessing Policy Quality in Multi-stage Stochastic Programming Anukal Chiralaksanakul and David P. Morton Graduate Program in Operations Research The University of Texas at Austin Austin, TX 78712 January

More information

Medium-Term Planning in Deregulated Energy Markets with Decision Rules

Medium-Term Planning in Deregulated Energy Markets with Decision Rules Imperial College London Department of Computing Medium-Term Planning in Deregulated Energy Markets with Decision Rules Paula Cristina Martins da Silva Rocha Submitted in part fullment of the requirements

More information

Sequential Decision Making

Sequential Decision Making Sequential Decision Making Dynamic programming Christos Dimitrakakis Intelligent Autonomous Systems, IvI, University of Amsterdam, The Netherlands March 18, 2008 Introduction Some examples Dynamic programming

More information

EARLY EXERCISE OPTIONS: UPPER BOUNDS

EARLY EXERCISE OPTIONS: UPPER BOUNDS EARLY EXERCISE OPTIONS: UPPER BOUNDS LEIF B.G. ANDERSEN AND MARK BROADIE Abstract. In this article, we discuss how to generate upper bounds for American or Bermudan securities by Monte Carlo methods. These

More information

Forecast Horizons for Production Planning with Stochastic Demand

Forecast Horizons for Production Planning with Stochastic Demand Forecast Horizons for Production Planning with Stochastic Demand Alfredo Garcia and Robert L. Smith Department of Industrial and Operations Engineering Universityof Michigan, Ann Arbor MI 48109 December

More information

Optimal Security Liquidation Algorithms

Optimal Security Liquidation Algorithms Optimal Security Liquidation Algorithms Sergiy Butenko Department of Industrial Engineering, Texas A&M University, College Station, TX 77843-3131, USA Alexander Golodnikov Glushkov Institute of Cybernetics,

More information

IEOR E4004: Introduction to OR: Deterministic Models

IEOR E4004: Introduction to OR: Deterministic Models IEOR E4004: Introduction to OR: Deterministic Models 1 Dynamic Programming Following is a summary of the problems we discussed in class. (We do not include the discussion on the container problem or the

More information

Risk Management for Chemical Supply Chain Planning under Uncertainty

Risk Management for Chemical Supply Chain Planning under Uncertainty for Chemical Supply Chain Planning under Uncertainty Fengqi You and Ignacio E. Grossmann Dept. of Chemical Engineering, Carnegie Mellon University John M. Wassick The Dow Chemical Company Introduction

More information

EE365: Risk Averse Control

EE365: Risk Averse Control EE365: Risk Averse Control Risk averse optimization Exponential risk aversion Risk averse control 1 Outline Risk averse optimization Exponential risk aversion Risk averse control Risk averse optimization

More information

Stochastic Programming in Gas Storage and Gas Portfolio Management. ÖGOR-Workshop, September 23rd, 2010 Dr. Georg Ostermaier

Stochastic Programming in Gas Storage and Gas Portfolio Management. ÖGOR-Workshop, September 23rd, 2010 Dr. Georg Ostermaier Stochastic Programming in Gas Storage and Gas Portfolio Management ÖGOR-Workshop, September 23rd, 2010 Dr. Georg Ostermaier Agenda Optimization tasks in gas storage and gas portfolio management Scenario

More information

An introduction on game theory for wireless networking [1]

An introduction on game theory for wireless networking [1] An introduction on game theory for wireless networking [1] Ning Zhang 14 May, 2012 [1] Game Theory in Wireless Networks: A Tutorial 1 Roadmap 1 Introduction 2 Static games 3 Extensive-form games 4 Summary

More information

Bounding Optimal Expected Revenues for Assortment Optimization under Mixtures of Multinomial Logits

Bounding Optimal Expected Revenues for Assortment Optimization under Mixtures of Multinomial Logits Bounding Optimal Expected Revenues for Assortment Optimization under Mixtures of Multinomial Logits Jacob Feldman School of Operations Research and Information Engineering, Cornell University, Ithaca,

More information

CSCI 1951-G Optimization Methods in Finance Part 07: Portfolio Optimization

CSCI 1951-G Optimization Methods in Finance Part 07: Portfolio Optimization CSCI 1951-G Optimization Methods in Finance Part 07: Portfolio Optimization March 9 16, 2018 1 / 19 The portfolio optimization problem How to best allocate our money to n risky assets S 1,..., S n with

More information

Stochastic Approximation Algorithms and Applications

Stochastic Approximation Algorithms and Applications Harold J. Kushner G. George Yin Stochastic Approximation Algorithms and Applications With 24 Figures Springer Contents Preface and Introduction xiii 1 Introduction: Applications and Issues 1 1.0 Outline

More information

Spot and forward dynamic utilities. and their associated pricing systems. Thaleia Zariphopoulou. UT, Austin

Spot and forward dynamic utilities. and their associated pricing systems. Thaleia Zariphopoulou. UT, Austin Spot and forward dynamic utilities and their associated pricing systems Thaleia Zariphopoulou UT, Austin 1 Joint work with Marek Musiela (BNP Paribas, London) References A valuation algorithm for indifference

More information

Economics 2010c: Lecture 4 Precautionary Savings and Liquidity Constraints

Economics 2010c: Lecture 4 Precautionary Savings and Liquidity Constraints Economics 2010c: Lecture 4 Precautionary Savings and Liquidity Constraints David Laibson 9/11/2014 Outline: 1. Precautionary savings motives 2. Liquidity constraints 3. Application: Numerical solution

More information

DM559/DM545 Linear and integer programming

DM559/DM545 Linear and integer programming Department of Mathematics and Computer Science University of Southern Denmark, Odense May 22, 2018 Marco Chiarandini DM559/DM55 Linear and integer programming Sheet, Spring 2018 [pdf format] Contains Solutions!

More information

Reinforcement Learning and Simulation-Based Search

Reinforcement Learning and Simulation-Based Search Reinforcement Learning and Simulation-Based Search David Silver Outline 1 Reinforcement Learning 2 3 Planning Under Uncertainty Reinforcement Learning Markov Decision Process Definition A Markov Decision

More information

Progressive Hedging for Multi-stage Stochastic Optimization Problems

Progressive Hedging for Multi-stage Stochastic Optimization Problems Progressive Hedging for Multi-stage Stochastic Optimization Problems David L. Woodruff Jean-Paul Watson Graduate School of Management University of California, Davis Davis, CA 95616, USA dlwoodruff@ucdavis.edu

More information

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models IEOR E4707: Foundations of Financial Engineering c 206 by Martin Haugh Martingale Pricing Theory in Discrete-Time and Discrete-Space Models These notes develop the theory of martingale pricing in a discrete-time,

More information

ROBUST OPTIMIZATION OF MULTI-PERIOD PRODUCTION PLANNING UNDER DEMAND UNCERTAINTY. A. Ben-Tal, B. Golany and M. Rozenblit

ROBUST OPTIMIZATION OF MULTI-PERIOD PRODUCTION PLANNING UNDER DEMAND UNCERTAINTY. A. Ben-Tal, B. Golany and M. Rozenblit ROBUST OPTIMIZATION OF MULTI-PERIOD PRODUCTION PLANNING UNDER DEMAND UNCERTAINTY A. Ben-Tal, B. Golany and M. Rozenblit Faculty of Industrial Engineering and Management, Technion, Haifa 32000, Israel ABSTRACT

More information

Dynamic Risk Management in Electricity Portfolio Optimization via Polyhedral Risk Functionals

Dynamic Risk Management in Electricity Portfolio Optimization via Polyhedral Risk Functionals Dynamic Risk Management in Electricity Portfolio Optimization via Polyhedral Risk Functionals A. Eichhorn and W. Römisch Humboldt-University Berlin, Department of Mathematics, Germany http://www.math.hu-berlin.de/~romisch

More information

CS 188: Artificial Intelligence

CS 188: Artificial Intelligence CS 188: Artificial Intelligence Markov Decision Processes Dan Klein, Pieter Abbeel University of California, Berkeley Non-Deterministic Search 1 Example: Grid World A maze-like problem The agent lives

More information

Provably Near-Optimal Balancing Policies for Multi-Echelon Stochastic Inventory Control Models

Provably Near-Optimal Balancing Policies for Multi-Echelon Stochastic Inventory Control Models Provably Near-Optimal Balancing Policies for Multi-Echelon Stochastic Inventory Control Models Retsef Levi Robin Roundy Van Anh Truong February 13, 2006 Abstract We develop the first algorithmic approach

More information

EE266 Homework 5 Solutions

EE266 Homework 5 Solutions EE, Spring 15-1 Professor S. Lall EE Homework 5 Solutions 1. A refined inventory model. In this problem we consider an inventory model that is more refined than the one you ve seen in the lectures. The

More information

Multi-armed bandit problems

Multi-armed bandit problems Multi-armed bandit problems Stochastic Decision Theory (2WB12) Arnoud den Boer 13 March 2013 Set-up 13 and 14 March: Lectures. 20 and 21 March: Paper presentations (Four groups, 45 min per group). Before

More information

Behavioral pricing of energy swing options by stochastic bilevel optimization

Behavioral pricing of energy swing options by stochastic bilevel optimization Energy Syst (2016) 7:637 662 DOI 10.1007/s12667-016-0190-z ORIGINAL PAPER Behavioral pricing of energy swing options by stochastic bilevel optimization Peter Gross 1 Georg Ch. Pflug 2,3 Received: 20 January

More information

Dynamic Asset and Liability Management Models for Pension Systems

Dynamic Asset and Liability Management Models for Pension Systems Dynamic Asset and Liability Management Models for Pension Systems The Comparison between Multi-period Stochastic Programming Model and Stochastic Control Model Muneki Kawaguchi and Norio Hibiki June 1,

More information

Fast Convergence of Regress-later Series Estimators

Fast Convergence of Regress-later Series Estimators Fast Convergence of Regress-later Series Estimators New Thinking in Finance, London Eric Beutner, Antoon Pelsser, Janina Schweizer Maastricht University & Kleynen Consultants 12 February 2014 Beutner Pelsser

More information

The Irrevocable Multi-Armed Bandit Problem

The Irrevocable Multi-Armed Bandit Problem The Irrevocable Multi-Armed Bandit Problem Ritesh Madan Qualcomm-Flarion Technologies May 27, 2009 Joint work with Vivek Farias (MIT) 2 Multi-Armed Bandit Problem n arms, where each arm i is a Markov Decision

More information

Dynamic Programming (DP) Massimo Paolucci University of Genova

Dynamic Programming (DP) Massimo Paolucci University of Genova Dynamic Programming (DP) Massimo Paolucci University of Genova DP cannot be applied to each kind of problem In particular, it is a solution method for problems defined over stages For each stage a subproblem

More information

On Complexity of Multistage Stochastic Programs

On Complexity of Multistage Stochastic Programs On Complexity of Multistage Stochastic Programs Alexander Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205, USA e-mail: ashapiro@isye.gatech.edu

More information

Chapter 21. Dynamic Programming CONTENTS 21.1 A SHORTEST-ROUTE PROBLEM 21.2 DYNAMIC PROGRAMMING NOTATION

Chapter 21. Dynamic Programming CONTENTS 21.1 A SHORTEST-ROUTE PROBLEM 21.2 DYNAMIC PROGRAMMING NOTATION Chapter 21 Dynamic Programming CONTENTS 21.1 A SHORTEST-ROUTE PROBLEM 21.2 DYNAMIC PROGRAMMING NOTATION 21.3 THE KNAPSACK PROBLEM 21.4 A PRODUCTION AND INVENTORY CONTROL PROBLEM 23_ch21_ptg01_Web.indd

More information

We formulate and solve two new stochastic linear programming formulations of appointment scheduling

We formulate and solve two new stochastic linear programming formulations of appointment scheduling Published online ahead of print December 7, 2011 INFORMS Journal on Computing Articles in Advance, pp. 1 17 issn 1091-9856 eissn 1526-5528 http://dx.doi.org/10.1287/ijoc.1110.0482 2011 INFORMS Dynamic

More information

Stochastic Optimization

Stochastic Optimization Stochastic Optimization Introduction and Examples Alireza Ghaffari-Hadigheh Azarbaijan Shahid Madani University (ASMU) hadigheha@azaruniv.edu Fall 2017 Alireza Ghaffari-Hadigheh (ASMU) Stochastic Optimization

More information

Energy Systems under Uncertainty: Modeling and Computations

Energy Systems under Uncertainty: Modeling and Computations Energy Systems under Uncertainty: Modeling and Computations W. Römisch Humboldt-University Berlin Department of Mathematics www.math.hu-berlin.de/~romisch Systems Analysis 2015, November 11 13, IIASA (Laxenburg,

More information

Introduction to Dynamic Programming

Introduction to Dynamic Programming Introduction to Dynamic Programming http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html Acknowledgement: this slides is based on Prof. Mengdi Wang s and Prof. Dimitri Bertsekas lecture notes Outline 2/65 1

More information

Handout 4: Deterministic Systems and the Shortest Path Problem

Handout 4: Deterministic Systems and the Shortest Path Problem SEEM 3470: Dynamic Optimization and Applications 2013 14 Second Term Handout 4: Deterministic Systems and the Shortest Path Problem Instructor: Shiqian Ma January 27, 2014 Suggested Reading: Bertsekas

More information

17 MAKING COMPLEX DECISIONS

17 MAKING COMPLEX DECISIONS 267 17 MAKING COMPLEX DECISIONS The agent s utility now depends on a sequence of decisions In the following 4 3grid environment the agent makes a decision to move (U, R, D, L) at each time step When the

More information

Optimal Policies for Distributed Data Aggregation in Wireless Sensor Networks

Optimal Policies for Distributed Data Aggregation in Wireless Sensor Networks Optimal Policies for Distributed Data Aggregation in Wireless Sensor Networks Hussein Abouzeid Department of Electrical Computer and Systems Engineering Rensselaer Polytechnic Institute abouzeid@ecse.rpi.edu

More information

Scenario Generation for Stochastic Programming Introduction and selected methods

Scenario Generation for Stochastic Programming Introduction and selected methods Michal Kaut Scenario Generation for Stochastic Programming Introduction and selected methods SINTEF Technology and Society September 2011 Scenario Generation for Stochastic Programming 1 Outline Introduction

More information

3. The Dynamic Programming Algorithm (cont d)

3. The Dynamic Programming Algorithm (cont d) 3. The Dynamic Programming Algorithm (cont d) Last lecture e introduced the DPA. In this lecture, e first apply the DPA to the chess match example, and then sho ho to deal ith problems that do not match

More information

On the Marginal Value of Water for Hydroelectricity

On the Marginal Value of Water for Hydroelectricity Chapter 31 On the Marginal Value of Water for Hydroelectricity Andy Philpott 21 31.1 Introduction This chapter discusses optimization models for computing prices in perfectly competitive wholesale electricity

More information

Optimal investments under dynamic performance critria. Lecture IV

Optimal investments under dynamic performance critria. Lecture IV Optimal investments under dynamic performance critria Lecture IV 1 Utility-based measurement of performance 2 Deterministic environment Utility traits u(x, t) : x wealth and t time Monotonicity u x (x,

More information

Markov Decision Processes

Markov Decision Processes Markov Decision Processes Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. AIMA 3. Chris Amato Stochastic domains So far, we have studied search Can use

More information

Foundations of Artificial Intelligence

Foundations of Artificial Intelligence Foundations of Artificial Intelligence 44. Monte-Carlo Tree Search: Introduction Thomas Keller Universität Basel May 27, 2016 Board Games: Overview chapter overview: 41. Introduction and State of the Art

More information

Flexible Demand Management under Time-Varying Prices. Yong Liang

Flexible Demand Management under Time-Varying Prices. Yong Liang Flexible Demand Management under Time-Varying Prices by Yong Liang A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Industrial Engineering

More information

Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Lecture 21 Successive Shortest Path Problem In this lecture, we continue our discussion

More information