Stochastic Dual Dynamic Programming Algorithm for Multistage Stochastic Programming
|
|
- Kristian Nash
- 6 years ago
- Views:
Transcription
1 Stochastic Dual Dynamic Programg Algorithm for Multistage Stochastic Programg Final presentation ISyE 8813 Fall 2011 Guido Lagos Wajdi Tekaya Georgia Institute of Technology November 30, 2011
2 Multistage Stochastic Linear Programg A 1 x 1 =b 1 x 1 0 c1 T x 1 +E Analysis of Stochastic Dual Dynamic Programg Method Alex Shapiro (2010) B 2 x 1 +A 2 x 2 =b 2 x 2 0 c2 T x 2 + E... + E B T x T 1 +A T x T =b T x T 0 Applications: planning problems in ing, energy, forestry, etc. ct T x T... Challenges: Tractability of E Stagewise dependence of data process {ξ t := (c t, B t, A t, b t)} t=1,...,t Curse of dimensionality
3 Summary 1 Two-stage stochastic programg Cutting-plane method SDDP algorithm for two-stage SP 2 Convergence and main contributions
4 Two-stage stochastic programg 1 Two-stage stochastic programg Cutting-plane method SDDP algorithm for two-stage SP 2 Convergence and main contributions
5 Two-stage stochastic programg True problem: c T x + E Ax=b x 0 T x+w y=h y 0 q T y Take random sample ξ 1,..., ξ N and approximate E 1 N N j=1 = Sample Average Approximation (SAA) problem: c T x + 1 N N q jt y Ax=b x 0 j=1 T j x+w j y=h j y 0 } {{ } } Q(x,ξ j ):= {{ } Q(x):=
6 Two-stage stochastic programg Cutting-plane method Basic Ingredients c T x + 1 N Ax=b x 0 N Q(x, ξ j ) j=1 } {{ } Q(x):= where Q(x, ξ) := max Assume: relatively complete recourse, i.e. feasible x, Q(x, ξ) < a.s. 1 N N j=1 Q(, ξj ) convex piecewise-linear, and problem is T x+w y=h y 0 q T y!!! For f convex: f (x 0) := { d : x f (x) f (x 0) + d T (x x } 0) Q(, [ ξ)(x 0) = T T {π : π opt. sol. of dual of Q(x 0, ξ)} ] 1 N N j=1 Q(, ξj ) (x 0) = 1 N N j=1 T j T {π :. π opt. sol. of dual of Q(x 0, ξ j )}
7 Two-stage stochastic programg Cutting-plane method Basic Ingredients c T x + 1 N Ax=b x 0 N Q(x, ξ j ) j=1 } {{ } Q(x):= where Q(x, ξ) := max Assume: relatively complete recourse, i.e. feasible x, Q(x, ξ) < a.s. Conclusion: c T x + 1 N Q(x, ξ j ) is: N j=1 easy to compute on given x difficult to optimize easy to compute subgradient on given x T x+w y=h y 0 q T y
8 Two-stage stochastic programg Cutting-plane method Cutting Plane Algorithm Given sample { ξ j = (q j, T j, W j, h j ) } j=1,...,n, 0. k 1; LB 1 ; UB 1 ; Q 1 (x) LB 1 x 1. ( Forward Step ) Let x k be the solution of: LB k c T x + Q k (x) Ax=b x 0 2. ( Backward Step ) Compute: Q(x k ) 1 N N j=1 Q(x k, ξ j ) g k 1 N N j=1 T jt π j,k (subgradient) Let UB k c T x k + Q(x k ). If UB k LB k ɛ, END. Else (LB k < UB k ), add plane Q(x k ) + g kt (x x k ) to Q k : Q k+1 (x) max{q k (x), Q(x k ) + g kt (x x k )} 3. k k + 1, iterate from 1.
9 Two-stage stochastic programg SDDP algorithm for two-stage SP SDDP algorithm I Given sample { ξ j = (q j, T j, W j, h j ) } j=1,...,n, 0. k 1; UB 1 ; LB 1 ; Q 1 (x) LB 1 x 1. Forward Step 1.1 Let x k be the solution of: LB k c T x + Q k (x) Ax=b x Take subsample {ξ (j) } M j=1 of {ξj } N j=1 (N >> M), and with values {ϑ j := c T x k + Q(x k, ξ (i) )} M j=1 compute (1 α) confidence upper bound of true problem opt. value ϑ : UB k ϑ + z α/2 σ ϑ / M where ϑ := 1 M M j=1 ϑ j and σ 2 ϑ := 1 M 1 M j=1 (ϑ j ϑ) If UB k LB k ɛ, END.
10 Two-stage stochastic programg SDDP algorithm for two-stage SP SDDP algorithm II 2. Backward Step 2.1 Compute: Q(x k ) 1 N N j=1 Q(x k, ξ j ) = 1 N N j=1 g k 1 N N j=1 T jt π j,k (subgradient) 2.2 Add plane Q(x k ) + g kt ( x k ) to Q k ( ): max T j x k +W j y=h j y 0 Q k+1 (x) max{q k (x), Q(x k ) + g kt (x x k )} 3. k k + 1, iterate from 1. q jt y
11 1 Two-stage stochastic programg Cutting-plane method SDDP algorithm for two-stage SP 2 Convergence and main contributions
12 True problem A 1 x 1 =b 1 x 1 0 c1 T x 1 +E B 2 x 1 +A 2 x 2 =b 2 x 2 0 c2 T x 2 + E... + E Equivalently (Dynamic Programg equations): where A 1 x 1 =b 1 x 1 0 c T 1 x 1 + E [Q 2(x 1, ξ 2)] }{{} Q 2 (x 1 ):= B T x T 1 +A T x T =b T x T 0 ct T x T Q t(x t 1, ξ t) := inf B t x t 1 +A t x t=b t x t 0 Q T (x T 1, ξ T ) := inf B T x T 1 +A T x T =b T x T 0 ct T x t + E [Q t+1(x t, ξ t+1)] }{{} Q t+1 (x t) ct T x T t = 2,..., T 1 Assumptions 1 Process is stagewise independent, i.e. ξ t+1 indep. of ξ 1,..., ξ t. 2 Problem has relatively complete recourse
13 SAA problem Take random sample { ξj t = ( c tj, Ãtj, B tj, b } tj) for each stage j=1,...,n t t = 2,..., T. SAA problem is: A 1 x 1 =b 1 x 1 0 c1 T x 1+ 1 N 2 N 2 j=1 B j2 x 1 +Ãj2x 2 = b j2 x 2 0 N T c j2 T x N 3 N N j=1 T j=1 Equivalently (Dynamic Programg equations): where A 1 x 1 =b 1 x 1 0 c1 T x N 2 Q 2,j(x 1) N 2 j=1 } {{ } Q 2 (x 1 ) B jt x T 1 +ÃjT x T = b jt x T 0 c jt T x T Q t,j(x t 1) := B tj x t 1 +Ãtj x t= b tj x t 0 Q T,j(x T 1) := B Tj x T 1 +ÃTj x T = b Tj x T 0 c T tj x t + 1 c T Tjx T N t+1 N t+1 j=1 Q t+1,j(x t) } {{ } Q t+1 (x t) t = 2,..., T 1
14 SDDP method: the idea!!! Cost-to-go functions Q t (x t 1 ) = 1 N t N t j=1 Q t,j (x t 1 ) t = 2,..., T are convex piecewise-linear Approximate A 1x 1 =b 1 x 1 0 B tj x t 1 +Ãtj xt= b tj x t 0 c1 Tx 1 + Q 2 (x 1 ) by A 1x 1 =b 1 x 1 0 c tj Tx t + Q t+1 (x t ) by B tj x t 1 +Ãtj xt= b tj x t 0 c T 1 x 1 + Q 2 (x 1 ) c T tj x t + Q t+1 (x t ) where Q 2 ( ),..., Q T ( ) are convex, piecewise-linear and Q t ( ) Q t ( ) t = 2,..., T In successive iterations, refine lower approximations Q t ( ) using subgradient of Q t ( ): Q k t ( ) Q k+1 t ( ) Q k+2 t ( )... Q t ( )
15 SDDP method: Forward step At iteration k 1, we have lower approximations Q 2,..., Q T Take subsample {( ξ (j) (j) 2,..., ξ T )}M j=1 of original sample For j = 1,..., M, take sampled process ( ξ (j) (j) 2,..., ξ ) and solve T A1x1=b1 x1 0 ξ (j) t, x t 1,j B t(j) x t 1,j +Ãt(j) xt = b t(j) xt 0 ξ (j) T, x T 1,j B T (j) x T 1,j +ÃT (j) x T = b T (j) x T 0 c1 T x1 + Q2(x1) c t(j) T xt + Qt+1(xt) c T T (j) x T x 1j x tj ( ξ (j) [t] ) t = 2,..., T 1 x T j ( ξ (j) [T ] ) obtaining candidate policy values x 1j, x 2j,..., x T j with cost T ϑ j ct(j) T x tj t=1 It s a (1 α) confidence upper bound of true opt. value ϑ : UB k ϑ + z α σ ϑ / M where ϑ := 1 M M j=1 ϑ j and σ ϑ 2 := 1 M M 1 j=1 (ϑ j ϑ) 2.
16 SDDP method: Backward step We have candidate values x 1, x 2,..., x T. At stage t = T : For x T 1 = x T 1 and for j = 1,..., N T, solve: Q T,j (x T 1 ) := B T,j x T 1 +ÃT,j x T = b T,j x T 0 and let π T,j be opt. dual solution. Let 1 NT Q T (x T 1 ) := N T Add cut g T := 1 N T c T T,jx T Q j=1 T,j (x T 1 ) NT B j=1 T T,j π T,j. L T (x T 1 ) := Q T (x T 1 ) + g T T (x T 1 x T 1 ) to lower approx. Q T used in stage t = T 1: Q T ( ) := max{q T ( ), L T ( )}.
17 SDDP method: Backward step At stage t = T 1: For x T 2 = x T 2 and for j = 1,..., N T 1, solve: Q T 1,j (x T 2 ) := B T 1,j x T 2 +ÃT 1,j x T 1 = b T 1,j x T 1 0 and let π T 1,j be opt. dual solution. Let Add cut c T T 1,jx T 1 +Q T (x T 1 ) 1 NT Q T 1 (x T 2 ) := 1 N T 1 j=1 Q T 1,j (x T 2 ) g T 1 := 1 NT 1 N T 1 j=1 B T T 1,j π T 1,j. L T 1 (x T 2 ) := Q T 1 (x T 2 ) + g T T 1(x T 2 x T 2 ) to lower approx. Q T 1 used in stage t = T 2: Q T 1 ( ) := max{q T 1 ( ), L T 1 ( )}.
18 SDDP method: Backward step At stage t = T 2: At stage t = 1: solve LB k A 1 x 1 =b 1 x 1 0 c T 1 x 1 + Q 2 (x 1 ) LB k is, on average, lower bound to ϑ true optimal value
19 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: True problem
20 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SAA problem
21 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Forward step
22 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Forward step
23 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Forward step
24 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Forward step
25 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Backward step
26 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Backward step
27 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 1: Backward step
28 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Forward step
29 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Forward step
30 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Forward step
31 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Backward step
32 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Backward step
33 SDDP method: Illustration (a) Stage 1 (b) Stage 2 (c) Stage 3 Figure: SDDP iteration 2: Backward step...
34 Summary: KEY IDEAS for SDDP algorithm Forward step: sample process ξ 1,..., ξ T implementable policy x 1,..., x T Repetitions Upper bound on true optimal value Backward step: Cost-to-go functions Q t (x t 1 ) = 1 N t N t max c t,jx T t + Q t+1 (x t ) j=1 B t,j x t 1+Ãt,j x t= b t,j x t 0 are convex piecewise-linear functions of x t 1 Refine lower approximations Q t ( ) using subgradient [ ] Qt ( ) (x t 1 ) = 1 N t N t j=1 Lower bound on true optimal value B T t,j {π t,j : opt. sol. of dual... }
35 Convergence and main contributions Proposition (Convergence) Assume i. At the forward step, process subsamples are taken independently of each other ii. At all iterations, approximated problems A 1 x 1 =b 1 x 1 0 c1 T x 1 + Q 2 (x 1 ) and B t,j x t 1 +Ãt,j x t = b t,j x t 0 have finite optimal value iii. In the backward step, basical solutions are used c T t,jx t + Q t+1 (x t ) Then, after a sufficiently but finite large amount of iterations of the SDDP algorithm, the forward procedure defines an optimal policy for the SAA problem.
36 Convergence and main contributions Tractability? Issue: total number of scenarios Π T t=2 N t Cutting plane for two-stage: bad SDDP algorithm: generalization of cutting plane method even worse? One run of Backward step: solve 1 + N N T LP s One run of Forward step: solve 1 + M(T 1) LP s Tractability : SDDP method construct feasible policy
37 Convergence and main contributions THE END
Worst-case-expectation approach to optimization under uncertainty
Worst-case-expectation approach to optimization under uncertainty Wajdi Tekaya Joint research with Alexander Shapiro, Murilo Pereira Soares and Joari Paulo da Costa : Cambridge Systems Associates; : Georgia
More informationRobust Dual Dynamic Programming
1 / 18 Robust Dual Dynamic Programming Angelos Georghiou, Angelos Tsoukalas, Wolfram Wiesemann American University of Beirut Olayan School of Business 31 May 217 2 / 18 Inspired by SDDP Stochastic optimization
More informationDASC: A DECOMPOSITION ALGORITHM FOR MULTISTAGE STOCHASTIC PROGRAMS WITH STRONGLY CONVEX COST FUNCTIONS
DASC: A DECOMPOSITION ALGORITHM FOR MULTISTAGE STOCHASTIC PROGRAMS WITH STRONGLY CONVEX COST FUNCTIONS Vincent Guigues School of Applied Mathematics, FGV Praia de Botafogo, Rio de Janeiro, Brazil vguigues@fgv.br
More informationReport for technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional do Sistema Elétrico Risk Averse Approach
Report for technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional do Sistema Elétrico Risk Averse Approach Alexander Shapiro and Wajdi Tekaya School of Industrial and
More informationOn solving multistage stochastic programs with coherent risk measures
On solving multistage stochastic programs with coherent risk measures Andy Philpott Vitor de Matos y Erlon Finardi z August 13, 2012 Abstract We consider a class of multistage stochastic linear programs
More informationMultistage risk-averse asset allocation with transaction costs
Multistage risk-averse asset allocation with transaction costs 1 Introduction Václav Kozmík 1 Abstract. This paper deals with asset allocation problems formulated as multistage stochastic programming models.
More informationStochastic Dual Dynamic Programming
1 / 43 Stochastic Dual Dynamic Programming Operations Research Anthony Papavasiliou 2 / 43 Contents [ 10.4 of BL], [Pereira, 1991] 1 Recalling the Nested L-Shaped Decomposition 2 Drawbacks of Nested Decomposition
More informationSOLVING ROBUST SUPPLY CHAIN PROBLEMS
SOLVING ROBUST SUPPLY CHAIN PROBLEMS Daniel Bienstock Nuri Sercan Özbay Columbia University, New York November 13, 2005 Project with Lucent Technologies Optimize the inventory buffer levels in a complicated
More informationStochastic Dual Dynamic integer Programming
Stochastic Dual Dynamic integer Programming Shabbir Ahmed Georgia Tech Jikai Zou Andy Sun Multistage IP Canonical deterministic formulation ( X T ) f t (x t,y t ):(x t 1,x t,y t ) 2 X t 8 t x t min x,y
More informationFinancial Optimization ISE 347/447. Lecture 15. Dr. Ted Ralphs
Financial Optimization ISE 347/447 Lecture 15 Dr. Ted Ralphs ISE 347/447 Lecture 15 1 Reading for This Lecture C&T Chapter 12 ISE 347/447 Lecture 15 2 Stock Market Indices A stock market index is a statistic
More informationIE 495 Lecture 11. The LShaped Method. Prof. Jeff Linderoth. February 19, February 19, 2003 Stochastic Programming Lecture 11 Slide 1
IE 495 Lecture 11 The LShaped Method Prof. Jeff Linderoth February 19, 2003 February 19, 2003 Stochastic Programming Lecture 11 Slide 1 Before We Begin HW#2 $300 $0 http://www.unizh.ch/ior/pages/deutsch/mitglieder/kall/bib/ka-wal-94.pdf
More informationApproximation of Continuous-State Scenario Processes in Multi-Stage Stochastic Optimization and its Applications
Approximation of Continuous-State Scenario Processes in Multi-Stage Stochastic Optimization and its Applications Anna Timonina University of Vienna, Abraham Wald PhD Program in Statistics and Operations
More informationA Robust Option Pricing Problem
IMA 2003 Workshop, March 12-19, 2003 A Robust Option Pricing Problem Laurent El Ghaoui Department of EECS, UC Berkeley 3 Robust optimization standard form: min x sup u U f 0 (x, u) : u U, f i (x, u) 0,
More informationScenario Generation and Sampling Methods
Scenario Generation and Sampling Methods Güzin Bayraksan Tito Homem-de-Mello SVAN 2016 IMPA May 9th, 2016 Bayraksan (OSU) & Homem-de-Mello (UAI) Scenario Generation and Sampling SVAN IMPA May 9 1 / 30
More informationMULTISTAGE STOCHASTIC PROGRAMS WITH A RANDOM NUMBER OF STAGES: DYNAMIC PROGRAMMING EQUATIONS, SOLUTION METHODS, AND APPLICATION TO PORTFOLIO SELECTION
MULTISTAGE STOCHASTIC PROGRAMS WITH A RANDOM NUMBER OF STAGES: DYNAMIC PROGRAMMING EQUATIONS, SOLUTION METHODS, AND APPLICATION TO PORTFOLIO SELECTION Vincent Guigues School of Applied Mathematics, FGV
More informationOn Complexity of Multistage Stochastic Programs
On Complexity of Multistage Stochastic Programs Alexander Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205, USA e-mail: ashapiro@isye.gatech.edu
More informationDynamic sampling algorithms for multi-stage stochastic programs with risk aversion
Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion A.B. Philpott y and V.L. de Matos z October 7, 2011 Abstract We consider the incorporation of a time-consistent coherent
More informationScenario tree generation for stochastic programming models using GAMS/SCENRED
Scenario tree generation for stochastic programming models using GAMS/SCENRED Holger Heitsch 1 and Steven Dirkse 2 1 Humboldt-University Berlin, Department of Mathematics, Germany 2 GAMS Development Corp.,
More informationInvestigation of the and minimum storage energy target levels approach. Final Report
Investigation of the AV@R and minimum storage energy target levels approach Final Report First activity of the technical cooperation between Georgia Institute of Technology and ONS - Operador Nacional
More informationContents Critique 26. portfolio optimization 32
Contents Preface vii 1 Financial problems and numerical methods 3 1.1 MATLAB environment 4 1.1.1 Why MATLAB? 5 1.2 Fixed-income securities: analysis and portfolio immunization 6 1.2.1 Basic valuation of
More informationQuasi-Convex Stochastic Dynamic Programming
Quasi-Convex Stochastic Dynamic Programming John R. Birge University of Chicago Booth School of Business JRBirge SIAM FM12, MSP, 10 July 2012 1 General Theme Many dynamic optimization problems dealing
More informationApproximations of Stochastic Programs. Scenario Tree Reduction and Construction
Approximations of Stochastic Programs. Scenario Tree Reduction and Construction W. Römisch Humboldt-University Berlin Institute of Mathematics 10099 Berlin, Germany www.mathematik.hu-berlin.de/~romisch
More informationMartingale Pricing Theory in Discrete-Time and Discrete-Space Models
IEOR E4707: Foundations of Financial Engineering c 206 by Martin Haugh Martingale Pricing Theory in Discrete-Time and Discrete-Space Models These notes develop the theory of martingale pricing in a discrete-time,
More informationThe Irrevocable Multi-Armed Bandit Problem
The Irrevocable Multi-Armed Bandit Problem Ritesh Madan Qualcomm-Flarion Technologies May 27, 2009 Joint work with Vivek Farias (MIT) 2 Multi-Armed Bandit Problem n arms, where each arm i is a Markov Decision
More informationOptimal Security Liquidation Algorithms
Optimal Security Liquidation Algorithms Sergiy Butenko Department of Industrial Engineering, Texas A&M University, College Station, TX 77843-3131, USA Alexander Golodnikov Glushkov Institute of Cybernetics,
More informationDynamic Programming: An overview. 1 Preliminaries: The basic principle underlying dynamic programming
Dynamic Programming: An overview These notes summarize some key properties of the Dynamic Programming principle to optimize a function or cost that depends on an interval or stages. This plays a key role
More informationAsset-Liability Management
Asset-Liability Management John Birge University of Chicago Booth School of Business JRBirge INFORMS San Francisco, Nov. 2014 1 Overview Portfolio optimization involves: Modeling Optimization Estimation
More informationAssessing Policy Quality in Multi-stage Stochastic Programming
Assessing Policy Quality in Multi-stage Stochastic Programming Anukal Chiralaksanakul and David P. Morton Graduate Program in Operations Research The University of Texas at Austin Austin, TX 78712 January
More information6.231 DYNAMIC PROGRAMMING LECTURE 3 LECTURE OUTLINE
6.21 DYNAMIC PROGRAMMING LECTURE LECTURE OUTLINE Deterministic finite-state DP problems Backward shortest path algorithm Forward shortest path algorithm Shortest path examples Alternative shortest path
More informationRisk aversion in multi-stage stochastic programming: a modeling and algorithmic perspective
Risk aversion in multi-stage stochastic programming: a modeling and algorithmic perspective Tito Homem-de-Mello School of Business Universidad Adolfo Ibañez, Santiago, Chile Joint work with Bernardo Pagnoncelli
More informationDynamic sampling algorithms for multi-stage stochastic programs with risk aversion
Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion A.B. Philpott y and V.L. de Matos z March 28, 2011 Abstract We consider the incorporation of a time-consistent coherent
More informationRisk Management for Chemical Supply Chain Planning under Uncertainty
for Chemical Supply Chain Planning under Uncertainty Fengqi You and Ignacio E. Grossmann Dept. of Chemical Engineering, Carnegie Mellon University John M. Wassick The Dow Chemical Company Introduction
More informationMultistage Stochastic Demand-side Management for Price-Making Major Consumers of Electricity in a Co-optimized Energy and Reserve Market
Multistage Stochastic Demand-side Management for Price-Making Major Consumers of Electricity in a Co-optimized Energy and Reserve Market Mahbubeh Habibian Anthony Downward Golbon Zakeri Abstract In this
More informationMultistage Stochastic Mixed-Integer Programs for Optimizing Gas Contract and Scheduling Maintenance
Multistage Stochastic Mixed-Integer Programs for Optimizing Gas Contract and Scheduling Maintenance Zhe Liu Siqian Shen September 2, 2012 Abstract In this paper, we present multistage stochastic mixed-integer
More informationDecomposition Methods
Decomposition Methods separable problems, complicating variables primal decomposition dual decomposition complicating constraints general decomposition structures Prof. S. Boyd, EE364b, Stanford University
More informationMarkov Decision Processes II
Markov Decision Processes II Daisuke Oyama Topics in Economic Theory December 17, 2014 Review Finite state space S, finite action space A. The value of a policy σ A S : v σ = β t Q t σr σ, t=0 which satisfies
More informationOptimal energy management and stochastic decomposition
Optimal energy management and stochastic decomposition F. Pacaud P. Carpentier J.P. Chancelier M. De Lara JuMP-dev workshop, 2018 ENPC ParisTech ENSTA ParisTech Efficacity 1/23 Motivation We consider a
More informationColumn generation to solve planning problems
Column generation to solve planning problems ALGORITMe Han Hoogeveen 1 Continuous Knapsack problem We are given n items with integral weight a j ; integral value c j. B is a given integer. Goal: Find a
More informationHigh-dimensional risk-constrained dynamic asset allocation via Markov stochastic dual dynamic programming
High-dimensional risk-constrained dynamic asset allocation via Markov stochastic dual dynamic programming Davi Valladão Industrial Engineering Department, PUC-Rio, davimv@puc-rio.br Thuener Silva Industrial
More informationLecture Notes 1
4.45 Lecture Notes Guido Lorenzoni Fall 2009 A portfolio problem To set the stage, consider a simple nite horizon problem. A risk averse agent can invest in two assets: riskless asset (bond) pays gross
More informationComputer Vision Group Prof. Daniel Cremers. 7. Sequential Data
Group Prof. Daniel Cremers 7. Sequential Data Bayes Filter (Rep.) We can describe the overall process using a Dynamic Bayes Network: This incorporates the following Markov assumptions: (measurement) (state)!2
More informationEllipsoid Method. ellipsoid method. convergence proof. inequality constraints. feasibility problems. Prof. S. Boyd, EE364b, Stanford University
Ellipsoid Method ellipsoid method convergence proof inequality constraints feasibility problems Prof. S. Boyd, EE364b, Stanford University Ellipsoid method developed by Shor, Nemirovsky, Yudin in 1970s
More informationOptimal monetary policy when asset markets are incomplete
Optimal monetary policy when asset markets are incomplete R. Anton Braun Tomoyuki Nakajima 2 University of Tokyo, and CREI 2 Kyoto University, and RIETI December 9, 28 Outline Introduction 2 Model Individuals
More informationPolicy iterated lower bounds and linear MC upper bounds for Bermudan style derivatives
Finance Winterschool 2007, Lunteren NL Policy iterated lower bounds and linear MC upper bounds for Bermudan style derivatives Pricing complex structured products Mohrenstr 39 10117 Berlin schoenma@wias-berlin.de
More informationModeling Time-dependent Randomness in Stochastic Dual Dynamic Programming
Modeling Time-dependent Randomness in Stochastic Dual Dynamic Programming Nils Löhndorf Department of Information Systems and Operations Vienna University of Economics and Business Vienna, Austria nils.loehndorf@wu.ac.at
More informationA Stochastic Approximation Algorithm for Making Pricing Decisions in Network Revenue Management Problems
A Stochastic Approximation Algorithm for Making ricing Decisions in Network Revenue Management roblems Sumit Kunnumkal Indian School of Business, Gachibowli, Hyderabad, 500032, India sumit kunnumkal@isb.edu
More informationAn Empirical Study of Optimization for Maximizing Diffusion in Networks
An Empirical Study of Optimization for Maximizing Diffusion in Networks Kiyan Ahmadizadeh Bistra Dilkina, Carla P. Gomes, Ashish Sabharwal Cornell University Institute for Computational Sustainability
More informationStochastic Optimization Methods in Scheduling. Rolf H. Möhring Technische Universität Berlin Combinatorial Optimization and Graph Algorithms
Stochastic Optimization Methods in Scheduling Rolf H. Möhring Technische Universität Berlin Combinatorial Optimization and Graph Algorithms More expensive and longer... Eurotunnel Unexpected loss of 400,000,000
More informationThe Values of Information and Solution in Stochastic Programming
The Values of Information and Solution in Stochastic Programming John R. Birge The University of Chicago Booth School of Business JRBirge ICSP, Bergamo, July 2013 1 Themes The values of information and
More informationSupport Vector Machines: Training with Stochastic Gradient Descent
Support Vector Machines: Training with Stochastic Gradient Descent Machine Learning Spring 2018 The slides are mainly from Vivek Srikumar 1 Support vector machines Training by maximizing margin The SVM
More informationApproximate Composite Minimization: Convergence Rates and Examples
ISMP 2018 - Bordeaux Approximate Composite Minimization: Convergence Rates and S. Praneeth Karimireddy, Sebastian U. Stich, Martin Jaggi MLO Lab, EPFL, Switzerland sebastian.stich@epfl.ch July 4, 2018
More informationNew Policies for Stochastic Inventory Control Models: Theoretical and Computational Results
OPERATIONS RESEARCH Vol. 00, No. 0, Xxxxx 0000, pp. 000 000 issn 0030-364X eissn 1526-5463 00 0000 0001 INFORMS doi 10.1287/xxxx.0000.0000 c 0000 INFORMS New Policies for Stochastic Inventory Control Models:
More informationFrom Discrete Time to Continuous Time Modeling
From Discrete Time to Continuous Time Modeling Prof. S. Jaimungal, Department of Statistics, University of Toronto 2004 Arrow-Debreu Securities 2004 Prof. S. Jaimungal 2 Consider a simple one-period economy
More informationMultistage Stochastic Programming
Multistage Stochastic Programming John R. Birge University of Michigan Models - Long and short term - Risk inclusion Approximations - stages and scenarios Computation Slide Number 1 OUTLINE Motivation
More informationForecast Horizons for Production Planning with Stochastic Demand
Forecast Horizons for Production Planning with Stochastic Demand Alfredo Garcia and Robert L. Smith Department of Industrial and Operations Engineering Universityof Michigan, Ann Arbor MI 48109 December
More informationMultistage Stochastic Programming
IE 495 Lecture 21 Multistage Stochastic Programming Prof. Jeff Linderoth April 16, 2003 April 16, 2002 Stochastic Programming Lecture 21 Slide 1 Outline HW Fixes Multistage Stochastic Programming Modeling
More informationSequential Decision Making
Sequential Decision Making Dynamic programming Christos Dimitrakakis Intelligent Autonomous Systems, IvI, University of Amsterdam, The Netherlands March 18, 2008 Introduction Some examples Dynamic programming
More informationInformation Relaxations and Duality in Stochastic Dynamic Programs
Information Relaxations and Duality in Stochastic Dynamic Programs David Brown, Jim Smith, and Peng Sun Fuqua School of Business Duke University February 28 1/39 Dynamic programming is widely applicable
More informationCPS 270: Artificial Intelligence Markov decision processes, POMDPs
CPS 270: Artificial Intelligence http://www.cs.duke.edu/courses/fall08/cps270/ Markov decision processes, POMDPs Instructor: Vincent Conitzer Warmup: a Markov process with rewards We derive some reward
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Other Miscellaneous Topics and Applications of Monte-Carlo Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com
More informationAllocation of Risk Capital via Intra-Firm Trading
Allocation of Risk Capital via Intra-Firm Trading Sean Hilden Department of Mathematical Sciences Carnegie Mellon University December 5, 2005 References 1. Artzner, Delbaen, Eber, Heath: Coherent Measures
More informationReinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration
Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration Piyush Rai CS5350/6350: Machine Learning November 29, 2011 Reinforcement Learning Supervised Learning: Uses explicit supervision
More informationReinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration
Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration Piyush Rai CS5350/6350: Machine Learning November 29, 2011 Reinforcement Learning Supervised Learning: Uses explicit supervision
More informationShape-Preserving Dynamic Programming
Shape-Preserving Dynamic Programming Kenneth Judd and Yongyang Cai July 20, 2011 1 Introduction The multi-stage decision-making problems are numerically challenging. When the problems are time-separable,
More informationTHE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE
THE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE GÜNTER ROTE Abstract. A salesperson wants to visit each of n objects that move on a line at given constant speeds in the shortest possible time,
More informationMarkov Decision Processes
Markov Decision Processes Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. AIMA 3. Chris Amato Stochastic domains So far, we have studied search Can use
More informationSome simple Bitcoin Economics
Some simple Bitcoin Economics Linda Schilling 1 and Harald Uhlig 2 1 École Polytechnique - CREST Department of Economics lin.schilling@gmail.com 2 University of Chicago Department of Economics huhlig@uchicago.edu
More informationDM559/DM545 Linear and integer programming
Department of Mathematics and Computer Science University of Southern Denmark, Odense May 22, 2018 Marco Chiarandini DM559/DM55 Linear and integer programming Sheet, Spring 2018 [pdf format] Contains Solutions!
More informationStochastic Optimal Control With Dynamic, Time-Consistent Risk Constraints
Stochastic Optimal Control With Dynamic, Time-Consistent Risk Constraints Yin-Lam Chow, Marco Pavone risk metric is applied to the future stream of costs; typical examples include variance-constrained
More informationCOMP331/557. Chapter 6: Optimisation in Finance: Cash-Flow. (Cornuejols & Tütüncü, Chapter 3)
COMP331/557 Chapter 6: Optimisation in Finance: Cash-Flow (Cornuejols & Tütüncü, Chapter 3) 159 Cash-Flow Management Problem A company has the following net cash flow requirements (in 1000 s of ): Month
More informationStochastic Programming and Financial Analysis IE447. Midterm Review. Dr. Ted Ralphs
Stochastic Programming and Financial Analysis IE447 Midterm Review Dr. Ted Ralphs IE447 Midterm Review 1 Forming a Mathematical Programming Model The general form of a mathematical programming model is:
More informationComparison of Static and Dynamic Asset Allocation Models
Comparison of Static and Dynamic Asset Allocation Models John R. Birge University of Michigan University of Michigan 1 Outline Basic Models Static Markowitz mean-variance Dynamic stochastic programming
More informationLecture 7: Bayesian approach to MAB - Gittins index
Advanced Topics in Machine Learning and Algorithmic Game Theory Lecture 7: Bayesian approach to MAB - Gittins index Lecturer: Yishay Mansour Scribe: Mariano Schain 7.1 Introduction In the Bayesian approach
More informationOnline Appendix: Extensions
B Online Appendix: Extensions In this online appendix we demonstrate that many important variations of the exact cost-basis LUL framework remain tractable. In particular, dual problem instances corresponding
More information6.231 DYNAMIC PROGRAMMING LECTURE 3 LECTURE OUTLINE
6.21 DYNAMIC PROGRAMMING LECTURE LECTURE OUTLINE Deterministic finite-state DP problems Backward shortest path algorithm Forward shortest path algorithm Shortest path examples Alternative shortest path
More informationInterpolation. 1 What is interpolation? 2 Why are we interested in this?
Interpolation 1 What is interpolation? For a certain function f (x we know only the values y 1 = f (x 1,,y n = f (x n For a point x different from x 1,,x n we would then like to approximate f ( x using
More informationOnline Appendix Optimal Time-Consistent Government Debt Maturity D. Debortoli, R. Nunes, P. Yared. A. Proofs
Online Appendi Optimal Time-Consistent Government Debt Maturity D. Debortoli, R. Nunes, P. Yared A. Proofs Proof of Proposition 1 The necessity of these conditions is proved in the tet. To prove sufficiency,
More informationCSCI 1951-G Optimization Methods in Finance Part 00: Course Logistics Introduction to Finance Optimization Problems
CSCI 1951-G Optimization Methods in Finance Part 00: Course Logistics Introduction to Finance Optimization Problems January 26, 2018 1 / 24 Basic information All information is available in the syllabus
More information17 MAKING COMPLEX DECISIONS
267 17 MAKING COMPLEX DECISIONS The agent s utility now depends on a sequence of decisions In the following 4 3grid environment the agent makes a decision to move (U, R, D, L) at each time step When the
More informationMarkov Decision Processes (MDPs) CS 486/686 Introduction to AI University of Waterloo
Markov Decision Processes (MDPs) CS 486/686 Introduction to AI University of Waterloo Outline Sequential Decision Processes Markov chains Highlight Markov property Discounted rewards Value iteration Markov
More informationEllipsoid Method. ellipsoid method. convergence proof. inequality constraints. feasibility problems. Prof. S. Boyd, EE392o, Stanford University
Ellipsoid Method ellipsoid method convergence proof inequality constraints feasibility problems Prof. S. Boyd, EE392o, Stanford University Challenges in cutting-plane methods can be difficult to compute
More informationStrategies and Nash Equilibrium. A Whirlwind Tour of Game Theory
Strategies and Nash Equilibrium A Whirlwind Tour of Game Theory (Mostly from Fudenberg & Tirole) Players choose actions, receive rewards based on their own actions and those of the other players. Example,
More informationEuropean Contingent Claims
European Contingent Claims Seminar: Financial Modelling in Life Insurance organized by Dr. Nikolic and Dr. Meyhöfer Zhiwen Ning 13.05.2016 Zhiwen Ning European Contingent Claims 13.05.2016 1 / 23 outline
More informationLarge-Scale SVM Optimization: Taking a Machine Learning Perspective
Large-Scale SVM Optimization: Taking a Machine Learning Perspective Shai Shalev-Shwartz Toyota Technological Institute at Chicago Joint work with Nati Srebro Talk at NEC Labs, Princeton, August, 2008 Shai
More informationOptimal investments under dynamic performance critria. Lecture IV
Optimal investments under dynamic performance critria Lecture IV 1 Utility-based measurement of performance 2 Deterministic environment Utility traits u(x, t) : x wealth and t time Monotonicity u x (x,
More informationDynamic Replication of Non-Maturing Assets and Liabilities
Dynamic Replication of Non-Maturing Assets and Liabilities Michael Schürle Institute for Operations Research and Computational Finance, University of St. Gallen, Bodanstr. 6, CH-9000 St. Gallen, Switzerland
More informationLecture 5 January 30
EE 223: Stochastic Estimation and Control Spring 2007 Lecture 5 January 30 Lecturer: Venkat Anantharam Scribe: aryam Kamgarpour 5.1 Secretary Problem The problem set-up is explained in Lecture 4. We review
More informationNote on Oblivious equilibrium
Note on Oblivious equilibrium November 29, 2016 1 Introduction Ericson and Pakes(1995) introduced a framework for dynamic oligopoly equilibrium with entry and exit, which can be used for policy analysis.
More informationAn adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity
An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity Coralia Cartis, Nick Gould and Philippe Toint Department of Mathematics,
More informationThe Uncertain Volatility Model
The Uncertain Volatility Model Claude Martini, Antoine Jacquier July 14, 008 1 Black-Scholes and realised volatility What happens when a trader uses the Black-Scholes (BS in the sequel) formula to sell
More informationHIGHER ORDER BINARY OPTIONS AND MULTIPLE-EXPIRY EXOTICS
Electronic Journal of Mathematical Analysis and Applications Vol. (2) July 203, pp. 247-259. ISSN: 2090-792X (online) http://ejmaa.6te.net/ HIGHER ORDER BINARY OPTIONS AND MULTIPLE-EXPIRY EXOTICS HYONG-CHOL
More informationROBUST OPTIMIZATION OF MULTI-PERIOD PRODUCTION PLANNING UNDER DEMAND UNCERTAINTY. A. Ben-Tal, B. Golany and M. Rozenblit
ROBUST OPTIMIZATION OF MULTI-PERIOD PRODUCTION PLANNING UNDER DEMAND UNCERTAINTY A. Ben-Tal, B. Golany and M. Rozenblit Faculty of Industrial Engineering and Management, Technion, Haifa 32000, Israel ABSTRACT
More informationTangent Lévy Models. Sergey Nadtochiy (joint work with René Carmona) Oxford-Man Institute of Quantitative Finance University of Oxford.
Tangent Lévy Models Sergey Nadtochiy (joint work with René Carmona) Oxford-Man Institute of Quantitative Finance University of Oxford June 24, 2010 6th World Congress of the Bachelier Finance Society Sergey
More informationADVANCED MACROECONOMIC TECHNIQUES NOTE 7b
316-406 ADVANCED MACROECONOMIC TECHNIQUES NOTE 7b Chris Edmond hcpedmond@unimelb.edu.aui Aiyagari s model Arguably the most popular example of a simple incomplete markets model is due to Rao Aiyagari (1994,
More informationOptimally Thresholded Realized Power Variations for Lévy Jump Diffusion Models
Optimally Thresholded Realized Power Variations for Lévy Jump Diffusion Models José E. Figueroa-López 1 1 Department of Statistics Purdue University University of Missouri-Kansas City Department of Mathematics
More informationEARLY EXERCISE OPTIONS: UPPER BOUNDS
EARLY EXERCISE OPTIONS: UPPER BOUNDS LEIF B.G. ANDERSEN AND MARK BROADIE Abstract. In this article, we discuss how to generate upper bounds for American or Bermudan securities by Monte Carlo methods. These
More informationChapter 3: Black-Scholes Equation and Its Numerical Evaluation
Chapter 3: Black-Scholes Equation and Its Numerical Evaluation 3.1 Itô Integral 3.1.1 Convergence in the Mean and Stieltjes Integral Definition 3.1 (Convergence in the Mean) A sequence {X n } n ln of random
More informationOptimal Switching Games for Emissions Trading
Optimal Switching Games for Emissions Trading Mike Department of Statistics & Applied Probability University of California Santa Barbara MSRI, May 4, 2009 1 / 29 Outline Cap-and-Trade: Producer Perspective
More informationDynamic Portfolio Choice II
Dynamic Portfolio Choice II Dynamic Programming Leonid Kogan MIT, Sloan 15.450, Fall 2010 c Leonid Kogan ( MIT, Sloan ) Dynamic Portfolio Choice II 15.450, Fall 2010 1 / 35 Outline 1 Introduction to Dynamic
More informationPOMDPs: Partially Observable Markov Decision Processes Advanced AI
POMDPs: Partially Observable Markov Decision Processes Advanced AI Wolfram Burgard Types of Planning Problems Classical Planning State observable Action Model Deterministic, accurate MDPs observable stochastic
More information