IEOR E4703: Monte-Carlo Simulation

Similar documents
IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation

AMH4 - ADVANCED OPTION PRICING. Contents

2 Control variates. λe λti λe e λt i where R(t) = t Y 1 Y N(t) is the time from the last event to t. L t = e λr(t) e e λt(t) Exercises

Ch4. Variance Reduction Techniques

Interval estimation. September 29, Outline Basic ideas Sampling variation and CLT Interval estimation using X More general problems

Monte Carlo Methods in Option Pricing. UiO-STK4510 Autumn 2015

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

V ar( X 1 + X 2. ) = ( V ar(x 1) + V ar(x 2 ) + 2 Cov(X 1, X 2 ) ),

3. Monte Carlo Simulation

IEOR E4602: Quantitative Risk Management

B. Maddah INDE 504 Discrete-Event Simulation. Output Analysis (3)

Chapter 7 - Lecture 1 General concepts and criteria

Simulating Stochastic Differential Equations

Estimating the Greeks

IEOR E4602: Quantitative Risk Management

MAFS Computational Methods for Pricing Structured Products

As we saw in Chapter 12, one of the many uses of Monte Carlo simulation by

Monte-Carlo Methods for Risk Management

Strategies for Improving the Efficiency of Monte-Carlo Methods

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Gamma. The finite-difference formula for gamma is

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

Machine Learning for Quantitative Finance

Chapter 5. Statistical inference for Parametric Models

Statistical analysis and bootstrapping

The Binomial Lattice Model for Stocks: Introduction to Option Pricing

2.1 Mathematical Basis: Risk-Neutral Pricing

Results for option pricing

Chapter 7: Estimation Sections

Importance Sampling for Option Pricing. Steven R. Dunbar. Put Options. Monte Carlo Method. Importance. Sampling. Examples.

The Binomial Lattice Model for Stocks: Introduction to Option Pricing

Asian Option Pricing: Monte Carlo Control Variate. A discrete arithmetic Asian call option has the payoff. S T i N N + 1

Forwards, Swaps, Futures and Options

CPSC 540: Machine Learning

Generating Random Variables and Stochastic Processes

Bias Reduction Using the Bootstrap

Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel

Chapter 8: Sampling distributions of estimators Sections

Equity correlations implied by index options: estimation and model uncertainty analysis

Chapter 7: Estimation Sections

MAFS5250 Computational Methods for Pricing Structured Products Topic 5 - Monte Carlo simulation

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

Econ 300: Quantitative Methods in Economics. 11th Class 10/19/09

Posterior Inference. , where should we start? Consider the following computational procedure: 1. draw samples. 2. convert. 3. compute properties

The histogram should resemble the uniform density, the mean should be close to 0.5, and the standard deviation should be close to 1/ 12 =

MATH4143: Scientific Computations for Finance Applications Final exam Time: 9:00 am - 12:00 noon, April 18, Student Name (print):

2 f. f t S 2. Delta measures the sensitivityof the portfolio value to changes in the price of the underlying

CPSC 540: Machine Learning

Hand and Spreadsheet Simulations

Extend the ideas of Kan and Zhou paper on Optimal Portfolio Construction under parameter uncertainty

Path-dependent inefficient strategies and how to make them efficient.

For these techniques to be efficient, we need to use. M then we introduce techniques to reduce Var[Y ] while. Var[Y ]

Confidence Intervals Introduction

STK 3505/4505: Summary of the course

1.1 Basic Financial Derivatives: Forward Contracts and Options

IEOR E4703: Monte-Carlo Simulation

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

Point Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased.

Computational Finance

Applied Statistics I

The method of Maximum Likelihood.

Introduction Dickey-Fuller Test Option Pricing Bootstrapping. Simulation Methods. Chapter 13 of Chris Brook s Book.

Lecture 1 Definitions from finance

Much of what appears here comes from ideas presented in the book:

8.1 Estimation of the Mean and Proportion

6. Numerical methods for option pricing

NEWCASTLE UNIVERSITY SCHOOL OF MATHEMATICS, STATISTICS & PHYSICS SEMESTER 1 SPECIMEN 2 MAS3904. Stochastic Financial Modelling. Time allowed: 2 hours

DRAFT. 1 exercise in state (S, t), π(s, t) = 0 do not exercise in state (S, t) Review of the Risk Neutral Stock Dynamics

Ch 5. Several Numerical Methods

FREDRIK BAJERS VEJ 7 G 9220 AALBORG ØST Tlf.: URL: Fax: Monte Carlo methods

University of California Berkeley

Chapter 8. Introduction to Statistical Inference

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

The Use of Importance Sampling to Speed Up Stochastic Volatility Simulations

MATH6911: Numerical Methods in Finance. Final exam Time: 2:00pm - 5:00pm, April 11, Student Name (print): Student Signature: Student ID:

Asymptotic methods in risk management. Advances in Financial Mathematics

Write legibly. Unreadable answers are worthless.

Department of Mathematics. Mathematics of Financial Derivatives

may be of interest. That is, the average difference between the estimator and the truth. Estimators with Bias(ˆθ) = 0 are called unbiased.

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by. Cov(X, Y ) = E(X E(X))(Y E(Y ))

10. Monte Carlo Methods

- 1 - **** d(lns) = (µ (1/2)σ 2 )dt + σdw t

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models

Using Monte Carlo Integration and Control Variates to Estimate π

Math 416/516: Stochastic Simulation

From Discrete Time to Continuous Time Modeling

Modeling of Price. Ximing Wu Texas A&M University

PhD Qualifier Examination

Asset Allocation and Risk Assessment with Gross Exposure Constraints

Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMSN50)

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example...

Mathematics of Finance Final Preparation December 19. To be thoroughly prepared for the final exam, you should

King s College London

Stochastic Modelling in Finance

Asymmetric Information: Walrasian Equilibria, and Rational Expectations Equilibria

Simulation Wrap-up, Statistics COS 323

Business Statistics 41000: Probability 3

Transcription:

IEOR E4703: Monte-Carlo Simulation Simulation Efficiency and an Introduction to Variance Reduction Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

Outline Simulation Efficiency Control Variates Multiple Control Variates Antithetic Variates Non-Uniform Antithetic Variates Conditional Monte Carlo 2 (Section 0)

Simulation Efficiency As usual we wish to estimate θ := E[h(X)]. Standard simulation algorithm is: 1. Generate X 1,..., X n 2. Estimate θ with ˆθ n = n j=1 Y j/n where Y j := h(x j ). 3. Approximate 100(1 α)% confidence intervals are then given by [ ] ˆσ n ˆσ ˆθ n z 1 α/2 n, ˆθn + z 1 α/2. n n Can measure quality of ˆθ n by the half-width HW of the CI Var(Y ) HW = z 1 α/2. n Would like HW to be small but sometimes this is difficult to achieve. So often imperative to address the issue of simulation efficiency. There are a number of things we can do: 3 (Section 1)

Simulation Efficiency 1. Develop a good simulation algorithm. 2. Program carefully to minimize storage requirements. e.g. Do not need to store all Y j s. Instead just store Y j and Yj 2 compute ˆθ n and approximate CI s. 3. Program carefully to minimize execution time. 4. Decrease variability of simulation output that we use to estimate θ. Techniques used to do this are called variance reduction techniques. Will now study some of the simplest variance reduction techniques, and assume we are doing items (1) to (3) as well as possible. But first we should first discuss a measure of simulation efficiency. to 4 (Section 1)

Measuring Simulation Efficiency Suppose there are two r.vars, W and Y, such that E[W ] = E[Y ] = θ. Let M w and M y denote methods of estimating θ by simulating the W i s and Y i s, respectively. Question: Which method is more efficient, M w or M y? To answer this, let n w and n y be the number of samples of W and Y, respectively, that are needed to achieve a half-width, HW. Then n w = n y = ( z1 α/2 HW ( z1 α/2 HW ) 2 Var(W ) ) 2 Var(Y ). Let E w and E y denote the amount of computational effort required to produce one sample of W and Y, respectively. 5 (Section 1)

Measuring Simulation Efficiency Then the total effort expended by M w and M y, respectively, to achieve HW are TE w = TE y = ( z1 α/2 HW ( z1 α/2 HW ) 2 Var(W ) Ew ) 2 Var(Y ) Ey. Say M w is more efficient than M y if TE w < TE y. This occurs if and only if Var(W )E w < Var(Y )E y. (1) Will therefore use Var(W )E w as a measure of M w s efficiency of the. Note that (1) implies we cannot conclude that M w is better than M y simply because Var(W ) < Var(Y ). But often the case that E w E y and Var(W ) << Var(Y ). In such cases it is clear that using M w provides a substantial improvement over using M y. 6 (Section 1)

Control Variates We wish to estimate θ := E[Y ] where Y = h(x) is the output of a simulation experiment. Suppose Z is also an output (or that we can easily output it if we wish). Also assume we know E[Z]. Then can construct many unbiased estimators of θ: 1. ˆθ = Y, our usual estimator 2. ˆθc := Y + c(z E[Z]). Variance of ˆθ c satisfies Var(ˆθ c ) = Var(Y ) + c 2 Var(Z) + 2c Cov(Y, Z). (2) Can choose c to minimize this quantity and optimal value given by c = Cov(Y, Z). Var(Z) 7 (Section 2)

Control Variates Minimized variance satisfies Var(ˆθ c ) = Var(Y ) = Var(ˆθ) Cov(Y, Z)2 Var(Z) Cov(Y, Z)2. Var(Z) In order to achieve a variance reduction therefore only necessary that Cov(Y, Z) 0. New resulting Monte Carlo algorithm proceeds by generating n samples of Y and Z and then setting n i=1 ˆθ c = (Y i + c (Z i E[Z])). n There is a problem with this, however, as we usually do not know Cov(Y, Z). Resolve this problem by doing p pilot simulations and setting p j=1 Ĉov(Y, Z) = (Y j Y p )(Z j E[Z]). p 1 8 (Section 2)

Control Variates If it is also the case that Var(Z) unknown, then can also estimate it with p j=1 (Z j E[Z]) 2 Var(Z) = p 1 and finally set ĉ =, Z) Ĉov(Y. Var(Z) Our control variate simulation algorithm is as follows: 9 (Section 2)

Control Variate Simulation Algorithm for Estimating E[Y ] / Do pilot simulation first / for i = 1 to p generate (Y i, Z i ) end for compute ĉ / Now do main simulation / for i = 1 to n generate (Y i, Z i ) set V i = Y i + ĉ (Z i E[Z]) end for set ˆθĉ = V n = n i=1 V i/n set ˆσ n,v 2 = (V i ˆθĉ ) 2 /(n 1) set 100(1 α)% [ ] ˆσ n,v ˆσ CI = ˆθĉ z 1 α/2 n,v, ˆθĉ + z 1 α/2 n n

e.g. Pricing an Asian Call Option Payoff of an Asian call option given by ( m i=1 h(x) := max 0, S it/m m ) K where X := {S it/m : i = 1,..., m}, K is the strike and T the expiration date. Price of option then given by C a = E Q 0 [e rt h(x)]. Will assume as usual that S t GBM (r, σ) under Q. Usual simulation method for estimating C a is to generate n independent samples of the payoff, Y i := e rt h(x i ), for i = 1,..., n, and to set n i=1 Ĉ a = Y i. n 11 (Section 2)

e.g. Pricing an Asian Call Option But could also estimate C a using control variates and there are many possible choices: 1. Z 1 = S T 2. Z 2 = e rt max(0, S T K) 3. Z 3 = m j=1 S it/m/m In each of the three cases, it is easy to compute E[Z]. Would also expect Z 1, Z 2 and Z 3 to have a positive covariance with Y, so that each one would be a suitable candidate for use as a control variate. Question: Which one would lead to the greatest variance reduction? Question: Explain why you could also use the value of the option with payoff ( m ) 1/m g(x) := max 0, S it/m K as a control variate. Do you think it would result in a substantial variance reduction? i=1 12 (Section 2)

e.g. The Barbershop A barbershop opens for business every day at 9am and closes at 6pm. There is only 1 barber but he s considering hiring another one. But first he would like to estimate the mean total time that customers spend waiting each day. Assume customers arrive at barbershop according to a non-homogeneous Poisson process, N (t), with intensity λ(t). Let W i denote waiting time of i th customer. The barber closes the shop after T = 9 hours but still serves any customers who have arrived before then. Quantity that he wants to estimate is θ := E[Y ] where Y := N(T) j=1 W j. 13 (Section 2)

e.g. The Barbershop Assume the service times of customers are IID with CDF, F(.), and that they are also independent of the arrival process, N (t). Usual simulation algorithm: simulate n days of operation in the barbershop, thereby obtaining n samples, Y 1,..., Y n, and then set n j=1 ˆθ n = Y j. n However, a better estimate could be obtained by using a control variate. Let Z denote the total time customers on a given day spend in service so that Z := N(T) where S j is the service time of the j th customer. Then easy to see that j=1 S j E[Z] = E[S]E[N (T)]. Intuition suggests that Z would be a good candidate to use as a control variate. 14 (Section 2)

Multiple Control Variates No reason why we should not use more than one control variate. So suppose again that we wish to estimate θ := E[Y ] where Y is the output of a simulation experiment. Also suppose that for i = 1,..., m, Z i is an output or that we can easily output it if we wish, and that E[Z i ] is known for each i. Can then construct unbiased estimators of θ by defining ˆθ c := Y + c 1 (Z 1 E[Z 1 ]) +... + c m (Z m E[Z m ]). Variance of ˆθ c satisfies m m m Var(ˆθ c ) = Var(Y ) + 2 c i Cov(Y, Z i ) + c i c j Cov(Z i, Z j ) (3) Can easily minimize Var(ˆθ c ) w.r.t the c i s. i=1 i=1 i=1 15 (Section 2)

Multiple Control Variates As before, however, optimal solutions ci will involve unknown covariances (and possibly variances of the Z i s) that will need to be estimated using a pilot simulation. A convenient way of doing this is to observe that ĉ i = b i where the b i s are the least squares solution to the following linear regression: Y = a + b 1 Z 1 +... + b m Z m + ɛ. (4) The simulation algorithm with multiple control variates is exactly analogous to the simulation algorithm with just a single control variate. 16 (Section 2)

Antithetic Variates As usual would like to estimate θ = E[h(X)] = E[Y ]. Suppose we have generated two samples, Y 1 and Y 2. Then an unbiased estimate of θ is given by with ˆθ := (Y 1 + Y 2 )/2 Var(ˆθ) = Var(Y 1) + Var(Y 2 ) + 2Cov(Y 1, Y 2 ). 4 If Y 1 and Y 2 are IID, then Var(ˆθ) = Var(Y )/2. However, we could reduce Var(ˆθ) if we could arrange it so that Cov(Y 1, Y 2 ) < 0. We now describe the method of antithetic variates for doing this. We will begin with the case where Y is a function of IID U (0, 1) random variables so that θ = E[h(U)] where U = (U 1,..., U m ) and the U i s are IID U (0, 1). 17 (Section 3)

Usual Simulation Algorithm for Estimating θ (with 2n Samples) for i = 1 to 2n generate U i set Y i = h(u i ) end for set ˆθ 2n = Y 2n = 2n i=1 Y i/2n set ˆσ 2 2n = 2n i=1 (Y i Y 2n ) 2 /(2n 1) set approx. 100(1 α) % CI = ˆθ 2n ± z 1 α/2 ˆσ 2n

Antithetic Variates In above algorithm, however, could also have used the 1 U i s to generate sample Y values. Can use this fact to construct another estimator of θ as follows: 1. As before, set Y i = h(u i ), where U i = (U (i) 1,..., U (i) m ). 2. Also set Ỹi = h(1 U i ), where we use 1 U i = (1 U (i) 1,..., 1 U (i) m ). 3. Set Z i := (Y i + Ỹi)/2. Note that E[Z i ] = θ so Z i also unbiased estimator of θ. If the U i s are IID, then so too are the Z i s and can use them as usual to compute approximate CI s for θ. We say that U i and 1 U i are antithetic variates. Have the following antithetic variate simulation algorithm. 19 (Section 3)

Antithetic Variate Simulation Algorithm for Estimating θ for i = 1 to n generate U i set Y i = h(u i ) and Ỹi = h(1 U i ) set Z i = (Y i + Ỹi)/2 end for set ˆθ n,a = Z n = n i=1 Z i/n set ˆσ 2 n,a = n i=1 (Z i Z n ) 2 /(n 1) set approx. 100(1 α) % CI = ˆθ a,n ± z 1 α/2 ˆσ n,a n ˆθ a,n is unbiased and SLLN implies ˆθ n,a θ w.p. 1 as n. Each of the two algorithms uses 2n samples so question arises as to which algorithm is better? Both algorithms require approximately same amount of effort so comparing the two algorithms amounts to computing which estimator has a smaller variance.

Comparing Estimator Variances Easy to see that Var(ˆθ 2n ) = Var ( 2n ) i=1 Y i = Var(Y ) 2n 2n and ( n i=1 Var(ˆθ n,a ) = Var Z ) i n = Var(Z) n = Var(Y + Ỹ ) 4n = Var(Y ) 2n = Var(ˆθ 2n ) + Cov(Y, Ỹ ). 2n Therefore Var(ˆθ n,a ) < Var(ˆθ 2n ) if and only Cov(Y, Ỹ ) < 0. + Cov(Y, Ỹ ) 2n Recalling that Y = h(u) and Ỹ = h(1 U), this means that Var(ˆθ n,a ) < Var(ˆθ 2n ) Cov (h(u), h(1 U)) < 0. 21 (Section 3)

When Can a Variance Reduction Be Guaranteed? Consider first the case where U is a scalar uniform so m = 1, U = U and θ = E[h(U )]. Suppose h(.) is a non-decreasing function of u over [0, 1]. Then if U is large, h(u ) will also tend to be large while 1 U and h(1 U ) will tend to be small. That is, Cov(h(U ), h(1 U )) < 0. Can similarly conclude that if h(.) is a non-increasing function of u then once again, Cov(h(U ), h(1 U )) < 0. So for the case where m = 1, a sufficient condition to guarantee a variance reduction is for h(.) to be a monotonic function of u on [0, 1]. Consider the more general case where m > 1, U = (U 1,..., U m ) and θ = E[h(U)]. Say h(u 1,..., u m ) is a monotonic function of each of its m arguments if, in each of its arguments, it is non-increasing or non-decreasing. 22 (Section 3)

Comparing Estimator Variances Theorem. If h(u 1,..., u m ) is a monotonic function of each of its arguments on [0, 1] m, then for a set U := (U 1,..., U m ) of IID U (0, 1) random variables Cov(h(U), h(1 U)) < 0 where Cov(h(U), h(1 U)) := Cov(h(U 1,..., U m ), h(1 U 1,..., 1 U m )). Proof See Sheldon M. Ross s Simulation. Note that the theorem specifies sufficient conditions for a variance reduction, but not necessary conditions. So still possible to obtain a variance reduction even if conditions of the theorem are not satisfied. For example, if h(.) is mostly monotonic, then a variance reduction might be still be obtained. 23 (Section 3)

Non-Uniform Antithetic Variates So far have only considered problems where θ = E[h(U)], for U a vector of IID U (0, 1) random variables. But often the case that θ = E[Y ] where Y = h(x 1,..., X m ), and where (X 1,..., X m ) is a vector of independent random variables. Can still use antithetic variable method for such problems if we can use the inverse transform method to generate the X i s. To see this, suppose F i (.) is the CDF of X i. If U i U (0, 1) then Fi 1 (U i ) has the same distribution as X i. So can generate Y by generating U 1,..., U m IID U (0, 1) and setting Z = h ( F 1 1 (U 1),..., F 1 m (U m ) ). Since the CDF of any random variable is non-decreasing, it follows that F 1 i (.) also non-decreasing. So if h(x 1,..., x m ) monotonic in each of its arguments, then h(f1 1 (U 1),..., Fm 1 (U m )) also monotonic function of the U i s. 24 (Section 3)

The Barbershop Revisited Consider again our barbershop example and suppose the barber now wants to estimate the average total waiting time, θ, of the first 100 customers. Then θ = E[Y ] where Y = 100 j=1 W j and where W j is the waiting time of the j th customer. For each customer, j, there is an inter-arrival time, I j = time between the (j 1) th and j th arrivals. There is also a service time, S j = amount of time the barber spends cutting the j th customer s hair. Therefore there is some function, h(.), for which Y = h(i 1,..., I 100, S 1,..., S 100 ). For many queueing systems, h(.) will be a monotonic function of its arguments. Why? Antithetic variates guaranteed to give a variance reduction in these systems. 25 (Section 3)

Normal Antithetic Variates Can also generate antithetic normal random variates without using the inverse transform technique. For if X N(µ, σ 2 ) then X N(µ, σ 2 ) also, where X := 2µ X. Clearly X and X are negatively correlated. So if θ = E[h(X 1,..., X m )] where the X i s are IID N(µ, σ 2 ) and h(.) is monotonic in its arguments, then we can again achieve a variance reduction by using antithetic variates. 26 (Section 3)

e.g. Normal Antithetic Variates Suppose we want to estimate θ = E[X 2 ] where X N(2, 1). Then easy to see that θ = 5, but can also estimate it using antithetic variates. Question: Is a variance reduction guaranteed? Why or why not? Question: What would you expect if Z N(10, 1)? Question: What would you expect if Z N(0.5, 1)? 27 (Section 3)

e.g. Estimating the Price of a Knock-In Option Wish to price a knock-in option where the payoff is given by h(s T ) = max(0, S T K)I {ST >B} where B is some fixed constant and S t GBM (r, σ 2 ) under Q. Option price may be then written as C 0 = E Q 0 [e rt max(0, S T K)I {ST >B}] Can write S T = S 0 e (r σ2 /2)T+σ TX where X N(0, 1) so we can use antithetic variates to estimate C 0. Question: Are we sure to get a variance reduction? Worth emphasizing that the variance reduction that can be achieved through the use of antithetic variates is rarely (if ever!) dramatic. 28 (Section 3)

Conditional Monte Carlo Let X and Z be random vectors, and let Y = h(x) be a random variable. Suppose we set V = E[Y Z]. Then V is itself a random variable that depends on Z, so can write V = g(z) for some function, g( ). Also know that E[V ] = E[E[Y Z]] = E[Y ] so if we are trying to estimate θ = E[Y ], could simulate V instead of Y. To determine the better estimator we compare variances of Y and V = E[Y Z]. To do this, recall the conditional variance formula: Var(Y ) = E[Var(Y Z)] + Var(E[Y Z]). (5) 29 (Section 4)

Conditional Monte Carlo Must have (why?) E[Var(Y Z)] 0 but then (5) implies Var(Y ) Var(E[Y Z]) = Var(V ) so we can conclude (can we?!) that V is a better estimator of θ than Y. To see this from another perspective, suppose that to estimate θ we first have to simulate Z and then simulate Y given Z. If we can compute E[Y Z] exactly, then we can eliminate the additional noise that comes from simulating Y given Z, thereby obtaining a variance reduction. Question: Why must Y and Z be dependent for the conditional Monte Carlo method to be worthwhile? 30 (Section 4)

Conditional Monte Carlo Summarizing, we want to estimate θ := E[h(X)] = E[Y ] using conditional MC. To do so, must have another variable or vector, Z, that satisfies: 1. Z can be easily simulated 2. V := g(z) := E[Y Z] can be computed exactly. If these two conditions satisfied then can simulate V by first simulating Z and then setting V = g(z) = E[Y Z]. Question: It may be possible to identify the distribution of V = g(z). What might we do in that case? Also possible that other variance reduction methods could be used in conjunction with conditioning. e.g. If g(.) a monotonic function of its arguments, then antithetic variates might be useful. 31 (Section 4)

Conditional Monte Carlo Algorithm for Estimating θ for i = 1 to n generate Z i compute g(z i ) = E[Y Z i ] set V i = g(z i ) end for set ˆθ n,cm = V n = n i=1 V i/n set ˆσ 2 n,cm = n i=1 (V i V n ) 2 /(n 1) set approx. 100(1 α) % CI = ˆθ ˆσ n,cm ± z n,cm 1 α/2 n

An Example of Conditional Monte Carlo We wish to estimate θ := P(U + Z > 4) where U Exp(1) and Z Exp(1/2). Let Y := I {U+Z>4} then θ = E[Y ] and can use conditional MC as follows. Set V = E[Y Z] so that E[Y Z = z] = P(U + Z > 4 Z = z) = P(U > 4 z) = 1 F u (4 z) where F u (.) is the CDF of U. Therefore which implies 1 F u (4 z) = { e (4 z) if 0 z 4, 1 if z > 4. V = E[Y Z] = { e (4 Z) if 0 Z 4, 1 if Z > 4. 33 (Section 4)

An Example of Conditional Monte Carlo Now the conditional Monte Carlo algorithm for estimating θ = E[V ] is: 1. Generate Z 1,..., Z n all independent 2. Set V i = E[Y Z i ] for i = 1,..., n 3. Set ˆθ n,cm = n i=1 V i/n 4. Compute approximate CI s as usual using the V i s. Could also use other variance reduction methods in conjunction with conditioning. 34 (Section 4)

Pricing a Barrier Option Definition. Let c(x, t, K, r, σ) be the Black-Scholes price of a European call option when the current stock price is x, the time to maturity is t, the strike is K, the risk-free interest rate is r and the volatility is σ. Want to estimate price of a European option with payoff { max(0, ST K h(x) = 1 ) if S T/2 L, max(0, S T K 2 ) otherwise. where X = (S T/2, S T ). Can write option price as [ ( )] C 0 = E Q 0 e rt max(0, S T K 1 )I {ST/2 L} + max(0, S T K 2 )I {ST/2 >L} where as usual S t GBM (r, σ 2 ) under Q. Question: How would you estimate C 0 using simulation and only one normal random variable per sample payoff? Question: Could you use antithetic variates as well? Would they be guaranteed to produce a variance reduction? 35 (Section 4)

An Exercise: Estimating Portfolio Credit Risk A bank has a portfolio of N = 100 loans to N companies and wants to evaluate its credit risk. Given that company n defaults, the loss for the bank is a N(µ, σ 2 ) random variable X n where µ = 3, σ 2 = 1. Defaults are dependent and described by indicators D 1,..., D N and a background random variable P, such that D 1,..., D N are IID Bernouilli(p) given P = p. P has a Beta(1, 19) distribution, i.e. P has density (1 p) 18 /19, 0 < p < 1. How would you estimate P(L > x), where L = N n=1 D nx n is the loss, using conditional Monte Carlo, where the conditioning is on N n=1 D n? 36 (Section 4)