Importance sampling and Monte Carlo-based calibration for time-changed Lévy processes Stefan Kassberger Thomas Liebmann BFS 2010
1 Motivation 2 Time-changed Lévy-models and Esscher transforms 3 Applications and examples Importance sampling Sensitivities w.r.t. Esscher transform parameters Simulation of a Normal Tempered Stable process Calibration Example 4 Summary
Motivation Variety of tractable Lévy-models can be represented as time-changed Brownian motion Esscher-transform well-established for Lévy-models Kassberger and Liebmann (2009) apply independent Esscher transforms to Brownian motion and subordinator in the TCL-context Exploit above idea in the context of Monte Carlo simulation: Variance reduction through importance sampling Calculating sensitivities by likelihood-ratio methods Extending sampling algorithms to more general classes of distributions Monte-Carlo based model calibration
Setup n-dimensional Brownian motion (B t ), subordinator (T (t)). Fix τ and denote by κ(u) = loge Q [exp(ut (τ))] the cumulant generating function of T (τ). Define equivalent measure S η,γ via ds η,γ dq = exp ( η B T (τ) 1 ) 2 η 2 T (τ) + γt (τ) κ(γ). This transform is composed of two Esscher transforms: 1 One with parameter η R n on the Brownian motion B. Shifts the drift of B from 0 under Q to η under S η,γ. 2 One with parameter γ on the subordinator. Its cgf under S η,γ is given by κ S η,γ (u) = κ(u + γ) κ(γ).
Outline 1 Motivation 2 Time-changed Lévy-models and Esscher transforms 3 Applications and examples Importance sampling Sensitivities w.r.t. Esscher transform parameters Simulation of a Normal Tempered Stable process Calibration Example 4 Summary
Importance sampling Set X(t) = B T (t) + mt (t). Take expectations under the transformed measure: [ E Q f((x(t))t [0,τ] ) ] ( = E Sη,γ [exp η B T (τ) + 1 ) ] 2 η 2 T (τ) γt (τ) + κ(γ) f((x(t)) t [0,τ] ). If the subordinator cannot be efficiently sampled under Q but under S η,γ for certain values of γ, we can choose η for variance reduction via importance sampling (compare example for Normal Tempered Stable process).
Importance sampling Goal: Estimate ( E Sη,γ [exp η B T (τ) + 1 ) ] 2 η 2 T (τ) γt (τ) + κ(γ) f((x(t)) t [0,τ] ) (1) by simulation. Simulate appropriately discretized version of (X(t)) : (X(t i )) i=0,...,n with 0 = t 0 < < t N = τ. Proceed as follows: Simulate (T (t i )) i=0,...,n and iid N n (0,diag(1 n )) rvs W 1,...,W N. Set Y (t i ) = Y (t i 1 ) + W i T (ti ) T (t i 1 ) + η(t i t i 1 ). Set X(t i ) = Y (t i ) + mt (t i ). Repeat M times to arrive at set of sample paths indexed by k. Estimator for (1) is ( M 1 exp η Y k (τ) + 1 ) M k=1 2 η 2 T k (τ) γt k (τ) + κ(γ) f((x k (t i )) i=0,...,n ).
Importance sampling: Variance of the Estimator Let f(x) ce β x for some c 0 and β R n. Bound for variance of the estimator [ ( M 1 Var Sη,γ exp η Y k (τ) + 1 ) ] M k=1 2 η 2 T k (τ) γt k (τ) + κ(γ) f(x k (τ)) c2 M exp( κ( β 2 + β η 2 γ + 2β m) + κ(γ) ) κ is increasing and convex. For given β, c and γ, the bound is minimal if η = β. If γ can be chosen freely, minimum is attained for γ = β 2 /2 + β m. If further m = 0, this holds for γ = β 2 /2 = η 2 /2, i.e. an Esscher transform on B T (τ) with parameter β.
Outline 1 Motivation 2 Time-changed Lévy-models and Esscher transforms 3 Applications and examples Importance sampling Sensitivities w.r.t. Esscher transform parameters Simulation of a Normal Tempered Stable process Calibration Example 4 Summary
Sensitivities with the LR-method Sensitivities w.r.t. Esscher transform parameters η and γ can be estimated by the likelihood ratio method If the expectations are finite in some neighborhood of γ 0, d dγ E [ S η0,γ f((x(t))t [0,τ] ) ] γ=γ0 = E Sη,γ [ dsη0,γ 0 dq dq ( T (τ) κ (γ 0 ) ) ] f((x(t)) t [0,τ] ) ds η,γ If the expectations are finite in some neighborhood of η 0, d dη E S η,γ0 [ f((x(t))t [0,τ] ) ] η=η0 = E Sη,γ [ dsη0,γ 0 dq dq ds η,γ ( BT (τ) η 0 T (τ) ) f((x(t)) t [0,τ] ) ]
Outline 1 Motivation 2 Time-changed Lévy-models and Esscher transforms 3 Applications and examples Importance sampling Sensitivities w.r.t. Esscher transform parameters Simulation of a Normal Tempered Stable process Calibration Example 4 Summary
Simulation of a Normal Tempered Stable process Let T be a Tempered Stable TS(κ,a,(2γ) κ ) random variable with κ (0,1), a,γ > 0 and cumulant generating function κ κ,a,γ (u) = a(2γ) κ a(2γ 2u) κ. Algorithm that allows direct sampling is unknown Idea: Simulate stable random variables and perform Esscher transform Algorithm: 1 Sample n iid stable TS(κ,a,0) rvs T k 2 Set importance weights to w k = exp( γt k + κ κ,a,γ (γ)) = exp( γt k + a(2γ) κ ) In Normal Tempered Stable model, η still available for IS
Outline 1 Motivation 2 Time-changed Lévy-models and Esscher transforms 3 Applications and examples Importance sampling Sensitivities w.r.t. Esscher transform parameters Simulation of a Normal Tempered Stable process Calibration Example 4 Summary
MC-based model calibration MC-based model calibration time-consuming and unstable Change of parameters often requires re-simulation Noise due to finite sample size often precludes gradient-based methods and slows down convergence of optimizer Idea: Use of two Esscher transforms often allows calibration based on a single set of paths or at least helps significantly reduce number of simulation runs Simulate BM and subordinator, keep paths in memory, and then apply Esscher transforms in conjunction with basic transformation (shifting & scaling) Facilitates use of gradient-based optimization algorithms (more exact & stable gradients)
Outline 1 Motivation 2 Time-changed Lévy-models and Esscher transforms 3 Applications and examples Importance sampling Sensitivities w.r.t. Esscher transform parameters Simulation of a Normal Tempered Stable process Calibration Example 4 Summary
Example: MC-calibration of NIG model Objective: Calibrate model based on NIG(α,β,δ,µ) paths, i.e. find α,β,δ,µ such that observed derivatives prices are replicated Proceed as follows (easily generalizes to appropriate time-scaling): 1 Fix c. Simulate IG(1,c) paths (IG t ) and standard Brownian paths (B t ) 2 Fix α,β,δ,µ 3 Calculate b = δ α 2 β 2 and choose γ such that c 2 2γ = b. 4 Calculate Esscher weights corresponding to γ. 5 Under S 0,γ, X = βδ 2 IG 1 + δ B IG1 + µ is NIG(α,β,δ,µ)-distributed. 6 Calculate option prices / evaluate objective function under S 0,γ, 7 Repeat steps 2-7 if necessary η (unused above) can be employed for importance sampling.
Example: Pricing a binary down-and-out call via MC Let the log-return process (X t ) be a Lévy-process with X 1 NIG(α,0,δ,µ) with µ = µ(α,δ) chosen such that (S 0 exp(x t )) is a martingale (assume riskless interest rate r = 0 and S 0 = 100). Price a binary down-and-out call (BDOC) in this model via Esscher-based MC. Payoff of BDOC maturing in 1 year: I(min(S t,t [0,1]) 80) I(S 1 100) Price BDOC via MC: Discretize using 250 time-steps. Simulate 100000 Brownian and IG paths. Use Esscher-based MC: Smooth dependence of BDOC-price on parameters (α, δ). Fast calculation of option price surface as only one simulation run is needed.
BDOC price surface BDOC price 0.4 0.35 0.3 0.25 0.2 0.15 0.1 0.05 0 0.25 0.5 0.75 1 1.25 δ 1.5 1 α 1.5 2 Figure: BDOC price as function of α and δ
Example: Calculating sensitivities w.r.t. model parameters Sensitivities based on plain MC notoriously noisy Esscher-based MC makes sensitivity estimates more stable and provides faster convergence In above setup, calculate sensitivity of BDOC-price w.r.t. model parameter δ for α 0 = 2 and δ 0 = 1 First method: Finite differences with resampling (same seed for random number generator in both runs) Second method: Finite differences with Esscher (only one sample plus Esscher transform)
Convergence of parameter sensitivities 0.4 0.3 0.2 0.1 Finite differences (Esscher) Finite differences (resampling) 0 1.0E+3 2.0E+3 4.1E+3 8.2E+3 1.6E+4 3.3E+4 6.6E+4 1.3E+5 2.6E+5 5.2E+5 1.0E+6-0.1-0.2-0.3-0.4-0.5 Figure: Sensitivities w.r.t. δ as function of sample size
Summary Apply independent Esscher transforms to subordinator and BM Gives higher degree of flexibility than Esscher transform applied to TCBM itself Subclass of structure preserving transforms (Kassberger and Liebmann (2009)) Variety of applications: 1 IS with upper bounds for variance 2 IS with a mixture of importance distributions 3 sensitivities w.r.t. Esscher parameters via LR-methods 4 development of sampling algorithms 5 model calibration Approach also works for subordinated stable processes
Thanks for your attention!