Stratified Sampling in Monte Carlo Simulation: Motivation, Design, and Sampling Error

Similar documents
Lecture 22. Survey Sampling: an Overview

Lecture Neyman Allocation vs Proportional Allocation and Stratified Random Sampling vs Simple Random Sampling

Chapter 7. Sampling Distributions and the Central Limit Theorem

Chapter 7. Sampling Distributions and the Central Limit Theorem

Chapter 2 Uncertainty Analysis and Sampling Techniques

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals

IEOR E4703: Monte-Carlo Simulation

The Binomial Lattice Model for Stocks: Introduction to Option Pricing

8.1 Estimation of the Mean and Proportion

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

χ 2 distributions and confidence intervals for population variance

The Binomial Lattice Model for Stocks: Introduction to Option Pricing

Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMSN50)

Strategies for Improving the Efficiency of Monte-Carlo Methods

Math Computational Finance Double barrier option pricing using Quasi Monte Carlo and Brownian Bridge methods

EE266 Homework 5 Solutions

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Uniform Probability Distribution. Continuous Random Variables &

Summary Sampling Techniques

Risk Measurement in Credit Portfolio Models

STA Module 3B Discrete Random Variables

Simulation Wrap-up, Statistics COS 323

Much of what appears here comes from ideas presented in the book:

Tutorial 6. Sampling Distribution. ENGG2450A Tutors. 27 February The Chinese University of Hong Kong 1/6

Gamma. The finite-difference formula for gamma is

University of California Berkeley

Scenario Generation and Sampling Methods

6 Central Limit Theorem. (Chs 6.4, 6.5)

5.3 Statistics and Their Distributions

CPSC 540: Machine Learning

Homework Assignments

Introduction to Sequential Monte Carlo Methods

Chapter 14 : Statistical Inference 1. Note : Here the 4-th and 5-th editions of the text have different chapters, but the material is the same.

Efficient Quantile Estimation When Applying Stratified Sampling and Conditional. Monte Carlo, With Applications to Nuclear Safety

Discrete Random Variables

Chapter 7: Point Estimation and Sampling Distributions

Binomial Random Variables. Binomial Random Variables

MAFS Computational Methods for Pricing Structured Products

Sampling and sampling distribution

Lecture outline. Monte Carlo Methods for Uncertainty Quantification. Importance Sampling. Importance Sampling

Slides for Risk Management

Ch4. Variance Reduction Techniques

IEOR E4703: Monte-Carlo Simulation

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

CPSC 540: Machine Learning

Distribution of the Sample Mean

Handout 8: Introduction to Stochastic Dynamic Programming. 2 Examples of Stochastic Dynamic Programming Problems

C.10 Exercises. Y* =!1 + Yz

A probability distribution shows the possible outcomes of an experiment and the probability of each of these outcomes.

ECON 214 Elements of Statistics for Economists 2016/2017

Statistics for Business and Economics

UNIVERSITY OF VICTORIA Midterm June 2014 Solutions

Probabilistic Analysis of the Economic Impact of Earthquake Prediction Systems

Statistics 13 Elementary Statistics

Examples: Random Variables. Discrete and Continuous Random Variables. Probability Distributions

Department of Mathematics. Mathematics of Financial Derivatives

MLLunsford 1. Activity: Central Limit Theorem Theory and Computations

Math Computational Finance Option pricing using Brownian bridge and Stratified samlping

II. Random Variables

MORE DATA OR BETTER DATA? A Statistical Decision Problem. Jeff Dominitz Resolution Economics. and. Charles F. Manski Northwestern University

STA Rev. F Learning Objectives. What is a Random Variable? Module 5 Discrete Random Variables

Week 1 Quantitative Analysis of Financial Markets Distributions B

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

Confidence Intervals Introduction

Random Variables and Probability Functions

Using Monte Carlo Integration and Control Variates to Estimate π

MATH 3200 Exam 3 Dr. Syring

Characterization of the Optimum

Portfolio Risk Management and Linear Factor Models

Lecture 3: Return vs Risk: Mean-Variance Analysis

SIMULATION OF ELECTRICITY MARKETS

Stochastic Differential Equations in Finance and Monte Carlo Simulations

On Complexity of Multistage Stochastic Programs

Statistics for Business and Economics

Estimating parameters 5.3 Confidence Intervals 5.4 Sample Variance

B. Consider the problem of evaluating the one dimensional integral

Lecture 9 - Sampling Distributions and the CLT

Commonly Used Distributions

Lecture 4: Return vs Risk: Mean-Variance Analysis

Probability. An intro for calculus students P= Figure 1: A normal integral

IEOR E4602: Quantitative Risk Management

Module 3: Sampling Distributions and the CLT Statistics (OA3102)

ADVANCED OPERATIONAL RISK MODELLING IN BANKS AND INSURANCE COMPANIES

Time Observations Time Period, t

Sampling Distributions

King s College London

MEMORIAL UNIVERSITY OF NEWFOUNDLAND DEPARTMENT OF MATHEMATICS AND STATISTICS MIDTERM EXAM - STATISTICS FALL 2014, SECTION 005

Option Pricing. Chapter Discrete Time

Unobserved Heterogeneity Revisited

5.3 Interval Estimation

Introduction to Algorithmic Trading Strategies Lecture 8

Determining Sample Size. Slide 1 ˆ ˆ. p q n E = z α / 2. (solve for n by algebra) n = E 2

Statistical analysis and bootstrapping

Sampling Distribution

MAS187/AEF258. University of Newcastle upon Tyne

MEASURING PORTFOLIO RISKS USING CONDITIONAL COPULA-AR-GARCH MODEL

Chapter 5. Statistical inference for Parametric Models

MATH4143: Scientific Computations for Finance Applications Final exam Time: 9:00 am - 12:00 noon, April 18, Student Name (print):

Monte Carlo Methods in Option Pricing. UiO-STK4510 Autumn 2015

Transcription:

South Texas Project Risk- Informed GSI- 191 Evaluation Stratified Sampling in Monte Carlo Simulation: Motivation, Design, and Sampling Error Document: STP- RIGSI191- ARAI.03 Revision: 1 Date: September 13, 2013 Prepared by: David Morton, The University of Texas at Austin Jeremy Tejada, The University of Texas at Austin Alexander Zolan, The University of Texas at Austin Pending Review by: Ernie J. Kee, South Texas Project Zahra Mohaghegh, University of Illinois at Urbana- Champaign Seyed A. Reihani, University of Illinois at Urbana- Champaign

Stratified Sampling in Monte Carlo Simulation: Motivation, Design, and Sampling Error David Morton, Jeremy Tejada, and Alexander Zolan The University of Texas at Austin Abstract We describe a stratified estimator in Monte Carlo simulation. We compare it to the standard sample mean estimator that arises from naive Monte Carlo sampling. We characterize the sampling error associated with a stratified estimator. And, we discuss how to design the stratified sampling estimator, in term of choosing the strata and allocating sample sizes, to reduce sampling error over that of the naive approach. 1 Overview Let Y denote a random variable, and let µ = EY denote a performance measure associated with a simulation model. For example, we could have X denote the (random) lifetime of a system, and Y = I(X t 0 ) denote the binary random variable, indicating whether the system lifetime exceeds a critical time threshold, t 0. Then µ = EY = P(X t 0 ) denotes the reliability of the system; i.e., the probability the system does not fail by time t 0. We assume that we cannot compute µ exactly. Instead we estimate µ by Monte Carlo sampling. We describe two Monte Carlo schemes and compare their relative merits. None of what we describe below requires that Y be a binary variable, but our results do require finite variance, σ 2 = var Y <, because our confidence interval statements rely on central limit theorems, which use assume finite variance. The ideas we present here are not new. For example, see Section 5.3 of Hammersley and Handscomb [2] and Section 5.7 of Asmussen and Glynn [1]. 2 Naive Monte Carlo Sampling Suppose we have n independent and identically distributed (iid) observations of Y, which we denote Y 1, Y 2,..., Y n. Typically to construct iid observations of Y in this way we run a simulation model, which may be computationally expensive and is based on a potentially large number of underlying variables. That is, Y = f(x 1, X 2,..., X d ), where (X 1, X 2,..., X d ) is the d-dimensional random inputs to the simulation model. Furthermore, it is possible that each Y is itself an average of further underlying batches. 1

Given the n iid observations, we can form the sample mean and sample variance: Ȳ n = 1 n S 2 n = n 1 n 1 Y i n (1a) ( ) 2 Yi Ȳn. (1b) Using the equations of (1) we can form a 100(1 α)% confidence interval: [Ȳn t n 1,α S n / n, Ȳ n + t n 1,α S n / n ], (2) where t n 1,α satisfies P( t n 1,α T n 1 t n 1,α ) = 1 α and T n 1 is a Student s t random variable with n 1 degrees of freedom. A typical value of α is 0.10 so that we obtain a 90% confidence interval. If α = 0.10 and n = 100 then t 99,0.10 = 1.660. For smaller values of n the Student s t quantile is larger; e.g., with n = 5 we have t 4,0.10 = 2.132. We know var Ȳn = σ 2 /n, where again, σ 2 = var Y. And, given the formulas for the sample mean in equation (1a) and the sample variance Sn 2 in equation (1b) we have: EȲn = µ (3a) ES 2 n = σ 2. (3b) Restated, both the sample mean and the sample variance are unbiased estimators of their population counterparts. 3 Stratified Sampling We describe stratified sampling for the case in which Y = f(x), and we stratify on X. That said, our more general setting of Y = f(x 1, X 2,..., X d ) still applies provided: (i) X corresponds to X 1 ; (ii) X 1 and (X 2,..., X d ) are independent; and, Y = E X2,...,X d f(x 1, X 2,..., X d ). In the context of GSI-191 and the CASA Grande simulation model, we can think of X as, say, the random initiating frequency or the random break size, or rather, as a randomly selected quantile associated with those random variables. Let S 1, S 2,..., S K denote a partition of X s support; i.e., P(X K S k) = 1 and P(S i S j ) = for i j. We have EY = Ef(X) = P(X S k ) E [f(x) X S }{{} k ]. p k 2

We assume that the probability mass, p k, of each cell (stratum) is known, and we let F X (x X S k ) denote the conditional probability distribution of X, given that X lies in cell k. Let X k 1, Xk 2,..., Xk be iid observations from F X (x X S k ), i.e., iid observations of X, given that X lies in the k-th cell. Then we can form a stratified estimator, which is the analog of equation (1a): where K = n. f n = [ ] 1 p k f(xi k ), (4) } {{ } f k While equation (4) yields the point estimate for the stratified sampling procedure, it remains to characterize its sampling error. We have: var f n = p 2 k var } [f(x) X {{ S k] /n } k. (5) σk 2 Let S 2 k, denote the sample variance for the k-th cell, i.e., and hence S 2 k, = 1 1 var f n = ( f(x k i ) f k ) 2, p 2 k S 2 k,. (6) Using equations (4) and (6) we can form a 100(1 α)% confidence interval: [ f n t n 1,α var f n, fn + t n 1,α var f n ]. (7) In similar fashion to equations (3), we have: 4 Designing a Stratified Sampling Procedure E f n = µ (8a) E var f n = var f n. (8b) Naive sampling and stratified sampling both yield an unbiased point estimate of µ. The motivation for using stratified sampling over the simpler naive sampling alternative is the hope that: var f n var Ȳn. 3

So, in designing a stratified sampling procedure, we seek to reduce variance. And we have two key choices to make in designing a stratified sampling procedure: 1. How should we choose the cells of the stratification, S 1, S 2,..., S K? 2. How should we choose the sample sizes, n 1, n 2,..., n K? We begin with the latter question. Assuming that the probability masses, p 1, p 2,..., p K, are cheap to calculate, the computational effort for the stratified sampling scheme is dominated by the n function evaluations, f(x k i ), i = 1, 2,...,, k = 1, 2,..., K, where K = n. The computational effort for the native sampling scheme of Section 2 is similarly dominated by the n function evaluations, f(x i ), i = 1, 2,..., n. To minimize the variance of the stratified sampler we solve the following optimization problem: min n 1,n 2,...,n K s.t. p 2 σk 2 k = n Z +. (9a) (9b) (9c) The objective function in (9a) follows from the formula for var f n in equation (5). In light of the discussion above, constraint (9b) limits the total computational effort to n function evaluations. Ignoring the integrality constraints (9c), the solution to model (9) is given by: ( ) p k σ k = K p n. (10) kσ k While the probability masses, p k, are known, the population variances, σk 2, are not. We could employ a two-stage procedure in which we first estimate σk 2 by sample variances in the first stage, and then allocate samples according to equation (10), with σ k replaced by the sample standard deviations. To answer the question of how to choose the cells, consider the variance decomposition formula: var f(x) = E [var [f(x) I]] + var [E [f(x) I]], (11) where I is an indicator random variable with realizations I(X S k ), k = 1, 2,..., K. We have var Ȳn = var f(x)/n; i.e., var Ȳn is the left-hand side of equation (11), within the factor of 1/n. The first term on the right-hand side of equation (11) is similar to var f n, except that the 4

probability weights are p k rather than p 2 k and, again within the scaling factors of 1/. Despite this difference, equation (11) gives us qualitative insight on how to attempt to achieve var f n var Ȳn. In particular, we would like the second term on the right-hand side of equation (11) to be as large as possible because this will decrease the magnitude of the first term, given that the left-hand side is a constant. This means that we would like the terms, E [f(x) I(X S k )], k = 1, 2,..., K, to be as variable as possible. To make this final observation concrete in the context of GSI-191 and the CASA Grande simulation model, consider a system reliability example, as we discuss in Section 1 with Y = I(X t 0 ). Then in forming cells for a stratification, we seek to form cells for which the conditional system reliability values, P(X t 0 S k ), k = 1, 2,..., K, are as spread out as possible. References [1] Asmussen, S. and P. W. Glynn (2007). Stochastic Simulation: Algorithms and Analysis. Springer, New York. [2] Hammersley, J. M. and D. C. Handscomb (1964). Monte Carlo Methods. Chapman and Hall, London. 5