Bias Reduction Using the Bootstrap

Similar documents
Introduction Dickey-Fuller Test Option Pricing Bootstrapping. Simulation Methods. Chapter 13 of Chris Brook s Book.

Strategies for Improving the Efficiency of Monte-Carlo Methods

Value at Risk Ch.12. PAK Study Manual

Week 7 Quantitative Analysis of Financial Markets Simulation Methods

Monte Carlo Methods for Uncertainty Quantification

Brooks, Introductory Econometrics for Finance, 3rd Edition

ELEMENTS OF MONTE CARLO SIMULATION

STAT 509: Statistics for Engineers Dr. Dewei Wang. Copyright 2014 John Wiley & Sons, Inc. All rights reserved.

Much of what appears here comes from ideas presented in the book:

Lecture outline. Monte Carlo Methods for Uncertainty Quantification. Importance Sampling. Importance Sampling

Gamma. The finite-difference formula for gamma is

Posterior Inference. , where should we start? Consider the following computational procedure: 1. draw samples. 2. convert. 3. compute properties

Ch4. Variance Reduction Techniques

Section 8.2: Monte Carlo Estimation

Statistical analysis and bootstrapping

Interval estimation. September 29, Outline Basic ideas Sampling variation and CLT Interval estimation using X More general problems

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

Results for option pricing

IEOR E4703: Monte-Carlo Simulation

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

2 Control variates. λe λti λe e λt i where R(t) = t Y 1 Y N(t) is the time from the last event to t. L t = e λr(t) e e λt(t) Exercises

Using Monte Carlo Integration and Control Variates to Estimate π

12 The Bootstrap and why it works

STRESS-STRENGTH RELIABILITY ESTIMATION

Generating Random Numbers

3. Monte Carlo Simulation

Alternative VaR Models

Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel

Module 4: Point Estimation Statistics (OA3102)

Statistics for Business and Economics

Optimal Search for Parameters in Monte Carlo Simulation for Derivative Pricing

Econ 300: Quantitative Methods in Economics. 11th Class 10/19/09

COMPARISON OF RATIO ESTIMATORS WITH TWO AUXILIARY VARIABLES K. RANGA RAO. College of Dairy Technology, SPVNR TSU VAFS, Kamareddy, Telangana, India

The Use of Importance Sampling to Speed Up Stochastic Volatility Simulations

Technische Universiteit Delft Faculteit Elektrotechniek, Wiskunde en Informatica Delft Institute of Applied Mathematics

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

Monte Carlo Based Reliability Analysis

STA 532: Theory of Statistical Inference

Chapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29

1 Inferential Statistic

Back to estimators...

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Chapter 8. Introduction to Statistical Inference

BIO5312 Biostatistics Lecture 5: Estimations

Homework Problems Stat 479

IEOR E4703: Monte-Carlo Simulation

The Two-Sample Independent Sample t Test

Chapter 7 - Lecture 1 General concepts and criteria

Chapter 8 Statistical Intervals for a Single Sample

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics

10. Monte Carlo Methods

As we saw in Chapter 12, one of the many uses of Monte Carlo simulation by

Machine Learning for Quantitative Finance

Financial Risk Management and Governance Other VaR methods. Prof. Hugues Pirotte

IEOR E4703: Monte-Carlo Simulation

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.

MONTE CARLO EXTENSIONS

Random Tree Method. Monte Carlo Methods in Financial Engineering

Chapter 14 : Statistical Inference 1. Note : Here the 4-th and 5-th editions of the text have different chapters, but the material is the same.

Chapter 7: Point Estimation and Sampling Distributions

Section 2.4. Properties of point estimators 135

Numerical Methods II

Lecture 22. Survey Sampling: an Overview

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

Calculating VaR. There are several approaches for calculating the Value at Risk figure. The most popular are the

The Values of Information and Solution in Stochastic Programming

AMH4 - ADVANCED OPTION PRICING. Contents

Review: Population, sample, and sampling distributions

Monte Carlo, Resampling, And Other Estimation Tricks. Mauricio Aguiar ti MÉTRICAS, President IFPUG Immediate Past President

Chapter 6: Point Estimation

Linear Regression with One Regressor

Ideal Bootstrapping and Exact Recombination: Applications to Auction Experiments

Monte Carlo Methods in Option Pricing. UiO-STK4510 Autumn 2015

STAT Chapter 6: Sampling Distributions

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Bootstrap Inference for Multiple Imputation Under Uncongeniality

Resampling Methods. Exercises.

UNIVERSITY OF VICTORIA Midterm June 2014 Solutions

ROM SIMULATION Exact Moment Simulation using Random Orthogonal Matrices

KARACHI UNIVERSITY BUSINESS SCHOOL UNIVERSITY OF KARACHI BS (BBA) VI

Chapter 2 Uncertainty Analysis and Sampling Techniques

Lecture 12: The Bootstrap

MM and ML for a sample of n = 30 from Gamma(3,2) ===============================================

1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by. Cov(X, Y ) = E(X E(X))(Y E(Y ))

Basics. STAT:5400 Computing in Statistics Simulation studies in statistics Lecture 9 September 21, 2016

FINITE SAMPLE DISTRIBUTIONS OF RISK-RETURN RATIOS

Statistics 431 Spring 2007 P. Shaman. Preliminaries

Alexander Marianski August IFRS 9: Probably Weighted and Biased?

Chapter 7: Estimation Sections

IEOR E4703: Monte-Carlo Simulation

A comment on Christoffersen, Jacobs and Ornthanalai (2012), Dynamic jump intensities and risk premiums: Evidence from S&P500 returns and options

UNIVERSITY OF NORTH CAROLINA Department of Statistics Chapel Hill, N. C. ON MONTE CARli) METHODS IN CONGESTION PROBLEMS

Modelling Returns: the CER and the CAPM

Multilevel Monte Carlo for VaR

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

FEEG6017 lecture: The normal distribution, estimation, confidence intervals. Markus Brede,

2.1 Mathematical Basis: Risk-Neutral Pricing

Chapter 5: Statistical Inference (in General)

Transcription:

Bias Reduction Using the Bootstrap Find f t (i.e., t) so that or E(f t (P, P n ) P) = 0 E(T(P n ) θ(p) + t P) = 0. Change the problem to the sample: whose solution is so the bias-reduced estimate is E(T(P (1) n ) T(P n ) + t 1 P n ) = 0, t 1 = T(P n ) E(T(P (1) n ) P n ), T 1 = 2T(P n ) E(T(P (1) n ) P n ).

Bias Reduction We may be able to compute E(T(P n ) P n ). If so, do so. If not, estimate by Monte Carlo.

Resampling to Estimate Bias Problem: given a random sample (x 1, x 2,..., x n ) from an unknown distribution with df P, we want to estimate a parameter, θ = θ(p). One of the most important properties of an estimator, T = t(x 1, x 2,..., x n ), is its bias, E P (T) θ. A good estimator, of course, has a small bias, probably even zero, i.e., the estimator is unbiased. The plug-in estimator, T = θ(p n ), may or may not be unbiased, but its bias is often small (e.g., the plug-in estimator of σ 2 ; the bias correction is n/(n 1)). Of course, without fairly strong assumptions on the underlying distribution, it is unlikely that we would know the bias of an estimator. Resampling methods can be used to estimate the bias. In particular, the bootstrap estimate of the bias is E Pn (T) θ(p n )

The Bootstrap Estimate of the Bias The bootstrap estimate of the bias is the plug-in estimate of E P (T) θ. The plug-in step occurs in two places, for estimating E P (T) and then for estimating θ. Consider the simple estimators (both plug-in estimators): Sample mean: E Pn (T) θ(p n ) = 0. The sample mean is unbiased for the population mean (if the population mean exists). Sample second central moment: E Pn (T) θ(p n ) = (x i x) 2 /n. This is also what we would want. This ideal bootstrap estimate must generally be approximated by Monte Carlo simulation.

The Monte Carlo Bootstrap Estimate of the Bias The ideal estimate is an expected value of a funtion: E Pn (T) θ(p n ); the Monte Carlo estimate of an expected value of a function is the Monte Carlo sample mean of that function. For the Monte Carlo simulation, we generate a number of bootstrap samples, (x 1, x 2,..., x n ), drawn from the empirical ditribution resulting from (x 1, x 2,..., x n ). Letting (x j 1, x j 2,..., x j n ) represent the j th bootstrap sample, the Monte Carlo estimate of the bootstrap estimate of the bias is j t(x 1, x j 2,..., x j n )/m θ(p n ) For the Monte Carlo estimate we can define a resampling vector, P, corresponding to each bootstrap sample as the vector of proportions of the elements of the original sample in the given bootstrap sample.

The Resampling Vector If the bootstrap sample (x 1, x 2, x 3, x 4 ) is really the sample (x 2, x 2, x 4, x 3 ), the resampling vector P is (0, 1/2, 1/4, 1/4) The resampling vector has random components that sum to 1. The bootstrap replication of the estimator T is a function of P. The Monte Carlo estimate of the bootstrap estimate of the bias can be improved if the estimator whose bias is being estimated is a plug-in estimator.

The Bootstrap Estimate of the Bias of a Plug-In Estimator Consider the resampling vector, P 0 = (1/n, 1/n,..., 1/n). Such a resampling vector corresponds to a permutation of the original sample. If the estimator is a plug-in estimator, then its value is invariant to permutations of the sample; and, in fact, θ(p 0 ) = θ(p n ), so the Monte Carlo estimate of the bootstrap estimate of the bias, j t(x 1, x j 2,..., x j n )/m θ(p n), can be written as j t(x 1, x j 2,..., x j n )/m θ(p 0 ). Instead of using θ(p 0 ), we can increase the precision of the Monte Carlo estimate by using the individual P s actually obtained: t(x j 1, x j 2,..., x j n )/m θ( P j /m), that is, by using the mean of the resampling vectors. Notice that for an unbiased plug-in estimator, e.g., the sample mean, this quantity is 0.

Variance Reduction in Monte Carlo The use of θ( P ) is a type of variance reduction in a Monte Carlo procedure. Remember that a Monte Carlo procedure is estimating an integral. Suppose the integral is (f(x) + g(x)) dx = f(x)dx + g(x))dx, and suppose we know g(x))dx. What is the best way to do the Monte Carlo, to do the integral of the sum, or just to do the integral of f(x) and add on the known value of the integral of g(x)? What makes one estimator better than another? Is the variance of f(x) + g(x) smaller than the variance of f(x)? Remember that the variance of the Monte Carlo estimator, Î, of an integral, h(x) dx, is generally proportional to the variance of h(x), where X is a random variable. (There are, of course, different ways of using random variables in the Monte Carlo estimation of the integral.)

Variance Reduction in Monte Carlo If the objective in Monte Carlo experimentation is to estimate some quantity, just as in any estimation procedure, we want to reduce the variance of our estimator (while preserving its other good qualities). The basic idea is usually to reduce the problem analytically as far as possible, and then to Monte Carlo what is left. Beyond that general reduction principle, in Monte Carlo experimentation, there are several possibilities for reducing the variance.

Variance Reduction in Monte Carlo judicious use of an auxilliary variable control variates (any correlated variable, either positively or negatively correlated) antithetic variates (in the basic uniform generator) regression methods use of probability sampling discrete: stratified sampling continuous: importance sampling

Balanced Resampling Another way of reducing the variance in Monte Carlo experimentation is to constrain the sampling so that some aspects of the samples reflect precisely some aspects of the population. What about constraining P so as to equal P 0? This makes θ( P ) = θ(p 0 ), and hopefully makes t(x j 1, x j 2,..., x j n )/m closer to its expected value, while preserving its correlation with θ( P ). This is called balanced resampling. Hall (1990) has shown that the balanced resampling Monte Carlo estimator of the bootstrap estimator has a bias O(m 1 ), but that the reduced variance generally more than makes up for it.