IEOR 3106: Introduction to OR: Stochastic Models. Fall 2013, Professor Whitt. Class Lecture Notes: Tuesday, September 10.

Similar documents
Homework Assignments

Chapter 5. Sampling Distributions

MATH 3200 Exam 3 Dr. Syring

Chapter 7: Point Estimation and Sampling Distributions

STAT Chapter 7: Central Limit Theorem

Random Variables Handout. Xavier Vilà

Tutorial 11: Limit Theorems. Baoxiang Wang & Yihan Zhang bxwang, April 10, 2017

Elementary Statistics Lecture 5

Central Limit Theorem (cont d) 7/28/2006

4.3 Normal distribution

Engineering Statistics ECIV 2305

Statistics 6 th Edition

MA : Introductory Probability

What was in the last lecture?

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

MLLunsford 1. Activity: Central Limit Theorem Theory and Computations

Math 489/Math 889 Stochastic Processes and Advanced Mathematical Finance Dunbar, Fall 2007

Normal distribution Approximating binomial distribution by normal 2.10 Central Limit Theorem

1 Geometric Brownian motion

STAT 241/251 - Chapter 7: Central Limit Theorem

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics

Chapter 5. Statistical inference for Parametric Models

Chapter 5: Statistical Inference (in General)

Binomial Random Variables. Binomial Random Variables

BROWNIAN MOTION Antonella Basso, Martina Nardon

Random Variables and Probability Functions

Chapter 3 - Lecture 5 The Binomial Probability Distribution

Reading: You should read Hull chapter 12 and perhaps the very first part of chapter 13.

Statistics, Measures of Central Tendency I

IEOR 165 Lecture 1 Probability Review

Statistical Tables Compiled by Alan J. Terry

Lecture 9 - Sampling Distributions and the CLT

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

ECON 214 Elements of Statistics for Economists 2016/2017

Lecture 17. The model is parametrized by the time period, δt, and three fixed constant parameters, v, σ and the riskless rate r.

BIO5312 Biostatistics Lecture 5: Estimations

Chapter 3 Discrete Random Variables and Probability Distributions

1 IEOR 4701: Notes on Brownian Motion

Discrete Random Variables and Probability Distributions. Stat 4570/5570 Based on Devore s book (Ed 8)

Statistical Methods in Practice STAT/MATH 3379

Part V - Chance Variability

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

The Binomial Distribution

Commonly Used Distributions

1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by. Cov(X, Y ) = E(X E(X))(Y E(Y ))

Business Statistics 41000: Probability 4

5.3 Statistics and Their Distributions

STATISTICAL LABORATORY, May 18th, 2010 CENTRAL LIMIT THEOREM ILLUSTRATION

Random Variable: Definition

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

Chapter 9: Sampling Distributions

Mathematics of Finance Final Preparation December 19. To be thoroughly prepared for the final exam, you should

5. In fact, any function of a random variable is also a random variable

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

Describing Uncertain Variables

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

MATH 264 Problem Homework I

Brownian Motion. Richard Lockhart. Simon Fraser University. STAT 870 Summer 2011

Chapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables

Central Limit Theorem, Joint Distributions Spring 2018

Some Discrete Distribution Families

IEOR 3106: Introduction to Operations Research: Stochastic Models SOLUTIONS to Final Exam, Sunday, December 16, 2012

Central Limit Thm, Normal Approximations

The Binomial Probability Distribution

From Discrete Time to Continuous Time Modeling

Central Limit Theorem 11/08/2005

Lecture 9. Probability Distributions. Outline. Outline

PROBABILITY DISTRIBUTIONS

Probability. An intro for calculus students P= Figure 1: A normal integral

CRRAO Advanced Institute of Mathematics, Statistics and Computer Science (AIMSCS) Research Report. B. L. S. Prakasa Rao

Chapter 3 Discrete Random Variables and Probability Distributions

1 The continuous time limit

Lecture 9. Probability Distributions

Chapter 7. Sampling Distributions and the Central Limit Theorem

Math489/889 Stochastic Processes and Advanced Mathematical Finance Homework 5

Lecture Stat 302 Introduction to Probability - Slides 15

Lecture 9 - Sampling Distributions and the CLT. Mean. Margin of error. Sta102/BME102. February 6, Sample mean ( X ): x i

15.063: Communicating with Data Summer Recitation 4 Probability III

Lecture 2. Probability Distributions Theophanis Tsandilas

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

Continuous random variables

Law of Large Numbers, Central Limit Theorem

AP Statistics Ch 8 The Binomial and Geometric Distributions

MA 1125 Lecture 12 - Mean and Standard Deviation for the Binomial Distribution. Objectives: Mean and standard deviation for the binomial distribution.

Module 10:Application of stochastic processes in areas like finance Lecture 36:Black-Scholes Model. Stochastic Differential Equation.

6.5: THE NORMAL APPROXIMATION TO THE BINOMIAL AND

Chapter 14 - Random Variables

Central limit theorems

STAT/MATH 395 PROBABILITY II

Section The Sampling Distribution of a Sample Mean

Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making

STA258H5. Al Nosedal and Alison Weir. Winter Al Nosedal and Alison Weir STA258H5 Winter / 41

Basic Data Analysis. Stephen Turnbull Business Administration and Public Policy Lecture 4: May 2, Abstract

4.2 Bernoulli Trials and Binomial Distributions

Week 1 Quantitative Analysis of Financial Markets Distributions B

ECON 214 Elements of Statistics for Economists 2016/2017

4 Random Variables and Distributions

Confidence Intervals Introduction

Probability Theory. Mohamed I. Riffi. Islamic University of Gaza

Transcription:

IEOR 3106: Introduction to OR: Stochastic Models Fall 2013, Professor Whitt Class Lecture Notes: Tuesday, September 10. The Central Limit Theorem and Stock Prices 1. The Central Limit Theorem (CLT See Section 2.7 of Ross. (a Time on My Hands: Suppose that I have a lot of time on my hands, e.g., because I am on a subway travelling the full length of the subway system. Fortunately, I have a coin in my pocket. And now I decide that this is an ideal time to see if heads will come up half the time in a large number of coin tosses. Specifically, I decide to see what happens if I toss a coin many times. Indeed, I toss my coin 1, 000, 000 times. Below are various possible outcomes, i.e., various possible numbers of heads that I might report having observed: 1. 500,000 2. 500,312 3. 501,013 4. 511,062 5. 598,372 What do you think of these reported outcomes? How believable are each of these possible outcomes? How likely are these outcomes? (Assume that I would report only one of these outcomes after doing the experiment. Assume that I would actually do such an experiment. We rule out outcome 5; there are clearly too many heads. We rule out outcome 1; it is too perfect. Even though 500, 000 is the most likely single outcome, it itself is extremely unlikely. But how do we think about the remaining three? The other three possibilities (alternatives 2 4 require more thinking. We introduce a probability model. We assume that successive coin tosses are independent and identically distributed (commonly denoted by IID with probability of 1/2 of coming out heads. Let S n denote the number of heads in n coin tosses. Observe that S n has exactly (according to the model a binomial distribution. Carefully examining the too-perfect case. First, here in these notes we quantify just how unlikely is the too perfect outcome of exactly the mean value 500, 000 in 10 6 tosses. If the probability of heads in one toss were p, then we can exploit properties of the binomial distribution to conclude that the probability of k heads in n tosses would be P (S n = k = b(k; n, p = n! k!(n k! pk (1 p (n k

Now we are interested in the case and p = 1/2 and k = n/2 (for n even, i.e., P (S n = n/2 = b(n/2; n, 1/2 = n! (n/2!(n/2! (1/2n. It is good to be able to roughly estimate these probabilities. To do so, we can use Stirling s formula (see p. 146: n! 2πn(n/e n. We thus see that P (S n = n/2 = b(n/2; n, 1/2 2πn(n/e n (πn(n/2e n (1/2n = 2/πn 0.8 n < 1 n. Hence, the probability of getting the outcome is approximately 0.8/1000, less than 1/1000. Of course, this special outcome is the most likely single outcome, and it could of course occur, but a probability less than 1/1000 is quite unlikely. a normal approximation. We now consider alternatives 2 4 above. We can answer the question by doing a normal approximation; see Section 2.7 of Ross, especially pages 79-83. A key property of the normal distribution is that it has only two parameters: its mean and variance. Thus all we should need to know are the mean and variance. The random variable S n is approximately normally distributed with mean np = 500, 000 and variance np(1 p = 250, 000, where p = 1/2. Thus S n has standard deviation SD(S n = = 500. Case 2 looks likely because it is less than 1 standard deviation from the mean; case 3 is not too likely, but not extremely unlikely, because it is just over 2 standard deviations from the mean. On the other hand, Case 4 is extremely unlikely, because it is over 20 standard deviations from the mean. See the Table on page 81 of the text. When we consider case 3, we do not look at the probability P (S n = 501, 013, because the probability of each value is necessarily small, as we have just seen. Instead, we want to look at P (S n = 501, 013. We want to see the probability of getting that large value or any larger value. In more detail we would do the following computation for case 3: P (S n 501, 013 = P ( 501, 013 E[S n ] = P ( 501, 013 E[S n] P (N(0, 1 501, 013 E[S n] 501, 013 500, 000 = P (N(0, 1 500 = P (N(0, 1 2.006 0.0225, where the first two lines follow from simple arithmetic, doing the same on both sides, while the third line is the normal approximation justified by the CLT and the final numerical value is obtained from the Table 2.3 on page 82 of the class textbook. Such a large result is unlikely but not extremely unlikely. Case 4 is worth additional discussion. An outcome of 20 standard deviations above the mean is extremely unlikely. However, there actually are two possible causes. On the one 2

hand, the report may be inaccurate. On the other hand the model may be inaccurate. Given that the model is accurate, it would be extremely unlikely that the reported outcome actually occurred. However, there is another possibility. It is possible that the model is not accurate. If the probability of heads coming up on each toss were actually around 0.51, then Case 4 would be reasonable, and Case 2 and 3 would not be reasonable. If Case 4 were the outcome of a proper experiment, we could conclude that the probability of heads must not actually be exactly 0.500 or anything less than that. (b The Power of the CLT The normal approximation for the binomial distribution with parameters (n, p when n is not too small and the normal approximation for the Poisson with mean λ when λ is not too small are both special cases of the central limit theorem (CLT. The CLT states that a properly normalized sum of random variables converges in distribution to the normal distribution. Of course there are conditions. We give a formal statement; see Theorem 2.2 on p. 79 of Ross. For that purpose, let N(m, σ 2 denote a random variable having a normal distribution with mean m and variance σ 2. Let denote convergence in distribution. Theorem 0.1 (central limit theorem (CLT Suppose that {X n : n 1} is a sequence of independent and identically distributed (IID random variables, each distributed as X. Form the partial sums S n X 1 + + X n for n 1. If E[X 2 ] < or, equivalently, if σ 2 V ar(x < (which implies that the mean is finite, then N(0, 1 as n, i.e., P as n for each x. ( x P (N(0, 1 x = 1 x 2π e y2 /2 dy Where does the sum appear in our application? A random variable that has a binomial distribution with parameters (n, p can be regarded as the sum of n IID random variables with a Bernoulli distribution having parameter p; each of these random variables X i assumes the value 1 with probability p and assumes the value 0 otherwise. A random variable having a Poisson distribution with mean nλ can be regarded as the sum of n IID random variables, each with a Poisson distribution with mean λ (for any n. And what about the normalization? We simply subtract the mean of S n and divide by the standard deviation of S n to make the normalized sum have mean 0 and variance 1. Note that = S n nµ (1 nσ 2 has mean 0 and variance 1 whenever S n X 1 + + X n, 3

where {X n : n 1} is a sequence of IID random variables with mean µ and variance σ 2. (It is crucial that the mean and variance be finite. Please note that the normalization is not a significant part of the CLT statement. For any random variable Z, the associated normalized random variable (Z E[Z]/SD(Z has mean 0 and variance 1. Since the normalized sums above have mean 0 and variance 1 for all n, there is some hope that there might be a limiting distribution, which we expect to have mean 0 and variance 1. But, for an arbitrary random variable Z, the associated normalized random variable (Z E[Z]/SD(Z does not need to be normally distributed. Indeed, it is not unless Z itself is normally distributed. The amazing part of the CLT is that the distribution of the normalized sum (/ V ar(s n does approach the normal distribution as n gets large. Moreover, the CLT applies much more generally; it has remarkably force. The random variables being added do not have to be Bernoulli or Poisson; they can have any distribution. We only require that the distribution have finite mean µ and variance σ 2. The statement of a basic CLT is given in Theorem 2.2 on p. 79 of Ross. The conclusion actually holds under even weaker conditions. The random variables being added do not actually have to be independent; it suffices for them to be weakly dependent; and the random variables do not have to be identically distributed; it suffices for no single random variable to be large compared to the sum. But the statement then need adjusting: the first expression in (1 remains valid, but the second does not. What does the CLT say? The precise mathematical statement is a limit as n. It says that, as n, the normalized sum in (1 converges in distribution to N(0, 1, a random variable that has a normal distribution with mean 0 and variance 1, whose distribution is given in the table on page 81 of our textbook. (Let N(a, b denote a normal distribution with mean a and variance b. What does convergence in distribution mean? It means that the cumulative distribution functions (cdf s converge to the cdf of the normal limit, denoted by N(0, 1, which means that ( P x P (N(0, 1 x 1 x e y2 /2 dy 2π for all x. Note that convergence in distribution means convergence of cdf s, which means convergence of functions. How do we apply the CLT? We approximate the distribution of the normalized sum in (1 by the distribution of N(0, 1. The standard normal (with mean 0 and variance 1 has no parameters at all; its distribution is given in the Table on page 81. By scaling, we can reduce other normal distributions to this one. The approximation is N(0, 1, which, upon undoing the normalization becomes S n E[S n ] + V ar(s n N(0, 1 d = N(E[S n ], V ar(s n. 4

As a consequence of the CLT, we conclude that S n is approximately normally distributed with its true mean and variance. The CLT states that the distribution is approximately normal, regardless of the distribution of the underlying random variables X i. The CLT helps explain why the normal distribution arises so often. We apply this normal approximation to approximate the distribution of S n. As we did for case 3 above, we write P (S n c = P ( c E[S n ] = P ( c E[S n] P (N(0, 1 c E[S n], and then calculate b = c E[S n ]/ V ar(s n and look up the value P (N(0, 1 b in Table 2.3 of the normal distribution (or use a program for that purpose. We can use or, but we have to be careful that we are consistent. We use P (N(0, 1 b = 1 P (N(0, 1 b. 2. An Application of the CLT: Modeling Stock Prices Given the generality of the CLT, it is nice to consider an application where the random variables being added in the CLT are not Bernoulli or Poisson, as in many applications. Hence we consider such an application now. (a An Additive Random Walk Model for Stock Prices We start by introducing a random-walk (RW model for a stock price. Let S n denote the price of some stock at the end of day n. We then can write S n = S 0 + X 1 + + X n, (2 where X i is the change in stock price between day i 1 and day i (over day i and S 0 is the initial stock price, presumably known (if we start at current time and contemplate the evolution of the stock price into the uncertain future. We are letting the index n count days, but we could have a different time unit. We now make a probability model. We do so by assuming that the successive changes come from a sequence {X n : n 1} of IID random variables, each with mean µ and variance σ 2. This is roughly reasonable. Moreover, we do not expect the distribution to be Bernoulli or Poisson. The stochastic process {S n : n 0} is a random walk with steps X n, but a general random walk. If the steps are Bernoulli random variables, then we have a simple random walk, as discussed in Chapter 4, in particular, in Example 4.5 on page 183 and Example 4.15. But here the steps can have an arbitrary distribution. We now can apply the CLT to deduce that the model implies that we can approximate the stock price on day n by a normal distribution. In particular, P (S n x P (N(S 0 + nµ, nσ 2 x = P (N(0, 1 (x S 0 nµ/σx. How do we do that last step? Just re-scale: subtract the mean from both sides and then divide by the standard deviation for both sides, inside the probabilities. The normal variable is then transformed into N(0, 1. We can clearly estimate the distribution of X n by looking at data. We can investigate if the stock prices are indeed normally distributed. 5

(b A Multiplicative Model for Stock Prices Actually, many people do not like the previous model, because they believe that the change in a stock price should be somehow proportional to the price. (There is much much more hardnosed empirical evidence, not just idle speculation. That leads to introducing an alternative multiplicative model of stock prices. Instead of (2 above, we assume that S n = S 0 X 1 X n, (3 where the random variables are again IID, but now they are random daily multipliers. Clearly, the random variable X n will have a different distribution if it is regarded as a multiplier instead of an additive increment. But, even with this modification, we can apply the CLT. We obtain an additive model again if we simply take logarithms (using any base, but think of standard base e = 2.71828... Note that log (S n = log (S 0 + log (X 1 + + log (X n, (4 so that, by virtue of the CLT above, where now (with this new interpretation of X n log (S n N(log (S 0 + nµ, nσ 2, (5 µ E[log (X 1 ] and σ 2 V ar(log (X 1. (6 As a consequence, we can now take exponentials of both sides of (5 to deduce that S n e (N(log (S 0+nµ,nσ 2. (7 That says that S n has a lognormal distribution. Some discussion of this model appears on page 608 of our textbook. It underlies geometric Brownian motion, one of the fundamental stochastic models in finance. 6