Business Statistics 41000: Probability 3

Similar documents
Business Statistics 41000: Probability 4

A useful modeling tricks.

Homework: Due Wed, Feb 20 th. Chapter 8, # 60a + 62a (count together as 1), 74, 82

Statistics for Business and Economics

ECO220Y Continuous Probability Distributions: Normal Readings: Chapter 9, section 9.10

Chapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables

Introduction to Business Statistics QM 120 Chapter 6

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

Homework: Due Wed, Nov 3 rd Chapter 8, # 48a, 55c and 56 (count as 1), 67a

Chapter 7 1. Random Variables

Statistical Intervals (One sample) (Chs )

The topics in this section are related and necessary topics for both course objectives.

Probability. An intro for calculus students P= Figure 1: A normal integral

Chapter 7 Sampling Distributions and Point Estimation of Parameters

Introduction to Statistics I

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

Topic 6 - Continuous Distributions I. Discrete RVs. Probability Density. Continuous RVs. Background Reading. Recall the discrete distributions

Section 0: Introduction and Review of Basic Concepts

2011 Pearson Education, Inc

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

ECON Introductory Econometrics. Lecture 1: Introduction and Review of Statistics

A random variable (r. v.) is a variable whose value is a numerical outcome of a random phenomenon.

Central Limit Theorem, Joint Distributions Spring 2018

MAKING SENSE OF DATA Essentials series

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman:

Chapter 4 Continuous Random Variables and Probability Distributions

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

MAS1403. Quantitative Methods for Business Management. Semester 1, Module leader: Dr. David Walshaw

Random Variables and Probability Distributions

Normal Probability Distributions

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Random Variables Handout. Xavier Vilà

Section 7.5 The Normal Distribution. Section 7.6 Application of the Normal Distribution

What was in the last lecture?

CS 237: Probability in Computing

Part V - Chance Variability

AP STATISTICS FALL SEMESTSER FINAL EXAM STUDY GUIDE

Biostatistics and Design of Experiments Prof. Mukesh Doble Department of Biotechnology Indian Institute of Technology, Madras

Commonly Used Distributions

Statistics for Business and Economics: Random Variables:Continuous

Statistics and Probability

4: Probability. What is probability? Random variables (RVs)

STAT Chapter 5: Continuous Distributions. Probability distributions are used a bit differently for continuous r.v. s than for discrete r.v. s.

STAT 201 Chapter 6. Distribution

Econ 6900: Statistical Problems. Instructor: Yogesh Uppal

Theoretical Foundations

MAS3904/MAS8904 Stochastic Financial Modelling

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics

Continuous Distributions

4.3 Normal distribution

Chapter 7: SAMPLING DISTRIBUTIONS & POINT ESTIMATION OF PARAMETERS

Chapter 4 Continuous Random Variables and Probability Distributions

Probability Distributions II

Basic Data Analysis. Stephen Turnbull Business Administration and Public Policy Lecture 4: May 2, Abstract

SYSM 6304 Risk and Decision Analysis Lecture 2: Fitting Distributions to Data

TOPIC: PROBABILITY DISTRIBUTIONS

Continuous random variables

Chapter 5: Probability models

THE UNIVERSITY OF TEXAS AT AUSTIN Department of Information, Risk, and Operations Management

STOR Lecture 15. Jointly distributed Random Variables - III

Data Analysis and Statistical Methods Statistics 651

Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is

Probability and Random Variables A FINANCIAL TIMES COMPANY

Lecture 6: Chapter 6

Figure 1: 2πσ is said to have a normal distribution with mean µ and standard deviation σ. This is also denoted

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Lecture 9. Probability Distributions. Outline. Outline

IEOR 165 Lecture 1 Probability Review

Statistics, Their Distributions, and the Central Limit Theorem

The Normal Distribution

A random variable (r. v.) is a variable whose value is a numerical outcome of a random phenomenon.

Copyright 2011 Pearson Education, Inc. Publishing as Addison-Wesley.

MLLunsford 1. Activity: Central Limit Theorem Theory and Computations

Lecture 9. Probability Distributions

Discrete Random Variables

AP Statistics Chapter 6 - Random Variables

The normal distribution is a theoretical model derived mathematically and not empirically.

2. The sum of all the probabilities in the sample space must add up to 1

Normal distribution Approximating binomial distribution by normal 2.10 Central Limit Theorem

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

STATISTICS and PROBABILITY

The Bernoulli distribution

Lecture 2. Probability Distributions Theophanis Tsandilas

NORMAL RANDOM VARIABLES (Normal or gaussian distribution)

MATH 264 Problem Homework I

Discrete Random Variables and Probability Distributions

Chapter 7: Random Variables

Statistics, Measures of Central Tendency I

Chapter 4 Probability and Probability Distributions. Sections

. (i) What is the probability that X is at most 8.75? =.875

Lecture 5 - Continuous Distributions

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals

Basic Procedure for Histograms

χ 2 distributions and confidence intervals for population variance

ECON 214 Elements of Statistics for Economists 2016/2017

I. Return Calculations (20 pts, 4 points each)

The Normal Distribution. (Ch 4.3)

Transcription:

Business Statistics 41000: Probability 3 Drew D. Creal University of Chicago, Booth School of Business February 7 and 8, 2014 1

Class information Drew D. Creal Email: dcreal@chicagobooth.edu Office: 404 Harper Center Office hours: email me for an appointment Office phone: 773.834.5249 http://faculty.chicagobooth.edu/drew.creal/teaching/index.html 2

Course schedule Week # 1: Plotting and summarizing univariate data Week # 2: Plotting and summarizing bivariate data Week # 3: Probability 1 Week # 4: Probability 2 Week # 5: Probability 3 Week # 6: In-class exam Week # 7: Statistical inference 1 Week # 8: Statistical inference 2 Week # 9: Simple linear regression Week # 10: Multiple linear regression 3

Outline of today s topics I. Continuous Random Variables (AWZ p. 215-216) II. The Uniform Distribution (AWZ p. 411) III. The Normal Distribution (AWZ p. 257-278) IV. Modeling Stock Returns: the i.i.d. Normal Model V. The Random Walk Model (again) VI. Cumulative Distribution Functions (AWZ p. 538) VII. Expected Value and Variance of Continuous Random Variables (AWZ p. 241-242) VIII. The Chi-square and Student s t distributions (AWZ p. 438, 453, 434) 4

Continuous Random Variables 5

Continuous Random Variables In the last two lectures, we considered discrete random variables. We used them to model uncertainty in situations where there were a countable number of outcomes of the random event. In many situations, a random event can take on a continuous range of values. This is a continuous random variable! 6

Continuous Random Variables Suppose we have a machine that cuts cloth. Each time the machine cuts the cloth, there are threads. We believe the length of a thread could be anything between 0 and 0.5 inches and any value in the interval is equally likely. The machine is about to cut and it will leave a new thread. In advance of the machine cutting the cloth, the length of the thread is a number we are uncertain about. It is a random variable. How do we describe our beliefs? 7

Continuous Random Variables This is an example of a continuous random variable. It can take on any value in an interval. In this example each value in the interval is equally likely. But in general, we might want to express beliefs that some values are more likely than others. 8

Continuous Random Variables In the discrete case, we listed the possible outcomes and assigned a probability to each outcome. Now, the number of possible outcomes is so large that we can t make a list and give each outcome a probability! Instead we talk about the probability of intervals. In the discrete case, we had P(X = x) = 0.1 whereas in the continuous case we have P(a < X < b) = 0.1 (NOTE: Here, the interval is (a, b), i.e. all the real numbers between a and b.) 9

Continuous Random Variables One convenient way to specify the probability of an interval is with a probability density function (pdf). The probability of an interval is the area under the pdf over that interval. This is an example of a pdf. Values close to zero are more likely because the area under the curve in that region is larger. f(x) 0.40 0.35 0.30 0.25 0.20 0.15 0.10 0.05 4 3 2 1 0 1 2 3 4 x 10

Continuous Random Variables 0.40 0.35 0.30 f(x) 0.25 0.20 0.15 0.10 0.05 4 3 2 1 0 1 2 3 4 x For this random variable the probability that it is in the interval [0, 2] is 0.477. (47.7 percent of the time it will fall in this interval). The area under the curve within the blue region is 0.477. 11

Continuous Random Variables 0.35 0.30 0.25 f(x) 0.20 0.15 0.10 0.05 x 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Here is another example of a probability density function. This random variable can only take on positive values because it gives zero probability mass to values less than zero. This kind of distribution is called skewed right. 12

Continuous Random Variables 0.35 0.30 0.25 f(x) 0.20 0.15 0.10 0.05 x 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 For this random variable, consider the interval (4, + ). What is the probability that it takes a value greater than 4? The probability is 0.0915. 13

Continuous Random Variables For a continuous random variable X, the probability that it takes a value in the interval (a, b) is the area under the probability density function from a to b. By definition, the area under the curve (pdf) must be equal to 1. (NOTE: Why?) The probability density function is a function that you may see written as f (x). 14

Continuous Random Variables For those of you who like math, an equivalent way of saying this is to use calculus. (NOTE: This will not be on an exam.) If the random variable X has a probability density function f (x), then P(a < X < b) = b a f (x)dx An integral allows us to compute the area under the curve f (x) over the interval (a, b). 15

Continuous Random Variables An integral is just a sum! P(a < X < b) = b a f (x)dx It is a fancy way of saying OR means ADD that we saw earlier. P(a < X < b) = a<x <b p(x) 16

The Uniform Distribution 17

The Uniform Distribution Let s return to our example of the machine that cuts cloth and leaves threads. Any value between 0 and 0.5 is equally likely. If X is the length of a thread, what is its pdf? 18

The Uniform Distribution The pdf is flat in the interval (0, 0.5). Its height must be 2 so that the total area is 1. Outside of (0, 0.5), the pdf is equal to zero. f(x) 2.5 2.0 1.5 1.0 0.5 0.0 0.5 0.4 0.3 0.2 0.1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 x Technically, any particular value has probability 0! But, we still express the idea that they are equally likely in (0,.5). 19

The Uniform Distribution In general, we have a family of uniform distributions describing the situation where any value in the interval (a, b) is equally likely. 1/(b a) a b The height must be 1/(b a) so that the area of the rectangle = 1. 20

The Uniform Distribution A random variable X is said to have a uniform distribution if it has probability density function f (x) = { 1 (b a) a < x < b 0 otherwise We write X U(a, b) The parameters of the Uniform distribution are the two endpoints a and b. 21

The Normal Distribution 22

The Normal Distribution This pdf describes the standard normal distribution. We often use the letter Z to denote the random variable which has this pdf. 4 f(x) 0.40 0.35 0.30 0.25 0.20 0.15 0.10 0.05 3 2 1 0 1 2 3 4 x The normal distribution has a bell-shaped curve. Any value in the interval (, ) is possible. It is the standard normal because it has mean 0 and variance 1. 23

The Standard Normal Distribution Some properties of the standard normal distribution, which has µ Z = 0 and σ Z = 1. P(0 < Z < 1) = 0.34 P( 1 < Z < 1) = 0.68 P( 2 < Z < 2) = 0.954 P( 1.96 < Z < 1.96) = 0.95 P( 3 < Z < 3) = 0.9974 (NOTE: In these notes I will usually act as if 1.96 = 2) 24

The Normal Distribution If X is a normal random variable, we write: X N(µ, σ 2 ) The normal distribution has two parameters µ and σ 2. It turns out that µ is the mean, σ 2 is the variance, and σ is the standard deviation. The standard normal is the special case when µ = 0 and σ 2 = 1. Here, we really have a family of normal distributions. One for each value of the parameters (µ, σ 2 ). The normal distribution is also known as the Gaussian distribution. 25

The Normal Distribution The normal family has two parameters. 0.8 0.7 µ : determines where the curve is centered. σ : determines how spread out the curve is. pdf 0.6 0.5 0.4 0.3 0.2 X N(0,4) Z N(0,1) Y N(1,0.5 2 ) 0.1 5 4 3 2 1 0 1 2 3 4 5 X,Y, and Z are all normally distributed. 26

Interpreting the Normal Distribution X N(µ, σ 2 ) There is a 95% chance X is in the interval µ ± 2σ. µ describes what you think will happen on average. ±2σ describes how wrong you could be. µ determines the location of the distribution. σ determines the spread. 27

The Empirical Rule In lecture #1, we suggested that for mound-shaped data an empirical rule was Approximately 68% of the data is in the interval (x s x, x + s x ) = x ± s x Approximately 95% of the data is in the interval (x 2s x, x + 2s x ) = x ± 2s x That s right. We are just using a normal distribution as an approximation. 28

Identically distributed normal random variables Consider two normal random variables: X N(1.2, 1) Y N(1.201, 1) Are X and Y identically distributed? Consider two more normal random variables: X N(µ, σ 2 ) Y N(µ, 2σ 2 ) Are X and Y identically distributed? 29

Example: Interpreting the Normal Distribution Let the random variable R t be the return next month on a mutual fund. Assume that R t N(0.1, 0.01). 4.0 3.5 3.0 There is a 95% chance that R t will be in the interval (µ 2σ, µ+2σ) = ( 0.1, 0.3) f(x) 2.5 2.0 1.5 1.0 0.5 95 % area x 0.4 0.3 0.2 0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 What is P( 0.1 < R t < 0.3)? (NOTE: The distribution is symmetric around µ.) 30

Interpreting the Normal Distribution Just so you know, if then f (x) = X N(µ, σ 2 ) ( 1 exp 1 ) (x µ)2 2πσ 2 2σ2 is the probability density function (pdf). In this class, we will not need this. But, we will see the normal distribution many more times throughout the course. It is probably the most important distribution in statistics. 31

One More Property of the Normal Distribution Any linear combination of normal random variables is also normally distributed. In statistics, many models are based on linear relationships Y = c 0 + c 1 X 1 +... + c k X k. In words, if each X i is normally distributed, then Y will be normally distributed as well. This is another reason why the normal distribution is so important. 32

Building Models Part II 1. The i.i.d. Normal Model 2. The random walk model 33

The i.i.d. Normal Model Remember how we used the idea of i.i.d draws from the Bernoulli(.5) distribution to model coin tosses and i.i.d. draws from the Bernoulli(.2) to model a manufacturing process that gives faulty and good parts? Now, we want to use the normal distribution to model real data. Surprisingly often, data looks like i.i.d. draws from a normal distribution. 34

The i.i.d. Normal Model We can have i.i.d. draws from any distribution. When we write X 1, X 2,..., X n N(µ, σ 2 ) i.i.d. we mean that each random variable X i for i = 1,..., n will be an independent draw from the same normal distribution. We have not formally defined independence for continuous distributions, but our intuition is the same as for discrete random variables. Each draw has no effect on the others, and the same normal distribution describes what we think each variable will turn out to be. 35

The i.i.d. Normal Model What do i.i.d. normal draws look like? We can simulate them using a computer. In this plot, there are 100 draws from a standard normal distribution. 3 2 1 0 1 2 There is no pattern in the draws. They look random. 3 0 10 20 30 40 50 60 70 80 90 100 36

The i.i.d. Normal Model This is the same graph only now I have added lines indicating µ = 0 and the interval (µ 2σ, µ + 2σ). 3 95% of the data will be in between these lines (in the long run). 2 1 0 1 2 3 0 10 20 30 40 50 60 70 80 90 100 37

The i.i.d. Normal Model Here are draws from a different normal distribution. These are draws from N(5, 4). 10 8 6 4 How would you predict the next draw? 2 0 0 10 20 30 40 50 60 70 80 90 100 38

The i.i.d. Normal Model Data that is approximately normal is very common. 5000 4000 3000 2000 1000 Histogram of the daily returns on the Dow Jones from 1928 to 2010. 0.08 0.06 0.04 0.02 0.00 0.02 0.04 0.06 0.08 39

The i.i.d. Normal Model Example: Consider the return data for Canada. We have monthly returns from Feb 1988 to Dec. 1996 1. Do the data look i.i.d.? 2. Do the data look normally distribution? 0.075 0.050 0.025 0.000 0.025 0.050 1. Look at a time series plot. There appears to be no pattern. 0.075 Canadian returns 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 40

The i.i.d. Normal Model 30 25 20 15 10 5 0.100 0.075 0.050 0.025 0.000 0.025 0.050 0.075 0.100 Canada 2. Look at a histogram. They appear approximately normal. The returns look i.i.d. normal (at least as a first approximation). 41

The i.i.d. Normal Model The sample mean and stand. dev. are R t = 0.01 and s Rt = 0.04. If we use these to approximate µ and σ, then our prediction for the next return is 0.01 and a 95% interval is (-0.07,0.09) 0.075 0.09 0.050 0.025 0.000 0.025 0.050 0.075 R t = 0.01 0.08 Canadian returns Our model is R t N(0.01, 0.04 2 ) i.i.d. 1989 1990 1991 1992 1993 1994 1995 1996 1997 42

The i.i.d. Normal Model It is important to realize that not all data is normal. 40 Remember the data set on arrival times of customers at a bank. 35 30 25 20 15 10 Histogram of Interarrival Times for Bank Customers 5 0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0 22.5 25.0 27.5 30.0 43

The i.i.d. Normal Model Not all time series are independent... Dow Jones Industrial Average in levels. 14000 13000 Dow Jones Industrial Average Index Sept. 29, 2008 12000 (NOTE: we can often 11000 transform data such that 10000 the i.i.d. Normal model is a better approximation.) 9000 8000 7000 2002 2002.5 2003 2003.5 2004 2004.5 2005 2005.5 2006 2006.5 2007 2007.5 2008 2008.5 2009 2009.5 44

The i.i.d. Normal Model Here, I have transformed the DJ Index into returns. If X t measures the level of the series, then the returns R t are computed as R t = log(x t) log(x t 1 ). 400 300 200 100 0.08 0.06 0.04 0.02 0.00 0.02 0.04 0.06 0.08 0.10 0.12 Dow Jones Returns The normal approximation appears to be better. 45

The i.i.d. Normal Model But, appearances can still be deceiving. Are the returns really even i.i.d.? Does it look like they have the same variance σ 2? 0.100 Returns on the Dow Jones Industrial Average Index 0.075 0.050 0.025 0.000 0.025 0.050 0.075 2002 2002.5 2003 2003.5 2004 2004.5 2005 2005.5 2006 2006.5 2007 2007.5 2008 2008.5 2009 2009.5 (NOTE: Robert Engle won the 2003 Nobel prize for modeling volatility clustering, which is a fancy phrase to describe a time-varying variance σ 2 t.) 46

The i.i.d. Normal Model Not all time series are independent... 12 Levels of Lake Michigan from 1872 to 1975. (see lakemichigan.xls ) 11 10 9 8 7 6 1880 1890 1900 1910 1920 1930 1940 1950 1960 1970 year 47

The Random Walk Model (again) In Lecture # 4, we looked at the random walk model. Let P t denote the price of an asset. The equation for a random walk is: P t+1 = P t + E t+1 (..., E t 1, E t, E t+1,...) are a sequence of i.i.d. random variables. Recall that in the previous lecture, we modeled E t+1 as a series of discrete random variables. Each random variable E t+1 could take on only two possible values 0.125 or 0.125 with equal probabilities. 48

The Random Walk Model (again) Here is the same graph of P t from our previous lecture. We can see that prices move by taking discrete jumps. 83.0 82.5 82.0 81.5 81.0 80.5 80.0 79.5 0 25 50 75 100 125 150 175 200 225 250 49

The Random Walk Model (again) P t+1 = P t + E t+1 In a random walk model, it is possible for E t+1 to be a random variable having a continuous distribution. Consider letting E t+1 be a sequence of i.i.d normal random variables. E t N(0, 0.04) for t = 1, 2,... 50

The Random Walk Model (again) Here is a plot of data from this model. At each point in time, prices changes are normally distributed. 82 81 80 79 78 77 0 25 50 75 100 125 150 175 200 225 250 51

Cumulative Distribution Functions 52

Cumulative Distribution Function The cumulative distribution function (c.d.f.) is another way (besides the p.d.f.) to specify the probability of intervals. For a random variable X, the c.d.f. is defined by F X (x) = P(X x) (NOTE: it is common to use capital F to denote the c.d.f. and lower case f to denote the p.d.f.) 53

Cumulative Distribution Function F X (x) = P(X x) = Area under the pdf to the left of the point x Two examples for the standard normal distribution. 0.40 0.40 0.35 0.30 F( 1) = 0.16 0.35 0.30 F(1.96) = 0.975 0.25 0.25 0.20 0.20 0.15 0.15 0.10 0.10 0.05 0.05 4 3 2 1 0 1 2 3 4 4 3 2 1 0 1 2 3 4 54

Cumulative Distribution Function The cumulative distribution function allows us to compute the probability of any interval (a, b) P(a < x < b) = P(X < b) P(X < a) = F X (b) F X (a) By taking the difference between two c.d.f.s we get the probability of the interval (a, b). 55

Cumulative Distribution Function As an example, consider a random variable Z that has a standard normal distribution. P( 1 < Z < 1) = P(Z < 1) P(Z < 1) = F Z (1) F Z ( 1) = 0.84 0.16 = 0.68 0.40 0.35 The area under the curve between -1 and 1 is 0.68. 0.30 0.25 0.20 0.15 0.10 0.05 4 3 2 1 0 1 2 3 4 56

Cumulative Distribution Function Instead of plotting the p.d.f., we can plot the c.d.f.. In the c.d.f., the probability of an interval is the jump in the c.d.f. over that interval. Probability is plotted on the vertical axis. Here is an example of Z N(0, 1). 1.0 Note that F Z (1) = 0.84 and F Z ( 1) = 0.16 as on the last slide. F(1) F( 1) 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 4 3 2 1 0 1 2 3 4 x 57

Cumulative Distribution Function For x big enough, F (x) must go to 1. (Why?) For x small enough, F (x) must go to 0. (Why?) Here is an example of a c.d.f. where we consider a generic interval (a, b). 1.0 F(b) 0.8 F(a) 0.6 0.4 0.2 x a b 58

Example: let R denote the return on our portfolio next month. We do not know the return in advance so it is random. Let s assume that R N(0.01, 0.04 2 ). 10 9 8 7 6 5 4 3 2 1 0.2 0.175 0.15 0.125 0.1 0.075 0.05 0.025 0 0.025 0.05 0.075 0.1 0.125 0.15 0.175 0.2 59

R N(0.01, 0.04 2 ) What is the probability of a negative return? P(R < 0) = F R (0) = 0.4013 NOTE: in Excel, =NORMDIST(0,0.01,0.04,TRUE) What is the probability of a return between 0 and 0.05? P(0 < R < 0.05) = F R (0.05) F R (0) = 0.8413 0.4013 = 0.44 NOTE: in Excel, =NORMDIST(0.05,0.01,0.04,TRUE) 60

CDF of the Uniform Distribution Let X U(a, b). Recall that its p.d.f. is: f (x) = { 1 (b a) a < x < b 0 otherwise What is the c.d.f. of F X (x)? F X (x) = P(X < x) = area under f (x) to the left of x. Can you write down a formula for this area? 61

Expected Value and Variance of Continuous Random Variables 62

Expected Value and Variance of Continuous RV s If X is a continuous random variable with p.d.f. f (x) then the expected value is E [X ] = x f (x)dx This is basically the same thing as we had in the discrete case E [X ] = all x x p(x) 63

Expected Value and Variance of Continuous RV s If X is a continuous random variable with p.d.f. f (x) then the variance is Var [X ] = E [(X µ) 2] = (x µ) 2 f (x)dx This is basically the same thing as we had in the discrete case Var [X ] = all x p(x) (x µ) 2 64

Expected Value and Variance of Continuous RV s If you know calculus, those formulas are intuitive. If you do not know calculus, they are basically jibberish. But remember: An integral is just a sum. Just replace with all x and then replace f (x)dx with p(x). If you did this, then the formulas would be the same as for discrete random variables from Lecture # 3. 65

Expected Value and Variance of Continuous RV s Practically, what does this mean? First, recognize that the intuition we developed for discrete random variables is the same. 1. The mean is a best guess or a measure of the center of the distribution. 2. The variance is the expected squared distance of X from its mean. It measures how spread out the distribution is. 66

Expected Value and Variance of Continuous RV s The formulas about means and variances of linear functions that we had for discrete random variables are also the same. Suppose Then Y = c 0 + c 1 X 1 + c 2 X 2 E [Y ] = c 0 + c 1 E [X 1 ] + c 2 E [X 2 ] V [Y ] = c 2 1 V [X 1] + c 2 2 V [X 2] + 2c 1 c 2 COV [X 1, X 2 ] 67

Expected Value and Variance of Continuous RV s We could also define joint probabilities, conditional probabilities, independence, covariance, and correlation for continuous random variables. The ideas behind them are the same. It is still true that for two rv s X and Y : ρ XY = σ XY σ X σ Y and it is still true that 1 ρ XY 1. The only thing that s different is how we specify the probabilities. We specify probabilities of intervals instead of individual outcomes. 68

The Random Walk Model (again) Example: Remember the random walk model P t+1 = P t + E t+1 E t+1 N(0, 0.04) E t N(0, 0.04) for t = 1.... Given we know the price today P t = p t, what is the expected value of the price tomorrow based on this model? In other words, what is the mean of the conditional probability distribution P(P t+1 P t = p t )? 69

The Random Walk Model (again) We take the (conditional) expected value of each side: E[P t+1 P t = p t ] = E[P t + E t+1 ] = p t + E[E t+1 ] = p t Given that we know today s price, the price P t = p t is a constant. The expectation of E t+1 is E[E t+1 ] = 0. The best guess of tomorrow s price is today s price. 70

The Chi-Square and Student s t Distributions 71

The Chi-Square and Student s t Distributions We will take a brief look at two continuous distributions: 1. the Chi-squared distribution: χ 2 2. the Student s t distribution: t Both of these distributions are related to the normal distribution. 72

The Chi-Square Distribution Suppose that we have a sequence of standard normal random variables Z 1, Z 2, Z 3,..., Z k i.i.d. N(0, 1). Next, define a new random variable X to be the sum of squared Z i s. X = Z 2 1 + Z 2 2 + + Z 2 k The random variable X has a Chi-squared distribution: X χ 2 (k) The χ 2 (k) distribution is a family of distributions for different values of the parameter k. The parameter k is known as the degrees of freedom. 73

The Chi-Square Distribution What does the χ 2 (k) density look like? 0.250 0.10 0.225 0.200 0.175 0.150 χ 2 (3) 0.09 0.08 0.07 0.06 χ 2 (10) 0.125 0.05 0.100 0.04 0.075 0.03 0.050 0.02 0.025 0.01 0 2 4 6 8 10 12 14 16 18 20 0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0 22.5 25.0 The χ 2 distribution is skewed to the right. However, as k gets larger, the density flattens out and becomes symmetric. 74

The Chi-Square Distribution Why is the χ 2 distribution useful? Look at the sample variance: s 2 x = 1 n 1 n (x i x) 2 i=1 If we assume our data x 1, x 2,..., x n are i.i.d. normal, then the sample variance looks chi-square! 75

The Chi-Square Distribution Just so you know, if then X χ 2 (k) f (x) = 1 2 k/2 Γ ( )x k 2 1 e x 2 k 2 is the probability density function (pdf). In this class, we will not need this. But, we will see the chi-squared distribution later in the course when we build confidence intervals for the sample variance. 76

The Chi-Square Distribution Finally, we note that the mean of a χ 2 (k) random variable is: E[X ] = k and the variance is V[X ] = 2k 77

The Student s t Distribution Suppose we have two independent random variables Z and X whose distributions are Z N(0, 1) and X χ 2 (k). Define a new random variable T as: T = Z X /k The random variable T has a Student s t distribution written as: T t(k) The parameter k is known as the degrees of freedom. 78

The Student s t Distribution The random variable T with a Student s t distribution T t(k) describes a standard normal random variable divided by the square root of a normalized chi-square (a chi-square with k d.f., divided by k). 79

The Student s t Distribution What does a Student s t distribution look like? 0.40 0.35 N(0,1) in red 0.30 0.25 t 30 in green 0.20 0.15 0.10 t 3 0.05 4 3 2 1 0 1 2 3 4 The t-distribution has fatter tails (higher kurtosis) than the normal distribution. 80

The Student s t Distribution 0.40 0.35 N(0,1) in red 0.30 0.25 t 30 in green 0.20 0.15 0.10 t 3 0.05 4 3 2 1 0 1 2 3 4 At about 30 degrees of freedom and higher, though, the t and the normal are pretty much the same. 81

The Student s t Distribution Why is the Student s t distribution useful? There are many uses of a Student s t distribution and in particular 1. It is used as a model for asset returns. 2. It is used to construct confidence intervals (we will see this in Lecture # 8). 82

An i.i.d. Student s t Model Earlier we stated that it is common for data to appear approximately normal. 5000 4000 3000 2000 1000 Histogram of the daily returns on the Dow Jones from 1928 to 2010. 0.08 0.06 0.04 0.02 0.00 0.02 0.04 0.06 0.08 83

An i.i.d. Student s t Model However, returns often have heavy tails. Occasionally, we observe very large (negative) returns, making the i.i.d. normal model unrealistic. 5000 4000 3000 2000 1000 0.08 0.06 0.04 0.02 0.00 0.02 0.04 0.06 0.08 84

An i.i.d. Student s t Model A Student s t distribution with a small degrees of freedom, say ν 4, assigns more probability mass to the tails. Consequently, it may be a better approximation than the normal model. 5000 4000 3000 2000 1000 0.08 0.06 0.04 0.02 0.00 0.02 0.04 0.06 0.08 85