Central Limit Thm, Normal Approximations

Similar documents
Bernoulli and Binomial Distributions

STAT Chapter 7: Central Limit Theorem

No, because np = 100(0.02) = 2. The value of np must be greater than or equal to 5 to use the normal approximation.

Geometric & Negative Binomial Distributions

STAT 241/251 - Chapter 7: Central Limit Theorem

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 10: Continuous RV Families. Prof. Vince Calhoun

Math 14 Lecture Notes Ch The Normal Approximation to the Binomial Distribution. P (X ) = nc X p X q n X =

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Engineering Statistics ECIV 2305

STAT 111 Recitation 4

MATH 3200 Exam 3 Dr. Syring

Tutorial 11: Limit Theorems. Baoxiang Wang & Yihan Zhang bxwang, April 10, 2017

5.3 Statistics and Their Distributions

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

STAT 111 Recitation 3

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Business Statistics 41000: Probability 4

Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is

6 Central Limit Theorem. (Chs 6.4, 6.5)

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

CIVL Discrete Distributions

Central Limit Theorem, Joint Distributions Spring 2018

Binomial Distribution. Normal Approximation to the Binomial

IEOR 3106: Introduction to OR: Stochastic Models. Fall 2013, Professor Whitt. Class Lecture Notes: Tuesday, September 10.

Statistics for Business and Economics

STAT/MATH 395 PROBABILITY II

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

Central Limit Theorem (cont d) 7/28/2006

BIOL The Normal Distribution and the Central Limit Theorem

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

Section 7.5 The Normal Distribution. Section 7.6 Application of the Normal Distribution

STAT 830 Convergence in Distribution

Statistical Tables Compiled by Alan J. Terry

The Bernoulli distribution

Exam 2 Spring 2015 Statistics for Applications 4/9/2015

BIO5312 Biostatistics Lecture 5: Estimations

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

AMS7: WEEK 4. CLASS 3

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

5. In fact, any function of a random variable is also a random variable

Central Limit Theorem (CLT) RLS

The Normal Probability Distribution

4.3 Normal distribution

STA258H5. Al Nosedal and Alison Weir. Winter Al Nosedal and Alison Weir STA258H5 Winter / 41

15.063: Communicating with Data Summer Recitation 4 Probability III

Elementary Statistics Lecture 5

Statistics 251: Statistical Methods Sampling Distributions Module

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman:

Lecture 8. The Binomial Distribution. Binomial Distribution. Binomial Distribution. Probability Distributions: Normal and Binomial

Chapter 5. Statistical inference for Parametric Models

CIVL Learning Objectives. Definitions. Discrete Distributions

Probability Distributions for Discrete RV

Normal distribution Approximating binomial distribution by normal 2.10 Central Limit Theorem

MA131 Lecture 9.1. = µ = 25 and σ X P ( 90 < X < 100 ) = = /// σ X

The graph of a normal curve is symmetric with respect to the line x = µ, and has points of

MidTerm 1) Find the following (round off to one decimal place):

Chapter 7: Point Estimation and Sampling Distributions

Random Variables Handout. Xavier Vilà

Sampling Distribution

Commonly Used Distributions

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Review. Binomial random variable

Favorite Distributions

Populations and Samples Bios 662

Homework Assignments

Taylor Series & Binomial Series

The Normal Approximation to the Binomial

What was in the last lecture?

Covariance and Correlation. Def: If X and Y are JDRVs with finite means and variances, then. Example Sampling

Statistics for Business and Economics: Random Variables:Continuous

Chapter 7: SAMPLING DISTRIBUTIONS & POINT ESTIMATION OF PARAMETERS

Statistics and Their Distributions

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

Binomial Random Variables. Binomial Random Variables

PROBABILITY DISTRIBUTIONS

Chapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables

Week 7. Texas A& M University. Department of Mathematics Texas A& M University, College Station Section 3.2, 3.3 and 3.4

Midterm Exam III Review

Introduction to Statistical Data Analysis II

Chapter 4 Continuous Random Variables and Probability Distributions

Introduction to Probability

ECO220Y Continuous Probability Distributions: Normal Readings: Chapter 9, section 9.10

Chapter 3 - Lecture 4 Moments and Moment Generating Funct

Math 227 Elementary Statistics. Bluman 5 th edition

Confidence Intervals Introduction

Discrete Random Variables and Probability Distributions. Stat 4570/5570 Based on Devore s book (Ed 8)

Probability Theory. Probability and Statistics for Data Science CSE594 - Spring 2016

Data Analysis and Statistical Methods Statistics 651

Probability Theory. Mohamed I. Riffi. Islamic University of Gaza

SOME MOST POPULAR DISTRIBUTIONS OF RANDOM VARIABLES

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

Drunken Birds, Brownian Motion, and Other Random Fun

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

STAT Chapter 5: Continuous Distributions. Probability distributions are used a bit differently for continuous r.v. s than for discrete r.v. s.

Lecture Stat 302 Introduction to Probability - Slides 15

Random Variable: Definition

Transcription:

Central Limit Thm, Normal Approximations Engineering Statistics Section 5.4 Josh Engwer TTU 23 March 2016 Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 1 / 26

PART I PART I: CENTRAL LIMIT THEOREM Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 2 / 26

Expected Value & Variance of a Sum of iid rv s Proposition Let X 1,..., X n be a random sample from a population and c 1,..., c n 0. Then: E[c 1 X 1 + + c n X n ] = c 1 E[X 1 ] + + c n E[X n ] V[c 1 X 1 + + c n X n ] = c 2 1 V[X 1] + + c 2 nv[x n ] iid PROOF: CASE I (n = 2): X 1, X 2 (discrete population) E[c 1 X 1 + c 2 X 2 ] := (c 1 j + c 2 k) p (X1,X 2)(j, k) iid = (j,k) Supp(X 1,X 2) j Supp(X 1) k Supp(X 2) = c 1 j Supp(X 1) := c 1 E[X 1 ] + c 2 E[X 2 ] j p X1 (j) + c 2 (c 1 j + c 2 k) p X1 (j) p X2 (k) k Supp(X 2) k p X2 (k) Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 3 / 26

Expected Value & Variance of a Sum of iid rv s Proposition Let X 1,..., X n be a random sample from a population and c 1,..., c n 0. Then: PROOF: E[c 1 X 1 + + c n X n ] = c 1 E[X 1 ] + + c n E[X n ] V[c 1 X 1 + + c n X n ] = c 2 1 V[X 1] + + c 2 nv[x n ] E[c 1 X 1 + c 2 X 2 ] := CASE II (n = 2): X 1, X 2 iid (continuous population) iid = Supp(X 1,X 2) = c 1 Supp(X 1) (c 1 x 1 + c 2 x 2 ) f (X1,X 2)(x 1, x 2 ) dx 1 dx 2 Supp(X 1) Supp(X 2) := c 1 E[X 1 ] + c 2 E[X 2 ] (requires independence) (c 1 x 1 + c 2 x 2 ) f X1 (x 1 ) f X2 (x 2 ) dx 1 dx 2 x 1 f X1 (x 1 ) dx 1 + c 2 Supp(X 2) x 2 f X2 (x 2 ) dx 2 Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 4 / 26

Properties of Sample Mean Proposition Let X 1,..., X n be a random sample from a distribution with mean µ and variance σ 2. Then: PROOF: µ X = E[X] = µ σ 2 X = V[X] = σ2 /n σ X = σ/ n E[X] = E [ 1 n (X 1 + + X n ) ] = 1 n [E[X 1] + + E[X n ]] = 1 n [µ + + µ] = 1 n [nµ] = µ V[X] = V [ 1 n (X 1 + + X n ) ] = 1 [V[X n 2 1 ] + + V[X n ]] = 1 [σ 2 + + σ 2 ] = 1 [nσ 2 ] = σ 2 /n n 2 n 2 σ X = V[X] = σ 2 /n = σ/ n Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 5 / 26

Properties of Sample Total Proposition Let X 1,..., X n be a random sample from a distribution with mean µ and variance σ 2. Then: PROOF: µ X1+ +X n = E[X 1 + + X n ] = nµ σ 2 X 1+ +X n = V[X 1 + + X n ] = nσ 2 σ X1+ +X n = σ n E[X 1 + + X n ] = E [ nx ] = ne [ X ] = n µ = nµ V[X 1 + + X n ] = V [ nx ] = n 2 V [ X ] = n 2 [σ 2 /n] = nσ 2 σ X1+ +X n = V[X 1 + + X n ] = nσ 2 = σ n Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 6 / 26

Sample Mean & Sample Total of a Random Sample from a Normal Population The proceeding properties of sample means & sample variances can be applied to particular population distributions: Proposition Let random sample X 1,..., X n iid Normal(µ, σ 2 ). For any sample size n > 1, Then: X Normal(µ, σ 2 /n) For any sample size n > 1, X 1 + + X n Normal(nµ, nσ 2 ) PROOF: X Normal(µ, σ 2 ) = E[X] = µ, V[X] = σ 2 E[X] = E [ 1 n (X 1 + + X n ) ] = 1 n [E[X 1] + + E[X n ]] = 1 n [µ + + µ] = 1 n [nµ] = µ V[X] = V [ 1 n (X 1 + + X n ) ] = 1 [V[X n 2 1 ] + + V[X n ]] = 1 [σ 2 + + σ 2 ] = 1 [nσ 2 ] = σ 2 /n n 2 n 2 X Normal(µ, σ 2 /n) Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 7 / 26

Sample Mean & Sample Total of a Random Sample from a Normal Population The proceeding properties of sample means & sample variances can be applied to particular population distributions: Proposition Let random sample X 1,..., X n iid Normal(µ, σ 2 ). For any sample size n > 1, Then: X Normal(µ, σ 2 /n) For any sample size n > 1, X 1 + + X n Normal(nµ, nσ 2 ) PROOF: X Normal(µ, σ 2 ) = E[X] = µ, V[X] = σ 2 E[X 1 + + X n ] = E [ nx ] = ne [ X ] = n µ = nµ V[X 1 + + X n ] = V [ nx ] = n 2 V [ X ] = n 2 [σ 2 /n] = nσ 2 X 1 + + X n Normal(nµ, nσ 2 ) Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 8 / 26

Central Limit Theorem (CLT) The Central Limit Thm is considered a fundamental theorem of Statistics: Theorem (Central Limit Theorem) Let X 1,..., X n be a random sample from a non-normal distribution with mean µ and variance σ 2. Then: The sample mean X is approximately normal as follows: X approx Normal(µ, σ 2 /n) The sample total X 1 + + X n is approximately normal as follows: X 1 + + X n approx Normal(nµ, nσ 2 ) Requirement for this normal approximation to be valid: n > 30 The larger the sample size n, the better the approximation. PROOF: Beyond the scope of this course. Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 9 / 26

PART II PART II: NORMAL APPROXIMATION TO THE BINOMIAL Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 10 / 26

The Sum of several iid Bernoulli(p) rv s Proposition Let random sample X 1,..., X m iid Bernoulli(p). Then: For any sample size n > 1, X 1 + + X n Binomial(n, p) PROOF: X Bernoulli(p) = E[X] = p, V[X] = pq (q = 1 p) E[X 1 + + X n ] = E[X 1 ] + + E[X n ] iid = n E[X 1 ] = n p = np V[X 1 + + X n ] iid = V[X 1 ] + + V[X n ] iid = n V[X 1 ] = n pq = npq X 1 + + X n Binomial(n, p) Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 11 / 26

Normal Approximation to the Binomial Corollary Let X Binomial(n, p). Then: X approx Normal(µ = np, σ 2 = npq) (where q = 1 p) Requirement for this normal approximation to be valid: min{np, nq} 10 i.e. It s required that both np 10 and nq 10. NOTE: Remember that normal distributions are symmetric. If min{np, nq} < 10, then the binomial distribution is too skewed. PROOF: Let X 1,..., X n Bernoulli(p) = µ X1 = E[X 1 ] = p and σ 2 X 1 = V[X 1 ] = pq. Then X := X 1 + + X n Binomial(n, p). Moreover, the CLT asserts that X 1 + + X n approx Normal(µ = nµ X1, σ 2 = nσ 2 X 1 ) X Binomial(n, p) = X approx Normal(µ = np, σ 2 = npq) Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 12 / 26

Normal Approximation to Binomial Probability Corollary Let X Binomial(n, p). Then: P(X x) = Bi(x; n, p) Φ ( ) x + 0.5 np npq (where q = 1 p) Requirement for this normal approximation to be valid: min{np, nq} 10 NOTE: The continuity correction term + 0.5 improves the approximation. PROOF: Assume that the requirement min{np, nq} 10 is satisfied. Then: X Binomial(n, p) = X approx Normal(µ = np, σ 2 = npq) ( = Bi(x; n, p) = P(X x) P Z x µ ) ( ) ( ) x µ x np = Φ = Φ σ σ npq Add continuity correction + 0.5 in numerator of Φ to better the approximation. Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 13 / 26

Binomial Density Plots (pmf s) for Sample Size n = 2 Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 14 / 26

Binomial Density Plots (pmf s) for Sample Size n = 5 Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 15 / 26

Binomial Density Plots (pmf s) for Sample Size n = 10 Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 16 / 26

Binomial Density Plots (pmf s) for Sample Size n = 20 Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 17 / 26

Binomial Density Plots (pmf s) for Sample Size n = 50 Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 18 / 26

PART III PART III: NORMAL APPROXIMATION TO THE POISSON Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 19 / 26

The Sum of several iid Poisson(λ) rv s The proceeding properties of sample totals can be applied to particular population distributions: Proposition Let random sample X 1,..., X n iid Poisson(λ). For any sample size n > 1, Then: X 1 + + X n Poisson(nλ) PROOF: X Poisson(λ) = E[X] = λ, V[X] = λ E[X 1 + + X n ] = E[X 1 ] + + E[X n ] iid = n E[X 1 ] = n λ = nλ V[X 1 + + X n ] iid = V[X 1 ] + + V[X n ] iid = n V[X 1 ] = n λ = nλ X 1 + + X n Poisson(nλ) Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 20 / 26

Normal Approximation to the Poisson Corollary Let X Poisson(λ). Then: X approx Normal(µ = λ, σ 2 = λ) Requirement for this normal approximation to be valid: λ > 20 NOTE: Remember that normal distributions are symmetric. If λ 20, then the binomial distribution is too skewed. PROOF: Let X 1,..., X n Poisson(λ/n) = µ X1 = E[X 1 ] = λ/n and σ 2 X 1 = V[X 1 ] = λ/n. Then X := X 1 + + X n Poisson(λ). Moreover, the CLT asserts that X 1 + + X n approx Normal(µ = nµ X1, σ 2 = nσ 2 X 1 ) X Poisson(λ) = X approx Normal(µ = λ, σ 2 = λ) Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 21 / 26

Normal Approximation to Poisson Probability Corollary Let X Poisson(λ). Then: P(X x) = Pois(x; λ) Φ ( ) x + 0.5 λ λ Requirement for this normal approximation to be valid: λ > 20 NOTE: The continuity correction term + 0.5 improves the approximation. PROOF: Assume that the requirement λ > 20 is satisfied. Then: X Poisson(λ) = X approx Normal(µ = λ, σ 2 = λ) ( = Pois(x; λ) = P(X x) P Z x µ ) ( ) ( ) x µ x λ = Φ = Φ σ σ λ Add continuity correction + 0.5 in numerator of Φ to better the approximation. Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 22 / 26

Poisson Density Plots (pmf s) Notice the Poisson density curves for λ = 1, 2, 5, 10, 15 are skewed. Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 23 / 26

Poisson Density Plots (pmf s) Notice the Poisson density curves for λ 20 are nearly symmetric. Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 24 / 26

Textbook Logistics for Section 5.4 Difference(s) in Notation: CONCEPT TEXTBOOK NOTATION SLIDES/OUTLINE NOTATION Probability of Event P(A) P(A) Support of a r.v. All possible values of X Supp(X) pmf of a r.v. p X (x) p X (k) Expected Value of r.v. E(X) E[X] Variance of r.v. V(X) V[X] Sample Total T o Xk pmf of Sample Mean p X (x) p X (k) pmf of Sample Variance p S 2(s 2 ) p S 2(k) Ignore Lognormal Approximation (bottom of pg 236) The Lognormal distribution was part of Section 4.5 that was skipped Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 25 / 26

Fin Fin. Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 26 / 26