Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is

Similar documents
continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 10: Continuous RV Families. Prof. Vince Calhoun

Commonly Used Distributions

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

The Normal Distribution

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

STAT Chapter 5: Continuous Distributions. Probability distributions are used a bit differently for continuous r.v. s than for discrete r.v. s.

Statistics for Business and Economics

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

Central Limit Theorem, Joint Distributions Spring 2018

What was in the last lecture?

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Probability Distributions for Discrete RV

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

STAT Chapter 5: Continuous Distributions. Probability distributions are used a bit differently for continuous r.v. s than for discrete r.v. s.

The Normal Distribution. (Ch 4.3)

Consider the following examples: ex: let X = tossing a coin three times and counting the number of heads

Topic 6 - Continuous Distributions I. Discrete RVs. Probability Density. Continuous RVs. Background Reading. Recall the discrete distributions

PROBABILITY DISTRIBUTIONS

4.3 Normal distribution

Business Statistics 41000: Probability 3

ECO220Y Continuous Probability Distributions: Normal Readings: Chapter 9, section 9.10

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Normal distribution Approximating binomial distribution by normal 2.10 Central Limit Theorem

Chapter 3 Statistical Quality Control, 7th Edition by Douglas C. Montgomery. Copyright (c) 2013 John Wiley & Sons, Inc.

MidTerm 1) Find the following (round off to one decimal place):

The Bernoulli distribution

Continuous random variables

Chapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Binomial Random Variables. Binomial Random Variables

AP STATISTICS FALL SEMESTSER FINAL EXAM STUDY GUIDE

Lecture Slides. Elementary Statistics Tenth Edition. by Mario F. Triola. and the Triola Statistics Series. Slide 1

Math 14 Lecture Notes Ch The Normal Approximation to the Binomial Distribution. P (X ) = nc X p X q n X =

2011 Pearson Education, Inc

The normal distribution is a theoretical model derived mathematically and not empirically.

Continuous Distributions

Section 7.5 The Normal Distribution. Section 7.6 Application of the Normal Distribution

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

ECON 214 Elements of Statistics for Economists 2016/2017

STATISTICS and PROBABILITY

6. Continous Distributions

Chapter 4 Continuous Random Variables and Probability Distributions

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman:

No, because np = 100(0.02) = 2. The value of np must be greater than or equal to 5 to use the normal approximation.

Exam 2 Spring 2015 Statistics for Applications 4/9/2015

Lecture 12. Some Useful Continuous Distributions. The most important continuous probability distribution in entire field of statistics.

Lecture 6: Chapter 6

CH 5 Normal Probability Distributions Properties of the Normal Distribution

X = x p(x) 1 / 6 1 / 6 1 / 6 1 / 6 1 / 6 1 / 6. x = 1 x = 2 x = 3 x = 4 x = 5 x = 6 values for the random variable X

Chapter 2 ( ) Fall 2012

Introduction to Business Statistics QM 120 Chapter 6

Week 7. Texas A& M University. Department of Mathematics Texas A& M University, College Station Section 3.2, 3.3 and 3.4

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

5.3 Statistics and Their Distributions

STA258H5. Al Nosedal and Alison Weir. Winter Al Nosedal and Alison Weir STA258H5 Winter / 41

Probability. An intro for calculus students P= Figure 1: A normal integral

14.30 Introduction to Statistical Methods in Economics Spring 2009

Introduction to Statistics I

Populations and Samples Bios 662

Lecture 8. The Binomial Distribution. Binomial Distribution. Binomial Distribution. Probability Distributions: Normal and Binomial

Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making

Chapter 7 1. Random Variables

5.4 Normal Approximation of the Binomial Distribution

Continuous Random Variables and Probability Distributions

Central Limit Thm, Normal Approximations

CHAPTERS 5 & 6: CONTINUOUS RANDOM VARIABLES

Statistics and Probability

Discrete Random Variables and Probability Distributions. Stat 4570/5570 Based on Devore s book (Ed 8)

The Normal Distribution

Bivariate Birnbaum-Saunders Distribution

Math 227 Elementary Statistics. Bluman 5 th edition

Statistical Methods in Practice STAT/MATH 3379

Chapter 4 Continuous Random Variables and Probability Distributions

A random variable (r. v.) is a variable whose value is a numerical outcome of a random phenomenon.

Statistics 431 Spring 2007 P. Shaman. Preliminaries

8.2 The Standard Deviation as a Ruler Chapter 8 The Normal and Other Continuous Distributions 8-1

MA : Introductory Probability

Homework Problems Stat 479

STAT Chapter 7: Central Limit Theorem

Basic notions of probability theory: continuous probability distributions. Piero Baraldi

Binomial Distribution. Normal Approximation to the Binomial

Lecture 10: Point Estimation

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Section Introduction to Normal Distributions

A random variable (r. v.) is a variable whose value is a numerical outcome of a random phenomenon.

Random Variables and Probability Distributions

Statistical Intervals. Chapter 7 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

MAS187/AEF258. University of Newcastle upon Tyne

Random variables. Contents

IEOR 165 Lecture 1 Probability Review

STAT 241/251 - Chapter 7: Central Limit Theorem

Homework: Due Wed, Nov 3 rd Chapter 8, # 48a, 55c and 56 (count as 1), 67a

Business Statistics. Chapter 5 Discrete Probability Distributions QMIS 120. Dr. Mohammad Zainal

Unit 5: Sampling Distributions of Statistics

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

Transcription:

Normal Distribution

Normal Distribution Definition A continuous rv X is said to have a normal distribution with parameter µ and σ (µ and σ 2 ), where < µ < and σ > 0, if the pdf of X is f (x; µ, σ) = 1 2πσ e (x µ)2 /(2σ 2 ) We use the notation X N(µ, σ 2 ) to denote that X is rormally distributed with parameters µ and σ 2.

Normal Distribution Proposition For X N(µ, σ 2 ), we have E(X ) = µ and V (X ) = σ 2

Normal Distribution Definition The normal distribution with parameter values µ = 0 and σ = 1 is called the standard normal distribution. A random variable having a standard normal distribution is called a standard normal random variable and will be denoted by Z. The pdf of Z is f (z; 0, 1) = 1 2π e z2 /2 < z < The graph of f (z; 0, 1) is called the standard normal (or z) curve. The cdf of Z is P(Z z) = z f (y; 0, 1)dy, which we will denote by Φ(z).

Normal Distribution Proposition If X has a normal distribution with mean µ and stadard deviation σ, then Z = X µ σ has a standard normal distribution. Thus P(a X b) = P( a µ Z b µ σ σ ) = Φ( b µ σ ) Φ(a µ σ ) P(X a) = Φ( a µ µ ) P(X b) = 1 Φ(b σ σ )

Normal Distribution

Normal Distribution Proposition {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} σ

Normal Distribution

Normal Distribution Proposition Let X be a binomial rv based on n trials with success probability p. Then if the binomial probability histogram is not too skewed, X has approximately a normal distribution with µ = np and σ = npq, where q = 1 p. In particular, for x = a posible value of X, ( ) area under the normal curve P(X x) = B(x; n, p) to the left of x+0.5 x+0.5 np = Φ( ) npq In practice, the approximation is adequate provided that both np 10 and nq 10, since there is then enough symmetry in the underlying binomial distribution.

Normal Distribution

Normal Distribution A graphical explanation for ( ) area under the normal curve P(X x) = B(x; n, p) to the left of x+0.5 x+0.5 np = Φ( ) npq

Normal Distribution A graphical explanation for ( ) area under the normal curve P(X x) = B(x; n, p) to the left of x+0.5 x+0.5 np = Φ( ) npq

Normal Distribution

Normal Distribution Example (Problem 54) Suppose that 10% of all steel shafts produced by a certain process are nonconforming but can be reworked (rather than having to be scrapped). Consider a random sample of 200 shafts, and let X denote the number among these that are nonconforming and can be reworked. What is the (approximate) probability that X is between 15 and 25 (inclusive)?

Normal Distribution Example (Problem 54) Suppose that 10% of all steel shafts produced by a certain process are nonconforming but can be reworked (rather than having to be scrapped). Consider a random sample of 200 shafts, and let X denote the number among these that are nonconforming and can be reworked. What is the (approximate) probability that X is between 15 and 25 (inclusive)? In this problem n = 200, p = 0.1 and q = 1 p = 0.9. Thus np = 20 > 10 and nq = 180 > 10

Normal Distribution Example (Problem 54) Suppose that 10% of all steel shafts produced by a certain process are nonconforming but can be reworked (rather than having to be scrapped). Consider a random sample of 200 shafts, and let X denote the number among these that are nonconforming and can be reworked. What is the (approximate) probability that X is between 15 and 25 (inclusive)? In this problem n = 200, p = 0.1 and q = 1 p = 0.9. Thus np = 20 > 10 and nq = 180 > 10 P(15 X 25) = Bin(25; 200, 0.1) Bin(14; 200, 0.1) 25 + 0.5 20 15 + 0.5 20 Φ( ) Φ( ) 200 0.1 0.9 200 0.1 0.9 = Φ(0.3056) Φ( 0.2500) = 0.6217 0.4013 = 0.2204

Exponential Distribution

Exponential Distribution Definition X is said to have an exponential distribution with parameter λ(λ > 0) if the pdf of X is { λe λx x 0 f (x; λ) = 0 otherwise

Exponential Distribution

Exponential Distribution Proposition If X EXP(λ), then And the cdf for X is E(X ) = 1 λ and V (X ) = 1 λ 2 F (x; λ) = { 1 e λx x 0 0 x < 0

Exponential Distribution

Exponential Distribution Memoryless Property for any positive t and t 0. P(X t) = P(X t + t 0 X t 0 )

Exponential Distribution Memoryless Property P(X t) = P(X t + t 0 X t 0 ) for any positive t and t 0. In words, the distribution of additional lifetime is exactly the same as the original distribution of lifetime, so at each point in time the component shows no effect of wear. In other words, the distribution of remaining lifetime is independent of current age.

Example: There is a machine available for cutting corks intended for use in wine bottles. We want to find out the distribution of the diameters of the corks produced by that machine. Assume we have 10 samples produced by that machine and the diameters is recorded as following: 3.0879 3.2546 2.8970 2.7377 2.7740 2.6030 3.5931 3.1253 2.4756 2.5133

3.0879 3.2546 2.8970 2.7377 2.7740 2.6030 3.5931 3.1253 2.4756 2.5133

3.0879 3.2546 2.8970 2.7377 2.7740 2.6030 3.5931 3.1253 2.4756 2.5133

Sample Percentile

Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X, denoted by η(p), is defined by p = F (η(p)) = η(p) f (y)dy.

Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X, denoted by η(p), is defined by p = F (η(p)) = η(p) f (y)dy. In words, the (100p)th percentile η(p) is the X value such that there are 100p% X values below η(p).

Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X, denoted by η(p), is defined by p = F (η(p)) = η(p) f (y)dy. In words, the (100p)th percentile η(p) is the X value such that there are 100p% X values below η(p). Similarly, we can define sample percentile in the same manner, i.e. the (100p)th percentile x p is the value such that there are 100p% sample values below x p.

Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X, denoted by η(p), is defined by p = F (η(p)) = η(p) f (y)dy. In words, the (100p)th percentile η(p) is the X value such that there are 100p% X values below η(p). Similarly, we can define sample percentile in the same manner, i.e. the (100p)th percentile x p is the value such that there are 100p% sample values below x p. Unfortunately, x p may not be a sample value for some p.

Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X, denoted by η(p), is defined by p = F (η(p)) = η(p) f (y)dy. In words, the (100p)th percentile η(p) is the X value such that there are 100p% X values below η(p). Similarly, we can define sample percentile in the same manner, i.e. the (100p)th percentile x p is the value such that there are 100p% sample values below x p. Unfortunately, x p may not be a sample value for some p. e.g. for the previous example, what is the 35th percentile for the ten sample values?

Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i 0.5)/n]th sample percentile.

Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i 0.5)/n]th sample percentile. Remark: 1. Why i 0.5?

Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i 0.5)/n]th sample percentile. Remark: 1. Why i 0.5? We regard the sample observation as being half in the lower group and half in the upper group.

Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i 0.5)/n]th sample percentile. Remark: 1. Why i 0.5? We regard the sample observation as being half in the lower group and half in the upper group. e.g. if n = 9, then the sample median is the 5th largest observation and this observation is regarded as two parts: one in the lower half and one in the upper half.

Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i 0.5)/n]th sample percentile. Remark: 1. Why i 0.5? We regard the sample observation as being half in the lower group and half in the upper group. e.g. if n = 9, then the sample median is the 5th largest observation and this observation is regarded as two parts: one in the lower half and one in the upper half. 2. Once the percentage values 100(i 0.5)/n(i = 1, 2,..., n) have been calculated, sample percentiles corresponding to intermediate percentages can be obtained by linear interpolation.

Example: for the previous example, the [100(i 0.5)/n]th sample percentile is tabulated as following: 2.4756 2.5133 2.6030 100(1-.5)/10 = 5% 100(2-.5)/10 = 15% 100(3-.5)/10 = 25% 2.7377 2.7740 100(4-.5)/10 = 35% 100(5-.5)/10 = 45% 2.8970 3.0879 3.1253 100(6-.5)/10 = 55% 100(7-.5)/10 = 65% 100(8-.5)/10 = 75% 3.2546 3.5931 100(9-.5)/10 = 85% 100(10-.5)/10 = 95%

Example: for the previous example, the [100(i 0.5)/n]th sample percentile is tabulated as following: 2.4756 2.5133 2.6030 100(1-.5)/10 = 5% 100(2-.5)/10 = 15% 100(3-.5)/10 = 25% 2.7377 2.7740 100(4-.5)/10 = 35% 100(5-.5)/10 = 45% 2.8970 3.0879 3.1253 100(6-.5)/10 = 55% 100(7-.5)/10 = 65% 100(8-.5)/10 = 75% 3.2546 3.5931 100(9-.5)/10 = 85% 100(10-.5)/10 = 95% The 10th percentile would be (2.4756 + 2.5133)/2 = 2.49445

Idea for Quantile-Quantile Plot: 1. Determine the [100(i 0.5)/n]th sample percentile for a given sample.

Idea for Quantile-Quantile Plot: 1. Determine the [100(i 0.5)/n]th sample percentile for a given sample. 2. Find the corresponding [100(i 0.5)/n]th percentile from the population with the assumed distribution; for example, if the assumed distribution is standard normal, then find corresponding [100(i 0.5)/n]th percentile from the standard normal distribution.

Idea for Quantile-Quantile Plot: 1. Determine the [100(i 0.5)/n]th sample percentile for a given sample. 2. Find the corresponding [100(i 0.5)/n]th percentile from the population with the assumed distribution; for example, if the assumed distribution is standard normal, then find corresponding [100(i 0.5)/n]th percentile from the standard normal distribution. 3. Consider the (population percentile, sample percentile) pairs, i.e. ([100(i ) 0.5)/n]th percentile, ith smallest sample of the distribution observation

Idea for Quantile-Quantile Plot: 1. Determine the [100(i 0.5)/n]th sample percentile for a given sample. 2. Find the corresponding [100(i 0.5)/n]th percentile from the population with the assumed distribution; for example, if the assumed distribution is standard normal, then find corresponding [100(i 0.5)/n]th percentile from the standard normal distribution. 3. Consider the (population percentile, sample percentile) pairs, i.e. ([100(i ) 0.5)/n]th percentile, ith smallest sample of the distribution observation 4. Each pair plotted as a point on a two-dimensional coordinate system should fall close to a 45 line.

4. Each pair plotted as a point on a two-dimensional coordinate system should fall close to a 45 line. Substantial deviations of the plotted points from a 45 line cast doubt on the assumption that the distribution under consideration is the correct one. Probability Plot Idea for Quantile-Quantile Plot: 1. Determine the [100(i 0.5)/n]th sample percentile for a given sample. 2. Find the corresponding [100(i 0.5)/n]th percentile from the population with the assumed distribution; for example, if the assumed distribution is standard normal, then find corresponding [100(i 0.5)/n]th percentile from the standard normal distribution. 3. Consider the (population percentile, sample percentile) pairs, i.e. ([100(i ) 0.5)/n]th percentile, ith smallest sample of the distribution observation

Example 4.29: The value of a certain physical constant is known to an experimenter. The experimenter makes n = 10 independent measurements of this value using a particular measurement device and records the resulting measurement errors (error = observed value - true value). These observations appear in the following table. Percentage 5 15 25 35 45 Sample Observation -1.91-1.25-0.75-0.53 0.20 Percentage 55 65 75 85 95 Sample Observation 0.35 0.72 0.87 1.40 1.56

Example 4.29: The value of a certain physical constant is known to an experimenter. The experimenter makes n = 10 independent measurements of this value using a particular measurement device and records the resulting measurement errors (error = observed value - true value). These observations appear in the following table. Percentage 5 15 25 35 45 Sample Observation -1.91-1.25-0.75-0.53 0.20 Percentage 55 65 75 85 95 Sample Observation 0.35 0.72 0.87 1.40 1.56 Is it plausible that the random variable measurement error has standard normal distribution?

We first find the corresponding population distribution percentiles, in this case, the z percentiles: Percentage 5 15 25 35 45 Sample Observation -1.91-1.25-0.75-0.53 0.20 z percentile -1.645-1.037-0.675-0.385-0.126 Percentage 55 65 75 85 95 Sample Observation 0.35 0.72 0.87 1.40 1.56 z percentile 0.126 0.385 0.675 1.037 1.645

What about the first example? We are only interested in whether the ten sample observations come from a normal distribution.

What about the first example? We are only interested in whether the ten sample observations come from a normal distribution. Recall: {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} σ

What about the first example? We are only interested in whether the ten sample observations come from a normal distribution. Recall: {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} σ If µ = 0, then the pairs (σ [z percentile], observation) fall on a 45 line, which has slope 1.

What about the first example? We are only interested in whether the ten sample observations come from a normal distribution. Recall: {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} σ If µ = 0, then the pairs (σ [z percentile], observation) fall on a 45 line, which has slope 1. Therefore the pairs ([z percentile], observation) fall on a line passing through (0,0) (i.e., one with y-intercept 0) but having slope σ rather than 1.

What about the first example? We are only interested in whether the ten sample observations come from a normal distribution. Recall: {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} σ If µ = 0, then the pairs (σ [z percentile], observation) fall on a 45 line, which has slope 1. Therefore the pairs ([z percentile], observation) fall on a line passing through (0,0) (i.e., one with y-intercept 0) but having slope σ rather than 1. Now for µ 0, the y-intercept is µ instead of 0.

Normal Probability Plot A plot of the n pairs ([100(i 0.5)/n]th z percentile, ith smallest observation) on a two-dimensional coordinate system is called a normal probability plot. If the sample observations are in fact drawn from a normal distribution with mean value µ and standard deviation σ, the points should fall close to a straight line with slope σ and y-intercept µ. Thus a plot for which the points fall close to some straight line suggests that the assumption of a normal population distribution is plausible.

Percentage 5 15 25 35 45 Sample Observation 2.4756 2.5133 2.6030 2.7377 2.7740 z percentile -1.645-1.037-0.675-0.385-0.126 Percentage 55 65 75 85 95 Sample Observation 2.8970 3.0879 3.1253 3.2546 3.5931 z percentile 0.126 0.385 0.675 1.037 1.645

A nonnormal population distribution can often be placed in one of the following three categories: 1. It is symmetric and has lighter tails than does a normal distribution; that is, the density curve declines more rapidly out in the tails than does a normal curve. 2. It is symmetric and heavy-tailed compared to a normal distribution. 3. It is skewed.

Symmetric and light-tailed : e.g. Uniform distribution

Symmetric and heavy-tailed: e.g. Cauchy distribution with pdf f (x) = 1/[π(1 + x 2 )] for < x <

Skewed: e.g. lognormal distribution

Some guidances for probability plot for normal distributions (from the book Fitting Equations to Data (2nd ed.) Daniel, Cuthbert, and Fed Wood, Wiley, New York, 1980)

Some guidances for probability plot for normal distributions (from the book Fitting Equations to Data (2nd ed.) Daniel, Cuthbert, and Fed Wood, Wiley, New York, 1980) 1. For sample size smaller than 30, there is typically greater variation in the apperance of the probability plot.

Some guidances for probability plot for normal distributions (from the book Fitting Equations to Data (2nd ed.) Daniel, Cuthbert, and Fed Wood, Wiley, New York, 1980) 1. For sample size smaller than 30, there is typically greater variation in the apperance of the probability plot. 2. Only for much larger sample sizes does a linear pattern generally predominate.

Some guidances for probability plot for normal distributions (from the book Fitting Equations to Data (2nd ed.) Daniel, Cuthbert, and Fed Wood, Wiley, New York, 1980) 1. For sample size smaller than 30, there is typically greater variation in the apperance of the probability plot. 2. Only for much larger sample sizes does a linear pattern generally predominate. Therefore, when a plot is based on a small sample size, only a very substantial departure from linearity should be taken as conclusive evidence of nonnorality.

Definition Consider a family of probability distributions involving two parameters, θ 1 and θ 2, and let F (x; θ 1, θ 2 ) denote the corresponding cdf s. The parameters θ 1 and θ 2 are said to be location and scale parameters, respectively, if F (x; θ 1, θ 2 ) is a function of (x θ 1 )/θ 2.

Definition Consider a family of probability distributions involving two parameters, θ 1 and θ 2, and let F (x; θ 1, θ 2 ) denote the corresponding cdf s. The parameters θ 1 and θ 2 are said to be location and scale parameters, respectively, if F (x; θ 1, θ 2 ) is a function of (x θ 1 )/θ 2. e.g. 1. Normal distributions N(µ, σ): F (x; µ, σ) = Φ( x µ σ ).

Definition Consider a family of probability distributions involving two parameters, θ 1 and θ 2, and let F (x; θ 1, θ 2 ) denote the corresponding cdf s. The parameters θ 1 and θ 2 are said to be location and scale parameters, respectively, if F (x; θ 1, θ 2 ) is a function of (x θ 1 )/θ 2. e.g. 1. Normal distributions N(µ, σ): F (x; µ, σ) = Φ( x µ σ ). 2. The extreme value distribution with cdf F (x; θ 1, θ 2 ) = 1 e e(x θ 1 )/θ 2

For Weibull distribution: F (x; α, β) = 1 e (x/β)α, the parameter β is a scale parameter but α is NOT a location parameter. α is usually referred to as a shape parameter.

For Weibull distribution: F (x; α, β) = 1 e (x/β)α, the parameter β is a scale parameter but α is NOT a location parameter. α is usually referred to as a shape parameter. Fortunately, if X has a Weibull distribution with shape parameter α and scale parameter β, then the transformed variable ln(x ) has an extreme value distribution with location parameter θ 1 = ln(β) and scale parameter θ 2 = 1/α.

The gamma distribution also has a shape parameter α. However, there is no transformation h( ) such that h(x ) has a distribution that depends only on location and scale parameters.

The gamma distribution also has a shape parameter α. However, there is no transformation h( ) such that h(x ) has a distribution that depends only on location and scale parameters. Thus, before we construct a probability plot, we have to estimate the shape parameter from the sample data.

Jointly Distributed Random Variables

Jointly Distributed Random Variables Consider tossing a fair die twice. Then the outcomes would be (1,1) (1,2) (1,3) (1,4) (1,5) (1,6) (2,1) (2,2) (2,3) (2,4) (2,5) (2,6) (6,1) (6,2) (6,3) (6,4) (6,5) (6,6) and the probability for each outcome is 1 36.

Jointly Distributed Random Variables Consider tossing a fair die twice. Then the outcomes would be (1,1) (1,2) (1,3) (1,4) (1,5) (1,6) (2,1) (2,2) (2,3) (2,4) (2,5) (2,6) (6,1) (6,2) (6,3) (6,4) (6,5) (6,6) and the probability for each outcome is 1 36. If we define two random variables by X = the outcome of the first toss and Y = the outcome of the second toss,

Jointly Distributed Random Variables Consider tossing a fair die twice. Then the outcomes would be (1,1) (1,2) (1,3) (1,4) (1,5) (1,6) (2,1) (2,2) (2,3) (2,4) (2,5) (2,6) (6,1) (6,2) (6,3) (6,4) (6,5) (6,6) and the probability for each outcome is 1 36. If we define two random variables by X = the outcome of the first toss and Y = the outcome of the second toss, then the outcome for this experiment (two tosses) can be describe by the random pair (X, Y ),

Jointly Distributed Random Variables Consider tossing a fair die twice. Then the outcomes would be (1,1) (1,2) (1,3) (1,4) (1,5) (1,6) (2,1) (2,2) (2,3) (2,4) (2,5) (2,6) (6,1) (6,2) (6,3) (6,4) (6,5) (6,6) and the probability for each outcome is 1 36. If we define two random variables by X = the outcome of the first toss and Y = the outcome of the second toss, then the outcome for this experiment (two tosses) can be describe by the random pair (X, Y ), and the probability for any possible value of that random pair (x, y) is 1 36.

Jointly Distributed Random Variables

Jointly Distributed Random Variables Definition Let X and Y be two discrete random variables defined on the sample space S of an experiment. The joint probability mass function p(x, y) is defined for each pair of numbers (x, y) by p(x, y) = P(X = x and Y = y) (It must be the case that p(x, y) 0 and x y p(x, y) = 1.) For any event A consisting of pairs of (x, y), the probability P[(X, Y ) A] is obtained by summing the joint pmf over pairs in A: P[(X, Y ) A] = p(x, y) (x,y) A

Jointly Distributed Random Variables

Jointly Distributed Random Variables Example (Problem 75) A restaurant serves three fixed-price dinners costing $12, $15, and $20. For a randomly selected couple dinning at this restaurant, let X = the cost of the man s dinner and Y = the cost of the woman s dinner. If the joint pmf of X and Y is assumed to be y p(x, y) 12 15 20 12.05.05.10 x 15.05.10.35 20 0.20.10

Jointly Distributed Random Variables Example (Problem 75) A restaurant serves three fixed-price dinners costing $12, $15, and $20. For a randomly selected couple dinning at this restaurant, let X = the cost of the man s dinner and Y = the cost of the woman s dinner. If the joint pmf of X and Y is assumed to be y p(x, y) 12 15 20 12.05.05.10 x 15.05.10.35 20 0.20.10 a. What is the probability for them to both have the $12 dinner?

Jointly Distributed Random Variables Example (Problem 75) A restaurant serves three fixed-price dinners costing $12, $15, and $20. For a randomly selected couple dinning at this restaurant, let X = the cost of the man s dinner and Y = the cost of the woman s dinner. If the joint pmf of X and Y is assumed to be y p(x, y) 12 15 20 12.05.05.10 x 15.05.10.35 20 0.20.10 a. What is the probability for them to both have the $12 dinner? b. What is the probability that they have the same price dinner?

Jointly Distributed Random Variables Example (Problem 75) A restaurant serves three fixed-price dinners costing $12, $15, and $20. For a randomly selected couple dinning at this restaurant, let X = the cost of the man s dinner and Y = the cost of the woman s dinner. If the joint pmf of X and Y is assumed to be y p(x, y) 12 15 20 12.05.05.10 x 15.05.10.35 20 0.20.10 a. What is the probability for them to both have the $12 dinner? b. What is the probability that they have the same price dinner? c. What is the probability that the man s dinner cost $12?

Jointly Distributed Random Variables

Jointly Distributed Random Variables Definition Let X and Y be two discrete random variables defined on the sample space S of an experiment with joint probability mass function p(x, y). Then the pmf s of each one of the variables alone are called the marginal probability mass functions, denoted by p X (x) and p Y (y), respectively. Furthermore, p X (x) = y p(x, y) and p Y (y) = x p(x, y)

Jointly Distributed Random Variables

Jointly Distributed Random Variables Example (Problem 75) continued: The marginal probability mass functions for the previous example is calculated as following: y p(x, y) 12 15 20 p(x, ) 12.05.05.10.20 x 15.05.10.35.50 20 0.20.10.30 p(, y).10.35.55

Jointly Distributed Random Variables

Jointly Distributed Random Variables Definition Let X and Y be continuous random variables. A joint probability density function f (x, y) for these two variables is a function satisfying f (x, y) 0 and f (x, y)dxdy = 1. For any two-dimensional set A P[(X, Y ) A] = f (x, y)dxdy In particular, if A is the two-dimensilnal rectangle {(x, y) : a x b, c y d}, then P[(X, Y ) A] = P(a X b, c Y d) = A b d a c f (x, y)dydx

Jointly Distributed Random Variables

Jointly Distributed Random Variables Definition Let X and Y be continuous random variables with joint pdf f (x, y). Then the marginal probability density functions of X and Y, denoted by f X (x) and f Y (y), respectively, are given by f X (x) = f Y (y) = f (x, y)dy f (x, y)dx for < x < for < y <

Jointly Distributed Random Variables

Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : { xe (x+y) x 0 and y 0 f (x, y) = 0 otherwise

Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : { xe (x+y) x 0 and y 0 f (x, y) = 0 otherwise a. What is the probability that the lifetimes of both components excceed 3?

Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : { xe (x+y) x 0 and y 0 f (x, y) = 0 otherwise a. What is the probability that the lifetimes of both components excceed 3? b. What are the marginal pdf s of X and Y?

Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : { xe (x+y) x 0 and y 0 f (x, y) = 0 otherwise a. What is the probability that the lifetimes of both components excceed 3? b. What are the marginal pdf s of X and Y? c. What is the probability that the lifetime X of the first component excceeds 3?

Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : { xe (x+y) x 0 and y 0 f (x, y) = 0 otherwise a. What is the probability that the lifetimes of both components excceed 3? b. What are the marginal pdf s of X and Y? c. What is the probability that the lifetime X of the first component excceeds 3? d. What is the probability that the lifetime of at least one component excceeds 3?