Lecture 10: Point Estimation

Similar documents
Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

Chapter 6: Point Estimation

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel

6. Genetics examples: Hardy-Weinberg Equilibrium

Chapter 4: Asymptotic Properties of MLE (Part 3)

Point Estimation. Copyright Cengage Learning. All rights reserved.

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Likelihood Methods of Inference. Toss coin 6 times and get Heads twice.

Chapter 7 - Lecture 1 General concepts and criteria

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Back to estimators...

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Chapter 8. Introduction to Statistical Inference

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

MATH 3200 Exam 3 Dr. Syring

Applied Statistics I

Chapter 8: Sampling distributions of estimators Sections

Exercise. Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1. Exercise Estimation

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems

The Bernoulli distribution

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

Statistical estimation

CSE 312 Winter Learning From Data: Maximum Likelihood Estimators (MLE)

BIO5312 Biostatistics Lecture 5: Estimations

Chapter 5. Statistical inference for Parametric Models

CS340 Machine learning Bayesian model selection

Point Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased.

Chapter 8: Sampling distributions of estimators Sections

Random Samples. Mathematics 47: Lecture 6. Dan Sloughter. Furman University. March 13, 2006

Chapter 7: Point Estimation and Sampling Distributions

Computer Statistics with R

Chapter 7: Estimation Sections

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

Learning From Data: MLE. Maximum Likelihood Estimators

Chapter 5: Statistical Inference (in General)

MVE051/MSG Lecture 7

Random Variables Handout. Xavier Vilà

EE641 Digital Image Processing II: Purdue University VISE - October 29,

A New Hybrid Estimation Method for the Generalized Pareto Distribution

Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

PROBABILITY AND STATISTICS

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

What was in the last lecture?

Practice Exam 1. Loss Amount Number of Losses

12. Conditional heteroscedastic models (ARCH) MA6622, Ernesto Mordecki, CityU, HK, 2006.

Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making

Random Variables and Probability Distributions

Chapter 7: Estimation Sections

Exam STAM Practice Exam #1

The Normal Distribution

Commonly Used Distributions

Bivariate Birnbaum-Saunders Distribution

LET us say we have a population drawn from some unknown probability distribution f(x) with some

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

3 ˆθ B = X 1 + X 2 + X 3. 7 a) Find the Bias, Variance and MSE of each estimator. Which estimator is the best according

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.

STRESS-STRENGTH RELIABILITY ESTIMATION

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION

Lecture III. 1. common parametric models 2. model fitting 2a. moment matching 2b. maximum likelihood 3. hypothesis testing 3a. p-values 3b.

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 10: Continuous RV Families. Prof. Vince Calhoun

Lecture 2. Probability Distributions Theophanis Tsandilas

Much of what appears here comes from ideas presented in the book:

A potentially useful approach to model nonlinearities in time series is to assume different behavior (structural break) in different subsamples

1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by. Cov(X, Y ) = E(X E(X))(Y E(Y ))

(Practice Version) Midterm Exam 1

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Random variables. Contents

Parameter Estimation for the Lognormal Distribution

Homework Problems Stat 479

Homework Assignments

Probability & Statistics

Reliability and Risk Analysis. Survival and Reliability Function

6. Continous Distributions

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics

Confidence Intervals Introduction

12 The Bootstrap and why it works

STAT/MATH 395 PROBABILITY II

Binomial Random Variables. Binomial Random Variables

Simulation Wrap-up, Statistics COS 323

χ 2 distributions and confidence intervals for population variance

Stochastic Models. Statistics. Walt Pohl. February 28, Department of Business Administration

The Vasicek Distribution

Hardy Weinberg Model- 6 Genotypes

may be of interest. That is, the average difference between the estimator and the truth. Estimators with Bias(ˆθ) = 0 are called unbiased.

Elementary Statistics Lecture 5

Statistical analysis and bootstrapping

Chapter 3 Discrete Random Variables and Probability Distributions

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman:

Chapter 3 - Lecture 4 Moments and Moment Generating Funct

Non-informative Priors Multiparameter Models

Objective Bayesian Analysis for Heteroscedastic Regression

Transcription:

Lecture 10: Point Estimation MSU-STT-351-Sum-17B (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 1 / 31

Basic Concepts of Point Estimation A point estimate of a parameter θ, denoted by ˆθ, is a single number that can be considered as a possible value for θ. Since it is computed from the sample X = (X 1,..., X n ), it is a function of X, that is, ˆθ = ˆθ(X). Some simple examples are: (i) If X 1,..., X n is from B(1, p) (Bernoulli data), then ˆp = 1 n n X i, the sample proportion of success. (ii) If X 1,..., X n is a random sample from a continuous population F(x) with mean µ and variance σ 2, then the commonly used estimators of µ and σ 2 are ˆµ = X; ˆσ 2 = 1 n (X i X) 2 = S 2. n 1 Some other estimators of µ are the sample median, the trimmed mean, etc. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 2 / 31

Next, we discuss some properties of the estimators. (i) The Unbiased Estimators Definition: An estimator ˆθ = ˆθ(X) for the parameter θ is said to be unbiased if E(ˆθ(X)) = θ for all θ. Result: Let X 1,..., X n be a random sample on X F(x) with mean µ and variance σ 2. Then the sample mean X and the sample varance S 2 are unbiased estimators of µ and σ 2, respectively. Proof: (i) Note that ( 1 E(X n ) = E n n ) X i = 1 n n E(X i ) = 1 (nµ) = µ. n (ii) Note S 2 = 1 n 1 n (X i X) 2. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 3 / 31

Then ( n ) E((n 1)S 2 ) = E (X i X) 2 = E ( n X 2 i n(x) 2 ) = ne(x 2 1 ) ne(x 2 ) ) = n(µ 2 + σ 2 ) n (µ 2 + σ2 n = (n 1)σ 2, using E(X 2 1 ) = Var(X 1) + (E(X 1 )) 2 and E(X 2 ) = Var(X) + (E(X)) 2. Thus, E(S 2 ) = σ 2. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 4 / 31

Example 1 (Ex. 4): Let X and Y denote the strength of concrete beams and cylinders. The following data are obtained X : 5.9, 7.2, 7.3, 6.3, 8.1, 6.8, 7.0, 7.6, 6.8, 6.5, 7.0, 6.3, 7.9, 9.0, 8.2, 8.7, 7.8, 9.7, 7.4, 7.7, 9.7, 7.8, 7.7, 11.6, 11.3, 11.8, 10.7. Y : 6.1, 5.8, 7.8, 7.1, 7.2, 9.2, 6.6, 8.3, 7.0, 8.3, 7.8, 8.1, 7.4, 8.5, 8.9, 9.8, 9.7, 14.1, 12.6, 11.2. Suppose E(X) = µ 1, V(X) = σ 2 1 ; E(Y) = µ 2, V(Y) = σ 2 2. (a) Show that X Y is an unbiased estimator of µ 1 µ 2. Calculate it for the given data. (b) Find the variance and standard deviation (standard error) of the estimator in Part(a), and then compute the estimated standard error. (c) Calculate an estimate of the ratio σ 1 /σ 2 of the two standard deviations. (d) Suppose a single beam X and a single cylinder Y are randomly selected. Calculate an estimate of the variance of the difference X Y. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 5 / 31

Solution: (a) E(X Y) = E(X) E(Y) = µ 1 µ 2. Hence, the unbiased estimate based on the given data is x y = 8.141 8.575 = 0.434 (b) V(X Y) = V(X) + V(Y) = σ 2 + X σ2 = σ2 1 Y n 1 + σ2 2 n 2. Thus, σ X Y = V(X Y) = An estimate would be S 2 1 S X Y = + S2 2 (1.666) 2 = n 1 n 2 27 σ 2 1 n 1 + σ2 2 n 2. + (2.104)2 20 = 0.5687. Note S 1 is not an unbised estimator of σ 1. Similarly, S 1 /S 2 is not an unbiased estimator of σ 1 /σ 2. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 6 / 31

(c) An estimate of σ 1 /σ 2 is (this is a biased estimate) (d) Note that S 1 S 2 = 1.660 2.104 = 0.7890. V(X Y) = V(X) + V(Y) = σ 2 1 + σ2 2. Hence, ˆσ 1 2 + ˆσ 2 2 = (1.66) 2 + (2.104) 2 = 7.1824 Example 2 (Ex 8): In a random sample of 80 components of a certain type, 12 are found to be defective. (a) Give a point estimate of the proportion of all not-defective units. (b) A system is to be constructed by randomly selecting two of these components and connecting them in series. Estimate the proportion of all such systems that work properly. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 7 / 31

Solution: (a) With p denoting the true proportion of non-defective components, ˆp = 80 12 80 = 0.85 (b) P(system works)=p 2, since the system works if and only if both components work. So, an estimate of this probability is ˆp = ( 68 ) 2 =.723 80 Variances of estimators The unbiased estimators are not in general unique. Given two unbiased estimators, it is natural to choose the one with less variance. In some cases, depending on the form of F(x θ), we can find the unbiased estimator with minimum variance, called the MVUE. For instance, in the N(µ, 1) case, the MVUE of µ is X. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 8 / 31

Example 3 (Ex 10): Using a rod of length µ, you lay out a square plot whose length of each side is µ. Thus, the area of the plot will be µ 2 (unknown). Based on n independent measurements X 1,..., X n of the length, estimate µ 2. Assume that each X i has mean µ and variance σ 2. (a) Show that X 2 is not an unbiased estimator for µ 2. (b) For what value of k is the estimator X 2 ks 2 unbiased for µ 2? Solution: (a) Note E(X 2 ) = Var(X) + [E(X)] 2 = σ2 n + µ2. So, the bias of the estimator X 2 is E(X 2 µ 2 ) = σ2 n. Also, X 2 tends to overestimate µ 2. (b) Also, E(X 2 ks 2 ) = E(X 2 ) ke(s 2 ) = µ 2 + σ2 n kσ2. Hence, with k = 1/n, E(X 2 ks 2 ) = µ 2. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 9 / 31

The Standard Error of an Estimator It is useful to report the standard error of the estimator, in addition to its value. Unfortunately, it depends on the unknown parameters, and hence its estimate is usually used. For a binomial model the estimator ˆp = S n /n of p, has the standard p(1 p) deviation n which depends on p (unknown). To estimate µ based on a random sample from a normal distribution, we use the estimator X, whose standard deviation σ n which depends on another unknown parameter σ. Using estimates of p and σ, we obtain s.e.(ˆp) = ˆp(1 ˆp) ; s.e.(x) = s. n n (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 10 / 31

Example 4 (Ex 12): Suppose fertilizer-1 has a mean yield per acre of µ 1 with variance σ 2, whereas the expected yield for fertilizer-2 is µ 2 with the same variance σ 2. Let S 2 denote the sample variances of yields based on i sample sizes n 1 and n 2, respectively, of the two fertilizers. Show that the pooled (combined) estimator ˆσ 2 = (n 1 1)S 2 1 + (n 2 1)S 2 2 n 1 + n 2 2 is an unbiased estimator of σ 2. Solution: [ (n1 1)S 2 1 E 2 1)S 2 ] 2 n 1 + n 2 2 = (n 1 1) n 1 + n 2 2 E(S2 1 ) + (n 2 1) n 1 + n 2 2 E(S2 2 ) = (n 1 1) n 1 + n 2 2 σ2 + (n 2 1) n 1 + n 2 2 σ2 = σ 2. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 11 / 31

Method of Estimation It is desirable to have some general methods of estimation which yield estimators with some good properties. One of the classical methods is the method of moments (MoM), though it is not frequently used these days. The maximum likelihood (ML) method is one of the popular methods and the resulting maximum likelihood estimators (MLEs) have several finite and large sample properties. The method of moments Early in the development of statistics, the moments of a distribution (mean, variance, skewness, kurtosis) were discussed in depth, and estimators were formulated by equating the sample moments (i.e., x, s 2,...) to the corresponding population moments, which are functions of the parameters. The number of equations should be equal to the number of parameters. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 12 / 31

Example 1: Consider the exponential distribution, E(λ), with density { λe f(x; λ) = λx, x 0 0, otherwise. Then E[X] = 1/λ, and so solving X = 1 λ, we obtain MoM as ˆλ = (1/X). Drawbacks of MoM estimators (i) A drawback of the MoM estimators is that it is difficult to solve the associated equations. Consider the parameters α and β in a Weibull distribution (see pp. 181-183). In this case, we need to solve ( µ = βγ 1 + 1 [ ( ), σ 2 = β 2 Γ 1 + 2 ) [ ( Γ 1 + 1 )] 2 ], α α α which is not an easy one. (ii) Since MoM estimators use only a few population moments and their sample counterparts, the resulting estimators may some times unreasonable, as in the following example. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 13 / 31

Example 5: Suppose X 1,..., X n is a random sample from uniform U(0, θ) distribution. Then solving E(X) = θ 2 = X, we get MoM estimator as ˆθ = 2X. It is possible ˆθ > max(x i ), while each X i < θ. Example 6 (Ex 22): Let X denote the proportion of allotted time that a randomly selected student spends working on a certain aptitude test. Suppose the pdf of X is { (θ + 1)x f(x; θ) = θ, 0 x 1 0, otherwise where 1 < θ. A random sample of ten students yields data x 1 = 0.92, x 2 = 0.79, x 3 = 0.90, x 4 = 0.65, x 5 = 0.86, x 6 = 0.47, x 7 = 0.73, x 8 = 0.97, x 9 = 0.94, x 10 = 0.77. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 14 / 31

(a) Obtain MoM estimator and find it from the above data. (b) Obtain MLE of θ, and compute it for the given data. Solution: (a) E(X) = 1 0 x(θ + 1)x θ dx = θ + 1 θ + 2 = 1 1 θ + 2. So, the moment estimator ˆθ is the solution to X = 1 1 ˆθ + 2, yielding ˆθ = 1 1 X 2. For the given data, x = 0.80, ˆθ = 5 2 = 3. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 15 / 31

Maximum Likelihood Estimators The ML method, introduced by R.A. Fisher, is based on the likelihood function of unknown parameter. Definition: Let X 1,..., X n be a random sample from f(x θ). Then the joint density n f(x 1,..., x n θ) = f(x i θ) = L(θ x) (veiwed as a function of θ) is called the likelihood function of θ, for an observed X = x. An estimate ˆθ(x) that maximizes the L(θ x) is called a maximum likelihood estimate of θ. Also, the estimator ˆθ(X) = ˆθ(X 1,..., X n ) is called the maximum likelihood estimator (MLE) of θ. Here, θ may be a vector. This method yields estimators that have many desirable properties; both finite as well as large sample properties. The basic idea to find an estimator ˆθ(x) which is the most likely given the data X = (X 1,..., X n ). (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 16 / 31

Example 7: Consider the density, discussed in Example 6, { (θ + 1)x f(x; θ) = θ, 0 x 1 0, otherwise. Obtain the MLE of θ and compute it for the data given there. Solution: Note the likelihood function is So, the log-likelihood is Taking d n dθ and equating to 0 yields f(x 1,..., x n ; θ) = (θ + 1) n (x 1 x 2... x n ) θ. n ln(θ + 1) + θ ln(x i ). θ + 1 = ln(x i ). Solve for θ to get n ˆθ = ln(xi ) 1. Taking ln(x i ) for each given x i yields ultimately ˆθ = 3.12. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 17 / 31

Example 8: Let X B(1, p), Bernoulli distribution, with pmf P(X = x p) = p(x p) = p x (1 p) 1 x, x = 0, 1, where p = P(X = 1). Find the MLE of p, based on X 1,..., X n. Solution: Aim is to estimate the population proportion p based on a random sample X = (X 1,..., X n ) of size n. Note X 1,..., X n are independent and identically distributed random variables. For x i {0, 1}, we have the joint pmf of X 1,..., X n is (using independence) ) ) ) P (X 1 = x 1,..., X n = x n = P (X 1 = x 1... P (X n = x n since X i s have identical pmf. = p x 1 (1 p) 1 x 1... p x n (1 p) 1 x n = p n 1 x i (1 p) n n 1 x i, (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 18 / 31

Write the above density as a function of p, the likelihood function is L(p x) = p n 1 x i (1 p) n n 1 x i = p s n (1 p) n s n, where s n = n 1 x i. Choose an estimator that maximizes L(p x). Take Now l = ln L = s n ln p + (n s n ) ln(1 p). ln L p = 0 s n p n s n 1 p = 0 ˆp = s n n = p, the sample mean (proportion). Also, it can be shown that 2 l p 2 ˆp < 0. Hence, ˆp = S n /n = n X i, the sample proportion, is the MLE of p. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 19 / 31

Example 9: Let X 1,..., X n be a random sample from N(µ, σ 2 ), where both mean µ and σ 2 are unknown. Find the MLE s of µ and σ 2. Solution: Let θ = (µ, σ 2 ). Then ( f(x i θ) = 1 1 xi µ ) 2 ( x µ ) 2 e 2 σ = (2πσ 2 ) 1/2 e 1 2 σ. 2πσ Hence, the joint density is f(x 1,..., x n θ) = f(x 1 θ)f(x 2 θ)... f(x n θ) 1 n ( xi µ = (2πσ 2 ) n 2 σ 2 e i = L(µ, σ 2 x). ) 2 (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 20 / 31

Then l = ln L(µ, σ 2 x) = n 2 ln(2πσ2 ) 1 2σ 2 Then, for all σ 2 > 0, n (x i µ) 2 = n 2 ln(2π) n 2 ln(σ2 ) 1 2σ 2 ln L µ Substituting ˆµ = x in l(µ, σ 2 ), we get = 0 ˆµ = x. l(µ, σ 2 ) = n 2 ln(2π) n 2 ln(σ2 ) 1 2σ 2 1 n (x i µ) 2. n (x i x) 2. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 21 / 31

Then Hence, l(µ, σ 2 ) σ 2 l(µ, σ 2 ) σ 2 = n 2σ 2 + 1 2(σ 2 ) 2 = 0 ˆσ 2 = 1 n n (x i x) 2. n (x i x) 2. Also, the Hessian matrix of second order partial derivatives of l(x, σ 2 ), calculated at ˆµ = x and ˆσ 2, can be shown to be nonnegative definite. Therefore, ˆµ and ˆσ 2 are the MLEs of µ and σ 2. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 22 / 31

Example 10: Let X 1,..., X n be a random sample from exponential density f(x λ) = λ e λx, x > 0, λ > 0. Find the MLE of λ. Solution: The joint density of X 1,..., X n (likelihood function) is n n f(x λ) = f(x i λ) = λ e λx i = λ n e λ n 1 x i. Hence, L(λ x) = λ n e nλx l = log(l) = n ln(λ) nλx; l λ = 0 ˆλ = 1 x. Thus, the MLE of λ is ˆλ = 1 X. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 23 / 31

Example 11 (Ex 29): Suppose n time head-ways X 1,..., X n in a traffic flow follow a shifted-exponential with pdf { λe f(x λ, θ) = λ(x θ), x θ; 0, otherwise. (a) Obtain the MLE s of θ and λ. (b) If n = 10 time headway observations are 3.11,.64, 2.55, 2.20, 5.44, 3.42, 10.39, 8.93, 17.82, 1.30 calculate the estimates of θ and λ. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 24 / 31

Solution: (a) The joint pdf of X 1,..., X n is f(x 1,..., x n λ, θ) = = n f(x i λ, θ) { λ n e λ n (xi θ), x 1 θ,..., x n θ 0, otherwise. Notice that x 1 θ,..., x n θ iff min(x i ) θ, and also n n λ (x i θ) = λ x i + nλθ. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 25 / 31

Hence, the likelihood function is { λ L(λ, θ x) = n e (nλθ λ n xi), min(x i ) θ 0, otherwise. Consider first the maximization with respect to θ. Because the exponent nλθ is positive, increasing θ will increase the likelihood provided that min(x i ) θ; if we make θ larger than min(x i ), the likelihood drops to 0. This implies that the mle of θ is ˆθ = min(x i ) = x (1). Now, substituting ˆθ in likelihood function L(λ, ˆθ x) = λ n e n n ) (x i x (1). (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 26 / 31

This implies n ) l(λ, ˆθ x) = ln(l(λ, ˆθ x)) = n ln(λ) n (x i x (1) Solving for λ, l λ = n n λ (x i x (1) ) = 0. ˆλ = n (xi x (1) ). (b) From the data, ˆθ = min(x i ) =.64 and n x i = 55.80. hence, ˆλ = 10 55.80 6.4 = 0.202. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 27 / 31

Properties of MLE s (i)for large n, the MLE ˆθ(X) is asymptotically normal, unbiased, and has variance smaller than any other estimator. (ii) Invariance property: If ˆθ is an MLE of θ, then g(ˆθ) is an MLE of g(θ) for any function g. Example 12: Let X 1,..., X n be a random sample from exponential distribution E(λ) with parameter λ. Find the MLE of the mean of the distribution. Solution: As seen in Example 10, the MLE of λ is ˆλ = 1 X. Then the MLE of g(λ) = 1 λ = E(X i) is ĝ(λ) = 1ˆλ = x, using the invariance property of the MLE. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 28 / 31

Example 11 (Ex 26): The following data represents shear strength (X) of the test spot weld 392, 376, 401, 367, 389, 362, 409, 415, 358, 375. (a) Assuming that X is normally distributed, estimate the true average shear strength and standard deviation of shear strength using the method of maximum likelihood. (b) Obtain the MLE of P(X 400). Solution: (a) The MLE s of µ and σ 2 are ˆµ = X; ˆσ 2 = 1 n Hence, the MLE of σ is ˆσ = n n 1 n S2. (X i X) 2 = n 1 n S2. (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 29 / 31

From the data given: ˆµ = x = 384.4; S 2 = 395.16. So, 1 (x i x) 2 = ˆσ 2 = 9 n 10 (395.16) = 355.64 and ˆσ = 355.64 = 18.86. (b) Let θ = P(X 400). Then σ 400 µ ) σ Z 400 µ σ ( 400 µ ) = Φ. σ ( X µ θ = P ( = P The MLE of θ, by invariance property, is ) (Z N(0, 1)) ( 400 ˆµ ) ( 400 384.4 ) ˆθ = Φ = Φ = 0.7881. ˆσ 18.86 (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 30 / 31

Home work: Sect 6.1: 3, 11, 13, 15, 16 Sect 6.2: 20, 23, 28, 30, 32 (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 31 / 31