Back to estimators...

Similar documents
Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Applied Statistics I

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

Statistical estimation

MATH 3200 Exam 3 Dr. Syring

Point Estimation. Copyright Cengage Learning. All rights reserved.

Chapter 7: Estimation Sections

Lecture 10: Point Estimation

3 ˆθ B = X 1 + X 2 + X 3. 7 a) Find the Bias, Variance and MSE of each estimator. Which estimator is the best according

Estimating parameters 5.3 Confidence Intervals 5.4 Sample Variance

Point Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased.

Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel

Chapter 6: Point Estimation

Chapter 8. Introduction to Statistical Inference

Statistical analysis and bootstrapping

Statistics and Probability

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Learning From Data: MLE. Maximum Likelihood Estimators

Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems

Confidence Intervals Introduction

MVE051/MSG Lecture 7

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

STAT Chapter 7: Confidence Intervals

Statistics 431 Spring 2007 P. Shaman. Preliminaries

Chapter 8: Sampling distributions of estimators Sections

Review of key points about estimators

CHAPTER 8. Confidence Interval Estimation Point and Interval Estimates

BIO5312 Biostatistics Lecture 5: Estimations

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example...

19. CONFIDENCE INTERVALS FOR THE MEAN; KNOWN VARIANCE

Chapter 4: Asymptotic Properties of MLE (Part 3)

Chapter 7: Estimation Sections

Interval estimation. September 29, Outline Basic ideas Sampling variation and CLT Interval estimation using X More general problems

Determining Sample Size. Slide 1 ˆ ˆ. p q n E = z α / 2. (solve for n by algebra) n = E 2

PROBABILITY AND STATISTICS

Chapter 7. Sampling Distributions and the Central Limit Theorem

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

Chapter 7. Sampling Distributions and the Central Limit Theorem

Chapter 7 presents the beginning of inferential statistics. The two major activities of inferential statistics are

Probability & Statistics

Chapter 4: Estimation

12 The Bootstrap and why it works

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

CSE 312 Winter Learning From Data: Maximum Likelihood Estimators (MLE)

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

Introduction to Probability and Inference HSSP Summer 2017, Instructor: Alexandra Ding July 19, 2017

Exercise. Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1. Exercise Estimation

Chapter 7 - Lecture 1 General concepts and criteria

Homework Problems Stat 479

Chapter 7: Point Estimation and Sampling Distributions

CS 361: Probability & Statistics

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.

Likelihood Methods of Inference. Toss coin 6 times and get Heads twice.

Statistical Intervals (One sample) (Chs )

8.1 Estimation of the Mean and Proportion

Data Analysis and Statistical Methods Statistics 651

Statistics for Business and Economics

Probability Theory. Probability and Statistics for Data Science CSE594 - Spring 2016

CS340 Machine learning Bayesian model selection

Point Estimation. Edwin Leuven

AMS 7 Sampling Distributions, Central limit theorem, Confidence Intervals Lecture 4

χ 2 distributions and confidence intervals for population variance

4-2 Probability Distributions and Probability Density Functions. Figure 4-2 Probability determined from the area under f(x).

6. Genetics examples: Hardy-Weinberg Equilibrium

STAT 111 Recitation 2

5.3 Interval Estimation

1 Inferential Statistic

Section 0: Introduction and Review of Basic Concepts

Chapter 5. Statistical inference for Parametric Models

Chapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29

Continuous Probability Distributions & Normal Distribution


A New Hybrid Estimation Method for the Generalized Pareto Distribution

Statistics 13 Elementary Statistics

Module 4: Point Estimation Statistics (OA3102)

Chapter 7: Estimation Sections

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

Review of key points about estimators

Week 7. Texas A& M University. Department of Mathematics Texas A& M University, College Station Section 3.2, 3.3 and 3.4

Chapter 8 Statistical Intervals for a Single Sample

Practice Exam 1. Loss Amount Number of Losses

Sampling & Confidence Intervals

STAT 509: Statistics for Engineers Dr. Dewei Wang. Copyright 2014 John Wiley & Sons, Inc. All rights reserved.

MATH 264 Problem Homework I

A useful modeling tricks.

. (i) What is the probability that X is at most 8.75? =.875

Chapter 4 Continuous Random Variables and Probability Distributions

σ 2 : ESTIMATES, CONFIDENCE INTERVALS, AND TESTS Business Statistics

Posterior Inference. , where should we start? Consider the following computational procedure: 1. draw samples. 2. convert. 3. compute properties

Statistics 6 th Edition

CH 5 Normal Probability Distributions Properties of the Normal Distribution

1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by. Cov(X, Y ) = E(X E(X))(Y E(Y ))

1. Statistical problems - a) Distribution is known. b) Distribution is unknown.

Lecture 6: Chapter 6

Data Analysis and Statistical Methods Statistics 651

The Bernoulli distribution

Transcription:

Back to estimators... So far, we have: Identified estimators for common parameters Discussed the sampling distributions of estimators Introduced ways to judge the goodness of an estimator (bias, MSE, etc.) Used estimators in confidence intervals, hypothesis testing, etc. We haven t spent a lot of time on developing/thinking of good estimators for a parameter. What can we do to estimate a parameter that we ve never considered before?

Binomial example Let s begin with a familiar case. We have a RV that we know has a binomial distribution. Familiar example: We have 10 people receiving a new treatment for a serious disease? What is the probability of recovery (success) p? Another familiar example: What proportion of those casting ballots voted for Gore? Try to estimate this based on a random sample of 1000 voters. The problem is that we don t know what p is, now that we have implemented a new treatment regimen.

Binomial probability histograms 0.0 0.15 0.30 0.0 0.15 0.30 Binomial prob. hist. with n=10 and p=0.25 Binomial prob. hist. with n=10 and p=0.4 0.0 0.15 0.30 0.0 0.15 0.30 Binomial prob. hist. with n=10 and p=0.55 Binomial prob. hist. with n=10 and p=0.7

Estimating p Pretend for a few moments that we haven t previously discussed using ˆp as an estimator for p. Our intuition would tell us that the best guess for p was the proportion of successes in our sample (ˆp). There is also a more mathematical way to approach the problem. 1. Think about the probability distribution for Y, the number of successes in your sample of size n. 2. Although the parameter p is unknown, we can substitute into the probability distribution the information we do know: Y and n. 3. Determine what value of p would give the highest probability of obtaining that sample with y successes in n trials.

Maximum likelihood methodology Assume that 6 of the 10 patients receiving the new treatment regimen recover. Write down the probability distribution for the number of successes in n trials. (This is the likelihood function.) Substitute in the information from your sample.

Likelihood function 0.0 0.05 0.10 0.15 0.20 0.25 Likelihood function for p 0.0 0.2 0.4 0.6 0.8 1.0 p

Maximum likelihood methodology Find the value of p that maximizes L(p). This is your maximum likehood estimate for p.

Maximum likelihood methodology Let s work through the more general case and find a formula for the maximum likelihood estimate for p - a formula which depends just on observed data. As before, write down likelihood function. How to take the derivative, set it equal to zero, and solve for p? Looks messy...

Maximum likelihood estimate of p Find the value of p that maximizes L(p). This is your maximum likehood estimate for p.

Invariance property of MLEs Say we have the MLE for parameter 1, but we want to know the MLE for parameter 2, which is a one-to-one function of parameter 1. To find the MLE for parameter 2, substitute the MLE for parameter 1 into the function that gives parameter 2. If t(θ) is a function of θ and ˆθ is the MLE for θ, then the MLE for t(θ) is given by t(ˆθ).

Using the invariance property We know that for a binomial proportion of successes p, the MLE is given by ˆp. What is the MLE for the variance of Y?

More about MLEs Can be used to help you find an estimator for a parameter that is new to you Invariance property of MLEs means that for many functions, you can build a MLE for the function using MLEs that are already known. MLEs are not always unbiased. However, in some cases adjusting the MLE by a constant can yield an unbiased estimator with minimum variance. The method of maximum likelihood is a commonly used procedure in statistics.

MLE for mean of the normal Assume that we have a random sample x 1,x 2,..., x n from a normally distributed population for which the variance σ 2 is known. What is the MLE for the mean µ? The probability density function for the normal distribution is: f(x) = 1 σ exp[ 1 2π 2σ (x µ) 2 ], <x<, <µ< 2

MLE for variance of the normal Assume that we have a random sample x 1,x 2,..., x n from a normally distributed population for which the mean µ is known. What is the MLE for the variance σ 2? The probability density function for the normal distribution is: f(x) = 1 σ exp[ 1 2π 2σ (x µ) 2 ], <x<, σ 2 > 0 2

Large sample properties of MLEs Already know from the invariance property that if ˆθ is the MLE for θ, then for a function of θ, t(θ), the MLE t(θ) ist(ˆθ) Under certain regularity conditions (that apply for the distributions that we will consider), t(θ) is a consistent estimator for t(θ) This means that as the sample size n grows, t(θ) tends to get closer to t(θ) For large sample sizes, we also know that t(ˆθ) is normally distributed, so that the following has an approximately standard normal distribution Z = t(ˆθ) t(θ) [ dt(θ) dθ ]2 ne[ d2 ln f(y θ) dθ 2 ]

Confidence intervals using MLEs For large sample sizes, we can obtain a confidence interval for t(θ) usingthe MLE estimator and its large sample properties: t(ˆθ) ± z α [ dt(θ) dθ ]2 2 ne[ d2 ln f(y θ) dθ ] 2 However, this formula can still depend on θ, so in practice, we can substitute the MLE ˆθ for θ: t(ˆθ) ± z α [ dt(θ) dθ ]2 2 ne[ d2 ln f(y θ) dθ ] 2 θ=ˆθ

CI for normal mean As before, assume that we have a random sample x 1,x 2,..., x n from a normally distributed population for which the variance σ 2 is known. Show that a 100(1 α)% confidence interval for µ, when σ is known, is given by x ± z α σ 2 n, as we learned before.

Example Suppose that X 1,X 2,...,X n form a random sample from a distribution for which the p.d.f. f(x θ) is given below. Also, suppose θ is unknown (θ >0). Find the MLE of θ. θx θ 1 for 0 <x<1 f(x θ) = 0 otherwise

Example continued Find a 100(1 α)% confidence interval for the θ 2 (reference the distribution given in the last slide). t(ˆθ) ± z α 2 [ dt(θ) dθ ]2 ne[ d2 ln f(y θ) dθ ] 2 θ=ˆθ