Review of key points about estimators

Similar documents
Review of key points about estimators

Point Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased.

Chapter 7 - Lecture 1 General concepts and criteria

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

Applied Statistics I

STAT 509: Statistics for Engineers Dr. Dewei Wang. Copyright 2014 John Wiley & Sons, Inc. All rights reserved.

Interval estimation. September 29, Outline Basic ideas Sampling variation and CLT Interval estimation using X More general problems

8.1 Estimation of the Mean and Proportion

Point Estimation. Edwin Leuven

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example...

Statistical analysis and bootstrapping

MATH 3200 Exam 3 Dr. Syring

Chapter 7 presents the beginning of inferential statistics. The two major activities of inferential statistics are

Statistical estimation

Chapter 8. Introduction to Statistical Inference

MVE051/MSG Lecture 7

STAT Chapter 6: Sampling Distributions

may be of interest. That is, the average difference between the estimator and the truth. Estimators with Bias(ˆθ) = 0 are called unbiased.

Contents. 1 Introduction. Math 321 Chapter 5 Confidence Intervals. 1 Introduction 1

Chapter 7: Point Estimation and Sampling Distributions

Chapter 8: Sampling distributions of estimators Sections

Statistics for Business and Economics

1 Introduction 1. 3 Confidence interval for proportion p 6

Confidence Intervals Introduction

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

Econ 300: Quantitative Methods in Economics. 11th Class 10/19/09

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Back to estimators...

Section The Sampling Distribution of a Sample Mean

Maximum Likelihood Estimation

1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by. Cov(X, Y ) = E(X E(X))(Y E(Y ))

Module 4: Point Estimation Statistics (OA3102)

The Constant Expected Return Model

Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel

Optimal Search for Parameters in Monte Carlo Simulation for Derivative Pricing

MATH 10 INTRODUCTORY STATISTICS

BIO5312 Biostatistics Lecture 5: Estimations

Chapter 8: Sampling distributions of estimators Sections

Measure of Variation

Sampling Distributions

Midterm Exam III Review

1. Statistical problems - a) Distribution is known. b) Distribution is unknown.

Chapter 5: Statistical Inference (in General)

σ 2 : ESTIMATES, CONFIDENCE INTERVALS, AND TESTS Business Statistics

Chapter 6: Point Estimation

Statistics 431 Spring 2007 P. Shaman. Preliminaries

Statistics for Managers Using Microsoft Excel 7 th Edition

Sampling and sampling distribution

ECO220Y Estimation: Confidence Interval Estimator for Sample Proportions Readings: Chapter 11 (skip 11.5)

MTH6154 Financial Mathematics I Stochastic Interest Rates

Chapter 5. Statistical inference for Parametric Models

Statistics vs. statistics

Shifting our focus. We were studying statistics (data, displays, sampling...) The next few lectures focus on probability (randomness) Why?

Chapter 4: Estimation

LET us say we have a population drawn from some unknown probability distribution f(x) with some

Estimating parameters 5.3 Confidence Intervals 5.4 Sample Variance

Class 16. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems

Unit 5: Sampling Distributions of Statistics

MLLunsford 1. Activity: Central Limit Theorem Theory and Computations

Unit 5: Sampling Distributions of Statistics

Chapter 7. Sampling Distributions and the Central Limit Theorem

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Section 0: Introduction and Review of Basic Concepts

Agricultural and Applied Economics 637 Applied Econometrics II

Lecture 6: Chapter 6

Key Objectives. Module 2: The Logic of Statistical Inference. Z-scores. SGSB Workshop: Using Statistical Data to Make Decisions

STAT Chapter 7: Confidence Intervals

VARIABILITY: Range Variance Standard Deviation

12 The Bootstrap and why it works

Statistics and Probability

Much of what appears here comes from ideas presented in the book:

UNIVERSITY OF VICTORIA Midterm June 2014 Solutions

Chapter 4 Variability

The Assumption(s) of Normality

CPSC 540: Machine Learning

19. CONFIDENCE INTERVALS FOR THE MEAN; KNOWN VARIANCE

ST440/550: Applied Bayesian Analysis. (5) Multi-parameter models - Summarizing the posterior

Probability & Statistics

3.3-Measures of Variation

Chapter 14 : Statistical Inference 1. Note : Here the 4-th and 5-th editions of the text have different chapters, but the material is the same.

Hypothesis Tests: One Sample Mean Cal State Northridge Ψ320 Andrew Ainsworth PhD

STA Module 3B Discrete Random Variables

IEOR E4703: Monte-Carlo Simulation

CPSC 540: Machine Learning

χ 2 distributions and confidence intervals for population variance

Linear Regression with One Regressor

Tutorial 11: Limit Theorems. Baoxiang Wang & Yihan Zhang bxwang, April 10, 2017

MA131 Lecture 9.1. = µ = 25 and σ X P ( 90 < X < 100 ) = = /// σ X

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Chapter 7. Sampling Distributions and the Central Limit Theorem

Math 140 Introductory Statistics

Chapter 5. Sampling Distributions

Statistical Intervals (One sample) (Chs )

The following content is provided under a Creative Commons license. Your support

IEOR E4703: Monte-Carlo Simulation

Sampling Distributions Chapter 18

Alexander Marianski August IFRS 9: Probably Weighted and Biased?

Stat 139 Homework 2 Solutions, Fall 2016

The Binomial Distribution

Transcription:

Review of key points about estimators Populations can be at least partially described by population parameters Population parameters include: mean, proportion, variance, etc. Because populations are often very large (maybe infinite, like the output of a process) or otherwise hard to investigate, we often have no way to know the exact values of the paramters Statistics or point estimators are used to estimate population pararmeters An estimator is calculated using a function that depends on information taken from a sample from the population We are interested in evaluating the goodness of our estimator - topic of sections 8.1-8.4 To evaluate goodness, it s important to understand facts about the estimator s sampling distribution, its mean, its variance, etc.

Different estimators are possible for same parameter In everyday life, people who are working with the same information arrive at different ideas/decisions based on the same information Given the same sample measurements/data, people may derive different estimators for the population parameter (mean, variance, etc.) For this reason, we need to evaluate the estimators on some criteria (bias, etc.) to determine which is best Complication: the criteria that are used to judge estimators may differ Example: For estimating σ 2 (variance), which is better: s 2 = 1 n 1 n i=1 (x i x) 2 (sample variance) or some other estimator s 2 = 1 n n i=1 (x i x) 2 (which more closely resembles population variance)

Repeated estimation yields sampling distribution If you use an estimator once, and it works well, is that enough proof for you that you should always use that estimator for that parameter? Visualize calculating an estimator over and over with different samples from the same population, i.e. take a sample, calculate an estimate using that rule, then repeat This process yields sampling distribution for the estimator We look at the mean of this sampling distribution to see what value our estimates are centered around We look at the spread of this sampling distribution to see how much our estimates vary

Bias We may want to make sure that the estimates are centered around the paramter of interest (the population parameter that we re trying to estimate) One measurement of center is the mean, so may want to see how far the mean of the estimates is from the parameter of interest bias Assume we re using the estimator ˆθ to estimate the population parameter θ Bias(ˆθ) =E(ˆθ) θ If bias equals 0, the estimator is unbiased Two common unbiased estimators are: 1. Sampling proportion ˆp for population proportion p 2. Sample mean X for population mean µ

Bias and the sample variance What is the bias of the sample variance, s 2 = 1 n 1 n i=1 (x i x) 2? Contrast this case with that of the estimator s 2 = 1 n n i=1 (x i x) 2, which looks more like the formula for population variance.

Variance of an estimator Say your considering two possible estimators for the same population parameter, and both are unbiased Variance is another factor that might help you choose between them. It s desirable to have the most precision possible when estimating a parameter, so you would prefer the estimator with smaller variance (given that both are unbiased). For two of the estimators that we have discussed so far, we have the variances: 1. Var(ˆp) = p(1 p) n 2. Var( X) = σ2 n

Mean square error of an estimator If one or more of the estimators are biased, it may be harder to choose between them. For example, one estimator may have a very small bias and a small variance, while another is unbiased but has a very large variance. In this case, you may prefer the biased estimator over the unbiased one. Mean square error (MSE) is a criterion which tries to take into account concerns about both bias and variance of estimators. MSE(ˆθ) =E[(ˆθ θ) 2 ] the expected size of the squared error, which is the difference between the estimate ˆθ and the actual parameter θ

MSE can be re-stated Show that the MSE of an estimate can be re-stated in terms of its variance and its bias, so that MSE(ˆθ) =Var(ˆθ)+[Bias(ˆθ)] 2

Moving from one population of interest to two Parameters and sample statistics that have been discussed so far only apply to one population. What if we want to compare two populations? Example: We want to calculate the difference in the mean income in the year after graduation between economics majors and other social science majors µ 1 µ 2 Example: We want to calculate the difference in the proportion of students who go on to grad school between economics majors and other social science majors p 1 p 2

Comparing two populations Try to develop a point estimate for these quantities based on estimators we already have For the difference between two means, µ 1 µ 2, we try the estimator x 1 x 2 For the difference between two proportions, p 1 p 2, we try the estimator ˆp 1 ˆp 2 We want to evaluate the goodness of these estimators. What do we know about the sampling distributions for these estimators? Are they unbiased? What is their variance?

Mean and variance of x 1 x 2 Show that x 1 x 2 is an unbiased estimator for µ 1 µ 2. Also show that the variance of this estimator is σ2 1 n 1 + σ2 2 n 2

Mean and variance of ˆp 1 ˆp 2 Show that ˆp 1 ˆp 2 is an unbiased estimator for p 1 p 2. Also show that the variance of this estimator is p 1 (1 p 1 ) n 1 + p 2(1 p 2 ) n 2

Summary of two sample estimators We have just shown that x 1 x 2 and ˆp 1 ˆp 2 are unbiased estimators, as were x and ˆp The CLT doesn t apply to these estimators since they are not sample means - they are differences of sample means Other theorems do state that given at least moderate (n 30) sample sizes, these estimators have sampling distributions that are approximately normal

Estimation errors Even with a good point estimate ˆθ, there is very likely to be some error (ˆθ = θ not likely) We can express this error of estimation, denoted ε, asε = ˆθ θ This is the number of units that our estimate, ˆθ, isofffromθ (doesn t take into account the direction of the error) We can use the sampling distribution of ˆθ to help place some bounds on our estimate