Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel

Size: px
Start display at page:

Download "Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel"

Transcription

1 STATISTICS Lecture no. 10 Department of Econometrics FEM UO Brno office 69a, tel

2 Introduction Suppose that we manufacture lightbulbs and we want to state the average lifetime on the box. Let us say that we have following five observed lifetimes (in hours) which have the average If it is all the information we have, it seems to be reasonable to state 1086 as the average lifetime.

3 Introduction Let the random variable X be the lifetime of a lightbulb, and let E(X ) = µ. Here µ is an unknown parameter. We decide to repeat the experiment to measure a lifetime 5 times and will then get an outcome on the five random variables X 1,..., X 5 that are i.i.d. (independent identically distributed). We now estimate µ by X = X i which is the sample mean.

4 Point Estimator Definition Let X 1,..., X n be a random sample. The statistic (random variable) T = T (X 1, X 2,..., X n ) = T (X), which is a function of the random sample and is used to estimate an unknown parameter θ, is called a point estimator of θ. We write T (X) = ˆθ.

5 Point Estimator Definition The estimator T (X) is said to be unbiased estimator the parameter θ if E [T (X)] = θ.

6 Point Estimator Definition The estimator T (X) is said to be unbiased estimator the parameter θ if E [T (X)] = θ. The difference B(θ, T ) = E [T (X)] θ is called a bias of the estimator T (X).

7 Example Point Estimator Let X 1, X 2,..., X n be a random sample from a distribution with the mean µ and the variance σ 2. The sample mean X is an unbiased estimator of µ, because ( ) 1 E(X ) = E X i = 1 E(X i ) = µ. n n

8 Example Point Estimator Let X 1, X 2,..., X n be a random sample from a distribution with the mean µ and the variance σ 2. The sample mean X is an unbiased estimator of µ, because ( ) 1 E(X ) = E X i = 1 E(X i ) = µ. n n The sample variance S 2 is an unbiased estimator of σ 2, because ( ) E(S 2 1 ) = E (X i X ) 2 = = σ 2. n 1

9 Example Point Estimator Let X 1, X 2,..., X n be a random sample from a distribution with the mean µ and the variance σ 2. The (moment) variance Sn 2 is a biased estimator of σ 2, because ( ) E(Sn) 2 1 = E (X i X ) 2 = = n 1 n n σ2.

10 Example Point Estimator Let X 1, X 2,..., X n be a random sample from a distribution with the mean µ and the variance σ 2. The (moment) variance Sn 2 is a biased estimator of σ 2, because ( ) E(Sn) 2 1 = E (X i X ) 2 = = n 1 n n σ2. The bias of the estimator S 2 n is B(σ 2, S 2 n) = E(S 2 n) σ 2 = n 1 n σ2 σ 2 = 1 n σ2. The bias decreases for large n.

11 Some estimators are biased but their bias decrease when n increases. Definition If lim E [T (X)] = θ, n then the estimator T (X) is said to be asymptotically unbiased estimator of the parameter θ. It easy to see that lim E [T (X) θ] = 0. n

12 Example Point Estimator The (moment) variance is an asymptotically unbiased estimator of σ 2, because lim E(S 2 n 1 n n) = lim n n σ2 = σ 2.

13 Definition The statistic T (X) is a consistent estimator of the parameter θ if for every ɛ > 0 lim P( T (X) θ < ɛ) = 1. n

14 Definition The statistic T (X) is a consistent estimator of the parameter θ if for every ɛ > 0 If lim P( T (X) θ < ɛ) = 1. n lim B(θ, T ) = 0 and lim D[T (X)] = 0, n n then T (X) is the consistent estimator of θ.

15 Example Point Estimator Prove that the sample mean is a consistent estimator of the expected value µ.

16 Example Point Estimator Prove that the sample mean is a consistent estimator of the expected value µ. According to E(X ) = µ and D(X ) = σ 2 /n we obtain σ 2 B(µ, X ) = E(X ) µ = 0 a lim D(X ) = lim n n n = 0.

17 If we have two unbiased estimators T 1 (X) = ˆθ and T 2 (X) = θ, which should we choose? Intuitively, we should choose the one that tends to be closer to θ, and since E(T 1 ) = E(T 2 ) = θ, it makes sense to choose the estimator with the smaller variance. Definition Suppose that T 1 (X) = ˆθ and T 2 (X) = θ are two unbiased estimators of θ. If D(T 1 (X)) < D(T 2 (X)) then T 1 (X) = ˆθ is said to be more efficient than T 2 (X) = θ.

18 Example Point Estimator We can find two unbiased estimators of a parameter λ of Poisson distribution E(X ) = λ and E(S 2 ) = λ. It is possible to calculate that D(X ) < D(S 2 ). The estimator X is more efficient then the estimator S 2.

19 How to Compare Estimators? Let us suppose we would like to compare unbiased and biased estimators of the parameter θ. In this case might not be suitable to choose one of the smallest variance. The estimator T has the smallest variance but has a large bias. Even the estimator with the smallest bias is not necessary the best one. The estimator U has no bias but its variance is to large. The estimator V seems to be the best.

20 Point Estimator Definition The mean square error of the estimator T of a parameter θ is defined as MSE(T ) = E(T θ) 2 = D(T ) + B 2 (θ, T ) (MSE of estimator = variance of estimator + bias 2 ), where T θ is a sample error.

21 Point Estimator The mean square error indicates the average sample error of estimates which can be calculated for all possible random sample of the size n.

22 Point Estimator The mean square error indicates the average sample error of estimates which can be calculated for all possible random sample of the size n. is a combination of 2 required properties (a small bias and a small variance), that why it is an universal criterion.

23 Point Estimator The mean square error indicates the average sample error of estimates which can be calculated for all possible random sample of the size n. is a combination of 2 required properties (a small bias and a small variance), that why it is an universal criterion. If T is an unbiased estimator then MSE(T ) = D(T ). Another possibility how to measure an accuracy of estimators is standard error SE = D(T ).

24 Example Point Estimator The sample mean is an unbiased estimator of the expected value µ, the standard error is equal to the standard deviation of the sample mean SE = D(X ) = σ(x ) = σ(x ). n σ(x ) is unknown, we have to estimate it by the sample standard deviation and we get the estimation ŜE = ˆσ(X ) n = S n.

25 Example Point Estimator Find the mean square error of S 2 and S 2 n. Let us start with the statistic S 2 which is an unbiased estimator of σ 2. MSE(S 2 ) = D(S 2 ) = E(S 2 σ 2 ) 2 = E(S 4 ) 2σ 2 E(σ 2 ) + σ 4 = = E(S 4 ) σ 4 = 2σ4 n 1.

26 Example Point Estimator Find the mean square error of S 2 and S 2 n. Let us start with the statistic S 2 which is an unbiased estimator of σ 2. MSE(S 2 ) = D(S 2 ) = E(S 2 σ 2 ) 2 = E(S 4 ) 2σ 2 E(σ 2 ) + σ 4 = = E(S 4 ) σ 4 = 2σ4 n 1. The MSE of the estimator S 2 n is MSE(Sn) 2 = 2 σ 2 ) 2 = E(Sn) 4 2 n 1 n σ4 + σ 4 = = E(Sn) 4 2 n n σ4 = 2n 1 σ 4, n 2. MSE(S 2 n) < MSE(S 2 ) because 2n 1 n 2 < 2 n 1

27 Method of Moments Method of Maximum Likelihood The definitions of unbiasness and other properties of estimators do not provide any guidance about how good estimators can be obtained. In this part, we discuss two methods for obtaining point estimators: the method of moments, the method of maximum likelihood. Maximum likelihood estimates are generally preferable to moment estimators because they have better efficiency properties. However, moment estimators are sometimes easier to compute. Both methods can produce unbiased point estimators.

28 Method of Moments Method of Maximum Likelihood Method of Moments The general idea behind the method of moments is to equate population moments, which are defined in terms of expected values, to the corresponding sample moments. The population moments will be functions of the unknown parameters. Then these equations are solved to yield estimators of the unknown parameters.

29 Method of Moments Method of Maximum Likelihood Method of Moments Let us assume the distribution with m 1 real parameters θ 1, θ 2,..., θ m and let X 1, X 2,..., X n be a random sample from this distribution. Let us suppose that exist moments µ r = E(X r i ) for r = 1, 2,..., m. These moments depend on the parameters θ 1, θ 2,..., θ m. Sample moments are defined by the formula M r = 1 n X r i, r = 1, 2....

30 Method of Moments Method of Maximum Likelihood Method of Moments Let X 1,..., X n be a random sample from either a probability function or probability density function with m unknown parameters θ 1,..., θ m. The moment estimators are found by equating the first m population moments to the first m sample moments and solving the resulting equations for the unknown parameters µ r = M r.

31 Method of Moments Method of Maximum Likelihood Example Estimation of the parameter λ Poisson distribution. Suppose that X 1,..., X n is a random sample from the Poisson distribution Po(λ), we get an equation µ 1 = M 1 E(X i ) = 1 n X i, the estimator ˆλ of the parameter λ is ˆλ = X.

32 Method of Moments Method of Maximum Likelihood Example Estimation of the parameters µ and σ 2 normal distribution. Suppose that X 1,..., X n is a random sample from the normal distribution N(µ, σ 2 ). µ 1 = M 1 E(X i ) = 1 n X i,

33 Method of Moments Method of Maximum Likelihood Example Estimation of the parameters µ and σ 2 normal distribution. Suppose that X 1,..., X n is a random sample from the normal distribution N(µ, σ 2 ). µ 2 = M 2 E(X 2 i ) = 1 n µ 1 = M 1 E(X i ) = 1 n X 2 i σ 2 + µ 2 = 1 n X i, D(X i ) + E(X i ) 2 = 1 n X 2 i X 2 i

34 Method of Moments Method of Maximum Likelihood Example Estimation of the parameters µ and σ 2 normal distribution. Suppose that X 1,..., X n is a random sample from the normal distribution N(µ, σ 2 ). µ 2 = M 2 E(X 2 i ) = 1 n We obtain estimators. ˆµ = X, ˆσ 2 = 1 n µ 1 = M 1 E(X i ) = 1 n X 2 i X 2 i σ 2 + µ 2 = 1 n X 2 = 1 n X i, D(X i ) + E(X i ) 2 = 1 n X 2 i X 2 i (X i X ) 2 = S 2 n = n 1 n S 2

35 Method of Moments Method of Maximum Likelihood Method of Maximum Likelihood Let X 1, X 2,..., X n be a random sample from either a probability density function f (x, θ) or a probability function p(x, θ) with an unknown parameter θ = (θ 1, θ 2,..., θ m ). A random vector X = (X 1, X 2,..., X n ) has either a joint probability density function or probability function or g(x, θ) = g(x 1, x 2,..., x n, θ) = f (x 1, θ)f (x 2, θ) f (x n, θ) g(x, θ) = g(x 1, x 2,..., x n, θ) = p(x 1, θ)p(x 2, θ) p(x n, θ).

36 Method of Moments Method of Maximum Likelihood Method of Maximum Likelihood The density g(x, θ) is a function of x with a given value of θ. If values x are given (observed data) than g(x, θ) is a function of a variable θ. We denote it L(θ, x) and call it a likelihood function. If exists some ˆθ which fulfils L(ˆθ, x) L(θ, x), then ˆθ is a maximum likelihood estimator of the parameter θ.

37 Method of Moments Method of Maximum Likelihood Method of Maximum Likelihood The density g(x, θ) is a function of x with a given value of θ. If values x are given (observed data) than g(x, θ) is a function of a variable θ. We denote it L(θ, x) and call it a likelihood function. If exists some ˆθ which fulfils L(ˆθ, x) L(θ, x), then ˆθ is a maximum likelihood estimator of the parameter θ. Sometimes is reasonable to use a logarithm of the likelihood function L(θ, x) = ln L(θ, x). For the maximum likelihood estimator we can write L(ˆθ, x) L(θ, x), because the logarithm is an increasing function.

38 Method of Moments Method of Maximum Likelihood Method of Maximum Likelihood The Maximum likelihood estimator of the vector θ = (θ 1, θ 2,..., θ m ) we obtain by solving a system of equations L(θ, x) θ i = 0, i = 1, 2,..., m.

39 Method of Moments Method of Maximum Likelihood Example Let X be a Bernoulli random variable. The probability function is { π p(x) = x (1 π) 1 x x = 0, 1, 0 otherwise. The likelihood function is L(π, x) = π x 1 (1 π) 1 x 1 π x 2 (1 π) 1 x 2... π xn (1 π) 1 xn = = π P n x i (1 π) n P n x i The logarithm of L(π, x) is ( L(π, x) = x i ln π + n ) x i ln(1 π).

40 Method of Moments Method of Maximum Likelihood Example We calculate the maximum of L(π, x) dl(π, x) dπ = n x i π n n x i 1 π = 0, and get the estimator ˆπ = n x i n = x.

41 Method of Moments Method of Maximum Likelihood Example Find a maximum likelihood estimator of a parameter λ of Poisson distribution Po(λ). λ P n x i L(λ, x) = e nλ x 1!x 2! x n!, L(λ, x) = ln L(λ, x) = nλ + dl(λ, x) dλ = n + ˆλ = 1 n x i ln λ ln(x 1!x 2! x n!) x i = x. x i 1 λ = 0

Chapter 7 - Lecture 1 General concepts and criteria

Chapter 7 - Lecture 1 General concepts and criteria Chapter 7 - Lecture 1 General concepts and criteria January 29th, 2010 Best estimator Mean Square error Unbiased estimators Example Unbiased estimators not unique Special case MVUE Bootstrap General Question

More information

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ. 9 Point estimation 9.1 Rationale behind point estimation When sampling from a population described by a pdf f(x θ) or probability function P [X = x θ] knowledge of θ gives knowledge of the entire population.

More information

Lecture 10: Point Estimation

Lecture 10: Point Estimation Lecture 10: Point Estimation MSU-STT-351-Sum-17B (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 1 / 31 Basic Concepts of Point Estimation A point estimate of a parameter θ,

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS Answer any FOUR of the SIX questions.

More information

Chapter 8. Introduction to Statistical Inference

Chapter 8. Introduction to Statistical Inference Chapter 8. Introduction to Statistical Inference Point Estimation Statistical inference is to draw some type of conclusion about one or more parameters(population characteristics). Now you know that a

More information

Chapter 7: Point Estimation and Sampling Distributions

Chapter 7: Point Estimation and Sampling Distributions Chapter 7: Point Estimation and Sampling Distributions Seungchul Baek Department of Statistics, University of South Carolina STAT 509: Statistics for Engineers 1 / 20 Motivation In chapter 3, we learned

More information

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage 6 Point Estimation Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Point Estimation Statistical inference: directed toward conclusions about one or more parameters. We will use the generic

More information

Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems

Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems Spring 2005 1. Which of the following statements relate to probabilities that can be interpreted as frequencies?

More information

Exercise. Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1. Exercise Estimation

Exercise. Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1. Exercise Estimation Exercise Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1 Exercise S 2 = = = = n i=1 (X i x) 2 n i=1 = (X i µ + µ X ) 2 = n 1 n 1 n i=1 ((X

More information

Chapter 4: Asymptotic Properties of MLE (Part 3)

Chapter 4: Asymptotic Properties of MLE (Part 3) Chapter 4: Asymptotic Properties of MLE (Part 3) Daniel O. Scharfstein 09/30/13 1 / 1 Breakdown of Assumptions Non-Existence of the MLE Multiple Solutions to Maximization Problem Multiple Solutions to

More information

Chapter 5: Statistical Inference (in General)

Chapter 5: Statistical Inference (in General) Chapter 5: Statistical Inference (in General) Shiwen Shen University of South Carolina 2016 Fall Section 003 1 / 17 Motivation In chapter 3, we learn the discrete probability distributions, including Bernoulli,

More information

Statistical estimation

Statistical estimation Statistical estimation Statistical modelling: theory and practice Gilles Guillot gigu@dtu.dk September 3, 2013 Gilles Guillot (gigu@dtu.dk) Estimation September 3, 2013 1 / 27 1 Introductory example 2

More information

MATH 3200 Exam 3 Dr. Syring

MATH 3200 Exam 3 Dr. Syring . Suppose n eligible voters are polled (randomly sampled) from a population of size N. The poll asks voters whether they support or do not support increasing local taxes to fund public parks. Let M be

More information

Point Estimation. Copyright Cengage Learning. All rights reserved.

Point Estimation. Copyright Cengage Learning. All rights reserved. 6 Point Estimation Copyright Cengage Learning. All rights reserved. 6.2 Methods of Point Estimation Copyright Cengage Learning. All rights reserved. Methods of Point Estimation The definition of unbiasedness

More information

EE641 Digital Image Processing II: Purdue University VISE - October 29,

EE641 Digital Image Processing II: Purdue University VISE - October 29, EE64 Digital Image Processing II: Purdue University VISE - October 9, 004 The EM Algorithm. Suffient Statistics and Exponential Distributions Let p(y θ) be a family of density functions parameterized by

More information

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality Point Estimation Some General Concepts of Point Estimation Statistical inference = conclusions about parameters Parameters == population characteristics A point estimate of a parameter is a value (based

More information

Applied Statistics I

Applied Statistics I Applied Statistics I Liang Zhang Department of Mathematics, University of Utah July 14, 2008 Liang Zhang (UofU) Applied Statistics I July 14, 2008 1 / 18 Point Estimation Liang Zhang (UofU) Applied Statistics

More information

Back to estimators...

Back to estimators... Back to estimators... So far, we have: Identified estimators for common parameters Discussed the sampling distributions of estimators Introduced ways to judge the goodness of an estimator (bias, MSE, etc.)

More information

STAT 509: Statistics for Engineers Dr. Dewei Wang. Copyright 2014 John Wiley & Sons, Inc. All rights reserved.

STAT 509: Statistics for Engineers Dr. Dewei Wang. Copyright 2014 John Wiley & Sons, Inc. All rights reserved. STAT 509: Statistics for Engineers Dr. Dewei Wang Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger 7 Point CHAPTER OUTLINE 7-1 Point Estimation 7-2

More information

Chapter 6: Point Estimation

Chapter 6: Point Estimation Chapter 6: Point Estimation Professor Sharabati Purdue University March 10, 2014 Professor Sharabati (Purdue University) Point Estimation Spring 2014 1 / 37 Chapter Overview Point estimator and point estimate

More information

Statistical analysis and bootstrapping

Statistical analysis and bootstrapping Statistical analysis and bootstrapping p. 1/15 Statistical analysis and bootstrapping Michel Bierlaire michel.bierlaire@epfl.ch Transport and Mobility Laboratory Statistical analysis and bootstrapping

More information

BIO5312 Biostatistics Lecture 5: Estimations

BIO5312 Biostatistics Lecture 5: Estimations BIO5312 Biostatistics Lecture 5: Estimations Yujin Chung September 27th, 2016 Fall 2016 Yujin Chung Lec5: Estimations Fall 2016 1/34 Recap Yujin Chung Lec5: Estimations Fall 2016 2/34 Today s lecture and

More information

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی یادگیري ماشین توزیع هاي نمونه و تخمین نقطه اي پارامترها Sampling Distributions and Point Estimation of Parameter (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی درس هفتم 1 Outline Introduction

More information

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems. Practice Exercises for Midterm Exam ST 522 - Statistical Theory - II The ACTUAL exam will consists of less number of problems. 1. Suppose X i F ( ) for i = 1,..., n, where F ( ) is a strictly increasing

More information

Point Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased.

Point Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased. Point Estimation Point Estimation Definition A point estimate of a parameter θ is a single number that can be regarded as a sensible value for θ. A point estimate is obtained by selecting a suitable statistic

More information

Random Variables Handout. Xavier Vilà

Random Variables Handout. Xavier Vilà Random Variables Handout Xavier Vilà Course 2004-2005 1 Discrete Random Variables. 1.1 Introduction 1.1.1 Definition of Random Variable A random variable X is a function that maps each possible outcome

More information

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ. Sufficient Statistics Lecture Notes 6 Sufficiency Data reduction in terms of a particular statistic can be thought of as a partition of the sample space X. Definition T is sufficient for θ if the conditional

More information

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. Random Variables 2 A random variable X is a numerical (integer, real, complex, vector etc.) summary of the outcome of the random experiment.

More information

Tutorial 11: Limit Theorems. Baoxiang Wang & Yihan Zhang bxwang, April 10, 2017

Tutorial 11: Limit Theorems. Baoxiang Wang & Yihan Zhang bxwang, April 10, 2017 Tutorial 11: Limit Theorems Baoxiang Wang & Yihan Zhang bxwang, yhzhang@cse.cuhk.edu.hk April 10, 2017 1 Outline The Central Limit Theorem (CLT) Normal Approximation Based on CLT De Moivre-Laplace Approximation

More information

Econ 300: Quantitative Methods in Economics. 11th Class 10/19/09

Econ 300: Quantitative Methods in Economics. 11th Class 10/19/09 Econ 300: Quantitative Methods in Economics 11th Class 10/19/09 Statistical thinking will one day be as necessary for efficient citizenship as the ability to read and write. --H.G. Wells discuss test [do

More information

Simulation Wrap-up, Statistics COS 323

Simulation Wrap-up, Statistics COS 323 Simulation Wrap-up, Statistics COS 323 Today Simulation Re-cap Statistics Variance and confidence intervals for simulations Simulation wrap-up FYI: No class or office hours Thursday Simulation wrap-up

More information

Chapter 5. Statistical inference for Parametric Models

Chapter 5. Statistical inference for Parametric Models Chapter 5. Statistical inference for Parametric Models Outline Overview Parameter estimation Method of moments How good are method of moments estimates? Interval estimation Statistical Inference for Parametric

More information

may be of interest. That is, the average difference between the estimator and the truth. Estimators with Bias(ˆθ) = 0 are called unbiased.

may be of interest. That is, the average difference between the estimator and the truth. Estimators with Bias(ˆθ) = 0 are called unbiased. 1 Evaluating estimators Suppose you observe data X 1,..., X n that are iid observations with distribution F θ indexed by some parameter θ. When trying to estimate θ, one may be interested in determining

More information

6. Genetics examples: Hardy-Weinberg Equilibrium

6. Genetics examples: Hardy-Weinberg Equilibrium PBCB 206 (Fall 2006) Instructor: Fei Zou email: fzou@bios.unc.edu office: 3107D McGavran-Greenberg Hall Lecture 4 Topics for Lecture 4 1. Parametric models and estimating parameters from data 2. Method

More information

Chapter 8: Sampling distributions of estimators Sections

Chapter 8: Sampling distributions of estimators Sections Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample variance Skip: p.

More information

MVE051/MSG Lecture 7

MVE051/MSG Lecture 7 MVE051/MSG810 2017 Lecture 7 Petter Mostad Chalmers November 20, 2017 The purpose of collecting and analyzing data Purpose: To build and select models for parts of the real world (which can be used for

More information

Random Samples. Mathematics 47: Lecture 6. Dan Sloughter. Furman University. March 13, 2006

Random Samples. Mathematics 47: Lecture 6. Dan Sloughter. Furman University. March 13, 2006 Random Samples Mathematics 47: Lecture 6 Dan Sloughter Furman University March 13, 2006 Dan Sloughter (Furman University) Random Samples March 13, 2006 1 / 9 Random sampling Definition We call a sequence

More information

Learning From Data: MLE. Maximum Likelihood Estimators

Learning From Data: MLE. Maximum Likelihood Estimators Learning From Data: MLE Maximum Likelihood Estimators 1 Parameter Estimation Assuming sample x1, x2,..., xn is from a parametric distribution f(x θ), estimate θ. E.g.: Given sample HHTTTTTHTHTTTHH of (possibly

More information

Statistics 6 th Edition

Statistics 6 th Edition Statistics 6 th Edition Chapter 5 Discrete Probability Distributions Chap 5-1 Definitions Random Variables Random Variables Discrete Random Variable Continuous Random Variable Ch. 5 Ch. 6 Chap 5-2 Discrete

More information

PROBABILITY AND STATISTICS

PROBABILITY AND STATISTICS Monday, January 12, 2015 1 PROBABILITY AND STATISTICS Zhenyu Ye January 12, 2015 Monday, January 12, 2015 2 References Ch10 of Experiments in Modern Physics by Melissinos. Particle Physics Data Group Review

More information

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions Frequentist Methods: 7.5 Maximum Likelihood Estimators

More information

Random variables. Contents

Random variables. Contents Random variables Contents 1 Random Variable 2 1.1 Discrete Random Variable............................ 3 1.2 Continuous Random Variable........................... 5 1.3 Measures of Location...............................

More information

Chapter 8: Sampling distributions of estimators Sections

Chapter 8: Sampling distributions of estimators Sections Chapter 8 continued Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample

More information

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved. 4-1 Chapter 4 Commonly Used Distributions 2014 by The Companies, Inc. All rights reserved. Section 4.1: The Bernoulli Distribution 4-2 We use the Bernoulli distribution when we have an experiment which

More information

Chapter 5. Sampling Distributions

Chapter 5. Sampling Distributions Lecture notes, Lang Wu, UBC 1 Chapter 5. Sampling Distributions 5.1. Introduction In statistical inference, we attempt to estimate an unknown population characteristic, such as the population mean, µ,

More information

Central Limit Theorem (cont d) 7/28/2006

Central Limit Theorem (cont d) 7/28/2006 Central Limit Theorem (cont d) 7/28/2006 Central Limit Theorem for Binomial Distributions Theorem. For the binomial distribution b(n, p, j) we have lim npq b(n, p, np + x npq ) = φ(x), n where φ(x) is

More information

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections 1 / 40 Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: Chapter 7 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods:

More information

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example...

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example... Chapter 4 Point estimation Contents 4.1 Introduction................................... 2 4.2 Estimating a population mean......................... 2 4.2.1 The problem with estimating a population mean

More information

Point Estimation. Edwin Leuven

Point Estimation. Edwin Leuven Point Estimation Edwin Leuven Introduction Last time we reviewed statistical inference We saw that while in probability we ask: given a data generating process, what are the properties of the outcomes?

More information

M.I.T Fall Practice Problems

M.I.T Fall Practice Problems M.I.T. 15.450-Fall 2010 Sloan School of Management Professor Leonid Kogan Practice Problems 1. Consider a 3-period model with t = 0, 1, 2, 3. There are a stock and a risk-free asset. The initial stock

More information

Much of what appears here comes from ideas presented in the book:

Much of what appears here comes from ideas presented in the book: Chapter 11 Robust statistical methods Much of what appears here comes from ideas presented in the book: Huber, Peter J. (1981), Robust statistics, John Wiley & Sons (New York; Chichester). There are many

More information

CSE 312 Winter Learning From Data: Maximum Likelihood Estimators (MLE)

CSE 312 Winter Learning From Data: Maximum Likelihood Estimators (MLE) CSE 312 Winter 2017 Learning From Data: Maximum Likelihood Estimators (MLE) 1 Parameter Estimation Given: independent samples x1, x2,..., xn from a parametric distribution f(x θ) Goal: estimate θ. Not

More information

STAT/MATH 395 PROBABILITY II

STAT/MATH 395 PROBABILITY II STAT/MATH 395 PROBABILITY II Distribution of Random Samples & Limit Theorems Néhémy Lim University of Washington Winter 2017 Outline Distribution of i.i.d. Samples Convergence of random variables The Laws

More information

Lecture Note 8 of Bus 41202, Spring 2017: Stochastic Diffusion Equation & Option Pricing

Lecture Note 8 of Bus 41202, Spring 2017: Stochastic Diffusion Equation & Option Pricing Lecture Note 8 of Bus 41202, Spring 2017: Stochastic Diffusion Equation & Option Pricing We shall go over this note quickly due to time constraints. Key concept: Ito s lemma Stock Options: A contract giving

More information

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise. Math 224 Q Exam 3A Fall 217 Tues Dec 12 Version A Problem 1. Let X be the continuous random variable defined by the following pdf: { 1 x/2 when x 2, f(x) otherwise. (a) Compute the mean µ E[X]. E[X] x

More information

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation IEOR E4703: Monte-Carlo Simulation Simulating Stochastic Differential Equations Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation IEOR E4703: Monte-Carlo Simulation Further Variance Reduction Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Outline

More information

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi Chapter 4: Commonly Used Distributions Statistics for Engineers and Scientists Fourth Edition William Navidi 2014 by Education. This is proprietary material solely for authorized instructor use. Not authorized

More information

Likelihood Methods of Inference. Toss coin 6 times and get Heads twice.

Likelihood Methods of Inference. Toss coin 6 times and get Heads twice. Methods of Inference Toss coin 6 times and get Heads twice. p is probability of getting H. Probability of getting exactly 2 heads is 15p 2 (1 p) 4 This function of p, is likelihood function. Definition:

More information

UNIVERSITY OF VICTORIA Midterm June 2014 Solutions

UNIVERSITY OF VICTORIA Midterm June 2014 Solutions UNIVERSITY OF VICTORIA Midterm June 04 Solutions NAME: STUDENT NUMBER: V00 Course Name & No. Inferential Statistics Economics 46 Section(s) A0 CRN: 375 Instructor: Betty Johnson Duration: hour 50 minutes

More information

3 ˆθ B = X 1 + X 2 + X 3. 7 a) Find the Bias, Variance and MSE of each estimator. Which estimator is the best according

3 ˆθ B = X 1 + X 2 + X 3. 7 a) Find the Bias, Variance and MSE of each estimator. Which estimator is the best according STAT 345 Spring 2018 Homework 9 - Point Estimation Name: Please adhere to the homework rules as given in the Syllabus. 1. Mean Squared Error. Suppose that X 1, X 2 and X 3 are independent random variables

More information

Homework Assignments

Homework Assignments Homework Assignments Week 1 (p. 57) #4.1, 4., 4.3 Week (pp 58 6) #4.5, 4.6, 4.8(a), 4.13, 4.0, 4.6(b), 4.8, 4.31, 4.34 Week 3 (pp 15 19) #1.9, 1.1, 1.13, 1.15, 1.18 (pp 9 31) #.,.6,.9 Week 4 (pp 36 37)

More information

8.1 Estimation of the Mean and Proportion

8.1 Estimation of the Mean and Proportion 8.1 Estimation of the Mean and Proportion Statistical inference enables us to make judgments about a population on the basis of sample information. The mean, standard deviation, and proportions of a population

More information

Lecture 22. Survey Sampling: an Overview

Lecture 22. Survey Sampling: an Overview Math 408 - Mathematical Statistics Lecture 22. Survey Sampling: an Overview March 25, 2013 Konstantin Zuev (USC) Math 408, Lecture 22 March 25, 2013 1 / 16 Survey Sampling: What and Why In surveys sampling

More information

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections 1 / 31 : Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods: 7.5 Maximum Likelihood

More information

IEOR 165 Lecture 1 Probability Review

IEOR 165 Lecture 1 Probability Review IEOR 165 Lecture 1 Probability Review 1 Definitions in Probability and Their Consequences 1.1 Defining Probability A probability space (Ω, F, P) consists of three elements: A sample space Ω is the set

More information

Random Variables and Probability Functions

Random Variables and Probability Functions University of Central Arkansas Random Variables and Probability Functions Directory Table of Contents. Begin Article. Stephen R. Addison Copyright c 001 saddison@mailaps.org Last Revision Date: February

More information

Week 1 Quantitative Analysis of Financial Markets Basic Statistics A

Week 1 Quantitative Analysis of Financial Markets Basic Statistics A Week 1 Quantitative Analysis of Financial Markets Basic Statistics A Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 October

More information

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION INSTITUTE AND FACULTY OF ACTUARIES Curriculum 2019 SPECIMEN EXAMINATION Subject CS1A Actuarial Statistics Time allowed: Three hours and fifteen minutes INSTRUCTIONS TO THE CANDIDATE 1. Enter all the candidate

More information

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 10: Continuous RV Families. Prof. Vince Calhoun

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 10: Continuous RV Families. Prof. Vince Calhoun ECE 340 Probabilistic Methods in Engineering M/W 3-4:15 Lecture 10: Continuous RV Families Prof. Vince Calhoun 1 Reading This class: Section 4.4-4.5 Next class: Section 4.6-4.7 2 Homework 3.9, 3.49, 4.5,

More information

Computer Statistics with R

Computer Statistics with R MAREK GAGOLEWSKI KONSTANCJA BOBECKA-WESO LOWSKA PRZEMYS LAW GRZEGORZEWSKI Computer Statistics with R 5. Point Estimation Faculty of Mathematics and Information Science Warsaw University of Technology []

More information

The Vasicek Distribution

The Vasicek Distribution The Vasicek Distribution Dirk Tasche Lloyds TSB Bank Corporate Markets Rating Systems dirk.tasche@gmx.net Bristol / London, August 2008 The opinions expressed in this presentation are those of the author

More information

STAT Chapter 7: Central Limit Theorem

STAT Chapter 7: Central Limit Theorem STAT 251 - Chapter 7: Central Limit Theorem In this chapter we will introduce the most important theorem in statistics; the central limit theorem. What have we seen so far? First, we saw that for an i.i.d

More information

Continuous random variables

Continuous random variables Continuous random variables probability density function (f(x)) the probability distribution function of a continuous random variable (analogous to the probability mass function for a discrete random variable),

More information

Commonly Used Distributions

Commonly Used Distributions Chapter 4: Commonly Used Distributions 1 Introduction Statistical inference involves drawing a sample from a population and analyzing the sample data to learn about the population. We often have some knowledge

More information

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial. Lecture 21,22, 23 Text: A Course in Probability by Weiss 8.5 STAT 225 Introduction to Probability Models March 31, 2014 Standard Sums of Whitney Huang Purdue University 21,22, 23.1 Agenda 1 2 Standard

More information

Review of key points about estimators

Review of key points about estimators Review of key points about estimators Populations can be at least partially described by population parameters Population parameters include: mean, proportion, variance, etc. Because populations are often

More information

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation IEOR E4703: Monte-Carlo Simulation Simulation Efficiency and an Introduction to Variance Reduction Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University

More information

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is: **BEGINNING OF EXAMINATION** 1. You are given: (i) A random sample of five observations from a population is: 0.2 0.7 0.9 1.1 1.3 (ii) You use the Kolmogorov-Smirnov test for testing the null hypothesis,

More information

Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making

Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making May 30, 2016 The purpose of this case study is to give a brief introduction to a heavy-tailed distribution and its distinct behaviors in

More information

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions April 9th, 2018 Lecture 20: Special distributions Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters 4, 6: Random variables Week 9 Chapter

More information

Confidence Intervals Introduction

Confidence Intervals Introduction Confidence Intervals Introduction A point estimate provides no information about the precision and reliability of estimation. For example, the sample mean X is a point estimate of the population mean μ

More information

ECE 295: Lecture 03 Estimation and Confidence Interval

ECE 295: Lecture 03 Estimation and Confidence Interval ECE 295: Lecture 03 Estimation and Confidence Interval Spring 2018 Prof Stanley Chan School of Electrical and Computer Engineering Purdue University 1 / 23 Theme of this Lecture What is Estimation? You

More information

Interval estimation. September 29, Outline Basic ideas Sampling variation and CLT Interval estimation using X More general problems

Interval estimation. September 29, Outline Basic ideas Sampling variation and CLT Interval estimation using X More general problems Interval estimation September 29, 2017 STAT 151 Class 7 Slide 1 Outline of Topics 1 Basic ideas 2 Sampling variation and CLT 3 Interval estimation using X 4 More general problems STAT 151 Class 7 Slide

More information

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as Lecture 0 on BST 63: Statistical Theory I Kui Zhang, 09/9/008 Review for the previous lecture Definition: Several continuous distributions, including uniform, gamma, normal, Beta, Cauchy, double exponential

More information

Lecture 18 Section Mon, Feb 16, 2009

Lecture 18 Section Mon, Feb 16, 2009 The s the Lecture 18 Section 5.3.4 Hampden-Sydney College Mon, Feb 16, 2009 Outline The s the 1 2 3 The 4 s 5 the 6 The s the Exercise 5.12, page 333. The five-number summary for the distribution of income

More information

The Bernoulli distribution

The Bernoulli distribution This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc. INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc. Summary of the previous lecture Moments of a distribubon Measures of

More information

Lecture 18 Section Mon, Sep 29, 2008

Lecture 18 Section Mon, Sep 29, 2008 The s the Lecture 18 Section 5.3.4 Hampden-Sydney College Mon, Sep 29, 2008 Outline The s the 1 2 3 The 4 s 5 the 6 The s the Exercise 5.12, page 333. The five-number summary for the distribution of income

More information

Comparing the Means of. Two Log-Normal Distributions: A Likelihood Approach

Comparing the Means of. Two Log-Normal Distributions: A Likelihood Approach Journal of Statistical and Econometric Methods, vol.3, no.1, 014, 137-15 ISSN: 179-660 (print), 179-6939 (online) Scienpress Ltd, 014 Comparing the Means of Two Log-Normal Distributions: A Likelihood Approach

More information

4.2 Probability Distributions

4.2 Probability Distributions 4.2 Probability Distributions Definition. A random variable is a variable whose value is a numerical outcome of a random phenomenon. The probability distribution of a random variable tells us what the

More information

Review of key points about estimators

Review of key points about estimators Review of key points about estimators Populations can be at least partially described by population parameters Population parameters include: mean, proportion, variance, etc. Because populations are often

More information

Section The Sampling Distribution of a Sample Mean

Section The Sampling Distribution of a Sample Mean Section 5.2 - The Sampling Distribution of a Sample Mean Statistics 104 Autumn 2004 Copyright c 2004 by Mark E. Irwin The Sampling Distribution of a Sample Mean Example: Quality control check of light

More information

Lecture 17: More on Markov Decision Processes. Reinforcement learning

Lecture 17: More on Markov Decision Processes. Reinforcement learning Lecture 17: More on Markov Decision Processes. Reinforcement learning Learning a model: maximum likelihood Learning a value function directly Monte Carlo Temporal-difference (TD) learning COMP-424, Lecture

More information

Statistics for Business and Economics

Statistics for Business and Economics Statistics for Business and Economics Chapter 7 Estimation: Single Population Copyright 010 Pearson Education, Inc. Publishing as Prentice Hall Ch. 7-1 Confidence Intervals Contents of this chapter: Confidence

More information

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics Unit 5: Sampling Distributions of Statistics Statistics 571: Statistical Methods Ramón V. León 6/12/2004 Unit 5 - Stat 571 - Ramon V. Leon 1 Definitions and Key Concepts A sample statistic used to estimate

More information

Modelling financial data with stochastic processes

Modelling financial data with stochastic processes Modelling financial data with stochastic processes Vlad Ardelean, Fabian Tinkl 01.08.2012 Chair of statistics and econometrics FAU Erlangen-Nuremberg Outline Introduction Stochastic processes Volatility

More information

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics Unit 5: Sampling Distributions of Statistics Statistics 571: Statistical Methods Ramón V. León 6/12/2004 Unit 5 - Stat 571 - Ramon V. Leon 1 Definitions and Key Concepts A sample statistic used to estimate

More information

Econometric Methods for Valuation Analysis

Econometric Methods for Valuation Analysis Econometric Methods for Valuation Analysis Margarita Genius Dept of Economics M. Genius (Univ. of Crete) Econometric Methods for Valuation Analysis Cagliari, 2017 1 / 25 Outline We will consider econometric

More information

STAT 241/251 - Chapter 7: Central Limit Theorem

STAT 241/251 - Chapter 7: Central Limit Theorem STAT 241/251 - Chapter 7: Central Limit Theorem In this chapter we will introduce the most important theorem in statistics; the central limit theorem. What have we seen so far? First, we saw that for an

More information