Chapter 5: Statistical Inference (in General)

Similar documents
Chapter 7: Point Estimation and Sampling Distributions

1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by. Cov(X, Y ) = E(X E(X))(Y E(Y ))

Chapter 7 - Lecture 1 General concepts and criteria

8.1 Estimation of the Mean and Proportion

Chapter 5. Statistical inference for Parametric Models

Chapter 8. Introduction to Statistical Inference

The normal distribution is a theoretical model derived mathematically and not empirically.

Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel

BIO5312 Biostatistics Lecture 5: Estimations

Chapter 7 Sampling Distributions and Point Estimation of Parameters

MVE051/MSG Lecture 7

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

Chapter 7: SAMPLING DISTRIBUTIONS & POINT ESTIMATION OF PARAMETERS

MATH 3200 Exam 3 Dr. Syring

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Confidence Intervals Introduction

Chapter 3 Discrete Random Variables and Probability Distributions

Introduction to Statistics I

Business Statistics 41000: Probability 4

Elementary Statistics Lecture 5

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

Binomial Random Variables. Binomial Random Variables

STAT 241/251 - Chapter 7: Central Limit Theorem

Chapter 5. Sampling Distributions

Statistical analysis and bootstrapping

MATH 264 Problem Homework I

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Review of the Topics for Midterm I

Probability Theory. Mohamed I. Riffi. Islamic University of Gaza

1. Statistical problems - a) Distribution is known. b) Distribution is unknown.

STAT Chapter 7: Central Limit Theorem

Section The Sampling Distribution of a Sample Mean

Chapter 3 Discrete Random Variables and Probability Distributions

Chapter 3 Statistical Quality Control, 7th Edition by Douglas C. Montgomery. Copyright (c) 2013 John Wiley & Sons, Inc.

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

Unit 5: Sampling Distributions of Statistics

Applied Statistics I

Unit 5: Sampling Distributions of Statistics

Chapter 7. Sampling Distributions and the Central Limit Theorem

Normal distribution Approximating binomial distribution by normal 2.10 Central Limit Theorem

IEOR 3106: Introduction to OR: Stochastic Models. Fall 2013, Professor Whitt. Class Lecture Notes: Tuesday, September 10.

Lecture 9 - Sampling Distributions and the CLT

Chapter 7. Sampling Distributions and the Central Limit Theorem

Central Limit Theorem (cont d) 7/28/2006

4-2 Probability Distributions and Probability Density Functions. Figure 4-2 Probability determined from the area under f(x).

Shifting our focus. We were studying statistics (data, displays, sampling...) The next few lectures focus on probability (randomness) Why?

Interval estimation. September 29, Outline Basic ideas Sampling variation and CLT Interval estimation using X More general problems

Chapter 9: Sampling Distributions

Point Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased.

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

Some Discrete Distribution Families

Random Variables CHAPTER 6.3 BINOMIAL AND GEOMETRIC RANDOM VARIABLES

Sampling Distributions and the Central Limit Theorem

Sampling and sampling distribution

Introduction to Statistical Data Analysis II

Sampling. Marc H. Mehlman University of New Haven. Marc Mehlman (University of New Haven) Sampling 1 / 20.

BIOL The Normal Distribution and the Central Limit Theorem

CHAPTER 5 SAMPLING DISTRIBUTIONS

4.3 Normal distribution

Statistics 6 th Edition

6 Central Limit Theorem. (Chs 6.4, 6.5)

Chapter 6: Random Variables

Bayesian Normal Stuff

Lecture 10: Point Estimation

2011 Pearson Education, Inc

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 10: Continuous RV Families. Prof. Vince Calhoun

Tutorial 6. Sampling Distribution. ENGG2450A Tutors. 27 February The Chinese University of Hong Kong 1/6

8.1 Binomial Distributions

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Commonly Used Distributions

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

Much of what appears here comes from ideas presented in the book:

E509A: Principle of Biostatistics. GY Zou

Statistical estimation

Random Variables. Chapter 6: Random Variables 2/2/2014. Discrete and Continuous Random Variables. Transforming and Combining Random Variables

The Binomial Probability Distribution

Random Variables Handout. Xavier Vilà

UNIVERSITY OF VICTORIA Midterm June 2014 Solutions

CVE SOME DISCRETE PROBABILITY DISTRIBUTIONS

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

Key Objectives. Module 2: The Logic of Statistical Inference. Z-scores. SGSB Workshop: Using Statistical Data to Make Decisions

Tutorial 11: Limit Theorems. Baoxiang Wang & Yihan Zhang bxwang, April 10, 2017

Statistics 431 Spring 2007 P. Shaman. Preliminaries

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Sampling Distributions

Review of commonly missed questions on the online quiz. Lecture 7: Random variables] Expected value and standard deviation. Let s bet...

Continuous random variables

Lecture 3. Sampling distributions. Counts, Proportions, and sample mean.

Stat 213: Intro to Statistics 9 Central Limit Theorem

Random variables. Contents

Midterm Exam III Review

Statistics and Their Distributions

Discrete Random Variables and Probability Distributions

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example...

Sampling Distributions

Point Estimation. Copyright Cengage Learning. All rights reserved.

Discrete Probability Distributions and application in Business

Data Analysis and Statistical Methods Statistics 651

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

Transcription:

Chapter 5: Statistical Inference (in General) Shiwen Shen University of South Carolina 2016 Fall Section 003 1 / 17

Motivation In chapter 3, we learn the discrete probability distributions, including Bernoulli, Binomial, Geometric, Negative Binomial, Hypergeometric, and Poisson. In chapter 4, we learn the continuous probability distributions, including Exponential, Weibull, and Normal. In chapter 3 and 4, we always assume that we know the parameter of the distribution. For example, we know the mean µ and variance σ 2 for a normal distributed random varialbe, so that we can calculate all kinds of probabilities with them. 2 / 17

Motivation For example, suppose we know the height of 18-year-old US male follows N(µ = 176.4, σ 2 = 9) in centimeters. Let Y = the height of one 18-year-old US male. We can calculate P(Y > 180) = 1 pnorm(180, 176.4, 3)= 0.115. However, it is natural that we do NOT know the population mean µ and population variance σ 2 in reality. What should we do? We use statistical inference! Statistical inference deals with making (probabilistic) statements about a population of individuals based on information that is contained in a sample taken from the population. 3 / 17

Terminology: population/sample A population refers to the entire group of individuals (e.g., people, parts, batteries, etc.) about which we would like to make a statement (e.g., height probability, median weight, defective proportion, mean lifetime, etc.). Problem: Population can not be measured (generally) Solution: We observe a sample of individuals from the population to draw inference We denote a random sample of observations by Y 1, Y 2,..., Y n n is the sample size Denote y1, y 2,..., y n to be one realization. 4 / 17

Example BATTERY DATA: Consider the following random sample of n = 50 battery lifetimes y 1, y 2,..., y 50 (measured in hours): 4285 2066 2584 1009 318 1429 981 1402 1137 414 564 604 14 4152 737 852 1560 1786 520 396 1278 209 349 478 3032 1461 701 1406 261 83 205 602 3770 726 3894 2662 497 35 2778 1379 3920 1379 99 510 582 308 3367 99 373 454 5 / 17

A histogram of battery lifetime data 6 / 17

Cont d on battery lifetime data The (empirical) distribution of the battery lifetimes is skewed towards the high side Which continuous probability distribution seems to display the same type of pattern that we see in histogram? An exponential(λ) models seems reasonable here (based in the histogram shape). What is λ? In this example, λ is called a (population) parameter (generally unknown). It describes the theoretical distribution which is used to model the entire population of battery lifetimes. 7 / 17

Terminology: parameter A parameter is a numerical quantity that describes a population. In general, population parameters are unknown. Some very common examples are: µ = population mean σ 2 = population variance σ = population standard deviation p = population proportion Connection: all of the probability distributions that we talked about in previous chapter are indexed by population parameters. 8 / 17

Terminology: statistics A statistic is a numerical quantity that can be calculated from a sample of data. Suppose Y 1, Y 2,..., Y n is a random sample from a population, some very common examples are: sample mean: n sample variance: Y = 1 n i=1 Y i S 2 = 1 n 1 n (Y i Y ) 2 i=1 sample standard deviation: S = S 2 1 sample proportion: p = n i=1 n Y i if Y i s are binary. 9 / 17

Back to battery lifetime data With the battery lifetime data (a random sample of n = 50 lifetimes), R code: y = 1274.14 hours s 2 = 1505156 (hours) 2 s 1226.85 hours > mean(battery) ## sample mean [1] 1274.14 > var(battery) ## sample variance [1] 1505156 > sd(battery) ## sample standard deviation [1] 1226.848 10 / 17

Parameters and Statistics Cont d SUMMARY: The table below succinctly summarizes the salient differences between a population and a sample (a parameter and a statistic): Comparison between parameters and statistics Statistics Parameters Describes a sample Describes a population Always known Usually unknown Random, changes upon repeated sampling Fixed Ex: X, S 2, S Ex: µ, σ 2, σ 11 / 17

Statistical Inference Statistical inference deals with making (probabilistic) statements about a population of individuals based on information that is contained in a sample taken from the population. We do this by estimating unknown poopulation parameters with sample statistics. quantifying the uncertainty (variability) that arises in the estimation process. 12 / 17

Point estimators and sampling distributions Let θ denote a population parameter. A point estimator ˆθ is a statistic that is used to estimate a population parameter θ. Common examples of point estimators are: θ = Y a point estimator for θ = µ θ = S 2 a point estimator for θ = σ 2 θ = S a point estimator for θ = σ Remark: In general, θ is a statistic, the value of θ will vary from sample to sample. 13 / 17

Terminology: sampling distribution The distribution of an estimator θ is called its sampling distribution. A sampling distribution describes mathematically how θ would vary in repeated sampling. What is a good estimator? And good in what sense? 14 / 17

Evaluate an estimator Accuracy: We say that θ is an unbiased estimator of θ if and only if E( θ) = θ RESULT: When Y 1,..., Y n is a random sample, E(Y ) = µ E(S 2 ) = σ 2 Precision: Suppose that θ 1 and θ 2 are unbiased estimators of θ. We would like to pick the estimator with smaller variance, since it is more likely to produce an estimate close to the true value θ. 15 / 17

Evaluate an estimator: cont d SUMMARY: We desire point estimators θ which are unbiased (perfectly accurate) and have small variance (highly precise). TERMINOLOGY: The standard error of a point estimator θ is equal to se( θ) = var( θ). Note: smaller se( θ) θ more precise. 16 / 17

Evaluate an estimator: cont d Which estimator is better? Why? 17 / 17