Welcome to Stat 410!

Similar documents
Random Variables Handout. Xavier Vilà

STOR Lecture 7. Random Variables - I

STAT Chapter 4/6: Random Variables and Probability Distributions

Chapter 3 - Lecture 5 The Binomial Probability Distribution

Commonly Used Distributions

IEOR 165 Lecture 1 Probability Review

Chapter 2. Random variables. 2.3 Expectation

5. In fact, any function of a random variable is also a random variable

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

4 Random Variables and Distributions

Chapter 3 Discrete Random Variables and Probability Distributions

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics

Lecture 9: Plinko Probabilities, Part III Random Variables, Expected Values and Variances

6. Continous Distributions

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

Statistics and Probability

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

Probability and Random Variables A FINANCIAL TIMES COMPANY

Probability Theory. Mohamed I. Riffi. Islamic University of Gaza

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

MATH MW Elementary Probability Course Notes Part IV: Binomial/Normal distributions Mean and Variance

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

Chapter 7: Random Variables

Binomial Random Variables. Binomial Random Variables

Chapter 6: Random Variables. Ch. 6-3: Binomial and Geometric Random Variables

STAT/MATH 395 PROBABILITY II

Probability Distributions for Discrete RV

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

Chapter 8 Homework Solutions Compiled by Joe Kahlig. speed(x) freq 25 x < x < x < x < x < x < 55 5

Engineering Statistics ECIV 2305

Review of the Topics for Midterm I

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Statistics & Flood Frequency Chapter 3. Dr. Philip B. Bedient

1/2 2. Mean & variance. Mean & standard deviation

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

Chapter 3 Discrete Random Variables and Probability Distributions

Random variables. Contents

Business Statistics 41000: Probability 4

Random Variable: Definition

Elementary Statistics Lecture 5

2011 Pearson Education, Inc

2 of PU_2015_375 Which of the following measures is more flexible when compared to other measures?

Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is

Lecture III. 1. common parametric models 2. model fitting 2a. moment matching 2b. maximum likelihood 3. hypothesis testing 3a. p-values 3b.

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

Probability. An intro for calculus students P= Figure 1: A normal integral

TOPIC: PROBABILITY DISTRIBUTIONS

Discrete Random Variables and Probability Distributions. Stat 4570/5570 Based on Devore s book (Ed 8)

Exam 2 Spring 2015 Statistics for Applications 4/9/2015

Statistics 6 th Edition

What was in the last lecture?

Statistics for Business and Economics

4.3 Normal distribution

Math489/889 Stochastic Processes and Advanced Mathematical Finance Homework 5

Part 1 In which we meet the law of averages. The Law of Averages. The Expected Value & The Standard Error. Where Are We Going?

MA : Introductory Probability

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

4.2 Bernoulli Trials and Binomial Distributions

STAT Chapter 7: Central Limit Theorem

The Normal Distribution

MA 1125 Lecture 14 - Expected Values. Wednesday, October 4, Objectives: Introduce expected values.

Random Variables and Probability Functions

Normal distribution Approximating binomial distribution by normal 2.10 Central Limit Theorem

CS145: Probability & Computing

Statistics. Marco Caserta IE University. Stats 1 / 56

A random variable (r. v.) is a variable whose value is a numerical outcome of a random phenomenon.

Law of Large Numbers, Central Limit Theorem

Simple Random Sample

Probability is the tool used for anticipating what the distribution of data should look like under a given model.

Section Random Variables and Histograms

MATH 264 Problem Homework I

Binomial Approximation and Joint Distributions Chris Piech CS109, Stanford University

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman:

Chapter 5 Discrete Probability Distributions. Random Variables Discrete Probability Distributions Expected Value and Variance

1 PMF and CDF Random Variable PMF and CDF... 4

INF FALL NATURAL LANGUAGE PROCESSING. Jan Tore Lønning, Lecture 3, 1.9

March 21, Fractal Friday: Sign up & pay now!

PROBABILITY DISTRIBUTIONS

Chapter 3 - Lecture 3 Expected Values of Discrete Random Va

Statistical Tables Compiled by Alan J. Terry

(Practice Version) Midterm Exam 1

Exam P September 2014 Study Sheet! copylefted by Jared Nakamura, 9/4/2014!

STAT 241/251 - Chapter 7: Central Limit Theorem

Statistics for Managers Using Microsoft Excel 7 th Edition

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Model Paper Statistics Objective. Paper Code Time Allowed: 20 minutes

CS134: Networks Spring Random Variables and Independence. 1.2 Probability Distribution Function (PDF) Number of heads Probability 2 0.

Chapter 5. Sampling Distributions

Chapter 3 - Lecture 4 Moments and Moment Generating Funct

7. For the table that follows, answer the following questions: x y 1-1/4 2-1/2 3-3/4 4

Data Analytics (CS40003) Practice Set IV (Topic: Probability and Sampling Distribution)

Introduction to Probability and Inference HSSP Summer 2017, Instructor: Alexandra Ding July 19, 2017

AP Statistics Test 5

STA Module 3B Discrete Random Variables

Homework: Due Wed, Feb 20 th. Chapter 8, # 60a + 62a (count together as 1), 74, 82

Econ 6900: Statistical Problems. Instructor: Yogesh Uppal

Chapter 7 1. Random Variables

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Transcription:

Welcome to Stat 410! Personnel Instructor: Liang, Feng TA: Gan, Gary (Lingrui) Instructors/TAs from two other sessions Websites: Piazza and Compass Homework When, where and how to submit your homework No late submissions will be accepted Remember to list your section and instructor on the 1st page Grading policy Three Exams, No Final. 1

Will keep you posted on where to pickup your graded hw; rules for reporting grading errors; rules for missing homework. 2

Communication For questions related to homework/lectures, please post your question on Piazza. By default, you are anonymous to your classmates, but not to the instructor. If you want to send email to my Illinois account, please 1. Write from your Illinois email account (so I would know who you are) 2. Start your subject line with the course number, e.g., [stat410] HW1 missing (since I m teaching two courses this semester) 3. Sign with your full name 4. Don t send unexpected attachments. 3

General Expectations Review the notes/book Go through practice problems Finish homework independently You can discuss homework problems with other students but should write your answers independently using your own words. Feedback and questions How to do well: Exercise, Exercise, Exercise 4

Discrete Random Variables X Bern(p): X = 0 or 1 P(X = 1) = p, P(X = 0) = 1 p. Toss a fair coin, X = 1 if head; 0 if tail, then X Bern(0.5). X Bin(n, p): X = number of 1 s from n independent Bern(p) trials. P(X = k) = p k (1 p) n k, k = 0, 1,..., n k where ( ) n k = n! k!(n k)! is the binomial coefficient represents the number of ways to select (unordered) k objects out of n given objects. 5

Note k =, n k 0 = n = 1, 1 = n 1 = n Using Binomial Theorem: (a + b) n = n k=0 a k b n k, n k check that n k=0 P(X = k) = (p + 1 p)n = 1. 6

X Geo(p): X = number of 0 s from independent Bern(p) trials before seeing the first 1. P(X = k) = (1 p) k p, k = 0, 1,... Sometimes Geo(p) is described as P(X = m) = (1 p) m 1 p, m = 1, 2,..., i.e., X =number of independent Bern(p) trials before seeing the first 1. X NB(r, p): X = number of 0 s from independent Bern(p) trials before seeing r 1 s. 7

X Po(λ): X can be viewed as the limit of Bern(n, p) when n is large and p is small and np λ. For example, X = number of Stat 410 students who have the same birthday as you. p k (1 p) n k = k n! k!(n k)! pk (1 p) n k = 1 k! n! (n k)! p k (1 p)n (1 p) k = 1 k! [np] [(n 1)p] [(n k + 1)p] 1 (1 p)n (1 p) k 1 k! λk e λ, k = 0, 1, 2,... 8

A jar contains 20 M&M candies and 7 red and 13 green. Sampling with replacement: randomly draw one M&M from the jar, record its color and put it back, and then repeat this 5 times. X = number of red candies you ve drew, which follows Bin(5, 7 20 ). Sampling without replacement: randomly draw one M&M from the jar, record its color and repeat this 5 times, i.e., randomly draw 5 candies from the jar. X = number of red candies you ve drew, which follows a Hypergometric distribution (20, 7, 5). 9

Uniform distribution over a discrete set, e.g., tossing a fair die will result a uniform distribution over set {1, 2,..., 6}. 10

Continuous Random Variables Uniform distribution, e.g., Unif(0, 1). Exponential distribution pdf f(x) = λe λx, x > 0, which is a special case of the Gamma distribution. Sometimes parameterized as f(x) = 1 β e x/β. Normal distribution: THE most important distribution in Statistics. Student t dist, F dist, and Chi-squared dist are related to the Normal distribution. Beta distribution over (0, 1). Unif(0, 1) is a special case of Beta. 11

Random Variables How to describe a random variable? pmf/pdf or CDF. Properties of pmf/pdf f(x): 1. f(x) 0; 2. its integral/summation over the range is equal to 1. Properties of CDF F (x) = P(X x) (Theorem 1.5.1, p.35) 1. F is nondecreasing, i.e., F (a) F (b) if a < b. 2. lim x F (x) = 0, lim x F (x) = 1. 3. F is right continuous, i.e., lim xn a F (x n ) = F (a). Mean, median (or quantiles), and mode of a distribution 12

Expectations E(X) = x xf(x) xf(x)dx, E[g(X)] = x g(x)f(x) g(x)f(x)dx, provided that those integrals and summations exist (i.e., E X or E g(x) finite). E(a) = a, where a is a constant. E(aX + by ) = ae(x) + be(y ) (E is a linear operator) k-th moment = E(X k ), µ = E(X), Var(X) = σ 2 = E(X µ) 2 = EX 2 (EX) 2. What does a rv with zero variance look like? Usually Eg(X) g(ex) when g is non-linear ( EX 2 [ E(X) ] 2 ) 13

Note and Tips Random variables are either discrete or continuous? NO In the discrete case, pmf f(x) = P(X = x), while in the continuous case, pdf f(x) is NOT equal to P(X = x) that is equal to zero. When computing probabilities involving continuous rvs, and < (or and > ) are interchangeable, but that s not the case for discrete rvs. Range! When describing a rv, remember to specify its range; when computing expectations, keep track of the range. 14

Moment Generating Functions M X (t) = Ee tx = x e tx f(x) or e tx f(x)dx, provided that the expectation exists for t in a small neighborhood of 0, otherwise we say that the mgf does not exist. We don t care about the mgf for any t, but t in an interval around zero, i.e., h < t < h for some h > 0. 15

The connection between moments and mgf: M X (t) = Ee tx = E [1 + tx 1 + t2 X 2 2! + t3 X 3 3! ] + = 1 + t EX 1 + t2 EX2 2! + t 3 EX3 3! + (1) Then, we have M (k) X (0) = E(Xk ), k = 0, 1,... 16

Another important use of mgf: it provides an alternative way to characterize a distribution. If the mgf for X 1 and the one for X 2 exist, and M X1 (t) = M X2 (t) = X 1, X 2 follow the same distribution. That is, the mgf uniquely determines a distribution. If we know all the moments of a r.v., then we know its distribution? Or in other words, do moments uniquely determine a distribution? The answer is No. You might be tempted to say Yes. If the moments are not too large, the mgf defined via (1) exists, then moments determine a distribution. 17