Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Size: px
Start display at page:

Download "Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ."

Transcription

1 Sufficient Statistics Lecture Notes 6 Sufficiency Data reduction in terms of a particular statistic can be thought of as a partition of the sample space X. Definition T is sufficient for θ if the conditional distribution of the sample X n T does not depend on θ, where we use X n (X,..., X n ). Thus, f(x n t; θ) f(x n t). Sample X X n X,, X n F. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ. We want to learn about θ and try to summarize the data without throwing any information about θ away. Calculate some statistic T (X,, X n ) that contains all available information about θ in the sample. If so, we say T is sufficient. Theorem 2 (Theorem 6.2.2, CB) If f(x n ; θ) is the joint pdf or pmf of X n and q(t; θ) is the pdf or pmf of T (X n ), then T (X n ) is a sufficient statistic for θ if, for every x n in the sample space, the ratio is constant as a function of θ. f(x n ; θ) q(t (x n ); θ) Exercise 2: Binomial sufficient statistic Let X,, X n be iid Bernoulli random variables with parameter θ, 0 < θ <. We show that T (X n ) n i X i is a sufficient statistic for θ. Note that T (X n ) has Binomial(n, θ) distribution. Thus for t n i x i, we have f(x n ; θ) q(t (x n ); θ) n i θx i ( θ) x i ( n ) θt ( θ) t n t θ n i x i ( θ) n ) θt ( θ) n t ( n t i ( x i) ( n t θ t ( θ) ) n t θt ( θ) n t ( n t) ( ) n n i x i

2 Interpretation: The total number of s in this Bernoulli sample contains all the information about θ that is in the data. Sum of Poisson Random variables Theorem 3 (Theorem 4.3.2, CB) if X Poisson(λ) and if Y Poisson(µ) and X and Y are independent, then X + Y Poisson(λ + µ) Example and exercise! X,, X n Poisson(θ). Denote by X n (X,..., X n ) and x n (x,..., x n ). Let T n i X i. Clearly T Poisson (nθ) by Theorem 3. Hence for t 0, P (T t) e nθ (nθ) t. t! We have where p X n T (x n t) P(X n x n T (X n ) t) P (Xn x n and T t), P (T t) { 0 if n P (X n x n and T t) i x n t P (X n x n ) if n i x n t Thus we focus on the case t n i x i, where x i 0, i,..., n: P (X n x n ) by independence of X,..., X n. Now n e θ θ x i i x i! e nθ θ x i (xi!) e nθ θ t (xi!) p X n T (x n t) P (Xn x n ) P (T t) t! xi! n i ( ) xi t! n xi! n t when n x i t i and p X n T (x n t) 0 otherwise; Clearly this is the Multinomial Distribution with parameter t and p i /n for all i,..., n. Note that this distribution does not depend on θ. So T i X i is a sufficient statistic for θ following the definition. 2 Factorization Theorem Theorem 4 (Theorem 6.2.6, CB) Let f(x n θ) denote the joint pdf or pmf of a sample X n. A statistic T (X n ) is a sufficient statistic for θ if and only if there exists functions g(t θ) and h(x n ) such that, for all sample points x n and all parameters θ, f(x n θ) g(t (x n ) θ)h(x n ). 2

3 Homework: read proof of Theorem 4 on CB P.276. exponential family of distributions is. Try also to understand what an Example 5 Let X,, X n Poisson. Then f(x n θ) e nθ θ X i (xi!) e nθ θ i Xi (xi!) Example 6 X,, X n iid N(µ, σ 2 ) f(x n µ, σ 2 ) ( ) n 2 2πσ 2 ( (a) If σ known f(x n µ) 2πσ 2 X n is sufficient for µ. ) n 2 { } (xi x) 2 + n(x µ) 2 exp 2σ 2 { } (xi x n ) 2 exp 2σ }{{ 2 } h(x n ) { } n(xn µ) 2 exp 2σ }{{ 2 } g(t (x n ) µ) () (b) If (µ, σ 2 ) unknown, by staring at (2), we have T (X n, S 2 n) is sufficient. So is T ( X i, X 2 i ). Note:. T is jointly sufficient for (µ, σ 2 ); 2. and the sample is sufficient also. 3 Sufficient Partitions Data reduction in terms of a particular statistic can be thought of as a partition of the sample space X. Let T {t : t T (x n ) for some x n X } be the image of X under T (x n ). Then T (x n ) partitions the sample space into sets B t, t T, defined by B t {x n : T (x n ) t} The statistic summarizes the data in that, rather than reporting the entire sample x n, it reports only that T (x n ) t or x n B t. Example 7 Let X, X 2, X 3 Bernoulli(θ). Let T X i. 3

4 B t x n t p(x n t) B 0 (0, 0, 0) t 0 B (0, 0, ) t /3 (0,, 0) t /3 (, 0, 0) t /3 B 2 (0,, ) t 2 /3 (, 0, ) t 2 /3 (,, 0) t 2 /3 B 3 (,, ) t 3 8 elements 4 elements. A partition B,..., B k is sufficient if f(x n X n B j ) does not depend on θ for all j. 2. A statistic T induces a partition. For each t, B t {x n : T (x n ) t} is one element of the partition. T is sufficient if and only if the partition is sufficient. 3. Two statistics can generate the same partition: example: i X i and 3 i X i. We wish to consider the conditional probability P (X n x n T (X n ) t). If x n is a sample point such that T (x n ) t, then clearly P θ (X n x n T (X n ) t) 0. So we are interested in P θ (X n x n T (X n ) T (x n )). We must verify that for any fixed values of x n and t, the conditional probability P θ (X n x n T (X n ) T (x n )) does not depend on θ. Since {X n x n } is a subset of {T (X n ) T (x n ), P θ (X n x n T (X n ) T (x n )) P θ(x n x n and T (X n ) T (x n )) P θ (T (X n ) T (x n )) P θ (X n x n ) P θ (T (X n ) T (x n )) f(x n ; θ) q(t (x n ); θ). This is exactly the content of Theorem 2. Example 8 For the discrete version of the example above, we showed that for all sample points x n, and Y n (X, X 2,..., X n) such that the mechanism of producing Y n is based on the following conditional probabilities given T (X n ), P (X n x n T (X n ) T (x n )) P (Y n x n T (X n ) T (x n )) Exercise: Show that P θ (X n x n ) P θ (Y n x n ). 4

5 Example 9 X,, X n iid N(µ, σ 2 ) ( ) n { } f(x n µ, σ 2 2 (xi x) 2 + n(x µ) 2 ) exp 2πσ 2 2σ 2 ( ) n { } 2 (xi ) 2 exp 2πσ 2 2σ }{{ 2 } g(t (x n ) µ0,σ 2 ) i X2 i is sufficient for σ 2 when µ is known to be zero. Note:. T is jointly sufficient for (µ, σ 2 ); 2. and the sample is sufficient also. Example 0 (Example.9.2 from Testing Statistical Hypotheses by E.L.Lehmann and J.P.Romano) If X,, X n iid N(0, σ 2 ), the conditional distribution of the sample point over each of the spheres, defined as {x n : n i x2 i constant} is uniform irrespective of σ 2. Hint: to understand this, you need to use the density function defined in the previous example to obtain the conditional distribution defined immediately above. One can therefore construct an equivalent sample X,..., X n from the knowledge of T (X n ) n i X2 i and a mechanism that can produce a point randomly distributed over a sphere. Equivalence here is in the following sense, For Y n (X, X 2,..., X n), f X n(x n ; θ) f Y n(x n ; θ). 4 Minimal Sufficient Statistics (MSS) Want the greatest reduction in dimension. Example X,, X n N(0, σ 2 ). Some sufficient statistics: T (X,, X n ) (X,, X n ) R n (X, 2, Xn) 2 R n+ m n ( Xi 2, Xi 2 ) R 2+ i X 2 i im+ minimal sufficient 5

6 Definition T is a Minimal Sufficient Statistic if (i) T is sufficient and (ii) T is a function of all other sufficient statistics. (It yields coarsest partition.) Suppose U is sufficient. Suppose T H(U) is also sufficient. T provides greater reduction than U unless H is a transformation, in which case T and U are equivalent. Example 2 X N(0, σ 2 ). X is sufficient. X is sufficient. X is MSS. So are X 2, X 4, e X2. 4. How to find a MSS Theorem 3 Suppose that T has the following property: for each pair x n, y n, R(x n, y n ; θ) f(yn ; θ) f(x n ; θ) does not depend on θ if and only if T (y n ) T (x n ). Then T is a MSS. Example 4 Y,, Y n iid Poisson (θ). f(y n θ) e nθ θ y i yi, f(y n ; θ) f(x n ; θ) θ yi x i yi!/ x i! which is independent of θ iff y i x i. This implies that T (Y n ) Y i is a minimal sufficient statistic for θ. The minimal sufficient statistic is not unique. But, the minimal sufficient partition is unique. Exercise 3: Understand the following example. Example 5 Let X, X 2, X 3 Bernoulli(θ). Then T X is not sufficient. Look at its partition: x n t p(x n t) (0, 0, 0) t 0 ( θ) 2 (0, 0, ) t 0 θ( θ) (0,, 0) t 0 θ( θ) (0,, ) t 0 θ 2 (, 0, 0) t ( θ) 2 (, 0, ) t θ( θ) (,, 0) t θ( θ) (,, ) t θ 2 8 elements 2 elements 6

7 5 Digression: Order Statistics Assume that the sample X, X 2,..., X n has been ordered from smallest to largest: X () X (2)... X (n). Theorem 6 Let X (), X (2),..., X (n) denote the order statistics of a random sample X,..., X n from a continuous population with cdf F X (x) and pdf f X (x). Then the CDF and pdf of X ( k) is: Proof. F X(k) (x) f X(k) (x) n jk ( ) n [F X (x)] j [ F X (x)] n j j n! (k )!(n k)! f X(x) [F X (x)] k [ F X (x)] n k. Let Y be the random variable counting the number of X,..., X n less or equal to x. Then, by defining a success as the event {X j x}, we see that Y binomial(n, F X (x)) as F X (x) P (X x). F X(k) (x) P (X (k) x) P (Y k) P (at least k of the X,..., X n x) n P (exactly j of the X,..., X n x) Differentiate to find the pdf. jk n jk ( ) n [F X (x)] j [ F X (x)] n j j Corollary 7 Let random variables X and X 2 be independent and both from a continuous distribution with cdf F X and pdf f X (x). Let Z : max{x, X 2 } X (2). The pdf of Z is: and the density is F Z (z) P (X z, X 2 z) F 2 X(z) f Z (z) df 2 X(z)/dz 2F X (z)f X (z). 7

8 6 Best Unbiased Estimators What is the smallest variance of an unbiased estimator? This was once considered an important question. Today we consider it not so important. There is no reason to require an estimator to be unbiased. Having small MSE is more important. Example 8 We have learned two unbiased estimators: Let X,..., X n be a random sample from a population with mean µ and variance σ 2 <. That is. µ E(X i ) and σ 2 Var(X i ). Then E(X n ) µ, Var(X n ) σ2 n, E(S2 n) σ 2. In your homework, you are asked to compute Var(S 2 n) in case X,..., X n N(µ, σ 2 ). Example 9 Let X,..., X n N(θ, σ 2 ). Then The MSE s are E(X) µ, E(S 2 ) σ 2. E(X µ) 2 σ2 n, E(S2 σ 2 ) 2 2σ4 n. We now calculate the MSE for the MLE estimators σ 2, namely, σ 2 n n (X i X) 2. i using MSE B 2 + V, for which we have for n, Exercise 2: E( σ 2 σ 2 ) 2 2n n 2 σ 4 < 2σ4 n E(S2 σ 2 ) 2. Verify everything involved in the last equation. Hint, you need to compute B and V of σ 2 based on those of S 2. However, for completeness, we will briefly consider the question. An estimator W is UMVUE (Uniform Minimum Variance Unbiased Estimator) for τ(θ) if (i) E θ (W ) τ(θ) for all θ and (ii) if E θ (W ) τ(θ) for all θ then Var θ (W ) Var θ (W ). The Cramer-Rao inequality gives a lower bound on the variance of any unbiased estimator, which we will study later. 8

9 Theorem 20 The Rao-Blackwell Theorem. Let W be an unbiased estimator of τ(θ) and let T be a sufficient statistic. Define W φ(t ) E(W T ). Then W is unbiased and Var θ (W ) Var θ (W ) for all θ. Proof. We have E θ (W ) E θ (E(W T )) E θ (W ) τ(θ) so W is unbiased. Also, Var θ (W ) Var θ (E(W T )) + E θ (Var(W T )) Var θ (W ) + E θ (Var(W T )) Var θ (W ). Implications: Conditioning any unbiased estimator on a sufficient statistic will result in a uniform improvement, so we need consider only statistics that are functions of a sufficient statistic in our serach for best unbiased estimators. Exercises: Convince yourself that W is an estimator that you can construct from the sample, that is, its expression is independent of θ by sufficiency. When does the eqaulity hold in the above formula? 7 Exponential Family We cover material from the text, Section 3.4. Example 2 Here are some examples of the Exponential Family: Binomial distribution when n is fixed and p (0, ) Double exponential distribution when µ is fixed and σ > 0 Log-normal distribution where < µ < and 0 < σ < Normal distribution N(µ, σ 2 ) where < µ < and 0 < σ < Gamma distribution where α, β > 0, More examples: Poisson, Multinomial distribution when n is fixed... 9

10 We will prove Theorem next week. We proved Theorem regarding the sufficient statistics for the Exponential Family. We can rewrite an exponential family in terms of a natural parameterization. For k we have f(x η) h(x) exp{ηt(x) A(η)} define A(η) log h(x) exp{ηt(x)}dx For example a Poisson can be written as f(x η) exp{ηx e η }/x! where the natural parameter is η log λ. Theorem. Let X have density in an exponential family. Then E (t(x)) A (η), Var (t(x)) A (η). This convenient representation and link to moments holds also for k >. 0

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as Lecture 0 on BST 63: Statistical Theory I Kui Zhang, 09/9/008 Review for the previous lecture Definition: Several continuous distributions, including uniform, gamma, normal, Beta, Cauchy, double exponential

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS Answer any FOUR of the SIX questions.

More information

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions April 9th, 2018 Lecture 20: Special distributions Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters 4, 6: Random variables Week 9 Chapter

More information

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ. 9 Point estimation 9.1 Rationale behind point estimation When sampling from a population described by a pdf f(x θ) or probability function P [X = x θ] knowledge of θ gives knowledge of the entire population.

More information

Chapter 8: Sampling distributions of estimators Sections

Chapter 8: Sampling distributions of estimators Sections Chapter 8 continued Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample

More information

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage 6 Point Estimation Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Point Estimation Statistical inference: directed toward conclusions about one or more parameters. We will use the generic

More information

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems. Practice Exercises for Midterm Exam ST 522 - Statistical Theory - II The ACTUAL exam will consists of less number of problems. 1. Suppose X i F ( ) for i = 1,..., n, where F ( ) is a strictly increasing

More information

STAT/MATH 395 PROBABILITY II

STAT/MATH 395 PROBABILITY II STAT/MATH 395 PROBABILITY II Distribution of Random Samples & Limit Theorems Néhémy Lim University of Washington Winter 2017 Outline Distribution of i.i.d. Samples Convergence of random variables The Laws

More information

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections 1 / 31 : Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods: 7.5 Maximum Likelihood

More information

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections 1 / 40 Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: Chapter 7 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods:

More information

MATH 3200 Exam 3 Dr. Syring

MATH 3200 Exam 3 Dr. Syring . Suppose n eligible voters are polled (randomly sampled) from a population of size N. The poll asks voters whether they support or do not support increasing local taxes to fund public parks. Let M be

More information

Exercise. Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1. Exercise Estimation

Exercise. Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1. Exercise Estimation Exercise Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1 Exercise S 2 = = = = n i=1 (X i x) 2 n i=1 = (X i µ + µ X ) 2 = n 1 n 1 n i=1 ((X

More information

Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems

Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems Spring 2005 1. Which of the following statements relate to probabilities that can be interpreted as frequencies?

More information

Hardy Weinberg Model- 6 Genotypes

Hardy Weinberg Model- 6 Genotypes Hardy Weinberg Model- 6 Genotypes Silvelyn Zwanzig Hardy -Weinberg with six genotypes. In a large population of plants (Mimulus guttatus there are possible alleles S, I, F at one locus resulting in six

More information

Chapter 4: Asymptotic Properties of MLE (Part 3)

Chapter 4: Asymptotic Properties of MLE (Part 3) Chapter 4: Asymptotic Properties of MLE (Part 3) Daniel O. Scharfstein 09/30/13 1 / 1 Breakdown of Assumptions Non-Existence of the MLE Multiple Solutions to Maximization Problem Multiple Solutions to

More information

Lecture 10: Point Estimation

Lecture 10: Point Estimation Lecture 10: Point Estimation MSU-STT-351-Sum-17B (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 1 / 31 Basic Concepts of Point Estimation A point estimate of a parameter θ,

More information

Statistical analysis and bootstrapping

Statistical analysis and bootstrapping Statistical analysis and bootstrapping p. 1/15 Statistical analysis and bootstrapping Michel Bierlaire michel.bierlaire@epfl.ch Transport and Mobility Laboratory Statistical analysis and bootstrapping

More information

What was in the last lecture?

What was in the last lecture? What was in the last lecture? Normal distribution A continuous rv with bell-shaped density curve The pdf is given by f(x) = 1 2πσ e (x µ)2 2σ 2, < x < If X N(µ, σ 2 ), E(X) = µ and V (X) = σ 2 Standard

More information

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. Random Variables 2 A random variable X is a numerical (integer, real, complex, vector etc.) summary of the outcome of the random experiment.

More information

Chapter 8: Sampling distributions of estimators Sections

Chapter 8: Sampling distributions of estimators Sections Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample variance Skip: p.

More information

STAT 830 Convergence in Distribution

STAT 830 Convergence in Distribution STAT 830 Convergence in Distribution Richard Lockhart Simon Fraser University STAT 830 Fall 2013 Richard Lockhart (Simon Fraser University) STAT 830 Convergence in Distribution STAT 830 Fall 2013 1 / 31

More information

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality Point Estimation Some General Concepts of Point Estimation Statistical inference = conclusions about parameters Parameters == population characteristics A point estimate of a parameter is a value (based

More information

Chapter 7 - Lecture 1 General concepts and criteria

Chapter 7 - Lecture 1 General concepts and criteria Chapter 7 - Lecture 1 General concepts and criteria January 29th, 2010 Best estimator Mean Square error Unbiased estimators Example Unbiased estimators not unique Special case MVUE Bootstrap General Question

More information

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial. Lecture 21,22, 23 Text: A Course in Probability by Weiss 8.5 STAT 225 Introduction to Probability Models March 31, 2014 Standard Sums of Whitney Huang Purdue University 21,22, 23.1 Agenda 1 2 Standard

More information

IEOR 165 Lecture 1 Probability Review

IEOR 165 Lecture 1 Probability Review IEOR 165 Lecture 1 Probability Review 1 Definitions in Probability and Their Consequences 1.1 Defining Probability A probability space (Ω, F, P) consists of three elements: A sample space Ω is the set

More information

Chapter 5. Statistical inference for Parametric Models

Chapter 5. Statistical inference for Parametric Models Chapter 5. Statistical inference for Parametric Models Outline Overview Parameter estimation Method of moments How good are method of moments estimates? Interval estimation Statistical Inference for Parametric

More information

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom Review for Final Exam 18.05 Spring 2014 Jeremy Orloff and Jonathan Bloom THANK YOU!!!! JON!! PETER!! RUTHI!! ERIKA!! ALL OF YOU!!!! Probability Counting Sets Inclusion-exclusion principle Rule of product

More information

Central Limit Theorem, Joint Distributions Spring 2018

Central Limit Theorem, Joint Distributions Spring 2018 Central Limit Theorem, Joint Distributions 18.5 Spring 218.5.4.3.2.1-4 -3-2 -1 1 2 3 4 Exam next Wednesday Exam 1 on Wednesday March 7, regular room and time. Designed for 1 hour. You will have the full

More information

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation IEOR E4703: Monte-Carlo Simulation Simulation Efficiency and an Introduction to Variance Reduction Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University

More information

Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel

Point Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel STATISTICS Lecture no. 10 Department of Econometrics FEM UO Brno office 69a, tel. 973 442029 email:jiri.neubauer@unob.cz 8. 12. 2009 Introduction Suppose that we manufacture lightbulbs and we want to state

More information

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial Lecture 23 STAT 225 Introduction to Probability Models April 4, 2014 approximation Whitney Huang Purdue University 23.1 Agenda 1 approximation 2 approximation 23.2 Characteristics of the random variable:

More information

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise. Math 224 Q Exam 3A Fall 217 Tues Dec 12 Version A Problem 1. Let X be the continuous random variable defined by the following pdf: { 1 x/2 when x 2, f(x) otherwise. (a) Compute the mean µ E[X]. E[X] x

More information

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی یادگیري ماشین توزیع هاي نمونه و تخمین نقطه اي پارامترها Sampling Distributions and Point Estimation of Parameter (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی درس هفتم 1 Outline Introduction

More information

Homework Assignments

Homework Assignments Homework Assignments Week 1 (p. 57) #4.1, 4., 4.3 Week (pp 58 6) #4.5, 4.6, 4.8(a), 4.13, 4.0, 4.6(b), 4.8, 4.31, 4.34 Week 3 (pp 15 19) #1.9, 1.1, 1.13, 1.15, 1.18 (pp 9 31) #.,.6,.9 Week 4 (pp 36 37)

More information

Probability Theory. Probability and Statistics for Data Science CSE594 - Spring 2016

Probability Theory. Probability and Statistics for Data Science CSE594 - Spring 2016 Probability Theory Probability and Statistics for Data Science CSE594 - Spring 2016 What is Probability? 2 What is Probability? Examples outcome of flipping a coin (seminal example) amount of snowfall

More information

Chapter 6: Point Estimation

Chapter 6: Point Estimation Chapter 6: Point Estimation Professor Sharabati Purdue University March 10, 2014 Professor Sharabati (Purdue University) Point Estimation Spring 2014 1 / 37 Chapter Overview Point estimator and point estimate

More information

Central Limit Thm, Normal Approximations

Central Limit Thm, Normal Approximations Central Limit Thm, Normal Approximations Engineering Statistics Section 5.4 Josh Engwer TTU 23 March 2016 Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 1 / 26 PART I PART I:

More information

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence continuous rv Let X be a continuous rv. Then a probability distribution or probability density function (pdf) of X is a function f(x) such that for any two numbers a and b with a b, P(a X b) = b a f (x)dx.

More information

Chapter 7: Point Estimation and Sampling Distributions

Chapter 7: Point Estimation and Sampling Distributions Chapter 7: Point Estimation and Sampling Distributions Seungchul Baek Department of Statistics, University of South Carolina STAT 509: Statistics for Engineers 1 / 20 Motivation In chapter 3, we learned

More information

6. Genetics examples: Hardy-Weinberg Equilibrium

6. Genetics examples: Hardy-Weinberg Equilibrium PBCB 206 (Fall 2006) Instructor: Fei Zou email: fzou@bios.unc.edu office: 3107D McGavran-Greenberg Hall Lecture 4 Topics for Lecture 4 1. Parametric models and estimating parameters from data 2. Method

More information

Qualifying Exam Solutions: Theoretical Statistics

Qualifying Exam Solutions: Theoretical Statistics Qualifying Exam Solutions: Theoretical Statistics. (a) For the first sampling plan, the expectation of any statistic W (X, X,..., X n ) is a polynomial of θ of degree less than n +. Hence τ(θ) cannot have

More information

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi Chapter 4: Commonly Used Distributions Statistics for Engineers and Scientists Fourth Edition William Navidi 2014 by Education. This is proprietary material solely for authorized instructor use. Not authorized

More information

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg :

More information

Chapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables

Chapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables Chapter 5 Continuous Random Variables and Probability Distributions 5.1 Continuous Random Variables 1 2CHAPTER 5. CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Probability Distributions Probability

More information

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions Frequentist Methods: 7.5 Maximum Likelihood Estimators

More information

4 Random Variables and Distributions

4 Random Variables and Distributions 4 Random Variables and Distributions Random variables A random variable assigns each outcome in a sample space. e.g. called a realization of that variable to Note: We ll usually denote a random variable

More information

Conjugate priors: Beta and normal Class 15, Jeremy Orloff and Jonathan Bloom

Conjugate priors: Beta and normal Class 15, Jeremy Orloff and Jonathan Bloom 1 Learning Goals Conjugate s: Beta and normal Class 15, 18.05 Jeremy Orloff and Jonathan Bloom 1. Understand the benefits of conjugate s.. Be able to update a beta given a Bernoulli, binomial, or geometric

More information

Statistics for Business and Economics

Statistics for Business and Economics Statistics for Business and Economics Chapter 5 Continuous Random Variables and Probability Distributions Ch. 5-1 Probability Distributions Probability Distributions Ch. 4 Discrete Continuous Ch. 5 Probability

More information

The Normal Distribution

The Normal Distribution The Normal Distribution The normal distribution plays a central role in probability theory and in statistics. It is often used as a model for the distribution of continuous random variables. Like all models,

More information

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc. INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc. Summary of the previous lecture Moments of a distribubon Measures of

More information

STAT 111 Recitation 4

STAT 111 Recitation 4 STAT 111 Recitation 4 Linjun Zhang http://stat.wharton.upenn.edu/~linjunz/ September 29, 2017 Misc. Mid-term exam time: 6-8 pm, Wednesday, Oct. 11 The mid-term break is Oct. 5-8 The next recitation class

More information

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved. 4-1 Chapter 4 Commonly Used Distributions 2014 by The Companies, Inc. All rights reserved. Section 4.1: The Bernoulli Distribution 4-2 We use the Bernoulli distribution when we have an experiment which

More information

Random Variables Handout. Xavier Vilà

Random Variables Handout. Xavier Vilà Random Variables Handout Xavier Vilà Course 2004-2005 1 Discrete Random Variables. 1.1 Introduction 1.1.1 Definition of Random Variable A random variable X is a function that maps each possible outcome

More information

CSE 312 Winter Learning From Data: Maximum Likelihood Estimators (MLE)

CSE 312 Winter Learning From Data: Maximum Likelihood Estimators (MLE) CSE 312 Winter 2017 Learning From Data: Maximum Likelihood Estimators (MLE) 1 Parameter Estimation Given: independent samples x1, x2,..., xn from a parametric distribution f(x θ) Goal: estimate θ. Not

More information

Exam 2 Spring 2015 Statistics for Applications 4/9/2015

Exam 2 Spring 2015 Statistics for Applications 4/9/2015 18.443 Exam 2 Spring 2015 Statistics for Applications 4/9/2015 1. True or False (and state why). (a). The significance level of a statistical test is not equal to the probability that the null hypothesis

More information

Probability. An intro for calculus students P= Figure 1: A normal integral

Probability. An intro for calculus students P= Figure 1: A normal integral Probability An intro for calculus students.8.6.4.2 P=.87 2 3 4 Figure : A normal integral Suppose we flip a coin 2 times; what is the probability that we get more than 2 heads? Suppose we roll a six-sided

More information

Back to estimators...

Back to estimators... Back to estimators... So far, we have: Identified estimators for common parameters Discussed the sampling distributions of estimators Introduced ways to judge the goodness of an estimator (bias, MSE, etc.)

More information

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION INSTITUTE AND FACULTY OF ACTUARIES Curriculum 2019 SPECIMEN EXAMINATION Subject CS1A Actuarial Statistics Time allowed: Three hours and fifteen minutes INSTRUCTIONS TO THE CANDIDATE 1. Enter all the candidate

More information

The Bernoulli distribution

The Bernoulli distribution This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Confidence Intervals Introduction

Confidence Intervals Introduction Confidence Intervals Introduction A point estimate provides no information about the precision and reliability of estimation. For example, the sample mean X is a point estimate of the population mean μ

More information

Homework Problems Stat 479

Homework Problems Stat 479 Chapter 10 91. * A random sample, X1, X2,, Xn, is drawn from a distribution with a mean of 2/3 and a variance of 1/18. ˆ = (X1 + X2 + + Xn)/(n-1) is the estimator of the distribution mean θ. Find MSE(

More information

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics Unit 5: Sampling Distributions of Statistics Statistics 571: Statistical Methods Ramón V. León 6/12/2004 Unit 5 - Stat 571 - Ramon V. Leon 1 Definitions and Key Concepts A sample statistic used to estimate

More information

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics Unit 5: Sampling Distributions of Statistics Statistics 571: Statistical Methods Ramón V. León 6/12/2004 Unit 5 - Stat 571 - Ramon V. Leon 1 Definitions and Key Concepts A sample statistic used to estimate

More information

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions. ME3620 Theory of Engineering Experimentation Chapter III. Random Variables and Probability Distributions Chapter III 1 3.2 Random Variables In an experiment, a measurement is usually denoted by a variable

More information

Engineering Statistics ECIV 2305

Engineering Statistics ECIV 2305 Engineering Statistics ECIV 2305 Section 5.3 Approximating Distributions with the Normal Distribution Introduction A very useful property of the normal distribution is that it provides good approximations

More information

Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is

Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is Normal Distribution Normal Distribution Definition A continuous rv X is said to have a normal distribution with parameter µ and σ (µ and σ 2 ), where < µ < and σ > 0, if the pdf of X is f (x; µ, σ) = 1

More information

Asymptotic results discrete time martingales and stochastic algorithms

Asymptotic results discrete time martingales and stochastic algorithms Asymptotic results discrete time martingales and stochastic algorithms Bernard Bercu Bordeaux University, France IFCAM Summer School Bangalore, India, July 2015 Bernard Bercu Asymptotic results for discrete

More information

Computer Statistics with R

Computer Statistics with R MAREK GAGOLEWSKI KONSTANCJA BOBECKA-WESO LOWSKA PRZEMYS LAW GRZEGORZEWSKI Computer Statistics with R 5. Point Estimation Faculty of Mathematics and Information Science Warsaw University of Technology []

More information

Chapter 5: Statistical Inference (in General)

Chapter 5: Statistical Inference (in General) Chapter 5: Statistical Inference (in General) Shiwen Shen University of South Carolina 2016 Fall Section 003 1 / 17 Motivation In chapter 3, we learn the discrete probability distributions, including Bernoulli,

More information

Commonly Used Distributions

Commonly Used Distributions Chapter 4: Commonly Used Distributions 1 Introduction Statistical inference involves drawing a sample from a population and analyzing the sample data to learn about the population. We often have some knowledge

More information

Chapter 3 - Lecture 4 Moments and Moment Generating Funct

Chapter 3 - Lecture 4 Moments and Moment Generating Funct Chapter 3 - Lecture 4 and s October 7th, 2009 Chapter 3 - Lecture 4 and Moment Generating Funct Central Skewness Chapter 3 - Lecture 4 and Moment Generating Funct Central Skewness The expected value of

More information

Conjugate Models. Patrick Lam

Conjugate Models. Patrick Lam Conjugate Models Patrick Lam Outline Conjugate Models What is Conjugacy? The Beta-Binomial Model The Normal Model Normal Model with Unknown Mean, Known Variance Normal Model with Known Mean, Unknown Variance

More information

ASSIGNMENT - 1, MAY M.Sc. (PREVIOUS) FIRST YEAR DEGREE STATISTICS. Maximum : 20 MARKS Answer ALL questions.

ASSIGNMENT - 1, MAY M.Sc. (PREVIOUS) FIRST YEAR DEGREE STATISTICS. Maximum : 20 MARKS Answer ALL questions. (DMSTT 0 NR) ASSIGNMENT -, MAY-04. PAPER- I : PROBABILITY AND DISTRIBUTION THEORY ) a) State and prove Borel-cantelli lemma b) Let (x, y) be jointly distributed with density 4 y(+ x) f( x, y) = y(+ x)

More information

Week 1 Quantitative Analysis of Financial Markets Basic Statistics A

Week 1 Quantitative Analysis of Financial Markets Basic Statistics A Week 1 Quantitative Analysis of Financial Markets Basic Statistics A Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 October

More information

Write legibly. Unreadable answers are worthless.

Write legibly. Unreadable answers are worthless. MMF 2021 Final Exam 1 December 2016. This is a closed-book exam: no books, no notes, no calculators, no phones, no tablets, no computers (of any kind) allowed. Do NOT turn this page over until you are

More information

Non-informative Priors Multiparameter Models

Non-informative Priors Multiparameter Models Non-informative Priors Multiparameter Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Prior Types Informative vs Non-informative There has been a desire for a prior distributions that

More information

MTH6154 Financial Mathematics I Stochastic Interest Rates

MTH6154 Financial Mathematics I Stochastic Interest Rates MTH6154 Financial Mathematics I Stochastic Interest Rates Contents 4 Stochastic Interest Rates 45 4.1 Fixed Interest Rate Model............................ 45 4.2 Varying Interest Rate Model...........................

More information

MA : Introductory Probability

MA : Introductory Probability MA 320-001: Introductory Probability David Murrugarra Department of Mathematics, University of Kentucky http://www.math.uky.edu/~dmu228/ma320/ Spring 2017 David Murrugarra (University of Kentucky) MA 320:

More information

INF FALL NATURAL LANGUAGE PROCESSING. Jan Tore Lønning, Lecture 3, 1.9

INF FALL NATURAL LANGUAGE PROCESSING. Jan Tore Lønning, Lecture 3, 1.9 INF5830 015 FALL NATURAL LANGUAGE PROCESSING Jan Tore Lønning, Lecture 3, 1.9 Today: More statistics Binomial distribution Continuous random variables/distributions Normal distribution Sampling and sampling

More information

STAT 111 Recitation 3

STAT 111 Recitation 3 STAT 111 Recitation 3 Linjun Zhang stat.wharton.upenn.edu/~linjunz/ September 23, 2017 Misc. The unpicked-up homeworks will be put in the STAT 111 box in the Stats Department lobby (It s on the 4th floor

More information

4-2 Probability Distributions and Probability Density Functions. Figure 4-2 Probability determined from the area under f(x).

4-2 Probability Distributions and Probability Density Functions. Figure 4-2 Probability determined from the area under f(x). 4-2 Probability Distributions and Probability Density Functions Figure 4-2 Probability determined from the area under f(x). 4-2 Probability Distributions and Probability Density Functions Definition 4-2

More information

Point Estimation. Copyright Cengage Learning. All rights reserved.

Point Estimation. Copyright Cengage Learning. All rights reserved. 6 Point Estimation Copyright Cengage Learning. All rights reserved. 6.2 Methods of Point Estimation Copyright Cengage Learning. All rights reserved. Methods of Point Estimation The definition of unbiasedness

More information

Chapter 2. Random variables. 2.3 Expectation

Chapter 2. Random variables. 2.3 Expectation Random processes - Chapter 2. Random variables 1 Random processes Chapter 2. Random variables 2.3 Expectation 2.3 Expectation Random processes - Chapter 2. Random variables 2 Among the parameters representing

More information

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman:

Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman: Math 224 Fall 207 Homework 5 Drew Armstrong Problems from 9th edition of Probability and Statistical Inference by Hogg, Tanis and Zimmerman: Section 3., Exercises 3, 0. Section 3.3, Exercises 2, 3, 0,.

More information

Statistics, Measures of Central Tendency I

Statistics, Measures of Central Tendency I Statistics, Measures of Central Tendency I We are considering a random variable X with a probability distribution which has some parameters. We want to get an idea what these parameters are. We perfom

More information

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 10: Continuous RV Families. Prof. Vince Calhoun

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 10: Continuous RV Families. Prof. Vince Calhoun ECE 340 Probabilistic Methods in Engineering M/W 3-4:15 Lecture 10: Continuous RV Families Prof. Vince Calhoun 1 Reading This class: Section 4.4-4.5 Next class: Section 4.6-4.7 2 Homework 3.9, 3.49, 4.5,

More information

Two Hours. Mathematical formula books and statistical tables are to be provided THE UNIVERSITY OF MANCHESTER. 22 January :00 16:00

Two Hours. Mathematical formula books and statistical tables are to be provided THE UNIVERSITY OF MANCHESTER. 22 January :00 16:00 Two Hours MATH38191 Mathematical formula books and statistical tables are to be provided THE UNIVERSITY OF MANCHESTER STATISTICAL MODELLING IN FINANCE 22 January 2015 14:00 16:00 Answer ALL TWO questions

More information

EE641 Digital Image Processing II: Purdue University VISE - October 29,

EE641 Digital Image Processing II: Purdue University VISE - October 29, EE64 Digital Image Processing II: Purdue University VISE - October 9, 004 The EM Algorithm. Suffient Statistics and Exponential Distributions Let p(y θ) be a family of density functions parameterized by

More information

SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS

SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS Questions 1-307 have been taken from the previous set of Exam C sample questions. Questions no longer relevant

More information

Ch4. Variance Reduction Techniques

Ch4. Variance Reduction Techniques Ch4. Zhang Jin-Ting Department of Statistics and Applied Probability July 17, 2012 Ch4. Outline Ch4. This chapter aims to improve the Monte Carlo Integration estimator via reducing its variance using some

More information

Introduction to Probability and Inference HSSP Summer 2017, Instructor: Alexandra Ding July 19, 2017

Introduction to Probability and Inference HSSP Summer 2017, Instructor: Alexandra Ding July 19, 2017 Introduction to Probability and Inference HSSP Summer 2017, Instructor: Alexandra Ding July 19, 2017 Please fill out the attendance sheet! Suggestions Box: Feedback and suggestions are important to the

More information

3 ˆθ B = X 1 + X 2 + X 3. 7 a) Find the Bias, Variance and MSE of each estimator. Which estimator is the best according

3 ˆθ B = X 1 + X 2 + X 3. 7 a) Find the Bias, Variance and MSE of each estimator. Which estimator is the best according STAT 345 Spring 2018 Homework 9 - Point Estimation Name: Please adhere to the homework rules as given in the Syllabus. 1. Mean Squared Error. Suppose that X 1, X 2 and X 3 are independent random variables

More information

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example...

4.1 Introduction Estimating a population mean The problem with estimating a population mean with a sample mean: an example... Chapter 4 Point estimation Contents 4.1 Introduction................................... 2 4.2 Estimating a population mean......................... 2 4.2.1 The problem with estimating a population mean

More information

IEOR 3106: Introduction to OR: Stochastic Models. Fall 2013, Professor Whitt. Class Lecture Notes: Tuesday, September 10.

IEOR 3106: Introduction to OR: Stochastic Models. Fall 2013, Professor Whitt. Class Lecture Notes: Tuesday, September 10. IEOR 3106: Introduction to OR: Stochastic Models Fall 2013, Professor Whitt Class Lecture Notes: Tuesday, September 10. The Central Limit Theorem and Stock Prices 1. The Central Limit Theorem (CLT See

More information

Exam M Fall 2005 PRELIMINARY ANSWER KEY

Exam M Fall 2005 PRELIMINARY ANSWER KEY Exam M Fall 005 PRELIMINARY ANSWER KEY Question # Answer Question # Answer 1 C 1 E C B 3 C 3 E 4 D 4 E 5 C 5 C 6 B 6 E 7 A 7 E 8 D 8 D 9 B 9 A 10 A 30 D 11 A 31 A 1 A 3 A 13 D 33 B 14 C 34 C 15 A 35 A

More information

Homework Problems Stat 479

Homework Problems Stat 479 Chapter 2 1. Model 1 is a uniform distribution from 0 to 100. Determine the table entries for a generalized uniform distribution covering the range from a to b where a < b. 2. Let X be a discrete random

More information

Two hours UNIVERSITY OF MANCHESTER. 23 May :00 16:00. Answer ALL SIX questions The total number of marks in the paper is 90.

Two hours UNIVERSITY OF MANCHESTER. 23 May :00 16:00. Answer ALL SIX questions The total number of marks in the paper is 90. Two hours MATH39542 UNIVERSITY OF MANCHESTER RISK THEORY 23 May 2016 14:00 16:00 Answer ALL SIX questions The total number of marks in the paper is 90. University approved calculators may be used 1 of

More information

STA258H5. Al Nosedal and Alison Weir. Winter Al Nosedal and Alison Weir STA258H5 Winter / 41

STA258H5. Al Nosedal and Alison Weir. Winter Al Nosedal and Alison Weir STA258H5 Winter / 41 STA258H5 Al Nosedal and Alison Weir Winter 2017 Al Nosedal and Alison Weir STA258H5 Winter 2017 1 / 41 NORMAL APPROXIMATION TO THE BINOMIAL DISTRIBUTION. Al Nosedal and Alison Weir STA258H5 Winter 2017

More information

Mean of a Discrete Random variable. Suppose that X is a discrete random variable whose distribution is : :

Mean of a Discrete Random variable. Suppose that X is a discrete random variable whose distribution is : : Dr. Kim s Note (December 17 th ) The values taken on by the random variable X are random, but the values follow the pattern given in the random variable table. What is a typical value of a random variable

More information

Tutorial 11: Limit Theorems. Baoxiang Wang & Yihan Zhang bxwang, April 10, 2017

Tutorial 11: Limit Theorems. Baoxiang Wang & Yihan Zhang bxwang, April 10, 2017 Tutorial 11: Limit Theorems Baoxiang Wang & Yihan Zhang bxwang, yhzhang@cse.cuhk.edu.hk April 10, 2017 1 Outline The Central Limit Theorem (CLT) Normal Approximation Based on CLT De Moivre-Laplace Approximation

More information