ECSE B Assignment 5 Solutions Fall (a) Using whichever of the Markov or the Chebyshev inequalities is applicable, estimate

Similar documents
Version A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

Chapter 2: Random Variables (Cont d)

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals

Homework Assignments

Probability. An intro for calculus students P= Figure 1: A normal integral

MidTerm 1) Find the following (round off to one decimal place):

STAT/MATH 395 PROBABILITY II

Random Variables and Probability Functions

Statistical Tables Compiled by Alan J. Terry

Chapter 2. Random variables. 2.3 Expectation

Discrete Probability Distribution

Tutorial 6. Sampling Distribution. ENGG2450A Tutors. 27 February The Chinese University of Hong Kong 1/6

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

Section Distributions of Random Variables

Chapter 4 Continuous Random Variables and Probability Distributions

Strategies for Improving the Efficiency of Monte-Carlo Methods

Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems

Chapter 5 Discrete Probability Distributions. Random Variables Discrete Probability Distributions Expected Value and Variance

4.2 Probability Distributions

Expected value and variance

ECE 295: Lecture 03 Estimation and Confidence Interval

Chapter 4 Continuous Random Variables and Probability Distributions

Statistics, Measures of Central Tendency I

Statistics (This summary is for chapters 17, 28, 29 and section G of chapter 19)

Chapter 2: Descriptive Statistics. Mean (Arithmetic Mean): Found by adding the data values and dividing the total by the number of data.

Theoretical Statistics. Lecture 3. Peter Bartlett

Section Distributions of Random Variables

Random Variables Handout. Xavier Vilà

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

2011 Pearson Education, Inc

Course information FN3142 Quantitative finance

Populations and Samples Bios 662

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

VI. Continuous Probability Distributions

Discrete Random Variables

IEOR 165 Lecture 1 Probability Review

MATH 3200 Exam 3 Dr. Syring

Central Limit Theorem, Joint Distributions Spring 2018

Advanced Financial Economics Homework 2 Due on April 14th before class

STATISTICS and PROBABILITY

Introduction to Probability

PhD Qualifier Examination

Statistics for Business and Economics

Outline Brownian Process Continuity of Sample Paths Differentiability of Sample Paths Simulating Sample Paths Hitting times and Maximum

Business Statistics 41000: Probability 3

Binomial Random Variables. Binomial Random Variables

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

Chapter 4: Asymptotic Properties of MLE (Part 3)

Data Analysis and Statistical Methods Statistics 651

MTH6154 Financial Mathematics I Stochastic Interest Rates

The rth moment of a real-valued random variable X with density f(x) is. x r f(x) dx

The Bernoulli distribution

Drunken Birds, Brownian Motion, and Other Random Fun

STAT 830 Convergence in Distribution

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

PROBABILITY AND STATISTICS

Lecture 2. Probability Distributions Theophanis Tsandilas

Asymptotic results discrete time martingales and stochastic algorithms

MATH MW Elementary Probability Course Notes Part IV: Binomial/Normal distributions Mean and Variance

Probability and Random Variables A FINANCIAL TIMES COMPANY

II - Probability. Counting Techniques. three rules of counting. 1multiplication rules. 2permutations. 3combinations

Lecture 12. Some Useful Continuous Distributions. The most important continuous probability distribution in entire field of statistics.

Section 7.1: Continuous Random Variables

Week 1 Quantitative Analysis of Financial Markets Basic Statistics A

EE641 Digital Image Processing II: Purdue University VISE - October 29,

The Conservative Expected Value: A New Measure with Motivation from Stock Trading via Feedback

Lecture 8. The Binomial Distribution. Binomial Distribution. Binomial Distribution. Probability Distributions: Normal and Binomial

3. Continuous Probability Distributions

Chapter 3 Discrete Random Variables and Probability Distributions

Class 12. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

STA 103: Final Exam. Print clearly on this exam. Only correct solutions that can be read will be given credit.

. 13. The maximum error (margin of error) of the estimate for μ (based on known σ) is:

E509A: Principle of Biostatistics. GY Zou

Econ 6900: Statistical Problems. Instructor: Yogesh Uppal

STAT Chapter 4/6: Random Variables and Probability Distributions

STOR Lecture 15. Jointly distributed Random Variables - III

ECE 340 Probabilistic Methods in Engineering M/W 3-4:15. Lecture 10: Continuous RV Families. Prof. Vince Calhoun

A continuous random variable is one that can theoretically take on any value on some line interval. We use f ( x)

Math489/889 Stochastic Processes and Advanced Mathematical Finance Homework 5

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.

Chapter 8 Homework Solutions Compiled by Joe Kahlig. speed(x) freq 25 x < x < x < x < x < x < 55 5

A probability distribution shows the possible outcomes of an experiment and the probability of each of these outcomes.

Chapter 3 - Lecture 3 Expected Values of Discrete Random Va

MATH 264 Problem Homework I

Normal distribution. We say that a random variable X follows the normal distribution if the probability density function of X is given by

ECO220Y Continuous Probability Distributions: Normal Readings: Chapter 9, section 9.10

L04: Homework Answer Key

Continuous Probability Distributions & Normal Distribution

Two Hours. Mathematical formula books and statistical tables are to be provided THE UNIVERSITY OF MANCHESTER. 22 January :00 16:00

Have you ever wondered whether it would be worth it to buy a lottery ticket every week, or pondered on questions such as If I were offered a choice

Probability without Measure!

(Practice Version) Midterm Exam 1

MA : Introductory Probability

Chapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29

Chapter 7 1. Random Variables

Exercise. Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1. Exercise Estimation

Transcription:

ECSE 304-305B Assignment 5 Solutions Fall 2008 Question 5.1 A positive scalar random variable X with a density is such that EX = µ <, EX 2 =. (a) Using whichever of the Markov or the Chebyshev inequalities is applicable, estimate the probability P (X 2 α 2 ), α > 0. (b) Is it the case that the distribution function F X (x) = P (X x), x IR, of X is necessarily a continuous function of x? Briefly explain your answer. (c) Does there exist a real number γ > 0 such that e γ = P (X γ) and if so why? Solution 5.1 (a) EX 2 = One cannot use Chebyshev Inequality; however: X 0 {X 2 α 2 } = {X α} P((X 2 α 2 )=P((X α) EX α = µ α (By the Markov Inequality) (b) Since X has density, call it f(x), F(x)= x 0 f(s)ds F(x) is continuuous because F(x+dx)-F(x)= x+dx x f(s)ds 0 as dx 0 (c) As γ goes from 0 to, e γ 0 is continuous and (strictly) monotonically decreasing from 1 to 0, but P (X γ) = F (γ) is continuous and (strictly) monotonically increasing from 0 to 1, and so there exists a unique point of intersection. (Here uniqueness follows from the fact that there is at least one strictly monotonic function.) 1

Question 5.2 (SG p 182)m What is the expected number, the variance and the standard deviation ( = the square root of the variance) of the number of spades in a poker hand? (A poker hand is a set of five cards that are randomly selected (i.e. the EPP applies) from an ordinary deck of 52 cards.) Give your answer to three decimal places. Solution 5.2: Define a random variable X as the number of spades in a poker hand. And we know So, P (X = n) = P (n are spades and 5 n are not spades) E(X) = = C13 n C 39 = 5 n C5 32, 0 n 5. 5 np (X = n) n=0 5 n=0 n C13 n C5 n 39 C5 52 1.248, 5 V ar(x) = [n E(X)] 2 P (X = n) 0.866, n=0 σ = V ar(x) 0.931. 2

Question 5.3 Let X be a continuous random variable (i.e. a not a discrete random random variable, hence it takes an uncountable set of values) whose probability distribution function has the density f(x) = 6x(1 x), 0 < x < 1. (1) What is the probability that X takes a value within two standard deviations of the mean? (That is to say, what is the probability that X is less than or equal to the mean plus two standard deviations but greater than or equal to the mean minus two standard deviations?) Solution 5.3: We have E(X) = V ar(x) = 1 0 1 0 xf(x)dx = 1 2 (x E(x)) 2 f(x)dx = 1 20, σ = V ar(x) = 1 20. So, P [E(X) 2σ X E(X) + 2σ] = E(X)+2σ E(X) 2σ f(x)dx 0.984. 3

Question 5.4 Each of an i.i.d. sequence of random variables X = {X n ; n Z 1 }, where Z 1 = {1, 2,...}, has the probability density (2π9) 1/2 exp( x2 + x 1 ), x IR. 18 3 2 (i) Find the mean µ = EX n, and the variance σ 2 of X n ; n Z 1. (Hint: it is not necessary to use integration, just use the standard form of the Gaussian density.) (ii) Use Chebychev s inequality to find an upper bound on the probability that any one of these random variables takes a value greater than or equal to 3 units away from its mean. Solution 5.4 : (i) The probability density of Gaussian distribution is equal to Here we have So µ = 3 and σ 2 = 9 f(x) = { } 1 exp (x µ)2. 2πσ 2 2σ 2 f(x) = 1 } (x 3)2 exp {. 2π9 2 9 (ii) Using Chebychev s inequality, P ( X µ k) σ2 k 2, when k = 3, the desired upper bound on the probability that anyone of these i.i.d. random variables takes a value greater than or equal to 3 units away from its mean is 1 P [(X n µ) < 3] n = 1 {1 P [(X n µ) 3]} n 1 ( 1 2 )n, 4

since for a random variable X with Gaussian distribution P [(X µ) 3] = P [(X µ) 3], and P [(X n µ) 3] 1 2. Solution if Q4(ii) had used 3σ instead of 3 units. Using Chebychev s inequality, P ( X µ kσ) 1 k 2, when k = 3, the desired upper bound on the probability that anyone of these i.i.d. random variables takes a value greater than or equal to 3σ away from its mean is 1 P [(X n µ) < 3σ] n = 1 {1 P [(X n µ) 3σ]} n 1 ( 17 18 )n, since for a random variable X with Gaussian distribution P [(X µ) 3σ] = P [(X µ) 3σ], and P [(X n µ) 3σ] 1 2 9 2 = 1 18. Question 5.5 One is attempting to estimate the distribution function F X (x) of the random variable X at x = 1, i.e. to estimate F X (1), on the basis of n independent and identically distributed observations {X 1, X 2,..., X n } of the random variable where each X i F X. This is done by computing the relative frequency r A (n) = 1 n Σn I A (X i ), where A is the event {X 1}. 5

Evaluate a Chebychev upper bound to P ( r A (n) P (A) ɛ.n 1 2 ) in terms of F X (1). Solution 5.5: A Chebychev upper bound to P ( r A (n) P (A) ɛn 1/2 ) is give by P ( r A (n) P (A) ɛn 1/2 ) σ 2 /(ɛn 1/2 ) 2, where σ 2 = E[rA 2 (n)] - E2 [r A (n)]. E[r A (n)] is given by [ ] n E[r A (n)] = E 1/n I A (X i ) = 1/n n E [I A (X i )] = P (A), and [ ( n ) ( n )] σ 2 = E (1/n 2 ) I A (X i ) I A (X i ) P 2 (A) { [ n ] [ = (1/n 2 ) E IA(X 2 i ) + E,j=1,i j = (1/n 2 ) {n P (A) + n(n 1) P 2 (A) } P 2 (A) ]} n I A (X i )I A (X j ) P 2 (A) = (1/n) P (A) (1/n) P 2 (A). Since F X (1) = P (X 1) = P (A), we have σ 2 = F X (1)/n (1 F X (1)). Hence, P ( r A (n) P (A) ɛn 1/2 ) σ 2 /(ɛn 1/2 ) 2 = F X (1)(1 F X (1))/ɛ 2. 6