Qualifying Exam Solutions: Theoretical Statistics
|
|
- Gwen Booker
- 5 years ago
- Views:
Transcription
1 Qualifying Exam Solutions: Theoretical Statistics. (a) For the first sampling plan, the expectation of any statistic W (X, X,..., X n ) is a polynomial of θ of degree less than n +. Hence τ(θ) cannot have an unbiased estimator from this plan. For the second plan, let Y be the number of X i s observed. Then P (Y = y) = θ( θ) y, y =,,... ; E θ (Y ) = /θ. Therefore, Y is an unbiased estimator of θ the second sampling plan should be employed. (b) Observe X i s till you get two X i s equal to. Let Y be the the number of X i s till the first, Y be the number of X i s between the first second ones. Then Y Y are iid, E θ (Y i ) = /θ implies that E θ (Y Y ) = /θ.. (a) The joint pmf is (/N) n I X(n) NI X() where X i s are in {,,..., N, N +,...} X (r) is the rth order statistic. By the Factorization Theorem, X (n) is sufficient. We show it is complete. if k N. And P (X (n) k N) = P (X k N) P (X n k N) = ( k N )n P (X (n) = k N) = P (X (n) k N) P (X (n) k N) = ( k N )n ( k n )n. Let h(x (n) ) have E N (h(x (n) ) = o for any N =,,.... Putting N =, h() = 0; N =, h()p (X (n) = ) + h()p (X (n) = ) = h() = 0 since P (X (n) = N) > 0 for N. Proceeding this way by induction, h = 0 identically. Therefore, X n is complete. (b)an unbiased estimator of N is X. Hence the BUE of N is E(X X (n) ) = E(X X (n) ). (Suggestion from Professor Ghosh: it is already acceptable if a student stops here) Incidentally, E(X X (n) = x (n) ) = x (n) P (X = x (n) X (n) = x (n) ) + xp (X = x X n = x (n) ) x<x (n) P (X = x (n) X (n) = x (n) ) = P (X = x < x (n) X (n) = x (n) ) = N n k= ( N )k ( n k n ( ( n N )k k k= ) ( x(n) N ) ( x(n) N ) n k ) n k
2 from which E(X X (n) = x (n) ) can be found. 3. (a) It can be shown or known that T is the complete sufficient statistic of λ. If we can find f such that E[f(T )] = e aλ, then f(t ) is the UMVUE. Consider f(u) = b u. Then E(b T ) = k=0 b k (nλ)k e nλ = e nλ k! k=0 (bnλ) k k! = e nλ+bnλ. Let b = + a/n. We have E[( + a/n) T ] = e aλ. Therefore, ( + a/n) T is the UMVUE of e aλ. (b) Note that E( X ) = k λk k! e λ = e λ. k=0 So X is an unbiased estimator of e λ. Because T is complete sufficient, E( X T ) must be the UMVUE of e λ. Since UMVUE is unique, using the result in (a), we have E( X T ) = ( + n )T. 4. (a) Use π a,b = Beta(a, b) as the prior distribution. Then the posterior distribution of p is Beta(a + x, n x + b), the Bayes estimator of p is ˆp a,b = a + x n + a + b = n n + a + b ˆp + a n + a + b. When α > 0, β < 0 α + β <, we can let a = nβ/α b = n( α β)/α, both of which are positive, so that δ(x) is a Bayes estimator thus is admissible. (More sophisticated proofs are possible, for example, by using Karlin s Theorem, which leads to stronger conclusions) (b) The risk function of δ(x) is R(δ, p) =E p [(αˆp + β p) ] =E p (αˆp αp) + [β ( α)p] = α p( p) n + [β ( α)p] =[( α) α n ]p + [ α n β( α)]p + β. Setting ( α) α n = α β( α) = 0, n
3 that is, setting α = n + n, β = ( + n), R(δ, p) becomes a constant. Therefore, the corresponding estimator n δ(x) = + n X + ( + n) is minimax. As a matter of factor, this δ(x) is the Bayes estimator corresponding to the prior distribution Beta( n/, n/). 5. (a) Apply the Neyman-Pearson Lemma, the rejection region of a size α UMP test includes x, x,..., x n such that f(x,..., x n H a ) f(x,..., x n H 0 ) = (πσ ) n/ exp( σ (xi θ i0 ) ) (πσ ) n/ exp( σ x i ) = exp( xi θ σ i0 ) exp( θ σ i0 ) > k, or equivalently, x i θ i0 > k, with P ( X i θ i0 > k H 0 ) = α. Since X i s are independent of each other, the distribution of X i θ i0 is N(0, σ θ i0) under H 0. Therefore, we can choose k = σz α θ i0, the rejection region of the size α UMP test is { (x,..., x n ) x i θ i0 > σz α θ i0 }. (b). The MLE of θ i is ˆθ i = X i for i n. So the LRT statistic is λ = (πσ ) n/ exp( σ x i ) (πσ ) n/ exp( σ (xi ˆθ i ) ) = exp( x σ i ) The reject region is λ < c for some constant c such that 0 < c <, or equivalently X i > c for some constant c. Let T = Xi. Under H 0, T/σ χ n. So H 0 is rejected when T = n i= X i > σ χ n,α. Using normal approximation, the rejection region is x i > nσ + σ nz α. (c). For the test obtained in (a), under H a, X i θ i0 = n /3 X i follows a normal distribution with mean n /3 variance n /3 σ. So its power is ( σzα n /6 n /3 ) Φ = Φ(z σn /6 α σ n /6 ) as n where Φ is the cdf of the stard normal distribution. 3
4 For the test obtained in (b), under H a, Xi approximately follows a normal distribution with mean nσ + n /3 variance nσ + 4n /3. So its power is, as n, ( nσ + σ nz α (nσ + n /3 ) ) ( zα n /6 /( σ) ) Φ = Φ α. nσ + 4n /3 + n /3 σ 6. (a) Method : Let µ = (µ, µ, µ 3, µ 4 ) Then, the null hypothesis is A =. 0 H 0 : A t µ = 0. Since the columns of A are not orthogonal, we need to write A as à = Then, we can write H 0 : Ãt µ = 0. Let ȳ = (ȳ,, ȳ 4 ). Note that the rank of Ã(Ãt Ã) à t is ȳ N(µ, σ I). Then under the null hypothesis, we have 5 ȳ t Ã(Ãt Ã) à t ȳ = 4 (ȳ ȳ ȳ 3 + ȳ 4 ) + 0 (ȳ + 3ȳ 3ȳ 3 ȳ 4 ) σ 5 χ. Let where F = MSE [5 8 (ȳ ȳ ȳ 3 + ȳ 4 ) + 8 (ȳ + 3ȳ 3ȳ 3 ȳ 4 ) ] Then, under the null hypothesis MSE = 5 4 i= 5 (y i y ). j= F F,5 4
5 H 0 is rejected if F F α,,5, where F α,,5 is the upper α quantile of F,5 distribution. Method : Directly use the result that the F statistic is F = ˆµ A [A(X X) A ] Aˆµ/ MSE where X is the design matrix ˆµ = (ȳ., ȳ., ȳ 3., ȳ 4. ). In fact, 3/ / A [A(X X) A ] A = / 7/ 7/ / / 3/ (b) Method Directly use the result that that is, ˆµ H0 = ˆµ + (X X) A [A(X X) A ] Aˆµ, ˆµ,h0 ȳ. 3/ / ȳ. ˆµ,h0 ˆµ 3,h0 = ȳ. ȳ 3. + / 7/ ȳ. 5 7/ / ȳ 3. ˆµ 4,h0 ȳ 4. / 3/ ȳ 4. Other methods are also acceptable, including using lagrange multipliers, orthogonal projections, etc 7. (a) The density of X i is f(x, x ) = ρ (π)σ σ exp{ ( ρ ) [(x µ ) σ ρ(x µ )(x µ ) + (x µ ) ]}. σ σ σ The conditional density of X i given X i is f X X (x x ) = exp{ π( ρ )σ ( ρ )σ [(x µ ) ρ σ σ (x µ )] } Therefore, τ = ( ρ )σ. 5
6 Then, we have the moment estimator as ˆµ = X, ˆµ = X, ˆσ = S, ˆσ = S, ρ = S /(S S ), ˆτ = ( ˆρ )ˆσ. (b) Clearly, V (X i ) = σ 4, V (X i ) = σ 4. Note that For the other terms, we need X i X i N(ρ σ σ X i, ( ρ )σ ). E(X ix i) =E[X ie(x i X i )] =E{Xi[( ρ )σ + ρ σ X σ i]} =( ρ )σσ + 3ρ σ σ 4 σ =( + ρ )σσ. We also need Therefore, we have E(X 3 ix i ) = E[X 3 ie(x i X i )] = ρσ σ E(X 3 i) = ρσ 3 σ E(X 3 ix i ) = 3ρσ σ 3. V (X i X i ) =( + ρ )σ σ Cov(X i, X i) =ρ σ σ Cov(X i, X i X i ) =ρσ 3 σ Cov(X i, X i X i ) =ρσ σ 3 Therefore, we have the covariance matrix of (Xi, Xi, X i X i ) equal to σ 4 ρ σσ 0 Σ = ρ σσ σ ( + ρ )σσ (c) To compute the limiting distribution of ρ M. We define g(x, y, z) = z/ xy. Then, x y z = z x 3 y = z xy 3 = xy. 6
7 Let x = σ, y = σ z = ρσ σ. We have x y = ρ σ = ρ σ =. z σ σ Then, we have the limiting variance of n(ˆρ ρ) as ( ) σ ρ ρ 4 ρ σσ ρσσ 3 σ σ σ σ ρ σσ σ 4 ρσ σ 3 ρσσ 3 ρσ σ 3 ( + ρ )σσ Therefore, we have n(ˆρ ρ) approx N(0, ( ρ ) ). ρ σ ρ σ σ σ = ( ρ ). To compute the limiting distribution of ˆτ, we define g(x, y, z) = y z /x. Then, we have = z x x = y = z z x Let x = σ, y = σ z = ρσ σ. We have Then, we have ( ρ σ σ =( ρ ) σ 4. x y z = ρ σ σ = = ρσ σ ) σ ρσ 4 ρ σσ ρσσ 3 σ ρ σσ σ 4 ρσ σ 3 ρσσ 3 ρσ σ 3 ( + ρ )σσ ρ σ σ ρσ σ 7
8 Thus, we have n(ˆτ ˆτ) approx N(0, ( ρ ) σ 4 ). 8
Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER
Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS Answer any FOUR of the SIX questions.
More informationLecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.
Sufficient Statistics Lecture Notes 6 Sufficiency Data reduction in terms of a particular statistic can be thought of as a partition of the sample space X. Definition T is sufficient for θ if the conditional
More informationExercise. Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1. Exercise Estimation
Exercise Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1 Exercise S 2 = = = = n i=1 (X i x) 2 n i=1 = (X i µ + µ X ) 2 = n 1 n 1 n i=1 ((X
More informationChapter 4: Asymptotic Properties of MLE (Part 3)
Chapter 4: Asymptotic Properties of MLE (Part 3) Daniel O. Scharfstein 09/30/13 1 / 1 Breakdown of Assumptions Non-Existence of the MLE Multiple Solutions to Maximization Problem Multiple Solutions to
More informationMATH 3200 Exam 3 Dr. Syring
. Suppose n eligible voters are polled (randomly sampled) from a population of size N. The poll asks voters whether they support or do not support increasing local taxes to fund public parks. Let M be
More informationPractice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.
Practice Exercises for Midterm Exam ST 522 - Statistical Theory - II The ACTUAL exam will consists of less number of problems. 1. Suppose X i F ( ) for i = 1,..., n, where F ( ) is a strictly increasing
More informationExam 2 Spring 2015 Statistics for Applications 4/9/2015
18.443 Exam 2 Spring 2015 Statistics for Applications 4/9/2015 1. True or False (and state why). (a). The significance level of a statistical test is not equal to the probability that the null hypothesis
More informationDefinition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.
9 Point estimation 9.1 Rationale behind point estimation When sampling from a population described by a pdf f(x θ) or probability function P [X = x θ] knowledge of θ gives knowledge of the entire population.
More information1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by. Cov(X, Y ) = E(X E(X))(Y E(Y ))
Correlation & Estimation - Class 7 January 28, 2014 Debdeep Pati Association between two variables 1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by Cov(X, Y ) = E(X E(X))(Y
More informationCS340 Machine learning Bayesian model selection
CS340 Machine learning Bayesian model selection Bayesian model selection Suppose we have several models, each with potentially different numbers of parameters. Example: M0 = constant, M1 = straight line,
More informationChapter 8. Introduction to Statistical Inference
Chapter 8. Introduction to Statistical Inference Point Estimation Statistical inference is to draw some type of conclusion about one or more parameters(population characteristics). Now you know that a
More informationActuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems
Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems Spring 2005 1. Which of the following statements relate to probabilities that can be interpreted as frequencies?
More informationLikelihood Methods of Inference. Toss coin 6 times and get Heads twice.
Methods of Inference Toss coin 6 times and get Heads twice. p is probability of getting H. Probability of getting exactly 2 heads is 15p 2 (1 p) 4 This function of p, is likelihood function. Definition:
More informationPoint Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased.
Point Estimation Point Estimation Definition A point estimate of a parameter θ is a single number that can be regarded as a sensible value for θ. A point estimate is obtained by selecting a suitable statistic
More information**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:
**BEGINNING OF EXAMINATION** 1. You are given: (i) A random sample of five observations from a population is: 0.2 0.7 0.9 1.1 1.3 (ii) You use the Kolmogorov-Smirnov test for testing the null hypothesis,
More informationVersion A. Problem 1. Let X be the continuous random variable defined by the following pdf: 1 x/2 when 0 x 2, f(x) = 0 otherwise.
Math 224 Q Exam 3A Fall 217 Tues Dec 12 Version A Problem 1. Let X be the continuous random variable defined by the following pdf: { 1 x/2 when x 2, f(x) otherwise. (a) Compute the mean µ E[X]. E[X] x
More informationStatistical analysis and bootstrapping
Statistical analysis and bootstrapping p. 1/15 Statistical analysis and bootstrapping Michel Bierlaire michel.bierlaire@epfl.ch Transport and Mobility Laboratory Statistical analysis and bootstrapping
More informationmay be of interest. That is, the average difference between the estimator and the truth. Estimators with Bias(ˆθ) = 0 are called unbiased.
1 Evaluating estimators Suppose you observe data X 1,..., X n that are iid observations with distribution F θ indexed by some parameter θ. When trying to estimate θ, one may be interested in determining
More informationLecture 10: Point Estimation
Lecture 10: Point Estimation MSU-STT-351-Sum-17B (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 1 / 31 Basic Concepts of Point Estimation A point estimate of a parameter θ,
More information6. Genetics examples: Hardy-Weinberg Equilibrium
PBCB 206 (Fall 2006) Instructor: Fei Zou email: fzou@bios.unc.edu office: 3107D McGavran-Greenberg Hall Lecture 4 Topics for Lecture 4 1. Parametric models and estimating parameters from data 2. Method
More informationChapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as
Lecture 0 on BST 63: Statistical Theory I Kui Zhang, 09/9/008 Review for the previous lecture Definition: Several continuous distributions, including uniform, gamma, normal, Beta, Cauchy, double exponential
More informationChapter 8: Sampling distributions of estimators Sections
Chapter 8 continued Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample
More informationApplied Statistics I
Applied Statistics I Liang Zhang Department of Mathematics, University of Utah July 14, 2008 Liang Zhang (UofU) Applied Statistics I July 14, 2008 1 / 18 Point Estimation Liang Zhang (UofU) Applied Statistics
More informationPoint Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel
STATISTICS Lecture no. 10 Department of Econometrics FEM UO Brno office 69a, tel. 973 442029 email:jiri.neubauer@unob.cz 8. 12. 2009 Introduction Suppose that we manufacture lightbulbs and we want to state
More informationThe Bernoulli distribution
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
More informationStatistics for Business and Economics
Statistics for Business and Economics Chapter 5 Continuous Random Variables and Probability Distributions Ch. 5-1 Probability Distributions Probability Distributions Ch. 4 Discrete Continuous Ch. 5 Probability
More informationBIO5312 Biostatistics Lecture 5: Estimations
BIO5312 Biostatistics Lecture 5: Estimations Yujin Chung September 27th, 2016 Fall 2016 Yujin Chung Lec5: Estimations Fall 2016 1/34 Recap Yujin Chung Lec5: Estimations Fall 2016 2/34 Today s lecture and
More informationChapter 8: Sampling distributions of estimators Sections
Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample variance Skip: p.
More informationChapter 7 - Lecture 1 General concepts and criteria
Chapter 7 - Lecture 1 General concepts and criteria January 29th, 2010 Best estimator Mean Square error Unbiased estimators Example Unbiased estimators not unique Special case MVUE Bootstrap General Question
More informationST440/550: Applied Bayesian Analysis. (5) Multi-parameter models - Summarizing the posterior
(5) Multi-parameter models - Summarizing the posterior Models with more than one parameter Thus far we have studied single-parameter models, but most analyses have several parameters For example, consider
More informationSTAT 830 Convergence in Distribution
STAT 830 Convergence in Distribution Richard Lockhart Simon Fraser University STAT 830 Fall 2013 Richard Lockhart (Simon Fraser University) STAT 830 Convergence in Distribution STAT 830 Fall 2013 1 / 31
More informationReview for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom
Review for Final Exam 18.05 Spring 2014 Jeremy Orloff and Jonathan Bloom THANK YOU!!!! JON!! PETER!! RUTHI!! ERIKA!! ALL OF YOU!!!! Probability Counting Sets Inclusion-exclusion principle Rule of product
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Generating Random Variables and Stochastic Processes Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com
More informationCERTIFICATE IN FINANCE CQF. Certificate in Quantitative Finance Subtext t here GLOBAL STANDARD IN FINANCIAL ENGINEERING
CERTIFICATE IN FINANCE CQF Certificate in Quantitative Finance Subtext t here GLOBAL STANDARD IN FINANCIAL ENGINEERING Certificate in Quantitative Finance Probability and Statistics June 2011 1 1 PROBABILITY
More informationEstimation after Model Selection
Estimation after Model Selection Vanja M. Dukić Department of Health Studies University of Chicago E-Mail: vanja@uchicago.edu Edsel A. Peña* Department of Statistics University of South Carolina E-Mail:
More informationBack to estimators...
Back to estimators... So far, we have: Identified estimators for common parameters Discussed the sampling distributions of estimators Introduced ways to judge the goodness of an estimator (bias, MSE, etc.)
More informationCentral Limit Theorem, Joint Distributions Spring 2018
Central Limit Theorem, Joint Distributions 18.5 Spring 218.5.4.3.2.1-4 -3-2 -1 1 2 3 4 Exam next Wednesday Exam 1 on Wednesday March 7, regular room and time. Designed for 1 hour. You will have the full
More informationMAS6012. MAS Turn Over SCHOOL OF MATHEMATICS AND STATISTICS. Sampling, Design, Medical Statistics
t r r r t s t SCHOOL OF MATHEMATICS AND STATISTICS Sampling, Design, Medical Statistics Spring Semester 206 207 3 hours t s 2 r t t t t r t t r s t rs t2 r t s s rs r t r t 2 r t st s rs q st s r rt r
More informationStatistical estimation
Statistical estimation Statistical modelling: theory and practice Gilles Guillot gigu@dtu.dk September 3, 2013 Gilles Guillot (gigu@dtu.dk) Estimation September 3, 2013 1 / 27 1 Introductory example 2
More informationNormal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.
Lecture 21,22, 23 Text: A Course in Probability by Weiss 8.5 STAT 225 Introduction to Probability Models March 31, 2014 Standard Sums of Whitney Huang Purdue University 21,22, 23.1 Agenda 1 2 Standard
More informationAMH4 - ADVANCED OPTION PRICING. Contents
AMH4 - ADVANCED OPTION PRICING ANDREW TULLOCH Contents 1. Theory of Option Pricing 2 2. Black-Scholes PDE Method 4 3. Martingale method 4 4. Monte Carlo methods 5 4.1. Method of antithetic variances 5
More information3 ˆθ B = X 1 + X 2 + X 3. 7 a) Find the Bias, Variance and MSE of each estimator. Which estimator is the best according
STAT 345 Spring 2018 Homework 9 - Point Estimation Name: Please adhere to the homework rules as given in the Syllabus. 1. Mean Squared Error. Suppose that X 1, X 2 and X 3 are independent random variables
More information2 of PU_2015_375 Which of the following measures is more flexible when compared to other measures?
PU M Sc Statistics 1 of 100 194 PU_2015_375 The population census period in India is for every:- quarterly Quinqennial year biannual Decennial year 2 of 100 105 PU_2015_375 Which of the following measures
More informationChapter 7: Estimation Sections
1 / 40 Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: Chapter 7 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods:
More informationThe Normal Distribution
The Normal Distribution The normal distribution plays a central role in probability theory and in statistics. It is often used as a model for the distribution of continuous random variables. Like all models,
More informationLecture 5: Fundamentals of Statistical Analysis and Distributions Derived from Normal Distributions
Lecture 5: Fundamentals of Statistical Analysis and Distributions Derived from Normal Distributions ELE 525: Random Processes in Information Systems Hisashi Kobayashi Department of Electrical Engineering
More information12 The Bootstrap and why it works
12 he Bootstrap and why it works For a review of many applications of bootstrap see Efron and ibshirani (1994). For the theory behind the bootstrap see the books by Hall (1992), van der Waart (2000), Lahiri
More informationStatistical Analysis of Data from the Stock Markets. UiO-STK4510 Autumn 2015
Statistical Analysis of Data from the Stock Markets UiO-STK4510 Autumn 2015 Sampling Conventions We observe the price process S of some stock (or stock index) at times ft i g i=0,...,n, we denote it by
More informationCentral Limit Thm, Normal Approximations
Central Limit Thm, Normal Approximations Engineering Statistics Section 5.4 Josh Engwer TTU 23 March 2016 Josh Engwer (TTU) Central Limit Thm, Normal Approximations 23 March 2016 1 / 26 PART I PART I:
More informationNon-informative Priors Multiparameter Models
Non-informative Priors Multiparameter Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Prior Types Informative vs Non-informative There has been a desire for a prior distributions that
More informationChapter 2. Random variables. 2.3 Expectation
Random processes - Chapter 2. Random variables 1 Random processes Chapter 2. Random variables 2.3 Expectation 2.3 Expectation Random processes - Chapter 2. Random variables 2 Among the parameters representing
More informationPopulations and Samples Bios 662
Populations and Samples Bios 662 Michael G. Hudgens, Ph.D. mhudgens@bios.unc.edu http://www.bios.unc.edu/ mhudgens 2008-08-22 16:29 BIOS 662 1 Populations and Samples Random Variables Random sample: result
More informationThe Vasicek Distribution
The Vasicek Distribution Dirk Tasche Lloyds TSB Bank Corporate Markets Rating Systems dirk.tasche@gmx.net Bristol / London, August 2008 The opinions expressed in this presentation are those of the author
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Simulation Efficiency and an Introduction to Variance Reduction Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University
More informationPractice Exam 1. Loss Amount Number of Losses
Practice Exam 1 1. You are given the following data on loss sizes: An ogive is used as a model for loss sizes. Determine the fitted median. Loss Amount Number of Losses 0 1000 5 1000 5000 4 5000 10000
More informationHomework Problems Stat 479
Chapter 10 91. * A random sample, X1, X2,, Xn, is drawn from a distribution with a mean of 2/3 and a variance of 1/18. ˆ = (X1 + X2 + + Xn)/(n-1) is the estimator of the distribution mean θ. Find MSE(
More information4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.
4-1 Chapter 4 Commonly Used Distributions 2014 by The Companies, Inc. All rights reserved. Section 4.1: The Bernoulli Distribution 4-2 We use the Bernoulli distribution when we have an experiment which
More informationChapter 5. Statistical inference for Parametric Models
Chapter 5. Statistical inference for Parametric Models Outline Overview Parameter estimation Method of moments How good are method of moments estimates? Interval estimation Statistical Inference for Parametric
More informationComputer Statistics with R
MAREK GAGOLEWSKI KONSTANCJA BOBECKA-WESO LOWSKA PRZEMYS LAW GRZEGORZEWSKI Computer Statistics with R 5. Point Estimation Faculty of Mathematics and Information Science Warsaw University of Technology []
More informationChapter 6: Point Estimation
Chapter 6: Point Estimation Professor Sharabati Purdue University March 10, 2014 Professor Sharabati (Purdue University) Point Estimation Spring 2014 1 / 37 Chapter Overview Point estimator and point estimate
More informationEcon 250 Fall Due at November 16. Assignment 2: Binomial Distribution, Continuous Random Variables and Sampling
Econ 250 Fall 2010 Due at November 16 Assignment 2: Binomial Distribution, Continuous Random Variables and Sampling 1. Suppose a firm wishes to raise funds and there are a large number of independent financial
More informationROM SIMULATION Exact Moment Simulation using Random Orthogonal Matrices
ROM SIMULATION Exact Moment Simulation using Random Orthogonal Matrices Bachelier Finance Society Meeting Toronto 2010 Henley Business School at Reading Contact Author : d.ledermann@icmacentre.ac.uk Alexander
More information1. Statistical problems - a) Distribution is known. b) Distribution is unknown.
Probability February 5, 2013 Debdeep Pati Estimation 1. Statistical problems - a) Distribution is known. b) Distribution is unknown. 2. When Distribution is known, then we can have either i) Parameters
More informationCommon one-parameter models
Common one-parameter models In this section we will explore common one-parameter models, including: 1. Binomial data with beta prior on the probability 2. Poisson data with gamma prior on the rate 3. Gaussian
More informationExam M Fall 2005 PRELIMINARY ANSWER KEY
Exam M Fall 005 PRELIMINARY ANSWER KEY Question # Answer Question # Answer 1 C 1 E C B 3 C 3 E 4 D 4 E 5 C 5 C 6 B 6 E 7 A 7 E 8 D 8 D 9 B 9 A 10 A 30 D 11 A 31 A 1 A 3 A 13 D 33 B 14 C 34 C 15 A 35 A
More informationIEOR 165 Lecture 1 Probability Review
IEOR 165 Lecture 1 Probability Review 1 Definitions in Probability and Their Consequences 1.1 Defining Probability A probability space (Ω, F, P) consists of three elements: A sample space Ω is the set
More informationHedging and Regression. Hedging and Regression
Returns The discrete return on a stock is the percentage change: S i S i 1 S i 1. The index i can represent days, weeks, hours etc. What happens if we compute returns at infinitesimally short intervals
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Further Variance Reduction Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Outline
More informationStatistical Tables Compiled by Alan J. Terry
Statistical Tables Compiled by Alan J. Terry School of Science and Sport University of the West of Scotland Paisley, Scotland Contents Table 1: Cumulative binomial probabilities Page 1 Table 2: Cumulative
More information1. You are given the following information about a stationary AR(2) model:
Fall 2003 Society of Actuaries **BEGINNING OF EXAMINATION** 1. You are given the following information about a stationary AR(2) model: (i) ρ 1 = 05. (ii) ρ 2 = 01. Determine φ 2. (A) 0.2 (B) 0.1 (C) 0.4
More informationcontinuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence
continuous rv Let X be a continuous rv. Then a probability distribution or probability density function (pdf) of X is a function f(x) such that for any two numbers a and b with a b, P(a X b) = b a f (x)dx.
More information2.1 Probability, stochastic variables and distribution functions
Chapter 2 Probability and statistics 2.1 Probability, stochastic variables and distribution functions The defining characteristic of a stochastic experiment E is that it produces different outcomes under
More information1 Bayesian Bias Correction Model
1 Bayesian Bias Correction Model Assuming that n iid samples {X 1,...,X n }, were collected from a normal population with mean µ and variance σ 2. The model likelihood has the form, P( X µ, σ 2, T n >
More informationInterval estimation. September 29, Outline Basic ideas Sampling variation and CLT Interval estimation using X More general problems
Interval estimation September 29, 2017 STAT 151 Class 7 Slide 1 Outline of Topics 1 Basic ideas 2 Sampling variation and CLT 3 Interval estimation using X 4 More general problems STAT 151 Class 7 Slide
More informationCalibration of Interest Rates
WDS'12 Proceedings of Contributed Papers, Part I, 25 30, 2012. ISBN 978-80-7378-224-5 MATFYZPRESS Calibration of Interest Rates J. Černý Charles University, Faculty of Mathematics and Physics, Prague,
More informationSTAT/MATH 395 PROBABILITY II
STAT/MATH 395 PROBABILITY II Distribution of Random Samples & Limit Theorems Néhémy Lim University of Washington Winter 2017 Outline Distribution of i.i.d. Samples Convergence of random variables The Laws
More informationBayesian Linear Model: Gory Details
Bayesian Linear Model: Gory Details Pubh7440 Notes By Sudipto Banerjee Let y y i ] n i be an n vector of independent observations on a dependent variable (or response) from n experimental units. Associated
More informationSection 7.1: Continuous Random Variables
Section 71: Continuous Random Variables Discrete-Event Simulation: A First Course c 2006 Pearson Ed, Inc 0-13-142917-5 Discrete-Event Simulation: A First Course Section 71: Continuous Random Variables
More informationTutorial 11: Limit Theorems. Baoxiang Wang & Yihan Zhang bxwang, April 10, 2017
Tutorial 11: Limit Theorems Baoxiang Wang & Yihan Zhang bxwang, yhzhang@cse.cuhk.edu.hk April 10, 2017 1 Outline The Central Limit Theorem (CLT) Normal Approximation Based on CLT De Moivre-Laplace Approximation
More informationA potentially useful approach to model nonlinearities in time series is to assume different behavior (structural break) in different subsamples
1.3 Regime switching models A potentially useful approach to model nonlinearities in time series is to assume different behavior (structural break) in different subsamples (or regimes). If the dates, the
More informationχ 2 distributions and confidence intervals for population variance
χ 2 distributions and confidence intervals for population variance Let Z be a standard Normal random variable, i.e., Z N(0, 1). Define Y = Z 2. Y is a non-negative random variable. Its distribution is
More informationSTA 103: Final Exam. Print clearly on this exam. Only correct solutions that can be read will be given credit.
STA 103: Final Exam June 26, 2008 Name: } {{ } by writing my name i swear by the honor code Read all of the following information before starting the exam: Print clearly on this exam. Only correct solutions
More informationEE641 Digital Image Processing II: Purdue University VISE - October 29,
EE64 Digital Image Processing II: Purdue University VISE - October 9, 004 The EM Algorithm. Suffient Statistics and Exponential Distributions Let p(y θ) be a family of density functions parameterized by
More informationApplications of Good s Generalized Diversity Index. A. J. Baczkowski Department of Statistics, University of Leeds Leeds LS2 9JT, UK
Applications of Good s Generalized Diversity Index A. J. Baczkowski Department of Statistics, University of Leeds Leeds LS2 9JT, UK Internal Report STAT 98/11 September 1998 Applications of Good s Generalized
More informationEVA Tutorial #1 BLOCK MAXIMA APPROACH IN HYDROLOGIC/CLIMATE APPLICATIONS. Rick Katz
1 EVA Tutorial #1 BLOCK MAXIMA APPROACH IN HYDROLOGIC/CLIMATE APPLICATIONS Rick Katz Institute for Mathematics Applied to Geosciences National Center for Atmospheric Research Boulder, CO USA email: rwk@ucar.edu
More informationINSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION
INSTITUTE AND FACULTY OF ACTUARIES Curriculum 2019 SPECIMEN EXAMINATION Subject CS1A Actuarial Statistics Time allowed: Three hours and fifteen minutes INSTRUCTIONS TO THE CANDIDATE 1. Enter all the candidate
More informationChapter 7: SAMPLING DISTRIBUTIONS & POINT ESTIMATION OF PARAMETERS
Chapter 7: SAMPLING DISTRIBUTIONS & POINT ESTIMATION OF PARAMETERS Part 1: Introduction Sampling Distributions & the Central Limit Theorem Point Estimation & Estimators Sections 7-1 to 7-2 Sample data
More informationHomework Problems Stat 479
Chapter 2 1. Model 1 is a uniform distribution from 0 to 100. Determine the table entries for a generalized uniform distribution covering the range from a to b where a < b. 2. Let X be a discrete random
More informationOptimizing Portfolios
Optimizing Portfolios An Undergraduate Introduction to Financial Mathematics J. Robert Buchanan 2010 Introduction Investors may wish to adjust the allocation of financial resources including a mixture
More informationPosterior Inference. , where should we start? Consider the following computational procedure: 1. draw samples. 2. convert. 3. compute properties
Posterior Inference Example. Consider a binomial model where we have a posterior distribution for the probability term, θ. Suppose we want to make inferences about the log-odds γ = log ( θ 1 θ), where
More informationSTA258 Analysis of Variance
STA258 Analysis of Variance Al Nosedal. University of Toronto. Winter 2017 The Data Matrix The following table shows last year s sales data for a small business. The sample is put into a matrix format
More informationWeek 1 Quantitative Analysis of Financial Markets Basic Statistics A
Week 1 Quantitative Analysis of Financial Markets Basic Statistics A Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 October
More informationLecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial
Lecture 23 STAT 225 Introduction to Probability Models April 4, 2014 approximation Whitney Huang Purdue University 23.1 Agenda 1 approximation 2 approximation 23.2 Characteristics of the random variable:
More informationNormal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is
Normal Distribution Normal Distribution Definition A continuous rv X is said to have a normal distribution with parameter µ and σ (µ and σ 2 ), where < µ < and σ > 0, if the pdf of X is f (x; µ, σ) = 1
More informationComparing the Means of. Two Log-Normal Distributions: A Likelihood Approach
Journal of Statistical and Econometric Methods, vol.3, no.1, 014, 137-15 ISSN: 179-660 (print), 179-6939 (online) Scienpress Ltd, 014 Comparing the Means of Two Log-Normal Distributions: A Likelihood Approach
More informationWhat was in the last lecture?
What was in the last lecture? Normal distribution A continuous rv with bell-shaped density curve The pdf is given by f(x) = 1 2πσ e (x µ)2 2σ 2, < x < If X N(µ, σ 2 ), E(X) = µ and V (X) = σ 2 Standard
More informationChapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29
Chapter 5 Univariate time-series analysis () Chapter 5 Univariate time-series analysis 1 / 29 Time-Series Time-series is a sequence fx 1, x 2,..., x T g or fx t g, t = 1,..., T, where t is an index denoting
More informationSTAT 425: Introduction to Bayesian Analysis
STAT 45: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 018 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 1) Fall 018 1 / 37 Lectures 9-11: Multi-parameter
More informationLecture Notes in Financial Econometrics (MBF, MSc course at UNISG)
Contents Lecture Notes in Financial Econometrics (MBF, MSc course at UNISG) Paul Söderlind 1 24 June 25 1 Review of Statistics 3 1.1 Random Variables and Distributions.................. 3 1.2 Hypothesis
More informationMS-E2114 Investment Science Lecture 5: Mean-variance portfolio theory
MS-E2114 Investment Science Lecture 5: Mean-variance portfolio theory A. Salo, T. Seeve Systems Analysis Laboratory Department of System Analysis and Mathematics Aalto University, School of Science Overview
More information