Hardy Weinberg Model- 6 Genotypes
|
|
- Ariel Sparks
- 5 years ago
- Views:
Transcription
1 Hardy Weinberg Model- 6 Genotypes Silvelyn Zwanzig Hardy -Weinberg with six genotypes. In a large population of plants (Mimulus guttatus there are possible alleles S, I, F at one locus resulting in six genotypes labeled SS, II, FF, SI, SF. Let θ, θ 2, θ 3 denote the probabilities of S,I,F with θ + θ 2 + θ 3. The Hardy Weinberg model specifies that the six genotypes have probabilities: Genotype Genotype SS II F F SI SF IF ( Probability θ 2 θ2 2 θ3 2 2θ θ 2 2θ θ 3 2θ 2 θ 3 Consider an i.i.d sample (X,.., X n according to the distribution (.. Belongs the distribution ( to a strictly(! k-parametric exponential family? Theorem 3.2 implies, that it is enough to consider one observation of X from (. It holds with P (X x θ 2I (x θ 2I 2(x 2 θ 2I 3(x 3 (2θ θ 2 I 4(x (2θ θ 3 I 5(x (2θ 2 θ 3 I 6(x θ 2I (x+i 4 (x+i 5 (x θ 2I 2(x+I 4 (x+i 6 (x 2 θ 2I 3(x+I 5 (x+i 6 (x 3 2 I 4(x+I 5 (x+i 6 (x I k (x { for x k 0 for x k for k,..., 6 (2
2 Introduce I S (x 2I (x + I 4 (x + I 5 (x I I (x 2I 2 (x + I 4 (x + I 6 (x 2 for x for x 4, 5 0 else 2 for x 2 for x 4, 6 0 else 2 for x 3 I F (x 2I 3 (x + I 5 (x + I 6 (x for x 5, 6 0 else and reformulate the probability function, P (X x θ I S(x θ I I(x 2 θ I F (x 3 2 I 4(x+I 5 (x+i 6 (x. (3 Note thus 6 I k (x for all x, k I S (x + I I (x + I F (x 2I (x + I 4 (x + I 5 (x + 2I 2 (x + I 4 (x + I 6 (x + 2I 3 (x + I 5 (x + I 6 (x 6 2 I k (x 2. k Furthermore θ 2 + θ2 2 + θ (θ θ 2 + θ 2 θ 3 + θ 3 θ 2. It holds θ 2 + θ2 2 + θ (θ θ 2 + θ θ 2 + θ θ 2 (θ + θ 2 + θ 3 2. We obtain θ + θ 2 + θ 3 2
3 Rewrite the probability function P (X x θ I S(x θ I I(x 2 ( θ θ 2 2 I S(x I I (x 2 I 4(x+I 5 (x+i 6 (x as 2-parametric exponential family P (X x h(xa(θ exp (ς (θt (x + ς 2 (θt 2 (x, θ (θ, θ 2 (4 with h(x 2 I 4(x+I 5 (x+i 6 (x, A(θ exp (2 ln( θ θ 2 (5 and ( T (x I S (x, ς (θ ln ( T 2 (x I I (x, ς 2 (θ ln θ θ θ 2 θ 2 θ θ 2 (6. (7 Properties, using (2 T (x 2I (x + I 4 (x + I 5 (x T 2 (x 2I 2 (x + I 5 (x + I 6 (x We know with p k P (X k that { for k k I k (XI k2 (X 2 k 0 else, { pk for k EI k (X p k, EI k (XI k2 (X k 2 k 0 else (8 Hence ET 2θ 2 + 2θ θ 2 + 2θ θ 3 2θ (θ + θ 2 + θ 3 2θ (9 ET 2 2θ θ θ 2 + 2θ 2 θ 3 2θ 2 (θ 2 + θ + θ 3 2θ 2 3
4 and ET 2 E (2I (x + I 4 (x + I 5 (x 2 (0 4EI (x + EI 4 (x + EI 5 (x 2EI (x + ET 2θ 2 + 2θ 2θ ( + θ Cov(T, T 2 E (T 2θ (T 2 2θ 2 E (T T 2 4θ θ 2 E (T T 2 E (2I (x + I 4 (x + I 5 (x (2I 2 (x + I 4 (x + I 6 (x ( EI 4 (x 2θ θ 2 Thus Cov(T, T 2 2θ θ 2 Summarizing, the distribution of (T, T 2, T 3 is multinomial with. multinomial(2, θ, θ 2, θ 3 where T 3 (x 2I 3 (x + I 5 (x + I 6 (x Solution: The i.i.d sample (X,.., X n according to the distribution ( belongs to a exponential family. 2. Derive k and the minimal sufficient statistics. Compare: (4 - (7 Apply Theorem 3.2 and Theorem 4.0. Solution: k 2 and T (X (T (X, T 2 (X, n T (X T (X i 2n + n 4 + n 5 T 2 (X i n T 2 (X i 2n 2 + n 4 + n 6 i 4
5 where n k # {x i k, i,..., n} 3. Are the minimal sufficient statistics complete? Solution: Yes. Because: Apply Theorem 5.5 and the natural parametrization with (6 and (7. The natural parameter space is defined by { } Z (ς, ς 2 : exp (ς T (x + ς 2 T 2 (x h(xdv < Here exp (ς T (x + ς 2 T 2 (x h(xdv thus Z R 2. x,2,3,4,5,6 exp (ς T (x + ς 2 T 2 (x h(x < 4. Derive the MLE for all unknown parameters: θ,mle, θ 2,MLE, θ 3,MLE. The log likelihood function is l(θ T (x ln(θ + T 2 (x ln(θ 2 + T 3 (x ln( θ θ 2 + const with T 3 (x 2n 3 + n 5 + n 6. l(θ T (x T 3 (x! 0, (2 θ θ θ θ 2 l(θ T 2 (x T 3 (x! 0 θ 2 θ 2 θ θ 2 Using θ + θ 2 + θ 3 θ 3 T (x θ T 2 (x 0 (3 θ 3 T 2 (x θ 2 T 3 (x 0 and adding both equations gives θ 3 (T + T 2 ( θ 3 T 3 0. Because T (x + T 2 (x + T 3 (x 2n we obtain θ 3,MLE T 3 2n 2n 3 + n 5 + n 6. 2n Analogously θ,mle T 2n 2+n 4 +n 5, θ2,mle T 2 2n 2+n 4 +n 6 2n 2n 2n 2n 5
6 5. Give the MLE for the probability of SI. We have p SI 2θ θ 2 Thus (Def. 5.2 p SI MLE 2 θ,mle θ2,mle T T 2 2n ( 2 6. Are the MLE θ,mle, θ 2,MLE, θ 3,MLE optimal? In which sense? (Give the definition! The MLE are functions of sufficient and complete statistics. We have to check the unbiasedness. Using (9 E θ,mle E T 2n ET 2n n2θ 2n θ. By symmetry θ 2,MLE, θ 3,MLE are also unbiased. Thus by Theorem of Lehman Scheffe the maximum likelihood estimators are best unbiased estimators. (Def These estimators have minimal variance in the class of all unbiased estimators. 7. Derive an alternative moment estimator for probability of SI, basing on the number of plants belonging to the genotype SI only. Number of plants belonging to the genotype SI : n i I 4 (x i n 4 Note the probability of SI equals p 4. It holds by (8 En 4 n EI 4 (x i np 4 i Thus p SI n 4 n is an reasonable unbiased estimator. 8. Compare the estimators in p SI MLE and p SI. Which is better? Why? First we have to study the unbiasedness of p SI MLE. E p SI MLE E T (x T 2 (x 2n 2 6
7 Because of (8 and ( it holds ( n E (T (x T 2 (x E T (x i i n T 2 (x j j ( ( E T (x i T 2 (x i + E T (x i T 2 (x j ne (T T 2 + (n 2 net ET 2 n2θ θ 2 + (n 2 n2θ 2θ 2 2θ θ 2 (2n 2 n. Thus E p SI MLE E T (x T 2 (x 2θ 2n 2 θ 2 ( 2n. The maximum likelihood estimator is not unbiased! Propose an bias corrected estimator: p SI T (x T 2 (x 2n 2 n This estimator p SI is BUE, because of the Theorem of Lehman Scheffe. 9. Calculate the Fisher information matrix for (. We have an i.i.d. sample. Consider only one observation. The score function is given by (compare (2 V (θ ( θ l(θ θ 2 l(θ ij ( θ T θ 3 T 3 θ 2 T 2 θ 3 T 3 ( Check: EV (θ 0. We have E θ T θ 3 T Fisher information matrix: I X (θ Cov (V (θ Using ( we calculate i j 7
8 and ( E T 2 T 3 θ θ 3 E (T θ ET T 3 + ET 2 θ θ 3 θ θ θ 2 ( + θ 2 2θ θ θ θ θ 3 θ3 2 3 ( + θ 3 ( 2 + θ θ 3 ( E T ( T 3 T 2 T 3 θ θ 3 θ 2 θ 3 ET T 2 ET T 3 ET 2 T 3 + ET 2 θ θ 2 θ θ 3 θ 2 θ 3 θ θ θ 2 2θ θ 3 2θ 2 θ 3 + 2θ θ θ 2 θ θ 3 θ 2 θ 3 θ3 2 3 ( + θ 3 2 θ 3 Because of Corollary 4. I X (θ ni X (θ I X (θ 2n ( θ + θ 3 θ 3 θ 3 θ 2 + θ 3 0. Derive the Cramer Rao Bound for unbiased estimators of p SI γ. Apply formulary ( for all unbiased estimators V ar ( γ D θ g T I X (θ D θ g. Here g(θ 2θ θ 2, thus ( 2θ2 D θ g 2θ and 8
9 ( θ + θ 3 θ 3 Thus θ 3 θ 2 + θ 3 ( ( (θ3 + θ 2 θ θ θ 2 ( θ θ θ θ 2 θ θ 2 (θ 3 + θ θ 2 θ θ 2 ( θ 2 θ 2 I X (θ ( ( θ θ θ θ 2 2n θ θ 2 ( θ 2 θ 2. (4 Remind : θ 3 + θ 2 + θ Then the CRB is ( T ( θ2 ( θ θ θ θ 2 2 n θ θ θ 2 ( θ 2 θ 2 ( θ2 θ 2θ θ 2 θ 2 4θ θ 2 +θ n. Calculate the efficiencies. Consider first the moment estimator p SI n 4 n. V ar ( p SI n 2 V ar ( n i I 4 (x i n V ar (I 4 (x n 2θ θ 2 ( 2θ θ 2 efficiency: e( p SI, p SI θ 2 4θ θ 2 +θ n CRB V ar ( p SI 2θ θ 2 2θ n θ 2 ( 2θ θ 2 θ 2 4θ θ 2 + θ 2θ θ 2 Consider now the bias corrected maximum likelihood estimator: p SI T (x T 2 (x 2n 2 n. It holds V ar ( p SI (2n 2 n 2 V ar (T (x T 2 (x Observe (T (x, T 2 (x, T 2 (x multinomial(2n, θ, θ 2, θ 3 9
10 Presentation as sums of independent Bernoulli variables T j (x i Z i,j, Z i,j Ber(θ j i.i.d., j, 2, 3, Z i, +Z i,2 +Z i,3 Need V ar(t (x T 2 (x E (T (x T 2 (x 2 (ET (x T 2 (x 2 ( 2 T 2 (x Z i, Thus T 2 (x T2 2 (x Z i, ET (x T 2 (x (2n 2 n2θ θ 2 Z i, Z i2, i i i 2 i ( i Z i, Z i, + Z i, + ( Z i2,2 + Z i,z i2, i i 2 i 3 Z i2,2 i i 3 + Z i, Z i3,2z i4,2 i i 3 i 4 + Z i,z i2, Z i2,2 i i 2 i 3 + Z i2,2 i i 3 i because: Z i,z i,2 0 Z i, i Z i,z i2, Z i3,2z i4,2 i i 2 i 3 i 4 Z i,z i,2 + i 3 i 4 Z i3,2z i4,2 i i 3 Z i,z i2,2 Z i,z i3,2z i4,2 i i 3 i 4 Z i,z i2, i i 2 Z i3,2z i4,2 i 3 i 4 Z i,z i2,2 i i 3 0
11 Z i,z i2, i i 2 i 3 i 4 Z i3,2z i4,2 Z i,z i2,z i3,2z i4,2 i i 2 i 3 i 4 T 2 (x T 2 2 (x Z i,z i2,2 i i Z i,z i3,2z i4,2 i i 3 i 4 Z i2,z i3,z i4,2 i 2 i 3 i 4 i i 2 i 3 i 4 Z i,z i2,z i3,2z i4,2 ET 2 (x T2 2 (x 2n (2n θ θ 2 (5 +2n (2n (2n 2θ θ2 2 (6 +2n (2n (2n 2θ 2 θ 2 +2n (2n (2n 2(2n 3θθ V ar ( T 2 (x T2 2 (x ET 2 (x T2 2 (x (n (2n 2θ θ 2 2 2n (2n (2n 2(2n 3 (2n (2n 2 V ar ( p SI (2n 2 n V ar (T 2 (x T 2 (x 2n (2n θ θ 2 + 2n (2n (2n 2θ θ2 2 +2n (2n (2n 2θ (2n 2 n 2 2 θ 2 + n (2n (2n 2(2n 32θθ (n (2n 2θ θ 2 2 2θ θ 2 (2n 2 n ( + (2n 2(θ 2 + θ + (6 8nθ θ 2 Efficiency e( p SI, p SI 2n2 n θ 2θ θ 2 4θ θ 2 +θ 2 n 2θ θ 2 ( + (2n 2(θ 2 + θ + (6 8nθ θ 2
12 (2n θ 2 4θ θ 2 +θ (+(2n 2(θ 2 +θ +( 8n+6θ θ 2?????? Conclusion, also the corrected maximumlikelihood estimator is not efficient. 2. Calculate the efficiences of θ,mle, θ 2,MLE. We have θ,mle T, θ2,mle T 2. Furthermore T 2n 2n, T 2 are multinomial distributed. Thus and (T (x, T 2 (x, T 2 (x multinomial(2n, θ, θ 2, θ 3 ( θ ( θ Cov (T, T 2 2n θ θ 2 θ θ 2 θ 2 ( θ 2 Cov ( θ,mle, θ 2,MLE Cov ( θ,mle, θ 2,MLE 2n 4n Cov (T, T 2 2 ( θ ( θ θ θ 2 θ θ 2 θ 2 ( θ 2 comparing with CRBI X (θ see (4. We get that θ,mle, θ 2,MLE are efficient. 2
Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.
Sufficient Statistics Lecture Notes 6 Sufficiency Data reduction in terms of a particular statistic can be thought of as a partition of the sample space X. Definition T is sufficient for θ if the conditional
More informationDefinition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.
9 Point estimation 9.1 Rationale behind point estimation When sampling from a population described by a pdf f(x θ) or probability function P [X = x θ] knowledge of θ gives knowledge of the entire population.
More informationChapter 8: Sampling distributions of estimators Sections
Chapter 8 continued Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample
More informationChapter 8: Sampling distributions of estimators Sections
Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample variance Skip: p.
More informationChapter 4: Asymptotic Properties of MLE (Part 3)
Chapter 4: Asymptotic Properties of MLE (Part 3) Daniel O. Scharfstein 09/30/13 1 / 1 Breakdown of Assumptions Non-Existence of the MLE Multiple Solutions to Maximization Problem Multiple Solutions to
More informationExercise. Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1. Exercise Estimation
Exercise Show the corrected sample variance is an unbiased estimator of population variance. S 2 = n i=1 (X i X ) 2 n 1 Exercise S 2 = = = = n i=1 (X i x) 2 n i=1 = (X i µ + µ X ) 2 = n 1 n 1 n i=1 ((X
More informationTwo hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER
Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS Answer any FOUR of the SIX questions.
More informationLecture 10: Point Estimation
Lecture 10: Point Estimation MSU-STT-351-Sum-17B (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 1 / 31 Basic Concepts of Point Estimation A point estimate of a parameter θ,
More informationApplied Statistics I
Applied Statistics I Liang Zhang Department of Mathematics, University of Utah July 14, 2008 Liang Zhang (UofU) Applied Statistics I July 14, 2008 1 / 18 Point Estimation Liang Zhang (UofU) Applied Statistics
More informationActuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems
Actuarial Mathematics and Statistics Statistics 5 Part 2: Statistical Inference Tutorial Problems Spring 2005 1. Which of the following statements relate to probabilities that can be interpreted as frequencies?
More informationPoint Estimation. Copyright Cengage Learning. All rights reserved.
6 Point Estimation Copyright Cengage Learning. All rights reserved. 6.2 Methods of Point Estimation Copyright Cengage Learning. All rights reserved. Methods of Point Estimation The definition of unbiasedness
More information6. Genetics examples: Hardy-Weinberg Equilibrium
PBCB 206 (Fall 2006) Instructor: Fei Zou email: fzou@bios.unc.edu office: 3107D McGavran-Greenberg Hall Lecture 4 Topics for Lecture 4 1. Parametric models and estimating parameters from data 2. Method
More informationدرس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی
یادگیري ماشین توزیع هاي نمونه و تخمین نقطه اي پارامترها Sampling Distributions and Point Estimation of Parameter (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی درس هفتم 1 Outline Introduction
More informationPoint Estimation. Some General Concepts of Point Estimation. Example. Estimator quality
Point Estimation Some General Concepts of Point Estimation Statistical inference = conclusions about parameters Parameters == population characteristics A point estimate of a parameter is a value (based
More informationPoint Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage
6 Point Estimation Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Point Estimation Statistical inference: directed toward conclusions about one or more parameters. We will use the generic
More informationChapter 7: Estimation Sections
1 / 40 Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: Chapter 7 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods:
More informationA New Hybrid Estimation Method for the Generalized Pareto Distribution
A New Hybrid Estimation Method for the Generalized Pareto Distribution Chunlin Wang Department of Mathematics and Statistics University of Calgary May 18, 2011 A New Hybrid Estimation Method for the GPD
More informationLearning From Data: MLE. Maximum Likelihood Estimators
Learning From Data: MLE Maximum Likelihood Estimators 1 Parameter Estimation Assuming sample x1, x2,..., xn is from a parametric distribution f(x θ), estimate θ. E.g.: Given sample HHTTTTTHTHTTTHH of (possibly
More informationStatistical estimation
Statistical estimation Statistical modelling: theory and practice Gilles Guillot gigu@dtu.dk September 3, 2013 Gilles Guillot (gigu@dtu.dk) Estimation September 3, 2013 1 / 27 1 Introductory example 2
More informationChapter 8. Introduction to Statistical Inference
Chapter 8. Introduction to Statistical Inference Point Estimation Statistical inference is to draw some type of conclusion about one or more parameters(population characteristics). Now you know that a
More informationLecture 17: More on Markov Decision Processes. Reinforcement learning
Lecture 17: More on Markov Decision Processes. Reinforcement learning Learning a model: maximum likelihood Learning a value function directly Monte Carlo Temporal-difference (TD) learning COMP-424, Lecture
More informationLikelihood Methods of Inference. Toss coin 6 times and get Heads twice.
Methods of Inference Toss coin 6 times and get Heads twice. p is probability of getting H. Probability of getting exactly 2 heads is 15p 2 (1 p) 4 This function of p, is likelihood function. Definition:
More informationChapter 7: Estimation Sections
Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions Frequentist Methods: 7.5 Maximum Likelihood Estimators
More informationASSIGNMENT - 1, MAY M.Sc. (PREVIOUS) FIRST YEAR DEGREE STATISTICS. Maximum : 20 MARKS Answer ALL questions.
(DMSTT 0 NR) ASSIGNMENT -, MAY-04. PAPER- I : PROBABILITY AND DISTRIBUTION THEORY ) a) State and prove Borel-cantelli lemma b) Let (x, y) be jointly distributed with density 4 y(+ x) f( x, y) = y(+ x)
More informationMATH 3200 Exam 3 Dr. Syring
. Suppose n eligible voters are polled (randomly sampled) from a population of size N. The poll asks voters whether they support or do not support increasing local taxes to fund public parks. Let M be
More informationAnalysis of truncated data with application to the operational risk estimation
Analysis of truncated data with application to the operational risk estimation Petr Volf 1 Abstract. Researchers interested in the estimation of operational risk often face problems arising from the structure
More informationChapter 7: Point Estimation and Sampling Distributions
Chapter 7: Point Estimation and Sampling Distributions Seungchul Baek Department of Statistics, University of South Carolina STAT 509: Statistics for Engineers 1 / 20 Motivation In chapter 3, we learned
More informationPoint Estimators. STATISTICS Lecture no. 10. Department of Econometrics FEM UO Brno office 69a, tel
STATISTICS Lecture no. 10 Department of Econometrics FEM UO Brno office 69a, tel. 973 442029 email:jiri.neubauer@unob.cz 8. 12. 2009 Introduction Suppose that we manufacture lightbulbs and we want to state
More informationM.Sc. (Previous) DEGREE EXAMINATION, DEC First Year STATISTICS. Paper - I : Probability and Distribution Theory
(DMSTT 01) M.Sc. (Previous) DEGREE EXAMINATION, DEC. - 2015 First Year STATISTICS Paper - I : Probability and Distribution Theory Time : 3 Hours Maximum Marks : 70 Answer any Five questions All questions
More informationPractice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.
Practice Exercises for Midterm Exam ST 522 - Statistical Theory - II The ACTUAL exam will consists of less number of problems. 1. Suppose X i F ( ) for i = 1,..., n, where F ( ) is a strictly increasing
More informationChapter 6: Point Estimation
Chapter 6: Point Estimation Professor Sharabati Purdue University March 10, 2014 Professor Sharabati (Purdue University) Point Estimation Spring 2014 1 / 37 Chapter Overview Point estimator and point estimate
More informationContents Part I Descriptive Statistics 1 Introduction and Framework Population, Sample, and Observations Variables Quali
Part I Descriptive Statistics 1 Introduction and Framework... 3 1.1 Population, Sample, and Observations... 3 1.2 Variables.... 4 1.2.1 Qualitative and Quantitative Variables.... 5 1.2.2 Discrete and Continuous
More informationComputer Statistics with R
MAREK GAGOLEWSKI KONSTANCJA BOBECKA-WESO LOWSKA PRZEMYS LAW GRZEGORZEWSKI Computer Statistics with R 5. Point Estimation Faculty of Mathematics and Information Science Warsaw University of Technology []
More informationPractice Exam 1. Loss Amount Number of Losses
Practice Exam 1 1. You are given the following data on loss sizes: An ogive is used as a model for loss sizes. Determine the fitted median. Loss Amount Number of Losses 0 1000 5 1000 5000 4 5000 10000
More informationWeb-based Supplementary Materials for. A space-time conditional intensity model. for invasive meningococcal disease occurence
Web-based Supplementary Materials for A space-time conditional intensity model for invasive meningococcal disease occurence by Sebastian Meyer 1,2, Johannes Elias 3, and Michael Höhle 4,2 1 Department
More informationA Stochastic Reserving Today (Beyond Bootstrap)
A Stochastic Reserving Today (Beyond Bootstrap) Presented by Roger M. Hayne, PhD., FCAS, MAAA Casualty Loss Reserve Seminar 6-7 September 2012 Denver, CO CAS Antitrust Notice The Casualty Actuarial Society
More informationWeierstrass Institute for Applied Analysis and Stochastics Maximum likelihood estimation for jump diffusions
Weierstrass Institute for Applied Analysis and Stochastics Maximum likelihood estimation for jump diffusions Hilmar Mai Mohrenstrasse 39 1117 Berlin Germany Tel. +49 3 2372 www.wias-berlin.de Haindorf
More informationCSC 411: Lecture 08: Generative Models for Classification
CSC 411: Lecture 08: Generative Models for Classification Richard Zemel, Raquel Urtasun and Sanja Fidler University of Toronto Zemel, Urtasun, Fidler (UofT) CSC 411: 08-Generative Models 1 / 23 Today Classification
More informationThe method of Maximum Likelihood.
Maximum Likelihood The method of Maximum Likelihood. In developing the least squares estimator - no mention of probabilities. Minimize the distance between the predicted linear regression and the observed
More informationChapter 7: Estimation Sections
1 / 31 : Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods: 7.5 Maximum Likelihood
More informationCSE 312 Winter Learning From Data: Maximum Likelihood Estimators (MLE)
CSE 312 Winter 2017 Learning From Data: Maximum Likelihood Estimators (MLE) 1 Parameter Estimation Given: independent samples x1, x2,..., xn from a parametric distribution f(x θ) Goal: estimate θ. Not
More informationChapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi
Chapter 4: Commonly Used Distributions Statistics for Engineers and Scientists Fourth Edition William Navidi 2014 by Education. This is proprietary material solely for authorized instructor use. Not authorized
More informationPoint Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased.
Point Estimation Point Estimation Definition A point estimate of a parameter θ is a single number that can be regarded as a sensible value for θ. A point estimate is obtained by selecting a suitable statistic
More informationCS340 Machine learning Bayesian model selection
CS340 Machine learning Bayesian model selection Bayesian model selection Suppose we have several models, each with potentially different numbers of parameters. Example: M0 = constant, M1 = straight line,
More informationBack to estimators...
Back to estimators... So far, we have: Identified estimators for common parameters Discussed the sampling distributions of estimators Introduced ways to judge the goodness of an estimator (bias, MSE, etc.)
More informationPROBABILITY AND STATISTICS
Monday, January 12, 2015 1 PROBABILITY AND STATISTICS Zhenyu Ye January 12, 2015 Monday, January 12, 2015 2 References Ch10 of Experiments in Modern Physics by Melissinos. Particle Physics Data Group Review
More informationChapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29
Chapter 5 Univariate time-series analysis () Chapter 5 Univariate time-series analysis 1 / 29 Time-Series Time-series is a sequence fx 1, x 2,..., x T g or fx t g, t = 1,..., T, where t is an index denoting
More informationSTRESS-STRENGTH RELIABILITY ESTIMATION
CHAPTER 5 STRESS-STRENGTH RELIABILITY ESTIMATION 5. Introduction There are appliances (every physical component possess an inherent strength) which survive due to their strength. These appliances receive
More informationFinancial Risk Forecasting Chapter 9 Extreme Value Theory
Financial Risk Forecasting Chapter 9 Extreme Value Theory Jon Danielsson 2017 London School of Economics To accompany Financial Risk Forecasting www.financialriskforecasting.com Published by Wiley 2011
More informationLog-Robust Portfolio Management
Log-Robust Portfolio Management Dr. Aurélie Thiele Lehigh University Joint work with Elcin Cetinkaya and Ban Kawas Research partially supported by the National Science Foundation Grant CMMI-0757983 Dr.
More informationEstimation of a parametric function associated with the lognormal distribution 1
Communications in Statistics Theory and Methods Estimation of a parametric function associated with the lognormal distribution Jiangtao Gou a,b and Ajit C. Tamhane c, a Department of Mathematics and Statistics,
More informationIntroduction to the Maximum Likelihood Estimation Technique. September 24, 2015
Introduction to the Maximum Likelihood Estimation Technique September 24, 2015 So far our Dependent Variable is Continuous That is, our outcome variable Y is assumed to follow a normal distribution having
More information12 The Bootstrap and why it works
12 he Bootstrap and why it works For a review of many applications of bootstrap see Efron and ibshirani (1994). For the theory behind the bootstrap see the books by Hall (1992), van der Waart (2000), Lahiri
More information8.1 Estimation of the Mean and Proportion
8.1 Estimation of the Mean and Proportion Statistical inference enables us to make judgments about a population on the basis of sample information. The mean, standard deviation, and proportions of a population
More informationSTAT 135 Solutions to Homework 3: 30 points
STAT 35 Solutios to Homework 3: 30 poits Sprig 205 The objective of this Problem Set is to study the Stei Pheomeo 955. Suppose that θ θ, θ 2,..., θ cosists of ukow parameters, with 3. We wish to estimate
More informationMachine Learning for Quantitative Finance
Machine Learning for Quantitative Finance Fast derivative pricing Sofie Reyners Joint work with Jan De Spiegeleer, Dilip Madan and Wim Schoutens Derivative pricing is time-consuming... Vanilla option pricing
More informationA Learning Theory of Ranking Aggregation
A Learning Theory of Ranking Aggregation France/Japan Machine Learning Workshop Anna Korba, Stephan Clémençon, Eric Sibony November 14, 2017 Télécom ParisTech Outline 1. The Ranking Aggregation Problem
More information1 Residual life for gamma and Weibull distributions
Supplement to Tail Estimation for Window Censored Processes Residual life for gamma and Weibull distributions. Gamma distribution Let Γ(k, x = x yk e y dy be the upper incomplete gamma function, and let
More informationPoint Estimation. Edwin Leuven
Point Estimation Edwin Leuven Introduction Last time we reviewed statistical inference We saw that while in probability we ask: given a data generating process, what are the properties of the outcomes?
More informationChapter 5. Sampling Distributions
Lecture notes, Lang Wu, UBC 1 Chapter 5. Sampling Distributions 5.1. Introduction In statistical inference, we attempt to estimate an unknown population characteristic, such as the population mean, µ,
More informationAsymptotic results discrete time martingales and stochastic algorithms
Asymptotic results discrete time martingales and stochastic algorithms Bernard Bercu Bordeaux University, France IFCAM Summer School Bangalore, India, July 2015 Bernard Bercu Asymptotic results for discrete
More informationHomework Problems Stat 479
Chapter 10 91. * A random sample, X1, X2,, Xn, is drawn from a distribution with a mean of 2/3 and a variance of 1/18. ˆ = (X1 + X2 + + Xn)/(n-1) is the estimator of the distribution mean θ. Find MSE(
More informationObjective Bayesian Analysis for Heteroscedastic Regression
Analysis for Heteroscedastic Regression & Esther Salazar Universidade Federal do Rio de Janeiro Colóquio Inter-institucional: Modelos Estocásticos e Aplicações 2009 Collaborators: Marco Ferreira and Thais
More informationChapter 5. Statistical inference for Parametric Models
Chapter 5. Statistical inference for Parametric Models Outline Overview Parameter estimation Method of moments How good are method of moments estimates? Interval estimation Statistical Inference for Parametric
More informationAsymmetric Type II Compound Laplace Distributions and its Properties
CHAPTER 4 Asymmetric Type II Compound Laplace Distributions and its Properties 4. Introduction Recently there is a growing trend in the literature on parametric families of asymmetric distributions which
More informationChapter 7. Sampling Distributions and the Central Limit Theorem
Chapter 7. Sampling Distributions and the Central Limit Theorem 1 Introduction 2 Sampling Distributions related to the normal distribution 3 The central limit theorem 4 The normal approximation to binomial
More informationSTAT 509: Statistics for Engineers Dr. Dewei Wang. Copyright 2014 John Wiley & Sons, Inc. All rights reserved.
STAT 509: Statistics for Engineers Dr. Dewei Wang Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger 7 Point CHAPTER OUTLINE 7-1 Point Estimation 7-2
More informationFinancial Econometrics
Financial Econometrics Volatility Gerald P. Dwyer Trinity College, Dublin January 2013 GPD (TCD) Volatility 01/13 1 / 37 Squared log returns for CRSP daily GPD (TCD) Volatility 01/13 2 / 37 Absolute value
More information2 of PU_2015_375 Which of the following measures is more flexible when compared to other measures?
PU M Sc Statistics 1 of 100 194 PU_2015_375 The population census period in India is for every:- quarterly Quinqennial year biannual Decennial year 2 of 100 105 PU_2015_375 Which of the following measures
More informationProbability & Statistics
Probability & Statistics BITS Pilani K K Birla Goa Campus Dr. Jajati Keshari Sahoo Department of Mathematics Statistics Descriptive statistics Inferential statistics /38 Inferential Statistics 1. Involves:
More informationChapter 7. Sampling Distributions and the Central Limit Theorem
Chapter 7. Sampling Distributions and the Central Limit Theorem 1 Introduction 2 Sampling Distributions related to the normal distribution 3 The central limit theorem 4 The normal approximation to binomial
More information(a) Is it possible for the rate of exit from OLF into E tobethesameastherateof exit from U into E? Why or why not?
Final (Take-Home) Exam The Formulation and Estimation of Equilibrium Models of Household and Labor Market Behavior Department of Economics, TAU Professor C. Flinn Please answer five (5) of the following
More informationBivariate Birnbaum-Saunders Distribution
Department of Mathematics & Statistics Indian Institute of Technology Kanpur January 2nd. 2013 Outline 1 Collaborators 2 3 Birnbaum-Saunders Distribution: Introduction & Properties 4 5 Outline 1 Collaborators
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Further Variance Reduction Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Outline
More informationQualifying Exam Solutions: Theoretical Statistics
Qualifying Exam Solutions: Theoretical Statistics. (a) For the first sampling plan, the expectation of any statistic W (X, X,..., X n ) is a polynomial of θ of degree less than n +. Hence τ(θ) cannot have
More information4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.
4-1 Chapter 4 Commonly Used Distributions 2014 by The Companies, Inc. All rights reserved. Section 4.1: The Bernoulli Distribution 4-2 We use the Bernoulli distribution when we have an experiment which
More informationChapter 7: Portfolio Theory
Chapter 7: Portfolio Theory 1. Introduction 2. Portfolio Basics 3. The Feasible Set 4. Portfolio Selection Rules 5. The Efficient Frontier 6. Indifference Curves 7. The Two-Asset Portfolio 8. Unrestriceted
More informationModeling of Price. Ximing Wu Texas A&M University
Modeling of Price Ximing Wu Texas A&M University As revenue is given by price times yield, farmers income risk comes from risk in yield and output price. Their net profit also depends on input price, but
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation The likelihood and log-likelihood functions are the basis for deriving estimators for parameters, given data. While the shapes of these two functions are different, they have
More informationCase Study: Heavy-Tailed Distribution and Reinsurance Rate-making
Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making May 30, 2016 The purpose of this case study is to give a brief introduction to a heavy-tailed distribution and its distinct behaviors in
More information10/1/2012. PSY 511: Advanced Statistics for Psychological and Behavioral Research 1
PSY 511: Advanced Statistics for Psychological and Behavioral Research 1 Pivotal subject: distributions of statistics. Foundation linchpin important crucial You need sampling distributions to make inferences:
More informationProbability Theory. Probability and Statistics for Data Science CSE594 - Spring 2016
Probability Theory Probability and Statistics for Data Science CSE594 - Spring 2016 What is Probability? 2 What is Probability? Examples outcome of flipping a coin (seminal example) amount of snowfall
More informationAN EXTREME VALUE APPROACH TO PRICING CREDIT RISK
AN EXTREME VALUE APPROACH TO PRICING CREDIT RISK SOFIA LANDIN Master s thesis 2018:E69 Faculty of Engineering Centre for Mathematical Sciences Mathematical Statistics CENTRUM SCIENTIARUM MATHEMATICARUM
More informationDesirable properties for a good model of portfolio credit risk modelling
3.3 Default correlation binomial models Desirable properties for a good model of portfolio credit risk modelling Default dependence produce default correlations of a realistic magnitude. Estimation number
More informationAsymmetric Information in Health Insurance: Evidence from the National Medical Expenditure Survey. Cardon and Hendel
Asymmetric Information in Health Insurance: Evidence from the National Medical Expenditure Survey. Cardon and Hendel This paper separately estimates adverse selection and moral hazard. Two-stage decision.
More informationComparing the Means of. Two Log-Normal Distributions: A Likelihood Approach
Journal of Statistical and Econometric Methods, vol.3, no.1, 014, 137-15 ISSN: 179-660 (print), 179-6939 (online) Scienpress Ltd, 014 Comparing the Means of Two Log-Normal Distributions: A Likelihood Approach
More informationA start of Variational Methods for ERGM Ranran Wang, UW
A start of Variational Methods for ERGM Ranran Wang, UW MURI-UCI April 24, 2009 Outline A start of Variational Methods for ERGM [1] Introduction to ERGM Current methods of parameter estimation: MCMCMLE:
More informationSTAT 830 Convergence in Distribution
STAT 830 Convergence in Distribution Richard Lockhart Simon Fraser University STAT 830 Fall 2013 Richard Lockhart (Simon Fraser University) STAT 830 Convergence in Distribution STAT 830 Fall 2013 1 / 31
More informationFinancial Econometrics Jeffrey R. Russell. Midterm 2014 Suggested Solutions. TA: B. B. Deng
Financial Econometrics Jeffrey R. Russell Midterm 2014 Suggested Solutions TA: B. B. Deng Unless otherwise stated, e t is iid N(0,s 2 ) 1. (12 points) Consider the three series y1, y2, y3, and y4. Match
More informationUnit 5: Sampling Distributions of Statistics
Unit 5: Sampling Distributions of Statistics Statistics 571: Statistical Methods Ramón V. León 6/12/2004 Unit 5 - Stat 571 - Ramon V. Leon 1 Definitions and Key Concepts A sample statistic used to estimate
More informationSTART HERE: Instructions. 1 Exponential Family [Zhou, Manzil]
START HERE: Instructions Thanks a lot to John A.W.B. Constanzo and Shi Zong for providing and allowing to use the latex source files for quick preparation of the HW solution. The homework was due at 9:00am
More information14.30 Introduction to Statistical Methods in Economics Spring 2009
MIT OpenCourseWare http://ocw.mit.edu 14.30 Introduction to Statistical Methods in Economics Spring 2009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationUnit 5: Sampling Distributions of Statistics
Unit 5: Sampling Distributions of Statistics Statistics 571: Statistical Methods Ramón V. León 6/12/2004 Unit 5 - Stat 571 - Ramon V. Leon 1 Definitions and Key Concepts A sample statistic used to estimate
More informationSTA 532: Theory of Statistical Inference
STA 532: Theory of Statistical Inference Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 2 Estimating CDFs and Statistical Functionals Empirical CDFs Let {X i : i n}
More informationOptimal Search for Parameters in Monte Carlo Simulation for Derivative Pricing
Optimal Search for Parameters in Monte Carlo Simulation for Derivative Pricing Prof. Chuan-Ju Wang Department of Computer Science University of Taipei Joint work with Prof. Ming-Yang Kao March 28, 2014
More informationChapter 3. Dynamic discrete games and auctions: an introduction
Chapter 3. Dynamic discrete games and auctions: an introduction Joan Llull Structural Micro. IDEA PhD Program I. Dynamic Discrete Games with Imperfect Information A. Motivating example: firm entry and
More informationSummary. Recap. Last Lecture. .1 If you know MLE of θ, can you also know MLE of τ(θ) for any function τ?
Last Lecture Biostatistics 60 - Statistical Iferece Lecture Cramer-Rao Theorem Hyu Mi Kag February 9th, 03 If you kow MLE of, ca you also kow MLE of τ() for ay fuctio τ? What are plausible ways to compare
More informationModelling financial data with stochastic processes
Modelling financial data with stochastic processes Vlad Ardelean, Fabian Tinkl 01.08.2012 Chair of statistics and econometrics FAU Erlangen-Nuremberg Outline Introduction Stochastic processes Volatility
More informationChapter 3 - Lecture 4 Moments and Moment Generating Funct
Chapter 3 - Lecture 4 and s October 7th, 2009 Chapter 3 - Lecture 4 and Moment Generating Funct Central Skewness Chapter 3 - Lecture 4 and Moment Generating Funct Central Skewness The expected value of
More informationM.Sc. ACTUARIAL SCIENCE. Term-End Examination
No. of Printed Pages : 15 LMJA-010 (F2F) M.Sc. ACTUARIAL SCIENCE Term-End Examination O CD December, 2011 MIA-010 (F2F) : STATISTICAL METHOD Time : 3 hours Maximum Marks : 100 SECTION - A Attempt any five
More information