STA 114: Statistics. Notes 10. Conjugate Priors
|
|
- Chrystal Lane
- 6 years ago
- Views:
Transcription
1 STA 114: Statistics Notes 10. Conjugate Priors Conjugate family Once we get a /pmf ξ(θ x) by combining a model X f(x θ) with a /pmf ξ(θ) on θ Θ, a report can be made by summarizing the. It helps to have the /pmf in a recognizable form so that we can easily compute its mean, spread, quantiles etc. This is not guaranteed to happen in general. For example, the model X Binomial(n, p) and the ξ(p) = e p /(e 1), p [0, 1] lead to the ξ(p x) = const p x (1 p) nx e p, p [0, 1], with a constant term that is fairly difficult to compute, making it difficult to get summaries of this (recall Lab 4). However, for certain models, certain s do lead to s that are analytically tractable. We already saw one in the female birth rate analysis, a uniform for the binomial model gives a beta. In fact more is true for this model: any beta leads to a beta! This phenomenon is called conjugacy. A formal definition is given below. Definition 1 (Conjugacy). A collection of s (or pmfs) is called a conjugate family for a model X f(x θ), θ Θ, if whenever a ξ(θ) is chosen from the collection, it leads to a ξ(θ x) that is also a member of the collection, for every observation X = x. Conjugacy in itself is not a very useful property. For example the collection of all s on Θ is surely conjugate to the model. It becomes useful when a small collection of s exhibit conjugacy to a certain statistical model. By a small collection we usually mean a collection of s/pmfs G = {g(θ a) : a A} indexed by a low-dimensional vector a. G is conjugate to a statistical model X f(x θ) if ξ(θ) = g(θ a) for some a A means for every x, ξ(θ x) = g(θ a ) for some a A. As mentioned before, the collection of beta s {Beta(a, b) : a > 0, b > 0} is a (-dimensional) conjugate family to the binomial model X Binomial(n, p), p [0, 1]. Table?? below gives a list of other common models with known, low-dimensional conjugate families. We will establish conjugacy for three of the listed models; you re required to do the maths for the remaining ones in HW5. The binomial-beta conjugacy The of bet(a, b) distribution, for a > 0, b > 0. equals g(p) = p a1 (1 p) b1 /B(a, b), p [0, 1], where B(a, b) = 1 0 qa1 (1 q) b1 dq is known as the Beta function. If we take ξ(p) = Beta(a, b) (for some a > 0, b > 0) as the for a binomial model X Binomial(n, p), p [0, 1], then for any observations x {0, 1,, n}, ξ(p x) = const f(x p)ξ(p) = const p x (1p) nx p a1 (1p) b1 = const p a+x1 (1p) b+nx1. 1
2 Model Parameter Prior Posterior X Binomial(n, p) 0 p 1 Beta(a, b) Beta(a, b ) a > 0, b > 0 a = a + x b = b + n x X = (X 1,, X n ) λ > 0 Gamma(a, b) Gamma(a, b ) X i Poisson(λ) a > 0, b > 0 a = a + n x b = b + n X = (X 1,, X n ) λ > 0 Gamma(a, b) Gamma(a, b ) X i Exponential(λ) a > 0, b > 0 a = a + n b = b + n x X = (X 1,, X n ) < < Normal(a, b ) Normal(a, b ) Normal(, σ ) < a < a = nb x+σ a nb +σ σ known b > 0 b = σ b nb +σ X = (X 1,, X n ) < < Nχ (m, k, r, s) Nχ (m, k, r, s ) Normal(, σ ) σ > 0 < m < m = km+n x X i X i k > 0, r > 0, s > 0 s = k+n k = k + n r = r + n rs+ kn k+n ( xm) +(n1)s x r+n Table 1: Conjugate and for some common models. But the of Beta(a+x, b+nx) (note: a+x > 0, b+nx > 0) is p a+x1 (1p) b+nx1 /B(a+ x, b + n x). Therefore ξ(p x) is a constant multiple of the Beta(a + x, b + n x). But if two s are constant multiples of each other, they must be identical (and the constant must be 1). So ξ(p x) = Beta(a + x, b + n x). The normal-normal conjugacy Next we show that for the model X 1,, X n Normal(, σ ), (, ), σ fixed, the ξ() = Normal(a, b ) gives a ξ( x) = Normal(a, b ) for some a and b [which we shall identify]. It suffices to show that ξ( x) is a constant multiple of the Normal(a, b ) density. This is equivalent to showing log ξ( x) = const + ( a ) b
3 by going in the log-scale. Now, by definition log ξ( x) = const + l x () + log ξ() n i=1 = const (x i ) ( a) σ b n i=1 = const (x i x) + n( x ) σ = const 1 [ ] n( x ) ( a) + σ b = const 1 ( nb x+σ a nb +σ ) b σ nb +σ ( a) b and therefore, ξ( x) = Normal( nb x+σ a completion of squares identity (give it a try!): b, σ ). nb +σ nb +σ The last equality above follows from a n( x ) ( a) + σ b = (nb + σ nb x+σa )( ) nb +σ + b σ A conjugate family for the full normal model n( x a) nb + σ. For the full normal model X 1,, X n Normal(, σ ), (, σ ) (, ) (0, ), we need a bivariate ξ(, σ ) on (, ) (0, ). There are several choices here. For example we could take ξ(, σ ) = g()h(σ ) where g() is a on (, ) and h(σ ) is a on (0, ). This is in fact widely used, with g() usually taken to be a normal and h(σ ) taken to be an inverse-gamma (i.e., the of 1/σ is a gamma ). However, the family of such s are not conjugate to the model. In particular, the ξ(, σ x) does not factor into a product g( x)h(σ x). There is however, a conjugate family of s, known as the normal-inverse-chi-square s, denotes Nχ (m, k, r, s) with parameters m (, ), k > 0, r > 0 and s > 0. The of this distributionis given by: g(w, v) = const. v r+3 exp ( k(w m) + rs v where the constant equals (rs/)r/ πγ(r/). Here are two important results. ), (w, v) (, ) (0, ) Result 1. A pair of random variables (W, V ) has a Nχ (m, k, r, s) if and only if 1. rs V χ (r), and. [W V = v] Normal(m, v/k). Result. If (W, V ) Nχ (m, k, r, s) then W m t(r). s/k 3
4 The first result can be proved by direct calculations (you re welcome to try it). The second result follows once you note that if (W, V ) Nχ (m, k, r, s) then by Result 1, we can write W = m + V/kZ where Z Normal(0, 1) is independent of V. Also U = rs V χ (r). Then, W m V/kZ = = Z t(r). s/k s/k U/r With this as the prelude, we re now ready to prove that for ξ(, σ ) = Nχ (m, k, r, s) we get ξ(, σ x) = Nχ (m, k (s), r, s ) where m = km+n x k+n k = n + k r = n + r s = 1 kn (rs + ( x n+r k+n m) + (n 1)s x). Working in the log-scale we have log ξ(, σ x) = l x (, σ ) + log ξ(, σ ) = const n n log σ i=1 (x i ) r + 3 log σ k( m) + rs σ σ = const n + r + 3 n log σ i=1 (x i ) + k( m) + rs σ = const n + r + 3 log σ (n 1)s x + n( x ) + k( m) + rs. σ There is another square-completion identity that we need (try this): ( n( x ) + k( m) = (k + n) Plugging this in we get log ξ(, σ x) = const r + 3 and hence ξ(, σ x) = Nχ (m, k, r, s ). Putting it to use ) km + n x + kn k + n k + n ( x m). log σ k ( m ) + r s σ It is nice to have low-dimensional conjugate family for a statistical model one is interested in. But we still have the job of deciding upon one member of the family to ultimately use as the /pmf. If a singular choice is hard to justify, we can run the analysis for a multitude of reasonable choices and present the analysis side by side. This can be done both graphically, and more important via tables. To summarize a or a, we can report a number representing the center of the, such as the mean or the median. To summarize the spread of the, we can 4
5 Prior Prior summary Posterior Posterior summary Mean [.05,.975 ] Mean [.05 (x),.975 (x)] Gamma(,.) 10 (1.1, 7.86) Gamma(15, 10.) 14.9 (1.63, 17.36) Gamma(5,.5) 10 (3.5, 0.48) Gamma(155, 10.5) (1.53, 17.17) Gamma(50, 5) 10 (7.4, 1.96) Gamma(00, 15) (11.55, 15.4) Gamma(500, 50) 10 (9.14, 10.9) Gamma(650, 60) (10.0, 11.68) Table : Poisson-gamma example, and summaries report an interval [a, b] such that the packs most of its area inside the interval. If we want an 1 α area inside, for some small α, then we can take a = the α/-th quantile of the and b = the (1 α/)-th quantile of it. For a ξ(θ), we ll denote its u-th quantile by θ u and for a ξ(θ x) we will use the notation θ u (x). We will usually work with α = 5% and hence report [θ.05, θ.975 ] from the and [θ.05 (x), θ.975 (x)] from the. Example. Consider hurricane counts X 1,, X n for n = 10 consecutive years in the north Atlantic basin, modeled as X i Poisson(). Figure?? and Table?? show - summaries for 4 choices of the from the conjugate family of gamma s. The observed data are x = (1, 14, 15, 1, 16, 14, 7, 10, 14, 16). 5
6 a =, b = 0. a = 5, b = a = 50, b = 5 a = 500, b = Figure 1: Prior- summary for hurricane count data with Gamma(a, b) conjugate s. 6
Chapter 7: Estimation Sections
1 / 40 Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: Chapter 7 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods:
More informationChapter 7: Estimation Sections
1 / 31 : Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods: 7.5 Maximum Likelihood
More informationMATH 3200 Exam 3 Dr. Syring
. Suppose n eligible voters are polled (randomly sampled) from a population of size N. The poll asks voters whether they support or do not support increasing local taxes to fund public parks. Let M be
More informationExam 2 Spring 2015 Statistics for Applications 4/9/2015
18.443 Exam 2 Spring 2015 Statistics for Applications 4/9/2015 1. True or False (and state why). (a). The significance level of a statistical test is not equal to the probability that the null hypothesis
More informationLecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.
Sufficient Statistics Lecture Notes 6 Sufficiency Data reduction in terms of a particular statistic can be thought of as a partition of the sample space X. Definition T is sufficient for θ if the conditional
More informationCommon one-parameter models
Common one-parameter models In this section we will explore common one-parameter models, including: 1. Binomial data with beta prior on the probability 2. Poisson data with gamma prior on the rate 3. Gaussian
More informationConjugate Models. Patrick Lam
Conjugate Models Patrick Lam Outline Conjugate Models What is Conjugacy? The Beta-Binomial Model The Normal Model Normal Model with Unknown Mean, Known Variance Normal Model with Known Mean, Unknown Variance
More informationConjugate priors: Beta and normal Class 15, Jeremy Orloff and Jonathan Bloom
1 Learning Goals Conjugate s: Beta and normal Class 15, 18.05 Jeremy Orloff and Jonathan Bloom 1. Understand the benefits of conjugate s.. Be able to update a beta given a Bernoulli, binomial, or geometric
More informationRandom Variables Handout. Xavier Vilà
Random Variables Handout Xavier Vilà Course 2004-2005 1 Discrete Random Variables. 1.1 Introduction 1.1.1 Definition of Random Variable A random variable X is a function that maps each possible outcome
More informationTwo hours UNIVERSITY OF MANCHESTER. 23 May :00 16:00. Answer ALL SIX questions The total number of marks in the paper is 90.
Two hours MATH39542 UNIVERSITY OF MANCHESTER RISK THEORY 23 May 2016 14:00 16:00 Answer ALL SIX questions The total number of marks in the paper is 90. University approved calculators may be used 1 of
More informationTwo hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER
Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS Answer any FOUR of the SIX questions.
More informationMVE051/MSG Lecture 7
MVE051/MSG810 2017 Lecture 7 Petter Mostad Chalmers November 20, 2017 The purpose of collecting and analyzing data Purpose: To build and select models for parts of the real world (which can be used for
More informationcontinuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence
continuous rv Let X be a continuous rv. Then a probability distribution or probability density function (pdf) of X is a function f(x) such that for any two numbers a and b with a b, P(a X b) = b a f (x)dx.
More informationSubject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018
` Subject CS1 Actuarial Statistics 1 Core Principles Syllabus for the 2019 exams 1 June 2018 Copyright in this Core Reading is the property of the Institute and Faculty of Actuaries who are the sole distributors.
More informationChapter 7: Estimation Sections
Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions Frequentist Methods: 7.5 Maximum Likelihood Estimators
More informationCS 361: Probability & Statistics
March 12, 2018 CS 361: Probability & Statistics Inference Binomial likelihood: Example Suppose we have a coin with an unknown probability of heads. We flip the coin 10 times and observe 2 heads. What can
More informationChapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as
Lecture 0 on BST 63: Statistical Theory I Kui Zhang, 09/9/008 Review for the previous lecture Definition: Several continuous distributions, including uniform, gamma, normal, Beta, Cauchy, double exponential
More informationThe Bernoulli distribution
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
More informationPractice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.
Practice Exercises for Midterm Exam ST 522 - Statistical Theory - II The ACTUAL exam will consists of less number of problems. 1. Suppose X i F ( ) for i = 1,..., n, where F ( ) is a strictly increasing
More informationPosterior Inference. , where should we start? Consider the following computational procedure: 1. draw samples. 2. convert. 3. compute properties
Posterior Inference Example. Consider a binomial model where we have a posterior distribution for the probability term, θ. Suppose we want to make inferences about the log-odds γ = log ( θ 1 θ), where
More informationBayesian Linear Model: Gory Details
Bayesian Linear Model: Gory Details Pubh7440 Notes By Sudipto Banerjee Let y y i ] n i be an n vector of independent observations on a dependent variable (or response) from n experimental units. Associated
More informationChapter 8: Sampling distributions of estimators Sections
Chapter 8 continued Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample
More informationChapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi
Chapter 4: Commonly Used Distributions Statistics for Engineers and Scientists Fourth Edition William Navidi 2014 by Education. This is proprietary material solely for authorized instructor use. Not authorized
More informationProbability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions
April 9th, 2018 Lecture 20: Special distributions Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters 4, 6: Random variables Week 9 Chapter
More information**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:
**BEGINNING OF EXAMINATION** 1. You are given: (i) A random sample of five observations from a population is: 0.2 0.7 0.9 1.1 1.3 (ii) You use the Kolmogorov-Smirnov test for testing the null hypothesis,
More informationGenerating Random Numbers
Generating Random Numbers Aim: produce random variables for given distribution Inverse Method Let F be the distribution function of an univariate distribution and let F 1 (y) = inf{x F (x) y} (generalized
More informationNon-informative Priors Multiparameter Models
Non-informative Priors Multiparameter Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Prior Types Informative vs Non-informative There has been a desire for a prior distributions that
More informationPart II: Computation for Bayesian Analyses
Part II: Computation for Bayesian Analyses 62 BIO 233, HSPH Spring 2015 Conjugacy In both birth weight eamples the posterior distribution is from the same family as the prior: Prior Likelihood Posterior
More informationCommonly Used Distributions
Chapter 4: Commonly Used Distributions 1 Introduction Statistical inference involves drawing a sample from a population and analyzing the sample data to learn about the population. We often have some knowledge
More informationChapter 5. Statistical inference for Parametric Models
Chapter 5. Statistical inference for Parametric Models Outline Overview Parameter estimation Method of moments How good are method of moments estimates? Interval estimation Statistical Inference for Parametric
More informationUQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.
UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. Random Variables 2 A random variable X is a numerical (integer, real, complex, vector etc.) summary of the outcome of the random experiment.
More informationAn Improved Skewness Measure
An Improved Skewness Measure Richard A. Groeneveld Professor Emeritus, Department of Statistics Iowa State University ragroeneveld@valley.net Glen Meeden School of Statistics University of Minnesota Minneapolis,
More informationPractice Exam 1. Loss Amount Number of Losses
Practice Exam 1 1. You are given the following data on loss sizes: An ogive is used as a model for loss sizes. Determine the fitted median. Loss Amount Number of Losses 0 1000 5 1000 5000 4 5000 10000
More informationME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.
ME3620 Theory of Engineering Experimentation Chapter III. Random Variables and Probability Distributions Chapter III 1 3.2 Random Variables In an experiment, a measurement is usually denoted by a variable
More informationReview for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom
Review for Final Exam 18.05 Spring 2014 Jeremy Orloff and Jonathan Bloom THANK YOU!!!! JON!! PETER!! RUTHI!! ERIKA!! ALL OF YOU!!!! Probability Counting Sets Inclusion-exclusion principle Rule of product
More informationChapter 3 Discrete Random Variables and Probability Distributions
Chapter 3 Discrete Random Variables and Probability Distributions Part 4: Special Discrete Random Variable Distributions Sections 3.7 & 3.8 Geometric, Negative Binomial, Hypergeometric NOTE: The discrete
More information6. Genetics examples: Hardy-Weinberg Equilibrium
PBCB 206 (Fall 2006) Instructor: Fei Zou email: fzou@bios.unc.edu office: 3107D McGavran-Greenberg Hall Lecture 4 Topics for Lecture 4 1. Parametric models and estimating parameters from data 2. Method
More informationIntroduction Recently the importance of modelling dependent insurance and reinsurance risks has attracted the attention of actuarial practitioners and
Asymptotic dependence of reinsurance aggregate claim amounts Mata, Ana J. KPMG One Canada Square London E4 5AG Tel: +44-207-694 2933 e-mail: ana.mata@kpmg.co.uk January 26, 200 Abstract In this paper we
More informationST440/550: Applied Bayesian Analysis. (5) Multi-parameter models - Summarizing the posterior
(5) Multi-parameter models - Summarizing the posterior Models with more than one parameter Thus far we have studied single-parameter models, but most analyses have several parameters For example, consider
More informationConfidence Intervals Introduction
Confidence Intervals Introduction A point estimate provides no information about the precision and reliability of estimation. For example, the sample mean X is a point estimate of the population mean μ
More informationM249 Diagnostic Quiz
THE OPEN UNIVERSITY Faculty of Mathematics and Computing M249 Diagnostic Quiz Prepared by the Course Team [Press to begin] c 2005, 2006 The Open University Last Revision Date: May 19, 2006 Version 4.2
More informationHomework Problems Stat 479
Chapter 10 91. * A random sample, X1, X2,, Xn, is drawn from a distribution with a mean of 2/3 and a variance of 1/18. ˆ = (X1 + X2 + + Xn)/(n-1) is the estimator of the distribution mean θ. Find MSE(
More informationSTART HERE: Instructions. 1 Exponential Family [Zhou, Manzil]
START HERE: Instructions Thanks a lot to John A.W.B. Constanzo and Shi Zong for providing and allowing to use the latex source files for quick preparation of the HW solution. The homework was due at 9:00am
More information1. You are given the following information about a stationary AR(2) model:
Fall 2003 Society of Actuaries **BEGINNING OF EXAMINATION** 1. You are given the following information about a stationary AR(2) model: (i) ρ 1 = 05. (ii) ρ 2 = 01. Determine φ 2. (A) 0.2 (B) 0.1 (C) 0.4
More informationData Analytics (CS40003) Practice Set IV (Topic: Probability and Sampling Distribution)
Data Analytics (CS40003) Practice Set IV (Topic: Probability and Sampling Distribution) I. Concept Questions 1. Give an example of a random variable in the context of Drawing a card from a deck of cards.
More informationChapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables
Chapter 5 Continuous Random Variables and Probability Distributions 5.1 Continuous Random Variables 1 2CHAPTER 5. CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Probability Distributions Probability
More informationProbability. An intro for calculus students P= Figure 1: A normal integral
Probability An intro for calculus students.8.6.4.2 P=.87 2 3 4 Figure : A normal integral Suppose we flip a coin 2 times; what is the probability that we get more than 2 heads? Suppose we roll a six-sided
More informationSimple Random Sample
Simple Random Sample A simple random sample (SRS) of size n consists of n elements from the population chosen in such a way that every set of n elements has an equal chance to be the sample actually selected.
More informationUsing Agent Belief to Model Stock Returns
Using Agent Belief to Model Stock Returns America Holloway Department of Computer Science University of California, Irvine, Irvine, CA ahollowa@ics.uci.edu Introduction It is clear that movements in stock
More informationChapter 3 Discrete Random Variables and Probability Distributions
Chapter 3 Discrete Random Variables and Probability Distributions Part 3: Special Discrete Random Variable Distributions Section 3.5 Discrete Uniform Section 3.6 Bernoulli and Binomial Others sections
More informationRules and Models 1 investigates the internal measurement approach for operational risk capital
Carol Alexander 2 Rules and Models Rules and Models 1 investigates the internal measurement approach for operational risk capital 1 There is a view that the new Basel Accord is being defined by a committee
More informationWeek 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals
Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg :
More informationSupplementary Material: Strategies for exploration in the domain of losses
1 Supplementary Material: Strategies for exploration in the domain of losses Paul M. Krueger 1,, Robert C. Wilson 2,, and Jonathan D. Cohen 3,4 1 Department of Psychology, University of California, Berkeley
More informationStatistical estimation
Statistical estimation Statistical modelling: theory and practice Gilles Guillot gigu@dtu.dk September 3, 2013 Gilles Guillot (gigu@dtu.dk) Estimation September 3, 2013 1 / 27 1 Introductory example 2
More informationThis is a open-book exam. Assigned: Friday November 27th 2009 at 16:00. Due: Monday November 30th 2009 before 10:00.
University of Iceland School of Engineering and Sciences Department of Industrial Engineering, Mechanical Engineering and Computer Science IÐN106F Industrial Statistics II - Bayesian Data Analysis Fall
More informationDescribing Uncertain Variables
Describing Uncertain Variables L7 Uncertainty in Variables Uncertainty in concepts and models Uncertainty in variables Lack of precision Lack of knowledge Variability in space/time Describing Uncertainty
More informationSTA 532: Theory of Statistical Inference
STA 532: Theory of Statistical Inference Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 2 Estimating CDFs and Statistical Functionals Empirical CDFs Let {X i : i n}
More informationCS 237: Probability in Computing
CS 237: Probability in Computing Wayne Snyder Computer Science Department Boston University Lecture 12: Continuous Distributions Uniform Distribution Normal Distribution (motivation) Discrete vs Continuous
More information4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.
4-1 Chapter 4 Commonly Used Distributions 2014 by The Companies, Inc. All rights reserved. Section 4.1: The Bernoulli Distribution 4-2 We use the Bernoulli distribution when we have an experiment which
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Generating Random Variables and Stochastic Processes Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com
More informationBayesian course - problem set 3 (lecture 4)
Bayesian course - problem set 3 (lecture 4) Ben Lambert November 14, 2016 1 Ticked off Imagine once again that you are investigating the occurrence of Lyme disease in the UK. This is a vector-borne disease
More informationChapter 6. Importance sampling. 6.1 The basics
Chapter 6 Importance sampling 6.1 The basics To movtivate our discussion consider the following situation. We want to use Monte Carlo to compute µ E[X]. There is an event E such that P(E) is small but
More informationNormal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is
Normal Distribution Normal Distribution Definition A continuous rv X is said to have a normal distribution with parameter µ and σ (µ and σ 2 ), where < µ < and σ > 0, if the pdf of X is f (x; µ, σ) = 1
More informationRandom variables. Contents
Random variables Contents 1 Random Variable 2 1.1 Discrete Random Variable............................ 3 1.2 Continuous Random Variable........................... 5 1.3 Measures of Location...............................
More informationBIOL The Normal Distribution and the Central Limit Theorem
BIOL 300 - The Normal Distribution and the Central Limit Theorem In the first week of the course, we introduced a few measures of center and spread, and discussed how the mean and standard deviation are
More informationModelling of Long-Term Risk
Modelling of Long-Term Risk Roger Kaufmann Swiss Life roger.kaufmann@swisslife.ch 15th International AFIR Colloquium 6-9 September 2005, Zurich c 2005 (R. Kaufmann, Swiss Life) Contents A. Basel II B.
More informationSome Discrete Distribution Families
Some Discrete Distribution Families ST 370 Many families of discrete distributions have been studied; we shall discuss the ones that are most commonly found in applications. In each family, we need a formula
More informationOutline. Review Continuation of exercises from last time
Bayesian Models II Outline Review Continuation of exercises from last time 2 Review of terms from last time Probability density function aka pdf or density Likelihood function aka likelihood Conditional
More informationGPD-POT and GEV block maxima
Chapter 3 GPD-POT and GEV block maxima This chapter is devoted to the relation between POT models and Block Maxima (BM). We only consider the classical frameworks where POT excesses are assumed to be GPD,
More informationBayesian Normal Stuff
Bayesian Normal Stuff - Set-up of the basic model of a normally distributed random variable with unknown mean and variance (a two-parameter model). - Discuss philosophies of prior selection - Implementation
More informationStatistics 431 Spring 2007 P. Shaman. Preliminaries
Statistics 4 Spring 007 P. Shaman The Binomial Distribution Preliminaries A binomial experiment is defined by the following conditions: A sequence of n trials is conducted, with each trial having two possible
More informationExam STAM Practice Exam #1
!!!! Exam STAM Practice Exam #1 These practice exams should be used during the month prior to your exam. This practice exam contains 20 questions, of equal value, corresponding to about a 2 hour exam.
More informationTOPIC: PROBABILITY DISTRIBUTIONS
TOPIC: PROBABILITY DISTRIBUTIONS There are two types of random variables: A Discrete random variable can take on only specified, distinct values. A Continuous random variable can take on any value within
More informationSOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS
SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS Questions 1-307 have been taken from the previous set of Exam C sample questions. Questions no longer relevant
More information1 Bayesian Bias Correction Model
1 Bayesian Bias Correction Model Assuming that n iid samples {X 1,...,X n }, were collected from a normal population with mean µ and variance σ 2. The model likelihood has the form, P( X µ, σ 2, T n >
More informationBasic Procedure for Histograms
Basic Procedure for Histograms 1. Compute the range of observations (min. & max. value) 2. Choose an initial # of classes (most likely based on the range of values, try and find a number of classes that
More informationPROBABILITY DISTRIBUTIONS
CHAPTER 3 PROBABILITY DISTRIBUTIONS Page Contents 3.1 Introduction to Probability Distributions 51 3.2 The Normal Distribution 56 3.3 The Binomial Distribution 60 3.4 The Poisson Distribution 64 Exercise
More informationAn Introduction to Bayesian Inference and MCMC Methods for Capture-Recapture
An Introduction to Bayesian Inference and MCMC Methods for Capture-Recapture Trinity River Restoration Program Workshop on Outmigration: Population Estimation October 6 8, 2009 An Introduction to Bayesian
More informationStochastic Components of Models
Stochastic Components of Models Gov 2001 Section February 5, 2014 Gov 2001 Section Stochastic Components of Models February 5, 2014 1 / 41 Outline 1 Replication Paper and other logistics 2 Data Generation
More informationHints on Some of the Exercises
Hints on Some of the Exercises of the book R. Seydel: Tools for Computational Finance. Springer, 00/004/006/009/01. Preparatory Remarks: Some of the hints suggest ideas that may simplify solving the exercises
More informationMAS187/AEF258. University of Newcastle upon Tyne
MAS187/AEF258 University of Newcastle upon Tyne 2005-6 Contents 1 Collecting and Presenting Data 5 1.1 Introduction...................................... 5 1.1.1 Examples...................................
More informationPoint Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage
6 Point Estimation Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Point Estimation Statistical inference: directed toward conclusions about one or more parameters. We will use the generic
More informationINSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION
INSTITUTE AND FACULTY OF ACTUARIES Curriculum 2019 SPECIMEN EXAMINATION Subject CS1A Actuarial Statistics Time allowed: Three hours and fifteen minutes INSTRUCTIONS TO THE CANDIDATE 1. Enter all the candidate
More informationIEOR 165 Lecture 1 Probability Review
IEOR 165 Lecture 1 Probability Review 1 Definitions in Probability and Their Consequences 1.1 Defining Probability A probability space (Ω, F, P) consists of three elements: A sample space Ω is the set
More informationChapter 9 Theoretical Probability Models
Making Hard Decisions Chapter 9 Theoretical Probability Models Slide 1 of 47 Theoretical Models Applied Theoretical Probability Models may be used when they describe the physical model "adequately" Examples:
More information12 The Bootstrap and why it works
12 he Bootstrap and why it works For a review of many applications of bootstrap see Efron and ibshirani (1994). For the theory behind the bootstrap see the books by Hall (1992), van der Waart (2000), Lahiri
More informationEE641 Digital Image Processing II: Purdue University VISE - October 29,
EE64 Digital Image Processing II: Purdue University VISE - October 9, 004 The EM Algorithm. Suffient Statistics and Exponential Distributions Let p(y θ) be a family of density functions parameterized by
More informationChapter 5 Discrete Probability Distributions. Random Variables Discrete Probability Distributions Expected Value and Variance
Chapter 5 Discrete Probability Distributions Random Variables Discrete Probability Distributions Expected Value and Variance.40.30.20.10 0 1 2 3 4 Random Variables A random variable is a numerical description
More informationUncertainty Analysis with UNICORN
Uncertainty Analysis with UNICORN D.A.Ababei D.Kurowicka R.M.Cooke D.A.Ababei@ewi.tudelft.nl D.Kurowicka@ewi.tudelft.nl R.M.Cooke@ewi.tudelft.nl Delft Institute for Applied Mathematics Delft University
More informationThe Delta Method. j =.
The Delta Method Often one has one or more MLEs ( 3 and their estimated, conditional sampling variancecovariance matrix. However, there is interest in some function of these estimates. The question is,
More informationMAS1403. Quantitative Methods for Business Management. Semester 1, Module leader: Dr. David Walshaw
MAS1403 Quantitative Methods for Business Management Semester 1, 2018 2019 Module leader: Dr. David Walshaw Additional lecturers: Dr. James Waldren and Dr. Stuart Hall Announcements: Written assignment
More informationCS340 Machine learning Bayesian statistics 3
CS340 Machine learning Bayesian statistics 3 1 Outline Conjugate analysis of µ and σ 2 Bayesian model selection Summarizing the posterior 2 Unknown mean and precision The likelihood function is p(d µ,λ)
More informationDefinition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.
9 Point estimation 9.1 Rationale behind point estimation When sampling from a population described by a pdf f(x θ) or probability function P [X = x θ] knowledge of θ gives knowledge of the entire population.
More informationThe Weibull in R is actually parameterized a fair bit differently from the book. In R, the density for x > 0 is
Weibull in R The Weibull in R is actually parameterized a fair bit differently from the book. In R, the density for x > 0 is f (x) = a b ( x b ) a 1 e (x/b) a This means that a = α in the book s parameterization
More information1. For a special whole life insurance on (x), payable at the moment of death:
**BEGINNING OF EXAMINATION** 1. For a special whole life insurance on (x), payable at the moment of death: µ () t = 0.05, t > 0 (ii) δ = 0.08 x (iii) (iv) The death benefit at time t is bt 0.06t = e, t
More informationDecision theoretic estimation of the ratio of variances in a bivariate normal distribution 1
Decision theoretic estimation of the ratio of variances in a bivariate normal distribution 1 George Iliopoulos Department of Mathematics University of Patras 26500 Rio, Patras, Greece Abstract In this
More informationWrite legibly. Unreadable answers are worthless.
MMF 2021 Final Exam 1 December 2016. This is a closed-book exam: no books, no notes, no calculators, no phones, no tablets, no computers (of any kind) allowed. Do NOT turn this page over until you are
More informationSTAT 111 Recitation 3
STAT 111 Recitation 3 Linjun Zhang stat.wharton.upenn.edu/~linjunz/ September 23, 2017 Misc. The unpicked-up homeworks will be put in the STAT 111 box in the Stats Department lobby (It s on the 4th floor
More informationIntroduction to Probability Theory and Stochastic Processes for Finance Lecture Notes
Introduction to Probability Theory and Stochastic Processes for Finance Lecture Notes Fabio Trojani Department of Economics, University of St. Gallen, Switzerland Correspondence address: Fabio Trojani,
More informationIs a Binomial Process Bayesian?
Is a Binomial Process Bayesian? Robert L. Andrews, Virginia Commonwealth University Department of Management, Richmond, VA. 23284-4000 804-828-7101, rlandrew@vcu.edu Jonathan A. Andrews, United States
More information