STA 114: Statistics. Notes 10. Conjugate Priors

Size: px
Start display at page:

Download "STA 114: Statistics. Notes 10. Conjugate Priors"

Transcription

1 STA 114: Statistics Notes 10. Conjugate Priors Conjugate family Once we get a /pmf ξ(θ x) by combining a model X f(x θ) with a /pmf ξ(θ) on θ Θ, a report can be made by summarizing the. It helps to have the /pmf in a recognizable form so that we can easily compute its mean, spread, quantiles etc. This is not guaranteed to happen in general. For example, the model X Binomial(n, p) and the ξ(p) = e p /(e 1), p [0, 1] lead to the ξ(p x) = const p x (1 p) nx e p, p [0, 1], with a constant term that is fairly difficult to compute, making it difficult to get summaries of this (recall Lab 4). However, for certain models, certain s do lead to s that are analytically tractable. We already saw one in the female birth rate analysis, a uniform for the binomial model gives a beta. In fact more is true for this model: any beta leads to a beta! This phenomenon is called conjugacy. A formal definition is given below. Definition 1 (Conjugacy). A collection of s (or pmfs) is called a conjugate family for a model X f(x θ), θ Θ, if whenever a ξ(θ) is chosen from the collection, it leads to a ξ(θ x) that is also a member of the collection, for every observation X = x. Conjugacy in itself is not a very useful property. For example the collection of all s on Θ is surely conjugate to the model. It becomes useful when a small collection of s exhibit conjugacy to a certain statistical model. By a small collection we usually mean a collection of s/pmfs G = {g(θ a) : a A} indexed by a low-dimensional vector a. G is conjugate to a statistical model X f(x θ) if ξ(θ) = g(θ a) for some a A means for every x, ξ(θ x) = g(θ a ) for some a A. As mentioned before, the collection of beta s {Beta(a, b) : a > 0, b > 0} is a (-dimensional) conjugate family to the binomial model X Binomial(n, p), p [0, 1]. Table?? below gives a list of other common models with known, low-dimensional conjugate families. We will establish conjugacy for three of the listed models; you re required to do the maths for the remaining ones in HW5. The binomial-beta conjugacy The of bet(a, b) distribution, for a > 0, b > 0. equals g(p) = p a1 (1 p) b1 /B(a, b), p [0, 1], where B(a, b) = 1 0 qa1 (1 q) b1 dq is known as the Beta function. If we take ξ(p) = Beta(a, b) (for some a > 0, b > 0) as the for a binomial model X Binomial(n, p), p [0, 1], then for any observations x {0, 1,, n}, ξ(p x) = const f(x p)ξ(p) = const p x (1p) nx p a1 (1p) b1 = const p a+x1 (1p) b+nx1. 1

2 Model Parameter Prior Posterior X Binomial(n, p) 0 p 1 Beta(a, b) Beta(a, b ) a > 0, b > 0 a = a + x b = b + n x X = (X 1,, X n ) λ > 0 Gamma(a, b) Gamma(a, b ) X i Poisson(λ) a > 0, b > 0 a = a + n x b = b + n X = (X 1,, X n ) λ > 0 Gamma(a, b) Gamma(a, b ) X i Exponential(λ) a > 0, b > 0 a = a + n b = b + n x X = (X 1,, X n ) < < Normal(a, b ) Normal(a, b ) Normal(, σ ) < a < a = nb x+σ a nb +σ σ known b > 0 b = σ b nb +σ X = (X 1,, X n ) < < Nχ (m, k, r, s) Nχ (m, k, r, s ) Normal(, σ ) σ > 0 < m < m = km+n x X i X i k > 0, r > 0, s > 0 s = k+n k = k + n r = r + n rs+ kn k+n ( xm) +(n1)s x r+n Table 1: Conjugate and for some common models. But the of Beta(a+x, b+nx) (note: a+x > 0, b+nx > 0) is p a+x1 (1p) b+nx1 /B(a+ x, b + n x). Therefore ξ(p x) is a constant multiple of the Beta(a + x, b + n x). But if two s are constant multiples of each other, they must be identical (and the constant must be 1). So ξ(p x) = Beta(a + x, b + n x). The normal-normal conjugacy Next we show that for the model X 1,, X n Normal(, σ ), (, ), σ fixed, the ξ() = Normal(a, b ) gives a ξ( x) = Normal(a, b ) for some a and b [which we shall identify]. It suffices to show that ξ( x) is a constant multiple of the Normal(a, b ) density. This is equivalent to showing log ξ( x) = const + ( a ) b

3 by going in the log-scale. Now, by definition log ξ( x) = const + l x () + log ξ() n i=1 = const (x i ) ( a) σ b n i=1 = const (x i x) + n( x ) σ = const 1 [ ] n( x ) ( a) + σ b = const 1 ( nb x+σ a nb +σ ) b σ nb +σ ( a) b and therefore, ξ( x) = Normal( nb x+σ a completion of squares identity (give it a try!): b, σ ). nb +σ nb +σ The last equality above follows from a n( x ) ( a) + σ b = (nb + σ nb x+σa )( ) nb +σ + b σ A conjugate family for the full normal model n( x a) nb + σ. For the full normal model X 1,, X n Normal(, σ ), (, σ ) (, ) (0, ), we need a bivariate ξ(, σ ) on (, ) (0, ). There are several choices here. For example we could take ξ(, σ ) = g()h(σ ) where g() is a on (, ) and h(σ ) is a on (0, ). This is in fact widely used, with g() usually taken to be a normal and h(σ ) taken to be an inverse-gamma (i.e., the of 1/σ is a gamma ). However, the family of such s are not conjugate to the model. In particular, the ξ(, σ x) does not factor into a product g( x)h(σ x). There is however, a conjugate family of s, known as the normal-inverse-chi-square s, denotes Nχ (m, k, r, s) with parameters m (, ), k > 0, r > 0 and s > 0. The of this distributionis given by: g(w, v) = const. v r+3 exp ( k(w m) + rs v where the constant equals (rs/)r/ πγ(r/). Here are two important results. ), (w, v) (, ) (0, ) Result 1. A pair of random variables (W, V ) has a Nχ (m, k, r, s) if and only if 1. rs V χ (r), and. [W V = v] Normal(m, v/k). Result. If (W, V ) Nχ (m, k, r, s) then W m t(r). s/k 3

4 The first result can be proved by direct calculations (you re welcome to try it). The second result follows once you note that if (W, V ) Nχ (m, k, r, s) then by Result 1, we can write W = m + V/kZ where Z Normal(0, 1) is independent of V. Also U = rs V χ (r). Then, W m V/kZ = = Z t(r). s/k s/k U/r With this as the prelude, we re now ready to prove that for ξ(, σ ) = Nχ (m, k, r, s) we get ξ(, σ x) = Nχ (m, k (s), r, s ) where m = km+n x k+n k = n + k r = n + r s = 1 kn (rs + ( x n+r k+n m) + (n 1)s x). Working in the log-scale we have log ξ(, σ x) = l x (, σ ) + log ξ(, σ ) = const n n log σ i=1 (x i ) r + 3 log σ k( m) + rs σ σ = const n + r + 3 n log σ i=1 (x i ) + k( m) + rs σ = const n + r + 3 log σ (n 1)s x + n( x ) + k( m) + rs. σ There is another square-completion identity that we need (try this): ( n( x ) + k( m) = (k + n) Plugging this in we get log ξ(, σ x) = const r + 3 and hence ξ(, σ x) = Nχ (m, k, r, s ). Putting it to use ) km + n x + kn k + n k + n ( x m). log σ k ( m ) + r s σ It is nice to have low-dimensional conjugate family for a statistical model one is interested in. But we still have the job of deciding upon one member of the family to ultimately use as the /pmf. If a singular choice is hard to justify, we can run the analysis for a multitude of reasonable choices and present the analysis side by side. This can be done both graphically, and more important via tables. To summarize a or a, we can report a number representing the center of the, such as the mean or the median. To summarize the spread of the, we can 4

5 Prior Prior summary Posterior Posterior summary Mean [.05,.975 ] Mean [.05 (x),.975 (x)] Gamma(,.) 10 (1.1, 7.86) Gamma(15, 10.) 14.9 (1.63, 17.36) Gamma(5,.5) 10 (3.5, 0.48) Gamma(155, 10.5) (1.53, 17.17) Gamma(50, 5) 10 (7.4, 1.96) Gamma(00, 15) (11.55, 15.4) Gamma(500, 50) 10 (9.14, 10.9) Gamma(650, 60) (10.0, 11.68) Table : Poisson-gamma example, and summaries report an interval [a, b] such that the packs most of its area inside the interval. If we want an 1 α area inside, for some small α, then we can take a = the α/-th quantile of the and b = the (1 α/)-th quantile of it. For a ξ(θ), we ll denote its u-th quantile by θ u and for a ξ(θ x) we will use the notation θ u (x). We will usually work with α = 5% and hence report [θ.05, θ.975 ] from the and [θ.05 (x), θ.975 (x)] from the. Example. Consider hurricane counts X 1,, X n for n = 10 consecutive years in the north Atlantic basin, modeled as X i Poisson(). Figure?? and Table?? show - summaries for 4 choices of the from the conjugate family of gamma s. The observed data are x = (1, 14, 15, 1, 16, 14, 7, 10, 14, 16). 5

6 a =, b = 0. a = 5, b = a = 50, b = 5 a = 500, b = Figure 1: Prior- summary for hurricane count data with Gamma(a, b) conjugate s. 6

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections 1 / 40 Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: Chapter 7 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods:

More information

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections 1 / 31 : Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods: 7.5 Maximum Likelihood

More information

MATH 3200 Exam 3 Dr. Syring

MATH 3200 Exam 3 Dr. Syring . Suppose n eligible voters are polled (randomly sampled) from a population of size N. The poll asks voters whether they support or do not support increasing local taxes to fund public parks. Let M be

More information

Exam 2 Spring 2015 Statistics for Applications 4/9/2015

Exam 2 Spring 2015 Statistics for Applications 4/9/2015 18.443 Exam 2 Spring 2015 Statistics for Applications 4/9/2015 1. True or False (and state why). (a). The significance level of a statistical test is not equal to the probability that the null hypothesis

More information

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ. Sufficient Statistics Lecture Notes 6 Sufficiency Data reduction in terms of a particular statistic can be thought of as a partition of the sample space X. Definition T is sufficient for θ if the conditional

More information

Common one-parameter models

Common one-parameter models Common one-parameter models In this section we will explore common one-parameter models, including: 1. Binomial data with beta prior on the probability 2. Poisson data with gamma prior on the rate 3. Gaussian

More information

Conjugate Models. Patrick Lam

Conjugate Models. Patrick Lam Conjugate Models Patrick Lam Outline Conjugate Models What is Conjugacy? The Beta-Binomial Model The Normal Model Normal Model with Unknown Mean, Known Variance Normal Model with Known Mean, Unknown Variance

More information

Conjugate priors: Beta and normal Class 15, Jeremy Orloff and Jonathan Bloom

Conjugate priors: Beta and normal Class 15, Jeremy Orloff and Jonathan Bloom 1 Learning Goals Conjugate s: Beta and normal Class 15, 18.05 Jeremy Orloff and Jonathan Bloom 1. Understand the benefits of conjugate s.. Be able to update a beta given a Bernoulli, binomial, or geometric

More information

Random Variables Handout. Xavier Vilà

Random Variables Handout. Xavier Vilà Random Variables Handout Xavier Vilà Course 2004-2005 1 Discrete Random Variables. 1.1 Introduction 1.1.1 Definition of Random Variable A random variable X is a function that maps each possible outcome

More information

Two hours UNIVERSITY OF MANCHESTER. 23 May :00 16:00. Answer ALL SIX questions The total number of marks in the paper is 90.

Two hours UNIVERSITY OF MANCHESTER. 23 May :00 16:00. Answer ALL SIX questions The total number of marks in the paper is 90. Two hours MATH39542 UNIVERSITY OF MANCHESTER RISK THEORY 23 May 2016 14:00 16:00 Answer ALL SIX questions The total number of marks in the paper is 90. University approved calculators may be used 1 of

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS Answer any FOUR of the SIX questions.

More information

MVE051/MSG Lecture 7

MVE051/MSG Lecture 7 MVE051/MSG810 2017 Lecture 7 Petter Mostad Chalmers November 20, 2017 The purpose of collecting and analyzing data Purpose: To build and select models for parts of the real world (which can be used for

More information

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence continuous rv Let X be a continuous rv. Then a probability distribution or probability density function (pdf) of X is a function f(x) such that for any two numbers a and b with a b, P(a X b) = b a f (x)dx.

More information

Subject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018

Subject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018 ` Subject CS1 Actuarial Statistics 1 Core Principles Syllabus for the 2019 exams 1 June 2018 Copyright in this Core Reading is the property of the Institute and Faculty of Actuaries who are the sole distributors.

More information

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions Frequentist Methods: 7.5 Maximum Likelihood Estimators

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics March 12, 2018 CS 361: Probability & Statistics Inference Binomial likelihood: Example Suppose we have a coin with an unknown probability of heads. We flip the coin 10 times and observe 2 heads. What can

More information

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as Lecture 0 on BST 63: Statistical Theory I Kui Zhang, 09/9/008 Review for the previous lecture Definition: Several continuous distributions, including uniform, gamma, normal, Beta, Cauchy, double exponential

More information

The Bernoulli distribution

The Bernoulli distribution This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems.

Practice Exercises for Midterm Exam ST Statistical Theory - II The ACTUAL exam will consists of less number of problems. Practice Exercises for Midterm Exam ST 522 - Statistical Theory - II The ACTUAL exam will consists of less number of problems. 1. Suppose X i F ( ) for i = 1,..., n, where F ( ) is a strictly increasing

More information

Posterior Inference. , where should we start? Consider the following computational procedure: 1. draw samples. 2. convert. 3. compute properties

Posterior Inference. , where should we start? Consider the following computational procedure: 1. draw samples. 2. convert. 3. compute properties Posterior Inference Example. Consider a binomial model where we have a posterior distribution for the probability term, θ. Suppose we want to make inferences about the log-odds γ = log ( θ 1 θ), where

More information

Bayesian Linear Model: Gory Details

Bayesian Linear Model: Gory Details Bayesian Linear Model: Gory Details Pubh7440 Notes By Sudipto Banerjee Let y y i ] n i be an n vector of independent observations on a dependent variable (or response) from n experimental units. Associated

More information

Chapter 8: Sampling distributions of estimators Sections

Chapter 8: Sampling distributions of estimators Sections Chapter 8 continued Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample

More information

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi Chapter 4: Commonly Used Distributions Statistics for Engineers and Scientists Fourth Edition William Navidi 2014 by Education. This is proprietary material solely for authorized instructor use. Not authorized

More information

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions April 9th, 2018 Lecture 20: Special distributions Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters 4, 6: Random variables Week 9 Chapter

More information

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is: **BEGINNING OF EXAMINATION** 1. You are given: (i) A random sample of five observations from a population is: 0.2 0.7 0.9 1.1 1.3 (ii) You use the Kolmogorov-Smirnov test for testing the null hypothesis,

More information

Generating Random Numbers

Generating Random Numbers Generating Random Numbers Aim: produce random variables for given distribution Inverse Method Let F be the distribution function of an univariate distribution and let F 1 (y) = inf{x F (x) y} (generalized

More information

Non-informative Priors Multiparameter Models

Non-informative Priors Multiparameter Models Non-informative Priors Multiparameter Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Prior Types Informative vs Non-informative There has been a desire for a prior distributions that

More information

Part II: Computation for Bayesian Analyses

Part II: Computation for Bayesian Analyses Part II: Computation for Bayesian Analyses 62 BIO 233, HSPH Spring 2015 Conjugacy In both birth weight eamples the posterior distribution is from the same family as the prior: Prior Likelihood Posterior

More information

Commonly Used Distributions

Commonly Used Distributions Chapter 4: Commonly Used Distributions 1 Introduction Statistical inference involves drawing a sample from a population and analyzing the sample data to learn about the population. We often have some knowledge

More information

Chapter 5. Statistical inference for Parametric Models

Chapter 5. Statistical inference for Parametric Models Chapter 5. Statistical inference for Parametric Models Outline Overview Parameter estimation Method of moments How good are method of moments estimates? Interval estimation Statistical Inference for Parametric

More information

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. Random Variables 2 A random variable X is a numerical (integer, real, complex, vector etc.) summary of the outcome of the random experiment.

More information

An Improved Skewness Measure

An Improved Skewness Measure An Improved Skewness Measure Richard A. Groeneveld Professor Emeritus, Department of Statistics Iowa State University ragroeneveld@valley.net Glen Meeden School of Statistics University of Minnesota Minneapolis,

More information

Practice Exam 1. Loss Amount Number of Losses

Practice Exam 1. Loss Amount Number of Losses Practice Exam 1 1. You are given the following data on loss sizes: An ogive is used as a model for loss sizes. Determine the fitted median. Loss Amount Number of Losses 0 1000 5 1000 5000 4 5000 10000

More information

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions. ME3620 Theory of Engineering Experimentation Chapter III. Random Variables and Probability Distributions Chapter III 1 3.2 Random Variables In an experiment, a measurement is usually denoted by a variable

More information

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom Review for Final Exam 18.05 Spring 2014 Jeremy Orloff and Jonathan Bloom THANK YOU!!!! JON!! PETER!! RUTHI!! ERIKA!! ALL OF YOU!!!! Probability Counting Sets Inclusion-exclusion principle Rule of product

More information

Chapter 3 Discrete Random Variables and Probability Distributions

Chapter 3 Discrete Random Variables and Probability Distributions Chapter 3 Discrete Random Variables and Probability Distributions Part 4: Special Discrete Random Variable Distributions Sections 3.7 & 3.8 Geometric, Negative Binomial, Hypergeometric NOTE: The discrete

More information

6. Genetics examples: Hardy-Weinberg Equilibrium

6. Genetics examples: Hardy-Weinberg Equilibrium PBCB 206 (Fall 2006) Instructor: Fei Zou email: fzou@bios.unc.edu office: 3107D McGavran-Greenberg Hall Lecture 4 Topics for Lecture 4 1. Parametric models and estimating parameters from data 2. Method

More information

Introduction Recently the importance of modelling dependent insurance and reinsurance risks has attracted the attention of actuarial practitioners and

Introduction Recently the importance of modelling dependent insurance and reinsurance risks has attracted the attention of actuarial practitioners and Asymptotic dependence of reinsurance aggregate claim amounts Mata, Ana J. KPMG One Canada Square London E4 5AG Tel: +44-207-694 2933 e-mail: ana.mata@kpmg.co.uk January 26, 200 Abstract In this paper we

More information

ST440/550: Applied Bayesian Analysis. (5) Multi-parameter models - Summarizing the posterior

ST440/550: Applied Bayesian Analysis. (5) Multi-parameter models - Summarizing the posterior (5) Multi-parameter models - Summarizing the posterior Models with more than one parameter Thus far we have studied single-parameter models, but most analyses have several parameters For example, consider

More information

Confidence Intervals Introduction

Confidence Intervals Introduction Confidence Intervals Introduction A point estimate provides no information about the precision and reliability of estimation. For example, the sample mean X is a point estimate of the population mean μ

More information

M249 Diagnostic Quiz

M249 Diagnostic Quiz THE OPEN UNIVERSITY Faculty of Mathematics and Computing M249 Diagnostic Quiz Prepared by the Course Team [Press to begin] c 2005, 2006 The Open University Last Revision Date: May 19, 2006 Version 4.2

More information

Homework Problems Stat 479

Homework Problems Stat 479 Chapter 10 91. * A random sample, X1, X2,, Xn, is drawn from a distribution with a mean of 2/3 and a variance of 1/18. ˆ = (X1 + X2 + + Xn)/(n-1) is the estimator of the distribution mean θ. Find MSE(

More information

START HERE: Instructions. 1 Exponential Family [Zhou, Manzil]

START HERE: Instructions. 1 Exponential Family [Zhou, Manzil] START HERE: Instructions Thanks a lot to John A.W.B. Constanzo and Shi Zong for providing and allowing to use the latex source files for quick preparation of the HW solution. The homework was due at 9:00am

More information

1. You are given the following information about a stationary AR(2) model:

1. You are given the following information about a stationary AR(2) model: Fall 2003 Society of Actuaries **BEGINNING OF EXAMINATION** 1. You are given the following information about a stationary AR(2) model: (i) ρ 1 = 05. (ii) ρ 2 = 01. Determine φ 2. (A) 0.2 (B) 0.1 (C) 0.4

More information

Data Analytics (CS40003) Practice Set IV (Topic: Probability and Sampling Distribution)

Data Analytics (CS40003) Practice Set IV (Topic: Probability and Sampling Distribution) Data Analytics (CS40003) Practice Set IV (Topic: Probability and Sampling Distribution) I. Concept Questions 1. Give an example of a random variable in the context of Drawing a card from a deck of cards.

More information

Chapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables

Chapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables Chapter 5 Continuous Random Variables and Probability Distributions 5.1 Continuous Random Variables 1 2CHAPTER 5. CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Probability Distributions Probability

More information

Probability. An intro for calculus students P= Figure 1: A normal integral

Probability. An intro for calculus students P= Figure 1: A normal integral Probability An intro for calculus students.8.6.4.2 P=.87 2 3 4 Figure : A normal integral Suppose we flip a coin 2 times; what is the probability that we get more than 2 heads? Suppose we roll a six-sided

More information

Simple Random Sample

Simple Random Sample Simple Random Sample A simple random sample (SRS) of size n consists of n elements from the population chosen in such a way that every set of n elements has an equal chance to be the sample actually selected.

More information

Using Agent Belief to Model Stock Returns

Using Agent Belief to Model Stock Returns Using Agent Belief to Model Stock Returns America Holloway Department of Computer Science University of California, Irvine, Irvine, CA ahollowa@ics.uci.edu Introduction It is clear that movements in stock

More information

Chapter 3 Discrete Random Variables and Probability Distributions

Chapter 3 Discrete Random Variables and Probability Distributions Chapter 3 Discrete Random Variables and Probability Distributions Part 3: Special Discrete Random Variable Distributions Section 3.5 Discrete Uniform Section 3.6 Bernoulli and Binomial Others sections

More information

Rules and Models 1 investigates the internal measurement approach for operational risk capital

Rules and Models 1 investigates the internal measurement approach for operational risk capital Carol Alexander 2 Rules and Models Rules and Models 1 investigates the internal measurement approach for operational risk capital 1 There is a view that the new Basel Accord is being defined by a committee

More information

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg :

More information

Supplementary Material: Strategies for exploration in the domain of losses

Supplementary Material: Strategies for exploration in the domain of losses 1 Supplementary Material: Strategies for exploration in the domain of losses Paul M. Krueger 1,, Robert C. Wilson 2,, and Jonathan D. Cohen 3,4 1 Department of Psychology, University of California, Berkeley

More information

Statistical estimation

Statistical estimation Statistical estimation Statistical modelling: theory and practice Gilles Guillot gigu@dtu.dk September 3, 2013 Gilles Guillot (gigu@dtu.dk) Estimation September 3, 2013 1 / 27 1 Introductory example 2

More information

This is a open-book exam. Assigned: Friday November 27th 2009 at 16:00. Due: Monday November 30th 2009 before 10:00.

This is a open-book exam. Assigned: Friday November 27th 2009 at 16:00. Due: Monday November 30th 2009 before 10:00. University of Iceland School of Engineering and Sciences Department of Industrial Engineering, Mechanical Engineering and Computer Science IÐN106F Industrial Statistics II - Bayesian Data Analysis Fall

More information

Describing Uncertain Variables

Describing Uncertain Variables Describing Uncertain Variables L7 Uncertainty in Variables Uncertainty in concepts and models Uncertainty in variables Lack of precision Lack of knowledge Variability in space/time Describing Uncertainty

More information

STA 532: Theory of Statistical Inference

STA 532: Theory of Statistical Inference STA 532: Theory of Statistical Inference Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 2 Estimating CDFs and Statistical Functionals Empirical CDFs Let {X i : i n}

More information

CS 237: Probability in Computing

CS 237: Probability in Computing CS 237: Probability in Computing Wayne Snyder Computer Science Department Boston University Lecture 12: Continuous Distributions Uniform Distribution Normal Distribution (motivation) Discrete vs Continuous

More information

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved.

4-1. Chapter 4. Commonly Used Distributions by The McGraw-Hill Companies, Inc. All rights reserved. 4-1 Chapter 4 Commonly Used Distributions 2014 by The Companies, Inc. All rights reserved. Section 4.1: The Bernoulli Distribution 4-2 We use the Bernoulli distribution when we have an experiment which

More information

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation IEOR E4703: Monte-Carlo Simulation Generating Random Variables and Stochastic Processes Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Bayesian course - problem set 3 (lecture 4)

Bayesian course - problem set 3 (lecture 4) Bayesian course - problem set 3 (lecture 4) Ben Lambert November 14, 2016 1 Ticked off Imagine once again that you are investigating the occurrence of Lyme disease in the UK. This is a vector-borne disease

More information

Chapter 6. Importance sampling. 6.1 The basics

Chapter 6. Importance sampling. 6.1 The basics Chapter 6 Importance sampling 6.1 The basics To movtivate our discussion consider the following situation. We want to use Monte Carlo to compute µ E[X]. There is an event E such that P(E) is small but

More information

Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is

Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is Normal Distribution Normal Distribution Definition A continuous rv X is said to have a normal distribution with parameter µ and σ (µ and σ 2 ), where < µ < and σ > 0, if the pdf of X is f (x; µ, σ) = 1

More information

Random variables. Contents

Random variables. Contents Random variables Contents 1 Random Variable 2 1.1 Discrete Random Variable............................ 3 1.2 Continuous Random Variable........................... 5 1.3 Measures of Location...............................

More information

BIOL The Normal Distribution and the Central Limit Theorem

BIOL The Normal Distribution and the Central Limit Theorem BIOL 300 - The Normal Distribution and the Central Limit Theorem In the first week of the course, we introduced a few measures of center and spread, and discussed how the mean and standard deviation are

More information

Modelling of Long-Term Risk

Modelling of Long-Term Risk Modelling of Long-Term Risk Roger Kaufmann Swiss Life roger.kaufmann@swisslife.ch 15th International AFIR Colloquium 6-9 September 2005, Zurich c 2005 (R. Kaufmann, Swiss Life) Contents A. Basel II B.

More information

Some Discrete Distribution Families

Some Discrete Distribution Families Some Discrete Distribution Families ST 370 Many families of discrete distributions have been studied; we shall discuss the ones that are most commonly found in applications. In each family, we need a formula

More information

Outline. Review Continuation of exercises from last time

Outline. Review Continuation of exercises from last time Bayesian Models II Outline Review Continuation of exercises from last time 2 Review of terms from last time Probability density function aka pdf or density Likelihood function aka likelihood Conditional

More information

GPD-POT and GEV block maxima

GPD-POT and GEV block maxima Chapter 3 GPD-POT and GEV block maxima This chapter is devoted to the relation between POT models and Block Maxima (BM). We only consider the classical frameworks where POT excesses are assumed to be GPD,

More information

Bayesian Normal Stuff

Bayesian Normal Stuff Bayesian Normal Stuff - Set-up of the basic model of a normally distributed random variable with unknown mean and variance (a two-parameter model). - Discuss philosophies of prior selection - Implementation

More information

Statistics 431 Spring 2007 P. Shaman. Preliminaries

Statistics 431 Spring 2007 P. Shaman. Preliminaries Statistics 4 Spring 007 P. Shaman The Binomial Distribution Preliminaries A binomial experiment is defined by the following conditions: A sequence of n trials is conducted, with each trial having two possible

More information

Exam STAM Practice Exam #1

Exam STAM Practice Exam #1 !!!! Exam STAM Practice Exam #1 These practice exams should be used during the month prior to your exam. This practice exam contains 20 questions, of equal value, corresponding to about a 2 hour exam.

More information

TOPIC: PROBABILITY DISTRIBUTIONS

TOPIC: PROBABILITY DISTRIBUTIONS TOPIC: PROBABILITY DISTRIBUTIONS There are two types of random variables: A Discrete random variable can take on only specified, distinct values. A Continuous random variable can take on any value within

More information

SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS

SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS Questions 1-307 have been taken from the previous set of Exam C sample questions. Questions no longer relevant

More information

1 Bayesian Bias Correction Model

1 Bayesian Bias Correction Model 1 Bayesian Bias Correction Model Assuming that n iid samples {X 1,...,X n }, were collected from a normal population with mean µ and variance σ 2. The model likelihood has the form, P( X µ, σ 2, T n >

More information

Basic Procedure for Histograms

Basic Procedure for Histograms Basic Procedure for Histograms 1. Compute the range of observations (min. & max. value) 2. Choose an initial # of classes (most likely based on the range of values, try and find a number of classes that

More information

PROBABILITY DISTRIBUTIONS

PROBABILITY DISTRIBUTIONS CHAPTER 3 PROBABILITY DISTRIBUTIONS Page Contents 3.1 Introduction to Probability Distributions 51 3.2 The Normal Distribution 56 3.3 The Binomial Distribution 60 3.4 The Poisson Distribution 64 Exercise

More information

An Introduction to Bayesian Inference and MCMC Methods for Capture-Recapture

An Introduction to Bayesian Inference and MCMC Methods for Capture-Recapture An Introduction to Bayesian Inference and MCMC Methods for Capture-Recapture Trinity River Restoration Program Workshop on Outmigration: Population Estimation October 6 8, 2009 An Introduction to Bayesian

More information

Stochastic Components of Models

Stochastic Components of Models Stochastic Components of Models Gov 2001 Section February 5, 2014 Gov 2001 Section Stochastic Components of Models February 5, 2014 1 / 41 Outline 1 Replication Paper and other logistics 2 Data Generation

More information

Hints on Some of the Exercises

Hints on Some of the Exercises Hints on Some of the Exercises of the book R. Seydel: Tools for Computational Finance. Springer, 00/004/006/009/01. Preparatory Remarks: Some of the hints suggest ideas that may simplify solving the exercises

More information

MAS187/AEF258. University of Newcastle upon Tyne

MAS187/AEF258. University of Newcastle upon Tyne MAS187/AEF258 University of Newcastle upon Tyne 2005-6 Contents 1 Collecting and Presenting Data 5 1.1 Introduction...................................... 5 1.1.1 Examples...................................

More information

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage 6 Point Estimation Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Point Estimation Statistical inference: directed toward conclusions about one or more parameters. We will use the generic

More information

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION INSTITUTE AND FACULTY OF ACTUARIES Curriculum 2019 SPECIMEN EXAMINATION Subject CS1A Actuarial Statistics Time allowed: Three hours and fifteen minutes INSTRUCTIONS TO THE CANDIDATE 1. Enter all the candidate

More information

IEOR 165 Lecture 1 Probability Review

IEOR 165 Lecture 1 Probability Review IEOR 165 Lecture 1 Probability Review 1 Definitions in Probability and Their Consequences 1.1 Defining Probability A probability space (Ω, F, P) consists of three elements: A sample space Ω is the set

More information

Chapter 9 Theoretical Probability Models

Chapter 9 Theoretical Probability Models Making Hard Decisions Chapter 9 Theoretical Probability Models Slide 1 of 47 Theoretical Models Applied Theoretical Probability Models may be used when they describe the physical model "adequately" Examples:

More information

12 The Bootstrap and why it works

12 The Bootstrap and why it works 12 he Bootstrap and why it works For a review of many applications of bootstrap see Efron and ibshirani (1994). For the theory behind the bootstrap see the books by Hall (1992), van der Waart (2000), Lahiri

More information

EE641 Digital Image Processing II: Purdue University VISE - October 29,

EE641 Digital Image Processing II: Purdue University VISE - October 29, EE64 Digital Image Processing II: Purdue University VISE - October 9, 004 The EM Algorithm. Suffient Statistics and Exponential Distributions Let p(y θ) be a family of density functions parameterized by

More information

Chapter 5 Discrete Probability Distributions. Random Variables Discrete Probability Distributions Expected Value and Variance

Chapter 5 Discrete Probability Distributions. Random Variables Discrete Probability Distributions Expected Value and Variance Chapter 5 Discrete Probability Distributions Random Variables Discrete Probability Distributions Expected Value and Variance.40.30.20.10 0 1 2 3 4 Random Variables A random variable is a numerical description

More information

Uncertainty Analysis with UNICORN

Uncertainty Analysis with UNICORN Uncertainty Analysis with UNICORN D.A.Ababei D.Kurowicka R.M.Cooke D.A.Ababei@ewi.tudelft.nl D.Kurowicka@ewi.tudelft.nl R.M.Cooke@ewi.tudelft.nl Delft Institute for Applied Mathematics Delft University

More information

The Delta Method. j =.

The Delta Method. j =. The Delta Method Often one has one or more MLEs ( 3 and their estimated, conditional sampling variancecovariance matrix. However, there is interest in some function of these estimates. The question is,

More information

MAS1403. Quantitative Methods for Business Management. Semester 1, Module leader: Dr. David Walshaw

MAS1403. Quantitative Methods for Business Management. Semester 1, Module leader: Dr. David Walshaw MAS1403 Quantitative Methods for Business Management Semester 1, 2018 2019 Module leader: Dr. David Walshaw Additional lecturers: Dr. James Waldren and Dr. Stuart Hall Announcements: Written assignment

More information

CS340 Machine learning Bayesian statistics 3

CS340 Machine learning Bayesian statistics 3 CS340 Machine learning Bayesian statistics 3 1 Outline Conjugate analysis of µ and σ 2 Bayesian model selection Summarizing the posterior 2 Unknown mean and precision The likelihood function is p(d µ,λ)

More information

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ. 9 Point estimation 9.1 Rationale behind point estimation When sampling from a population described by a pdf f(x θ) or probability function P [X = x θ] knowledge of θ gives knowledge of the entire population.

More information

The Weibull in R is actually parameterized a fair bit differently from the book. In R, the density for x > 0 is

The Weibull in R is actually parameterized a fair bit differently from the book. In R, the density for x > 0 is Weibull in R The Weibull in R is actually parameterized a fair bit differently from the book. In R, the density for x > 0 is f (x) = a b ( x b ) a 1 e (x/b) a This means that a = α in the book s parameterization

More information

1. For a special whole life insurance on (x), payable at the moment of death:

1. For a special whole life insurance on (x), payable at the moment of death: **BEGINNING OF EXAMINATION** 1. For a special whole life insurance on (x), payable at the moment of death: µ () t = 0.05, t > 0 (ii) δ = 0.08 x (iii) (iv) The death benefit at time t is bt 0.06t = e, t

More information

Decision theoretic estimation of the ratio of variances in a bivariate normal distribution 1

Decision theoretic estimation of the ratio of variances in a bivariate normal distribution 1 Decision theoretic estimation of the ratio of variances in a bivariate normal distribution 1 George Iliopoulos Department of Mathematics University of Patras 26500 Rio, Patras, Greece Abstract In this

More information

Write legibly. Unreadable answers are worthless.

Write legibly. Unreadable answers are worthless. MMF 2021 Final Exam 1 December 2016. This is a closed-book exam: no books, no notes, no calculators, no phones, no tablets, no computers (of any kind) allowed. Do NOT turn this page over until you are

More information

STAT 111 Recitation 3

STAT 111 Recitation 3 STAT 111 Recitation 3 Linjun Zhang stat.wharton.upenn.edu/~linjunz/ September 23, 2017 Misc. The unpicked-up homeworks will be put in the STAT 111 box in the Stats Department lobby (It s on the 4th floor

More information

Introduction to Probability Theory and Stochastic Processes for Finance Lecture Notes

Introduction to Probability Theory and Stochastic Processes for Finance Lecture Notes Introduction to Probability Theory and Stochastic Processes for Finance Lecture Notes Fabio Trojani Department of Economics, University of St. Gallen, Switzerland Correspondence address: Fabio Trojani,

More information

Is a Binomial Process Bayesian?

Is a Binomial Process Bayesian? Is a Binomial Process Bayesian? Robert L. Andrews, Virginia Commonwealth University Department of Management, Richmond, VA. 23284-4000 804-828-7101, rlandrew@vcu.edu Jonathan A. Andrews, United States

More information