Modelling Environmental Extremes

Similar documents
Modelling Environmental Extremes

An Introduction to Statistical Extreme Value Theory

Measuring Financial Risk using Extreme Value Theory: evidence from Pakistan

Financial Risk Forecasting Chapter 9 Extreme Value Theory

Generalized MLE per Martins and Stedinger

AN EXTREME VALUE APPROACH TO PRICING CREDIT RISK

EVA Tutorial #1 BLOCK MAXIMA APPROACH IN HYDROLOGIC/CLIMATE APPLICATIONS. Rick Katz

Estimate of Maximum Insurance Loss due to Bushfires

GPD-POT and GEV block maxima

Analysis of extreme values with random location Abstract Keywords: 1. Introduction and Model

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

NEWCASTLE UNIVERSITY SCHOOL OF MATHEMATICS & STATISTICS SEMESTER /2013 MAS8304. Environmental Extremes: Mid semester test

Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making

Introduction to Algorithmic Trading Strategies Lecture 8

Simulation of Extreme Events in the Presence of Spatial Dependence

A New Hybrid Estimation Method for the Generalized Pareto Distribution

Modelling catastrophic risk in international equity markets: An extreme value approach. JOHN COTTER University College Dublin

An Application of Extreme Value Theory for Measuring Financial Risk in the Uruguayan Pension Fund 1

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -5 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

Continuous random variables

QUANTIFYING THE RISK OF EXTREME EVENTS IN A CHANGING CLIMATE. Rick Katz. Joint Work with Holger Rootzén Chalmers and Gothenburg University, Sweden

Mongolia s TOP-20 Index Risk Analysis, Pt. 3

Lecture 3: Probability Distributions (cont d)

QQ PLOT Yunsi Wang, Tyler Steele, Eva Zhang Spring 2016

Random Variables and Probability Distributions

Chapter 8: Sampling distributions of estimators Sections

Generalized Additive Modelling for Sample Extremes: An Environmental Example

Business Statistics 41000: Probability 3

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

Stochastic model of flow duration curves for selected rivers in Bangladesh

John Cotter and Kevin Dowd

SYSM 6304 Risk and Decision Analysis Lecture 2: Fitting Distributions to Data

Describing Uncertain Variables

2002 Statistical Research Center for Complex Systems International Statistical Workshop 19th & 20th June 2002 Seoul National University

Scaling conditional tail probability and quantile estimators

Modelling insured catastrophe losses

Probability Weighted Moments. Andrew Smith

Exam 2 Spring 2015 Statistics for Applications 4/9/2015

Frequency Distribution Models 1- Probability Density Function (PDF)

An Application of Extreme Value Theory for Measuring Risk

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

IEOR E4602: Quantitative Risk Management

REINSURANCE RATE-MAKING WITH PARAMETRIC AND NON-PARAMETRIC MODELS

Course information FN3142 Quantitative finance

Financial Econometrics

Likelihood Methods of Inference. Toss coin 6 times and get Heads twice.

Extreme Value Theory with an Application to Bank Failures through Contagion

The Normal Distribution. (Ch 4.3)

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

Analysis of the Oil Spills from Tanker Ships. Ringo Ching and T. L. Yip

Normal Distribution. Definition A continuous rv X is said to have a normal distribution with. the pdf of X is

Lecture 5: Fundamentals of Statistical Analysis and Distributions Derived from Normal Distributions

Gamma Distribution Fitting

Commonly Used Distributions

Dependence Structure and Extreme Comovements in International Equity and Bond Markets

1. You are given the following information about a stationary AR(2) model:

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

Estimating the Parameters of Closed Skew-Normal Distribution Under LINEX Loss Function

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -26 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

Market Risk Analysis Volume I

Chapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29

Modelling of extreme losses in natural disasters

Extreme Values Modelling of Nairobi Securities Exchange Index

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Unconventional Resources in US: Potential & Lessons Learned

Risk Management Performance of Alternative Distribution Functions

MODELLING OF INCOME AND WAGE DISTRIBUTION USING THE METHOD OF L-MOMENTS OF PARAMETER ESTIMATION

Window Width Selection for L 2 Adjusted Quantile Regression

The normal distribution is a theoretical model derived mathematically and not empirically.

Modeling Co-movements and Tail Dependency in the International Stock Market via Copulae

Statistics 431 Spring 2007 P. Shaman. Preliminaries

Modelling Joint Distribution of Returns. Dr. Sawsan Hilal space

High-Frequency Data Analysis and Market Microstructure [Tsay (2005), chapter 5]

Statistical Modeling Techniques for Reserve Ranges: A Simulation Approach

Note. Everything in today s paper is new relative to the paper Stigler accepted

Advanced Extremal Models for Operational Risk

ECON 6022B Problem Set 2 Suggested Solutions Fall 2011

Master s in Financial Engineering Foundations of Buy-Side Finance: Quantitative Risk and Portfolio Management. > Teaching > Courses

Homework Problems Stat 479

Subject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018

Folded- and Log-Folded-t Distributions as Models for Insurance Loss Data

A Markov Chain Monte Carlo Approach to Estimate the Risks of Extremely Large Insurance Claims

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

Introduction to Computational Finance and Financial Econometrics Descriptive Statistics

Model Uncertainty in Operational Risk Modeling

MEASURING EXTREME RISKS IN THE RWANDA STOCK MARKET

Operational Risk: Evidence, Estimates and Extreme Values from Austria

Risk Analysis for Three Precious Metals: An Application of Extreme Value Theory

Graduate School of Business, University of Chicago Business 41202, Spring Quarter 2007, Mr. Ruey S. Tsay. Solutions to Final Exam

BROWNIAN MOTION Antonella Basso, Martina Nardon

STOCHASTIC MODELING OF HURRICANE DAMAGE UNDER CLIMATE CHANGE

STAT 830 Convergence in Distribution

Value at Risk Analysis of Gold Price Returns Using Extreme Value Theory

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

. (i) What is the probability that X is at most 8.75? =.875

THRESHOLD PARAMETER OF THE EXPECTED LOSSES

Statistics 13 Elementary Statistics

Chapter 7: Point Estimation and Sampling Distributions

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

Transcription:

19th TIES Conference, Kelowna, British Columbia 8th June 2008

Topics for the day 1. Classical models and threshold models 2. Dependence and non stationarity 3. R session: weather extremes 4. Multivariate extremes 5. Bayesian inference for Extremes 6. R session: multivariate analysis and Bayesian inference

Session 1. Classical models and threshold models 1.1 Introduction 1.2 Classical models 1.3 Threshold models

1.1 Introduction Statistical modelling of extreme weather has a very practical motivation: reliability

1.1 Introduction Statistical modelling of extreme weather has a very practical motivation: reliability anything we build needs to have a good chance of surviving the weather/environment for the whole of its working life.

1.1 Introduction Statistical modelling of extreme weather has a very practical motivation: reliability anything we build needs to have a good chance of surviving the weather/environment for the whole of its working life. This has obvious implications for civil engineers and planners. They need to know: etc. how strong to make buildings; how high to build sea walls; how tall to build reservoire dams; how much fuel to stockpile;

This motivates the need to estimate what the: strongest wind; highest tide; heaviest rainfall; most severe cold-spell; etc. will be over some fixed period of future time.

This motivates the need to estimate what the: strongest wind; highest tide; heaviest rainfall; most severe cold-spell; etc. will be over some fixed period of future time. The only sensible way to do this is to use data on the variable of interest (wind, rain etc.) and fit an appropiate statistical model.

This motivates the need to estimate what the: strongest wind; highest tide; heaviest rainfall; most severe cold-spell; etc. will be over some fixed period of future time. The only sensible way to do this is to use data on the variable of interest (wind, rain etc.) and fit an appropiate statistical model. The models themselves are motivated by asymptotic theory, and this is our starting point...

1.2 Classical models Extreme value modelling has a central theoretical result, analagous to the Central Limit Theorem... Suppose X 1,X 2,..., is an independent and identically distributed sequence of random variables. Define M n = max{x 1,...,X n }.

1.2 Classical models Extreme value modelling has a central theoretical result, analagous to the Central Limit Theorem... Suppose X 1,X 2,..., is an independent and identically distributed sequence of random variables. Define M n = max{x 1,...,X n }. We are interested in the limiting distribution of M n as n. As with the mean, X, of {X 1,...,X n }, the limiting distribution of M n as n is degenerate, and we need to work with a normalized version...

The Extremal Types Theorem (Fisher and Tippett, 1928) If there exist sequences of constants {a n > 0} and {b n } such that Pr{(M n b n )/a n z} G(z) as n, where G is a non degenerate distribution function, then G belongs to one of the following families:

The Extremal Types Theorem (Fisher and Tippett, 1928) If there exist sequences of constants {a n > 0} and {b n } such that Pr{(M n b n )/a n z} G(z) as n, where G is a non degenerate distribution function, then G belongs to one of the following families: { [ ( )]} z β I : G(z) = exp exp, < z < ; γ { ( ) } z β α II : G(z) = exp, z > β; [G(z) = 0,z β]; γ { [ ( ) z β α ]} III : G(z) = exp, z < β; [G(z) = 1,z β], γ for parameters γ > 0,β, and α > 0.

The Generalized Extreme Value Distribution (GEV) Families I, II and III are widely referred to as Gumbel, Frechet and Weibull (or Extreme Value Types I, II and III) respectively.

The Generalized Extreme Value Distribution (GEV) Families I, II and III are widely referred to as Gumbel, Frechet and Weibull (or Extreme Value Types I, II and III) respectively. Fortunately they can be combined into a single family, known as the Generalized Extreme Value Distribution (GEV), with c.d.f. { [ ( )] } z µ 1/ξ G(z) = exp 1 + ξ, (1) σ defined on the set {z : 1 + ξ(z µ)/σ > 0}, and where µ, σ > 0 and ξ are location, scale and shape parameters respectively.

The Generalized Extreme Value Distribution (GEV) Families I, II and III are widely referred to as Gumbel, Frechet and Weibull (or Extreme Value Types I, II and III) respectively. Fortunately they can be combined into a single family, known as the Generalized Extreme Value Distribution (GEV), with c.d.f. { [ ( )] } z µ 1/ξ G(z) = exp 1 + ξ, (1) σ defined on the set {z : 1 + ξ(z µ)/σ > 0}, and where µ, σ > 0 and ξ are location, scale and shape parameters respectively. So the Extremal Types Theorem can be restated with (1) as the limiting form, and this provides the basis for our first modelling approach...

Note that the Extreme Value Types I, II and III correspond to the cases ξ = 0, ξ > 0 and ξ < 0 respectively.

Note that the Extreme Value Types I, II and III correspond to the cases ξ = 0, ξ > 0 and ξ < 0 respectively. For Type I, we need to take the limiting form of Equation (1) as ξ 0, which gives { [ ( )]} z µ G(z) = exp exp, (2) σ defined for all z.

Note that the Extreme Value Types I, II and III correspond to the cases ξ = 0, ξ > 0 and ξ < 0 respectively. For Type I, we need to take the limiting form of Equation (1) as ξ 0, which gives { [ ( )]} z µ G(z) = exp exp, (2) σ defined for all z.

Approach 1: Block maxima Break up our sequence X 1,X 2,... into blocks of size n (with n reasonably large), and extract only the maximum observation from each block.

Approach 1: Block maxima Break up our sequence X 1,X 2,... into blocks of size n (with n reasonably large), and extract only the maximum observation from each block. Now fit Model (1) to the sequence of extracted maxima M (1),M (2),...,M (N) and use this as the basis for statistical inference.

Approach 1: Block maxima Break up our sequence X 1,X 2,... into blocks of size n (with n reasonably large), and extract only the maximum observation from each block. Now fit Model (1) to the sequence of extracted maxima M (1),M (2),...,M (N) and use this as the basis for statistical inference. The most common implementation of this approach for weather data is to take our block size to be one year.

Approach 1: Block maxima Break up our sequence X 1,X 2,... into blocks of size n (with n reasonably large), and extract only the maximum observation from each block. Now fit Model (1) to the sequence of extracted maxima M (1),M (2),...,M (N) and use this as the basis for statistical inference. The most common implementation of this approach for weather data is to take our block size to be one year. This rough and ready approach has shown itself to be surprisingly robust!

Approach 1: Example Consider the annual maxima of daily rainfall accumulations (mm) at a location in SW England, from 1914 to 1961. Figure 1. Annual Maxima for Rain Data Rainfall (mm) 30 40 50 60 70 80 1920 1930 1940 1950 1960 Year

Approach 1: Inferences Here our blocks have n = 365, which is reasonably large, so we fit Model (1) to the N = 49 annual maxima (e.g. using maximum likelhood estimation). We obtain fitted parameter values (standard errors in parentheses): µ = 40.7(1.5) σ = 9.4(1.2) ξ = 0.14(0.12).

Approach 1: Inferences Here our blocks have n = 365, which is reasonably large, so we fit Model (1) to the N = 49 annual maxima (e.g. using maximum likelhood estimation). We obtain fitted parameter values (standard errors in parentheses): µ = 40.7(1.5) σ = 9.4(1.2) ξ = 0.14(0.12). More importantly, we can make inferences on the quantities most useful to practitioners...

Approach 1: Inferences Here our blocks have n = 365, which is reasonably large, so we fit Model (1) to the N = 49 annual maxima (e.g. using maximum likelhood estimation). We obtain fitted parameter values (standard errors in parentheses): µ = 40.7(1.5) σ = 9.4(1.2) ξ = 0.14(0.12). More importantly, we can make inferences on the quantities most useful to practitioners... For example, the 99th percentile in the distribution of annual maxima is known as the 100 year return level.

Approach 1: Inferences Here our blocks have n = 365, which is reasonably large, so we fit Model (1) to the N = 49 annual maxima (e.g. using maximum likelhood estimation). We obtain fitted parameter values (standard errors in parentheses): µ = 40.7(1.5) σ = 9.4(1.2) ξ = 0.14(0.12). More importantly, we can make inferences on the quantities most useful to practitioners... For example, the 99th percentile in the distribution of annual maxima is known as the 100 year return level. The fitted value of this is easily obtained on inversion of Model (1): q 100 = 101.3(18.9).

Approach 1: Remarks

Approach 1: Remarks We don t need to deal explicitly with normalization constants. We don t even need to know n!

Approach 1: Remarks We don t need to deal explicitly with normalization constants. We don t even need to know n! The assumption of n independent and identically distributed variables in each block is cavalier, but inferences are surprisingly robust.

Approach 1: Remarks We don t need to deal explicitly with normalization constants. We don t even need to know n! The assumption of n independent and identically distributed variables in each block is cavalier, but inferences are surprisingly robust. The inferences on return levels are crucial for designers and engineers, to the extent they are built into legally binding codes of practice.

In actual fact, the existing codes of practice are usually based on a very primitive version of the methods just described. Fits are often based on restricting to one of the Fisher Tippett types, ignoring estimation uncertainty, and using an ad hoc interpolation of return levels across a network of sites.

In actual fact, the existing codes of practice are usually based on a very primitive version of the methods just described. Fits are often based on restricting to one of the Fisher Tippett types, ignoring estimation uncertainty, and using an ad hoc interpolation of return levels across a network of sites. In any case the block maxima approach is often very wasteful of data, leading to large uncertainties on return level estimates. This motivates a different approach (see later).

Approach 1: Diagnostics The goodness of fit of the GEV model is most easily assessed using various diagnostic plots. Here we consider four plots: 1 Probability plot: the fitted value of the c.d.f. is plotted against the empirical value of the c.d.f. for each data point. 2 Quantile plot: the empirical quantile is plotted against the fitted quantile for each data point. 3 Return level plot: the return level (with error bars) is plotted against the return period. Each data point defines a sample point. 4 Density plot: the fitted p.d.f. is supereimposed on a histogram of the data.

For our rainfall example, the diagnostic plots look like this... Probability Plot Quantile Plot 0.2 40 Model 0.4 0.6 Empirical 50 60 70 0.8 80 1.0 0.0 0.0 0.2 0.4 0.6 0.8 Empirical 1.0 30 30 40 50 60 70 Model 80 90 Return Level 40 60 80 100120140 Return Level Plot Density Plot f(z) 0.00 0.01 0.02 0.03 0.04 0.01 1 10 100 1000 20 30 40 50 60 70 80 90 Return Period z

Approach 1: Confidence intervals for return levels Although we could construct a symmetrical confidence interval for the r year return level using classical likelihood theory (ˆq r ± 1.96 standard error), this is not recommended. This practice assumes the limiting quadratic behaviour of the likelihood surface near the maximum, whereas in fact the surface is usually very asymmetrical.

Approach 1: Confidence intervals for return levels Although we could construct a symmetrical confidence interval for the r year return level using classical likelihood theory (ˆq r ± 1.96 standard error), this is not recommended. This practice assumes the limiting quadratic behaviour of the likelihood surface near the maximum, whereas in fact the surface is usually very asymmetrical. We recommend using the method of profile likelihood to take this into account: by reparametrization of Equation (1) to replace one of the parameters by q r, we can maximize the likelihood conditional on q r taking each possible value. We plot this constrained value against q r...

Approach 1: Profile likelihood confidence interval for q 100 For the rainfall example we get... Profile Log-likelihood -190.0-189.5-189.0-188.5-188.0-187.5 80 100 120 140 160 180 200 Return Level The likelihood ratio test can be applied directly to this likelihood surface by using a cut off equal to 0.5 χ 2 1 ( ). Here we see that the 95% confidence interval is approximately (78,176).

1.3 Threshold methods Threshold methods use a more natural way of determining whether an observation is extreme - all values greater than some high value (threshold) are considered.

1.3 Threshold methods Threshold methods use a more natural way of determining whether an observation is extreme - all values greater than some high value (threshold) are considered. This allows more efficient use of data, but brings its own problems.

1.3 Threshold methods Threshold methods use a more natural way of determining whether an observation is extreme - all values greater than some high value (threshold) are considered. This allows more efficient use of data, but brings its own problems. We must first go back and consider the asymptotic theory appropriate for this new situation.

The Generalized Pareto Distribution (GPD) The appropriate limit theorem can be stated as follows:

The Generalized Pareto Distribution (GPD) The appropriate limit theorem can be stated as follows: Under very broad conditions, if it exists, any limiting distribution as u of (X u X > u) is of Generalized Pareto Distribution (GPD) form (setting Y = X u):

The Generalized Pareto Distribution (GPD) The appropriate limit theorem can be stated as follows: Under very broad conditions, if it exists, any limiting distribution as u of (X u X > u) is of Generalized Pareto Distribution (GPD) form (setting Y = X u): ( H(y) = 1 1 + ξy ) 1/ξ, (3) σ +

The Generalized Pareto Distribution (GPD) The appropriate limit theorem can be stated as follows: Under very broad conditions, if it exists, any limiting distribution as u of (X u X > u) is of Generalized Pareto Distribution (GPD) form (setting Y = X u): ( H(y) = 1 1 + ξy ) 1/ξ, (3) σ + where a + = max(0,a) and σ (σ > 0) and ξ ( < ξ < ) are scale and shape parameters respectively.

Once again the GPD exists for ξ = 0, and is given by taking the limit of (3) as ξ 0. This time we get ( ) y H(y) = 1 exp, (4) σ defined for y > 0. This shows that when ξ = 0, the GPD is in fact the Exponential Distribution with mean equal to the scale parameter σ (σ > 0).

Return levels for the threshold excesses approach If the GPD is a suitable model for exceedances of a threshold u by a random variable X, then for x > u, [ ( )] x u 1/ξ Pr{X > x X > u} = 1 + ξ. σ

Return levels for the threshold excesses approach If the GPD is a suitable model for exceedances of a threshold u by a random variable X, then for x > u, [ ( )] x u 1/ξ Pr{X > x X > u} = 1 + ξ. σ It follows that ( )] x u 1/ξ Pr{X > x} = λ u [1 + ξ, (5) σ where λ u = Pr{X > u}. So the level x m that is exceeded once every m observations is the solution of ( )] x u 1/ξ λ u [1 + ξ = 1 σ m.

Rearranging this we obtain x m = u + σ ξ [(mλ u) ξ 1], so long as m is large enough to ensure that x m > u.

Rearranging this we obtain x m = u + σ ξ [(mλ u) ξ 1], so long as m is large enough to ensure that x m > u. Now if there are n y observations per year, then by setting m = N n y, the N year return level is obtained as or when ξ = 0, z N = µ + σ ξ [(Nn yλ u ) ξ 1] (6) z N = u + σ log(nn y λ u ), and standard errors can be obtained using the delta method.

Approach 2: Exceedances over thresholds

Approach 2: Exceedances over thresholds In practice, modelling might typically proceed as follows:

Approach 2: Exceedances over thresholds In practice, modelling might typically proceed as follows: 1 Choose some threshold u 0 which is high enough so that the GPD (3) is a good model for (X u 0 X > u 0 ).

Approach 2: Exceedances over thresholds In practice, modelling might typically proceed as follows: 1 Choose some threshold u 0 which is high enough so that the GPD (3) is a good model for (X u 0 X > u 0 ). 2 Fit the GPD to the observed excesses x u 0.

Approach 2: Exceedances over thresholds In practice, modelling might typically proceed as follows: 1 Choose some threshold u 0 which is high enough so that the GPD (3) is a good model for (X u 0 X > u 0 ). 2 Fit the GPD to the observed excesses x u 0. 3 Use the fitted GPD, together with some model for the rate of exceedances X > u 0, to provide estimates for return levels.

Approach 2: Example For the rainfall data we used before, now consider the daily totals themselves. Figure 2. Daily Rainfall (1914-1961) Daily Rainfall (mm) 0 20 40 60 80 1920 1930 1940 1950 1960 Year

Approach 2: Threshold choice Mean residual life plot We make use of the fact that if the GPD is the correct model for all the exceedances x i above some high threshold u 0, then the mean excess, i.e. the mean value of (x i u), plotted against u > u 0, should give a linear plot (Davison and Smith, 1990). [Because E[X i u 0 ] is a linear function of u : u > u 0 ] By producing such a plot for values of u starting at zero, we can select reasonable candidate values for u 0.

Approach 2: Mean residual life plot for daily rainfall replacements 0 Mean Excess 5 10 15 20 25 0 20 40 u 60 80

Approach 2: Inferences Model (3) turns out to work reasonably well for all the excesses above u 0 = 30mm. This gives 152 exceedancs x i ;i = 1,...,152, and Model (3) is fitted to the values (x i u), again using maximum likelihood. We get σ = 7.44(0.96) ξ = 0.18(0.10).

Approach 2: Inferences Model (3) turns out to work reasonably well for all the excesses above u 0 = 30mm. This gives 152 exceedancs x i ;i = 1,...,152, and Model (3) is fitted to the values (x i u), again using maximum likelihood. We get σ = 7.44(0.96) ξ = 0.18(0.10). Assuming a uniform rate of exceedances, we estimate the 100 year return level: q 100 = 106.3(20.8).

Approach 2: Diagnostics Probability Plot Quantile Plot 0.2 40 Model 0.4 0.6 0.8 0.0 Empirical 50 60 70 80 1.0 0.0 0.2 0.4 0.6 Empirical 0.8 1.0 30 30 40 50 60 70 Model 80 90 Return Level Plot Density Plot Return level 100 200 300 400 0.1 1 10 100 1000 Return period (years) f(x) 0.00 0.04 0.08 0.12 30 40 50 60 x 70 80 90

Approach 2: Profile likelilhood confidence interval for q 100 From the graph below, this is approximately (81,184). Profile Log-likelihood -487.0-486.5-486.0-485.5-487.5-485.0 80 100 120 140 160 180 200 Return Level

Approach 2: Threshold choice revisited If the GPD with shape parameter ξ and scale parameter σ u0 is the correct model for excesses over u 0, then for any threshold u > u 0, the excesses will be GPD with shape parameter ξ, and scale parameter σ u = σ u0 + ξ(u u 0 ).

Approach 2: Threshold choice revisited If the GPD with shape parameter ξ and scale parameter σ u0 is the correct model for excesses over u 0, then for any threshold u > u 0, the excesses will be GPD with shape parameter ξ, and scale parameter σ u = σ u0 + ξ(u u 0 ). If we now use a modified version of the scale parameter, σ = σ u ξu, we can see that both σ and ξ should be constant over thresholds greater than u 0 if we model excesses x i u for u > u 0 using the GPD.

This provides us with a further tool for assessing our original choice of threshold u 0. We refit the GPD for a range of thresholds upwards of u 0, and investigate the stability of our estimates of ξ and σ.

Approach 2: Parameter stability plots 50 0 100 150 Shape Modified Scale 50 2.0 1.0 0.0 30 30 35 35 40 Threshold 40 45 45 50 50 Threshold We can be reassured about our original choice of u 0 = 30!