Flexible modeling of frequency-severity data

Size: px
Start display at page:

Download "Flexible modeling of frequency-severity data"

Transcription

1 Faculty of Science Flexible modeling of frequency-severity data Olivier Vermassen Master dissertation submitted to obtain the degree of Master of Statistical Data Analysis Promotor: Prof. Dr. Christophe Ley Co-promotor: Dr. Robin van Oirbeek Statistics Academic year

2 The author and the promoter give permission to consult this master dissertation and to copy it or parts of it for personal use. Each other use falls under the restrictions of the copyright, in particular concerning the obligation to mention explicitly the source when using results of this master dissertation. Olivier Vermassen August 27, 2017

3 Foreword There are two main risk drivers in general insurance, namely claim frequency and claim severity. Many insurance models make restrictive distribution assumptions on latter random variables and even assume them to be independent. As the title of our master thesis suggests, we will explore flexible models for frequency and severity data and relax the commonly made assumptions. A case study is presented with non-confidential data from an unknown private insurer. A special thanks goes to my co-promotor Dr. Robin van Oirbeek for introducing me to this topic and his valuable remarks. In addition, I would like to thank Professor Dr. Christophe Ley for his guidance and enthusiasm throughout this project.

4

5 Table of Contents 1 Introduction 1 2 Methods The classical tarification strategy Univariate modeling of random variables Modeling dependence with copula models A more flexible pricing strategy Model validation Case study Exploratory Data Analysis Claim Frequency Analysis Claim Severity Analysis Dependence analysis Conclusion 37 References 37

6

7 Abstract Claim frequency and claim severity are the main two risk drivers in general insurance. Many insurance models make restrictive distribution assumptions on latter random variables and even assume them to be independent. The premium charged to a customer is given by the multiplication of the expected claim frequency and severity. This master thesis will explore flexible models for frequency and severity data and relax the commonly made assumptions. A motor insurance case study is provided to illustrate latter strategy. Insurance data is characterized by semi-continuous data. In fact, most customers did not incur a loss, hence resulting in a mixture of a degenerate distribution at 0 and a positive-continuous part. We start by fitting a binomial Generalized Additive Model to the claim probability component, i.e. modeling the probability a customer will incur a loss. The positive-continuous part is split into two components, namely claim severity and conditional claim frequency. The conditional claim frequency models the actual number of claims a customer will have, i.e. 1, 2, 3,... claims. We continue by fitting Generalized Additive Models for Location, Scale and Shape to latter random variables. Special care is taken for large losses by using Extreme Value Theory. Finally, dependence between the claim severity and conditional claim frequency is modeled by means of a Gaussian copula model. We find that our approach provides a better fit to the data, but does not improve predictive performance over the classical approach. KEY WORDS: General insurance pricing; Generalized Additive Models for Location, Scale and Shape (GAMLSS); Extreme Value Theory (EVT); Dependence modeling

8

9 1 Introduction Risk is a condition in which there is a possibility of an adverse deviation from a desired outcome that is expected or hoped for. [34] Possibility refers to the fact that we are dealing with probabilities and adverse deviation implies that there is a potential loss. For instance, drivers have a positive probability of being involved in a car crash and if the risk is realized, a potential cost is associated with it. How can members of society deal with such risks? The most well-known solution is insurance (see Figure 1.1). In principle, insurance is about exchanging uncertainty for certainty. The customer pays a fixed premium and obtains benefits in return, conditional on the occurrence of a pre-specified event that induces a financial loss. Financial losses insurers are confronted with are spread among a large community that contains the whole customer base of the insurer. In sum, an individual can transfer his risk to a group and potential losses are shared among the whole group. The latter principles are known as risk transfer and risk sharing, respectively. Risk sharing is related to the Law of Large Numbers, as by... expanding the number of risks in a pool, the average loss (or gain) in an expanding pool of risks eventually becomes certain (or predictable) [23, p. 7]. Certainty Premium New customer Insurer Community Benefits Uncertainty Figure 1.1: Principle of insurance. The insurer faces two potential problems with the construction of a community. Anti-selection implies that customers have more information about the risk than the insurer and is thus a form of information asymmetry. The concept of moral hazard is about customers behaving differently just because they have taken out insurance. How can the insurer deal with the latter problems? The most feasible solution for solving anti-selection is searching for rating variables that could explain the risk of a customer. This is also known as segmentation and implies that two distinct customers will be offered two distinct prices. Insurance data is in general reliable because fraud

10 2 would lead to a loss of coverage. Furthermore, the data available is usually only correlated to the risk and not causal. For instance, data on swiftness of reflexes cannot be obtained. Moral hazard can be solved by adding a time dimension to the modeling strategy. Customers that behaved better than expected last year, will receive a discount in the current year and vice versa. Non-life insurance is characterized by an inverted product cycle, which implies that premiums are received before any costs have to be paid. The premium in non-life insurance generally corresponds to an insurance contract that provides one-year coverage. Randomness comes into play as the insurer does not know how many claims a customer will have in the following year, nor how large the claims will be. The former is known as claim frequency and the latter is known as claim severity. A tariff is constructed with the multiplication of claim frequency and claim severity. The statisical challenge lies in the estimation of both expected frequency and severity from historic data. Why does insurance work? It is based on the Weak Law of Large Numbers (WLLN). For independent and identically distributed random variables X 1,..., X n, n N with E[X i ] = µ <, we have for all ɛ > 0: P [ 1 n ] n X i µ > ɛ i=1 n 0 In insurance, n will correspond to the number of policies in a portfolio and X i is the total cost related to policy i. It is usual to assume policies are independent. The WLLN tells that we can use the expected value as a premium, as long as the portfolio is large enough. In practice, the expected value of the claim frequency and severity are modelled by means of a Poisson and gamma distribution, respectively. Furthermore, both random variables are assumed to be independent. The concept of this thesis is flexible modelling and aims at relaxing the rather restrictive assumptions that are made in practice. For that reason, the main subject of this thesis is twofold: 1. investigate what other distributions/methods are appropriate for modelling the Frequency and Severity component; 2. investigate how dependence can be modelled between the Frequency and Severity component. Having both allows constructing a bivariate density. We will determine with a case study to what extent these rather restrictive assumptions lead to over- or underestimation of the risk.

11 2 Methods 2.1 The classical tarification strategy A key principle in non-life ratemaking is cost-based pricing. The costs related to customer i in a future year are random and can be decomposed in two components: 1. Claim frequency N i ; 2. Claim severity Y i,j where j refers to the j th payment. Every customer has a random aggregate loss L i = N i j=1 Y i,j, where L i 0 if N i = 0. The expected value of latter random sum equals [35, p. 25]: E[L i ] = E = E = E [ E [ Ni j=1 [ Ni [ Ni ]] Y i,j N i, j=1 ] E [Y i,j N i ], ] E [Y i,j ], j=1 = E[N i E[Y i,1 ]], = E[N i ] E[Y i,1 ]. under the assumption that Y i,j is independent from N i and Y i,j j are independent and identically distributed. E[L i ] will be charged to customer i as a premium, and relies upon the WLLN as explained in Section 1. Note that we ignore expenses, taxes and profit margins that would be added on top of the premium in practice. In the remainder of the master thesis, we will focus on the average claim severity Y i L i N i. The latter set-up is known as frequency-severity modeling, since there is one component modeling the event of a claim, whilst the other component focuses on the claim amount, conditional on there being at least one claim. A covariate vector x i will be used to segment the tariff and charge customer i a premium in line with his risk. Practitioners almost always assume that

12 4 N i Poi(λ i E i ) and Y i Ga(α i, γ i ), where Poi and Ga denote the Poisson and gamma distribution, respectively. Although insurance coverage is provided for one year, data records are broken down by year for accounting reasons. Hence, the logarithm of the exposure E i, i.e. the proportion of the year insured, is added as an offset. The expected value of random variables N i and Y i is modeled by means of a linear structure and logarithmic link function: λ i = e β x i +log E i, α i γ i = e η x i, with β the parameter vector for claim frequency and η the parameter vector for claim severity. Note that the latter pricing strategy allows a different covariate vector for claim frequency and claim severity. The premium P i is given by the multiplicative tariff under the assumption of independence: λ i αi γ i, and will be charged to the i th customer. As explained in Section 1, the expected value of one customer is not sufficient to cover claims suffered by one person, yet the premium income of the whole customer base should be able to cover all claims suffered due to the Weak Law of Large Numbers. The remainder of Chapter 2 will focus on more flexible modeling strategies including Extreme Value Theory (EVT), Generalized Additive Models for Location, Scale and Shape and dependence modeling via copulas.

13 5 2.2 Univariate modeling of random variables Generalized Linear Model Generalized Linear Models (GLM) are omnipresent in actuarial science, as they allow to understand the effects covariates have on a response variable Y i (see for instance [28] and [19]). A GLM allows strictly monotone transformations of the mean to depend on covariates in a linear manner. Even more, GLMs have been developed for members of the exponential family: ( ) yi θ i b(θ i ) f(y i ; θ i, φ) = exp + c(y i, φ) a(φ) where θ i is the location parameter, φ the dispersion paramater and a( ), b( ) and c( ) known functions. GLMs relax the normality assumption of the standard linear regression model. Wellknown distributions that are part of the exponential family are the Poisson, gamma, normal and binomial distributions. More distributions can be obtained by considering a transformation of the response variable Y i. For instance, a logarithmic transformation of the response variable allows working with a lognormal distribution. The real-valued functions a( ), b( ) and c( ) determine the distribution and the moments can be defined as: E[Y i ] = b (θ i ), Var(Y i ) = b (θ i )a(φ). A GLM consists of three components. Firstly, the random component Y i has a density belonging to the exponential family. Secondly, the systematic component η i = β x i corresponds to the linear predictor. The covariate vector x i of dimension p is the covariate information for observation i and β is the parameter vector (β 0, β 1,..., β p ). Finally, there is a link function g( ) that connects the expected value E[Y i ] = µ i to the linear predictor: g(µ i ) = η i., Section will introduce a flexible alternative for the GLM. Often of interest in actuarial modeling is a Poisson regression in order to model claim counts (see Section 2.1). A Poisson random variable N i Poi(λ i ) has probability mass function: P(N i = k i ) = λk i i e λ i k i! = e k i log λ i λ i log(k i!), and clearly belongs to the exponential family with a(φ) = 1, b(θ i ) = λ i and c(k i, φ) = log(k i!). We have that λ i > 0, hence a function g : R + R is needed in order to avoid

14 6 generating negative predictions. g is often chosen to be the logarithmic link function for Poisson regression models. Parameter estimates can be found by using maximum likelihood estimation, i.e. maximization of the log-likelihood. The log-likelihood for a member of the exponential family is: L(β, φ, y) = n [ yi θ i b(θ i ) i=1 a(φ) ] + c(y i, φ) and will be maximized over β, resulting in parameter estimates ˆβ. Using that b (θ i ) = µ i θ i = (b ) 1 (µ i ), the log-likelihood can be parameterized in terms of the mean µ i = g 1 (β x i ): L(β, φ, y i ) = n [ yi (b ) 1 (µ i ) b((b ) 1 (µ i )) i=1 a(φ), ] + c(y i, φ). The deviance is defined as: [ 2 L( ˆβ, ] φ, y) L(y, φ, y), with L(y, φ, y) the log-likelihood of the saturated model where ˆµ i = y i. The deviance essentially measures the closeness of the observed and predicted values. The smaller the deviance, the better the fit. For a Poisson regression model, the Deviance can be shown to equal: 2 n i=1 ( y i log ( yi ˆµ i ) ) (y i ˆµ i ) and can be used for hypothesis testing or out-of-sample validation., Whenever confronted with competing models, an information criterium can be defined that penalizes the likelihood: 2 L(β, φ, y) + c (degrees of freedom) The Akaike Information Criterium (AIC) corresponds to c = 2 and the Bayesian Information Criterion to ln n, where n denotes the sample size.

15 Generalized Additive Model of Location, Scale and Shape An extension of the GLM is the Generalized Additive Model (GAM) and allows for a smooth effect of the parameters. Generalized Additive Model for Location, Scale and Shape (GAMLSS) is even more flexible and has both the GLM and GAM as sub-components. GAMLSS was initially developed by Stasinopoulos & Rigby [33]. It has not received a great deal of attention, but has for instance been used in Denuit et al. [21] and Antonio et al. [1] to model insurance data. A major feature of GAMLSS is that it allows to relax the distribution assumption of the error distribution since it supports the choice among a wide array of flexible parametric distributions. Even more, the distributional family encompasses all parametric distributions that are part of the exponential family. Some examples are provided in Tables 2.1 and 2.2 for discrete and continuous distributions, respectively. A GAMLSS allows modeling up to four parameters, where modelization implies that the parameter values can depend upon covariate effects. Following the notation of Stasinopoulos & Rigby [33], a parametric linear GAMLSS assumes that the response Y i has a density f ( Y i θ i), with θ i = (θ 1i, θ 2i, θ 3i, θ 4i ). Note that latter vector θ i can be of a lower dimension than four, depending on the distribution assumption made for the error distribution. By convention, the first four parameters are the location µ i, scale σ i, skewness ν i and kurtosis τ i respectively. For instance, the one parameter exponential distribution will only have the first parameter with θ 1i = µ i, while the Box-Cox t distribution will have four parameters with θ 1i = µ i, θ 2i = σ i, θ 3i = ν i and θ 4i = τ i. Note that the location parameter in the Box-Cox t distribution corresponds to the median. Every model parameter can depend on covariates. The systematic component of the GAMLSS is: g k (θ ki ) = β kx i for k = 1, 2, 3, 4 ; with g k ( ) a monotonic link function that connects the systematic component β kx i with the parameter θ ki, β k the parameter vector and x i the covariate information for observation i. GAMLSS can be further extended to include additive terms, non-linear parametric terms etc. for which we refer the reader to the original paper [33, p. 2-4]. For instance, smoothing techniques can be used, which move us to semi-parametric GAMLSS [33, p. 3]: r g k (θ ki ) = β kx i + f jk (x jk ) for k = 1, 2, 3, 4 ; j=1 with r the number of covariates modeled by means of a smooth function f such as a cubic

16 8 spline, fractional polynomial or other. Linear parametric GAMLSS can be fit by using maximum likelihood. Often of interest is a truncated distribution that is cut off at deterministic points δ L, δ R for δ L < δ R. This can be fit by f(y i θ maximum likelihood, but the contribution for each observation y i will be i ) P(δ L <Y i <δ R θ i ) instead of f ( y i θ i). In other words, it corresponds to a conditional density that has all probability mass between the truncation points. In contrast, models that include non-linear terms or nonparametric smoothing functions are estimated by a penalized likelihood function [33, p. 4]. Distribution Number of parameters Poisson 1 Negative Binomial 2 Zero inflated Poisson 2 Zero adjusted Poisson 2 Poisson-inverse Gaussian 2 Sichel 3 Table 2.1: Some discrete distributions belonging to the GAMLSS family. Distribution Number of parameters Exponential 1 Gamma 2 Inverse Gaussian 2 Lognormal 2 Weibull 2 Generalized-Gamma 3 Generalized Inverse Gaussian 3 Box-Cox t 4 Generalized Beta type 2 4 Table 2.2: Some continuous distributions belonging to the GAMLSS family with support on R +.

17 Extreme value theory Insurance data is often characterized by right-skewed and heavy-tailed data. Classical statistical analysis focuses on the expected value of a random variable and is dominated by the normal distribution, a symmetric and light-tailed distribution. For instance, for a sum S n = X X n, the Central Limit Theorem (CLT) is concerned with finding constants a n > 0 and b n such that: Y n := S n b n a n n Y, and a normal distribution typically appears. In contrast, Extreme Value Theory (EVT) focuses on the limiting distribution of sample maxima X n,n = max(x 1,..., X n ). The Fisher-Tippett theorem, the EVT equivalent of the CLT, dictates that for suitable choices of a n > 0 and b n, the limiting distribution of normalized sample maxima: is the extreme value distribution: ( Xn,n b n P a n ) n x G(x), G ξ (x) = e (1+ξx) 1/ξ, with the Extreme Value Index ξ R and 1+ξx > 0. The sign of latter parameter will determine the domain of attraction of the distribution G. The Fréchet-Pareto domain corresponds to ξ > 0 and implies that the distribution is heavy tailed, which is often observed in insurance data as stated before. According to the Pickands-Balkema-de Haan theorem, data follows a Generalized Pareto Distribution (GPD) from a high enough threshold u on (see [20] and [4]). The GPD has the following cumulative distribution function: 1 ( ) 1 + ξx 1 ξ if ξ 0 σ G(x; ξ, σ) =, 1 e x σ if ξ = 0 with scale parameter σ > 0, shape parameter ξ (= Extreme Value Index) and conditions x 0 if ξ 0 and 0 x σ when ξ < 0. A location parameter µ can be added as well: ξ G(x µ; ξ, σ). Latter parameters can be found by using maximum likelihood. EVT is yet to determine the right approach to estimate the threshold u. McNeil [24, p. 132] does not discuss any method to find the threshold by a statistical procedure, but already hints that there is a bias-variance trade-off. If the threshold is too high, only few data points remain to fit the model, leading to high variance. In contrast, if the threshold is too low, the Pickands-

18 10 Balkema-de Haan theorem might not hold and bias will occur as a result. Which methods can be used to find the threshold u? Scarrot et al. [30] gives an overview of both the traditional and novel threshold estimation techniques. The traditional approaches are often 1 graphical, inspired by EVT, require investigating the data and allow for the inclusion of expert opinion. In contrast, the novel approaches focus on automation and uncertainty quantification. For instance, there exist likelihood-based approaches as well as clustering approaches with mixture models. Still, the novel approaches are not necessarily consistent with the theorems put forward by EVT. Which techniques are being used in the applied insurance literature? Antonio et al. [1], Reynkens et al. [29] and Ganegoda et al. [16] rely on an extreme value analysis with the related graphical tools. The first choose the threshold where the Hill estimator, an estimator of ξ, enters a stable region. The others study the mean excess plot in order to decide upon the optimal threshold u. In fact, the mean excess plot is linearly increasing for Pareto-type behavior. Similarly, this master thesis will perform a graphical EVT analysis and decide upon the threshold before the fitting procedure. The most well-known estimator for ξ is the Hill estimator H k,n : 1 k k log X n j+1,n log X n k,n, j=1 with k the threshold, n the sample size and X i,n the i th largest observation. A threshold stability plot can be created by plotting the Hill estimator as a function of the threshold. Yet another graphical approach is the mean excess plot. For a random variable X, the mean excess function is: and can be estimated as follows: e(t) = E[X t X > t], ê n (t) = n i=1 x ii (t, ) (x i ) n i=1 I (t, )(x i ) t, where I (t, ) (x i ) = 1 if x i > t and 0 otherwise. The mean excess plot of the exponential distribution is constant and therefore, interpretation in terms of tails is either heavier than exponential or lighter than exponential. Pareto-type behavior corresponds to a linearly increasing mean excess plot [8]. 1 Some researchers have proposed automated threshold selection techniques consistent with EVT. For instance, Pickands [20, p. 120] has proposed to choose as threshold u the value that minimizes the absolute value between the empirical distribution and the fitted GPD for values greater than u. Another example is Beirlant et al. [5], who proposes as threshold the value that minimizes the asymptotic mean squared error (AMSE) of an estimator for the extreme-value index. The AMSE can be decomposed in both the asymptotic squared bias and asymptotic variance.

19 Splicing for claim severity We motivated in Section the necessity to isolate large losses and to model them separate from the attritional losses, whereas attritional refers to small to medium sized losses. This section will explain the technique of splicing 2 or the body-tail approach to join the models for attritional and large losses into a single probabilistic model (see [22]). Latter technique was also used by Pigeon et al. [27] with a log-normal and Pareto distribution for the attritional and large losses, respectively. Ganegoda et al. [16] perform a splicing analysis as well in the context of operational loss modelling, but allow the distribution for attritional losses to come from the GAMLSS family. Similary, Antonio et al. [1], in the context of reserving, use a GAMLSS distribution for the attritional losses and a GPD for the large losses. In contrast, Bakar et al. [3] model loss data using spliced models and consider a Burr distribution for the tail. It was found that spliced models provide a significantly better fit. However, a caveat of their research is that the distribution for attritional losses is restricted to the Weibull distribution. Also, the threshold was considered as a model parameter, whilst imposing continuity and differentiability restrictions at the threshold. In fact, the latter is not consistent with EVT. Several other researchers have used K component mixture distributions. Miljkovic & Grun [25] have found such a mixture distribution to provide a better fit than the composite models used in Bakar et al. [3]. Reynkens et al. [29] use an even more flexible approach as the number of components K in the mixture distribution is considered to be random. In this section, we continue with the notation introduced in Section 2.1, namely L i N i Y i, a random variable that denotes the average claim severity of customer i in a particular year. The severity distribution f Yi (y) will be split up in two separate 3 density functions f 1 and f 2 by the splicing technique [22]: q 1 f 1 (y), y (0, u] f Yi (y) = q 2 f 2 (y), y (u, ) where both q 1 and q 2 0 and satisfy q 1 + q 2 = 1. Also, f 1 (y) and f 2 (y) are well-defined densities on the interval [0, u) and [u, ) respectively. In practice, one will often use truncated versions of existing probability distributions. The weights q 1 and q 2 ensure that the density integrates to 1: 0 u f Yi (y)dy = q 1 f 1 (y)dy + q 2 f 2 (y)dy = 1. 0 We refer to the first component as the body and the second component as the tail. The body 2 Spliced models are sometimes referred to as composite models. 3 We do not impose continuity nor differentiability at the splicing point as this would reduce the number of parameters and thus the flexibility. u,

20 12 component models the data up to the extreme value threshold u. This data can be modelled by a right truncated GAMLSS (see Section for more information about GAMLSS and truncation). The tail component will be modeled by means of a Generalized Pareto Distribution, as motivated in Section The upper limit was chosen to be +, but could be a policy limit or some maximal probable loss as well.

21 Modeling dependence with copula models General introduction to Copulas This section will give an introduction to dependence modeling with copulas. Since the interest lies in jointly modeling two random variables, we limit ourselves to the bivariate case. Consider continuous random variables X 1 and X 2 that are described separately by the cumulative distribution functions F X1 and F X2, respectively. X = (X 1, X 2 ) is a bivariate random vector that describes the joint behavior of latter random variables. The bivariate cumulative distribution function F X of X is defined as: F X (x) = P (X 1 x 1, X 2 x 2 ), where x = (x 1, x 2 ) R 2. In general, there exist three extreme dependence structures, namely independence, comonotonicity and counter monotonicity. Independence between X 1 and X 2 implies that the realized value for X 1 is not influenced by the realization of X 2 and vice versa. The random variables X 1 and X 2 are independent if and only if F X (x) satisfies: F X (x) = F X1 (x 1 ) F X2 (x 2 ), (x 1, x 2 ) R 2. A comonotonic dependence structure implies that all risk factors are driven by the same source of randomness. For a comonotonic random vector, we have that 4 : (X 1, X 2 ) d = ( F 1 X 1 (U), F 1 X 2 (U) ), where U d = Unif(0, 1). The joint cumulative distribution function under comonotonic dependence can be derived by making use of the quantile transform theorem: We obtain that: X i d = F 1 X i (U). F X (x) = P (X 1 x 1, X 2 x 2 ) = P ( ) F 1 X 1 (U) x 1, F 1 X 2 (U) x 2 = P (U F X1 (x 1 ), U F X2 (x 2 )) = min{f X1 (x 1 ), F X2 (x 2 )}, (x 1, x 2 ) R 2. 4 d = denotes equality in distribution.

22 14 See for instance Dhaene & Goovaerts [9], Dhaene et al. [11] and Dhaene et al. [12] for more information about comonotonicity. Counter monotonicity implies that all random variables are driven by the same source of randomness, but move in the opposite direction. The random variables X 1 and X 2 are counter monotonic if and only if: (X 1, X 2 ) d = ( F 1 X 1 (U), F 1 X 2 (1 U) ). The joint cumulative distribution function can again be derived by making use of the quantile transform theorem: F X (x) = P (X 1 x 1, X 2 x 2 ) = P ( ) F 1 X 1 (U) x 1, F 1 X 2 (1 U) x 2 = P (U F X1 (x 1 ), 1 U F X2 (x 2 )) = max{f X1 (x 1 ) + F X2 (x 2 ) 1, 0}, (x 1, x 2 ) R 2. Note that it is not straightforward to generalize counter monotonicity in a dimension larger than 2. A Fréchet space R 2 (F X1, F X2 ) that consists of all bivariate random vectors Y with fixed univariate distribution functions F X1, F X2 is used to study dependence structures: R 2 (F X1, F X2 ) = {Y F Y1 F X1, F Y2 F X2 }. Any joint distribution F Y with Y R 2 (F X1, F X2 ) can describe the random vector X. Note that the Fréchet upper and lower bound are given by the distribution functions with a comonotonic and countermonotonic dependence structure respectively. A copula is a function that allows to determine the dependence structure independently from the univariate distribution functions. The latter is in our interest as insurers usually have a model available for both claim frequency and claim severity. According to Sklar s theorem [32], there exists a unique copula C : [0, 1] [0, 1] [0, 1] so that: F X1, F X2 continuous distribution functions; F X (x 1, x 2 ) = C (F X1 (x 1 ), F X2 (x 2 )) with F X R 2 (F X1, F X2 ); (x 1, x 2 ) R 2. A copula satisfies several properties:

23 15 1. lim ui 0 C(u 1, u 2 ) = 0 for i = 1, 2; 2. lim u1 1 C(u 1, u 2 ) = u 2 and lim u2 1 C(u 1, u 2 ) = u 1 ; 3. C is supermodular so that the inequality C(v 1, v 2 ) C(u 1, v 2 ) C(v 1, u 2 )+C(u 1, u 2 ) 0 is valid for any u 1 v 1 and u 2 v 2. The first property implies that the joint probability equals 0 if one of the events has a zero probability of occurring. In contrast, the joint probability reduces to one event s probability of occurring if the other event will happen for sure. Finally, supermodular is a property that implies the joint probability should not decrease if the probabilities of both events are non-decreasing. We refer the reader to [15], [26] and [10] for more information about copulas Copula families The most well-known dependence structure is linear dependence and corresponds to the Gaussian copula with correlation parameter ρ. Note that the univariate distributions do not need to follow a Gaussian distribution. The strength of dependence increases with ρ as ρ = 1 corresponds to counter-monotonicity, ρ = 0 to independence and ρ = 1 to comonotonicity. A generalization of the Gaussian copula is the Student copula with an additional degrees of freedom parameter ν. Note that ρ = 0 only corresponds to independence as ν, which is equivalent to the Gaussian copula. Both the Gaussian and Student copula are elliptical copulas. An Archimedean copula with a strict generator φ is a copula of the form: C(u 1, u 2 ) = φ 1 (φ(u 1 ) + φ(u 2 )), and φ satisfies: φ(0) = + ; φ(1) = 0; φ is a continuous, strictly decreasing and convex function mapping [0, 1] onto [0, ). The Frank, Clayton and Gumbel copula satisfy these conditions and are thus part of the Archimedean copula family (see for instance [15] for an overview). Next to parametric copula families, there are non-parametric and empirical estimators as well. The empirical copula is simply the multivariate empirical distribution function of the transformed random variables F X1 (x 1 ) and F X2 (x 2 ).

24 Applications in Insurance This section will discuss the few research papers available that have relaxed the independence assumption between claim frequency and severity. Czado et al. [18] have performed a study on spatial (Bayesian) modelling for both frequency and severity data. Dependence was induced by using the number of claims as a covariate in the severity model. Garrido et al. [17] followed a similar strategy to relax the independence assumption. Although it was found to improve the results, it cannot be used in practice as the number of claims is unknown for new customers. In the fact, the latter is the most important source of randomness in pricing insurance contracts. Czado et al. [6] link a Poisson and gamma regression with a Gaussian copula. It was found that a copula approach did not outperform the independence model for their dataset. Czado et al. [7] generalize this approach to the class of Archimedean copulas. It was found that latter copulas are more appropriate than the Gaussian one. A drawback of their study is the use of a gamma regression, which, in the case of heavy-tailed data, could lead to misspecification of the percentiles. In addition, no out-of-sample estimation was performed, which is increasingly important as insurers try to predict claims in future years to come. The research of Shi et al. [31] is closest to the work that this master thesis will present. They work in a hurdle framework where the hurdle component is the probability to have at least one claim, and the conditional component is the number of claims and their respective sizes. They do not limit the univariate distributions to the GLM family, and model claim counts with a truncated negative binomial regression and the claim sizes are assumed to follow a generalized gamma distribution. The dependence assumption is relaxed by either the copula approach or using the number of claims as a covariate in the severity component. The copula approach was found to outperform the independence model on a hold-out sample.

25 A more flexible pricing strategy We have seen that the classical approach of the premium for customer i consists of the multiplication of the claim frequency and severity component, i.e. P i = E[Y i x i ] E[N i x i ]. We now want to relax the independence assumption between claim frequency and claim severity. The main problem is that lots of customers did not experience a claim, whilst claim severity is only defined if there is at least one accident. Hence, this motivates working in a conditional framework as explained in the paragraph below (see also Shi et al. [31] for a similar approach). Let I i be an indicator random variable that takes the value 1 if there is at least one claim (also called claim incidence) and 0 otherwise. N i is the claim frequency and has as support 5 S Ni = {0, 1, 2, 3,... }. Y i Y i I i = 1 with support S Yi = (0, ) is the claim severity and is conditional on I i = 1 by convention. We decompose N i by conditioning on I i : P(N i = k x i ) = P(N i = k I i = 1, x i ) P(I i = 1 x i )+P(N i = k I i = 0, x i ) P(I i = 0 x i ). Hence, we have that: P(I i = 0 x i ) if k = 0 ; P(N i = k x i ) = P(N i = k I i = 1, x i ) P(I i = 1 x i ) if k 1. The latter result shows that we can decompose claim frequency into two components, namely the probability to have a claim and the conditional claim frequency. Note that S Ni I i =1 = {1, 2, 3,... }, which implies it is truncated at 0. It is now feasible to define a dependence structure between Y i and N i I i = 1. The pricing strategy corresponding to the flexible approach is as follows: P i = E [N i Y i x i ] = E [N i Y i I i = 1, x i ] P(I i = 1 x i ) ; since E [N i Y i I i = 0, x i ] = 0. The advantages of the model over the traditional approach are twofold. Firstly, it is possible to consider a different set of covariates for claim incidence and claim frequency greater than or equal to 1. Next, it is allowed that Y i and N i are dependent so that: E [Y i N i x i ] E [Y i x i ] E [N i x i ]. The claim probability P(I i = 1 x i ) p(x i ) can be modeled by means of a binomial GAM. According to Baetschmann [2], the complementary log-log link can be used to deal with varying 5 The support S X of random variable X with probability density function f is the set of values for which f(x) > 0.

26 18 time exposure E i : ν(x i ) = 1 e eβ xi + r j=1 f j (x j )+log E i. Note that, in contrast to the Poisson case, the exposure is proportional to the risk in a non-linear fashion. Furthermore, it is assumed that: Y i I i = 1, x i ( ) G µ f i, σf i, νf i, τ f i ; N i I i = 1, x i H (µ s i, σ s i, ν s i, τ s i ). Note that f refers to frequency and s to severity. Distribution functions G and H are distributions existing within the GAMLSS framework and every parameter can potentially be modelled as a function of the covariate vector x i. Even more, Y i I i = 1, x i can be modeled with a truncated GAMLSS model for the attritional losses and a GPD for the tail as outlined in Section In addition, one can model a probability q(x i ) to have either an attritional or large loss. Finally, a copula is used to link the conditional claim frequency and claim size component: C ( F Ni I i =1,x i (k i ), F Yi I i =1,x i (y i ) ). A comprehensive summary of the model structure is provided in Figure 2.1 for customer i with covariate information x i. Frequency 1 N i I i = 1, x i Claim incidence Body p (x i ) Severity q (x i ) Customer i Y i I i = 1, x i 1 p(x i ) 1 q (x i ) No claim incidence Tail Figure 2.1: A more flexible pricing strategy that relaxes the classical assumption of independence. The contribution of this approach over the existing literature is threefold. Firstly, we investigate whether more predictive accuracy can be obtained by using GAMLSS in a hurdle framework. The latter could either be due to a more appropriate distribution assumption or that the risks for claim incidence and the conditional claim frequency are different. Secondly, large losses are

27 19 explicitly accounted for by using EVT and the splicing technique. Finally, this master thesis does not restrict dependence models to parametric copulas. 2.5 Model validation Insurance data is difficult to validate due to its semi-continuous nature. In fact, most customers did not incur a loss, hence resulting in a mixture of a degenerate distribution at 0 and a positivecontinuous part. As a result, any frequency-severity model is difficult to validate with standard performance metrics such as the Root Mean Square Error. Frees et al. [14] introduce the Lorentz curve, a well-known concept in economics, to insurance modeling. Observations y i are ordered by the riskiness measured by a relativity R. A straightforward choice would be the predicted premium as it indicates how risky customers are compared to each other. A plot can be generated: ( ˆFP (s), ˆF L (s)) with ˆF P (s) = ˆF L (s) = n i=1 P I(R i s) n i=1 P n i=1 y i I(R i s) n i=1 y i where I(R i s) = 1 if R i s and 0 otherwise. The x axis is the population of interest P and the y axis the distribution of incurred losses L. An example of the Lorentz Curve is displayed in Figure 2.2. The population of interest is the exposure of policyholders insurance policy. The observations y i are ordered by the predictions. As a result, the more concave the Lorentz curve, the better risks are classified. For instance, the point (20%, 35%), represented by the red lines, implies that 20% of the risks the insurer was exposed to, correspond to 35% of the incurred losses. Lorentz Curve, Incurred Losses Population Figure 2.2: Example: Ordered Lorentz Curve and Gini coefficient. The dotted line represents the 45 degree line, and the solid black line the Lorentz curve.

28 20 The related Gini coefficient is twice the area between the ordered Lorentz curve and the 45 degree line. This metric has for instance been used by Frees et al. [13] and Shi et al. [31] to validate pricing models.

29 3 Case study 3.1 Exploratory Data Analysis We have available a non-confidential dataset of 159, 911 records from an unknown private insurer. The available variables are listed in Table 3.1 and we visualize the distributions of the claim count Clm_Count, exposure TLength and claim severity AvPayment = TotPayment / Clm_Count in Figure 3.1. Most policyholders did not file a claim (145, 719 or 91.1%), those who do file usually have 1 claim only (12, 874 or 8.05%) and the remaining ones file two or more claims (1, 354 or 0.85%). The overall claim frequency Clm_Count TLength equals 15.47%. Variable Type Description Clm_Count Integer Number of claims filed by the policyholder. TLength Numeric The fraction of the year during which the policyholder was exposed to the risk. TotPayment Numeric The total amount claimed by the policyholder (in EUR). AgeInsured Integer Age of the policyholder in years. SexInsured Categorical Gender of the policyholder: Male (M) or Female (F). PrivateCar Binary 1 if it is a car for private use, 0 otherwise. Experience Integer Experience of the policyholder in years. VAge Continuous Age of the vehicle in years. Cover_C Binary 1 if the policyholder has this particular cover, 0 otherwise. NCD_0 Binary 1 if the no-claims history equals 0, 0 otherwise. Table 3.1: Overview of the variables. Relative frequency Relative frequency Density Clm_Count TLength AvPayment Figure 3.1: Distribution of the claim count Clm_Count, exposure TLength and kernel density estimate of the Average Payment AvPayment = TotPayment/Clm_Count.

30 22 A large proportion of policyholders have an exposure equal to 1. The remaining peaks are due to the underwriting process of the insurer, which happen at the end of the month. For instance, if a policyholder aged 26 subscribes an insurance contract 15 th of August 2016, the contract will end 31 st of July 2017 and potentially renewed 1 st of August The example presented in Table 3.2 shows how the data is structured and how variables are aged. In fact, variables are usually aged at the renewal date and observations broken down by accounting year. Observation Accounting year Renewal year Exposure Age /08/ /12/ /01/ /07/ /08/ /12/ /01/ /07/ Table 3.2: Example data structuring The claim severity, of which the density is visualized in Figure 3.1, is on average 4,438 EUR, whilst the median equals 2,575 EUR, indicating the right skewness of the data. It is clear that most policyholders file small claims, as only 3.16% of claims exceed 20,000 EUR. Finally, Figure 3.2 illustrates how the distribution of AvPayment evolves according to the number of claims filed. There seems to be an increasing trend, which might indicate some dependence is present between claim frequency and claim severity Count 1 density Loss Figure 3.2: Kernel Density Estimate of the severity distribution by Claim Count Figure 3.3 illustrates how the risk factors are distributed (right axis) and how they are related to the response variable (left axis). The response variables are the claim probability, conditional

31 23 claim frequency and claim severity for the left, middle and right panel, respectively. Claim Probability AgeInsured Exposure Claim Frequency AgeInsured Exposure Claim Severity AgeInsured Exposure Claim Probability SexInsured Exposure Claim Frequency SexInsured Exposure Claim Severity SexInsured Exposure F M F M F M Claim Probability Experience Exposure Claim Frequency Experience Exposure Claim Severity Experience Exposure Claim Probability VAge Exposure Claim Frequency VAge Exposure Claim Severity VAge Exposure Claim Probability PrivateCar Exposure Claim Frequency PrivateCar Exposure Claim Severity PrivateCar Exposure Claim Probability NCD_ Exposure Claim Frequency NCD_ Exposure Claim Severity NCD_ Exposure Claim Probability Cover_C Exposure Claim Frequency Cover_C Exposure Claim Severity Cover_C Exposure Figure 3.3: Exploratory Data Analysis : The crosses indicate the observed average of the response per level (left axis) and the bars the exposure per level (right axis). The response variables are the claim probability, conditional claim frequency and claim severity for the left, middle and right panel, respectively.

32 Claim Frequency Analysis This section will focus on modeling the claim frequency Clm_Count per unit of exposure TLength and provide a comparative analysis between the classical and hurdle approach. The classical approach will model Clm_Count directly with a Poisson error distribution and by using the logarithm of the exposure as offset. In contrast, the hurdle approach decomposes Clm_Count in the following two components: 1. Claim probability I Clm_Count 1, taking the value of 1 if there is at least one claim, and 0 otherwise; 2. Conditional claim frequency Clm_Count Clm_Count 1. The claim probability is modeled with a binomial GAM 1 with complementary log-log function, whilst using the logarithm of the exposure as offset. Similar to the classical approach, the conditional claim frequency component is modeled with a Poisson distribution and the logarithm of the exposure as offset. We use the following rule to model the conditional claim frequency, rather than using a truncated Poisson regression model: E [Clm_Count Clm_Count 1] = 1 + E [Clm_Count 1 Clm_Count 1]. For all three models, categorical variables are coded as dummy variables and the penalty for the smooth effects is estimated with generalized cross-validation. The optimal model structure is determined by doing an exhaustive search over all possible combinations 2, and by using 5 fold cross validation. On one hand, the AIC is calculated on the training folds, while on the other hand, the deviance between the observed and predicted values is calculated on the test folds. More weight is given to out-of-sample performance, as our aim is predicting the behavior of new clients. The following model structure, corresponding to the classical approach, attained the lowest AIC on all training folds: log E[Clm_Count] = β 0 + β 1 SexInsured + β 2 PrivateCar + β 3 Cover_C + β 4 NCD_0 + f 1 (AgeInsured) + f 2 (Experience) + f 3 (VAge) + log(tlength). In addition, latter structure also minimized the average Poisson deviance over the test folds. The exact same results were found for the claim probability component with the Binomial deviance. 1 The gam function from the mgcv package was used in R. 2 The model could be further enriched by searching for interaction effects. Yet, this would over-complicate the analysis and not add any value given the techniques we aim to demonstrate.

33 25 The model structure is as follows: log ( log (1 E[I Clm_Count 1 ])) = α 0 + α 1 SexInsured + α 2 PrivateCar + α 3 Cover_C + α 4 NCD_0 + g 1 (AgeInsured) + g 2 (Experience) + g 3 (VAge) + log(tlength). The optimal model structure for the conditional claim frequency on the training data included the variables SexInsured, Cover_C, NCD_0 and smooth effects for AgeInsured and VAge. In fact, latter structure attained the lowest AIC on 3 out of 5 training folds. Yet, the same model without a smooth effect for VAge attained the lowest average deviance on the test folds. Since more weight is given to out-of-sample performance, we continue with the following model structure: log E[Clm_Count 1 Clm_Count 1] = γ 0 + γ 1 SexInsured + γ 2 Cover_C + γ 3 NCD_0 + h 1 (AgeInsured) + log(tlength). Claim frequency Claim probability Conditional claim frequency f(ageinsured) f(ageinsured) f(ageinsured) AgeInsured AgeInsured AgeInsured Claim frequency Claim probability f(experience) f(experience) Experience Experience Claim frequency Claim probability f(vage) f(vage) VAge Figure 3.4: Fitted smooth effects for AgeInsured, Experience and VAge with 95% confidence intervals for the claim frequency, claim probability and conditional claim frequency. VAge

34 26 Figure 3.4 displays the fitted smooth functions with 95% confidence intervals. ˆf1, ˆf 2 & ˆf 3 correspond to the claim frequency model, ĝ 1, ĝ 2 & ĝ 3 to the claim probability model and ĥ1 to the conditional claim frequency model. It appears there are no tremendous differences in the fitted smooth effects of the three models. The fitted smooth effect of AgeInsured in any of the three models shows that young drivers are riskier. Latter effect stabilizes around the age of 40, after which the so-called accident hump occurs, i.e. parents with children that start driving and file claims under their parents name. The effect for the old ages is less clear due to scarceness of data, as the widening confidence intervals illustrate. Next, the smooth effect for Experience is monotonically decreasing and indicates that more experienced drivers have fewer claims on average. Finally, the effect of VAge is less intuitive due to the two peaks. It is known that older vehicles are less safe, but the latter does not explain the decrease from year 7 on. Claim frequency Model term Estimate Std. Error t statistic p value (Intercept) < 0.01 SexInsured - Male < 0.01 PrivateCar < 0.01 NCD_ < 0.01 Cover_C < 0.01 Claim probability Model term Estimate Std. Error t statistic p value (Intercept) < 0.01 SexInsured - Male < 0.01 PrivateCar < 0.01 NCD_ < 0.01 Cover_C < 0.01 Conditional claim frequency Model term Estimate Std. Error t statistic p value (Intercept) < 0.01 SexInsured - Male NCD_ < 0.01 Cover_C < 0.01 Table 3.3: Parameter estimates and corresponding statistics for the claim frequency, claim probability and conditional claim frequency. Next, we discuss the parameteric model terms as displayed in Table 3.3. It is clear that male

Subject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018

Subject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018 ` Subject CS1 Actuarial Statistics 1 Core Principles Syllabus for the 2019 exams 1 June 2018 Copyright in this Core Reading is the property of the Institute and Faculty of Actuaries who are the sole distributors.

More information

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is: **BEGINNING OF EXAMINATION** 1. You are given: (i) A random sample of five observations from a population is: 0.2 0.7 0.9 1.1 1.3 (ii) You use the Kolmogorov-Smirnov test for testing the null hypothesis,

More information

Financial Risk Forecasting Chapter 9 Extreme Value Theory

Financial Risk Forecasting Chapter 9 Extreme Value Theory Financial Risk Forecasting Chapter 9 Extreme Value Theory Jon Danielsson 2017 London School of Economics To accompany Financial Risk Forecasting www.financialriskforecasting.com Published by Wiley 2011

More information

1. You are given the following information about a stationary AR(2) model:

1. You are given the following information about a stationary AR(2) model: Fall 2003 Society of Actuaries **BEGINNING OF EXAMINATION** 1. You are given the following information about a stationary AR(2) model: (i) ρ 1 = 05. (ii) ρ 2 = 01. Determine φ 2. (A) 0.2 (B) 0.1 (C) 0.4

More information

Window Width Selection for L 2 Adjusted Quantile Regression

Window Width Selection for L 2 Adjusted Quantile Regression Window Width Selection for L 2 Adjusted Quantile Regression Yoonsuh Jung, The Ohio State University Steven N. MacEachern, The Ohio State University Yoonkyung Lee, The Ohio State University Technical Report

More information

Modeling. joint work with Jed Frees, U of Wisconsin - Madison. Travelers PASG (Predictive Analytics Study Group) Seminar Tuesday, 12 April 2016

Modeling. joint work with Jed Frees, U of Wisconsin - Madison. Travelers PASG (Predictive Analytics Study Group) Seminar Tuesday, 12 April 2016 joint work with Jed Frees, U of Wisconsin - Madison Travelers PASG (Predictive Analytics Study Group) Seminar Tuesday, 12 April 2016 claim Department of Mathematics University of Connecticut Storrs, Connecticut

More information

Keywords Akiake Information criterion, Automobile, Bonus-Malus, Exponential family, Linear regression, Residuals, Scaled deviance. I.

Keywords Akiake Information criterion, Automobile, Bonus-Malus, Exponential family, Linear regression, Residuals, Scaled deviance. I. Application of the Generalized Linear Models in Actuarial Framework BY MURWAN H. M. A. SIDDIG School of Mathematics, Faculty of Engineering Physical Science, The University of Manchester, Oxford Road,

More information

ADVANCED OPERATIONAL RISK MODELLING IN BANKS AND INSURANCE COMPANIES

ADVANCED OPERATIONAL RISK MODELLING IN BANKS AND INSURANCE COMPANIES Small business banking and financing: a global perspective Cagliari, 25-26 May 2007 ADVANCED OPERATIONAL RISK MODELLING IN BANKS AND INSURANCE COMPANIES C. Angela, R. Bisignani, G. Masala, M. Micocci 1

More information

Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making

Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making Case Study: Heavy-Tailed Distribution and Reinsurance Rate-making May 30, 2016 The purpose of this case study is to give a brief introduction to a heavy-tailed distribution and its distinct behaviors in

More information

Introduction to Algorithmic Trading Strategies Lecture 8

Introduction to Algorithmic Trading Strategies Lecture 8 Introduction to Algorithmic Trading Strategies Lecture 8 Risk Management Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com Outline Value at Risk (VaR) Extreme Value Theory (EVT) References

More information

Homework Problems Stat 479

Homework Problems Stat 479 Chapter 2 1. Model 1 is a uniform distribution from 0 to 100. Determine the table entries for a generalized uniform distribution covering the range from a to b where a < b. 2. Let X be a discrete random

More information

Practice Exam 1. Loss Amount Number of Losses

Practice Exam 1. Loss Amount Number of Losses Practice Exam 1 1. You are given the following data on loss sizes: An ogive is used as a model for loss sizes. Determine the fitted median. Loss Amount Number of Losses 0 1000 5 1000 5000 4 5000 10000

More information

Dependence Structure and Extreme Comovements in International Equity and Bond Markets

Dependence Structure and Extreme Comovements in International Equity and Bond Markets Dependence Structure and Extreme Comovements in International Equity and Bond Markets René Garcia Edhec Business School, Université de Montréal, CIRANO and CIREQ Georges Tsafack Suffolk University Measuring

More information

Random Variables and Probability Distributions

Random Variables and Probability Distributions Chapter 3 Random Variables and Probability Distributions Chapter Three Random Variables and Probability Distributions 3. Introduction An event is defined as the possible outcome of an experiment. In engineering

More information

Financial Risk Management

Financial Risk Management Financial Risk Management Professor: Thierry Roncalli Evry University Assistant: Enareta Kurtbegu Evry University Tutorial exercices #4 1 Correlation and copulas 1. The bivariate Gaussian copula is given

More information

A Multivariate Analysis of Intercompany Loss Triangles

A Multivariate Analysis of Intercompany Loss Triangles A Multivariate Analysis of Intercompany Loss Triangles Peng Shi School of Business University of Wisconsin-Madison ASTIN Colloquium May 21-24, 2013 Peng Shi (Wisconsin School of Business) Intercompany

More information

Market Risk Analysis Volume I

Market Risk Analysis Volume I Market Risk Analysis Volume I Quantitative Methods in Finance Carol Alexander John Wiley & Sons, Ltd List of Figures List of Tables List of Examples Foreword Preface to Volume I xiii xvi xvii xix xxiii

More information

SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS

SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS Questions 1-307 have been taken from the previous set of Exam C sample questions. Questions no longer relevant

More information

Modelling Environmental Extremes

Modelling Environmental Extremes 19th TIES Conference, Kelowna, British Columbia 8th June 2008 Topics for the day 1. Classical models and threshold models 2. Dependence and non stationarity 3. R session: weather extremes 4. Multivariate

More information

Modelling Environmental Extremes

Modelling Environmental Extremes 19th TIES Conference, Kelowna, British Columbia 8th June 2008 Topics for the day 1. Classical models and threshold models 2. Dependence and non stationarity 3. R session: weather extremes 4. Multivariate

More information

EVA Tutorial #1 BLOCK MAXIMA APPROACH IN HYDROLOGIC/CLIMATE APPLICATIONS. Rick Katz

EVA Tutorial #1 BLOCK MAXIMA APPROACH IN HYDROLOGIC/CLIMATE APPLICATIONS. Rick Katz 1 EVA Tutorial #1 BLOCK MAXIMA APPROACH IN HYDROLOGIC/CLIMATE APPLICATIONS Rick Katz Institute for Mathematics Applied to Geosciences National Center for Atmospheric Research Boulder, CO USA email: rwk@ucar.edu

More information

Risk Classification In Non-Life Insurance

Risk Classification In Non-Life Insurance Risk Classification In Non-Life Insurance Katrien Antonio Jan Beirlant November 28, 2006 Abstract Within the actuarial profession a major challenge can be found in the construction of a fair tariff structure.

More information

Modelling Joint Distribution of Returns. Dr. Sawsan Hilal space

Modelling Joint Distribution of Returns. Dr. Sawsan Hilal space Modelling Joint Distribution of Returns Dr. Sawsan Hilal space Maths Department - University of Bahrain space October 2011 REWARD Asset Allocation Problem PORTFOLIO w 1 w 2 w 3 ASSET 1 ASSET 2 R 1 R 2

More information

An Application of Extreme Value Theory for Measuring Financial Risk in the Uruguayan Pension Fund 1

An Application of Extreme Value Theory for Measuring Financial Risk in the Uruguayan Pension Fund 1 An Application of Extreme Value Theory for Measuring Financial Risk in the Uruguayan Pension Fund 1 Guillermo Magnou 23 January 2016 Abstract Traditional methods for financial risk measures adopts normal

More information

SYSM 6304 Risk and Decision Analysis Lecture 2: Fitting Distributions to Data

SYSM 6304 Risk and Decision Analysis Lecture 2: Fitting Distributions to Data SYSM 6304 Risk and Decision Analysis Lecture 2: Fitting Distributions to Data M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu September 5, 2015

More information

EX-POST VERIFICATION OF PREDICTION MODELS OF WAGE DISTRIBUTIONS

EX-POST VERIFICATION OF PREDICTION MODELS OF WAGE DISTRIBUTIONS EX-POST VERIFICATION OF PREDICTION MODELS OF WAGE DISTRIBUTIONS LUBOŠ MAREK, MICHAL VRABEC University of Economics, Prague, Faculty of Informatics and Statistics, Department of Statistics and Probability,

More information

Asymmetric Price Transmission: A Copula Approach

Asymmetric Price Transmission: A Copula Approach Asymmetric Price Transmission: A Copula Approach Feng Qiu University of Alberta Barry Goodwin North Carolina State University August, 212 Prepared for the AAEA meeting in Seattle Outline Asymmetric price

More information

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions. ME3620 Theory of Engineering Experimentation Chapter III. Random Variables and Probability Distributions Chapter III 1 3.2 Random Variables In an experiment, a measurement is usually denoted by a variable

More information

REINSURANCE RATE-MAKING WITH PARAMETRIC AND NON-PARAMETRIC MODELS

REINSURANCE RATE-MAKING WITH PARAMETRIC AND NON-PARAMETRIC MODELS REINSURANCE RATE-MAKING WITH PARAMETRIC AND NON-PARAMETRIC MODELS By Siqi Chen, Madeleine Min Jing Leong, Yuan Yuan University of Illinois at Urbana-Champaign 1. Introduction Reinsurance contract is an

More information

The mean-variance portfolio choice framework and its generalizations

The mean-variance portfolio choice framework and its generalizations The mean-variance portfolio choice framework and its generalizations Prof. Massimo Guidolin 20135 Theory of Finance, Part I (Sept. October) Fall 2014 Outline and objectives The backward, three-step solution

More information

Modeling of Price. Ximing Wu Texas A&M University

Modeling of Price. Ximing Wu Texas A&M University Modeling of Price Ximing Wu Texas A&M University As revenue is given by price times yield, farmers income risk comes from risk in yield and output price. Their net profit also depends on input price, but

More information

Homework Problems Stat 479

Homework Problems Stat 479 Chapter 10 91. * A random sample, X1, X2,, Xn, is drawn from a distribution with a mean of 2/3 and a variance of 1/18. ˆ = (X1 + X2 + + Xn)/(n-1) is the estimator of the distribution mean θ. Find MSE(

More information

MODELLING OF INCOME AND WAGE DISTRIBUTION USING THE METHOD OF L-MOMENTS OF PARAMETER ESTIMATION

MODELLING OF INCOME AND WAGE DISTRIBUTION USING THE METHOD OF L-MOMENTS OF PARAMETER ESTIMATION International Days of Statistics and Economics, Prague, September -3, MODELLING OF INCOME AND WAGE DISTRIBUTION USING THE METHOD OF L-MOMENTS OF PARAMETER ESTIMATION Diana Bílková Abstract Using L-moments

More information

Subject CS2A Risk Modelling and Survival Analysis Core Principles

Subject CS2A Risk Modelling and Survival Analysis Core Principles ` Subject CS2A Risk Modelling and Survival Analysis Core Principles Syllabus for the 2019 exams 1 June 2018 Copyright in this Core Reading is the property of the Institute and Faculty of Actuaries who

More information

An Improved Skewness Measure

An Improved Skewness Measure An Improved Skewness Measure Richard A. Groeneveld Professor Emeritus, Department of Statistics Iowa State University ragroeneveld@valley.net Glen Meeden School of Statistics University of Minnesota Minneapolis,

More information

Continuous random variables

Continuous random variables Continuous random variables probability density function (f(x)) the probability distribution function of a continuous random variable (analogous to the probability mass function for a discrete random variable),

More information

Modelling catastrophic risk in international equity markets: An extreme value approach. JOHN COTTER University College Dublin

Modelling catastrophic risk in international equity markets: An extreme value approach. JOHN COTTER University College Dublin Modelling catastrophic risk in international equity markets: An extreme value approach JOHN COTTER University College Dublin Abstract: This letter uses the Block Maxima Extreme Value approach to quantify

More information

Copulas? What copulas? R. Chicheportiche & J.P. Bouchaud, CFM

Copulas? What copulas? R. Chicheportiche & J.P. Bouchaud, CFM Copulas? What copulas? R. Chicheportiche & J.P. Bouchaud, CFM Multivariate linear correlations Standard tool in risk management/portfolio optimisation: the covariance matrix R ij = r i r j Find the portfolio

More information

MEASURING PORTFOLIO RISKS USING CONDITIONAL COPULA-AR-GARCH MODEL

MEASURING PORTFOLIO RISKS USING CONDITIONAL COPULA-AR-GARCH MODEL MEASURING PORTFOLIO RISKS USING CONDITIONAL COPULA-AR-GARCH MODEL Isariya Suttakulpiboon MSc in Risk Management and Insurance Georgia State University, 30303 Atlanta, Georgia Email: suttakul.i@gmail.com,

More information

Modeling Co-movements and Tail Dependency in the International Stock Market via Copulae

Modeling Co-movements and Tail Dependency in the International Stock Market via Copulae Modeling Co-movements and Tail Dependency in the International Stock Market via Copulae Katja Ignatieva, Eckhard Platen Bachelier Finance Society World Congress 22-26 June 2010, Toronto K. Ignatieva, E.

More information

Supplemental Online Appendix to Han and Hong, Understanding In-House Transactions in the Real Estate Brokerage Industry

Supplemental Online Appendix to Han and Hong, Understanding In-House Transactions in the Real Estate Brokerage Industry Supplemental Online Appendix to Han and Hong, Understanding In-House Transactions in the Real Estate Brokerage Industry Appendix A: An Agent-Intermediated Search Model Our motivating theoretical framework

More information

Lecture 5: Fundamentals of Statistical Analysis and Distributions Derived from Normal Distributions

Lecture 5: Fundamentals of Statistical Analysis and Distributions Derived from Normal Distributions Lecture 5: Fundamentals of Statistical Analysis and Distributions Derived from Normal Distributions ELE 525: Random Processes in Information Systems Hisashi Kobayashi Department of Electrical Engineering

More information

AN EXTREME VALUE APPROACH TO PRICING CREDIT RISK

AN EXTREME VALUE APPROACH TO PRICING CREDIT RISK AN EXTREME VALUE APPROACH TO PRICING CREDIT RISK SOFIA LANDIN Master s thesis 2018:E69 Faculty of Engineering Centre for Mathematical Sciences Mathematical Statistics CENTRUM SCIENTIARUM MATHEMATICARUM

More information

Intro to GLM Day 2: GLM and Maximum Likelihood

Intro to GLM Day 2: GLM and Maximum Likelihood Intro to GLM Day 2: GLM and Maximum Likelihood Federico Vegetti Central European University ECPR Summer School in Methods and Techniques 1 / 32 Generalized Linear Modeling 3 steps of GLM 1. Specify the

More information

Paper Series of Risk Management in Financial Institutions

Paper Series of Risk Management in Financial Institutions - December, 007 Paper Series of Risk Management in Financial Institutions The Effect of the Choice of the Loss Severity Distribution and the Parameter Estimation Method on Operational Risk Measurement*

More information

Log-Robust Portfolio Management

Log-Robust Portfolio Management Log-Robust Portfolio Management Dr. Aurélie Thiele Lehigh University Joint work with Elcin Cetinkaya and Ban Kawas Research partially supported by the National Science Foundation Grant CMMI-0757983 Dr.

More information

Exam STAM Practice Exam #1

Exam STAM Practice Exam #1 !!!! Exam STAM Practice Exam #1 These practice exams should be used during the month prior to your exam. This practice exam contains 20 questions, of equal value, corresponding to about a 2 hour exam.

More information

UNIVERSITY OF OSLO. Please make sure that your copy of the problem set is complete before you attempt to answer anything.

UNIVERSITY OF OSLO. Please make sure that your copy of the problem set is complete before you attempt to answer anything. UNIVERSITY OF OSLO Faculty of Mathematics and Natural Sciences Examination in: STK4540 Non-Life Insurance Mathematics Day of examination: Wednesday, December 4th, 2013 Examination hours: 14.30 17.30 This

More information

Contents Part I Descriptive Statistics 1 Introduction and Framework Population, Sample, and Observations Variables Quali

Contents Part I Descriptive Statistics 1 Introduction and Framework Population, Sample, and Observations Variables Quali Part I Descriptive Statistics 1 Introduction and Framework... 3 1.1 Population, Sample, and Observations... 3 1.2 Variables.... 4 1.2.1 Qualitative and Quantitative Variables.... 5 1.2.2 Discrete and Continuous

More information

3.4 Copula approach for modeling default dependency. Two aspects of modeling the default times of several obligors

3.4 Copula approach for modeling default dependency. Two aspects of modeling the default times of several obligors 3.4 Copula approach for modeling default dependency Two aspects of modeling the default times of several obligors 1. Default dynamics of a single obligor. 2. Model the dependence structure of defaults

More information

IEOR E4602: Quantitative Risk Management

IEOR E4602: Quantitative Risk Management IEOR E4602: Quantitative Risk Management Basic Concepts and Techniques of Risk Management Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Measuring Financial Risk using Extreme Value Theory: evidence from Pakistan

Measuring Financial Risk using Extreme Value Theory: evidence from Pakistan Measuring Financial Risk using Extreme Value Theory: evidence from Pakistan Dr. Abdul Qayyum and Faisal Nawaz Abstract The purpose of the paper is to show some methods of extreme value theory through analysis

More information

Mongolia s TOP-20 Index Risk Analysis, Pt. 3

Mongolia s TOP-20 Index Risk Analysis, Pt. 3 Mongolia s TOP-20 Index Risk Analysis, Pt. 3 Federico M. Massari March 12, 2017 In the third part of our risk report on TOP-20 Index, Mongolia s main stock market indicator, we focus on modelling the right

More information

CHOICE THEORY, UTILITY FUNCTIONS AND RISK AVERSION

CHOICE THEORY, UTILITY FUNCTIONS AND RISK AVERSION CHOICE THEORY, UTILITY FUNCTIONS AND RISK AVERSION Szabolcs Sebestyén szabolcs.sebestyen@iscte.pt Master in Finance INVESTMENTS Sebestyén (ISCTE-IUL) Choice Theory Investments 1 / 65 Outline 1 An Introduction

More information

Fitting financial time series returns distributions: a mixture normality approach

Fitting financial time series returns distributions: a mixture normality approach Fitting financial time series returns distributions: a mixture normality approach Riccardo Bramante and Diego Zappa * Abstract Value at Risk has emerged as a useful tool to risk management. A relevant

More information

Capital Allocation Principles

Capital Allocation Principles Capital Allocation Principles Maochao Xu Department of Mathematics Illinois State University mxu2@ilstu.edu Capital Dhaene, et al., 2011, Journal of Risk and Insurance The level of the capital held by

More information

Statistical Tables Compiled by Alan J. Terry

Statistical Tables Compiled by Alan J. Terry Statistical Tables Compiled by Alan J. Terry School of Science and Sport University of the West of Scotland Paisley, Scotland Contents Table 1: Cumulative binomial probabilities Page 1 Table 2: Cumulative

More information

UPDATED IAA EDUCATION SYLLABUS

UPDATED IAA EDUCATION SYLLABUS II. UPDATED IAA EDUCATION SYLLABUS A. Supporting Learning Areas 1. STATISTICS Aim: To enable students to apply core statistical techniques to actuarial applications in insurance, pensions and emerging

More information

Risk Measurement in Credit Portfolio Models

Risk Measurement in Credit Portfolio Models 9 th DGVFM Scientific Day 30 April 2010 1 Risk Measurement in Credit Portfolio Models 9 th DGVFM Scientific Day 30 April 2010 9 th DGVFM Scientific Day 30 April 2010 2 Quantitative Risk Management Profit

More information

Computational Statistics Handbook with MATLAB

Computational Statistics Handbook with MATLAB «H Computer Science and Data Analysis Series Computational Statistics Handbook with MATLAB Second Edition Wendy L. Martinez The Office of Naval Research Arlington, Virginia, U.S.A. Angel R. Martinez Naval

More information

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg :

More information

PIVOTAL QUANTILE ESTIMATES IN VAR CALCULATIONS. Peter Schaller, Bank Austria Creditanstalt (BA-CA) Wien,

PIVOTAL QUANTILE ESTIMATES IN VAR CALCULATIONS. Peter Schaller, Bank Austria Creditanstalt (BA-CA) Wien, PIVOTAL QUANTILE ESTIMATES IN VAR CALCULATIONS Peter Schaller, Bank Austria Creditanstalt (BA-CA) Wien, peter@ca-risc.co.at c Peter Schaller, BA-CA, Strategic Riskmanagement 1 Contents Some aspects of

More information

Probability and Statistics

Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be CHAPTER 3: PARAMETRIC FAMILIES OF UNIVARIATE DISTRIBUTIONS 1 Why do we need distributions?

More information

CHAPTER II LITERATURE STUDY

CHAPTER II LITERATURE STUDY CHAPTER II LITERATURE STUDY 2.1. Risk Management Monetary crisis that strike Indonesia during 1998 and 1999 has caused bad impact to numerous government s and commercial s bank. Most of those banks eventually

More information

M249 Diagnostic Quiz

M249 Diagnostic Quiz THE OPEN UNIVERSITY Faculty of Mathematics and Computing M249 Diagnostic Quiz Prepared by the Course Team [Press to begin] c 2005, 2006 The Open University Last Revision Date: May 19, 2006 Version 4.2

More information

Lecture 2. Probability Distributions Theophanis Tsandilas

Lecture 2. Probability Distributions Theophanis Tsandilas Lecture 2 Probability Distributions Theophanis Tsandilas Comment on measures of dispersion Why do common measures of dispersion (variance and standard deviation) use sums of squares: nx (x i ˆµ) 2 i=1

More information

2. Copula Methods Background

2. Copula Methods Background 1. Introduction Stock futures markets provide a channel for stock holders potentially transfer risks. Effectiveness of such a hedging strategy relies heavily on the accuracy of hedge ratio estimation.

More information

High-Frequency Data Analysis and Market Microstructure [Tsay (2005), chapter 5]

High-Frequency Data Analysis and Market Microstructure [Tsay (2005), chapter 5] 1 High-Frequency Data Analysis and Market Microstructure [Tsay (2005), chapter 5] High-frequency data have some unique characteristics that do not appear in lower frequencies. At this class we have: Nonsynchronous

More information

Contents. An Overview of Statistical Applications CHAPTER 1. Contents (ix) Preface... (vii)

Contents. An Overview of Statistical Applications CHAPTER 1. Contents (ix) Preface... (vii) Contents (ix) Contents Preface... (vii) CHAPTER 1 An Overview of Statistical Applications 1.1 Introduction... 1 1. Probability Functions and Statistics... 1..1 Discrete versus Continuous Functions... 1..

More information

[D7] PROBABILITY DISTRIBUTION OF OUTSTANDING LIABILITY FROM INDIVIDUAL PAYMENTS DATA Contributed by T S Wright

[D7] PROBABILITY DISTRIBUTION OF OUTSTANDING LIABILITY FROM INDIVIDUAL PAYMENTS DATA Contributed by T S Wright Faculty and Institute of Actuaries Claims Reserving Manual v.2 (09/1997) Section D7 [D7] PROBABILITY DISTRIBUTION OF OUTSTANDING LIABILITY FROM INDIVIDUAL PAYMENTS DATA Contributed by T S Wright 1. Introduction

More information

Optimal Allocation of Policy Limits and Deductibles

Optimal Allocation of Policy Limits and Deductibles Optimal Allocation of Policy Limits and Deductibles Ka Chun Cheung Email: kccheung@math.ucalgary.ca Tel: +1-403-2108697 Fax: +1-403-2825150 Department of Mathematics and Statistics, University of Calgary,

More information

Probability Weighted Moments. Andrew Smith

Probability Weighted Moments. Andrew Smith Probability Weighted Moments Andrew Smith andrewdsmith8@deloitte.co.uk 28 November 2014 Introduction If I asked you to summarise a data set, or fit a distribution You d probably calculate the mean and

More information

Statistical Analysis of Life Insurance Policy Termination and Survivorship

Statistical Analysis of Life Insurance Policy Termination and Survivorship Statistical Analysis of Life Insurance Policy Termination and Survivorship Emiliano A. Valdez, PhD, FSA Michigan State University joint work with J. Vadiveloo and U. Dias Sunway University, Malaysia Kuala

More information

the display, exploration and transformation of the data are demonstrated and biases typically encountered are highlighted.

the display, exploration and transformation of the data are demonstrated and biases typically encountered are highlighted. 1 Insurance data Generalized linear modeling is a methodology for modeling relationships between variables. It generalizes the classical normal linear model, by relaxing some of its restrictive assumptions,

More information

Can we use kernel smoothing to estimate Value at Risk and Tail Value at Risk?

Can we use kernel smoothing to estimate Value at Risk and Tail Value at Risk? Can we use kernel smoothing to estimate Value at Risk and Tail Value at Risk? Ramon Alemany, Catalina Bolancé and Montserrat Guillén Riskcenter - IREA Universitat de Barcelona http://www.ub.edu/riskcenter

More information

Measuring Risk Dependencies in the Solvency II-Framework. Robert Danilo Molinari Tristan Nguyen WHL Graduate School of Business and Economics

Measuring Risk Dependencies in the Solvency II-Framework. Robert Danilo Molinari Tristan Nguyen WHL Graduate School of Business and Economics Measuring Risk Dependencies in the Solvency II-Framework Robert Danilo Molinari Tristan Nguyen WHL Graduate School of Business and Economics 1 Overview 1. Introduction 2. Dependency ratios 3. Copulas 4.

More information

Monotone, Convex and Extrema

Monotone, Convex and Extrema Monotone Functions Function f is called monotonically increasing, if Chapter 8 Monotone, Convex and Extrema x x 2 f (x ) f (x 2 ) It is called strictly monotonically increasing, if f (x 2) f (x ) x < x

More information

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi Chapter 4: Commonly Used Distributions Statistics for Engineers and Scientists Fourth Edition William Navidi 2014 by Education. This is proprietary material solely for authorized instructor use. Not authorized

More information

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی یادگیري ماشین توزیع هاي نمونه و تخمین نقطه اي پارامترها Sampling Distributions and Point Estimation of Parameter (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی درس هفتم 1 Outline Introduction

More information

KURTOSIS OF THE LOGISTIC-EXPONENTIAL SURVIVAL DISTRIBUTION

KURTOSIS OF THE LOGISTIC-EXPONENTIAL SURVIVAL DISTRIBUTION KURTOSIS OF THE LOGISTIC-EXPONENTIAL SURVIVAL DISTRIBUTION Paul J. van Staden Department of Statistics University of Pretoria Pretoria, 0002, South Africa paul.vanstaden@up.ac.za http://www.up.ac.za/pauljvanstaden

More information

discussion Papers Some Flexible Parametric Models for Partially Adaptive Estimators of Econometric Models

discussion Papers Some Flexible Parametric Models for Partially Adaptive Estimators of Econometric Models discussion Papers Discussion Paper 2007-13 March 26, 2007 Some Flexible Parametric Models for Partially Adaptive Estimators of Econometric Models Christian B. Hansen Graduate School of Business at the

More information

Operational Risk Aggregation

Operational Risk Aggregation Operational Risk Aggregation Professor Carol Alexander Chair of Risk Management and Director of Research, ISMA Centre, University of Reading, UK. Loss model approaches are currently a focus of operational

More information

Session 5. Predictive Modeling in Life Insurance

Session 5. Predictive Modeling in Life Insurance SOA Predictive Analytics Seminar Hong Kong 29 Aug. 2018 Hong Kong Session 5 Predictive Modeling in Life Insurance Jingyi Zhang, Ph.D Predictive Modeling in Life Insurance JINGYI ZHANG PhD Scientist Global

More information

Bloomberg. Portfolio Value-at-Risk. Sridhar Gollamudi & Bryan Weber. September 22, Version 1.0

Bloomberg. Portfolio Value-at-Risk. Sridhar Gollamudi & Bryan Weber. September 22, Version 1.0 Portfolio Value-at-Risk Sridhar Gollamudi & Bryan Weber September 22, 2011 Version 1.0 Table of Contents 1 Portfolio Value-at-Risk 2 2 Fundamental Factor Models 3 3 Valuation methodology 5 3.1 Linear factor

More information

Institute of Actuaries of India Subject CT6 Statistical Methods

Institute of Actuaries of India Subject CT6 Statistical Methods Institute of Actuaries of India Subject CT6 Statistical Methods For 2014 Examinations Aim The aim of the Statistical Methods subject is to provide a further grounding in mathematical and statistical techniques

More information

Logit Models for Binary Data

Logit Models for Binary Data Chapter 3 Logit Models for Binary Data We now turn our attention to regression models for dichotomous data, including logistic regression and probit analysis These models are appropriate when the response

More information

On Some Test Statistics for Testing the Population Skewness and Kurtosis: An Empirical Study

On Some Test Statistics for Testing the Population Skewness and Kurtosis: An Empirical Study Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 8-26-2016 On Some Test Statistics for Testing the Population Skewness and Kurtosis:

More information

Cambridge University Press Risk Modelling in General Insurance: From Principles to Practice Roger J. Gray and Susan M.

Cambridge University Press Risk Modelling in General Insurance: From Principles to Practice Roger J. Gray and Susan M. adjustment coefficient, 272 and Cramér Lundberg approximation, 302 existence, 279 and Lundberg s inequality, 272 numerical methods for, 303 properties, 272 and reinsurance (case study), 348 statistical

More information

Dynamic Replication of Non-Maturing Assets and Liabilities

Dynamic Replication of Non-Maturing Assets and Liabilities Dynamic Replication of Non-Maturing Assets and Liabilities Michael Schürle Institute for Operations Research and Computational Finance, University of St. Gallen, Bodanstr. 6, CH-9000 St. Gallen, Switzerland

More information

Lecture 6: Non Normal Distributions

Lecture 6: Non Normal Distributions Lecture 6: Non Normal Distributions and their Uses in GARCH Modelling Prof. Massimo Guidolin 20192 Financial Econometrics Spring 2015 Overview Non-normalities in (standardized) residuals from asset return

More information

Characterization of the Optimum

Characterization of the Optimum ECO 317 Economics of Uncertainty Fall Term 2009 Notes for lectures 5. Portfolio Allocation with One Riskless, One Risky Asset Characterization of the Optimum Consider a risk-averse, expected-utility-maximizing

More information

The Bernoulli distribution

The Bernoulli distribution This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Operational Risk: Evidence, Estimates and Extreme Values from Austria

Operational Risk: Evidence, Estimates and Extreme Values from Austria Operational Risk: Evidence, Estimates and Extreme Values from Austria Stefan Kerbl OeNB / ECB 3 rd EBA Policy Research Workshop, London 25 th November 2014 Motivation Operational Risk as the exotic risk

More information

Labor Economics Field Exam Spring 2014

Labor Economics Field Exam Spring 2014 Labor Economics Field Exam Spring 2014 Instructions You have 4 hours to complete this exam. This is a closed book examination. No written materials are allowed. You can use a calculator. THE EXAM IS COMPOSED

More information

Chapter 5. Statistical inference for Parametric Models

Chapter 5. Statistical inference for Parametric Models Chapter 5. Statistical inference for Parametric Models Outline Overview Parameter estimation Method of moments How good are method of moments estimates? Interval estimation Statistical Inference for Parametric

More information

An Information Based Methodology for the Change Point Problem Under the Non-central Skew t Distribution with Applications.

An Information Based Methodology for the Change Point Problem Under the Non-central Skew t Distribution with Applications. An Information Based Methodology for the Change Point Problem Under the Non-central Skew t Distribution with Applications. Joint with Prof. W. Ning & Prof. A. K. Gupta. Department of Mathematics and Statistics

More information

2.1 Random variable, density function, enumerative density function and distribution function

2.1 Random variable, density function, enumerative density function and distribution function Risk Theory I Prof. Dr. Christian Hipp Chair for Science of Insurance, University of Karlsruhe (TH Karlsruhe) Contents 1 Introduction 1.1 Overview on the insurance industry 1.1.1 Insurance in Benin 1.1.2

More information

Some Characteristics of Data

Some Characteristics of Data Some Characteristics of Data Not all data is the same, and depending on some characteristics of a particular dataset, there are some limitations as to what can and cannot be done with that data. Some key

More information

P VaR0.01 (X) > 2 VaR 0.01 (X). (10 p) Problem 4

P VaR0.01 (X) > 2 VaR 0.01 (X). (10 p) Problem 4 KTH Mathematics Examination in SF2980 Risk Management, December 13, 2012, 8:00 13:00. Examiner : Filip indskog, tel. 790 7217, e-mail: lindskog@kth.se Allowed technical aids and literature : a calculator,

More information

FE670 Algorithmic Trading Strategies. Stevens Institute of Technology

FE670 Algorithmic Trading Strategies. Stevens Institute of Technology FE670 Algorithmic Trading Strategies Lecture 4. Cross-Sectional Models and Trading Strategies Steve Yang Stevens Institute of Technology 09/26/2013 Outline 1 Cross-Sectional Methods for Evaluation of Factor

More information