A NOTE ON FULL CREDIBILITY FOR ESTIMATING CLAIM FREQUENCY

Similar documents
EDUCATION COMMITTEE OF THE SOCIETY OF ACTUARIES SHORT-TERM ACTUARIAL MATHEMATICS STUDY NOTE CHAPTER 8 FROM

CAS Course 3 - Actuarial Models

[D7] PROBABILITY DISTRIBUTION OF OUTSTANDING LIABILITY FROM INDIVIDUAL PAYMENTS DATA Contributed by T S Wright

ECON 214 Elements of Statistics for Economists

1 Asset Pricing: Bonds vs Stocks

Week 2 Quantitative Analysis of Financial Markets Hypothesis Testing and Confidence Intervals

Stochastic Claims Reserving _ Methods in Insurance

Subject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018

On the Distribution and Its Properties of the Sum of a Normal and a Doubly Truncated Normal

Probability and Statistics

Institute of Actuaries of India Subject CT6 Statistical Methods

Review of the Topics for Midterm I

Contents. An Overview of Statistical Applications CHAPTER 1. Contents (ix) Preface... (vii)

2.1 Random variable, density function, enumerative density function and distribution function

Estimation Parameters and Modelling Zero Inflated Negative Binomial

Mathematical Methods in Risk Theory

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

Statistics 13 Elementary Statistics

Class 16. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Chapter 3 Statistical Quality Control, 7th Edition by Douglas C. Montgomery. Copyright (c) 2013 John Wiley & Sons, Inc.

ECON 214 Elements of Statistics for Economists 2016/2017

Chapter 7 presents the beginning of inferential statistics. The two major activities of inferential statistics are

SYLLABUS OF BASIC EDUCATION SPRING 2018 Construction and Evaluation of Actuarial Models Exam 4

An Application of Extreme Value Theory for Measuring Financial Risk in the Uruguayan Pension Fund 1

MULTISTAGE PORTFOLIO OPTIMIZATION AS A STOCHASTIC OPTIMAL CONTROL PROBLEM

Tolerance Intervals for Any Data (Nonparametric)

TABLE OF CONTENTS - VOLUME 2

Edgeworth Binomial Trees

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

Probability & Statistics

Continuous-Time Pension-Fund Modelling

The ruin probabilities of a multidimensional perturbed risk model

Key Objectives. Module 2: The Logic of Statistical Inference. Z-scores. SGSB Workshop: Using Statistical Data to Make Decisions

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

Session 178 TS, Stats for Health Actuaries. Moderator: Ian G. Duncan, FSA, FCA, FCIA, FIA, MAAA. Presenter: Joan C. Barrett, FSA, MAAA

The internal rate of return (IRR) is a venerable technique for evaluating deterministic cash flow streams.

ECO220Y Estimation: Confidence Interval Estimator for Sample Proportions Readings: Chapter 11 (skip 11.5)

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

10/1/2012. PSY 511: Advanced Statistics for Psychological and Behavioral Research 1

Contents Utility theory and insurance The individual risk model Collective risk models

Department of Agricultural Economics. PhD Qualifier Examination. August 2010

SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MATHEMATICS EXAM STAM SAMPLE QUESTIONS

Subject CS2A Risk Modelling and Survival Analysis Core Principles

PROBABILITY DISTRIBUTIONS

Probability. An intro for calculus students P= Figure 1: A normal integral

1/2 2. Mean & variance. Mean & standard deviation

Shifting our focus. We were studying statistics (data, displays, sampling...) The next few lectures focus on probability (randomness) Why?

Chapter 3 Common Families of Distributions. Definition 3.4.1: A family of pmfs or pdfs is called exponential family if it can be expressed as

A probability distribution shows the possible outcomes of an experiment and the probability of each of these outcomes.

Chapter 4 Probability Distributions

UNIT 4 MATHEMATICAL METHODS

Chapter 4 Random Variables & Probability. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables

Financial Risk Forecasting Chapter 9 Extreme Value Theory

Confidence Intervals for the Median and Other Percentiles

Expected utility inequalities: theory and applications

Continuous random variables

Modelling catastrophic risk in international equity markets: An extreme value approach. JOHN COTTER University College Dublin

I BASIC RATEMAKING TECHNIQUES

value BE.104 Spring Biostatistics: Distribution and the Mean J. L. Sherley

Characterization of the Optimum

Using Monte Carlo Analysis in Ecological Risk Assessments

A Markov Chain Monte Carlo Approach to Estimate the Risks of Extremely Large Insurance Claims

Exam P Flashcards exams. Key concepts. Important formulas. Efficient methods. Advice on exam technique

Confidence Intervals Introduction

M.Sc. ACTUARIAL SCIENCE. Term-End Examination

Statistical Intervals. Chapter 7 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Homework 1 posted, due Friday, September 30, 2 PM. Independence of random variables: We say that a collection of random variables

Financial Economics. Runs Test

Credibility. Chapters Stat Loss Models. Chapters (Stat 477) Credibility Brian Hartman - BYU 1 / 31

Chapter 7: Point Estimation and Sampling Distributions

ESTIMATION OF MODIFIED MEASURE OF SKEWNESS. Elsayed Ali Habib *

Mortality Rates Estimation Using Whittaker-Henderson Graduation Technique

A lower bound on seller revenue in single buyer monopoly auctions

Mathematics of Finance Final Preparation December 19. To be thoroughly prepared for the final exam, you should

Australian Journal of Basic and Applied Sciences. Conditional Maximum Likelihood Estimation For Survival Function Using Cox Model

The normal distribution is a theoretical model derived mathematically and not empirically.

Chapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables

ASTIN Colloquium Understanding Split Credibility. Ira Robbin, PhD AVP and Senior Pricing Actuary Endurance US Insurance Operations

Discrete-time Asset Pricing Models in Applied Stochastic Finance

Random Variables and Probability Distributions

Cambridge University Press Risk Modelling in General Insurance: From Principles to Practice Roger J. Gray and Susan M.

King Saud University Academic Year (G) College of Sciences Academic Year (H) Solutions of Homework 1 : Selected problems P exam

Exam STAM Practice Exam #1

Chapter 3 Discrete Random Variables and Probability Distributions

Overview. Definitions. Definitions. Graphs. Chapter 4 Probability Distributions. probability distributions

Statistics for Managers Using Microsoft Excel/SPSS Chapter 6 The Normal Distribution And Other Continuous Distributions

1. You are given the following information about a stationary AR(2) model:

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Chapter 8: Sampling distributions of estimators Sections

THE CHANGING SIZE DISTRIBUTION OF U.S. TRADE UNIONS AND ITS DESCRIPTION BY PARETO S DISTRIBUTION. John Pencavel. Mainz, June 2012

Statistics 431 Spring 2007 P. Shaman. Preliminaries

A First Course in Probability

GPD-POT and GEV block maxima

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION

Market Risk Analysis Volume I

Keller: Stats for Mgmt & Econ, 7th Ed July 17, 2006

CABARRUS COUNTY 2008 APPRAISAL MANUAL

Simulating Continuous Time Rating Transitions

The Credibility Theory applied to backtesting Counterparty Credit Risk. Matteo Formenti

Changes to Exams FM/2, M and C/4 for the May 2007 Administration

Transcription:

51 A NOTE ON FULL CREDIBILITY FOR ESTIMATING CLAIM FREQUENCY J. ERNEST HANSEN* The conventional standards for full credibility are known to be inadequate. This inadequacy has been well treated in the Mayerson, et. al. paper and in the ensuing discussions, where the general problem of estimating pure premium was considered. However, in spite of this previous treatment, that old, familiar number, 1,082, still enjoys widespread patronage. If, instead of estimating pure premium, we ignore claim severity and estimate only claim frequency, 1,082 claims, with the precision in estimation which it promises, is an acceptable standard, providing we are sampling from a homogeneous risk population and accept the usual assumption of mutual independence among risks having Poisson claim processes. However, we know the insureds are not a homogeneous population. We must provide for a distribution of the Poisson parameter over the population, referred to as the structure function. The structure function introduces additional variation into the claim process which reduces the precision of estimation promised by the conventional credibility standards. The More General Model Let m denote the number of claims of an insured selected at random, and let h denote the parameter of his Poisson claim process for a given interval of time, i.e., the experience period. Then the unconditional probability distribution of m can be represented as: *J. Ernest Hansen has submitted this paper in response to a presidential invitation. yierznsen is a Research Associate with the Insurance Company of North 1 A. L. Mayerson, D. A. Jones, N. L. Bowers, Jr., On the Credibility of the Pure Premium, PCAS Vol. LV, p. 175. 2 PCAS Vol. LVI, pp. 63-82. 3 H. Biihlmann, Mathematical Methods in Risk Theory, Springer-Verlag, New York, 1970, p. 65.

I 52 CREDIRILITY P(m) = F(mlh)f(A)dh, m = 0, I, 2, * * * s A=0 where P(mlA) is the Poisson claim process of an individual insured conditional upon A, and f(a) is the structure function describing the manner in which A is distributed over a population of insureds. The gamma distribution is often used as an example of a structure function, where we have: p(m) = s O,-AA -. &P e-lxx AI+-1 da A=0 m! (/I - I)! where (Y and p are the parameters for the gamma distribution. Upon integrating, we have: Ptm)= (m+p-i)! m! (p - I)! (*)P~~)m,m=o,I,2,... The above representation of P(m) is in the form of a negative binomial distribution. Therefore, if we assume that individual insureds have independent Poisson claim processes and that the Poisson parameter is gamma distributed over the population of insureds, and if we select insureds at random to observe m, then the number of claims from an insured, m has the negative binomial distribution. A number of researchers have found the negative binomial distribution satisfactorily fits automobile claims data4. From the general representation of P(m), again assuming a Poisson claim process for individual insureds, and using well known results from conditional probability, we can readily determine: Also: E(m) = EIEfm)lAA = E(A) Var(m) = Var[E(mjA)] + E[Var(m/A)] = Var(A) + E(A) 4 H. L. Seal, Stochastic Theory of (I Risk Business, John Wiley and Sons, Inc., New York, 1969, p. 16.

These results are immediate when we remember E(mIA) = Var(mlA) = A for the Poisson variable m with parameter A. Therefore, even though the mixed Poisson process P(m) can be mathematically difficult to work with, depending upon what structure function is selected, the mean and variance of m are simply related to those of A. Considering a random sample of II insureds, we have, by invoking the Central Limit Theorem, the consequence that the sample mean, F is approximately normally distributed with a mean of E(A) and a variance of [Var(A) + E(A)]/n. The exponential distribution, with only one parameter, is a convenient choice for f(a) for a numerical example. If E(h) =.35, i.e., we expect.35 claims per insured for the experience period, then Var(A) = E (h) =.1225 for an exponential distribution and FR is approximately normally distributed with the following parameter values: E(Z) = E(A) =.35 Var(Fi) = [Var(A) + E(A)]/n = (.1225 +.35)/n =.4725/n If we want the estimator Fi? to be within 5% of E(m) with probability.90, we determine n as follows:6 standard normal deviate, Z = 1.645 = iii - E(Z) ml.05 x.35 - -.4725 n n = 4,175 The number of claims we could expect in a sample of this size would be: n l E(m) = 4,175 x.35 = 1,461 Therefore, assuming an exponential structure function and an expectation of.35 claims per insured, we find that the standard for full credibility would be 1,461 claims in contrast to the conventional standard of 1,082 claims. The difference is attributable to the additional variation in m introduced by the structure function. 6 For an exponential structure function, iti is a maximum likelihood estimator.

54 CREI>IHII.ITY However, it is difficult to estimate the shape of the structure function for a particular population of insureds since an insured s risk parameter is not an observable random variable. We can observe the number of claims of a particular insured over time for purposes of estimating this risk parameter, the insured s expected claim frcqucncy, but the true expected claim frequency is never known with certainty. The conventional standards for full credibility are derived by assuming the structure function is concentrated at a single point,g i.e., the risk parameter A is assumed to be constant over the population of insured3 and, therefore, Var(A) = 0. If we reconsider the previous numerical example. with E(A) =.35, but assume the structure function is concentrated at this point, we have: Then : Var(m) = (0 +.35)/n 1.645 =.05 x.35.35-4 n n = 3,092 The number of claims we would expect in a sample of this size would be 3,092 x.35 = 1,082 claims. This is the answer we should have anticipated, the conventional standard for full credibility. The conventional standard, being adequate only for an extreme and improbable case, violates one of the basic purposes for employing the techniques of statistical inference. This is to establish procedures for estimation which guarantee a level of precision in the estimates, e.g., a probability of at least.90 of being within 5% of the true average claim frequency. The levels of precision associated with conventional standards represent the most precision possible using these standards rather than the least; we have a ceiling on possible precision when what we want is a floor. Choice of the Structure Function An ideal choice for a structure function is one that leads to gen- 0 Simon has used the term isohazardous to characterize such a population of insureds in his paper, The Negative Binomial and Poisson Distributions Compared, PCAS Vol. XLVII, p. 20.

CRIJI~IHII.I1 S 55 erally conservative standards for estimating claim frequency. Toward this end, we can use the following result from reliability theory: the coefficient of variation for all distributions which have an Increasing Failure Rate is bounded above by that of the exponential distribution. In particular, gamma distributions which may be used as structure functions have increasing failure rates. 8 For such distributions, Var(A) is maximized, and Vat-(m) = I/ar(A) + E(A) is maximized by assuming an exponential structure function for a given value of E(A). The maximum variance of m will then be: max. Var(m) = EP(h) + E(A) Credibility standards based on this variance will be adequate for the entire set of structure functions; the standards will be based on the maximum possible rather than the minimum possible variance. In practice, the actuary is sufficiently familiar with the data he works with to select an upper bound for E(A), the expected claim frequency. Then, using an exponential prior distribution, a more adequate standard for full credibility can be easily computed, as in the previous numerical example. Using the above expression for max. Var(m) and letting E denote the tolerance of error as a proportion of E(A), we can rearrange the formula of the example as: n=li-c!.ze E(A) 2 7 R. Barlow and F. Proschan, Matl~ematical Theory of Reliability, John Wiley and Sons, New York, 1965, p. 33. A distribution, f(x), is IFR, has an increasing failure rate, if f(x)/[i-f(x)] increases as x increases. If we restrict a population of insureds to lust those for which A > A., an arbitrary value of A, then the chance that an insured is close to A, is given by the conditional probability f(h,,)da/[i-ffh.)]. If f(a) is IFR, this conditional probability increases as A. increases. 8 Gamma distributions which are asymptotic to the vertical axis are not intuitively appealing candidates for structure functions.

56 CKI.I>IRILIT\ The following table was constructed using this formula with 2 = 1.645 and E =.05: Full Credibility Standards with a Tolerunce of Error of 5% und 90% Confidence Upper Bound for Sample Size: No. Expected No. Cluim Frequency of Exposure Units of Claims.05 22,73 1 1,137.I0 11,907 1,191.15 8,298 1,245.25 5,412 1,353.35 4,175 1,461.50 3,247 1,624.75 2,526 1,894 1.od 2,165 2,165 1.50 1,804 2,706 2.00 1,624 3,247 3.00 1,443 4,330 5.00 1,299 6,494 Conclusion.The conventional standards for full credibility are known to be minimal for the estimation of claim frequency. They are adequate only when the structure function is concentrated at a point. The exponential distribution appears to present a reasonable bounding case with respect to the additional variance introduced by the structure function. With the assumption of an exponential structure function and the selection of a maximum possible mean value for the Poisson risk parameter, an adequate sample size for estimating claim frequency can be computed.