WC-5 Just How Credible Is That Employer? Exploring GLMs and Multilevel Modeling for NCCI s Excess Loss Factor Methodology

Similar documents
UPDATED IAA EDUCATION SYLLABUS

NCCI s New ELF Methodology

Antitrust Notice. Copyright 2010 National Council on Compensation Insurance, Inc. All Rights Reserved.

Bayesian Trend Selection

Bayesian and Hierarchical Methods for Ratemaking

Subject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018

TABLE OF CONTENTS - VOLUME 2

And The Winner Is? How to Pick a Better Model

Exploring the Fundamental Insurance Equation

Obtaining Predictive Distributions for Reserves Which Incorporate Expert Opinions R. Verrall A. Estimation of Policy Liabilities

Computational Statistics Handbook with MATLAB

3/10/2014. Exploring the Fundamental Insurance Equation. CAS Antitrust Notice. Fundamental Insurance Equation

The Casualty Actuarial Society is committed to adhering strictly to the letter and spirit of the antitrust laws. Seminars conducted under the

2017 IAA EDUCATION SYLLABUS

Loss Cost Modeling vs. Frequency and Severity Modeling

Session 5. Predictive Modeling in Life Insurance

Institute of Actuaries of India Subject CT6 Statistical Methods

A Total Credibility Approach to Pool Reserving

SYLLABUS OF BASIC EDUCATION SPRING 2018 Construction and Evaluation of Actuarial Models Exam 4

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

Demand modeling for commercial lines: enhanced pricing, business projections, and customer experience. CAS RPM Seminar March 31, 2014

Individual Claims Reserving with Stan

And The Winner Is? How to Pick a Better Model

Predictive modelling around the world Peter Banthorpe, RGA Kevin Manning, Milliman

Workers Compensation Ratemaking An Overview

Maximum Likelihood Estimation

PRE CONFERENCE WORKSHOP 3

Model 0: We start with a linear regression model: log Y t = β 0 + β 1 (t 1980) + ε, with ε N(0,

Window Width Selection for L 2 Adjusted Quantile Regression

Evidence from Large Workers

Market Risk: FROM VALUE AT RISK TO STRESS TESTING. Agenda. Agenda (Cont.) Traditional Measures of Market Risk

Chapter 3. Numerical Descriptive Measures. Copyright 2016 Pearson Education, Ltd. Chapter 3, Slide 1

XLSTAT TIP SHEET FOR BUSINESS STATISTICS CENGAGE LEARNING

Evidence from Large Indemnity and Medical Triangles

Discussion of Using Tiers for Insurance Segmentation from Pricing, Underwriting and Product Management Perspectives

Quantile Regression. By Luyang Fu, Ph. D., FCAS, State Auto Insurance Company Cheng-sheng Peter Wu, FCAS, ASA, MAAA, Deloitte Consulting

A Stochastic Reserving Today (Beyond Bootstrap)

Data Analysis and Statistical Methods Statistics 651

Clark. Outside of a few technical sections, this is a very process-oriented paper. Practice problems are key!

Commercial Line Price Monitoring

Content Added to the Updated IAA Education Syllabus

Using Monte Carlo Analysis in Ecological Risk Assessments

Contents. An Overview of Statistical Applications CHAPTER 1. Contents (ix) Preface... (vii)

the display, exploration and transformation of the data are demonstrated and biases typically encountered are highlighted.

Subject CS2A Risk Modelling and Survival Analysis Core Principles

CAS antitrust notice CAS RPM Seminar Excess Loss Modeling. Page 1

Chapter ! Bell Shaped

Stochastic Analysis Of Long Term Multiple-Decrement Contracts

Contents Utility theory and insurance The individual risk model Collective risk models

ADVENTURES IN RATE CAPPING ACTUARIAL AND BUSINESS CONSIDERATIONS. Antitrust Notice

Five Things You Should Know About Quantile Regression

Some Characteristics of Data

On the Use of Stock Index Returns from Economic Scenario Generators in ERM Modeling

Exploring Data and Graphics

Where s the Beef Does the Mack Method produce an undernourished range of possible outcomes?

[D7] PROBABILITY DISTRIBUTION OF OUTSTANDING LIABILITY FROM INDIVIDUAL PAYMENTS DATA Contributed by T S Wright

Antitrust Notice 31/05/2016. Evaluating a Commercial Umbrella Rating Plan Using ISO. Table of Contents / Agenda

Workers compensation: what about frequency?

Basic Procedure for Histograms

Perspectives on European vs. US Casualty Costing

Introduction to Algorithmic Trading Strategies Lecture 8

The Leveled Chain Ladder Model. for Stochastic Loss Reserving

The histogram should resemble the uniform density, the mean should be close to 0.5, and the standard deviation should be close to 1/ 12 =

Cambridge University Press Risk Modelling in General Insurance: From Principles to Practice Roger J. Gray and Susan M.

STATISTICAL DISTRIBUTIONS AND THE CALCULATOR

Antitrust Notice The Casualty Actuarial Society is committed to adhering strictly to the letter and spirit i of the antitrust laws. Seminars conducted

Changes to Exams FM/2, M and C/4 for the May 2007 Administration

Week 7 Quantitative Analysis of Financial Markets Simulation Methods

Statistical Case Estimation Modelling

CAT Pricing: Making Sense of the Alternatives Ira Robbin. CAS RPM March page 1. CAS Antitrust Notice. Disclaimers

MAS1403. Quantitative Methods for Business Management. Semester 1, Module leader: Dr. David Walshaw

Homework: Due Wed, Nov 3 rd Chapter 8, # 48a, 55c and 56 (count as 1), 67a

Much of what appears here comes from ideas presented in the book:

starting on 5/1/1953 up until 2/1/2017.

Estimation of a Ramsay-Curve IRT Model using the Metropolis-Hastings Robbins-Monro Algorithm

Monte Carlo Simulation (General Simulation Models)

3/6/2017. Private Passenger Auto Plans RPM Seminar March 28 29, 2017 San Diego, CA. Residual Markets: Last Resort Coverage.

1. You are given the following information about a stationary AR(2) model:

Description of Data I

Introduction Dickey-Fuller Test Option Pricing Bootstrapping. Simulation Methods. Chapter 13 of Chris Brook s Book.

Summary of Information from Recapitulation Report Submittals (DR-489 series, DR-493, Central Assessment, Agricultural Schedule):

AP STATISTICS FALL SEMESTSER FINAL EXAM STUDY GUIDE

Key Features Asset allocation, cash flow analysis, object-oriented portfolio optimization, and risk analysis

Math 2311 Bekki George Office Hours: MW 11am to 12:45pm in 639 PGH Online Thursdays 4-5:30pm And by appointment

KARACHI UNIVERSITY BUSINESS SCHOOL UNIVERSITY OF KARACHI BS (BBA) VI

Pricing Analytics for the Small and Medium Sized Company

Statistics 431 Spring 2007 P. Shaman. Preliminaries

TEACHERS RETIREMENT BOARD INVESTMENT COMMITTEE

Presented at the 2012 SCEA/ISPA Joint Annual Conference and Training Workshop -

Homework: Due Wed, Feb 20 th. Chapter 8, # 60a + 62a (count together as 1), 74, 82

Session 5. A brief introduction to Predictive Modeling

Measuring Policyholder Behavior in Variable Annuity Contracts

Week 1 Variables: Exploration, Familiarisation and Description. Descriptive Statistics.

2011 Pearson Education, Inc

Interpolation Along a Curve

Chapter 7 1. Random Variables

Truth About Exposure Curves

Casualty Loss Reserve Seminar. Trends in Professional Liability. Gregory Larcher, FCAS, MAAA Aon Risk Solutions Global Risk Consulting

Probability and Statistics

value BE.104 Spring Biostatistics: Distribution and the Mean J. L. Sherley

Transcription:

Antitrust Notice The Casualty Actuarial Society is committed to adhering strictly to the letter and spirit of the antitrust laws. Seminars conducted under the auspices of the CAS are designed solely to provide a forum for the expression of various points of view on topics described in the programs or agendas for such meetings. Under no circumstances shall CAS seminars be used as a means for competing companies or firms to reach any understanding expressed or implied that restricts competition or in any way impairs the ability of members to exercise independent business judgment regarding matters affecting competition. It is the responsibility of all seminar participants to be aware of antitrust regulations, to prevent any written or verbal discussions that appear to violate these laws, and to adhere in every respect to the CAS antitrust compliance policy.

WC-5 Just How Credible Is That Employer? Exploring GLMs and Multilevel Modeling for NCCI s Excess Loss Factor Methodology CAS RPM Seminar Philadelphia, PA March 21, 2012 Presented by Chris Laws 2012 NCCI Holdings, Inc. Overview ELF Primer Motivation Alternatives to Current Approach Preliminary Approaches Results Next Steps Conclusion 2

Introduction NCCI is updating the methodology it uses to calculate Excess Loss Factors (ELFs) NCCI produces ELFs by state and hazard group ELFs are separated into the same two major components Excess Ratio Curves Severities and Loss Weights This presentation focuses on the improvements to the methodology used to arrive at the Severities and Loss Weights 3 What Are ELFs? An Excess Loss Factor (ELF) is the ratio of the expected portion of losses greater than a particular loss limit to standard premium For example, given a loss limit of $200,000 and an associated ELF of 10%, the expected losses over the deductible or retention of $200,000 per occurrence is equal to 10% of standard premium An ELF is the product of the Excess Ratio at a particular loss limit and the ratio of expected ground up losses to standard premium Let be the Excess Ratio for the loss random variable with density function at loss limit is defined as the ratio of expected losses in excess of to expected ground up losses 4

ELF Primer The heart of NCCI s ELF calculation is the Excess Ratio Curve Underlying curves are only updated once every 5 to 10 years However ELFs are generally updated annually There are two design features in NCCI s ELF methodology which allow our ELFs to be responsive on an annual timescale while holding the underlying curves constant 1. The curves are normalized to the average cost per case and are thus unitless 2. Different curves are created for each of the following injury types: Medical Only, Temporary Total, Permanent Partial, Permanent Total, and Fatal 5 ELF Primer Entry Ratios An entry ratio is defined as the ratio of a particular loss amount to the mean If the mean loss is $250,000 an entry ratio of 2.0 would correspond to a loss of $500,000 NCCI calculates and stores the excess ratio curves underlying the ELF calculation in terms of entry ratios The implicit assumption is that losses of all sizes (within a category described by a single underlying curve) share a common severity trend When calculating excess ratios corresponding to the loss amounts needed for ELFs The dollar amounts are normalized by the average cost per case (i.e. severity) to produce entry ratios The entry ratio is then used to find the excess ratio corresponding to the dollar amount The result is ELFs are responsive to annual severity trends Underlying curves are comparable between states Annual updates of ELFs require a sound severity estimate for each underlying state, hazard group, and injury type combination 6

ELF Primer Curves by Injury Type Final ELFs are intended to represent the loss experience for the entire state, hazard group combination Hazard groups are industry classifications which range from A (the least hazardous) to G (the most hazardous) NCCI calculates excess ratios for injury types and averages these excess ratios together using loss weights Injury Types are assumed to Represent homogeneous losses Separate heterogeneous losses Medical Only claims have A low average severity Permanent Total claims have A thick tail A high average severity Some changes in shape at the state hazard group level can be captured By changes in loss weights at the injury type level By relative changes in severity at the injury type level Annual updates of ELFs require sound loss weight estimates 7 ELF Primer From Injury Types to Claim Groups NCCI is switching from injury types to claim groups as shown below Current Grouping Fatal Permanent Total Permanent Partial Temporary Total Medical Only Proposed Grouping Fatal Permanent Total Likely to Develop * Permanent Partial & Temporary Total Not Likely to Develop * Permanent Partial & Temporary Total Medical Only *Claim groupings are differentiated based upon combinations of the injury type, claim status (open or closed) and the injured part of body. The various combinations are mapped to determine Likely to Develop or Not Likely to Develop claims. 8

Motivation The most important ingredient in the annual ELF update is the severities and loss weights for each combination of state, hazard group, and claim group Such partitioning can result in extremely small sample sizes Over 20 percent of NCCI states have zero Permanent Total claims for hazard group A (the least hazardous group) over a 5 year period Empirical statistics derived from such small samples generally have little resemblance to the true underlying data generating process The heavy tail of the loss distribution for some claim groups only exacerbates the problem The smallest sample sizes are seen in the claim groups with thicker tails and thus a disproportionate impact on ELFs As such, when deriving loss weights and severities, one needs a method to introduce a measure of stability balanced with responsiveness to the data 9 Current Approach Tempering for Large Fluctuations The current approach uses tempering to stabilize the effect of large fluctuations in empirical severities and loss weights Methods for tempering include Removing development from large losses when calculating severities and loss weights (manually done as needed) Taking weighted averages of indicated severities and/or loss weights with prior values (manually done as needed) Averaging calculated excess ratio with prior trended excess ratios (done automatically as part of the ELF calculation) If we can reduce the amount of tempering required We can streamline the production process Produce more objective ELFs 10

Alternative Generalized Linear Model GLMs are one approach currently in use by the insurance industry to address problems similar to the one at hand GLMs extend least squares regression by allowing for The assumption that observations follow a non-normal distribution The assumption of a multiplicative (as opposed to additive) relation A large set of ready made tools exist for GLMs Diagnostic and goodness-of-fit tests Model fitting software and algorithms GLMs have interpretable parameters Can be used to gain insight Can be used to describe the approach to less technical audiences GLMs also allow for other than multiplicative relations, but those are not of interest for this particular application 11 Alternative Multilevel Models\Random Effects GLMs do not automatically reduce the uncertainty surrounding parameters estimated in small samples Multilevel models are models in which the parameters are themselves modeled; such models can serve to mitigate the problem of parameter uncertainty Multilevel models are similar in concept to Bühlmann s credibility Both concepts rest on the ability to discern the variance within a group from the variance between groups in order to determine the appropriate level of credibility individual groups should receive All else being equal Larger groups receive more credibility Low between group variation points toward decreased individual group credibility While actuaries generally speak of the concept of credibility, multilevel modelers generally speak of the concept of shrinkage 12

Alternative Multilevel Models\Random Effects and Bühlmann s Credibility Following Gelman and Hill (2007), let be a normally distributed variable: ~, (where indicates a category (state, etc.) and indicates the observation) A multilevel model would assume that the parameter that governs the process in category is a draw from a distribution common for all levels of this category: ~, 1,, It can be shown that the multilevel estimator for reads: 1, 1 Gelman, Andrew, and Jennifer Hill, Data Analysis Using Regression and Multilevel/Hierarchical Models, Cambridge (MA): Cambridge University Press, 2007 13 Alternative Multilevel Models\Random Effects: Radon Example No pooling Multilevel model avg. log radon in county j 0 1 2 3 avg. log radon in county j 0 1 2 3 1 3 10 30 100 sample size in county j 1 3 10 30 100 sample size in county j The chart on the following slide displays a canonical example of multilevel modeling taken from Gelman and Hill (2007). The aim is to estimate radon level by county from several samples within each county. Many samples are taken in some counties. Few samples are taken in others. The chart on the left displays the sample mean. The chart on the right illustrates multilevel modeling. In both charts: The x-axis shows (on the log scale) the (jittered) number of observations in each county. The y-axis measures the estimated county radon level. Each bands represent a one standard deviation interval from the mean. The highlighted county has the highest sample mean. 14

Preliminary Severity Approach The Model A multilevel generalized linear model is used to model loss severity The model is linear on the log-scale and all covariates are categorical in nature Thus, one can think of the model in the terms of rating factors, where the model attempts to estimate a base rate and multiplicative factors for the categories of interest The model estimates such rating factors for state, hazard group, and claim group The model allows for state specific claim group factors (interaction between state and claim group) and takes into account the correlation between these factors Accounts for differences in state benefits by claim group 15 Preliminary Severity Approach The Model Severities are assumed to follow a Gamma distribution This simplifies model specification since the arithmetic mean of independent Gamma distributed losses is also Gamma distributed Given the state, hazard group, and claim group of a claim, individual losses Are assumed to be independent and to follow a Gamma distribution, thus Their empirical arithmetic average will follow Gamma distribution with the same underlying mean and a variance scaled by A decrease in variance corresponds to an increase in credibility 16

Possible Approaches For Loss Weights Using a GLM to estimate loss weights is not as straight forward as it is for estimating severities Three possible ways to estimate loss weights using GLMs are Option 1 Model loss weights directly Option 2 Model total losses and compute the necessary loss weights from the indicated total losses Option 3 Model claim counts and compute the necessary loss weights from the product of the indicated claim counts and indicated severities from a separate model 17 Possible Approaches For Loss Weights Pros and Cons Options Pros Cons 1 Models values of interest directly Accounts for correlation between claim counts and severities 2 Parameters have a more intuitive interpretation Distributions are commonly used Accounts for correlation between claim counts and severities 3 Parameters have a more intuitive interpretation Distributions are commonly used Handles observed $0 total losses Difficult to model due to uncommon distributions and support space Difficult to interpret parameters Cannot handle observed $0 losses without sophisticated techniques Does not account for correlation between claim counts and severities such correlation should be mild as severities and claim counts refer to the aggregation of many risks Option 3 was selected as the best option to pursue. 18

Preliminary Loss Weight Approach The Model Claim counts are assumed to follow a Negative Binomial distribution Assume represents total claim counts for claim group, state, hazard group, and report and, then The Negative Binomial is parameterized such that + is a parameter estimated from the observed data which varies by claim group For, the model reduces to a Poisson distribution The expected number of claims for each claim group, state, hazard group, and report combination is estimated as represents the (unadjusted) payroll and serves as a proxy for exposure represents an error term for each claim group, state, and hazard group combination Credibility is introduced on the estimated error terms by assuming that ~ 0,,4, where represents the distribution and is estimated from the data All else begin equal, the larger the estimated relative variation for observed claim counts within a claim group, state, and hazard group, the closer will be to zero,,, and are parameters to be estimated 19 Preliminary Loss Weight Approach The Negative Binomial Distribution Claim Count Coefficient of Variation 0.01 0.02 0.05 0.10 0.20 0.50 1.00 5 50 500 5000 1 10 100 1000 10000 Expected Claim Counts The above chart displays the resulting relation between the expected claim counts and the implied standard deviation for select values of. As approaches the Negative Binomial converges to a Poisson distribution. 20

Summary of Preliminary Approaches The severity model requires as input Observed severities (medical plus indemnity) by state, hazard group, and claim group Developed Trended On-leveled Observed claim counts by state, hazard group, and claim group Developed The claim count model requires as input Observed claim counts by state, hazard group, claim group, and report Developed Observed payroll by state, hazard group, and report Simple trending is currently handled by the model The estimated claim counts will then be combined with estimated severities from the severity model to produce the required loss weights 21 Implementation R is used in the pre and post-estimation process The model is estimated in JAGS R, http://www.r-project.org/ Open source software environment for statistical computing and graphics Implementation of the S language, which was developed at Bell Laboratories JAGS Just Another Gibbs Sampler, http://sourceforge.net/projects/mcmc-jags/files/ Open source program for the statistical analysis of Bayesian hierarchical models by Markov Chain Monte Carlo simulation Called from R using the package rjags, http://cran.rproject.org/web/packages/rjags/index.html 22

Evaluations Performed Both models were evaluated for Model Fit: The closeness of the indicated values to observed values Sensitivity: The impact of random fluctuations on indicated values Residual Plots were examined and Goodness of Fit Test were preformed to evaluate the Model Fit To evaluate the sensitivity of the severity model, a bootstrap analysis was performed To evaluate the sensitivity of the total claim count model, a remove-one report analysis was performed The implemented sensitivity evaluations also guard against overfitting If the model over-fits, the indicated values will follow the random functions 23 Evaluation of the Severity Model 24

Severity Model Fit Standardized Residual Charts: By Claim Group Standarized Residual -4-2 0 2 4 Not Likely Likely Fatal Permanent Total Claim Grouping Each point represents an observed state, hazard group, and claim group combination. 25 Severity Model Fit Standardized Residual Charts: By Standarized Residual -4-2 0 2 4 Each point represents an observed state, hazard group, and claim group combination. 26

Severity Model Fit Standardized Residual Charts: By Standarized Residual -4-2 0 2 4 Each point represents an observed state, hazard group, and claim group combination. 27 Bootstrapping Simple bootstrapped samples are generated by resampling, with replacement, from the observed dataset The theoretical motivation is to generate data from a process similar to the true underlying process with the assumption that the empirical distribution is such a process These samples can then be used to evaluate, among other things, the volatility of an estimator For example, suppose one observes loss of: $15k, $12k, $2k, $10k, $7k and $5k with a mean of $8.5k One randomly generated bootstrapped sample might be: $15k, $2k, $2k, $5k, $7k and $5k with a mean of $6k Another sample might be: $15k, $15k, $2k, $10k, $7k and $7k with a mean of $9.3k Claim characteristics (such as state and hazard group) are maintained throughout the sampling process Categories with more claims in the empirical sample are likely to have more claims in any given bootstrapped sample 28

Bootstrapped Results Large Group: Not Likely Average Claim Count 2368.6 6815 11473.3 4351.8 7124.4 4741.5 1326.8 Group: Likely Average Claim Count 584.2 1931.5 3613.5 1105.2 1915.7 1297.3 381.1 10 15 20 25 40 60 80 100 Group: Fatal Average Claim Count 3.1 12.8 36.5 19.6 32 39.6 24.9 Group: Permanent Total Average Claim Count 2.8 19.5 56.2 15.9 29.3 17.1 9.8 0 400 800 1200 0 500 1000 1500 Range of Bootstrapped Samples Range of Model Output Over Bootstrapped Samples Black polygons represent the range of model estimated severities over 100 bootstrapped samples Gray polygons represent the range of empirical severities over 100 bootstrapped samples The top and bottom of each polygon represents the max and min The widest point on the polygon represents the median The top and bottom notches represent the 75 th and 25 th percentiles 29 Bootstrapped Results Small Group: Not Likely Average Claim Count 2659.8 4619.2 4967.1 1977.7 3763.8 1496.3 276.1 Group: Likely Average Claim Count 388.4 1001.3 1207.9 373.4 847.9 345.9 78.9 10 15 20 25 40 60 80 Group: Fatal Average Claim Count 1 1.5 3.9 1 10.9 7.6 3 Group: Permanent Total Average Claim Count 1.9 1.1 0 0 4.8 2.9 1.1 0 200 400 600 0 500 1000 1500 Range of Bootstrapped Samples Range of Model Output Over Bootstrapped Samples Black polygons represent the range of model estimated severities over 100 bootstrapped samples Gray polygons represent the range of empirical severities over 100 bootstrapped samples The top and bottom of each polygon represents the max and min The widest point on the polygon represents the median The top and bottom notches represent the 75 th and 25 th percentiles 30

Bootstrapped Results Across s F Group: Not Likely : F Group: Likely 20 30 40 50 50 100 150 200 Group: Fatal Group: Permanent Total 0 500 1000 1500 0 2000 4000 6000 Range of Bootstrapped Samples Range of Model Output Over Bootstrapped Samples Black polygons represent the range of model estimated severities over 100 bootstrapped samples Gray polygons represent the range of empirical severities over 100 bootstrapped samples The top and bottom of each polygon represents the max and min The widest point on the polygon represents the median The top and bottom notches represent the 75 th and 25 th percentiles 31 Evaluation of the Claim Count Model 32

Claim Count Model Fit Standardized Residual Charts: By Standarized Residual -3-2 -1 0 1 2 3 4 Each point represents an observed state, hazard group, claim group, and report combination. The red line indicates the median residual. 33 Claim Count Model Fit Standardized Residual Charts: By Standarized Residual -3-2 -1 0 1 2 3 4 Each point represents an observed state, hazard group, claim group, and report combination. The red line indicates the median residual. 34

Claim Count Model Fit Standardized Residual Charts: By Claim Group Standarized Residual -3-2 -1 0 1 2 3 4 Not Likely Likely Fatal Permanent Total Claim Grouping Each point represents an observed state, hazard group, claim group, and report combination. The red line indicates the median residual. 35 Claim Count Model Fit Standardized Residual Charts: By Report Standarized Residual -3-2 -1 0 1 2 3 4 1 2 3 4 5 Report Each point represents an observed state, hazard group, claim group, and report combination. The red line indicates the median residual. Reports represent policy periods developed to 5th report. 36

Remove-One Report To assess the influence of statistical noise in the annual update, the model is estimated for the 5 sets of 4 reports created by removing, in turn, each report from the 5 reports included in the full dataset The range of the 5 predicted values is then compared to the 5 observed values and to the range of the empirical mean calculated on the 5 sets of 4 For example, suppose that we have observed claim counts of 0, 1, 5, 7, and 10 The 5 sets of 4 would then be 0, 1, 5, and 7; with an arithmetic mean of 3.25 0, 1, 5, and 10; with an arithmetic mean of 4 0, 1, 7, and 10; with an arithmetic mean of 4.5 0, 5, 7, and 10; with an arithmetic mean of 5.5 1, 5, 7, and 10; with an arithmetic mean of 5.75 37 Remove-One Results Large Group: Not Likely Group: Likely Group: Fatal Group: Permanent Total 2000 8000 14000 1000 3000 0 20 40 60 0 20 40 60 1.00 1.04 1.08 Group: Not Likely 1.00 1.06 1.12 Group: Likely 1.00 1.15 1.30 Group: Fatal 1.0 1.2 1.4 Group: Permanent Total Range of Remove-One Fits Observed Counts (1 per Report) Range of Remove-One Empirical Means Gray X-s in the first row represent observed claim counts one for each report. The height of the black rectangles in the first row represent the range of remove-one predicted values. Gray circles in the bottom row represent the ratio of the maximum to the minimum remove-one empirical means. The center of the black boxes in the bottom row represent the ratio of the maximum to the minimum remove-one fitted means. 38

Remove-One Results Small Group: Not Likely Group: Likely Group: Fatal Group: Permanent Total 0 400 800 0 50 150 250 0 1 2 3 4 5 0 1 2 3 4 1.00 1.10 Group: Not Likely 1.00 1.15 1.30 Group: Likely 1.0 2.0 3.0 Group: Fatal 1.0 1.5 2.0 2.5 Group: Permanent Total Range of Remove-One Fits Observed Counts (1 per Report) Range of Remove-One Empirical Means Gray X-s in the first row represent observed claim counts one for each report. The height of the black rectangles in the first row represent the range of remove-one predicted values. Gray circles in the bottom row represent the ratio of the maximum to the minimum remove-one empirical means. The center of the black boxes in the bottom row represent the ratio of the maximum to the minimum remove-one fitted means. 39 Remove-One Results Across s F Group: Not Likely : F Group: Likely 0 2000 4000 6000 Group: Not Likely 1.00 1.04 1.08 0 500 1000 1500 Group: Likely 1.00 1.10 1.20 Range of Remove-One Fits Observed Counts (1 per Report) Range of Remove-One Empirical Means Gray X-s in the first row represent observed claim counts one for each report. The height of the black rectangles in the first row represent the range of remove-one predicted values. Gray circles in the bottom row represent the ratio of the maximum to the minimum remove-one empirical means. The center of the black boxes in the bottom row represent the ratio of the maximum to the minimum remove-one fitted means. 40

Remove-One Results Across s F Group: Fatal : F Group: Permanent Total 0 10 30 50 Group: Fatal 1.0 1.1 1.2 1.3 1.4 1.5 0 10 20 30 40 50 Group: Permanent Total 1.0 2.0 3.0 4.0 Range of Remove-One Fits Observed Counts (1 per Report) Range of Remove-One Empirical Means Gray X-s in the first row represent observed claim counts one for each report. The height of the black rectangles in the first row represent the range of remove-one predicted values. Gray circles in the bottom row represent the ratio of the maximum to the minimum remove-one empirical means. The center of the black boxes in the bottom row represent the ratio of the maximum to the minimum remove-one fitted means. 41 Next Steps Endogenous model improvements The claim count model uses reports and has an error term for each state, hazard group, and claim group combination As such, it is more flexible than the severity model We are currently exploring incorporating such flexibility into the severity model We are seeking final structural form for both models Implementation Simple tempering of the data prior to model estimation, e.g. remove development from large claims Integration with production process Determining the appropriate spread of values across hazard groups 42

Conclusion This presentation introduces a new approach to calculating severities and loss weights by state, hazard group, and claim group for the ELF methodology The approach uses commonly employed techniques to introduce a measure of stability The proposed approach offers the opportunity for Increased automation A decreased need for manual tempering Allows for a more streamlined ELF calculation 43 Questions? 44