Actuarial Applications of A Hierarchical Insurance Claims Model
Actuarial Applications of A Hierarchical Insurance Claims Model
BY
ABSTRACT
KEYWORDS
Astin Bulletin 39(1), 165-197. doi: 10.2143/AST.39.1.2038061 © 2009 by Astin Bulletin. All rights reserved.
166 E.W. FREES, P. SHI AND E.A. VALDEZ
1. INTRODUCTION
Actuaries and other financial analysts that work with short term coverages
such as automobile insurance and healthcare expenditures typically have
massive amounts of in-company data. With modern computing equipment,
analysts can readily access data at the individual policyholder level that we
term “micro-level”. Actuaries use statistical models to summarize micro-level
data that subsequently need to be interpreted properly for financial decision-
making. For example, automobile insurers typically differentiate premium rates
based on policyholder characteristics such as age, gender and driving history.
Gourieroux and Jasiak (2007) have dubbed this emerging field the “micro-
econometrics of individual risk”.
Developing a statistical model to substantiate rate differentials may not be
straightforward because policyholder characteristics can affect the frequency
(number of accidents) and the severity (amount or intensity of the accident)
in different ways. Moreover, policies differ based on the type of coverage offered
as well as payment modifications (such as deductibles, upper limits and coin-
surance). Actuaries are also interested in other financial measures in addition
to those used for pricing. For example, actuaries use measures that summarize
portfolio risks for capital allocation and solvency purposes. As another exam-
ple, actuaries involved in managing risk through reinsurance use statistical
models to calibrate reinsurance treaties.
This paper demonstrates analyses that can be used for pricing, economic cap-
ital allocation, solvency and reinsurance based on a statistical model of a specific
database. The structure of the database encompasses policyholder and claim
files and is widely used. To illustrate, all property and casualty (general) insurers
domiciled in Singapore use this structure to report their data to a quasi-govern-
mental organization, the General Insurance Association (G.I.A.) of Singapore.
Our data are from a Singaporean insurance company. From the policyholder
file, we have available several policyholder characteristics that can be used to
anticipate automobile insurance claims. Additional data characteristics are in
Section 2.2. For each policy i, we are interested in predicting:
• Ni – the number of losses and
• yijk – the loss amount, available for each loss, j = 1, …, Ni , and the type of
loss k = 1, 2, 3.
• risk rating factors to be used as explanatory variables that predict both the
frequency and multivariate severity,
• the long-tail nature of the distribution of insurance claims through the GB2
(generalized beta of the second kind) distribution and
• the “two-part” distribution of losses. When a claim occurs, we do not
necessarily realize all three types of losses. Each type of loss may equal zero
or may be continuous and hence be comprised of “two parts”. Further, we
allow for
• losses to be related through a t-copula specification.
Thus, the tools that we propose can be used to determine economic capital for
insurers as well as for solvency testing. For example, we are able to quantify
the effect of imposing an upper limit on injury protection on the VaR for the
portfolio. Although we allow for long-tail severity for individual policies, we
can give practical guidance as to when approximate normality (via a central
limit theorem) holds for a portfolio of policies.
Section 5 retains the focus on portfolio of policies but now with an eye
towards quantifying the impact of reinsurance treaties. Essentially, we are
interested in “macro-level” changes from our micro-level (individual policy)
model. We discuss proportional reinsurance as well as non-proportional rein-
surance, and examine the effect of reinsurance agreements on the losses dis-
tribution. An example of portfolio reinsurance is given, and the combined
effects of reinsurance agreements are investigated by showing the tail summary
measures of losses for both insurer and reinsurer.
Section 6 provides a summary and closing remarks.
was interested in two lines of business, claims at fault and not at fault with
respect to a third party. For each line, Pinquet hypothesized a frequency and
severity component that were allowed to be correlated to one another. In par-
ticular, the claims frequency distribution was assumed to be bivariate Poisson.
Severities were modeled using lognormal and gamma distributions. Also at the
individual policyholder level, Frangos and Vrontos (2001) examined a claim fre-
quency and severity model, using negative binomial and Pareto distributions,
respectively. They used their statistical model to develop experience rated
(bonus-malus) premiums.
2.2. Data
Our statistical model of insurance claims is based on detailed, micro-level
automobile insurance records. Specifically, we analyzed information from two
databases: the policy and claims files. The policy file consists of all policy-
holders with vehicle insurance coverage purchased from a general insurer during
the observation period. Each vehicle is identified with a unique code. This file
provides characteristics of the policyholder and the vehicle insured, such as age
and gender, and type and age of vehicle insured. The claims file provides a
record of each accident claim that has been filed with the insurer during the
observation period and is linked to the policyholder file. For this analysis, we
ignored claims where no payments are made. Unfortunately, there was no infor-
mation in the files as to whether the claim was open or settled.
Some insurers also use a payment file that consists of information on each
payment that has been made during the observation period and is linked to the
claims file. Although it is common to see that a claim will have multiple pay-
ments made, we do not use that information for this paper and consider the
aggregate of all payments that arise from each accident event. See Antonio et
al. (2006) for a recent description of the claims “run-off ” problem.
The policyholder file provides several characteristics to help explain and
predict automobile accident frequency, type and severity. These characteristics
include information about the vehicle, such as type and age, as well as person
level characteristics, such as age, gender and prior driving experience. Table 1
summarizes these characteristics. These characteristics are denoted by the
vector xit and will serve as explanatory variables in our analysis. Person level
characteristics are largely unavailable for commercial use vehicles, so the
explanatory variables of personal characteristics are only used for observations
having non-commercial purposes. We also have the exposure eit , measured in
(a fraction of) years, which provides the length of time throughout the calen-
dar year for which the vehicle had insurance coverage.
With the information in the claims file, we potentially observe a trivariate
claim amount, one claim for each type. For each accident, it is possible to have
more than a single type of claim incurred; for example, an automobile accident
can result in damages to a driver’s own property as well as damages to a third
party who might be involved in the accident.
170 E.W. FREES, P. SHI AND E.A. VALDEZ
TABLE 1
DESCRIPTION OF RISK RATING FACTORS
Covariates Description
The statistical model, from Frees and Valdez (2008), consists of three compo-
nents – each component uses explanatory variables to account for heterogeneity
among policyholders. The first component uses a negative binomial regression
model to predict accident probability. The second component uses a multinomial
logit model to predict type of losses, either third party injury, own damage, third
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 171
party property or some combination. The third component is for severity; here,
a GB2 distribution is used to fit the marginal distributions and a t-copula is
used to model the dependence of the three types of claims.
It is customary in the actuarial literature to condition on the frequency
component when analyzing the joint frequency and severity distributions. See,
for example, Klugman, Panjer and Willmot (2004). Frees and Valdez (2008)
incorporate an additional claims type layer to handle the many zeros in each
distribution (known as “two-part” data) as well as accounting for the possibility
of multivariate claims. Specifically, conditional on having observed at least one
type of claim, the random variable M describes which of the seven combinations
is observed. Table 2 provides potential values of M.
TABLE 2
VALUE OF M, BY CLAIM TYPE.
Value of M 1 2 3 4 5 6 7
Claim by Combination Observed (y1 ) ( y2 ) (y3 ) ( y1, y2 ) ( y1, y3 ) ( y2, y3 ) (y1, y2, y3 )
We are now in a position to describe the full predictive model. Suppressing the
{i} subscripts, the joint distribution of the dependent variables is:
f (N, M, y) = f (N ) ≈ f (M| N ) ≈ f (y | N, M)
joint = frequency ≈ conditional claim type ≈ conditional severity,
Pr(M = m) = exp(Vm ) / {!7s =1 exp(Vs )}, where Vm = Vit, m = xM, it b M, m . Note that
for our application, the covariates in xM,it do not depend on the accident num-
ber j nor on the claim type m although we allow parameters ( b M, m ) to depend
on m. This portion of the model was proposed by Terza and Wilson (1990).
exp ^a 1 zh
f ^ yh = , (1)
y s B ^a 1 , a 2h 61 + exp ]zg @ a1 + a 2
where z = (ln y – m) / s and B(a1, a2 ) = G(a1) G(a2) / G(a1 + a2 ), the usual beta
function. Here, m is a location parameter, s is a scale parameter and a1 and a2
are shape parameters. This distribution is well known in actuarial modeling of
univariate loss distributions (see for example, Klugman, Panjer and Willmot,
2004). With four parameters, the distribution has great flexibility for fitting heavy
tailed data. Many distributions useful for fitting long-tailed distributions can
be written as special or limiting cases of the GB2 distribution; see, for exam-
ple, McDonald and Xu (1995).
We use this distribution but allow scale and shape parameters to vary by
type and thus consider a1k , a2k and sk for k = 1, 2, 3. Despite the prominence
of the GB2 in fitting distributions to univariate data, there are relatively few
applications that use the GB2 in a regression context. Recently, Sun et al.
(2008) used the GB2 in a longitudinal data context to forecast nursing home
utilization.
To accommodate dependencies among claim types, we use a parametric
copula. See Frees and Valdez (1998) for an introduction to copulas. Suppressing
the {i} subscripts, we may write the joint distribution of claims ( y1, y2, y3 ) as
Here, the marginal distribution of yk is given by Fk (·) and H(·) is the copula
linking the marginals to the joint distribution. We use a trivariate t-copula with
an unstructured correlation matrix. See Frees and Valdez (2008) for a further
motivation of the use of the t-copula.
Given a set of risk rating factors such as in Table 1, one basic task is to arrive
at a fair price for a contract. In setting prices, often the actuary is called upon
to quantify the effects of certain policy modifications. As a basis of fair pricing
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 173
is the predictive mean, this section calculates predictive means for several alter-
native policy designs that may be of interest.
For alternative designs, we consider four random variables:
• individuals losses, yijk
• the sum of losses from a type, Si, k = yi,1, k + … + yi,Ni, k
• the sum of losses from a specific accident, SACC, i, j = yi, j,1 + yi, j, 2 + yi, j,3 , and
• an overall loss per policy, Si = Si,1 + Si,2 + Si,3 = SACC, i,1 + ... + SACC, i, Ni .
From the conditional severity model, we define mik = E( yijk | Ni , Ki = k). The
random variable Ki indicates the type, for Ki = 1, 2, 3. Then, basic probability
calculations show that:
case, from the severity distribution, we have an analytic expression for the con-
ditional mean of the form
B ^a 1 k + s k , a 2 k - s k h
mik = exp (xik bk) , (6)
B ^ a 1 k , a 1 kh
where bk, ajk, sk are parameters of the GB2 severity distribution (see Sec-
tion 2.3.3). With policy modifications, we approximate mik via simulation (see
Section A.3).
The predictive means in equations (2)-(5), by level of no claims discount (NCD)
and insured’s age, are shown in Tables 3 and 4, respectively. The calculation is
based on a randomly selected observation from the 2001 portfolio. The policy-
holder is a 50-year old female driver who owns a Toyota Corolla manufactured
in year 2000 with a 1332 cubic inch capacity. For losses based on a coverage
type, we chose own damage because the risk factors NCD and age turned out
to be statistically significant for this coverage type. (Appendix A.2 shows
that both NCD and age are statistically significant variables in at least one
component of the predictive model, either frequency, type or severity.) The
point of this section is to understand their economic significance.
Using equation (6), Table 3 shows that the insured who enjoys a higher no
claim discount has a lower expected loss; this result holds for all four random
TABLE 3
PREDICTIVE MEAN BY LEVEL OF NCD
Level of NCD
Type of Random Variable
0 10 20 30 40 50
Individual Loss (Own Damage) 330.67 305.07 267.86 263.44 247.15 221.76
Sum of Losses from a Type (Own Damage) 436.09 391.53 339.33 332.11 306.18 267.63
Sum of Losses from a Specific Event 495.63 457.25 413.68 406.85 381.70 342.48
Overall Loss per Policy 653.63 586.85 524.05 512.90 472.86 413.31
TABLE 4
PREDICTIVE MEAN BY INSURED’S AGE
Insured’s Age
Type of Random Variable
# 21 22-25 26-35 36-45 46-55 56-65 $ 66
Individual Loss (Own Damage) 258.41 238.03 198.87 182.04 221.76 236.23 238.33
Sum of Losses from a Type (Own Damage) 346.08 309.48 247.67 221.72 267.63 281.59 284.62
Sum of Losses from a Specific Event 479.46 441.66 375.35 343.59 342.48 350.20 353.31
Overall Loss per Policy 642.14 574.24 467.45 418.47 413.31 417.44 421.93
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 175
variables. This is consistent with our intuition and, as shown in the Appendix,
with the statistical model fitting results.
Table 4 presents a more complex nonlinear pattern for insured’s age. For
each random variable, predictive means are at their highest at the youngest
age group, decrease as age increases and remain relatively stable thereafter.
Figure 1 presents a graphical display in the lower left-hand corner.
When different coverage modifications are incorporated, we need to simulate
the amount of losses to calculate predictive means. Tables 5 and 6 show the
effects of different policy designs under various combinations of NCD and
insured’s age. The first row under each of the four random variables corresponds
to the predictive mean for the policy without any coverage modification;
here, readers will notice a slight difference between these entries and the cor-
responding entries in Tables 3 and 4. This is due to simulation error. We used
5,000 simulated values.
To understand the simulation error, Figure 1 compares the analytic and sim-
ulated predictive means. Here, we consider the case of no deductible, policy
limits and coinsurance, so that the analytic result in equation (6) is available.
The upper two panels shows the relationship between predictive mean and
NCD, whereas the lower two panels are for insured’s age. The two panels on
the left are for the analytic result, whereas the two panels on the right are for
the simulation results. For the simulated results, the lines provide the 95%
confidence intervals. The width of these lines show that the simulation error
is negligible for our purposes – for other purposes, one can always reduce the
simulation error by increasing the simulation size.
Table 5 shows the simulated predictive mean at different levels of NCD
under various coverage modifications. As expected, any of a greater deductible,
lower policy limit or smaller coinsurance results in a lower predictive mean.
Deductibles and policy limits change the predictive mean nonlinearly, whereas
coinsurance changes the predictive mean linearly. For example, the predictive
mean decreases less when the deductible increases from 250 to 500, compared
to the decrease when deductible increases from 0 to 250. This pattern applies
to all the four loss variables at all NCD levels. The effect of policy limit depends
on the expected loss. For random variables with a small expected loss (e.g.
individual loss and sum of losses from a type), there is little difference in pre-
dictive means between policies with a 50,000 limit and no limit. In contrast,
for random variables with large expected losses (e.g. sum of losses from a specific
event and overall losses), the difference in predictive means can be greater when
limit increases from 25,000 to 50,000 than an increase from 50,000 to no limit.
Table 6 shows the effect of coverage modifications on the predictive mean for
the insured at different age categories. As with Table 5, any of a higher deductible,
lower coverage limit or lower coinsurance percentage results in a lower pre-
dictive mean. The combined effect of three kinds of coverage modifications
can be derived from the three marginal effects. For example, when the insured’s
age is in the 26-35 category, the predictive mean of individual loss with
deductible 250, coverage limit 25,000 and coinsurance 0.75 can be calculated
176 E.W. FREES, P. SHI AND E.A. VALDEZ
700
700
600
600
Simulated Mean
Analytic Mean
500
500
400
400
300
300
200
200
0 10 20 30 40 50 0 10 20 30 40 50
NCD NCD
100 200 300 400 500 600 700
<21 22−25 26−35 36−45 46−55 56−65 >65 <21 22−25 26−35 36−45 46− 55 56−65 >65
from predictive mean of individual loss with deductible 250 and predictive
mean of individual loss with coverage limit 25,000, that is (170.54 + 189.64 –
191.13) * 0.75 = 126.79. Similar results can be derived for all the four random
variables under different NCD or insured’s age values.
Comparing Tables 5 and 6, the effect of NCD has a greater effect on the
predictive mean than that of insured’s age, in the sense that the range of predictive
means is greater under alternative NCD levels compared to age levels. For
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 177
TABLE 5
SIMULATED PREDICTIVE MEAN BY LEVEL OF NCD AND COVERAGE MODIFICATIONS
TABLE 6
SIMULATED PREDICTIVE MEAN BY INSURED’S AGE AND COVERAGE MODIFICATIONS
example, for a policy with a 500 deductible, 50,000 policy limit and 0.75 coin-
surance, the predictive mean of individual losses is 138.07 from Table 5 and
132.12 from Table 6 (the difference is due to simulation error). For this policy, the
predictive mean varies between 138.07 and 208.05 under various NCD levels, and
varies between 107.95 and 151.42 under alterative insured’s age categories.
Under other coverage modifications we observe similar results.
8e−06
Density
4e− 06
0e+00
0 250000 500000
Portfolio Losses
own damage
2e−05
Predicted Losses
FIGURE 3: Simulated Density of Losses for Third Party Injury, Own Damage and
Third Party Property of a Randomly Selected Portfolio.
the normal is a good approximation for the distribution of the sum based on 30
to 50 i.i.d. draws. The distribution of the sum shown in Figure 2 is not approx-
imately normal; this is because (1) the policies are not identical, (2) have discrete
and continuous components and (3) have long-tailed continuous components.
As described in Section 3, there may be situations in which the analyst is
interested in the claims from each type of coverage instead of the total claims
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 181
in a comprehensive policy. In this case, one should consider the random variables
Sk = !1000
i =1 Si, k , where k = 1, 2, 3 corresponds to third party injury, own damage
and third party property claims.
As shown in Figure 3, the distributions for the three types of losses are quite
different in terms of their skewness and kurtosis as well as other properties.
The density for third party injury and own damage have higher peaks and are
more positively skewed than third party property. The density of third party
injury has a heavier tail. This type of analysis can provide useful information
about the risk profile of different coverages within a comprehensive policy, as
well as the risk profile of different lines of business. For example, the analyst
may consider which type of coverage should be incorporated and which type
should be eliminated, so that the product can be tuned to meet the company’s
risk management requirement, regulatory policy or other strategic goals.
TABLE 8
CTE BY PERCENTILE AND COVERAGE MODIFICATION WITH A CORRESPONDING STANDARD DEVIATION
on VaR. Fourth, the policy limit exerts a greater effect than a deductible on
the confidence interval capturing the VaR.
Table 8 shows CTEs by percentile and coverage modification with a corre-
sponding standard deviation. The results are consistent with Table 7. Either
a larger deductible or a smaller policy limit results in a lower CTE and the effect
is nonlinear. The range of policy limits explored has a greater influence on
CTE than the range of deductibles. The sparsity of data combined with the
heavy tailed nature of the distribution result in the greater standard deviations
at higher percentiles. Finally, the decrease in the policy limits reduces the stan-
dard deviation of the CTE substantially whereas changes in the deductible
have little influence.
4.3. Dependence
One way to examine the role of dependence is to decompose the comprehensive
coverage into more “primitive” coverages for the three types of claims (third
party injury, own damage and third party property). As in derivative securities,
we call this “unbundling” of coverages. We are able to calculate risk measures
for each unbundled coverage, as if separate financial institutions owned each
coverage, and compare them to risk measures for the bundled coverage that the
insurance company is responsible for. The results are shown in Table 9. For the
bundled (comprehensive) coverage, the VaR and CTE are from the first row of
Tables 7 and 8, respectively, for policies without coverage modifications.
In general, risk measures such as VaR need not be subadditive and so there
is no guarantee that bundling diversifies risk. However, for our application,
our results show that the risk measures for bundled coverages are smaller than
the sum of unbundled coverages, for both risk measures and all percentiles.
One of the important purposes of risk measures is to determine economic cap-
ital which is the amount of capital that banks and insurance companies set
aside as a buffer against potential losses from extreme risk event. The impli-
cation of this example is that by bundling different types of coverage into one
comprehensive policy, the insurers can reduce the economic capital for risk
management or regulatory purpose. Another perspective is that this example
simply demonstrates the effectiveness of economies of scales; three small
financially independent institutions (one for each coverage) require in total
more capital than a single combined institution (one for the bundled coverage).
The interesting thing is that this is true even though the dependencies among
the three coverages are positive, as shown in Appendix A.2.
The statistical model described in Section 2.3 with parameter estimates in
Appendix A.2 show strong significance evidence of positive relations among
the three coverage types, third party injury, own damage and third party prop-
erty. However, the model is complex, using copulas to assess this nonlinear
dependence. How important is the dependence for the financial risk measures?
To quantify this issue, Table 10 shows the VaR and CTE for different copula
models. The independence copula comes from treating the three lines as unrelated,
184 E.W. FREES, P. SHI AND E.A. VALDEZ
TABLE 9
VAR AND CTE BY PERCENTILE FOR UNBUNDLED AND BUNDLED COVERAGES
VaR CTE
Unbundled Coverages
90% 95% 99% 90% 95% 99%
the normal copula captures the correlation among different coverages within
a comprehensive policy compared with independence copula, whereas the t-cop-
ula captures the heavy-tail features of the risk compared to the normal copula.
As a sensitivity analysis, we incorporated the copulas in two ways. In the top
portion of the table, we assumed that the specified copula was consistently
used for the estimation and prediction portions. For the bottom portion, we
assumed that the (correct but more complex) t-copula was used for estimation
with the specified copula used for prediction. The idea behind the bottom
portion was that a statistical analysis unit of a company may perform a more
rigorous analysis using a t-copula and another unit within a company may
wish to use this output for quicker calculations about their financial impact.
Table 10 shows that the copula effect is large and increases with the per-
centile. Of course, the upper percentiles are the most important to the actuary
for many financial implications.
Table 10 demonstrates a large difference between assuming independence
among coverages and using a t-copula to quantify the dependence. We found,
when re-estimating the full model under alternative copulas, that the marginal
parameters changed to produce significant differences in the risk measures.
Intuitively, one can think of estimation and prediction under the independence
copula to be similar to “unbundled” coverages in Table 9, where we imagine
separate financial institutions accepting responsibility for each coverage. In one
sense, the results for the independence copula are somewhat counterintuitive.
For most portfolios, with positive correlations among claims, one typically
needs to go out further in the tail to achieve a desired percentile, suggesting
that the VaR should be larger for the t-copula than the independence copula.
To reinforce these findings, the bottom half of the table reports results
when the marginal distributions are unchanged, yet the copula differs. Table 10
shows that the VaR is not affected by the choice of copula; the differences in
Table 10 are due to simulation error. In contrast, for the CTEs, the normal and
t-copula give higher values than the independence copula. This result is due
to the higher losses in the tail under the normal and t-copula models. Although
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 185
TABLE 10
VAR AND CTE FOR BUNDLED COVERAGE BY COPULA
VaR CTE
Copula
90% 95% 99% 90% 95% 99%
not displayed here, we re-ran this portion of the analysis with 50,000 simulation
replications (in lieu of 5,000) and verified these patterns.
When applied to economic capital, these results indicate that the independence
copula leads to more conservative risk measures while the t-copula leads to more
aggressive risk measures. To determine which model to be used to calculate the
economic capital may depend on the purpose of the capital, either for interval
estimation, risk management or regulatory use. It may also depend on the trade-
off among model simplicity, estimation accuracy and computational complexity.
Reinsurance, an important risk management tool for property and casualty insur-
ers, is another area of application where predictive distributions can be utilized.
We examine two types of reinsurance agreements: quota share reinsurance and
excess-of-loss reinsurance. In addition, we investigate a simple portfolio rein-
surance. We simulate the number of accident, type of losses and severity of
losses, and then allocate the losses between insurer and reinsurer according to
different reinsurance agreements.
Quota share reinsurance is a form of proportional reinsurance which
specifies that a fixed percentage of each policy written will be transferred to
the reinsurer. The effect of different quota shares on the retained claims for the
ceding company is examined and presented in Figure 4. The distributions of
retained claims are derived assuming the insurer retains 25%, 50%, 75% and
100% of its business. In Figure 4 a quota of 0.25 means the insurer retains 25%
of losses and cedes 75% to the reinsurer. The curve corresponding to quota
of 1 represent the losses of insurer without a reinsurance agreement, as in
Figure 2. As we can see, the quota share reinsurance does not change the shape
186 E.W. FREES, P. SHI AND E.A. VALDEZ
of the retained losses, only the location and scale. For example, if the insurer
ceded 75% of losses, the retained losses will shift left and the variance of
retained losses will be 1/16 times the variance of original losses. We did not
present the distribution of losses for reinsurer, because under quota share rein-
surance, the insurer and reinsurer share the losses proportionally.
Excess-of-loss is a nonproportional reinsurance under which the reinsurer
will pay the ceding company for all the losses above a specified dollar amount,
the retention limit. The retention limit is similar to the deductible in a primary
policy, the reinsurer will assume all the losses above it. Figure 5 shows the
effect of different retention limits on the losses of insurer and reinsurer. Losses
are simulated and limits of 5,000, 10,000 and 20,000 per policy are imposed.
Unlike quota share arrangements, the retention limit changes the shape of the
distribution for both the insurer and reinsurer. The lower the retention limit,
the more business the insurer cedes so that losses for insurer become less skewed
with thinner tails because the losses in the tail of distribution become the
responsibility of reinsurer. Correspondingly, as the retention limit decreases,
the distribution of losses for the reinsurer exhibits fatter tails. This is because
the reinsurer retains a larger portion of the claim.
Figures 4 and 5 provide insights on how the various types of reinsurance
agreement will affect their risk profile of the insurer and reinsurer. Through
such analyses, the insurer can choose proper forms of reinsurance to manage
its risk portfolio, and the reinsurer can decide upon the amount of risk to be
underwritten. In practice, there are many other types of reinsurance contracts.
The above analysis focused on the reinsurance agreements where reim-
bursements are based on losses for each policy. As another example, we
4e−05
3e−05
Quota=0.25
Density
2e−05
Quota=0.5
Quota=0.75
1e−05
Quota=1
0e+00
0 250000 500000
Retained Claims
FIGURE 4: Distribution of Retained Claims for the Insurer under Quota Share Reinsurance.
The insurer retains 25%, 50%, 75% and 100% of losses, respectively.
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 187
4e−05
Density
2e−05 Retention=5,000
Retention=10,000
Retention=20,000
0e+00
0 250000 500000
Retained Claims
4e−05
Retention=20,000
Retention=10,000
Density
2e−05
Retention=5,000
0e+00
0 250000 500000
Ceded Claims
FIGURE 5: Distribution of Losses for the Insurer and Reinsurer under Excess-of-Loss Reinsurance.
The losses are simulated under different primary company retention limits.
The upper panel is for the insurer and lower panel is for the reinsurer.
consider a case where both policy reinsurance and portfolio reinsurance are
involved. Portfolio reinsurance generally refers to the situation where the rein-
surer is insuring all risks in a portfolio of policies, such as a particular line of
business. For the purpose of illustration, we still consider the randomly selected
portfolio which consists of 1000 policies from held-out-sample in year 2001.
TABLE 11
188
Quota Policy Retention Portfolio Retention 1% 5% 10% 25% 50% 75% 90% 95% 99%
0.25 none 100,000 22,518 26,598 29,093 34,196 40,943 50,657 64,819 83,500 100,000
0.5 none 100,000 45,036 53,197 58,187 68,393 81,885 100,000 100,000 100,000 100,000
0.75 none 100,000 67,553 79,795 87,280 100,000 100,000 100,000 100,000 100,000 100,000
1 10,000 100,000 86,083 99,747 100,000 100,000 100,000 100,000 100,000 100,000 100,000
1 10,000 200,000 86,083 99,747 108,345 122,927 140,910 159,449 177,013 188,813 200,000
1 20,000 200,000 89,605 105,578 114,512 132,145 154,858 177,985 200,000 200,000 200,000
0.25 10,000 100,000 21,521 24,937 27,086 30,732 35,228 39,862 44,253 47,203 53,352
0.5 20,000 100,000 44,803 52,789 57,256 66,072 77,429 88,993 100,000 100,000 100,000
0.75 10,000 200,000 64,562 74,810 81,259 92,195 105,683 119,586 132,760 141,610 160,056
1 20,000 200,000 89,605 105,578 114,512 132,145 154,858 177,985 200,000 200,000 200,000
Quota Policy Retention Portfolio Retention 1% 5% 10% 25% 50% 75% 90% 95% 99%
0.25 none 100,000 67,553 79,795 87,280 102,589 122,828 151,972 194,458 250,499 486,743
E.W. FREES, P. SHI AND E.A. VALDEZ
0.5 none 100,000 45,036 53,197 58,187 68,393 81,885 102,630 159,277 233,998 486,743
0.75 none 100,000 22,518 26,598 29,093 36,785 63,771 102,630 159,277 233,998 486,743
1 10,000 100,000 0 8,066 16,747 36,888 63,781 102,630 159,277 233,998 486,743
1 10,000 200,000 0 0 992 5,878 18,060 43,434 97,587 171,377 426,367
1 20,000 200,000 0 0 0 0 2,482 24,199 78,839 151,321 412,817
0.25 10,000 100,000 68,075 80,695 88,555 104,557 127,652 161,743 215,407 292,216 541,818
0.5 20,000 100,000 45,132 53,298 58,383 68,909 84,474 111,269 167,106 245,101 491,501
0.75 10,000 200,000 23,536 28,055 31,434 39,746 54,268 81,443 135,853 209,406 462,321
1 20,000 200,000 0 0 0 0 2,482 24,199 78,839 151,321 412,817
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 189
We now assume that there is a limit for the entire portfolio of business, the
reinsurer will assume all the losses exceeding the portfolio limit. In addition,
both quota share and policy limit have been applied to the individual policy
in the portfolio. The combined effect of these coverage modifications on the
risk profile for insurer and reinsurer are investigated and the results are pre-
sented in Table 11.
Table 11 presents percentiles of losses for the insurer and reinsurer under
various reinsurance arrangements. As before, we see the long tail nature in the
total losses of the portfolio and the interactive effect of coverage modifications
on the claim distribution. In the first three rows where there is no policy reten-
tion limit, the insurer and reinsurer share the losses proportionally before the
total losses reach the portfolio limit; that is why the 25th percentile for the
insurer is equal to the 75th percentile for the reinsurer. When the total losses
reach portfolio limit, the amount above it should be ceded to reinsurer. The
second three rows for insurer and reinsurer shows the interactive effect of policy
and portfolio retention limits. Both limits reduce the heavy tailed nature of
losses for the insurer. Under a policy retention limit of 10,000, a greater port-
folio retention limit has less effect on changing the tail behavior of losses for
insurer. However the effect of portfolio retention limit (say 200,000) depends
on the policy limit in the sense that the effect will be bigger under a greater
policy limit (20,000). The last four rows of percentiles for both insurer and
reinsurer shows the combined effect of coinsurance, policy limit and portfolio
limit.
Table 11 provides useful information about the distribution of potential
losses for insurer and reinsurer. This can help the ceding company understand
the risk characteristics of retained business. The insurer can choose appropriate
reinsurance agreements including setting proper policy retention limit, select-
ing the right quota and deciding the aggregate stop loss, to manage the risk
more efficiently. The results also can be helpful in determine the reinsurance
premium and to help reinsurer to assess the risk of the business they assumed
from ceding company.
and conditional tail expectation, CTE, although our approach could be easily
extended to other risk measures. We assessed the effects of some of statistical
assumptions, such as the copula, on these measures.
The third type of application was to provide predictive distributions for
reinsurance. We examined the combined effect of coverage modifications for
policies and portfolios on the claim distributions of the insurer and reinsurer,
by examining the tail summary for the losses of portfolio generated using
Monte Carlo simulation.
This paper demonstrates some of the interesting financial analyses of concern
to actuaries that can be accomplished with detailed micro-level data and advanced
statistical models. With simulation, we can easily calculate predictive distrib-
utions for several financial risk measures. A limitation of this paper is that we
did not explicitly incorporate estimation error into our predictive distributions
(see, for example, Cairns (2000)). On one hand, one might argue that we had
many observations available for estimation and that the resulting standard errors
for the statistical model were inherently small. On the other hand, one could say
that that the statistical model incorporates many parameters whose joint uncertainty
should be accounted for. We view this as an interesting area for future research.
ACKNOWLEDGMENTS
The first author thanks the National Science Foundation (Grant Number SES-
0436274) and the Assurant Health Insurance Professorship for funding to sup-
port this research. We also thank Richard Derrig, seminar participants at the
Risk Theory Society, Université de Montreal and the Universitat de Barcelona
as well as anonymous reviewers for comments that helped improve the paper.
The third author wishes to acknowledge hospitality from hosts Louis Doray,
Montserratt Guillen, Carmen Ribas and Oriol Roch.
REFERENCES
ANGERS, J.-F., DESJARDINS, D., DIONNE, G. and GUERTIN, F. (2006) Vehicle and fleet random
effects in a model of insurance rating for fleets of vehicles. ASTIN Bulletin 36(1), 25-77.
ANTONIO, K., BEIRLANT, J., HOEDEMAKERS, T. and VERLAAK, R. (2006) Lognormal mixed models
for reported claims reserves. North American Actuarial Journal 10(1), 30-48.
BOUCHER, J.-Ph. and DENUIT, M. (2006) Fixed versus random effects in Poisson regression models
for claim counts: A case study with motor insurance. ASTIN Bulletin 36(1), 285-301.
BROCKMAN, M.J. and WRIGHT, T.S. (1992) Statistical motor rating: making effective use of your
data. Journal of the Institute of Actuaries 119, 457-543.
CAIRNS, A.J.G. (2000) A discussion of parameter and model uncertainty in insurance. Insurance:
Mathematics and Economics 27(3), 313-330.
COUTTS, S.M. (1984) Motor insurance rating, an actuarial approach. Journal of the Institute of
Actuaries 111, 87-148.
DESJARDINS, D., DIONNE, G. and PINQUET, J. (2001) Experience rating schemes for fleets of vehi-
cles. ASTIN Bulletin 31(1), 81-105.
FRANGOS, N.E. and VRONTOS, S.D. (2001) Design of optimal bonus-malus systems with a frequency
and a severity component on an individual basis in automobile insurance. ASTIN Bulletin 31(1),
1-22.
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 191
FREES, E.W. and VALDEZ, E.A. (1998) Understanding relationships using copulas. North American
Actuarial Journal 2(1), 1-25.
FREES, E.W. and VALDEZ, E.A. (2008) Hierarchical insurance claims modeling. Journal of the
American Statistical Association 103(484), 1457-1469.
GOURIEROUX, Ch. and JASIAK, J. (2007) The Econometrics of Individual Risk: Credit, Insurance,
and Marketing. Princeton University Press, Princeton, NJ.
HARDY, M. (2003) Investment Guarantees: Modeling and Risk Management for Equity-Linked
Life Insurance. John Wiley & Sons, New York.
HSIAO, C., KIM, C. and TAYLOR, G. (1990) A statistical perspective on insurance rate-making.
Journal of Econometrics 44, 5-24.
KAHANE, Y. and HAIM Levy, H. (1975) Regulation in the insurance industry: determination of
premiums in automobile insurance. Journal of Risk and Insurance 42, 117-132.
KLUGMAN, S., PANJER, H. and WILLMOT, G. (2004) Loss Models: From Data to Decisions (Second
Edition), Wiley, New York.
MCDONALD, J.B. and XU, Y.J. (1995) A generalization of the beta distribution with applications.
Journal of Econometrics 66, 133-152.
PINQUET, J. (1997) Allowance for cost of claims in bonus-malus systems. ASTIN Bulletin 27(1),
33-57.
PINQUET, J. (1998) Designing optimal bonus-malus systems from different types of claims. ASTIN
Bulletin 28(2), 205-229.
RENSHAW, A.E. (1994) Modeling the claims process in the presence of covariates. ASTIN Bulletin
24(2), 265-285.
SUN, J., FREES, E.W. and ROSENBERG, M.A. (2008) Heavy-tailed longitudinal data modeling
using copulas. Insurance: Mathematics and Economics, 42(2), 817-830.
TERZA, J.V. and WILSON, P.W. (1990) Analyzing frequencies of several types of events: A mixed
multinomial-Poisson approach. The Review of Economics and Statistics 108-115.
WEISBERG, H.I. and TOMBERLIN, T.J. (1982) A statistical perspective on actuarial methods for
estimating pure premiums from cross-classified data. Journal of Risk and Insurance 49,
539-563.
WEISBERG, H.I., TOMBERLIN, T.J. and CHATTERJEE, S. (1984) Predicting insurance losses under
cross-classification: A comparison of alternative approaches. Journal of Business & Economic
Statistics 2(2), 170-178.
EDWARD W. FREES
School of Business
University of Wisconsin
Madison, Wisconsin 53706 USA
E-Mail: [email protected]
PENG SHI
School of Business
University of Wisconsin
Madison, Wisconsin 53706 USA
E-Mail: [email protected]
EMILIANO A. VALDEZ
Department of Mathematics
College of Liberal Arts and Sciences
University of Connecticut
Storrs, Connecticut 06269-3009 USA
E-Mail: [email protected]
192 E.W. FREES, P. SHI AND E.A. VALDEZ
To provide readers with a feel for the data, Table A.1 describes the frequency
of claims, Tables A.2 and A.3 describe the claim frequency relationship with
covariates and Table A.4 displays the distribution by type of claim. Figure 6
gives a density of losses by type.
TABLE A.1
FREQUENCY OF CLAIMS
Count 0 1 2 3 4 Total
TABLE A.2
NUMBER AND PERCENTAGES OF CLAIMS, BY VEHICLE TYPE AND AGE
VEHICLE TYPE
Automobile 90.06 121,249 92.31
Other 87.75 10,097 7.69
VEHICLE AGE (IN YEARS)
0 93.26 7,330 5.58
1 to 2 89.11 25,621 19.51
3 to 5 89.76 48,964 37.28
6 to 10 89.87 48,226 36.72
11 to 15 92.02 1,103 0.84
16 and older 87.25 102 0.08
Total Number 131,346 100
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 193
TABLE A.3
NUMBER AND PERCENTAGES OF GENDER, AGE AND NCD FOR AUTOMOBILE POLICIES
GENDER
Female 90.77 23,783 19.62
Male 89.89 97,466 80.38
PERSON AGE (IN YEARS)
21 and yonger 85.19 27 0.02
22-25 87.24 948 0.78
26-35 89.28 29,060 23.97
36-45 90.21 44,494 36.7
46-55 90.34 30,737 25.35
56-65 90.76 13,209 10.89
66 and over 90.56 2,774 2.29
Total Number 121,249 100
TABLE A.4
DISTRIBUTION OF CLAIMS, BY CLAIM TYPE OBSERVED
Value of M 1 2 3 4 5 6 7 Total
Claim Type (y1) (y2) (y3) (y1, y2) (y1, y3) (y2, y3) (y1, y2, y3 )
Number 160 9,928 1,660 184 30 2,513 92 14,567
Percentage 1.1 68.15 11.4 1.26 0.21 17.25 0.63 100
0.30
0.10
0.00
0 10 20 30 40 50
TABLE A.5
FITTED NEGATIVE BINOMIAL MODEL
TABLE A.6
FITTED MULTI LOGIT MODEL
Parameter Estimates
TABLE A.7
MAXIMUM LIKELIHOOD ANALYSIS OF VARIANCE
TABLE A.8
FITTED SEVERITY MODEL BY COPULAS
Types of Copula
Types of Copula
OWN DAMAGE
s2 0.671 0.007 0.670 0.002 0.660 0.004
a12 5.570 0.151 5.541 0.144 5.758 0.103
a22 12.383 0.628 12.555 0.277 13.933 0.750
intercept 1.987 0.115 2.005 0.094 2.183 0.112
year –0.016 0.006 –0.015 0.006 –0.013 0.006
vehicle capacity 0.116 0.031 0.129 0.022 0.144 0.012
vehicle age %5 0.107 0.034 0.106 0.031 0.107 0.003
automobile*NCD 0-10 0.102 0.029 0.099 0.039 0.087 0.031
automobile*age 26-55 –0.047 0.027 –0.042 0.044 –0.037 0.005
automobile*age &56 0.101 0.050 0.080 0.018 0.084 0.050
THIRD PARTY PROPERTY
s3 1.320 0.068 1.309 0.066 1.349 0.068
a13 0.677 0.088 0.615 0.080 0.617 0.079
a23 1.383 0.253 1.528 0.271 1.324 0.217
intercept 1.071 0.134 1.035 0.132 0.841 0.120
vehicle age 1-10 –0.008 0.098 –0.054 0.094 –0.036 0.092
vehicle age &11 –0.022 0.198 0.030 0.194 0.078 0.193
year 0.031 0.007 0.043 0.007 0.046 0.007
COPULA
r12 – – 0.250 0.049 0.241 0.054
r13 – – 0.163 0.063 0.169 0.074
r23 – – 0.310 0.017 0.330 0.019
n – – – – 6.013 0.688
The simulation procedure used in this paper can be described in terms the three
component hierarchical model.
We start the simulation for the frequency component of the hierarchical pre-
dictive model. The number of accidents Ni for policyholder i follows negative
binomial distribution described in Section 2.3.1. We generate Ni using the prob-
abilities Pr(Ni = k) = a k r+-r 1- 1k pir (1 – pi ) k.
Then we generate the type of losses given an accident by simulating claim
type variable Mi for each policyholder from the distribution Pr(Mi = m) =
exp ]Vm g
7 , which is described in Section 2.3.2.
!s = 1exp ]Vs g
APPLICATIONS OF A HIERARCHICAL INSURANCE CLAIMS MODEL 197
where mk is the location parameter for a type k loss, defined by equation (6),
and sk represents the scale parameter for a type k loss.
After generating the three components of the predictive model, we can calculate
the simulated losses of the three type from j th accident for policy holder i:
Incorporating the number of accidents, we have the claims of each type from
a single policy:
3
Si, k = yijk ! 1 (Ni > n), k = 1, 2, 3
n=0