Robertson Flashcards
Calculate Excess Losses at limit L
Expected XS Losses = E(L)*ELF(L)
ELF = XS Ratio = LER
Briefly describe the 1993 NCCI procedure to map classes into His
- NCCI defined 7 variables indicator of XS loss potential
- Because corr between them, the 7 variables were grouped into 3 subsets based on their partial correlations
- Principal components analysis was done and linear combination of the first 2 and last one was used to best explain variance
- Since there were 4 HGs prior to study, NCCI decided to keep 4 and thus continued to map each class to 1 of 4 HGs.
Briefly describe the 2 variables used by WC insurance ranting bureau of California (WCIRB) in mapping classes to HGs in 2001 analysis
- % claims in XS of $150K
- Diff between class loss distribution & avg loss distribution across all classes
Define Hazard Group (HG)
Collection of WC classifications that relatively similar ELFs over broad range of limits
Calculate XS Ratio (2007 analysis)
r = L/E(X)
S(r) = (E(y) - E(y;r)) / E(y)
Calculate credibility-weighted class XS ratio
Rc(final) = ZRc + (1-Z)Rhg
Rhg is the HG that contained the class prior to analysis
Z varies by class but not by limit
How was the credibility standard selected in 2007 analysis
Several alternatives were considered and none significantly impacted end results
Decided to stick with same crew formula used in prior review:
Z = min(1.5n/(n+k), 1)
n is the number of claims in class
k is avg number of claims per class
Discuss 3 other credibility options considered
- Using median instead of average for k
- Excluding med only claims from calculation of n and k (no impact on ELF)
- Including only serious claims in calculation of n and k
- Requiring min number of claims for classes used in calc of k
- Various square root formulas (ex: Z = (n/384)^0.5
Discuss why NCCI used 5 limits for cluster analysis
Prior to 2005, NCCI used and published ELFs for 17 different limits
Ultimately, they choose to use 5 of those limits for cluster analysis ($100K, $200K, $500K, $1M and $5M) for 2 reasons:
1. XS ratios at any pair of limits are highly corr across classes (especially for limits closer together)
2. Limits below $100K were heavily represented in 17 limits
Using fewer than 5 was also considered: while 1 limit would have not captured full variability in XS ratios, 2 principal components of 5 limits still explained 99% of variation in data.
Results of using 5 limits were not so different than 17 limits.
Ultimate driver decision: wanted to cover the range of limits commonly used for retro rating
Calculate Euclidean distance (L2)
L2 = square root of sum of (xi - yi)^2
Explain why NCCI decided to use L2 over L1
L1 = sum of abs value of (xi - yi) was also considered, this would have minimized relative error in estimating XS premiums.
L2 has 2 advantages:
1. Penalize large deviations: one big is worse and many small deviations (outliers have more impact)
2. Minimize squared error
Ultimate driver: analysis was not sensitive to distance measure so NCCI used traditional L2.
Discuss how XS ratios could have been standardized before clustering
XS ratios at lower limits are higher so without standardization the lower limits would naturally end up with more weight in clustering procedure
Standardization ensures each variable has similar impact on clusters.
NCCI explored 2 methods:
1. Zi = (xi - xbar)/s
2. Zi = (xi - minxi)/(maxxi - minxi)
Zi is standardization of xi
s is sample std dev
Discuss why NCCI did not standardize XS ratios
- Resulting HGs did to differ much from not using it
- XS ratios at different limits have similar unit of measure, which is $ of XS loss per $ of total loss
- All XS ratios are between 0 and 1, while standardization have to led to results outside of range
- There is greater range of XS ratios at lower limits & this is good thing since it is based on actual data
Standardization is appropriate when spread of values in data is due to normal random variation, but quite inappropriate if spread is due to presence of subclasses.
Describe the steps in using cluster analysis with k-means
Goal is to group classes with similar vectors of XS ratios as measured by L2 distance into HGs using premium weights
- Decide on k number of clusters (potential HGs)
- Start with arbitrary initial assignment of classes to k clusters
- Compute centroid of each cluster:
sumprod(prem, R(L))/sum(prem) - For each class, find closest centroid using L2 distance and assign class to that cluster
If any class has been re-assigned during step 4, go back to step 3 (iterative process)
Continue process until no classes is reassigned
What is the purpose of cluster analysis
Minimize variance within clusters and maximize variance between clusters
Means HGs will be homogeneous and well separated