Mahler - shifting risk parameters Flashcards
Mahler uses baseball team win/loss records to
illustrate the potentially significant impact of shifting risk parameters on credibilities and experience rating
-Mahler notes 3 simplifications that are found in baseball set that are not found in insurance data
- constant set of risks (teams)
- baseball loss data is readily available, accurate, and not subject to development
- each risk (team) is of equal size; they each roughly play the same # of games each year
standard formula used for credibility weighting
new estimate = data*credibility+(1-credibility)*prior estimate
-often prior estimate is class average or previous estimate for particular insured/group
to answer whether there is an inherent difference between baseball teams
-Mahler calculates average and std dev of losing percentage of each team
std dev = sqrt(np(1-p))
p=losing %, assumed to be 505
-since there are many teams outside of 2 std dev of mean losing percentage (divide std dev by n to put in % terms), concludes that there are inherent differences between teams
observes that a team that has been worse than average over 1 period of time is likely to
continue to be worse than average over another period of time which implies there is value in using past experience of team to predict future experience of that team
2 methods to test: are the observed team loss percentages over time random fluctuations?
- essentially asking whether risk parameters for team are changing over time
1. chi squared
2. compare correlation between years aka lagged correlations test
chi squared test
Null hypothesis: risk parameters do no shift over time
- group data into appropriate intervals, calculate overall expected value
- calculate χ2=(A-E)^2/E
- sum up for all interval
- if there are n time period groups, χ2 table value to compare against n-1 DoF
- if test statistic > tabular value, then reject null and conclude risk parameters have shifted over time
Compare correlation between years
aka
lagged correlations test
- group data by pairs based on time lag (separation in time)
- calculate correlation between for each pair
- calculate average correlation for each time lag
- if correlation decreases as time lag increases, risk parameters shift over time
*implies risk parameters are shifting over time and recent years can help predict future
using both tests, concluded
risk parameters of teams are changing over time -> they are not random fluctuations in the same distribution
3 criteria to evaluate the quality of solutions (different credibility weighted options)
- Least Squared Error
- Limited Fluctuation
- Meyers/Dorweiler
Least Squared Error
- calculate the mean squared error of prediction compared with actual observed result
- Buhlmann/Bayesian credibility methods attempt to minimize this criteria
*minimize squared error between actual and predicted result
final points
- Mahler notes that when there are shifting risk parameters over time, older years of data will be less relevant predicting the future and so should be given less or no credibility compared with recent years
- Related to prior item, Mahler looks at the impact of having a delay in getting historical data as is the case with experience rating; notes that not having the most recent year of historical data significantly increases the squared error of the estimate
Limited Fluctuation
- AKA Small Chance of Large Errors
- measures the probability that observed result differs by more than a certain percent from predicted result
- classical credibility method targets this criteria
*minimize likelihood that any one actual observation will be certain percent different from predicted result
Meyers/Dorweiler
- calculates correlation between ratio of actual to expected losses and ratio of predicted losses to overall losses
- 2 vectors used: V1=actual team t losing %/estimated team t losing %
V2=estimated team t losing %/mean losing %
-Kendall tau statistic is used to compute the correlation between 2 vectors
*minimize correlation between ratio of actual/predicted and predicted/average actual
Meyers/Dorweiler criteria confirms
that there is no evidence that large predictions lead to large errors and small predictions lead to small errors