A5 - A10 HARD Flashcards
Fully describe the bootstrapping process. Assume the data does not require any modifications
Shapland
- Calculate the fitted incremental claims using the GLM framework or the chain ladder age to age factors
- Calculate the residuals between fitted and actual incremental claims
- Create a triangle of random residuals by sampling with replacement from the set of non zero residuals
- Create a sample incremental triangle using the random residual triangle
- Accumulate the sample incremental triangle to cumulative
- Project the sample cumulative data to ultimate using the chain ladder method
- Calculate the reserve point estimate for each accident year using the projected data
- Iterate through this process to create a distribution of reserved for each accident year
Identify one deterministic method for reducing the variability in the extrapolation of future incremental values
Explain how this method can be stochastic
Shapland
Bornhuetter Ferguson method In addition to specifying a priori loss ratios for the BF method we can add a vector of standard deviations to go with these means. We can then assume a distribution and stimulate a different priori loss ratio for every iteration of the model.
Describe the MCMC methodology.
Verrall MCMC (Markov chain Monte Carlo) methods simulate the posterior distribution of a random variable by breaking the process down in a number of simulations. This is achieved by using the conditional dis- tribution of each parameter (given all of the others), making the simulation a univariate distribution. Considering each parameter in turn creates a Markov chain
Fully describe the steps required to implement a Bayesian model for the Bornhuetter/Ferguson method. Assume that prior distributions are defined for the column parameters and the row parameters.
Verrall ⇧ Define improper prior distributions for the column parameters and estimate the column parameters first. Since we are using improper prior distributions with large variances, the estimates will be those implied by the chain-ladder method ⇧ Define prior distributions for the row parameters x(i). In practice, these are often defined as gamma distributions where the beta parameter controls the level of confidence in the prior information ⇧ Using the x(i), re-parameterize the model in terms of lamda (i). We must do this since we defined prior distributions for the column parameters
Negative incremental values can cause extreme outcomes in early development periods. In partic- ular, they can cause large age-to-age factors. Describe four options for dealing with these extreme outcomes.
Shapland
- Identify the extreme iterations and remove them • Only remove unreasonable extreme iterations so that the probability of extreme out- comes is not understated
- Recalibrate the model • Identify the source of the negative incremental losses and remove it if necessary. For example, if the first row has negative incremental values due to sparse data, remove it and reparameterize the model
- Limit incremental losses to zero • This involves replacing negative incremental values with zeroes in the original triangles, zeroes in the sampled triangles OR zeroes in the projected future incremental losses. We can also replace negative incremental losses with zeroes based on their development column
- Use more than one model • For example, if negative values are caused by salvage/subrogation, we can model the gross losses and salvage/subrogation separately. Then, we can combine the iterations assuming 100% correlation
Describe the process for using an N-year weighted average of losses when determining development factors under the following frameworks: a) GLM framework
Shapland
We use N years of data by excluding the first few diagonals in the triangle (which leaves us with N + 1 included diagonals). This changes the shape of the triangle to a trapezoid. The excluded diagonals are given zero weight in the model and fewer calendar year parameters are required
Describe the process for using an N-year weighted average of losses when determining development factors under the following frameworks: b) Simplified GLM framework
Shapland
First, we calculate N-year average factors instead of all-year factors. Then, we exclude the first few diagonals when calculating residuals. However, when running the bootstrap simulations, we must still sample from the entire triangle so that we can calculate cumulative values. We use N-year average factors for projecting the future expected values as well
Provide two reasons why the coe fficient of variation may rise in the most recent accident years.
Shapland
- With an increasing number of parameters in the model, parameter uncertainty increases when moving from the oldest years to the most recent years. This parameter uncertainty may overpower the process uncertainty, causing an increase in variability
- The model may simply be overestimating the variability in the most recent years
Describe two methods for combining the results of multiple stochastic models.
Shapland
- Run models with the same random variables • Each model is run with the exact same random variables. Once all of the models have been run, the incremental values for each model are weighted together (for each iteration by accident year)
- Run models with independent random variables • Each model is run with its own random variables. Once all of the models have been run, weights are used to select a model (for each iteration by accident year).The result is a weighted mixture of models
Provide four reasons for fitting a curve to unpaid claim distributions.
Shapland
- Assess the quality of the fit
- Parameterize a DFA (dynamic financial analysis) model
- Estimate extreme values
- Estimate TVaR
Describe the process for creating distribution graphs. Include a discussion on kernel density func- tions.
Shapland
We can create a total unpaid distribution histogram by dividing the range of all values generated from the simulation into 100 equally sized buckets, and then counting the number of simulations that fall within each bucket. Since simulation results tend to appear jagged, a Kernel density function can be fit to the data to provide a smoothed distribution. Each point of a Kernel density function is estimated by weighting all of the values near that point, with less weight given to points further away
a) Briefly describe two methods for including correlation between bootstrap distributions for different business segments.
b) For each method, identify two advantages.
Shapland
Part a:
- Location mapping • Pick a business segment. For each bootstrap iteration, sample a residual and then note where it belonged in the original residual triangle. Then, sample each of the segments using the residuals at the same locations for their respective residual triangles. This preserves the correlation of the original residuals in the sampling process
- Re-sorting • To induce correlation among business segments in a bootstrap model, re-sort the resid- uals for each business segment until the rank correlation between each segment matches the desired correlation
Part b:
- Location mapping • Can be easily implemented in a spreadsheet • Does not require a correlation matrix
- Re-sorting • Works for residual triangles with different shapes/sizes • Different correlation assumptions can be employed
Define the following terms:
- Claims portfolio
- Valuation classes
- Claim group
Marshall
Claims portfolio – the aggregate portfolio for which the risk margins must be estimated
Valuation classes – the portfolios that are considered individually as part of the risk margin analysis
Claim group – a group of claims with common risk characteristics
Explain how scenario testing can be used in conjunction with risk margin analyses.
Marshall
Scenario testing can be used to determine how key assumptions underlying the risk margin calculation would need to change in order to produce a risk-loaded actuarial central estimate. These scenarios include changes in claim frequencies, claim severities, loss ratios, etc.
Describe three tests for uniformity for n predicted percentiles.
Meyers Histogram - if the percentiles are uniformly distributed, the height of the bars should be equal. p-p plot - we first sort them into increasing order. The expected value of these percentiles is given by {ei} = 100*{1/(n+1),2/(n+1),…,n/(n+1)}. We then plot the expected percentiles on the x-axis and the sorted predicted percentiles on the y-axis. If these predicted percentiles are uniformly distributed, we expect this plot to lie along a 45 degree line. K-S test - we can test the significane of the percentiles visually. Using the K-S test, we can reject the hypothesis that a set of percentiles is significant at the 5% level if the K-S statistic D = max(abs(pi - fi)) is greater than its critical value, 136/sqrt(n) where {fi} = 100*{1/n,2/n,…,n/n}. On a p-p plot, these critical values appear as 45 degree bands that run parallel to the line y=x. We reject the hypothesis of uniformity if the p-p plot lies outside the bands.