Articles - Connecting business volume and volatility


A key challenge faced by capital modellers is how to appropriately connect changes in premium and reserve volume with the volatility parameters used in capital models. On a theoretical level, we know that volatility typically decreases as volume increases. Despite this, we regularly see modelling processes where the business volumes evolve and the parameters don’t. This article explores why volume and volatility aren’t well joined up and suggests a research-driven formula for quickly adjusting your volatility parameters as volumes change.

 By Adam Smylie, Associate Consultant, Neil Gedalla, Principal and Ed Harrison, Partner from LCP

 Why are volume and volatility often not joined up?

 There are a number of reasons why it is surprisingly difficult to create a strong link between volume and volatility.

 Tight timescales: Modelling timescales are often very tight, so parameterisation is often done well in advance of the period to be modelled. This means that the modelled volumes of business may be very different to those in the historical data used to parameterise volatility.

 Parameter roll-forward: Volatility parameters (usually coefficients of variation, or “CoVs”), are often reviewed on eg a 3 year cycle. In the interim, CoVs tend to remain unchanged, even if the new business volumes or reserve volumes change. A full re-parameterisation of a risk area requires fresh data, judgement and validation.

 Lack of validation challenge: Common validation tests used to assess volatility assumptions include benchmarking CoVs against the market and back-testing against the historically observed volatility of the portfolio. These tests do not automatically allow for differences in volume and can potentially give false comfort if changes in volume are not factored in.

 Common examples of mismatched volume & volatility
 
 Reserve risk
 Reserve risk CoVs are often set in Q1 based on the previous year-end reserving data, so the volatility is assessed based on past reserve volumes. Typically these volatility parameters are then applied to expected reserve volumes at the forthcoming year-end, which are forecast based on Q2 projections.

 If there is a big difference in reserve volumes between the two successive year-ends, the CoVs set may no longer be appropriate if they are not volume-adjusted.

 New classes of business
 
 It can often be tricky to come up with an initial volatility assumption for a new class of business. This is mainly due to the lack of data available. A common solution for firms is to use volatility assumptions for another similar class of business written by the firm, ie use a benchmark.

 Again, this may cause a volume/volatility mismatch because the established class being used as a benchmark usually has a significantly greater volume than the new business line.
 
 Can we simply adjust volatility parameters to allow for volume?

 Challenges around model parameterisation timescales, resource and prioritisation are unlikely to go away. Given this, can we adjust volatility for volume without doing a full re-parameterisation? Having a simple data-driven formula that links volume and volatility would allow capital modellers to build dynamic volume-driven adjustment of CoVs into the capital process. This could include adjustments:

 To reflect differences in volume between past data used to set the CoV, and forward-looking business volumes the CoV is applied to. To keep the parameter in line with volumes of business written or reserves held in years when the CoV would otherwise not be updated.

 What might such a formula look like? A simple example might be expressed as follows:

 

 where V represents volume of business. A key judgement is the fine-tuning of the b parameter. A higher parameter implies the CoV is more sensitive to volume change.

 How well does this work in practice?
 
 In order to test the validity of the formula and to find sensible values for b, we carried out the following research:

 We performed two different types of analysis on the Schedule P dataset of the US National Association of Insurance Commissions (NAIC) return.

 Dataset: The NAIC Schedule P dataset contains historical reserving triangles for all US insurers over the period 2011 - 2022, with total reserves of $937 bn.
 
 Approach 1: We analysed how the volatility of the historical reserve deteriorations changed with reserve volume. We bucketed the data and fitted a trend line with the above relationship and a value for b of 0.22, resulting in an r-squared value of 91%, implying a strong model fit.
 
 Approach 2: We bootstrapped around 5,600 reserving triangles in the Schedule P dataset to obtain a CoV for each triangle. Plotting the CoV against the total reserves for the triangle allowed us to fit a similar trend line through the points. Bucketing the data in a similar way to the first approach resulted in a value for b of 0.21, with a corresponding r-squared value of 95%.

 Validation: We cross-checked our results against other data sources and regions. For example, we observed a similar relationship in our annual LCP capital benchmarking survey when comparing participants’ selected CoVs against their reserve volumes.

 In summary, our research suggested that the power relationship in the formula fits the data really well and that a value for b of around 0.22 is broadly applicable across a range of different datasets.

 As expected, we also found that b varied by class of business, for example, we found that it tended to be lower for classes more exposed to systemic risks.

 Next steps

 We think that adopting a scaling approach that links CoVs to volume addresses a current weakness of many firms’ modelling processes in a simple and efficient way.

 Overall, firms who create the strongest link between the capital modelling and the risks that business face will be best placed to understand those risks and manage them in line with the overall strategy.

 A more comprehensive summary of our research was presented at the IFoA General Insurance Spring Conference and is available to watch on-demand.
  

Back to Index


Similar News to this Story

The reserving actuary natural vs artificial intelligence
Why human actuaries still have the upper hand over AI when it comes to the nuanced art of reserving in the insurance industry. Every year, we take in
Five step approach vital for DB schemes looking to buyout
Insurers may refuse to quote and provide pricing for buy-ins and buy-outs where the DB pension schemes’ data is of a poor quality, warns Hymans Robert
What insurers must know about the hidden risks of silent AI
Anja Vischer, Senior Emerging Risk Manager at Swiss Re Institute, discusses the emerging risks of AI for insurers. She stresses the need to reassess c

Site Search

Exact   Any  

Latest Actuarial Jobs

Actuarial Login

Email
Password
 Jobseeker    Client
Reminder Logon

APA Sponsors

Actuarial Jobs & News Feeds

Jobs RSS News RSS

WikiActuary

Be the first to contribute to our definitive actuarial reference forum. Built by actuaries for actuaries.