1. Introduction
Traditional auto insurance premiums have been based on
driver-related risk (demographic) factors such as age, gender, marital status, claim history, credit risk and living district, and
vehicle-related risk factors such as vehicle year/make/model, which represent the residual value of an insured vehicle. Although these
traditional variables or factors indicate claim frequency and size, they do not reflect true driving risk and often lead to cross-subsidising higher-risk drivers by lower-risk drivers to balance the claim cost. These premiums have been criticised for being inefficient and socially unfair because they do not punish aggressive driving nor encourage prudent driving.
Chassagnon and Chiappori (
1997) reported that the accident risk depends not only on demographic variables but also on driver behaviour that reflects how drivers drive cautiously to reduce accident risk.
Usage-based insurance (UBI) relies on telematic data, often augmented by global positioning systems (GPSs), to gather vehicle information. UBI encompasses two primary models: Pay As You Drive (PAYD) and Pay How You Drive (PHYD). PAYD operates on a drive-less-pay-less principle, taking into account
driving habits and travel details such as route choices, travel time, and mileage. This model represents a significant advancement over traditional auto insurance approaches. For instance,
Ayuso et al. (
2019) utilised a Poisson regression model to analyse a combination of seven traditional and six travel-related variables. However,
Kantor and Stárek (
2014) highlighted limitations in PAYD policies, notably their sole focus on kilometres driven, neglecting crucial driver behaviour aspects.
By integrating a telematics device into the vehicle, PHYD extends the principles of PAYD to encompass the monitoring of driving behaviour profiles over a specified policy period.
Driving behaviour, encompassing operational choices such as speeding, harsh braking, hard acceleration, or sharp cornering in varying road types, traffic conditions, and weather, serves as a defining aspect of drivers’
styles (
Tselentis et al. 2016;
Winlaw et al. 2019). These collected driving data offer valuable insights into assessing true driving risks, enabling the calculation of the subsequent UBI experience rating premium. This advancement over traditional premiums incorporates both historical claim experiences and current driving risks. The UBI premium can undergo regular updates to provide feedback to drivers, incentivising improvements in driving skills through premium reductions. Research by
Soleymanian et al. (
2019) indicated that individuals drive less and safer when incentivised by UBI premiums. Moreover,
Bolderdijk et al. (
2011) demonstrated that monitoring driving behaviours can effectively reduce speeding and accidents by promoting drivers’ awareness and behavioural changes.
Wouters and Bos (
2000) showed that monitoring of driving resulted in a 20% reduction in accidents. The monitoring system enables early intervention for risky drivers, potentially saving lives (
Hurley et al. 2015). Finally,
Ellison et al. (
2015) concluded that personalised feedback coupled with financial incentives yields the most significant changes in driving behaviour, emphasising the importance of a multifaceted approach to risk reduction.
The popularity of PHYD policies has surged in recent years, driven by the promise of lower premiums for safe driving behaviour. QBE Australia (Q for Queensland Insurance, B for Bankers’ and Traders’ and E for The Equitable Probate and General Insurance Company), renowned for its innovative approaches, introduced a product called the
Insurance Box, a PHYD policy featuring in-vehicle telematics. This product not only offers lower premiums to good drivers but also delivers risk scores as actionable feedback on driving performance. These risk scores directly influence the calculation of insurance premiums. In essence, PHYD policies epitomise personalised insurance (
Barry and Charpentier 2020), nurturing a culture of traffic safety while concurrently reducing congestion and environmental impact by curbing oil demand and pollutant emissions.
To assess UBI premiums, extensive driving data are initially gathered via telematics technology. Subsequently, a comprehensive set of driving behaviour variables, termed driving variables (DVs), is generated. These variables encompass four main categories: driver-related, vehicle-related, driving habits, and driving behaviours. These DVs are then analysed through regression against insurance claims data to unveil correlations between driving habits and associated risks, which is a process commonly referred to as knowledge discovery (
Murphy 2012).
Stipancic et al. (
2018) determined drivers’ risk by analysing the correlations of accident frequency and accident severity with specific driving behaviours such as hard braking and acceleration.
To forecast and model accident frequencies,
Guillen et al. (
2021) proposed utilising Poisson regression models applied to both traditional variables (related to drivers, vehicles, and driving habits) and critical incidents, which encompass risky driving behaviours. Through this approach, the study delineates insurance premiums into a baseline component and supplemental charges tailored to
near-miss events—defined as critical incidents like abrupt braking, acceleration, and smartphone usage while driving—which have the potential to precipitate accidents. Building on this,
Guillen et al. (
2020) employed negative binomial (NB) regression, regressing seven traditional variables, five travel-related factors, and three DVs to the frequency of near-miss events attributable to acceleration, braking, and cornering. Notably, the study suggests that these supplementary charges stemming from near-miss events could be dynamically updated on a weekly basis.
To comprehensively assess the potential nonlinear impacts of DVs on claim frequencies,
Verbelen et al. (
2018) utilised Poisson and negative binomial regression models within generalised additive models (GAMs). They focused on traditional variables, as well as telematic risk exposure DVs, such as total distance driven, yearly distance, number of trips, distance per trip, distance segmented by road type (urban, other, motorways, and abroad), time slot, and weekday/weekend. While these exposure-centric DVs can serve as offsets in regression models, they fail to capture the subtle details of actual driving behaviour. To attain a deeper understanding of driving behaviour, it becomes essential to extract a broader array of DVs that can discern between safe and risky driving practices while also delineating claim risk. For instance, rather than merely registering a braking event, a more comprehensive approach involves constructing a detailed ’braking story’ that accounts for various factors such as road characteristics (location, lanes, angles, etc.), braking style (abrupt, continuous, repeated, intensity, etc.), braking time (time of day, day of the week, etc.), road type (speed limit, normative speed, etc.), weather conditions, preceding actions (turning, lane changing, etc.), and more. Furthermore, the inclusion of additional environmental and traffic variables obtained through GPS enhances the richness of available information, facilitating a more thorough analysis of driving behaviour and associated risk factors.
As the number of variables describing driving behaviour increases, the data can become voluminous, volatile, and noisy. Managing this influx of variables is crucial to mitigate computational costs and address the issue of multicollinearity among them. Multicollinearity arises due to significant overlap in the predictive power of certain variables. For instance, a driver residing in an area with numerous traffic lights might engage in more forceful braking, or an elderly driver might tend to drive more frequently during midday rather than during typical rush hours or late nights. Consequently, it is possible for confusion to arise between factors such as location and aggressive braking or age and preferred driving times. Multicollinearity can lead to overfitting and instability in predictive models, diminishing their effectiveness. Thus, streamlining the variables to a manageable number is not only essential for computational efficiency but also critical for addressing multicollinearity and enhancing the reliability of predictive models.
Machine learning, employing statistical algorithms, holds remarkable potential in mitigating overfitting and bolstering the stability and predictability of various predictive models. These algorithms are typically categorised as supervised or unsupervised. In the realm of unsupervised learning,
Osafune et al. (
2017) conducted an analysis wherein they developed classifiers to distinguish between safe and risky drivers based on acceleration, deceleration, and left-acceleration frequencies gleaned from smartphone-equipped sensor data from over 800 drivers. By labelling drivers with at least 20 years of driving experience and no accident records as safe and those with more than two accident records as risky, they achieved a validation accuracy of 70%.
Wüthrich (
2017) introduced pattern recognition techniques utilising two-dimensional velocity and acceleration (VA) heat maps via K-means clustering. However, it is worth noting that neither study offers predictions related to insurance claims.
With claim risk information such as claim making, claim frequency, and claim size, supervised machine learning models embedded within generalised linear models (GLMs) can be constructed to unfold the hidden patterns in big data and predict future claims for premium pricing. Various machine learning techniques are widely utilised in predictive modelling, including clustering, decision trees, random forests, gradient boosting, and neural networks.
Gao et al. (
2019) investigated the effectiveness of Poisson GAMs, integrating two-dimensional speed–acceleration heat maps alongside traditional risk factors for predicting claim frequencies. They employed feature extraction methods outlined in their previous work (
Gao and Wüthrich 2018), such as K-medoids clustering to group drivers with similar heatmaps and principal component analysis (PCA) to reduce the dimensionality of the design matrix, thereby enhancing computational efficiency. Furthermore,
Gao et al. (
2019) conducted an extensive analysis focusing on the predictive power of additional driving style and habit covariates using Poisson GAMs. In a similar vein,
Makov and Weiss (
2016) integrated decision trees into Poisson predictive models, expanding the repertoire of predictive algorithms in insurance claim forecasting.
In assessing various machine learning techniques,
Paefgen et al. (
2013) discovered that neural networks outperformed logistic regression and decision tree classifiers when analysing claim events using 15 travel-related variables.
Ma et al. (
2018) employed logistic regression to explore accident probabilities based on four traditional variables and 13 DVs, linking these probabilities to insurance premium ratings.
Weerasinghe and Wijegunasekara (
2016) categorised claim frequencies as low, fair, and high and compared neural networks, decision trees, and multinomial logistic regression models. Their findings indicated that neural networks achieved the best predictive performance, although logistic regression was recommended for its interpretability. Additionally,
Huang and Meng (
2019) utilised logistic regression for claim probability and Poisson regression for claim frequency, incorporating support vector machine, random forest, advanced gradient boosting, and neural networks. They examined seven traditional variables and 30 DVs grouped by travel habits, driving behaviour, and critical incidents, employing stepwise feature selection and providing an overview of UBI pricing models integrated with machine learning techniques.
However, machine learning techniques often encounter challenges with overfitting. One strategy to address both multicollinearity and overfitting is to regularise the loss function by penalising the likelihood based on the number of predictors. While ridge regression primarily offers shrinkage properties, it does not inherently select an optimal set of predictors to capture the best driving behaviours.
Tibshirani (
1996) introduced the Least Absolute Shrinkage and Selection Operator (lasso) regression, incorporating an L1 penalty for the predictors. Subsequently, the lasso regression framework underwent enhancements to improve model fitting and variable selection processes. For instance,
Zou and Hastie (
2005) proposed the elastic net, which combines the L1 and L2 penalties of lasso and ridge methods linearly.
Zou (
2006) introduced the adaptive lasso, employing adaptive weights to penalise different predictor coefficients in the L1 penalty. Moreover,
Park and Casella (
2008) presented the Bayesian implementation of lasso regression, wherein lasso estimates can be interpreted as Bayesian posterior mode estimates under the assumption of independent double-exponential (Laplace) distributions as priors on the regression parameters. This approach allows for the derivation of Bayesian credible intervals of parameters to guide variable selection.
Jeong and Valdez (
2018) expanded upon the Bayesian lasso framework proposed by
Park and Casella (
2008) by introducing conjugate hyperprior distributional assumptions. This extension led to the development of a new penalty function known as log-adjusted absolute deviation, enabling variable selection while ensuring the consistency of the estimator. While the Bayesian approach is appliable, the running of MCMC is often time-consuming.
When modelling claim frequencies, a common issue arises from an abundance of zero claims, which Poisson or negative binomial regression models may not effectively capture. These zero claims do not necessarily signify an absence of accidents during policy terms but rather indicate that some policyholders, particularly those pursuing no-claim discounts, may refrain from reporting accidents. To identify factors influencing zero and nonzero claims,
Winlaw et al. (
2019) employed logistic regression with lasso regularisation on a case–control study, assessing the impact of 24 DVs on acceleration, braking, speeding, and cornering. Their findings highlighted speeding as the most significant driver behaviour linked to accident risk. In a different approach,
Guillen et al. (
2019) and
Deng et al. (
2024) utilised zero-inflated Poisson (ZIP) regression models to model claim frequencies directly and
Tang et al. (
2014) further integrated the model with the EM algorithm and adaptive lasso penalty. However,
Tang et al. (
2014) observed suboptimal variable selection results for the zero-inflation component, suggesting a lower signal-to-noise ratio compared to the Poisson component.
Banerjee et al. (
2018) proposed a multicollinearity-adjusted adaptive lasso approach employing ZIP regression. They explored two data-adaptive weighting schemes: inverse of maximum likelihood estimates and inverse estimates divided by their standard errors. For a comprehensive overview of various modelling approaches in UBI, refer to Table A1 in
Eling and Kraft (
2020).
Numerous studies in UBI have employed a limited number of DVs to characterise a broad spectrum of driver behaviours. For instance,
Jeong (
2022) analysed synthetic telematic data sourced from
So et al. (
2021), encompassing 10 traditional variables and 39 DVs, including metrics like sudden acceleration and abrupt braking. While
Jeong (
2022) utilised PCA to reduce dimensionality and enhance predictive model stability, the interpretability of the principal components derived from PCA remains constrained. Regularisation provides a promising alternative for dimension reduction while addressing the challenge of overfitting. The literature on UBI predictive models employing GLMs with machine learning techniques, such as lasso regularisation to mitigate overfitting, is still relatively sparse, particularly concerning forecasting claim frequencies and addressing challenges like excessive zero claims and overdispersion arising from heterogeneous driving behaviours.
Our main objective is to propose predictive models to capture the impact of driving behaviours (safe or risky) on claim frequencies, aiming to enhance prediction accuracy, identify relevant DVs, and classify drivers based on their driving behaviours. This segmentation will enable the application of differential UBI premiums for safe and risky drivers. More importantly, we advocate for the regular updating of these UBI premiums to provide ongoing feedback to drivers through the relevant DVs and encourage safer driving habits.
We demonstrate the applicability of the proposed predictive models through a case study using a representative telematics dataset comprising 65 DVs. The proposed predictive models includes two-stage threshold Poisson (TP), Poisson mixture (PM), and ZIP regression models with lasso regularisation. We extend the regularisation technique to include adaptive lasso and elastic net, facilitating the identification of distinct sets of DVs that differentiate safe and risky behaviours. In the initial stage of regularised TP models, drivers are classified into risky (safe) group if their annual predicted claim frequencies, estimated by a single-component Poisson model, exceed (not exceed) predefined thresholds. Subsequently, in stage two, regularised Poisson regression models are refitted to each driver subgroup (exceeding thresholds or not) using different sets of selected DVs in each group. Alternatively, PM models simultaneously estimate parameters and classify drivers. Our findings reveal that PM models offer greater efficiency compared to TP models, providing added flexibility and capturing overdispersion akin to NB distributions.
In ZIP models, we observe that the structural zero component may not necessarily indicate safe drivers, as safe drivers may claim less frequently but not necessarily abstain from claims altogether, while risky drivers may avoid claims due to luck or incentives like bonus rewards.
So et al. (
2021) investigated the cost-sensitive multiclass adaptive boosting method, defining classes based on zero claims, one claim, and two or more claims, differing from our proposed safe and risky driver classes. We argue that the level of accident risk may not solely correlate with the number of claims but rather with driving behaviours. Hence, the regularised PM model proves more efficient in tracking the impact of DVs on claim frequencies, allowing for divergent effects between safe and risky drivers. This proposed PM model constitutes the primary contribution of this paper, addressing a critical research gap in telematics data analysis.
Our second contribution is to bolster the robustness of our approach and mitigate overfitting by incorporating resampling and cross-validation (CV) apart from lasso regularisation. These techniques help us attain more stable and reliable results. Additionally, we utilise the area under curve (AUC) from the receiver operating characteristic (ROC) curve as one of the performance metrics, which evaluates classification accuracy highlighting the contribution of predictive models in classifying drivers.
Our third contribution involves introducing an innovative UBI experience rating premium method. This method extends the traditional experience rating premium method by integrating classified claim groups and predicted claim frequencies derived from the best-trained model. This dynamic pricing approach also enables more frequent update of premiums to incentivise safer and reduced driving. Moreover, averaged and individual driving scores from the identified relevant DVs for each driver can inform their driving behaviour possibly with warnings and encourage skill improvement. By leveraging these advanced premium pricing models, we can improve loss reserving practices, and we can even evaluate the legitimacy of reported accidents based on driving behaviours.
Lastly, we highlight a recent paper (
Duval et al. 2023) with similar aims to this paper. They applied logistic regression with elastic net regularisation to predict the probability of claims, but this paper considers two-group PM regression instead of logistic regression to predict claim frequency and allow different DVs to capture the distinct safe and risky driving behaviours. For predictive variables, they used driving habits information (when, where, and how much the insured drives) from telematics, as well as traditional risk factors such as gender and vehicle age, whereas this paper focuses on driving behaviour/style (how the insured drives). To avoid handcrafting of telematics information, they proposed measures using the Mahalanobis method, Local Outlier Factor, and Isolation Forest to summarise trip information into local/routine anomaly scores by trips and global/peculiar anomaly scores by drivers, which were used as features. This is an
innovative idea in the literature. On the other hand, this paper uses common handcraft practices to summarise driving behaviour by drivers, using both driving habits (where and when) and driving styles (how) information by defining driving events such as braking and turning considering time, location, and severity of events.
Duval et al. (
2023) demonstrated that the improvement in classification using lower global/peculiar Mahalanobis anomaly scores enables a more precise pure premium (product of claim probability from logistic regression to insured amount) calculation. As stated above, this paper provides differential contributions by classifying drivers into safe and risky groups, predicting claims for drivers in their groups using regularised PM models (among regularised TP and ZIP models), which is
pioneering in the UBI literature, and calculating premiums using the proposed
innovative UBI experience rating premium based on drivers’ classifications (safe/risky) and predicted annual claims.
The paper is structured as follows:
Section 2 outlines the GLMs, including Poisson, PM, and ZIP regression models, alongside lasso regularisation and its extensions.
Section 3 presents the telematics data and conducts an extensive empirical analysis of the two-stage TP, PM, and ZIP models.
Section 4 introduces the proposed UBI experience rating premium method. Lastly,
Section 5 offers concluding remarks and implementation guidelines and explores potential avenues for future research.
4. UBI Experience Rating Premium
In the context of general insurance, a common approach for assessing risk in the typical short-tail portfolio involves multiplying predicted claims frequency by claims severity to determine the risk premium. This derived risk premium is subsequently factored into the profit margin, alongside operating expenses, to determine the final premium charged to customers. This paper centers on claims frequency, and in the premium calculation discussed herein, we assume that claim severity remains constant. Consequently, the premium calculation relies on predicting claims frequency.
The traditional experience rating method prices premiums using historical claims and offers the same rate for drivers within the same risk group (low/high or safe/risky). If individual claim history is available, premiums can be calculated using individual claims relative to overall claims—both from historical records. However, although this extended historical experience rating method can capture the individual differences of risk within a group, it still fails to reflect drivers’ recent driving risk. The integration of telematic data enables us to tailor pricing to current individual risks. This enhanced method is called the UBI experience rating method. We leverage premium pricing as a strategic approach to refine our pricing methodology.
Suppose that a new driver
i was classified to claim group
g with index set
of all drivers in this group and
. Let
be his premium for year
t,
be the historical claim/loss in year
,
be the total claim/loss from the claim group
g that driver
i was classified to, and
be the total premium from the claim group
g. Moreover, suppose that the best PMA model was trained using the sample of drivers. Let
be the observed DVs for driver
i at time
t,
safe group (
risky group) be the classified group if the group indicator
(otherwise),
in (
4) be the predicted claim frequency given
,
be the average predicted claim frequencies from the claim group
g that driver
i was classified to, and
be the size of group
g.
Using the proposed
UBI experience rating method, the premium
for driver
i in year
t is given by
where
is the group average annual premium in period
t from the group data,
is the average annual premium from all data or some other data source,
F is the credibility factor (
Dean 1997),
is the exposure of driver
i, and
is the individual adjustment factor to the overall group loss ratio given by
which is the sum of the
historical loss rate change adjustment
and weighted
UBI predicted loss rate change adjustment
;
is the
UBI policy parameter to determine how much UBI adjustment is applied to
in
when updating the premium to account for current driving behaviour. The
historical loss rate change
, historical individual loss ratio
, and historical group loss ratio
are, respectively,
The
UBI predicted loss rate change
,
UBI predicted individual loss ratio
, and
UBI predicted group loss ratio
are, respectively,
The credibility factor
F is the weight of the best linear combination between the premium estimate
using the sample data to the premium estimate
using all data or data from another source to improve the reliability of the premium estimate
. The credibility factor increases with the business size and, hence, the number of drivers in the sample.
Dean (
1997) provided some methods to estimate
F and suggested full credibility
when the sample size
N is large enough, such as above 10,000 in an example. As this requirement is fulfilled for the telematic data with size
14,157, and all data are used to estimate the chosen PMA model, a full credibility of
was applied. In cases where insured vehicles are less in number in the sample, the credibility factor
F may vary, and external data sources may be used to improve the reliability of the premium estimate. Moreover, as the selected PMA model can classify drivers, the premium calculation can focus on the classified driver group to provide a more precise premium calculation.
We give an example to demonstrate the experience rating method and its extension to UBI. Suppose that driver
i is classified as a safe driver (
) in a driving test and wants to buy auto insurance for the next period (
). As summarised in
Table 4, the annual premium for the safe group is
and for the risky group is
. Driver
i has recorded
in annual claim frequency and paid an annual premium of
before. The safe group has recorded an average of
in annual claim frequency and paid an annual premium if
per driver before. The risky group has recorded an average of
claims/loss and paid
in annual premium per driver before. Driver
i has more claims than the average of a safe group before. According to these historical claim frequencies, driver
i is expected to be relatively more risky than the average of the safe group, so he should pay more.
To illustrate the UBI experience rating method, additional assumptions about the predicted annual claim frequencies for driver
i have been added to the last row of
Table 4. Assume that driver
i has a predicted annual claim frequency
before; then, that of the safe group is
, and that of the risk group is
. This suggests that driver
i operates his vehicle more safely than his historical claims indicate. This information is summarised in
Table 4.
Taking the policy parameter
, the UBI experience rating premium is given by
where the
historical loss rate change
, the historical loss ratio
for driver
i, the historical loss ratio for safe group
, the
UBI predicted loss rate change
, the UBI predicted loss ratio
for driver
i, and the UBI predicted loss ratio
for the safe group are, respectively,
using (
12). So, the premium for driver
i using the UBI experience rating method is
$337.80. This premium is higher than the premium
$300 for the safe group because the loss ratio for driver
i is higher relative to the overall ratio in the safe group using historical claims. However, his current loss ratio due to current safe driving reduces the adverse effect due to the higher historical claims.
Nevertheless, we recognise that not all insured vehicles are equipped with telematic devices, introducing potential data gaps in the telematics insights. In response to this challenge, the
UBI policy parameter
in (
13) can be set to 0. This adaptation to the UBI pricing model in (
12) also allows for the application to newly insured drivers with only historical records (traditional demographic variables). This premium called
historical experience rating premium for driver
i during period
t is
where the
historical loss rate change
is given by (
16). This loss rate change can capture individual differences within a claim group using historical claims but fails to reflect the recent driving risk. Hence, this premium is higher than the UBI experience rating premium calculated using both historical and current driving experience. Thus, the historical experience rating method is unable to provide immediate compensation/reward for safe driving.
Moreover, the UBI premium can track driving behaviour more frequently and closely using regularly updated claim class and annual claim frequency prediction
. The updating period can be reduced to monthly or even weekly to provide more instant feedback using the live telematic data. In summary, the proposed UBI experience rating premium provides a correction of the loss rate change
of the experience rating only premium using the sum of both the
historical loss rate change
and the
UBI predicted loss rate change
. Here, the proposed PMA model can predict more instantly the annual claim frequencies
using live telematic data. Hence, the UBI premium can be updated more frequently to provide incentives for safe driving. The proposed UBI experience rating premium provides an incremental innovation to the business processes allowing the company to gradually transit to the new regime of UBI by adjusting the UBI policy factor
in (
13) such that
can gradually increase from 0 to 1 if driver
i wants his premium to gradually account for his current driving.
We remark that our analyses made a few assumptions. Firstly, we assumed that the annual premium
covers the total cost with possibly some profit, and the expectations of loss ratios
and
across drivers
i in group
g are around zero. To assess the validity of the assumptions on expectations, one can obtain the distributions of
based on the most recent data. If their means
are not zero, the overall loss ratio
in (
13) can be adjusted as
for group
g. For conservative purposes, the means
can be replaced by say 75% quantiles
of the distributions. Secondly, it also implicitly assumes perfect or near-perfect monitoring. However, the advent of monitoring technologies reduces the extent of asymmetric information between insureds and insurers and reduces moral hazard costs.