Explainable Boosting Machine for Predicting Claim Severity and Frequency in Car Insurance
- URL: http://arxiv.org/abs/2503.21321v1
- Date: Thu, 27 Mar 2025 09:59:45 GMT
- Title: Explainable Boosting Machine for Predicting Claim Severity and Frequency in Car Insurance
- Authors: Markéta Krùpovà, Nabil Rachdi, Quentin Guibert,
- Abstract summary: We introduce an Explainable Boosting Machine (EBM) model that combines intrinsically interpretable characteristics and high prediction performance.<n>We implement this approach on car insurance frequency and severity data and extensively compare the performance of this approach with classical competitors.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In a context of constant increase in competition and heightened regulatory pressure, accuracy, actuarial precision, as well as transparency and understanding of the tariff, are key issues in non-life insurance. Traditionally used generalized linear models (GLM) result in a multiplicative tariff that favors interpretability. With the rapid development of machine learning and deep learning techniques, actuaries and the rest of the insurance industry have adopted these techniques widely. However, there is a need to associate them with interpretability techniques. In this paper, our study focuses on introducing an Explainable Boosting Machine (EBM) model that combines intrinsically interpretable characteristics and high prediction performance. This approach is described as a glass-box model and relies on the use of a Generalized Additive Model (GAM) and a cyclic gradient boosting algorithm. It accounts for univariate and pairwise interaction effects between features and provides naturally explanations on them. We implement this approach on car insurance frequency and severity data and extensively compare the performance of this approach with classical competitors: a GLM, a GAM, a CART model and an Extreme Gradient Boosting (XGB) algorithm. Finally, we examine the interpretability of these models to capture the main determinants of claim costs.
Related papers
- Supervised Optimism Correction: Be Confident When LLMs Are Sure [91.7459076316849]
We establish a novel theoretical connection between supervised fine-tuning and offline reinforcement learning.
We show that the widely used beam search method suffers from unacceptable over-optimism.
We propose Supervised Optimism Correction, which introduces a simple yet effective auxiliary loss for token-level $Q$-value estimations.
arXiv Detail & Related papers (2025-04-10T07:50:03Z) - Latenrgy: Model Agnostic Latency and Energy Consumption Prediction for Binary Classifiers [0.0]
Machine learning systems increasingly drive innovation across scientific fields and industry.<n>Yet challenges in compute overhead, specifically during inference, limit their scalability and sustainability.<n>This study addresses critical gaps in the literature, chiefly the lack of generalized predictive techniques for latency and energy consumption.
arXiv Detail & Related papers (2024-12-26T14:51:24Z) - GenFollower: Enhancing Car-Following Prediction with Large Language Models [11.847589952558566]
We propose GenFollower, a novel zero-shot prompting approach that leverages large language models (LLMs) to address these challenges.
We reframe car-following behavior as a language modeling problem and integrate heterogeneous inputs into structured prompts for LLMs.
Experiments on Open datasets demonstrate GenFollower's superior performance and ability to provide interpretable insights.
arXiv Detail & Related papers (2024-07-08T04:54:42Z) - Cycles of Thought: Measuring LLM Confidence through Stable Explanations [53.15438489398938]
Large language models (LLMs) can reach and even surpass human-level accuracy on a variety of benchmarks, but their overconfidence in incorrect responses is still a well-documented failure mode.
We propose a framework for measuring an LLM's uncertainty with respect to the distribution of generated explanations for an answer.
arXiv Detail & Related papers (2024-06-05T16:35:30Z) - LoRA-Ensemble: Efficient Uncertainty Modelling for Self-attention Networks [52.46420522934253]
We introduce LoRA-Ensemble, a parameter-efficient deep ensemble method for self-attention networks.<n>By employing a single pre-trained self-attention network with weights shared across all members, we train member-specific low-rank matrices for the attention projections.<n>Our method exhibits superior calibration compared to explicit ensembles and achieves similar or better accuracy across various prediction tasks and datasets.
arXiv Detail & Related papers (2024-05-23T11:10:32Z) - Model-based Causal Bayesian Optimization [74.78486244786083]
We introduce the first algorithm for Causal Bayesian Optimization with Multiplicative Weights (CBO-MW)
We derive regret bounds for CBO-MW that naturally depend on graph-related quantities.
Our experiments include a realistic demonstration of how CBO-MW can be used to learn users' demand patterns in a shared mobility system.
arXiv Detail & Related papers (2023-07-31T13:02:36Z) - Enhanced Gradient Boosting for Zero-Inflated Insurance Claims and Comparative Analysis of CatBoost, XGBoost, and LightGBM [0.0]
CatBoost is the best library for developing auto claim frequency models based on predictive performance.
We propose a new zero-inflated Poisson boosted tree model, with variation in the assumption about the relationship between inflation probability $p$ and distribution mean $mu$.
arXiv Detail & Related papers (2023-07-15T10:54:46Z) - Model Transparency and Interpretability : Survey and Application to the
Insurance Industry [1.6058099298620423]
This paper introduces the importance of model tackles interpretation and the notion of model transparency.
Within an insurance context, it illustrates how some tools can be used to enforce the control of actuarial models.
arXiv Detail & Related papers (2022-09-01T16:12:54Z) - ER: Equivariance Regularizer for Knowledge Graph Completion [107.51609402963072]
We propose a new regularizer, namely, Equivariance Regularizer (ER)
ER can enhance the generalization ability of the model by employing the semantic equivariance between the head and tail entities.
The experimental results indicate a clear and substantial improvement over the state-of-the-art relation prediction methods.
arXiv Detail & Related papers (2022-06-24T08:18:05Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - When stakes are high: balancing accuracy and transparency with
Model-Agnostic Interpretable Data-driven suRRogates [0.0]
Highly regulated industries, like banking and insurance, ask for transparent decision-making algorithms.
We present a procedure to develop a Model-Agnostic Interpretable Data-driven suRRogate (maidrr)
Knowledge is extracted from a black box via partial dependence effects.
This results in a segmentation of the feature space with automatic variable selection.
arXiv Detail & Related papers (2020-07-14T08:10:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.