Bayesian CART models for insurance claims frequency
- URL: http://arxiv.org/abs/2303.01923v3
- Date: Fri, 1 Dec 2023 16:37:36 GMT
- Title: Bayesian CART models for insurance claims frequency
- Authors: Yaojun Zhang, Lanpeng Ji, Georgios Aivaliotis, and Charles Taylor
- Abstract summary: classification and regression trees (CARTs) and their ensembles have gained popularity in the actuarial literature.
We introduce Bayesian CART models for insurance pricing, with a particular focus on claims frequency modelling.
Some simulations and real insurance data will be discussed to illustrate the applicability of these models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Accuracy and interpretability of a (non-life) insurance pricing model are
essential qualities to ensure fair and transparent premiums for policy-holders,
that reflect their risk. In recent years, the classification and regression
trees (CARTs) and their ensembles have gained popularity in the actuarial
literature, since they offer good prediction performance and are relatively
easily interpretable. In this paper, we introduce Bayesian CART models for
insurance pricing, with a particular focus on claims frequency modelling.
Additionally to the common Poisson and negative binomial (NB) distributions
used for claims frequency, we implement Bayesian CART for the zero-inflated
Poisson (ZIP) distribution to address the difficulty arising from the
imbalanced insurance claims data. To this end, we introduce a general MCMC
algorithm using data augmentation methods for posterior tree exploration. We
also introduce the deviance information criterion (DIC) for the tree model
selection. The proposed models are able to identify trees which can better
classify the policy-holders into risk groups. Some simulations and real
insurance data will be discussed to illustrate the applicability of these
models.
Related papers
- Combining Structural and Unstructured Data: A Topic-based Finite Mixture Model for Insurance Claim Prediction [0.0]
This paper introduces a novel approach by developing a joint mixture model that integrates both claim descriptions and claim amounts.
Our method establishes a probabilistic link between textual descriptions and loss amounts, enhancing the accuracy of claims clustering and prediction.
arXiv Detail & Related papers (2024-10-07T01:37:07Z) - Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Mitigating Nonlinear Algorithmic Bias in Binary Classification [0.0]
This paper proposes the use of causal modeling to detect and mitigate bias that is nonlinear in the protected attribute.
We show that the probability of getting correctly classified as "low risk" is lowest among young people.
Based on the fitted causal model, the debiased- probability estimates are computed, showing improved fairness with little impact on overall accuracy.
arXiv Detail & Related papers (2023-12-09T01:26:22Z) - Enhanced Gradient Boosting for Zero-Inflated Insurance Claims and Comparative Analysis of CatBoost, XGBoost, and LightGBM [0.0]
CatBoost is the best library for developing auto claim frequency models based on predictive performance.
We propose a new zero-inflated Poisson boosted tree model, with variation in the assumption about the relationship between inflation probability $p$ and distribution mean $mu$.
arXiv Detail & Related papers (2023-07-15T10:54:46Z) - Rethinking Missing Data: Aleatoric Uncertainty-Aware Recommendation [59.500347564280204]
We propose a new Aleatoric Uncertainty-aware Recommendation (AUR) framework.
AUR consists of a new uncertainty estimator along with a normal recommender model.
As the chance of mislabeling reflects the potential of a pair, AUR makes recommendations according to the uncertainty.
arXiv Detail & Related papers (2022-09-22T04:32:51Z) - Model Transparency and Interpretability : Survey and Application to the
Insurance Industry [1.6058099298620423]
This paper introduces the importance of model tackles interpretation and the notion of model transparency.
Within an insurance context, it illustrates how some tools can be used to enforce the control of actuarial models.
arXiv Detail & Related papers (2022-09-01T16:12:54Z) - A Fair Pricing Model via Adversarial Learning [3.983383967538961]
At the core of insurance business lies classification between risky and non-risky insureds.
The distinction between a fair actuarial classification and "discrimination" is subtle.
We show that debiasing the predictor alone may be insufficient to maintain adequate accuracy.
arXiv Detail & Related papers (2022-02-24T10:42:20Z) - Self-Certifying Classification by Linearized Deep Assignment [65.0100925582087]
We propose a novel class of deep predictors for classifying metric data on graphs within PAC-Bayes risk certification paradigm.
Building on the recent PAC-Bayes literature and data-dependent priors, this approach enables learning posterior distributions on the hypothesis space.
arXiv Detail & Related papers (2022-01-26T19:59:14Z) - Risk Minimization from Adaptively Collected Data: Guarantees for
Supervised and Policy Learning [57.88785630755165]
Empirical risk minimization (ERM) is the workhorse of machine learning, but its model-agnostic guarantees can fail when we use adaptively collected data.
We study a generic importance sampling weighted ERM algorithm for using adaptively collected data to minimize the average of a loss function over a hypothesis class.
For policy learning, we provide rate-optimal regret guarantees that close an open gap in the existing literature whenever exploration decays to zero.
arXiv Detail & Related papers (2021-06-03T09:50:13Z) - Post-Contextual-Bandit Inference [57.88785630755165]
Contextual bandit algorithms are increasingly replacing non-adaptive A/B tests in e-commerce, healthcare, and policymaking.
They can both improve outcomes for study participants and increase the chance of identifying good or even best policies.
To support credible inference on novel interventions at the end of the study, we still want to construct valid confidence intervals on average treatment effects, subgroup effects, or value of new policies.
arXiv Detail & Related papers (2021-06-01T12:01:51Z) - Characterizing Fairness Over the Set of Good Models Under Selective
Labels [69.64662540443162]
We develop a framework for characterizing predictive fairness properties over the set of models that deliver similar overall performance.
We provide tractable algorithms to compute the range of attainable group-level predictive disparities.
We extend our framework to address the empirically relevant challenge of selectively labelled data.
arXiv Detail & Related papers (2021-01-02T02:11:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.