Machine Learning with High-Cardinality Categorical Features in Actuarial
Applications
- URL: http://arxiv.org/abs/2301.12710v1
- Date: Mon, 30 Jan 2023 07:35:18 GMT
- Title: Machine Learning with High-Cardinality Categorical Features in Actuarial
Applications
- Authors: Benjamin Avanzi and Greg Taylor and Melantha Wang and Bernard Wong
- Abstract summary: High-cardinality categorical features are pervasive in actuarial data.
Standard categorical encoding methods like one-hot encoding are inadequate in these settings.
We present a novel _Generalised Linear Mixed Model Neural Network_ ("GLMMNet") approach to the modelling of high-cardinality categorical features.
- Score: 0.31133049660590606
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High-cardinality categorical features are pervasive in actuarial data (e.g.
occupation in commercial property insurance). Standard categorical encoding
methods like one-hot encoding are inadequate in these settings.
In this work, we present a novel _Generalised Linear Mixed Model Neural
Network_ ("GLMMNet") approach to the modelling of high-cardinality categorical
features. The GLMMNet integrates a generalised linear mixed model in a deep
learning framework, offering the predictive power of neural networks and the
transparency of random effects estimates, the latter of which cannot be
obtained from the entity embedding models. Further, its flexibility to deal
with any distribution in the exponential dispersion (ED) family makes it widely
applicable to many actuarial contexts and beyond.
We illustrate and compare the GLMMNet against existing approaches in a range
of simulation experiments as well as in a real-life insurance case study.
Notably, we find that the GLMMNet often outperforms or at least performs
comparably with an entity embedded neural network, while providing the
additional benefit of transparency, which is particularly valuable in practical
applications.
Importantly, while our model was motivated by actuarial applications, it can
have wider applicability. The GLMMNet would suit any applications that involve
high-cardinality categorical variables and where the response cannot be
sufficiently modelled by a Gaussian distribution.
Related papers
- A PAC-Bayesian Perspective on the Interpolating Information Criterion [54.548058449535155]
We show how a PAC-Bayes bound is obtained for a general class of models, characterizing factors which influence performance in the interpolating regime.
We quantify how the test error for overparameterized models achieving effectively zero training error depends on the quality of the implicit regularization imposed by e.g. the combination of model, parameter-initialization scheme.
arXiv Detail & Related papers (2023-11-13T01:48:08Z) - Smoothness and monotonicity constraints for neural networks using ICEnet [0.0]
We present a novel method for enforcing constraints within deep neural network models.
We show how these models can be trained and provide example applications using real-world datasets.
arXiv Detail & Related papers (2023-05-15T17:14:52Z) - MISNN: Multiple Imputation via Semi-parametric Neural Networks [9.594714330925703]
Multiple imputation (MI) has been widely applied to missing value problems in biomedical, social and econometric research.
We propose MISNN, a novel and efficient algorithm that incorporates feature selection for MI.
arXiv Detail & Related papers (2023-05-02T21:45:36Z) - Bayesian community detection for networks with covariates [16.230648949593153]
"Community detection" has arguably received the most attention in the scientific community.
We propose a block model with a co-dependent random partition prior.
Our model has the ability to learn the number of the communities via posterior inference without having to assume it to be known.
arXiv Detail & Related papers (2022-03-04T01:58:35Z) - Robustness Certificates for Implicit Neural Networks: A Mixed Monotone
Contractive Approach [60.67748036747221]
Implicit neural networks offer competitive performance and reduced memory consumption.
They can remain brittle with respect to input adversarial perturbations.
This paper proposes a theoretical and computational framework for robustness verification of implicit neural networks.
arXiv Detail & Related papers (2021-12-10T03:08:55Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Efficient Construction of Nonlinear Models over Normalized Data [21.531781003420573]
We show how it is possible to decompose in a systematic way both for binary joins and for multi-way joins to construct mixture models.
We present algorithms that can conduct the training of the network in a factorized way and offer performance advantages.
arXiv Detail & Related papers (2020-11-23T19:20:03Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.