Calibrating the Lee-Carter and the Poisson Lee-Carter models via Neural
Networks
- URL: http://arxiv.org/abs/2106.12312v2
- Date: Thu, 24 Jun 2021 20:40:06 GMT
- Title: Calibrating the Lee-Carter and the Poisson Lee-Carter models via Neural
Networks
- Authors: Salvatore Scognamiglio
- Abstract summary: This paper introduces a neural network approach for fitting the Lee-Carter and the Poisson Lee-Carter model on multiple populations.
We develop some neural networks that replicate the structure of the individual LC models and allow their joint fitting by analysing the mortality data of all the considered populations simultaneously.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper introduces a neural network approach for fitting the Lee-Carter
and the Poisson Lee-Carter model on multiple populations. We develop some
neural networks that replicate the structure of the individual LC models and
allow their joint fitting by analysing the mortality data of all the considered
populations simultaneously. The neural network architecture is specifically
designed to calibrate each individual model using all available information
instead of using a population-specific subset of data as in the traditional
estimation schemes. A large set of numerical experiments performed on all the
countries of the Human Mortality Database (HMD) shows the effectiveness of our
approach. In particular, the resulting parameter estimates appear smooth and
less sensitive to the random fluctuations often present in the mortality rates'
data, especially for low-population countries. In addition, the forecasting
performance results significantly improved as well.
Related papers
- Modeling COVID-19 spread in the USA using metapopulation SIR models coupled with graph convolutional neural networks [0.0]
Graph convolutional neural networks (GCNs) have shown tremendous promise in addressing data-intensive challenges.
Some attempts have been made to improve predictions of Susceptible-Infected-Recovered (SIR) models by incorporating human mobility between metapopulations.
Recently, researchers have found that a hybrid GCN-SIR approach outperformed existing methodologies when used on the data collected on a precinct level in Japan.
arXiv Detail & Related papers (2025-01-03T18:06:26Z) - Screening for Diabetes Mellitus in the U.S. Population Using Neural Network Models and Complex Survey Designs [2.3849116823891987]
This paper proposes a general predictive framework for regression and classification using neural network (NN) modeling.
It applies this framework to develop robust risk score models for assessing the risk of Diabetes Mellitus in the US population.
Although the focus is on diabetes, this NN predictive framework is adaptable for developing clinical models across a diverse range of diseases and medical cohorts.
arXiv Detail & Related papers (2024-03-28T18:06:11Z) - Towards Theoretical Understandings of Self-Consuming Generative Models [56.84592466204185]
This paper tackles the emerging challenge of training generative models within a self-consuming loop.
We construct a theoretical framework to rigorously evaluate how this training procedure impacts the data distributions learned by future models.
We present results for kernel density estimation, delivering nuanced insights such as the impact of mixed data training on error propagation.
arXiv Detail & Related papers (2024-02-19T02:08:09Z) - Continuous time recurrent neural networks: overview and application to
forecasting blood glucose in the intensive care unit [56.801856519460465]
Continuous time autoregressive recurrent neural networks (CTRNNs) are a deep learning model that account for irregular observations.
We demonstrate the application of these models to probabilistic forecasting of blood glucose in a critical care setting.
arXiv Detail & Related papers (2023-04-14T09:39:06Z) - Clinical Deterioration Prediction in Brazilian Hospitals Based on
Artificial Neural Networks and Tree Decision Models [56.93322937189087]
An extremely boosted neural network (XBNet) is used to predict clinical deterioration (CD)
The XGBoost model obtained the best results in predicting CD among Brazilian hospitals' data.
arXiv Detail & Related papers (2022-12-17T23:29:14Z) - On the Generalization and Adaption Performance of Causal Models [99.64022680811281]
Differentiable causal discovery has proposed to factorize the data generating process into a set of modules.
We study the generalization and adaption performance of such modular neural causal models.
Our analysis shows that the modular neural causal models outperform other models on both zero and few-shot adaptation in low data regimes.
arXiv Detail & Related papers (2022-06-09T17:12:32Z) - Differentiable Reasoning over Long Stories -- Assessing Systematic
Generalisation in Neural Models [12.479512369785082]
We consider two classes of neural models: "E-GNN", the graph-based models that can process graph-structured data and consider the edge attributes simultaneously; and "L-Graph", the sequence-based models which can process linearized version of the graphs.
We found that the modified recurrent neural network yield surprisingly accurate results across every systematic generalisation tasks which outperform the graph neural network.
arXiv Detail & Related papers (2022-03-20T18:34:42Z) - Parameter estimation for WMTI-Watson model of white matter using
encoder-decoder recurrent neural network [0.0]
In this study, we evaluate the performance of NLLS, the RNN-based method and a multilayer perceptron (MLP) on datasets rat and human brain.
We showed that the proposed RNN-based fitting approach had the advantage of highly reduced computation time over NLLS.
arXiv Detail & Related papers (2022-03-01T16:33:15Z) - Adversarially-regularized mixed effects deep learning (ARMED) models for
improved interpretability, performance, and generalization on clustered data [0.974672460306765]
Mixed effects models separate cluster-invariant, population-level fixed effects from cluster-specific random effects.
We propose a general-purpose framework for building Adversarially-Regularized Mixed Effects Deep learning (ARMED) models through 3 non-intrusive additions to existing networks.
We apply this framework to dense feedforward neural networks (DFNNs), convolutional neural networks, and autoencoders on 4 applications including simulations, dementia prognosis and diagnosis, and cell microscopy.
arXiv Detail & Related papers (2022-02-23T20:58:22Z) - Post-mortem on a deep learning contest: a Simpson's paradox and the
complementary roles of scale metrics versus shape metrics [61.49826776409194]
We analyze a corpus of models made publicly-available for a contest to predict the generalization accuracy of neural network (NN) models.
We identify what amounts to a Simpson's paradox: where "scale" metrics perform well overall but perform poorly on sub partitions of the data.
We present two novel shape metrics, one data-independent, and the other data-dependent, which can predict trends in the test accuracy of a series of NNs.
arXiv Detail & Related papers (2021-06-01T19:19:49Z) - LocalDrop: A Hybrid Regularization for Deep Neural Networks [98.30782118441158]
We propose a new approach for the regularization of neural networks by the local Rademacher complexity called LocalDrop.
A new regularization function for both fully-connected networks (FCNs) and convolutional neural networks (CNNs) has been developed based on the proposed upper bound of the local Rademacher complexity.
arXiv Detail & Related papers (2021-03-01T03:10:11Z) - Multi-Sample Online Learning for Spiking Neural Networks based on
Generalized Expectation Maximization [42.125394498649015]
Spiking Neural Networks (SNNs) capture some of the efficiency of biological brains by processing through binary neural dynamic activations.
This paper proposes to leverage multiple compartments that sample independent spiking signals while sharing synaptic weights.
The key idea is to use these signals to obtain more accurate statistical estimates of the log-likelihood training criterion, as well as of its gradient.
arXiv Detail & Related papers (2021-02-05T16:39:42Z) - Ensembles of Spiking Neural Networks [0.3007949058551534]
This paper demonstrates how to construct ensembles of spiking neural networks producing state-of-the-art results.
We achieve classification accuracies of 98.71%, 100.0%, and 99.09%, on the MNIST, NMNIST and DVS Gesture datasets respectively.
We formalize spiking neural networks as GLM predictors, identifying a suitable representation for their target domain.
arXiv Detail & Related papers (2020-10-15T17:45:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.