Modern Multiple Imputation with Functional Data
- URL: http://arxiv.org/abs/2011.12509v1
- Date: Wed, 25 Nov 2020 04:22:30 GMT
- Title: Modern Multiple Imputation with Functional Data
- Authors: Aniruddha Rajendra Rao, Matthew Reimherr
- Abstract summary: This work considers the problem of fitting functional models with sparsely and irregularly sampled functional data.
It overcomes the limitations of the state-of-the-art methods, which face major challenges in the fitting of more complex non-linear models.
- Score: 6.624726878647541
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This work considers the problem of fitting functional models with sparsely
and irregularly sampled functional data. It overcomes the limitations of the
state-of-the-art methods, which face major challenges in the fitting of more
complex non-linear models. Currently, many of these models cannot be
consistently estimated unless the number of observed points per curve grows
sufficiently quickly with the sample size, whereas, we show numerically that a
modified approach with more modern multiple imputation methods can produce
better estimates in general. We also propose a new imputation approach that
combines the ideas of {\it MissForest} with {\it Local Linear Forest} and
compare their performance with {\it PACE} and several other multivariate
multiple imputation methods. This work is motivated by a longitudinal study on
smoking cessation, in which the Electronic Health Records (EHR) from Penn State
PaTH to Health allow for the collection of a great deal of data, with highly
variable sampling. To illustrate our approach, we explore the relation between
relapse and diastolic blood pressure. We also consider a variety of simulation
schemes with varying levels of sparsity to validate our methods.
Related papers
- Towards Learning Stochastic Population Models by Gradient Descent [0.0]
We show that simultaneous estimation of parameters and structure poses major challenges for optimization procedures.
We demonstrate accurate estimation of models but find that enforcing the inference of parsimonious, interpretable models drastically increases the difficulty.
arXiv Detail & Related papers (2024-04-10T14:38:58Z) - Latent variable model for high-dimensional point process with structured missingness [4.451479907610764]
Longitudinal data are important in numerous fields, such as healthcare, sociology and seismology.
Real-world datasets can be high-dimensional, contain structured missingness patterns, and measurement time points can be governed by an unknown process.
We propose a flexible and efficient latent-variable model that is capable of addressing all these limitations.
arXiv Detail & Related papers (2024-02-08T15:41:48Z) - Handling Overlapping Asymmetric Datasets -- A Twice Penalized P-Spline
Approach [0.40964539027092917]
This research aims to develop a new method which can model the smaller cohort against a particular response.
We find our twice penalized approach offers an enhanced fit over a linear B-Spline and once penalized P-Spline approximation.
Applying to a real-life dataset relating to a person's risk of developing Non-Alcoholic Steatohepatitis, we see an improved model fit performance of over 65%.
arXiv Detail & Related papers (2023-11-17T12:41:07Z) - A model-free feature selection technique of feature screening and random
forest based recursive feature elimination [0.0]
We propose a model-free feature selection method for ultra-high dimensional data with mass features.
We show that the proposed method is selection consistent and $L$ consistent under weak regularity conditions.
arXiv Detail & Related papers (2023-02-15T03:39:16Z) - Continuous-Time Modeling of Counterfactual Outcomes Using Neural
Controlled Differential Equations [84.42837346400151]
Estimating counterfactual outcomes over time has the potential to unlock personalized healthcare.
Existing causal inference approaches consider regular, discrete-time intervals between observations and treatment decisions.
We propose a controllable simulation environment based on a model of tumor growth for a range of scenarios.
arXiv Detail & Related papers (2022-06-16T17:15:15Z) - Multi-scale Attention Flow for Probabilistic Time Series Forecasting [68.20798558048678]
We propose a novel non-autoregressive deep learning model, called Multi-scale Attention Normalizing Flow(MANF)
Our model avoids the influence of cumulative error and does not increase the time complexity.
Our model achieves state-of-the-art performance on many popular multivariate datasets.
arXiv Detail & Related papers (2022-05-16T07:53:42Z) - Deep Bayesian Estimation for Dynamic Treatment Regimes with a Long
Follow-up Time [28.11470886127216]
Causal effect estimation for dynamic treatment regimes (DTRs) contributes to sequential decision making.
We combine outcome regression models with treatment models for high dimensional features using uncensored subjects that are small in sample size.
Also, the developed deep Bayesian models can model uncertainty and output the prediction variance which is essential for the safety-aware applications, such as self-driving cars and medical treatment design.
arXiv Detail & Related papers (2021-09-20T13:21:39Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - A Twin Neural Model for Uplift [59.38563723706796]
Uplift is a particular case of conditional treatment effect modeling.
We propose a new loss function defined by leveraging a connection with the Bayesian interpretation of the relative risk.
We show our proposed method is competitive with the state-of-the-art in simulation setting and on real data from large scale randomized experiments.
arXiv Detail & Related papers (2021-05-11T16:02:39Z) - Improving the Reconstruction of Disentangled Representation Learners via Multi-Stage Modeling [54.94763543386523]
Current autoencoder-based disentangled representation learning methods achieve disentanglement by penalizing the ( aggregate) posterior to encourage statistical independence of the latent factors.
We present a novel multi-stage modeling approach where the disentangled factors are first learned using a penalty-based disentangled representation learning method.
Then, the low-quality reconstruction is improved with another deep generative model that is trained to model the missing correlated latent variables.
arXiv Detail & Related papers (2020-10-25T18:51:15Z) - Evaluating the Disentanglement of Deep Generative Models through
Manifold Topology [66.06153115971732]
We present a method for quantifying disentanglement that only uses the generative model.
We empirically evaluate several state-of-the-art models across multiple datasets.
arXiv Detail & Related papers (2020-06-05T20:54:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.