BayesLDM: A Domain-Specific Language for Probabilistic Modeling of
Longitudinal Data
- URL: http://arxiv.org/abs/2209.05581v1
- Date: Mon, 12 Sep 2022 20:10:02 GMT
- Title: BayesLDM: A Domain-Specific Language for Probabilistic Modeling of
Longitudinal Data
- Authors: Karine Tung, Steven De La Torre, Mohamed El Mistiri, Rebecca Braga De
Braganca, Eric Hekler, Misha Pavel, Daniel Rivera, Pedja Klasnja, Donna
Spruijt-Metz, Benjamin M. Marlin
- Abstract summary: BayesLDM is a system for modeling complex longitudinal data.
BayesLDM combines a model specification with inspection of available data and outputs code for performing Bayesian inference for unknown parameters.
We describe the BayesLDM system components, evaluate the efficiency of representation and inference optimizations, and provide an illustrative example of the application of the system to analyzing heterogeneous and partially observed mobile health data.
- Score: 7.6623098641742615
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper we present BayesLDM, a system for Bayesian longitudinal data
modeling consisting of a high-level modeling language with specific features
for modeling complex multivariate time series data coupled with a compiler that
can produce optimized probabilistic program code for performing inference in
the specified model. BayesLDM supports modeling of Bayesian network models with
a specific focus on the efficient, declarative specification of dynamic
Bayesian Networks (DBNs). The BayesLDM compiler combines a model specification
with inspection of available data and outputs code for performing Bayesian
inference for unknown model parameters while simultaneously handling missing
data. These capabilities have the potential to significantly accelerate
iterative modeling workflows in domains that involve the analysis of complex
longitudinal data by abstracting away the process of producing computationally
efficient probabilistic inference code. We describe the BayesLDM system
components, evaluate the efficiency of representation and inference
optimizations and provide an illustrative example of the application of the
system to analyzing heterogeneous and partially observed mobile health data.
Related papers
- D-Models and E-Models: Diversity-Stability Trade-offs in the Sampling Behavior of Large Language Models [91.21455683212224]
In large language models (LLMs), the probability of relevance for the next piece of information is linked to the probability of relevance for the next product.<n>But whether fine-grained sampling probabilities faithfully align with task requirements remains an open question.<n>We identify two model types: D-models, whose P_token exhibits large step-to-step variability and poor alignment with P_task; and E-models, whose P_token is more stable and better aligned with P_task.
arXiv Detail & Related papers (2026-01-25T14:59:09Z) - Meta-probabilistic Modeling [36.339664748241944]
We propose meta-probabilistic modeling (MPM), a meta-learning algorithm that learns generative model structure directly from multiple related datasets.<n>For learning and inference, we propose a tractable VAE-inspired surrogate objective, and optimize it through bi-level optimization.<n>We evaluate MPM on object-centric image modeling and sequential text modeling, demonstrating that it adapts generative models to data while recovering meaningful latent representations.
arXiv Detail & Related papers (2026-01-08T00:34:06Z) - Fast Riemannian-manifold Hamiltonian Monte Carlo for hierarchical Gaussian-process models [0.0]
We show that, compared with the slow inference achieved with existing program libraries, the performance can be drastically improved.<n>We demonstrate that RMHMC effectively samples from the posterior, allowing the calculation of model evidence.<n>We highlight the need to develop a customisable library set that allows users to incorporate dynamically programmed objects.
arXiv Detail & Related papers (2025-11-09T14:44:13Z) - Every Step Counts: Decoding Trajectories as Authorship Fingerprints of dLLMs [63.82840470917859]
We show that the decoding mechanism of dLLMs can be used as a powerful tool for model attribution.<n>We propose a novel information extraction scheme called the Directed Decoding Map (DDM), which captures structural relationships between decoding steps and better reveals model-specific behaviors.
arXiv Detail & Related papers (2025-10-02T06:25:10Z) - Weighted Support Points from Random Measures: An Interpretable Alternative for Generative Modeling [0.0]
We introduce a generative modeling framework based on random weighted support points.<n>The proposed method generates diverse and interpretable sample sets from a fixed dataset.<n>We show that our approach produces high-quality and diverse outputs at a fraction of the computational cost of black-box alternatives.
arXiv Detail & Related papers (2025-08-28T22:58:39Z) - Multi-fidelity Parameter Estimation Using Conditional Diffusion Models [6.934199382834925]
We present a multi-fidelity method for uncertainty quantification of parameter estimates in complex systems.
We use conditional generative models trained to sample the target conditional distribution.
We demonstrate the effectiveness of the proposed method on several numerical examples.
arXiv Detail & Related papers (2025-04-02T16:54:47Z) - Latent Thought Models with Variational Bayes Inference-Time Computation [52.63299874322121]
Latent Thought Models (LTMs) incorporate explicit latent thought vectors that follow an explicit prior model in latent space.<n>LTMs demonstrate superior sample and parameter efficiency compared to autoregressive models and discrete diffusion models.
arXiv Detail & Related papers (2025-02-03T17:50:34Z) - Multi-modal Auto-regressive Modeling via Visual Words [96.25078866446053]
We propose the concept of visual tokens, which maps the visual features to probability distributions over Large Multi-modal Models' vocabulary.
We further explore the distribution of visual features in the semantic space within LMM and the possibility of using text embeddings to represent visual information.
arXiv Detail & Related papers (2024-03-12T14:58:52Z) - Latent Semantic Consensus For Deterministic Geometric Model Fitting [109.44565542031384]
We propose an effective method called Latent Semantic Consensus (LSC)
LSC formulates the model fitting problem into two latent semantic spaces based on data points and model hypotheses.
LSC is able to provide consistent and reliable solutions within only a few milliseconds for general multi-structural model fitting.
arXiv Detail & Related papers (2024-03-11T05:35:38Z) - Sample Complexity Characterization for Linear Contextual MDPs [67.79455646673762]
Contextual decision processes (CMDPs) describe a class of reinforcement learning problems in which the transition kernels and reward functions can change over time with different MDPs indexed by a context variable.
CMDPs serve as an important framework to model many real-world applications with time-varying environments.
We study CMDPs under two linear function approximation models: Model I with context-varying representations and common linear weights for all contexts; and Model II with common representations for all contexts and context-varying linear weights.
arXiv Detail & Related papers (2024-02-05T03:25:04Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - T1: Scaling Diffusion Probabilistic Fields to High-Resolution on Unified
Visual Modalities [69.16656086708291]
Diffusion Probabilistic Field (DPF) models the distribution of continuous functions defined over metric spaces.
We propose a new model comprising of a view-wise sampling algorithm to focus on local structure learning.
The model can be scaled to generate high-resolution data while unifying multiple modalities.
arXiv Detail & Related papers (2023-05-24T03:32:03Z) - BayesFlow can reliably detect Model Misspecification and Posterior
Errors in Amortized Bayesian Inference [0.0]
We conceptualize the types of model misspecification arising in simulation-based inference and systematically investigate the performance of the BayesFlow framework under these misspecifications.
We propose an augmented optimization objective which imposes a probabilistic structure on the latent data space and utilize maximum mean discrepancy (MMD) to detect potentially catastrophic misspecifications.
arXiv Detail & Related papers (2021-12-16T13:25:27Z) - Relational VAE: A Continuous Latent Variable Model for Graph Structured
Data [0.0]
We show applications on the problem of structured probability density modeling for simulated and real wind farm monitoring data.
We release the source code, along with the simulated datasets.
arXiv Detail & Related papers (2021-06-30T13:24:27Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - VAULT: VAriable Unified Long Text Representation for Machine Reading
Comprehension [31.639069657951747]
Existing models on Machine Reading require complex model architecture for modeling long texts with paragraph representation and classification.
We propose VAULT: a light-weight and parallel-efficient paragraph representation for MRC based on contextualized representation from long document input.
arXiv Detail & Related papers (2021-05-07T13:03:43Z) - Tensor Networks for Probabilistic Sequence Modeling [7.846449972735859]
We use a uniform matrix product state (u-MPS) model for probabilistic modeling of sequence data.
We then introduce a novel generative algorithm giving trained u-MPS the ability to efficiently sample from a wide variety of conditional distributions.
Experiments on sequence modeling with synthetic and real text data show u-MPS outperforming a variety of baselines.
arXiv Detail & Related papers (2020-03-02T17:16:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.