Multi-Model Probabilistic Programming
- URL: http://arxiv.org/abs/2208.06329v1
- Date: Fri, 12 Aug 2022 15:38:15 GMT
- Title: Multi-Model Probabilistic Programming
- Authors: Ryan Bernstein
- Abstract summary: We present an extension of probabilistic programming that lets each program represent a network of interrelated probabilistic models.
We give a formal semantics for these multi-model probabilistic programs, a collection of efficient algorithms for network-of-model operations, and an example implementation built on top of the popular probabilistic programming language Stan.
This network-of-models representation opens many doors, including search and automation in model-space, tracking and communication of model development, and explicit modeler degrees of freedom to mitigate issues like p-hacking.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Probabilistic programming makes it easy to represent a probabilistic model as
a program. Building an individual model, however, is only one step of
probabilistic modeling. The broader challenge of probabilistic modeling is in
understanding and navigating spaces of alternative models. There is currently
no good way to represent these spaces of alternative models, despite their
central role. We present an extension of probabilistic programming that lets
each program represent a network of interrelated probabilistic models. We give
a formal semantics for these multi-model probabilistic programs, a collection
of efficient algorithms for network-of-model operations, and an example
implementation built on top of the popular probabilistic programming language
Stan. This network-of-models representation opens many doors, including search
and automation in model-space, tracking and communication of model development,
and explicit modeler degrees of freedom to mitigate issues like p-hacking. We
demonstrate automatic model search and model development tracking using our
Stan implementation, and we propose many more possible applications.
Related papers
- Automated Statistical Model Discovery with Language Models [34.03743547761152]
We introduce a method for language model driven automated statistical model discovery.
We cast our automated procedure within the principled framework of Box's Loop.
Our results highlight the promise of LM-driven model discovery.
arXiv Detail & Related papers (2024-02-27T20:33:22Z) - Language Model Cascades [72.18809575261498]
Repeated interactions at test-time with a single model, or the composition of multiple models together, further expands capabilities.
Cases with control flow and dynamic structure require techniques from probabilistic programming.
We formalize several existing techniques from this perspective, including scratchpads / chain of thought, verifiers, STaR, selection-inference, and tool use.
arXiv Detail & Related papers (2022-07-21T07:35:18Z) - Stochastic Parameterizations: Better Modelling of Temporal Correlations
using Probabilistic Machine Learning [1.5293427903448025]
We show that by using a physically-informed recurrent neural network within a probabilistic framework, our model for the 96 atmospheric simulation is competitive.
This is due to a superior ability to model temporal correlations compared to standard first-order autoregressive schemes.
We evaluate across a number of metrics from the literature, but also discuss how the probabilistic metric of likelihood may be a unifying choice for future climate models.
arXiv Detail & Related papers (2022-03-28T14:51:42Z) - Model Reprogramming: Resource-Efficient Cross-Domain Machine Learning [65.268245109828]
In data-rich domains such as vision, language, and speech, deep learning prevails to deliver high-performance task-specific models.
Deep learning in resource-limited domains still faces multiple challenges including (i) limited data, (ii) constrained model development cost, and (iii) lack of adequate pre-trained models for effective finetuning.
Model reprogramming enables resource-efficient cross-domain machine learning by repurposing a well-developed pre-trained model from a source domain to solve tasks in a target domain without model finetuning.
arXiv Detail & Related papers (2022-02-22T02:33:54Z) - Low-Rank Constraints for Fast Inference in Structured Models [110.38427965904266]
This work demonstrates a simple approach to reduce the computational and memory complexity of a large class of structured models.
Experiments with neural parameterized structured models for language modeling, polyphonic music modeling, unsupervised grammar induction, and video modeling show that our approach matches the accuracy of standard models at large state spaces.
arXiv Detail & Related papers (2022-01-08T00:47:50Z) - Sampling from Arbitrary Functions via PSD Models [55.41644538483948]
We take a two-step approach by first modeling the probability distribution and then sampling from that model.
We show that these models can approximate a large class of densities concisely using few evaluations, and present a simple algorithm to effectively sample from these models.
arXiv Detail & Related papers (2021-10-20T12:25:22Z) - flip-hoisting: Exploiting Repeated Parameters in Discrete Probabilistic
Programs [25.320181572646135]
We present a program analysis and associated optimization, flip-hoisting, that collapses repetitious parameters in discrete probabilistic programs to improve inference performance.
We implement flip-hoisting in an existing probabilistic programming language and show empirically that it significantly improves inference performance.
arXiv Detail & Related papers (2021-10-19T22:04:26Z) - Probabilistic Modeling for Human Mesh Recovery [73.11532990173441]
This paper focuses on the problem of 3D human reconstruction from 2D evidence.
We recast the problem as learning a mapping from the input to a distribution of plausible 3D poses.
arXiv Detail & Related papers (2021-08-26T17:55:11Z) - How to Design Sample and Computationally Efficient VQA Models [53.65668097847456]
We find that representing the text as probabilistic programs and images as object-level scene graphs best satisfy these desiderata.
We extend existing models to leverage these soft programs and scene graphs to train on question answer pairs in an end-to-end manner.
arXiv Detail & Related papers (2021-03-22T01:48:16Z) - Transforming Probabilistic Programs for Model Checking [0.0]
We apply static analysis to probabilistic programs to automate large parts of two crucial model checking methods.
Our method transforms a probabilistic program specifying a density function into an efficient forward-sampling form.
We present an implementation targeting the popular Stan probabilistic programming language.
arXiv Detail & Related papers (2020-08-21T21:06:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.