Semi-Modular Inference: enhanced learning in multi-modular models by
tempering the influence of components
- URL: http://arxiv.org/abs/2003.06804v1
- Date: Sun, 15 Mar 2020 11:55:55 GMT
- Title: Semi-Modular Inference: enhanced learning in multi-modular models by
tempering the influence of components
- Authors: Chris U. Carmona and Geoff K. Nicholls
- Abstract summary: We show existing Modular/Cut-model inference is coherent, and write down a new family of Semi-Modular Inference schemes.
We give a meta-learning criterion and estimation procedure to choose the inference scheme.
We illustrate our methods on two standard test cases from the literature and a motivating archaeological data set.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bayesian statistical inference loses predictive optimality when generative
models are misspecified.
Working within an existing coherent loss-based generalisation of Bayesian
inference, we show existing Modular/Cut-model inference is coherent, and write
down a new family of Semi-Modular Inference (SMI) schemes, indexed by an
influence parameter, with Bayesian inference and Cut-models as special cases.
We give a meta-learning criterion and estimation procedure to choose the
inference scheme. This returns Bayesian inference when there is no
misspecification.
The framework applies naturally to Multi-modular models. Cut-model inference
allows directed information flow from well-specified modules to misspecified
modules, but not vice versa. An existing alternative power posterior method
gives tunable but undirected control of information flow, improving prediction
in some settings. In contrast, SMI allows tunable and directed information flow
between modules.
We illustrate our methods on two standard test cases from the literature and
a motivating archaeological data set.
Related papers
- Influence Functions for Scalable Data Attribution in Diffusion Models [52.92223039302037]
Diffusion models have led to significant advancements in generative modelling.
Yet their widespread adoption poses challenges regarding data attribution and interpretability.
In this paper, we aim to help address such challenges by developing an textitinfluence functions framework.
arXiv Detail & Related papers (2024-10-17T17:59:02Z) - Federated Bayesian Deep Learning: The Application of Statistical Aggregation Methods to Bayesian Models [0.9940108090221528]
Aggregation strategies have been developed to pool or fuse the weights and biases of distributed deterministic models.
We show that simple application of the aggregation methods associated with FL schemes for deterministic models is either impossible or results in sub-optimal performance.
arXiv Detail & Related papers (2024-03-22T15:02:24Z) - Data-driven Nonlinear Model Reduction using Koopman Theory: Integrated
Control Form and NMPC Case Study [56.283944756315066]
We propose generic model structures combining delay-coordinate encoding of measurements and full-state decoding to integrate reduced Koopman modeling and state estimation.
A case study demonstrates that our approach provides accurate control models and enables real-time capable nonlinear model predictive control of a high-purity cryogenic distillation column.
arXiv Detail & Related papers (2024-01-09T11:54:54Z) - Diffusion models for probabilistic programming [56.47577824219207]
Diffusion Model Variational Inference (DMVI) is a novel method for automated approximate inference in probabilistic programming languages (PPLs)
DMVI is easy to implement, allows hassle-free inference in PPLs without the drawbacks of, e.g., variational inference using normalizing flows, and does not make any constraints on the underlying neural network model.
arXiv Detail & Related papers (2023-11-01T12:17:05Z) - Modular Deep Learning [120.36599591042908]
Transfer learning has recently become the dominant paradigm of machine learning.
It remains unclear how to develop models that specialise towards multiple tasks without incurring negative interference.
Modular deep learning has emerged as a promising solution to these challenges.
arXiv Detail & Related papers (2023-02-22T18:11:25Z) - Meta-Causal Feature Learning for Out-of-Distribution Generalization [71.38239243414091]
This paper presents a balanced meta-causal learner (BMCL), which includes a balanced task generation module (BTG) and a meta-causal feature learning module (MCFL)
BMCL effectively identifies the class-invariant visual regions for classification and may serve as a general framework to improve the performance of the state-of-the-art methods.
arXiv Detail & Related papers (2022-08-22T09:07:02Z) - Scalable Semi-Modular Inference with Variational Meta-Posteriors [0.0]
The Cut posterior and Semi-Modular Inference are Generalised Bayes methods for Modular Bayesian evidence combination.
We show that analysis of models with multiple cuts is feasible using a new Variational Meta-Posterior.
This approximates a family of SMI posteriors indexed by $eta$ using a single set of variational parameters.
arXiv Detail & Related papers (2022-04-01T09:01:20Z) - CARE: Coherent Actionable Recourse based on Sound Counterfactual
Explanations [0.0]
This paper introduces CARE, a modular explanation framework that addresses the model- and user-level desiderata.
As a model-agnostic approach, CARE generates multiple, diverse explanations for any black-box model.
arXiv Detail & Related papers (2021-08-18T15:26:59Z) - MINIMALIST: Mutual INformatIon Maximization for Amortized Likelihood
Inference from Sampled Trajectories [61.3299263929289]
Simulation-based inference enables learning the parameters of a model even when its likelihood cannot be computed in practice.
One class of methods uses data simulated with different parameters to infer an amortized estimator for the likelihood-to-evidence ratio.
We show that this approach can be formulated in terms of mutual information between model parameters and simulated data.
arXiv Detail & Related papers (2021-06-03T12:59:16Z) - Deep Conditional Transformation Models [0.0]
Learning the cumulative distribution function (CDF) of an outcome variable conditional on a set of features remains challenging.
Conditional transformation models provide a semi-parametric approach that allows to model a large class of conditional CDFs.
We propose a novel network architecture, provide details on different model definitions and derive suitable constraints.
arXiv Detail & Related papers (2020-10-15T16:25:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.