Entropy-based Characterization of Modeling Constraints
- URL: http://arxiv.org/abs/2206.14105v1
- Date: Mon, 27 Jun 2022 17:25:49 GMT
- Title: Entropy-based Characterization of Modeling Constraints
- Authors: Orestis Loukas, Ho Ryun Chung
- Abstract summary: In most data-scientific approaches, the principle of Entropy (MaxEnt) is used to justify some parametric model.
We derive the distribution over all viable distributions that satisfy the provided set of constraints.
The appropriate parametric model which is supported by the data can be always deduced at the end of model selection.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In most data-scientific approaches, the principle of Maximum Entropy (MaxEnt)
is used to a posteriori justify some parametric model which has been already
chosen based on experience, prior knowledge or computational simplicity. In a
perpendicular formulation to conventional model building, we start from the
linear system of phenomenological constraints and asymptotically derive the
distribution over all viable distributions that satisfy the provided set of
constraints. The MaxEnt distribution plays a special role, as it is the most
typical among all phenomenologically viable distributions representing a good
expansion point for large-N techniques. This enables us to consistently
formulate hypothesis testing in a fully-data driven manner. The appropriate
parametric model which is supported by the data can be always deduced at the
end of model selection. In the MaxEnt framework, we recover major scores and
selection procedures used in multiple applications and assess their ability to
capture associations in the data-generating process and identify the most
generalizable model. This data-driven counterpart of standard model selection
demonstrates the unifying prospective of the deductive logic advocated by
MaxEnt principle, while potentially shedding new insights to the inverse
problem.
Related papers
- Continuous Bayesian Model Selection for Multivariate Causal Discovery [22.945274948173182]
Current causal discovery approaches require restrictive model assumptions or assume access to interventional data to ensure structure identifiability.
Recent work has shown that Bayesian model selection can greatly improve accuracy by exchanging restrictive modelling for more flexible assumptions.
We demonstrate the competitiveness of our approach on both synthetic and real-world datasets.
arXiv Detail & Related papers (2024-11-15T12:55:05Z) - A Likelihood Based Approach to Distribution Regression Using Conditional Deep Generative Models [6.647819824559201]
We study the large-sample properties of a likelihood-based approach for estimating conditional deep generative models.
Our results lead to the convergence rate of a sieve maximum likelihood estimator for estimating the conditional distribution.
arXiv Detail & Related papers (2024-10-02T20:46:21Z) - Scaling and renormalization in high-dimensional regression [72.59731158970894]
This paper presents a succinct derivation of the training and generalization performance of a variety of high-dimensional ridge regression models.
We provide an introduction and review of recent results on these topics, aimed at readers with backgrounds in physics and deep learning.
arXiv Detail & Related papers (2024-05-01T15:59:00Z) - Likelihood Based Inference in Fully and Partially Observed Exponential Family Graphical Models with Intractable Normalizing Constants [4.532043501030714]
Probabilistic graphical models that encode an underlying Markov random field are fundamental building blocks of generative modeling.
This paper is to demonstrate that full likelihood based analysis of these models is feasible in a computationally efficient manner.
arXiv Detail & Related papers (2024-04-27T02:58:22Z) - Diffusion posterior sampling for simulation-based inference in tall data settings [53.17563688225137]
Simulation-based inference ( SBI) is capable of approximating the posterior distribution that relates input parameters to a given observation.
In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model.
We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.
arXiv Detail & Related papers (2024-04-11T09:23:36Z) - On Least Square Estimation in Softmax Gating Mixture of Experts [78.3687645289918]
We investigate the performance of the least squares estimators (LSE) under a deterministic MoE model.
We establish a condition called strong identifiability to characterize the convergence behavior of various types of expert functions.
Our findings have important practical implications for expert selection.
arXiv Detail & Related papers (2024-02-05T12:31:18Z) - PAC Reinforcement Learning for Predictive State Representations [60.00237613646686]
We study online Reinforcement Learning (RL) in partially observable dynamical systems.
We focus on the Predictive State Representations (PSRs) model, which is an expressive model that captures other well-known models.
We develop a novel model-based algorithm for PSRs that can learn a near optimal policy in sample complexity scalingly.
arXiv Detail & Related papers (2022-07-12T17:57:17Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - The Predictive Normalized Maximum Likelihood for Over-parameterized
Linear Regression with Norm Constraint: Regret and Double Descent [12.929639356256928]
We show that modern machine learning models do not obey a trade-off between the complexity of a prediction rule and its ability to generalize.
We use the recently proposed predictive normalized maximum likelihood (pNML) which is the min-max regret solution for individual data.
We demonstrate the use of the pNML regret as a point-wise learnability measure on synthetic data and that it can successfully predict the double-decent phenomenon.
arXiv Detail & Related papers (2021-02-14T15:49:04Z) - On the minmax regret for statistical manifolds: the role of curvature [68.8204255655161]
Two-part codes and the minimum description length have been successful in delivering procedures to single out the best models.
We derive a sharper expression than the standard one given by the complexity, where the scalar curvature of the Fisher information metric plays a dominant role.
arXiv Detail & Related papers (2020-07-06T17:28:19Z) - On a computationally-scalable sparse formulation of the multidimensional
and non-stationary maximum entropy principle [0.0]
We derive a non-stationary formulation of the MaxEnt-principle and show that its solution can be approximated through a numerical maximisation of the sparse constrained optimization problem with regularization.
We show that all of the considered seven major financial benchmark time series are better described by conditionally memoryless MaxEnt-models.
This analysis also reveals a sparse network of statistically-significant temporal relations for the positive and negative latent variance changes among different markets.
arXiv Detail & Related papers (2020-05-07T05:22:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.