On a computationally-scalable sparse formulation of the multidimensional
and non-stationary maximum entropy principle
- URL: http://arxiv.org/abs/2005.03253v1
- Date: Thu, 7 May 2020 05:22:46 GMT
- Title: On a computationally-scalable sparse formulation of the multidimensional
and non-stationary maximum entropy principle
- Authors: Horenko Illia and Marchenko Ganna and Gagliardini Patrick
- Abstract summary: We derive a non-stationary formulation of the MaxEnt-principle and show that its solution can be approximated through a numerical maximisation of the sparse constrained optimization problem with regularization.
We show that all of the considered seven major financial benchmark time series are better described by conditionally memoryless MaxEnt-models.
This analysis also reveals a sparse network of statistically-significant temporal relations for the positive and negative latent variance changes among different markets.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data-driven modelling and computational predictions based on maximum entropy
principle (MaxEnt-principle) aim at finding as-simple-as-possible - but not
simpler then necessary - models that allow to avoid the data overfitting
problem. We derive a multivariate non-parametric and non-stationary formulation
of the MaxEnt-principle and show that its solution can be approximated through
a numerical maximisation of the sparse constrained optimization problem with
regularization. Application of the resulting algorithm to popular financial
benchmarks reveals memoryless models allowing for simple and qualitative
descriptions of the major stock market indexes data. We compare the obtained
MaxEnt-models to the heteroschedastic models from the computational
econometrics (GARCH, GARCH-GJR, MS-GARCH, GARCH-PML4) in terms of the model
fit, complexity and prediction quality. We compare the resulting model
log-likelihoods, the values of the Bayesian Information Criterion, posterior
model probabilities, the quality of the data autocorrelation function fits as
well as the Value-at-Risk prediction quality. We show that all of the
considered seven major financial benchmark time series (DJI, SPX, FTSE, STOXX,
SMI, HSI and N225) are better described by conditionally memoryless
MaxEnt-models with nonstationary regime-switching than by the common
econometric models with finite memory. This analysis also reveals a sparse
network of statistically-significant temporal relations for the positive and
negative latent variance changes among different markets. The code is provided
for open access.
Related papers
- Continuous Bayesian Model Selection for Multivariate Causal Discovery [22.945274948173182]
Current causal discovery approaches require restrictive model assumptions or assume access to interventional data to ensure structure identifiability.
Recent work has shown that Bayesian model selection can greatly improve accuracy by exchanging restrictive modelling for more flexible assumptions.
We demonstrate the competitiveness of our approach on both synthetic and real-world datasets.
arXiv Detail & Related papers (2024-11-15T12:55:05Z) - Latent Semantic Consensus For Deterministic Geometric Model Fitting [109.44565542031384]
We propose an effective method called Latent Semantic Consensus (LSC)
LSC formulates the model fitting problem into two latent semantic spaces based on data points and model hypotheses.
LSC is able to provide consistent and reliable solutions within only a few milliseconds for general multi-structural model fitting.
arXiv Detail & Related papers (2024-03-11T05:35:38Z) - On the Consistency of Maximum Likelihood Estimation of Probabilistic
Principal Component Analysis [1.0528389538549636]
PPCA has a broad spectrum of applications ranging from science and engineering to quantitative finance.
Despite this wide applicability in various fields, hardly any theoretical guarantees exist to justify the soundness of the maximal likelihood (ML) solution for this model.
We propose a novel approach using quotient topological spaces and in particular, we show that the maximum likelihood solution is consistent in an appropriate quotient Euclidean space.
arXiv Detail & Related papers (2023-11-08T22:40:45Z) - Stable Training of Probabilistic Models Using the Leave-One-Out Maximum Log-Likelihood Objective [0.7373617024876725]
Kernel density estimation (KDE) based models are popular choices for this task, but they fail to adapt to data regions with varying densities.
An adaptive KDE model is employed to circumvent this, where each kernel in the model has an individual bandwidth.
A modified expectation-maximization algorithm is employed to accelerate the optimization speed reliably.
arXiv Detail & Related papers (2023-10-05T14:08:42Z) - Entropy-based Characterization of Modeling Constraints [0.0]
In most data-scientific approaches, the principle of Entropy (MaxEnt) is used to justify some parametric model.
We derive the distribution over all viable distributions that satisfy the provided set of constraints.
The appropriate parametric model which is supported by the data can be always deduced at the end of model selection.
arXiv Detail & Related papers (2022-06-27T17:25:49Z) - Improved Prediction and Network Estimation Using the Monotone Single
Index Multi-variate Autoregressive Model [34.529641317832024]
We develop a semi-parametric approach based on the monotone single-index multi-variate autoregressive model (SIMAM)
We provide theoretical guarantees for dependent data and an alternating projected gradient descent algorithm.
We demonstrate the superior performance both on simulated data and two real data examples.
arXiv Detail & Related papers (2021-06-28T12:32:29Z) - Post-mortem on a deep learning contest: a Simpson's paradox and the
complementary roles of scale metrics versus shape metrics [61.49826776409194]
We analyze a corpus of models made publicly-available for a contest to predict the generalization accuracy of neural network (NN) models.
We identify what amounts to a Simpson's paradox: where "scale" metrics perform well overall but perform poorly on sub partitions of the data.
We present two novel shape metrics, one data-independent, and the other data-dependent, which can predict trends in the test accuracy of a series of NNs.
arXiv Detail & Related papers (2021-06-01T19:19:49Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Autoregressive Score Matching [113.4502004812927]
We propose autoregressive conditional score models (AR-CSM) where we parameterize the joint distribution in terms of the derivatives of univariable log-conditionals (scores)
For AR-CSM models, this divergence between data and model distributions can be computed and optimized efficiently, requiring no expensive sampling or adversarial training.
We show with extensive experimental results that it can be applied to density estimation on synthetic data, image generation, image denoising, and training latent variable models with implicit encoders.
arXiv Detail & Related papers (2020-10-24T07:01:24Z) - Accounting for Unobserved Confounding in Domain Generalization [107.0464488046289]
This paper investigates the problem of learning robust, generalizable prediction models from a combination of datasets.
Part of the challenge of learning robust models lies in the influence of unobserved confounders.
We demonstrate the empirical performance of our approach on healthcare data from different modalities.
arXiv Detail & Related papers (2020-07-21T08:18:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.