Learn2Extend: Extending sequences by retaining their statistical
properties with mixture models
- URL: http://arxiv.org/abs/2312.01507v1
- Date: Sun, 3 Dec 2023 21:05:50 GMT
- Title: Learn2Extend: Extending sequences by retaining their statistical
properties with mixture models
- Authors: Dimitris Vartziotis, George Dasoulas, Florian Pausinger
- Abstract summary: This paper addresses the challenge of extending general finite sequences of real numbers within a subinterval of the real line.
Our focus lies on preserving the gap distribution and pair correlation function of these point sets.
Leveraging advancements in deep learning applied to point processes, this paper explores the use of an auto-regressive textitSequence Extension Mixture Model.
- Score: 7.15769102504304
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper addresses the challenge of extending general finite sequences of
real numbers within a subinterval of the real line, maintaining their inherent
statistical properties by employing machine learning. Our focus lies on
preserving the gap distribution and pair correlation function of these point
sets. Leveraging advancements in deep learning applied to point processes, this
paper explores the use of an auto-regressive \textit{Sequence Extension Mixture
Model} (SEMM) for extending finite sequences, by estimating directly the
conditional density, instead of the intensity function. We perform comparative
experiments on multiple types of point processes, including Poisson, locally
attractive, and locally repelling sequences, and we perform a case study on the
prediction of Riemann $\zeta$ function zeroes. The results indicate that the
proposed mixture model outperforms traditional neural network architectures in
sequence extension with the retention of statistical properties. Given this
motivation, we showcase the capabilities of a mixture model to extend
sequences, maintaining specific statistical properties, i.e. the gap
distribution, and pair correlation indicators.
Related papers
- On learning higher-order cumulants in diffusion models [6.610338540492242]
We study the behaviour of higher-order cumulants, or connected n-point functions, under both the forward and backward process.
We derive explicit expressions for the moment- and cumulant-generating functionals.
We confirm our results in an exactly solvable toy model with nonzero cumulants and in scalar lattice field theory.
arXiv Detail & Related papers (2024-10-28T16:57:02Z) - Statistical Inference for Temporal Difference Learning with Linear Function Approximation [62.69448336714418]
Temporal Difference (TD) learning, arguably the most widely used for policy evaluation, serves as a natural framework for this purpose.
In this paper, we study the consistency properties of TD learning with Polyak-Ruppert averaging and linear function approximation, and obtain three significant improvements over existing results.
arXiv Detail & Related papers (2024-10-21T15:34:44Z) - Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis [56.442307356162864]
We study the theoretical aspects of score-based discrete diffusion models under the Continuous Time Markov Chain (CTMC) framework.
We introduce a discrete-time sampling algorithm in the general state space $[S]d$ that utilizes score estimators at predefined time points.
Our convergence analysis employs a Girsanov-based method and establishes key properties of the discrete score function.
arXiv Detail & Related papers (2024-10-03T09:07:13Z) - Convergence Analysis of Discrete Diffusion Model: Exact Implementation
through Uniformization [17.535229185525353]
We introduce an algorithm leveraging the uniformization of continuous Markov chains, implementing transitions on random time points.
Our results align with state-of-the-art achievements for diffusion models in $mathbbRd$ and further underscore the advantages of discrete diffusion models in comparison to the $mathbbRd$ setting.
arXiv Detail & Related papers (2024-02-12T22:26:52Z) - Learning Linear Causal Representations from Interventions under General
Nonlinear Mixing [52.66151568785088]
We prove strong identifiability results given unknown single-node interventions without access to the intervention targets.
This is the first instance of causal identifiability from non-paired interventions for deep neural network embeddings.
arXiv Detail & Related papers (2023-06-04T02:32:12Z) - Score-based Continuous-time Discrete Diffusion Models [102.65769839899315]
We extend diffusion models to discrete variables by introducing a Markov jump process where the reverse process denoises via a continuous-time Markov chain.
We show that an unbiased estimator can be obtained via simple matching the conditional marginal distributions.
We demonstrate the effectiveness of the proposed method on a set of synthetic and real-world music and image benchmarks.
arXiv Detail & Related papers (2022-11-30T05:33:29Z) - Mutual Exclusivity Training and Primitive Augmentation to Induce
Compositionality [84.94877848357896]
Recent datasets expose the lack of the systematic generalization ability in standard sequence-to-sequence models.
We analyze this behavior of seq2seq models and identify two contributing factors: a lack of mutual exclusivity bias and the tendency to memorize whole examples.
We show substantial empirical improvements using standard sequence-to-sequence models on two widely-used compositionality datasets.
arXiv Detail & Related papers (2022-11-28T17:36:41Z) - Functional Mixtures-of-Experts [0.24578723416255746]
We consider the statistical analysis of heterogeneous data for prediction in situations where the observations include functions.
We first present a new family of ME models, named functional ME (FME) in which the predictors are potentially noisy observations.
We develop dedicated expectation--maximization algorithms for Lasso-like (EM-Lasso) regularized maximum-likelihood parameter estimation strategies to fit the models.
arXiv Detail & Related papers (2022-02-04T17:32:28Z) - Optimal regularizations for data generation with probabilistic graphical
models [0.0]
Empirically, well-chosen regularization schemes dramatically improve the quality of the inferred models.
We consider the particular case of L 2 and L 1 regularizations in the Maximum A Posteriori (MAP) inference of generative pairwise graphical models.
arXiv Detail & Related papers (2021-12-02T14:45:16Z) - Slice Sampling for General Completely Random Measures [74.24975039689893]
We present a novel Markov chain Monte Carlo algorithm for posterior inference that adaptively sets the truncation level using auxiliary slice variables.
The efficacy of the proposed algorithm is evaluated on several popular nonparametric models.
arXiv Detail & Related papers (2020-06-24T17:53:53Z) - The continuous categorical: a novel simplex-valued exponential family [23.983555024375306]
We show that standard choices for simplex-valued data suffer from a number of limitations, including bias and numerical issues.
We resolve these limitations by introducing a novel exponential family of distributions for modeling simplex-valued data.
Unlike the Dirichlet and other typical choices, the continuous categorical results in a well-behaved probabilistic loss function.
arXiv Detail & Related papers (2020-02-20T04:28:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.