Gaussian Mixture Estimation from Weighted Samples
- URL: http://arxiv.org/abs/2106.05109v1
- Date: Wed, 9 Jun 2021 14:38:46 GMT
- Title: Gaussian Mixture Estimation from Weighted Samples
- Authors: Daniel Frisch and Uwe D. Hanebeck
- Abstract summary: We consider estimating the parameters of a Gaussian mixture density with a given number of components best representing a given set of weighted samples.
We adopt a density interpretation of the samples by viewing them as a discrete Dirac mixture density over a continuous domain with weighted components.
An expectation-maximization method is proposed that properly considers not only the sample locations, but also the corresponding weights.
- Score: 9.442139459221785
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider estimating the parameters of a Gaussian mixture density with a
given number of components best representing a given set of weighted samples.
We adopt a density interpretation of the samples by viewing them as a discrete
Dirac mixture density over a continuous domain with weighted components. Hence,
Gaussian mixture fitting is viewed as density re-approximation. In order to
speed up computation, an expectation-maximization method is proposed that
properly considers not only the sample locations, but also the corresponding
weights. It is shown that methods from literature do not treat the weights
correctly, resulting in wrong estimates. This is demonstrated with simple
counterexamples. The proposed method works in any number of dimensions with the
same computational load as standard Gaussian mixture estimators for unweighted
samples.
Related papers
- Density Ratio Estimation via Sampling along Generalized Geodesics on Statistical Manifolds [0.951494089949975]
We geometrically reinterpret existing methods for density ratio estimation based on incremental mixtures.
To achieve such a method requires Monte Carlo sampling along geodesics via transformations of the two distributions.
Our experiments demonstrate that the proposed approach outperforms the existing approaches.
arXiv Detail & Related papers (2024-06-27T00:44:46Z) - Sobolev Space Regularised Pre Density Models [51.558848491038916]
We propose a new approach to non-parametric density estimation that is based on regularizing a Sobolev norm of the density.
This method is statistically consistent, and makes the inductive validation model clear and consistent.
arXiv Detail & Related papers (2023-07-25T18:47:53Z) - Estimating Joint Probability Distribution With Low-Rank Tensor
Decomposition, Radon Transforms and Dictionaries [3.0892724364965005]
We describe a method for estimating the joint probability density from data samples by assuming that the underlying distribution can be decomposed as a mixture of product densities with few mixture components.
We combine two key ideas: dictionaries to represent 1-D densities, and random projections to estimate the joint distribution from 1-D marginals.
Our algorithm benefits from improved sample complexity over the previous dictionary-based approach by using 1-D marginals for reconstruction.
arXiv Detail & Related papers (2023-04-18T05:37:15Z) - Mean-Square Analysis of Discretized It\^o Diffusions for Heavy-tailed
Sampling [17.415391025051434]
We analyze the complexity of sampling from a class of heavy-tailed distributions by discretizing a natural class of Ito diffusions associated with weighted Poincar'e inequalities.
Based on a mean-square analysis, we establish the iteration complexity for obtaining a sample whose distribution is $epsilon$ close to the target distribution in the Wasserstein-2 metric.
arXiv Detail & Related papers (2023-03-01T15:16:03Z) - Unsupervised Learning of Sampling Distributions for Particle Filters [80.6716888175925]
We put forward four methods for learning sampling distributions from observed measurements.
Experiments demonstrate that learned sampling distributions exhibit better performance than designed, minimum-degeneracy sampling distributions.
arXiv Detail & Related papers (2023-02-02T15:50:21Z) - Importance sampling for stochastic quantum simulations [68.8204255655161]
We introduce the qDrift protocol, which builds random product formulas by sampling from the Hamiltonian according to the coefficients.
We show that the simulation cost can be reduced while achieving the same accuracy, by considering the individual simulation cost during the sampling stage.
Results are confirmed by numerical simulations performed on a lattice nuclear effective field theory.
arXiv Detail & Related papers (2022-12-12T15:06:32Z) - A Robust and Flexible EM Algorithm for Mixtures of Elliptical
Distributions with Missing Data [71.9573352891936]
This paper tackles the problem of missing data imputation for noisy and non-Gaussian data.
A new EM algorithm is investigated for mixtures of elliptical distributions with the property of handling potential missing data.
Experimental results on synthetic data demonstrate that the proposed algorithm is robust to outliers and can be used with non-Gaussian data.
arXiv Detail & Related papers (2022-01-28T10:01:37Z) - Unrolling Particles: Unsupervised Learning of Sampling Distributions [102.72972137287728]
Particle filtering is used to compute good nonlinear estimates of complex systems.
We show in simulations that the resulting particle filter yields good estimates in a wide range of scenarios.
arXiv Detail & Related papers (2021-10-06T16:58:34Z) - Relative Entropy Gradient Sampler for Unnormalized Distributions [14.060615420986796]
Relative entropy gradient sampler (REGS) for sampling from unnormalized distributions.
REGS is a particle method that seeks a sequence of simple nonlinear transforms iteratively pushing the initial samples from a reference distribution into the samples from an unnormalized target distribution.
arXiv Detail & Related papers (2021-10-06T14:10:38Z) - Consistent Estimation of Identifiable Nonparametric Mixture Models from
Grouped Observations [84.81435917024983]
This work proposes an algorithm that consistently estimates any identifiable mixture model from grouped observations.
A practical implementation is provided for paired observations, and the approach is shown to outperform existing methods.
arXiv Detail & Related papers (2020-06-12T20:44:22Z) - Uniform Convergence Rates for Maximum Likelihood Estimation under
Two-Component Gaussian Mixture Models [13.769786711365104]
We derive uniform convergence rates for the maximum likelihood estimator and minimax lower bounds for parameter estimation.
We assume the mixing proportions of the mixture are known and fixed, but make no separation assumption on the underlying mixture components.
arXiv Detail & Related papers (2020-06-01T04:13:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.