Consistent Estimation of Identifiable Nonparametric Mixture Models from
Grouped Observations
- URL: http://arxiv.org/abs/2006.07459v1
- Date: Fri, 12 Jun 2020 20:44:22 GMT
- Title: Consistent Estimation of Identifiable Nonparametric Mixture Models from
Grouped Observations
- Authors: Alexander Ritchie, Robert A. Vandermeulen, Clayton Scott
- Abstract summary: This work proposes an algorithm that consistently estimates any identifiable mixture model from grouped observations.
A practical implementation is provided for paired observations, and the approach is shown to outperform existing methods.
- Score: 84.81435917024983
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent research has established sufficient conditions for finite mixture
models to be identifiable from grouped observations. These conditions allow the
mixture components to be nonparametric and have substantial (or even total)
overlap. This work proposes an algorithm that consistently estimates any
identifiable mixture model from grouped observations. Our analysis leverages an
oracle inequality for weighted kernel density estimators of the distribution on
groups, together with a general result showing that consistent estimation of
the distribution on groups implies consistent estimation of mixture components.
A practical implementation is provided for paired observations, and the
approach is shown to outperform existing methods, especially when mixture
components overlap significantly.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Collaborative Heterogeneous Causal Inference Beyond Meta-analysis [68.4474531911361]
We propose a collaborative inverse propensity score estimator for causal inference with heterogeneous data.
Our method shows significant improvements over the methods based on meta-analysis when heterogeneity increases.
arXiv Detail & Related papers (2024-04-24T09:04:36Z) - Mixture Proportion Estimation Beyond Irreducibility [9.99177526976127]
We present a resampling-based meta-algorithm that takes any existing MPE algorithm designed to work under irreducibility and adapts it to work under our more general condition.
Our approach exhibits improved estimation performance relative to baseline methods and to a recently proposed regrouping-based algorithm.
arXiv Detail & Related papers (2023-06-02T03:32:44Z) - Estimating Joint Probability Distribution With Low-Rank Tensor
Decomposition, Radon Transforms and Dictionaries [3.0892724364965005]
We describe a method for estimating the joint probability density from data samples by assuming that the underlying distribution can be decomposed as a mixture of product densities with few mixture components.
We combine two key ideas: dictionaries to represent 1-D densities, and random projections to estimate the joint distribution from 1-D marginals.
Our algorithm benefits from improved sample complexity over the previous dictionary-based approach by using 1-D marginals for reconstruction.
arXiv Detail & Related papers (2023-04-18T05:37:15Z) - On the inadequacy of nominal assortativity for assessing homophily in
networks [0.0]
We show that nominal assortativity fails to account for group imbalance and asymmetric group interactions.
We propose adjusted nominal assortativity and show that this adjustment recovers the expected assortativity in networks with various level of mixing.
arXiv Detail & Related papers (2022-11-18T14:04:39Z) - RegMixup: Mixup as a Regularizer Can Surprisingly Improve Accuracy and
Out Distribution Robustness [94.69774317059122]
We show that the effectiveness of the well celebrated Mixup can be further improved if instead of using it as the sole learning objective, it is utilized as an additional regularizer to the standard cross-entropy loss.
This simple change not only provides much improved accuracy but also significantly improves the quality of the predictive uncertainty estimation of Mixup.
arXiv Detail & Related papers (2022-06-29T09:44:33Z) - Clustering consistency with Dirichlet process mixtures [0.0]
We study the posterior distribution induced by Dirichlet process mixtures as the sample size increases.
We show that consistency for the number of clusters can be achieved if the concentration parameter is adapted in a fully Bayesian way.
arXiv Detail & Related papers (2022-05-25T17:21:42Z) - A Robust and Flexible EM Algorithm for Mixtures of Elliptical
Distributions with Missing Data [71.9573352891936]
This paper tackles the problem of missing data imputation for noisy and non-Gaussian data.
A new EM algorithm is investigated for mixtures of elliptical distributions with the property of handling potential missing data.
Experimental results on synthetic data demonstrate that the proposed algorithm is robust to outliers and can be used with non-Gaussian data.
arXiv Detail & Related papers (2022-01-28T10:01:37Z) - Uniform Convergence Rates for Maximum Likelihood Estimation under
Two-Component Gaussian Mixture Models [13.769786711365104]
We derive uniform convergence rates for the maximum likelihood estimator and minimax lower bounds for parameter estimation.
We assume the mixing proportions of the mixture are known and fixed, but make no separation assumption on the underlying mixture components.
arXiv Detail & Related papers (2020-06-01T04:13:48Z) - Nonparametric Score Estimators [49.42469547970041]
Estimating the score from a set of samples generated by an unknown distribution is a fundamental task in inference and learning of probabilistic models.
We provide a unifying view of these estimators under the framework of regularized nonparametric regression.
We propose score estimators based on iterative regularization that enjoy computational benefits from curl-free kernels and fast convergence.
arXiv Detail & Related papers (2020-05-20T15:01:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.