Information Theoretical Importance Sampling Clustering
- URL: http://arxiv.org/abs/2302.04421v2
- Date: Tue, 30 May 2023 07:03:38 GMT
- Title: Information Theoretical Importance Sampling Clustering
- Authors: Jiangshe Zhang, Lizhen Ji, Meng Wang
- Abstract summary: A current assumption of most clustering methods is that the training data and future data are taken from the same distribution.
We propose an information theoretical importance sampling based approach for clustering problems (ITISC)
Experiment results on synthetic datasets and a real-world load forecasting problem validate the effectiveness of the proposed model.
- Score: 18.248246885248733
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A current assumption of most clustering methods is that the training data and
future data are taken from the same distribution. However, this assumption may
not hold in most real-world scenarios. In this paper, we propose an information
theoretical importance sampling based approach for clustering problems (ITISC)
which minimizes the worst case of expected distortions under the constraint of
distribution deviation. The distribution deviation constraint can be converted
to the constraint over a set of weight distributions centered on the uniform
distribution derived from importance sampling. The objective of the proposed
approach is to minimize the loss under maximum degradation hence the resulting
problem is a constrained minimax optimization problem which can be reformulated
to an unconstrained problem using the Lagrange method. The optimization problem
can be solved by both an alternative optimization algorithm or a general
optimization routine by commercially available software. Experiment results on
synthetic datasets and a real-world load forecasting problem validate the
effectiveness of the proposed model. Furthermore, we show that fuzzy c-means is
a special case of ITISC with the logarithmic distortion, and this observation
provides an interesting physical interpretation for fuzzy exponent $m$.
Related papers
- Error Feedback under $(L_0,L_1)$-Smoothness: Normalization and Momentum [56.37522020675243]
We provide the first proof of convergence for normalized error feedback algorithms across a wide range of machine learning problems.
We show that due to their larger allowable stepsizes, our new normalized error feedback algorithms outperform their non-normalized counterparts on various tasks.
arXiv Detail & Related papers (2024-10-22T10:19:27Z) - Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Distributed Markov Chain Monte Carlo Sampling based on the Alternating
Direction Method of Multipliers [143.6249073384419]
In this paper, we propose a distributed sampling scheme based on the alternating direction method of multipliers.
We provide both theoretical guarantees of our algorithm's convergence and experimental evidence of its superiority to the state-of-the-art.
In simulation, we deploy our algorithm on linear and logistic regression tasks and illustrate its fast convergence compared to existing gradient-based methods.
arXiv Detail & Related papers (2024-01-29T02:08:40Z) - Optimal Multi-Distribution Learning [88.3008613028333]
Multi-distribution learning seeks to learn a shared model that minimizes the worst-case risk across $k$ distinct data distributions.
We propose a novel algorithm that yields an varepsilon-optimal randomized hypothesis with a sample complexity on the order of (d+k)/varepsilon2.
arXiv Detail & Related papers (2023-12-08T16:06:29Z) - Flow-based Distributionally Robust Optimization [23.232731771848883]
We present a framework, called $textttFlowDRO$, for solving flow-based distributionally robust optimization (DRO) problems with Wasserstein uncertainty sets.
We aim to find continuous worst-case distribution (also called the Least Favorable Distribution, LFD) and sample from it.
We demonstrate its usage in adversarial learning, distributionally robust hypothesis testing, and a new mechanism for data-driven distribution perturbation differential privacy.
arXiv Detail & Related papers (2023-10-30T03:53:31Z) - Constrained Reweighting of Distributions: an Optimal Transport Approach [8.461214317999321]
We introduce a nonparametrically imbued distributional constraints on the weights, and develop a general framework leveraging the maximum entropy principle and tools from optimal transport.
The framework is demonstrated in the context of three disparate applications: portfolio allocation, semi-parametric inference for complex surveys, and ensuring algorithmic fairness in machine learning algorithms.
arXiv Detail & Related papers (2023-10-19T03:54:31Z) - An Optimization-based Deep Equilibrium Model for Hyperspectral Image
Deconvolution with Convergence Guarantees [71.57324258813675]
We propose a novel methodology for addressing the hyperspectral image deconvolution problem.
A new optimization problem is formulated, leveraging a learnable regularizer in the form of a neural network.
The derived iterative solver is then expressed as a fixed-point calculation problem within the Deep Equilibrium framework.
arXiv Detail & Related papers (2023-06-10T08:25:16Z) - Prescriptive PCA: Dimensionality Reduction for Two-stage Stochastic
Optimization [1.1612308609123565]
We develop a prescriptive dimensionality reduction framework that aims to minimize the degree of suboptimality in the optimization phase.
For the case where the downstream optimization problem has an expected value objective, we show that prescriptive dimensionality reduction can be performed via solving a distributionally-robust optimization problem.
Our approach significantly outperforms principal component analysis with real and synthetic data sets.
arXiv Detail & Related papers (2023-06-04T00:50:35Z) - A data variation robust learning model based on importance sampling [11.285259001286978]
We propose an importance sampling based data variation robust loss (ISloss) for learning problems which minimizes the worst case of loss under the constraint of distribution deviation.
We show that the proposed method is robust under large distribution deviations.
arXiv Detail & Related papers (2023-02-09T04:50:06Z) - Amortized Conditional Normalized Maximum Likelihood: Reliable Out of
Distribution Uncertainty Estimation [99.92568326314667]
We propose the amortized conditional normalized maximum likelihood (ACNML) method as a scalable general-purpose approach for uncertainty estimation.
Our algorithm builds on the conditional normalized maximum likelihood (CNML) coding scheme, which has minimax optimal properties according to the minimum description length principle.
We demonstrate that ACNML compares favorably to a number of prior techniques for uncertainty estimation in terms of calibration on out-of-distribution inputs.
arXiv Detail & Related papers (2020-11-05T08:04:34Z) - Optimization in Machine Learning: A Distribution Space Approach [16.038814087205143]
We present the viewpoint that optimization problems in machine learning can often be interpreted as minimizing a convex functional over a function space.
We repose such problems via a suitable relaxation as convex optimization problems in the space distributions.
We develop a numerical algorithm based on mixture distributions to perform approximate optimization directly in distribution space.
arXiv Detail & Related papers (2020-04-18T13:38:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.