CLUB: A Contrastive Log-ratio Upper Bound of Mutual Information
- URL: http://arxiv.org/abs/2006.12013v6
- Date: Thu, 23 Jul 2020 20:52:37 GMT
- Title: CLUB: A Contrastive Log-ratio Upper Bound of Mutual Information
- Authors: Pengyu Cheng, Weituo Hao, Shuyang Dai, Jiachang Liu, Zhe Gan, Lawrence
Carin
- Abstract summary: We propose a novel Contrastive Log-ratio Upper Bound (CLUB) of mutual information.
We provide a theoretical analysis of the properties of CLUB and its variational approximation.
Based on this upper bound, we introduce a MI minimization training scheme and further accelerate it with a negative sampling strategy.
- Score: 105.73798100327667
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mutual information (MI) minimization has gained considerable interests in
various machine learning tasks. However, estimating and minimizing MI in
high-dimensional spaces remains a challenging problem, especially when only
samples, rather than distribution forms, are accessible. Previous works mainly
focus on MI lower bound approximation, which is not applicable to MI
minimization problems. In this paper, we propose a novel Contrastive Log-ratio
Upper Bound (CLUB) of mutual information. We provide a theoretical analysis of
the properties of CLUB and its variational approximation. Based on this upper
bound, we introduce a MI minimization training scheme and further accelerate it
with a negative sampling strategy. Simulation studies on Gaussian distributions
show the reliable estimation ability of CLUB. Real-world MI minimization
experiments, including domain adaptation and information bottleneck,
demonstrate the effectiveness of the proposed method. The code is at
https://github.com/Linear95/CLUB.
Related papers
- Improving Mutual Information Estimation with Annealed and Energy-Based
Bounds [20.940022170594816]
Mutual information (MI) is a fundamental quantity in information theory and machine learning.
We present a unifying view of existing MI bounds from the perspective of importance sampling.
We propose three novel bounds based on this approach.
arXiv Detail & Related papers (2023-03-13T10:47:24Z) - An Information Minimization Based Contrastive Learning Model for
Unsupervised Sentence Embeddings Learning [19.270283247740664]
We present an information minimization based contrastive learning (InforMin-CL) model for unsupervised sentence representation learning.
We find that information minimization can be achieved by simple contrast and reconstruction objectives.
arXiv Detail & Related papers (2022-09-22T12:07:35Z) - Cooperative Distribution Alignment via JSD Upper Bound [7.071749623370137]
Unsupervised distribution alignment estimates a transformation that maps two or more source distributions to a shared aligned distribution.
This task has many applications including generative modeling, unsupervised domain adaptation, and socially aware learning.
We propose to unify and generalize previous flow-based approaches under a single non-adversarial framework.
arXiv Detail & Related papers (2022-07-05T20:09:03Z) - Minimax Optimization: The Case of Convex-Submodular [50.03984152441271]
Minimax problems extend beyond the continuous domain to mixed continuous-discrete domains or even fully discrete domains.
We introduce the class of convex-submodular minimax problems, where the objective is convex with respect to the continuous variable and submodular with respect to the discrete variable.
Our proposed algorithms are iterative and combine tools from both discrete and continuous optimization.
arXiv Detail & Related papers (2021-11-01T21:06:35Z) - Learning with Multiclass AUC: Theory and Algorithms [141.63211412386283]
Area under the ROC curve (AUC) is a well-known ranking metric for problems such as imbalanced learning and recommender systems.
In this paper, we start an early trial to consider the problem of learning multiclass scoring functions via optimizing multiclass AUC metrics.
arXiv Detail & Related papers (2021-07-28T05:18:10Z) - Tight Mutual Information Estimation With Contrastive Fenchel-Legendre
Optimization [69.07420650261649]
We introduce a novel, simple, and powerful contrastive MI estimator named as FLO.
Empirically, our FLO estimator overcomes the limitations of its predecessors and learns more efficiently.
The utility of FLO is verified using an extensive set of benchmarks, which also reveals the trade-offs in practical MI estimation.
arXiv Detail & Related papers (2021-07-02T15:20:41Z) - Decomposed Mutual Information Estimation for Contrastive Representation
Learning [66.52795579973484]
Recent contrastive representation learning methods rely on estimating mutual information (MI) between multiple views of an underlying context.
We propose decomposing the full MI estimation problem into a sum of smaller estimation problems by splitting one of the views into progressively more informed subviews.
This expression contains a sum of unconditional and conditional MI terms, each measuring modest chunks of the total MI.
We show that DEMI can capture a larger amount of MI than standard non-decomposed contrastive bounds in a synthetic setting.
arXiv Detail & Related papers (2021-06-25T03:19:25Z) - Mutual Information Gradient Estimation for Representation Learning [56.08429809658762]
Mutual Information (MI) plays an important role in representation learning.
Recent advances establish tractable and scalable MI estimators to discover useful representation.
We propose the Mutual Information Gradient Estimator (MIGE) for representation learning based on the score estimation of implicit distributions.
arXiv Detail & Related papers (2020-05-03T16:05:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.