Bottlenecks CLUB: Unifying Information-Theoretic Trade-offs Among
Complexity, Leakage, and Utility
- URL: http://arxiv.org/abs/2207.04895v1
- Date: Mon, 11 Jul 2022 14:07:48 GMT
- Title: Bottlenecks CLUB: Unifying Information-Theoretic Trade-offs Among
Complexity, Leakage, and Utility
- Authors: Behrooz Razeghi, Flavio P. Calmon, Deniz Gunduz, Slava Voloshynovskiy
- Abstract summary: Bottleneck problems are an important class of optimization problems that have recently gained increasing attention in the domain of machine learning and information theory.
We propose a general family of optimization problems, termed as complexity-leakage-utility bottleneck (CLUB) model.
We show that the CLUB model generalizes all these problems as well as most other information-theoretic privacy models.
- Score: 8.782250973555026
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Bottleneck problems are an important class of optimization problems that have
recently gained increasing attention in the domain of machine learning and
information theory. They are widely used in generative models, fair machine
learning algorithms, design of privacy-assuring mechanisms, and appear as
information-theoretic performance bounds in various multi-user communication
problems. In this work, we propose a general family of optimization problems,
termed as complexity-leakage-utility bottleneck (CLUB) model, which (i)
provides a unified theoretical framework that generalizes most of the
state-of-the-art literature for the information-theoretic privacy models, (ii)
establishes a new interpretation of the popular generative and discriminative
models, (iii) constructs new insights to the generative compression models, and
(iv) can be used in the fair generative models. We first formulate the CLUB
model as a complexity-constrained privacy-utility optimization problem. We then
connect it with the closely related bottleneck problems, namely information
bottleneck (IB), privacy funnel (PF), deterministic IB (DIB), conditional
entropy bottleneck (CEB), and conditional PF (CPF). We show that the CLUB model
generalizes all these problems as well as most other information-theoretic
privacy models. Then, we construct the deep variational CLUB (DVCLUB) models by
employing neural networks to parameterize variational approximations of the
associated information quantities. Building upon these information quantities,
we present unified objectives of the supervised and unsupervised DVCLUB models.
Leveraging the DVCLUB model in an unsupervised setup, we then connect it with
state-of-the-art generative models, such as variational auto-encoders (VAEs),
generative adversarial networks (GANs), as well as the Wasserstein GAN (WGAN),
Wasserstein auto-encoder (WAE), and adversarial auto-encoder (AAE) models
through the optimal transport (OT) problem. We then show that the DVCLUB model
can also be used in fair representation learning problems, where the goal is to
mitigate the undesired bias during the training phase of a machine learning
model. We conduct extensive quantitative experiments on colored-MNIST and
CelebA datasets, with a public implementation available, to evaluate and
analyze the CLUB model.
Related papers
- DiffSG: A Generative Solver for Network Optimization with Diffusion Model [75.27274046562806]
Diffusion generative models can consider a broader range of solutions and exhibit stronger generalization by learning parameters.
We propose a new framework, which leverages intrinsic distribution learning of diffusion generative models to learn high-quality solutions.
arXiv Detail & Related papers (2024-08-13T07:56:21Z) - Improving Generalization of Neural Vehicle Routing Problem Solvers Through the Lens of Model Architecture [9.244633039170186]
We propose a plug-and-play Entropy-based Scaling Factor (ESF) and a Distribution-Specific (DS) decoder.
ESF adjusts the attention weight pattern of the model towards familiar ones discovered during training when solving VRPs of varying sizes.
DS decoder explicitly models VRPs of multiple training distribution patterns through multiple auxiliary light decoders, expanding the model representation space.
arXiv Detail & Related papers (2024-06-10T09:03:17Z) - MONAL: Model Autophagy Analysis for Modeling Human-AI Interactions [11.972017738888825]
We propose Model Autophagy Analysis (MONAL) for large models' self-consumption explanation.
MONAL employs two distinct autophagous loops to elucidate the suppression of human-generated information in the exchange between human and AI systems.
We evaluate the capacities of generated models as both creators and disseminators of information.
arXiv Detail & Related papers (2024-02-17T13:02:54Z) - A PAC-Bayesian Perspective on the Interpolating Information Criterion [54.548058449535155]
We show how a PAC-Bayes bound is obtained for a general class of models, characterizing factors which influence performance in the interpolating regime.
We quantify how the test error for overparameterized models achieving effectively zero training error depends on the quality of the implicit regularization imposed by e.g. the combination of model, parameter-initialization scheme.
arXiv Detail & Related papers (2023-11-13T01:48:08Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Correlation Information Bottleneck: Towards Adapting Pretrained
Multimodal Models for Robust Visual Question Answering [63.87200781247364]
Correlation Information Bottleneck (CIB) seeks a tradeoff between compression and redundancy in representations.
We derive a tight theoretical upper bound for the mutual information between multimodal inputs and representations.
arXiv Detail & Related papers (2022-09-14T22:04:10Z) - De-Biasing Generative Models using Counterfactual Methods [0.0]
We propose a new decoder based framework named the Causal Counterfactual Generative Model (CCGM)
Our proposed method combines a causal latent space VAE model with specific modification to emphasize causal fidelity.
We explore how better disentanglement of causal learning and encoding/decoding generates higher causal intervention quality.
arXiv Detail & Related papers (2022-07-04T16:53:20Z) - On the Generalization and Adaption Performance of Causal Models [99.64022680811281]
Differentiable causal discovery has proposed to factorize the data generating process into a set of modules.
We study the generalization and adaption performance of such modular neural causal models.
Our analysis shows that the modular neural causal models outperform other models on both zero and few-shot adaptation in low data regimes.
arXiv Detail & Related papers (2022-06-09T17:12:32Z) - Application of Federated Learning in Building a Robust COVID-19 Chest
X-ray Classification Model [0.0]
Federated Learning (FL) helps AI models to generalize better without moving all the data to a central server.
We trained a deep learning model to solve a binary classification problem of predicting the presence or absence of COVID-19.
arXiv Detail & Related papers (2022-04-22T05:21:50Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - The Conditional Entropy Bottleneck [8.797368310561058]
We characterize failures of robust generalization as failures of accuracy or related metrics on a held-out set.
We propose the Minimum Necessary Information (MNI) criterion for evaluating the quality of a model.
In order to train models that perform well with respect to the MNI criterion, we present a new objective function, the Conditional Entropy Bottleneck (CEB)
We experimentally test our hypothesis by comparing the performance of CEB models with deterministic models and Variational Information Bottleneck (VIB) models on a variety of different datasets.
arXiv Detail & Related papers (2020-02-13T07:46:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.