Supermodular $\mf$-divergences and bounds on lossy compression and
generalization error with mutual $\mf$-information
- URL: http://arxiv.org/abs/2206.11042v1
- Date: Tue, 21 Jun 2022 09:17:06 GMT
- Title: Supermodular $\mf$-divergences and bounds on lossy compression and
generalization error with mutual $\mf$-information
- Authors: Saeed Masiha, Amin Gohari, Mohammad Hossein Yassaee
- Abstract summary: We introduce super-modular $mf$-divergences and provide three applications for them.
We provide a connection between the generalization error of algorithms with bounded input/output mutual $mf$-information and a generalized rate-distortion problem.
Our bound is based on a new lower bound on the rate-distortion function that strictly improves over previously best-known bounds.
- Score: 17.441807469515254
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we introduce super-modular $\mf$-divergences and provide three
applications for them: (i) we introduce Sanov's upper bound on the tail
probability of sum of independent random variables based on super-modular
$\mf$-divergence and show that our generalized Sanov's bound strictly improves
over ordinary one, (ii) we consider the lossy compression problem which studies
the set of achievable rates for a given distortion and code length. We extend
the rate-distortion function using mutual $\mf$-information and provide new and
strictly better bounds on achievable rates in the finite blocklength regime
using super-modular $\mf$-divergences, and (iii) we provide a connection
between the generalization error of algorithms with bounded input/output mutual
$\mf$-information and a generalized rate-distortion problem. This connection
allows us to bound the generalization error of learning algorithms using lower
bounds on the rate-distortion function. Our bound is based on a new lower bound
on the rate-distortion function that (for some examples) strictly improves over
previously best-known bounds. Moreover, super-modular $\mf$-divergences are
utilized to reduce the dimension of the problem and obtain single-letter
bounds.
Related papers
- Error Feedback under $(L_0,L_1)$-Smoothness: Normalization and Momentum [56.37522020675243]
We provide the first proof of convergence for normalized error feedback algorithms across a wide range of machine learning problems.
We show that due to their larger allowable stepsizes, our new normalized error feedback algorithms outperform their non-normalized counterparts on various tasks.
arXiv Detail & Related papers (2024-10-22T10:19:27Z) - Bridging the Gap Between General and Down-Closed Convex Sets in
Submodular Maximization [8.225819874406238]
We show that Mualem citemualem23re shows that this approach cannot smooth between down- and non-down-closed constraints.
In this work, we suggest novel offline and online algorithms based on a natural decomposition of the body into two distinct convex bodies.
We also empirically demonstrate the superiority of our proposed algorithms across three offline and two online applications.
arXiv Detail & Related papers (2024-01-17T14:56:42Z) - Multi-Grid Tensorized Fourier Neural Operator for High-Resolution PDEs [93.82811501035569]
We introduce a new data efficient and highly parallelizable operator learning approach with reduced memory requirement and better generalization.
MG-TFNO scales to large resolutions by leveraging local and global structures of full-scale, real-world phenomena.
We demonstrate superior performance on the turbulent Navier-Stokes equations where we achieve less than half the error with over 150x compression.
arXiv Detail & Related papers (2023-09-29T20:18:52Z) - Distributed Extra-gradient with Optimal Complexity and Communication
Guarantees [60.571030754252824]
We consider monotone variational inequality (VI) problems in multi-GPU settings where multiple processors/workers/clients have access to local dual vectors.
Extra-gradient, which is a de facto algorithm for monotone VI problems, has not been designed to be communication-efficient.
We propose a quantized generalized extra-gradient (Q-GenX), which is an unbiased and adaptive compression method tailored to solve VIs.
arXiv Detail & Related papers (2023-08-17T21:15:04Z) - Randomized Greedy Learning for Non-monotone Stochastic Submodular
Maximization Under Full-bandit Feedback [98.29086113546045]
We investigate the problem of unconstrained multi-armed bandits with full-bandit feedback and rewards for submodularity.
We show that RGL empirically outperforms other full-bandit variants in submodular and non-submodular settings.
arXiv Detail & Related papers (2023-02-02T18:52:14Z) - A Non-Asymptotic Moreau Envelope Theory for High-Dimensional Generalized
Linear Models [33.36787620121057]
We prove a new generalization bound that shows for any class of linear predictors in Gaussian space.
We use our finite-sample bound to directly recover the "optimistic rate" of Zhou et al. (2021)
We show that application of our bound generalization using localized Gaussian width will generally be sharp for empirical risk minimizers.
arXiv Detail & Related papers (2022-10-21T16:16:55Z) - Bregman divergence based em algorithm and its application to classical
and quantum rate distortion theory [61.12008553173672]
We address the minimization problem of the Bregman divergence between an exponential subfamily and a mixture subfamily in a Bregman divergence system.
We apply this algorithm to rate distortion and its variants including the quantum setting.
arXiv Detail & Related papers (2022-01-07T13:33:28Z) - Revisiting Modified Greedy Algorithm for Monotone Submodular
Maximization with a Knapsack Constraint [75.85952446237599]
We show that a modified greedy algorithm can achieve an approximation factor of $0.305$.
We derive a data-dependent upper bound on the optimum.
It can also be used to significantly improve the efficiency of such algorithms as branch and bound.
arXiv Detail & Related papers (2020-08-12T15:40:21Z) - Fast and Private Submodular and $k$-Submodular Functions Maximization
with Matroid Constraints [27.070004659301162]
We study the problem of maximizing monotone submodular functions subject to matroid constraints in the framework of differential privacy.
We give the first $frac12$-approximation algorithm that preserves $k$submodular functions subject to matroid constraints.
arXiv Detail & Related papers (2020-06-28T23:18:58Z) - Batch greedy maximization of non-submodular functions: Guarantees and
applications to experimental design [0.0]
We analyze greedys for cardinality constrained of non-submodular non-decreasing set functions.
Our theoretical guarantees are characterized by the combination of submodularity and supermodularity ratios.
arXiv Detail & Related papers (2020-06-03T18:58:06Z) - Streaming Submodular Maximization under a $k$-Set System Constraint [42.31117997337689]
We propose a novel framework that converts streaming for monotone submodular into streaming for non-monotone submodular.
We also propose the first algorithm for monotone submodular streaming subject to $k$ible $k$-set system constraints.
arXiv Detail & Related papers (2020-02-09T12:32:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.