Data-Driven Optimization of Directed Information over Discrete Alphabets
- URL: http://arxiv.org/abs/2301.00621v1
- Date: Mon, 2 Jan 2023 12:25:40 GMT
- Title: Data-Driven Optimization of Directed Information over Discrete Alphabets
- Authors: Dor Tsur, Ziv Aharoni, Ziv Goldfeld and Haim Permuter
- Abstract summary: Directed information (DI) is a fundamental measure for the study and analysis of sequential analytic models.
We propose a novel estimation-optimization framework for DI over discrete input spaces.
- Score: 15.372626012233736
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Directed information (DI) is a fundamental measure for the study and analysis
of sequential stochastic models. In particular, when optimized over input
distributions it characterizes the capacity of general communication channels.
However, analytic computation of DI is typically intractable and existing
optimization techniques over discrete input alphabets require knowledge of the
channel model, which renders them inapplicable when only samples are available.
To overcome these limitations, we propose a novel estimation-optimization
framework for DI over discrete input spaces. We formulate DI optimization as a
Markov decision process and leverage reinforcement learning techniques to
optimize a deep generative model of the input process probability mass function
(PMF). Combining this optimizer with the recently developed DI neural
estimator, we obtain an end-to-end estimation-optimization algorithm which is
applied to estimating the (feedforward and feedback) capacity of various
discrete channels with memory. Furthermore, we demonstrate how to use the
optimized PMF model to (i) obtain theoretical bounds on the feedback capacity
of unifilar finite-state channels; and (ii) perform probabilistic shaping of
constellations in the peak power-constrained additive white Gaussian noise
channel.
Related papers
- The Nature of Mathematical Modeling and Probabilistic Optimization Engineering in Generative AI [0.0]
We explore and discuss some potential further enhancement for current state of the art methods for some key underlying technologies of generative AI models.
In particular, we present an optimal solution for sub-word encoding (SWE) based on similar initial settings as that of byte-pair encoding (BPE) algorithm.
arXiv Detail & Related papers (2024-10-24T05:29:20Z) - Differentially Private Optimization with Sparse Gradients [60.853074897282625]
We study differentially private (DP) optimization problems under sparsity of individual gradients.
Building on this, we obtain pure- and approximate-DP algorithms with almost optimal rates for convex optimization with sparse gradients.
arXiv Detail & Related papers (2024-04-16T20:01:10Z) - Enhancing Gaussian Process Surrogates for Optimization and Posterior Approximation via Random Exploration [2.984929040246293]
novel noise-free Bayesian optimization strategies that rely on a random exploration step to enhance the accuracy of Gaussian process surrogate models.
New algorithms retain the ease of implementation of the classical GP-UCB, but an additional exploration step facilitates their convergence.
arXiv Detail & Related papers (2024-01-30T14:16:06Z) - Simulation Based Bayesian Optimization [0.6526824510982799]
This paper introduces Simulation Based Bayesian Optimization (SBBO) as a novel approach to optimizing acquisition functions.
SBBO allows the use of surrogate models tailored for spaces with discrete variables.
We demonstrate empirically the effectiveness of SBBO method using various choices of surrogate models.
arXiv Detail & Related papers (2024-01-19T16:56:11Z) - Functional Graphical Models: Structure Enables Offline Data-Driven Optimization [111.28605744661638]
We show how structure can enable sample-efficient data-driven optimization.
We also present a data-driven optimization algorithm that infers the FGM structure itself.
arXiv Detail & Related papers (2024-01-08T22:33:14Z) - Optimization of Annealed Importance Sampling Hyperparameters [77.34726150561087]
Annealed Importance Sampling (AIS) is a popular algorithm used to estimates the intractable marginal likelihood of deep generative models.
We present a parameteric AIS process with flexible intermediary distributions and optimize the bridging distributions to use fewer number of steps for sampling.
We assess the performance of our optimized AIS for marginal likelihood estimation of deep generative models and compare it to other estimators.
arXiv Detail & Related papers (2022-09-27T07:58:25Z) - Approximate Bayesian Optimisation for Neural Networks [6.921210544516486]
A body of work has been done to automate machine learning algorithm to highlight the importance of model choice.
The necessity to solve the analytical tractability and the computational feasibility in a idealistic fashion enables to ensure the efficiency and the applicability.
arXiv Detail & Related papers (2021-08-27T19:03:32Z) - Offline Model-Based Optimization via Normalized Maximum Likelihood
Estimation [101.22379613810881]
We consider data-driven optimization problems where one must maximize a function given only queries at a fixed set of points.
This problem setting emerges in many domains where function evaluation is a complex and expensive process.
We propose a tractable approximation that allows us to scale our method to high-capacity neural network models.
arXiv Detail & Related papers (2021-02-16T06:04:27Z) - An AI-Assisted Design Method for Topology Optimization Without
Pre-Optimized Training Data [68.8204255655161]
An AI-assisted design method based on topology optimization is presented, which is able to obtain optimized designs in a direct way.
Designs are provided by an artificial neural network, the predictor, on the basis of boundary conditions and degree of filling as input data.
arXiv Detail & Related papers (2020-12-11T14:33:27Z) - Global Optimization of Gaussian processes [52.77024349608834]
We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
arXiv Detail & Related papers (2020-05-21T20:59:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.