Sample-Then-Optimize Batch Neural Thompson Sampling
- URL: http://arxiv.org/abs/2210.06850v1
- Date: Thu, 13 Oct 2022 09:01:58 GMT
- Title: Sample-Then-Optimize Batch Neural Thompson Sampling
- Authors: Zhongxiang Dai, Yao Shu, Bryan Kian Hsiang Low, Patrick Jaillet
- Abstract summary: We introduce two algorithms for black-box optimization based on the Thompson sampling (TS) policy.
To choose an input query, we only need to train an NN and then choose the query by maximizing the trained NN.
Our algorithms sidestep the need to invert the large parameter matrix yet still preserve the validity of the TS policy.
- Score: 50.800944138278474
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bayesian optimization (BO), which uses a Gaussian process (GP) as a surrogate
to model its objective function, is popular for black-box optimization.
However, due to the limitations of GPs, BO underperforms in some problems such
as those with categorical, high-dimensional or image inputs. To this end,
recent works have used the highly expressive neural networks (NNs) as the
surrogate model and derived theoretical guarantees using the theory of neural
tangent kernel (NTK). However, these works suffer from the limitations of the
requirement to invert an extremely large parameter matrix and the restriction
to the sequential (rather than batch) setting. To overcome these limitations,
we introduce two algorithms based on the Thompson sampling (TS) policy named
Sample-Then-Optimize Batch Neural TS (STO-BNTS) and STO-BNTS-Linear. To choose
an input query, we only need to train an NN (resp. a linear model) and then
choose the query by maximizing the trained NN (resp. linear model), which is
equivalently sampled from the GP posterior with the NTK as the kernel function.
As a result, our algorithms sidestep the need to invert the large parameter
matrix yet still preserve the validity of the TS policy. Next, we derive regret
upper bounds for our algorithms with batch evaluations, and use insights from
batch BO and NTK to show that they are asymptotically no-regret under certain
conditions. Finally, we verify their empirical effectiveness using practical
AutoML and reinforcement learning experiments.
Related papers
- Quantum Bayesian Optimization [64.58749619145908]
We introduce the quantum-Gaussian process-upper confidence bound (Q-GP-UCB) algorithm.
It is the first BO algorithm able to achieve a regret upper bound of O(polylog T), which is significantly smaller than its regret lower bound of Omega(sqrt(T)) in the classical setting.
Thanks to our novel analysis of the confidence ellipsoid, our Q-GP-UCB with the linear kernel achieves a smaller regret than the quantum linear UCB algorithm.
arXiv Detail & Related papers (2023-10-09T03:10:42Z) - Benign Overfitting in Deep Neural Networks under Lazy Training [72.28294823115502]
We show that when the data distribution is well-separated, DNNs can achieve Bayes-optimal test error for classification.
Our results indicate that interpolating with smoother functions leads to better generalization.
arXiv Detail & Related papers (2023-05-30T19:37:44Z) - Globally Optimal Training of Neural Networks with Threshold Activation
Functions [63.03759813952481]
We study weight decay regularized training problems of deep neural networks with threshold activations.
We derive a simplified convex optimization formulation when the dataset can be shattered at a certain layer of the network.
arXiv Detail & Related papers (2023-03-06T18:59:13Z) - Bayesian Kernelized Tensor Factorization as Surrogate for Bayesian
Optimization [13.896697187967545]
Kernel optimization (BO) primarily uses Gaussian processes (GP) as the key surrogate model.
In this paper, we propose to use Bayesian Factorization (BKTF) as a new surrogate model -- for BO in a $D$-dimensional product space.
BKTF offers a flexible and highly effective approach for characterizing complex functions with uncertainty quantification.
arXiv Detail & Related papers (2023-02-28T12:00:21Z) - Surrogate modeling for Bayesian optimization beyond a single Gaussian
process [62.294228304646516]
We propose a novel Bayesian surrogate model to balance exploration with exploitation of the search space.
To endow function sampling with scalability, random feature-based kernel approximation is leveraged per GP model.
To further establish convergence of the proposed EGP-TS to the global optimum, analysis is conducted based on the notion of Bayesian regret.
arXiv Detail & Related papers (2022-05-27T16:43:10Z) - Misspecified Gaussian Process Bandit Optimization [59.30399661155574]
Kernelized bandit algorithms have shown strong empirical and theoretical performance for this problem.
We introduce a emphmisspecified kernelized bandit setting where the unknown function can be $epsilon$--uniformly approximated by a function with a bounded norm in some Reproducing Kernel Hilbert Space (RKHS)
We show that our algorithm achieves optimal dependence on $epsilon$ with no prior knowledge of misspecification.
arXiv Detail & Related papers (2021-11-09T09:00:02Z) - An Empirical Study of Neural Kernel Bandits [17.92492092034792]
Research on neural kernels (NK) has recently established a correspondence between deep networks and GPs that take into account all the parameters of a NN.
We show that NK bandits achieve state-of-the-art performance on highly non-linear structured data.
arXiv Detail & Related papers (2021-11-05T15:06:05Z) - Neural Contextual Bandits without Regret [47.73483756447701]
We propose algorithms for contextual bandits harnessing neural networks to approximate the unknown reward function.
We show that our approach converges to the optimal policy at a $tildemathcalO(T-1/2d)$ rate, where $d$ is the dimension of the context.
arXiv Detail & Related papers (2021-07-07T11:11:34Z) - Exploring the Uncertainty Properties of Neural Networks' Implicit Priors
in the Infinite-Width Limit [47.324627920761685]
We use recent theoretical advances that characterize the function-space prior to an ensemble of infinitely-wide NNs as a Gaussian process.
This gives us a better understanding of the implicit prior NNs place on function space.
We also examine the calibration of previous approaches to classification with the NNGP.
arXiv Detail & Related papers (2020-10-14T18:41:54Z) - Improving the Backpropagation Algorithm with Consequentialism Weight
Updates over Mini-Batches [0.40611352512781856]
We show that it is possible to consider a multi-layer neural network as a stack of adaptive filters.
We introduce a better algorithm by predicting then emending the adverse consequences of the actions that take place in BP even before they happen.
Our experiments show the usefulness of our algorithm in the training of deep neural networks.
arXiv Detail & Related papers (2020-03-11T08:45:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.