Conflict-Averse Gradient Optimization of Ensembles for Effective Offline
Model-Based Optimization
- URL: http://arxiv.org/abs/2303.17934v1
- Date: Fri, 31 Mar 2023 10:00:27 GMT
- Title: Conflict-Averse Gradient Optimization of Ensembles for Effective Offline
Model-Based Optimization
- Authors: Sathvik Kolli
- Abstract summary: We evaluate two algorithms for combining gradient information: multiple gradient descent algorithm (MGDA) and conflict-averse gradient descent (CAGrad)
Our results suggest that MGDA and CAGrad strike a desirable balance between conservatism and optimality and can help robustify data-driven offline MBO without compromising optimality of designs.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Data-driven offline model-based optimization (MBO) is an established
practical approach to black-box computational design problems for which the
true objective function is unknown and expensive to query. However, the
standard approach which optimizes designs against a learned proxy model of the
ground truth objective can suffer from distributional shift. Specifically, in
high-dimensional design spaces where valid designs lie on a narrow manifold,
the standard approach is susceptible to producing out-of-distribution, invalid
designs that "fool" the learned proxy model into outputting a high value. Using
an ensemble rather than a single model as the learned proxy can help mitigate
distribution shift, but naive formulations for combining gradient information
from the ensemble, such as minimum or mean gradient, are still suboptimal and
often hampered by non-convergent behavior.
In this work, we explore alternate approaches for combining gradient
information from the ensemble that are robust to distribution shift without
compromising optimality of the produced designs. More specifically, we explore
two functions, formulated as convex optimization problems, for combining
gradient information: multiple gradient descent algorithm (MGDA) and
conflict-averse gradient descent (CAGrad). We evaluate these algorithms on a
diverse set of five computational design tasks. We compare performance of
ensemble MBO with MGDA and ensemble MBO with CAGrad with three naive baseline
algorithms: (a) standard single-model MBO, (b) ensemble MBO with mean gradient,
and (c) ensemble MBO with minimum gradient.
Our results suggest that MGDA and CAGrad strike a desirable balance between
conservatism and optimality and can help robustify data-driven offline MBO
without compromising optimality of designs.
Related papers
- Offline Model-Based Optimization by Learning to Rank [26.21886715050762]
We argue that regression models trained with mean squared error (MSE) are not well-aligned with the primary goal of offline model-based optimization.
We propose learning a ranking-based model that leverages learning to rank techniques to prioritize promising designs based on their relative scores.
arXiv Detail & Related papers (2024-10-15T11:15:03Z) - Functional Graphical Models: Structure Enables Offline Data-Driven Optimization [111.28605744661638]
We show how structure can enable sample-efficient data-driven optimization.
We also present a data-driven optimization algorithm that infers the FGM structure itself.
arXiv Detail & Related papers (2024-01-08T22:33:14Z) - ROMO: Retrieval-enhanced Offline Model-based Optimization [14.277672372460785]
Data-driven black-box model-based optimization (MBO) problems arise in a number of practical application scenarios.
We propose retrieval-enhanced offline model-based optimization (ROMO)
ROMO is simple to implement and outperforms state-of-the-art approaches in the CoMBO setting.
arXiv Detail & Related papers (2023-10-11T15:04:33Z) - Model-based Causal Bayesian Optimization [78.120734120667]
We propose model-based causal Bayesian optimization (MCBO)
MCBO learns a full system model instead of only modeling intervention-reward pairs.
Unlike in standard Bayesian optimization, our acquisition function cannot be evaluated in closed form.
arXiv Detail & Related papers (2022-11-18T14:28:21Z) - Spike-and-Slab Generalized Additive Models and Scalable Algorithms for
High-Dimensional Data [0.0]
We propose hierarchical generalized additive models (GAMs) to accommodate high-dimensional data.
We consider the smoothing penalty for proper shrinkage of curve and separation of smoothing function linear and nonlinear spaces.
Two and deterministic algorithms, EM-Coordinate Descent and EM-Iterative Weighted Least Squares, are developed for different utilities.
arXiv Detail & Related papers (2021-10-27T14:11:13Z) - Conservative Objective Models for Effective Offline Model-Based
Optimization [78.19085445065845]
Computational design problems arise in a number of settings, from synthetic biology to computer architectures.
We propose a method that learns a model of the objective function that lower bounds the actual value of the ground-truth objective on out-of-distribution inputs.
COMs are simple to implement and outperform a number of existing methods on a wide range of MBO problems.
arXiv Detail & Related papers (2021-07-14T17:55:28Z) - A Bi-Level Framework for Learning to Solve Combinatorial Optimization on
Graphs [91.07247251502564]
We propose a hybrid approach to combine the best of the two worlds, in which a bi-level framework is developed with an upper-level learning method to optimize the graph.
Such a bi-level approach simplifies the learning on the original hard CO and can effectively mitigate the demand for model capacity.
arXiv Detail & Related papers (2021-06-09T09:18:18Z) - Zeroth-Order Hybrid Gradient Descent: Towards A Principled Black-Box
Optimization Framework [100.36569795440889]
This work is on the iteration of zero-th-order (ZO) optimization which does not require first-order information.
We show that with a graceful design in coordinate importance sampling, the proposed ZO optimization method is efficient both in terms of complexity as well as as function query cost.
arXiv Detail & Related papers (2020-12-21T17:29:58Z) - Cogradient Descent for Bilinear Optimization [124.45816011848096]
We introduce a Cogradient Descent algorithm (CoGD) to address the bilinear problem.
We solve one variable by considering its coupling relationship with the other, leading to a synchronous gradient descent.
Our algorithm is applied to solve problems with one variable under the sparsity constraint.
arXiv Detail & Related papers (2020-06-16T13:41:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.