Modeling All Response Surfaces in One for Conditional Search Spaces
- URL: http://arxiv.org/abs/2501.04260v2
- Date: Sun, 26 Jan 2025 10:26:44 GMT
- Title: Modeling All Response Surfaces in One for Conditional Search Spaces
- Authors: Jiaxing Li, Wei Liu, Chao Xue, Yibing Zhan, Xiaoxing Wang, Weifeng Liu, Dacheng Tao,
- Abstract summary: This paper proposes a novel approach to model the response surfaces of all subspaces in one.
We introduce an attention-based deep feature extractor, capable of projecting configurations with different structures from various subspaces into a unified feature space.
- Score: 69.90317997694218
- License:
- Abstract: Bayesian Optimization (BO) is a sample-efficient black-box optimizer commonly used in search spaces where hyperparameters are independent. However, in many practical AutoML scenarios, there will be dependencies among hyperparameters, forming a conditional search space, which can be partitioned into structurally distinct subspaces. The structure and dimensionality of hyperparameter configurations vary across these subspaces, challenging the application of BO. Some previous BO works have proposed solutions to develop multiple Gaussian Process models in these subspaces. However, these approaches tend to be inefficient as they require a substantial number of observations to guarantee each GP's performance and cannot capture relationships between hyperparameters across different subspaces. To address these issues, this paper proposes a novel approach to model the response surfaces of all subspaces in one, which can model the relationships between hyperparameters elegantly via a self-attention mechanism. Concretely, we design a structure-aware hyperparameter embedding to preserve the structural information. Then, we introduce an attention-based deep feature extractor, capable of projecting configurations with different structures from various subspaces into a unified feature space, where the response surfaces can be formulated using a single standard Gaussian Process. The empirical results on a simulation function, various real-world tasks, and HPO-B benchmark demonstrate that our proposed approach improves the efficacy and efficiency of BO within conditional search spaces.
Related papers
- Maintaining Structural Integrity in Parameter Spaces for Parameter Efficient Fine-tuning [78.39310274926535]
Adapting pre-trained foundation models for various downstream tasks has been prevalent in artificial intelligence.
To mitigate this, several fine-tuning techniques have been developed to update the pre-trained model weights in a more resource-efficient manner.
This paper introduces a generalized parameter-efficient fine-tuning framework, designed for various dimensional parameter space.
arXiv Detail & Related papers (2024-05-23T16:04:42Z) - Parameter Efficient Fine-tuning via Cross Block Orchestration for Segment Anything Model [81.55141188169621]
We equip PEFT with a cross-block orchestration mechanism to enable the adaptation of the Segment Anything Model (SAM) to various downstream scenarios.
We propose an intra-block enhancement module, which introduces a linear projection head whose weights are generated from a hyper-complex layer.
Our proposed approach consistently improves the segmentation performance significantly on novel scenarios with only around 1K additional parameters.
arXiv Detail & Related papers (2023-11-28T11:23:34Z) - Energy-efficient Task Adaptation for NLP Edge Inference Leveraging
Heterogeneous Memory Architectures [68.91874045918112]
adapter-ALBERT is an efficient model optimization for maximal data reuse across different tasks.
We demonstrate the advantage of mapping the model to a heterogeneous on-chip memory architecture by performing simulations on a validated NLP edge accelerator.
arXiv Detail & Related papers (2023-03-25T14:40:59Z) - Bayesian Optimization over High-Dimensional Combinatorial Spaces via
Dictionary-based Embeddings [36.60636056219264]
We consider the problem of optimizing black-box functions over high-dimensional spaces in science, engineering, and ML applications.
Key idea is to select a number of discrete structures from the input space and use them to define an ordinal embedding for high-dimensional structures.
We develop a principled approach based on binary wavelets to construct dictionaries for binary spaces, and propose a randomized construction method that generalizes to categorical spaces.
arXiv Detail & Related papers (2023-03-03T08:31:42Z) - Different Tunes Played with Equal Skill: Exploring a Unified
Optimization Subspace for Delta Tuning [95.72622659619445]
Delta tuning (DET) is deemed as the new paradigm for using pre-trained language models (PLMs)
Up to now, various DETs with distinct design elements have been proposed, achieving performance on par with fine-tuning.
arXiv Detail & Related papers (2022-10-24T14:57:35Z) - A local approach to parameter space reduction for regression and
classification tasks [0.0]
We propose a new method called local active subspaces (LAS), which explores the synergies of active subspaces with supervised clustering techniques.
LAS is particularly useful for the community working on surrogate modelling.
arXiv Detail & Related papers (2021-07-22T18:06:04Z) - High-Dimensional Bayesian Optimization with Sparse Axis-Aligned
Subspaces [14.03847432040056]
We argue that a surrogate model defined on sparse axis-aligned subspaces offer an attractive compromise between flexibility and parsimony.
We demonstrate that our approach, which relies on Hamiltonian Monte Carlo for inference, can rapidly identify sparse subspaces relevant to modeling the unknown objective function.
arXiv Detail & Related papers (2021-02-27T23:06:24Z) - MOFA: Modular Factorial Design for Hyperparameter Optimization [47.779983311833014]
MOdular FActorial Design (MOFA) is a novel HPO method that exploits evaluation results through factorial analysis.
We prove that the inference of MOFA achieves higher confidence than other sampling schemes.
Empirical results show that MOFA achieves better effectiveness and efficiency compared with state-of-the-art methods.
arXiv Detail & Related papers (2020-11-18T20:54:28Z) - High-Dimensional Bayesian Optimization via Nested Riemannian Manifolds [0.0]
We propose to exploit the geometry of non-Euclidean search spaces, which often arise in a variety of domains, to learn structure-preserving mappings.
Our approach features geometry-aware Gaussian processes that jointly learn a nested-manifold embedding and a representation of the objective function in the latent space.
arXiv Detail & Related papers (2020-10-21T11:24:11Z) - Misspecification-robust likelihood-free inference in high dimensions [13.934999364767918]
We introduce an extension of the popular Bayesian optimisation based approach to approximate discrepancy functions in a probabilistic manner.
Our approach achieves computational scalability for higher dimensional parameter spaces by using separate acquisition functions and discrepancies for each parameter.
The method successfully performs computationally efficient inference in a 100-dimensional space on canonical examples and compares favourably to existing modularised ABC methods.
arXiv Detail & Related papers (2020-02-21T16:06:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.