Batch Multi-Fidelity Active Learning with Budget Constraints
- URL: http://arxiv.org/abs/2210.12704v1
- Date: Sun, 23 Oct 2022 11:39:56 GMT
- Title: Batch Multi-Fidelity Active Learning with Budget Constraints
- Authors: Shibo Li, Jeff M. Phillips, Xin Yu, Robert M. Kirby, and Shandian Zhe
- Abstract summary: Batch Multi-Fidelity Active Learning with Budget Constraints (BMFAL-BC)
We propose a novel batch acquisition function that measures the mutual information between a batch of multi-fidelity queries and the target function.
We show the advantage of our method in several computational physics and engineering applications.
- Score: 37.420149663263835
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning functions with high-dimensional outputs is critical in many
applications, such as physical simulation and engineering design. However,
collecting training examples for these applications is often costly, e.g. by
running numerical solvers. The recent work (Li et al., 2022) proposes the first
multi-fidelity active learning approach for high-dimensional outputs, which can
acquire examples at different fidelities to reduce the cost while improving the
learning performance. However, this method only queries at one pair of fidelity
and input at a time, and hence has a risk to bring in strongly correlated
examples to reduce the learning efficiency. In this paper, we propose Batch
Multi-Fidelity Active Learning with Budget Constraints (BMFAL-BC), which can
promote the diversity of training examples to improve the benefit-cost ratio,
while respecting a given budget constraint for batch queries. Hence, our method
can be more practically useful. Specifically, we propose a novel batch
acquisition function that measures the mutual information between a batch of
multi-fidelity queries and the target function, so as to penalize highly
correlated queries and encourages diversity. The optimization of the batch
acquisition function is challenging in that it involves a combinatorial search
over many fidelities while subject to the budget constraint. To address this
challenge, we develop a weighted greedy algorithm that can sequentially
identify each (fidelity, input) pair, while achieving a near $(1 -
1/e)$-approximation of the optimum. We show the advantage of our method in
several computational physics and engineering applications.
Related papers
- Optimization by Parallel Quasi-Quantum Annealing with Gradient-Based Sampling [0.0]
This study proposes a different approach that integrates gradient-based update through continuous relaxation, combined with Quasi-Quantum Annealing (QQA)
Numerical experiments demonstrate that our method is a competitive general-purpose solver, achieving performance comparable to iSCO and learning-based solvers.
arXiv Detail & Related papers (2024-09-02T12:55:27Z) - Multi-Resolution Active Learning of Fourier Neural Operators [33.63483360957646]
We propose Multi-Resolution Active learning of FNO (MRA-FNO), which can dynamically select the input functions and resolutions to lower the data cost as much as possible.
Specifically, we propose a probabilistic multi-resolution FNO and use ensemble Monte-Carlo to develop an effective posterior inference algorithm.
We have shown the advantage of our method in several benchmark operator learning tasks.
arXiv Detail & Related papers (2023-09-29T04:41:27Z) - BatchGFN: Generative Flow Networks for Batch Active Learning [80.73649229919454]
BatchGFN is a novel approach for pool-based active learning that uses generative flow networks to sample sets of data points proportional to a batch reward.
We show our approach enables principled sampling near-optimal utility batches at inference time with a single forward pass per point in the batch in toy regression problems.
arXiv Detail & Related papers (2023-06-26T20:41:36Z) - OverPrompt: Enhancing ChatGPT through Efficient In-Context Learning [49.38867353135258]
We propose OverPrompt, leveraging the in-context learning capability of LLMs to handle multiple task inputs.
Our experiments show that OverPrompt can achieve cost-efficient zero-shot classification without causing significant detriment to task performance.
arXiv Detail & Related papers (2023-05-24T10:08:04Z) - Multi-task Bias-Variance Trade-off Through Functional Constraints [102.64082402388192]
Multi-task learning aims to acquire a set of functions that perform well for diverse tasks.
In this paper we draw intuition from the two extreme learning scenarios -- a single function for all tasks, and a task-specific function that ignores the other tasks.
We introduce a constrained learning formulation that enforces domain specific solutions to a central function.
arXiv Detail & Related papers (2022-10-27T16:06:47Z) - Mitigating Gradient Bias in Multi-objective Learning: A Provably Convergent Stochastic Approach [38.76462300149459]
We develop a Multi-objective Correction (MoCo) method for multi-objective gradient optimization.
The unique feature of our method is that it can guarantee convergence without increasing the non fairness gradient.
arXiv Detail & Related papers (2022-10-23T05:54:26Z) - Batch Active Learning at Scale [39.26441165274027]
Batch active learning, which adaptively issues batched queries to a labeling oracle, is a common approach for addressing this problem.
In this work, we analyze an efficient active learning algorithm, which focuses on the large batch setting.
We show that our sampling method, which combines notions of uncertainty and diversity, easily scales to batch sizes (100K-1M) several orders of magnitude larger than used in previous studies.
arXiv Detail & Related papers (2021-07-29T18:14:05Z) - Efficient Model-Based Multi-Agent Mean-Field Reinforcement Learning [89.31889875864599]
We propose an efficient model-based reinforcement learning algorithm for learning in multi-agent systems.
Our main theoretical contributions are the first general regret bounds for model-based reinforcement learning for MFC.
We provide a practical parametrization of the core optimization problem.
arXiv Detail & Related papers (2021-07-08T18:01:02Z) - Deep Multi-Fidelity Active Learning of High-dimensional Outputs [17.370056935194786]
We develop a deep neural network-based multi-fidelity model for learning with high-dimensional outputs.
We then propose a mutual information-based acquisition function that extends the predictive entropy principle.
We show the advantage of our method in several applications of computational physics and engineering design.
arXiv Detail & Related papers (2020-12-02T00:02:31Z) - Multi-task Supervised Learning via Cross-learning [102.64082402388192]
We consider a problem known as multi-task learning, consisting of fitting a set of regression functions intended for solving different tasks.
In our novel formulation, we couple the parameters of these functions, so that they learn in their task specific domains while staying close to each other.
This facilitates cross-fertilization in which data collected across different domains help improving the learning performance at each other task.
arXiv Detail & Related papers (2020-10-24T21:35:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.