Green Machine Learning via Augmented Gaussian Processes and
Multi-Information Source Optimization
- URL: http://arxiv.org/abs/2006.14233v1
- Date: Thu, 25 Jun 2020 08:04:48 GMT
- Title: Green Machine Learning via Augmented Gaussian Processes and
Multi-Information Source Optimization
- Authors: Antonio Candelieri, Riccardo Perego, Francesco Archetti
- Abstract summary: Strategy to drastically reduce computational time and energy consumed is to exploit the availability of different information sources.
An Augmented Gaussian Process method exploiting multiple information sources (namely, AGP-MISO) is proposed.
A novel acquisition function is defined according to the Augmented Gaussian Process.
- Score: 0.19116784879310028
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Searching for accurate Machine and Deep Learning models is a computationally
expensive and awfully energivorous process. A strategy which has been gaining
recently importance to drastically reduce computational time and energy
consumed is to exploit the availability of different information sources, with
different computational costs and different "fidelity", typically smaller
portions of a large dataset. The multi-source optimization strategy fits into
the scheme of Gaussian Process based Bayesian Optimization. An Augmented
Gaussian Process method exploiting multiple information sources (namely,
AGP-MISO) is proposed. The Augmented Gaussian Process is trained using only
"reliable" information among available sources. A novel acquisition function is
defined according to the Augmented Gaussian Process. Computational results are
reported related to the optimization of the hyperparameters of a Support Vector
Machine (SVM) classifier using two sources: a large dataset - the most
expensive one - and a smaller portion of it. A comparison with a traditional
Bayesian Optimization approach to optimize the hyperparameters of the SVM
classifier on the large dataset only is reported.
Related papers
- Computation-Aware Gaussian Processes: Model Selection And Linear-Time Inference [55.150117654242706]
We show that model selection for computation-aware GPs trained on 1.8 million data points can be done within a few hours on a single GPU.
As a result of this work, Gaussian processes can be trained on large-scale datasets without significantly compromising their ability to quantify uncertainty.
arXiv Detail & Related papers (2024-11-01T21:11:48Z) - Discovering Preference Optimization Algorithms with and for Large Language Models [50.843710797024805]
offline preference optimization is a key method for enhancing and controlling the quality of Large Language Model (LLM) outputs.
We perform objective discovery to automatically discover new state-of-the-art preference optimization algorithms without (expert) human intervention.
Experiments demonstrate the state-of-the-art performance of DiscoPOP, a novel algorithm that adaptively blends logistic and exponential losses.
arXiv Detail & Related papers (2024-06-12T16:58:41Z) - Optimizing the Optimal Weighted Average: Efficient Distributed Sparse Classification [50.406127962933915]
ACOWA allows an extra round of communication to achieve noticeably better approximation quality with minor runtime increases.
Results show that ACOWA obtains solutions that are more faithful to the empirical risk minimizer and attain substantially higher accuracy than other distributed algorithms.
arXiv Detail & Related papers (2024-06-03T19:43:06Z) - Sparsity-Aware Distributed Learning for Gaussian Processes with Linear
Multiple Kernel [22.23550794664218]
This paper presents a novel GP linear multiple kernel (LMK) and a generic sparsity-aware distributed learning framework.
The framework incorporates a quantized alternating direction method of multipliers (ADMM) for collaborative learning among multiple agents.
Experiments on diverse datasets demonstrate the superior prediction performance and efficiency of our proposed methods.
arXiv Detail & Related papers (2023-09-15T07:05:33Z) - Improved Distribution Matching for Dataset Condensation [91.55972945798531]
We propose a novel dataset condensation method based on distribution matching.
Our simple yet effective method outperforms most previous optimization-oriented methods with much fewer computational resources.
arXiv Detail & Related papers (2023-07-19T04:07:33Z) - Pre-training helps Bayesian optimization too [49.28382118032923]
We seek an alternative practice for setting functional priors.
In particular, we consider the scenario where we have data from similar functions that allow us to pre-train a tighter distribution a priori.
Our results show that our method is able to locate good hyper parameters at least 3 times more efficiently than the best competing methods.
arXiv Detail & Related papers (2022-07-07T04:42:54Z) - Fair and Green Hyperparameter Optimization via Multi-objective and
Multiple Information Source Bayesian Optimization [0.19116784879310028]
FanG-HPO uses subsets of the large dataset (aka information sources) to obtain cheap approximations of both accuracy and fairness.
Experiments consider two benchmark (fairness) datasets and two machine learning algorithms.
arXiv Detail & Related papers (2022-05-18T10:07:21Z) - Hyperparameter optimization of data-driven AI models on HPC systems [0.0]
This work is part of RAISE's work on data-driven use cases which leverages AI- and HPC cross-methods.
It is shown that in the case of Machine-Learned Particle reconstruction in High Energy Physics, the ASHA algorithm in combination with Bayesian optimization gives the largest performance increase per compute resources spent out of the investigated algorithms.
arXiv Detail & Related papers (2022-03-02T14:02:59Z) - Automatic tuning of hyper-parameters of reinforcement learning
algorithms using Bayesian optimization with behavioral cloning [0.0]
In reinforcement learning (RL), the information content of data gathered by the learning agent is dependent on the setting of many hyper- parameters.
In this work, a novel approach for autonomous hyper- parameter setting using Bayesian optimization is proposed.
Experiments reveal promising results compared to other manual tweaking and optimization-based approaches.
arXiv Detail & Related papers (2021-12-15T13:10:44Z) - Global Optimization of Gaussian processes [52.77024349608834]
We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
arXiv Detail & Related papers (2020-05-21T20:59:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.