Stale Profile Matching
- URL: http://arxiv.org/abs/2401.17168v1
- Date: Tue, 30 Jan 2024 16:56:32 GMT
- Title: Stale Profile Matching
- Authors: Amir Ayupov and Maksim Panchenko and Sergey Pupyrev
- Abstract summary: Profile-guided optimizations rely on profile data for directing compilers to generate optimized code.
In practice, there is typically a gap between the profile collection and the release, which makes a portion of the profile invalid for optimizations.
We propose the first practical solution for utilizing profiles collected on binaries built from several revisions behind the release.
- Score: 1.1091463365252643
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Profile-guided optimizations rely on profile data for directing compilers to
generate optimized code. To achieve the maximum performance boost, profile data
needs to be collected on the same version of the binary that is being
optimized. In practice however, there is typically a gap between the profile
collection and the release, which makes a portion of the profile invalid for
optimizations. This phenomenon is known as profile staleness, and it is a
serious practical problem for data-center workloads both for compilers and
binary optimizers.
In this paper we thoroughly study the staleness problem and propose the first
practical solution for utilizing profiles collected on binaries built from
several revisions behind the release. Our algorithm is developed and
implemented in a mainstream open-source post-link optimizer, BOLT. An extensive
evaluation on a variety of standalone benchmarks and production services
indicates that the new method recovers up to $0.8$ of the maximum BOLT benefit,
even when most of the input profile data is stale and would have been discarded
by the optimizer otherwise.
Related papers
- Brevity is the soul of wit: Pruning long files for code generation [19.61423412870527]
We find that a simple--pruning long files--outperforms other methods in compute-limited regimes.
Our method can yield up to a 2x efficiency benefit in training (while matching performance) or a 3.5% absolute performance improvement on HumanEval.
arXiv Detail & Related papers (2024-06-29T13:08:24Z) - Discovering Preference Optimization Algorithms with and for Large Language Models [50.843710797024805]
offline preference optimization is a key method for enhancing and controlling the quality of Large Language Model (LLM) outputs.
We perform objective discovery to automatically discover new state-of-the-art preference optimization algorithms without (expert) human intervention.
Experiments demonstrate the state-of-the-art performance of DiscoPOP, a novel algorithm that adaptively blends logistic and exponential losses.
arXiv Detail & Related papers (2024-06-12T16:58:41Z) - Optimistic Optimization of Gaussian Process Samples [30.226274682578172]
A competing, computationally more efficient, global optimization framework is optimistic optimization, which exploits prior knowledge about the geometry of the search space in form of a dissimilarity function.
We argue that there is a new research domain between geometric and probabilistic search, i.e. methods that run drastically faster than traditional Bayesian optimization, while retaining some of the crucial functionality of Bayesian optimization.
arXiv Detail & Related papers (2022-09-02T09:06:24Z) - Profile Guided Optimization without Profiles: A Machine Learning
Approach [0.0]
Profile guided optimization is an effective technique for improving the optimization ability of compilers based on dynamic behavior.
We present a novel statistical approach to inferring branch probabilities that improves the performance of programs that are compiled without profile guided optimizations.
arXiv Detail & Related papers (2021-12-24T22:49:21Z) - Bayesian Algorithm Execution: Estimating Computable Properties of
Black-box Functions Using Mutual Information [78.78486761923855]
In many real world problems, we want to infer some property of an expensive black-box function f, given a budget of T function evaluations.
We present a procedure, InfoBAX, that sequentially chooses queries that maximize mutual information with respect to the algorithm's output.
On these problems, InfoBAX uses up to 500 times fewer queries to f than required by the original algorithm.
arXiv Detail & Related papers (2021-04-19T17:22:11Z) - Revisiting Bayesian Optimization in the light of the COCO benchmark [1.4467794332678539]
This article reports a large investigation about the effects on the performance of (Gaussian process based) BO of common and less common design choices.
The code developed for this study makes the new version (v2.1.1) of the R package DiceOptim available on CRAN.
arXiv Detail & Related papers (2021-03-30T19:45:18Z) - Learning to Optimize: A Primer and A Benchmark [94.29436694770953]
Learning to optimize (L2O) is an emerging approach that leverages machine learning to develop optimization methods.
This article is poised to be the first comprehensive survey and benchmark of L2O for continuous optimization.
arXiv Detail & Related papers (2021-03-23T20:46:20Z) - Stochastic Optimization with Laggard Data Pipelines [65.20044914532221]
We show that "dataechoed" extensions of common optimization methods exhibit provable improvements over their synchronous counterparts.
Specifically, we show that in convex optimization with minibatches, data echoing affords speedups on the curvature-dominated part of the convergence rate, while maintaining the optimal statistical rate.
arXiv Detail & Related papers (2020-10-26T14:55:31Z) - Bayesian Optimization with a Prior for the Optimum [41.41323474440455]
We introduce Bayesian Optimization with a Prior for the Optimum (BOPrO)
BOPrO allows users to inject their knowledge into the optimization process in the form of priors about which parts of the input space will yield the best performance.
We show that BOPrO is around 6.67x faster than state-of-the-art methods on a common suite of benchmarks.
arXiv Detail & Related papers (2020-06-25T17:49:24Z) - Global Optimization of Gaussian processes [52.77024349608834]
We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
arXiv Detail & Related papers (2020-05-21T20:59:11Z) - Incorporating Expert Prior in Bayesian Optimisation via Space Warping [54.412024556499254]
In big search spaces the algorithm goes through several low function value regions before reaching the optimum of the function.
One approach to subside this cold start phase is to use prior knowledge that can accelerate the optimisation.
In this paper, we represent the prior knowledge about the function optimum through a prior distribution.
The prior distribution is then used to warp the search space in such a way that space gets expanded around the high probability region of function optimum and shrinks around low probability region of optimum.
arXiv Detail & Related papers (2020-03-27T06:18:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.