Detecting and Understanding Real-World Differential Performance Bugs in
Machine Learning Libraries
- URL: http://arxiv.org/abs/2006.01991v1
- Date: Wed, 3 Jun 2020 00:23:06 GMT
- Title: Detecting and Understanding Real-World Differential Performance Bugs in
Machine Learning Libraries
- Authors: Saeid Tizpaz-Niari and Pavol Cern\'y and Ashutosh Trivedi
- Abstract summary: We find inputs for which the performance varies widely, despite having the same size.
We compare the performance of not only single inputs, but of classes of inputs, where each class has similar inputs parameterized by their size.
Importantly, we also provide an explanation for why the performance differs in a form that can be readily used to fix a performance bug.
- Score: 2.879036956042183
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Programming errors that degrade the performance of systems are widespread,
yet there is little tool support for analyzing these bugs. We present a method
based on differential performance analysis---we find inputs for which the
performance varies widely, despite having the same size. To ensure that the
differences in the performance are robust (i.e. hold also for large inputs), we
compare the performance of not only single inputs, but of classes of inputs,
where each class has similar inputs parameterized by their size. Thus, each
class is represented by a performance function from the input size to
performance. Importantly, we also provide an explanation for why the
performance differs in a form that can be readily used to fix a performance
bug.
The two main phases in our method are discovery with fuzzing and explanation
with decision tree classifiers, each of which is supported by clustering.
First, we propose an evolutionary fuzzing algorithm to generate inputs. For
this fuzzing task, the unique challenge is that we not only need the input
class with the worst performance, but rather a set of classes exhibiting
differential performance. We use clustering to merge similar input classes
which significantly improves the efficiency of our fuzzer. Second, we explain
the differential performance in terms of program inputs and internals. We adapt
discriminant learning approaches with clustering and decision trees to localize
suspicious code regions.
We applied our techniques to a set of applications. On a set of
micro-benchmarks, we show that our approach outperforms state-of-the-art
fuzzers in finding inputs to characterize the differential performance. On a
set of case-studies, we discover and explain multiple performance bugs in
popular machine learning frameworks. Four of these bugs, reported first in this
paper, have since been fixed by the developers.
Related papers
- Adapting Vision-Language Models to Open Classes via Test-Time Prompt Tuning [50.26965628047682]
Adapting pre-trained models to open classes is a challenging problem in machine learning.
In this paper, we consider combining the advantages of both and come up with a test-time prompt tuning approach.
Our proposed method outperforms all comparison methods on average considering both base and new classes.
arXiv Detail & Related papers (2024-08-29T12:34:01Z) - Multiclass Learning from Noisy Labels for Non-decomposable Performance Measures [15.358504449550013]
We design algorithms to learn from noisy labels for two broad classes of non-decomposable performance measures.
In both cases, we develop noise-corrected versions of the algorithms under the widely studied class-conditional noise models.
Our experiments demonstrate the effectiveness of our algorithms in handling label noise.
arXiv Detail & Related papers (2024-02-01T23:03:53Z) - Convolutional autoencoder-based multimodal one-class classification [80.52334952912808]
One-class classification refers to approaches of learning using data from a single class only.
We propose a deep learning one-class classification method suitable for multimodal data.
arXiv Detail & Related papers (2023-09-25T12:31:18Z) - Learning Context-aware Classifier for Semantic Segmentation [88.88198210948426]
In this paper, contextual hints are exploited via learning a context-aware classifier.
Our method is model-agnostic and can be easily applied to generic segmentation models.
With only negligible additional parameters and +2% inference time, decent performance gain has been achieved on both small and large models.
arXiv Detail & Related papers (2023-03-21T07:00:35Z) - Unified Functional Hashing in Automatic Machine Learning [58.77232199682271]
We show that large efficiency gains can be obtained by employing a fast unified functional hash.
Our hash is "functional" in that it identifies equivalent candidates even if they were represented or coded differently.
We show dramatic improvements on multiple AutoML domains, including neural architecture search and algorithm discovery.
arXiv Detail & Related papers (2023-02-10T18:50:37Z) - Class-Incremental Learning: A Survey [84.30083092434938]
Class-Incremental Learning (CIL) enables the learner to incorporate the knowledge of new classes incrementally.
CIL tends to catastrophically forget the characteristics of former ones, and its performance drastically degrades.
We provide a rigorous and unified evaluation of 17 methods in benchmark image classification tasks to find out the characteristics of different algorithms.
arXiv Detail & Related papers (2023-02-07T17:59:05Z) - Counterfactual Explanations for Oblique Decision Trees: Exact, Efficient
Algorithms [0.0]
We consider counterfactual explanations, the problem of minimally adjusting features in a source input instance so that it is classified as a target class under a given classification.
This has become a topic of recent interest as a way to query a trained model and suggest possible actions to overturn its decision.
arXiv Detail & Related papers (2021-03-01T16:04:33Z) - Unsupervised Feature Learning for Event Data: Direct vs Inverse Problem
Formulation [53.850686395708905]
Event-based cameras record an asynchronous stream of per-pixel brightness changes.
In this paper, we focus on single-layer architectures for representation learning from event data.
We show improvements of up to 9 % in the recognition accuracy compared to the state-of-the-art methods.
arXiv Detail & Related papers (2020-09-23T10:40:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.