Error Discovery by Clustering Influence Embeddings
- URL: http://arxiv.org/abs/2312.04712v1
- Date: Thu, 7 Dec 2023 21:42:55 GMT
- Title: Error Discovery by Clustering Influence Embeddings
- Authors: Fulton Wang, Julius Adebayo, Sarah Tan, Diego Garcia-Olano, Narine
Kokhlikyan
- Abstract summary: We present a method for identifying groups of test examples -- slices -- on which a model under-performs.
We formalize coherence as a key property that any slice discovery method should satisfy.
We derive a new slice discovery method, InfEmbed, which satisfies coherence by returning slices whose examples are influenced similarly by the training data.
- Score: 7.27282591214364
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a method for identifying groups of test examples -- slices -- on
which a model under-performs, a task now known as slice discovery. We formalize
coherence -- a requirement that erroneous predictions, within a slice, should
be wrong for the same reason -- as a key property that any slice discovery
method should satisfy. We then use influence functions to derive a new slice
discovery method, InfEmbed, which satisfies coherence by returning slices whose
examples are influenced similarly by the training data. InfEmbed is simple, and
consists of applying K-Means clustering to a novel representation we deem
influence embeddings. We show InfEmbed outperforms current state-of-the-art
methods on 2 benchmarks, and is effective for model debugging across several
case studies.
Related papers
- Error Slice Discovery via Manifold Compactness [47.57891946791078]
There is no proper metric of slice coherence without relying on extra information like predefined slice labels.
We propose manifold compactness, a coherence metric without reliance on extra information by incorporating the data geometry property into its design.
Then we develop Manifold Compactness based error Slice Discovery (MCSD), a novel algorithm that directly treats risk and coherence as the optimization objective.
arXiv Detail & Related papers (2025-01-31T11:02:07Z) - Exploring Query Efficient Data Generation towards Data-free Model Stealing in Hard Label Setting [38.755154033324374]
Data-free model stealing involves replicating the functionality of a target model into a substitute model without accessing the target model's structure, parameters, or training data.
This paper presents a new data-free model stealing approach called Query Efficient Data Generation (textbfQEDG)
We introduce two distinct loss functions to ensure the generation of sufficient samples that closely and uniformly align with the target model's decision boundary.
arXiv Detail & Related papers (2024-12-18T03:03:15Z) - Scalable Influence and Fact Tracing for Large Language Model Pretraining [14.598556308631018]
Training data attribution (TDA) methods aim to attribute model outputs back to specific training examples.
This paper refines existing gradient-based methods to work effectively at scale.
arXiv Detail & Related papers (2024-10-22T20:39:21Z) - Simple Ingredients for Offline Reinforcement Learning [86.1988266277766]
offline reinforcement learning algorithms have proven effective on datasets highly connected to the target downstream task.
We show that existing methods struggle with diverse data: their performance considerably deteriorates as data collected for related but different tasks is simply added to the offline buffer.
We show that scale, more than algorithmic considerations, is the key factor influencing performance.
arXiv Detail & Related papers (2024-03-19T18:57:53Z) - Comprehensive OOD Detection Improvements [46.46252643210326]
Out-of-distribution (OOD) detection methods have been created for this task.
We employ dimensionality reduction on feature embeddings in representation-based methods for both time speedups and improved performance.
We demonstrate the effectiveness of our methods on the OpenOODv1.5 benchmark framework.
arXiv Detail & Related papers (2024-01-18T18:05:35Z) - Where Does My Model Underperform? A Human Evaluation of Slice Discovery
Algorithms [24.127380328812855]
New slice discovery algorithms aim to group together coherent and high-error subsets of data.
We show 40 slices output by two state-of-the-art slice discovery algorithms to users, and ask them to form hypotheses about an object detection model.
Our results provide positive evidence that these tools provide some benefit over a naive baseline, and also shed light on challenges faced by users during the hypothesis formation step.
arXiv Detail & Related papers (2023-06-13T22:44:53Z) - Generalization Properties of Retrieval-based Models [50.35325326050263]
Retrieval-based machine learning methods have enjoyed success on a wide range of problems.
Despite growing literature showcasing the promise of these models, the theoretical underpinning for such models remains underexplored.
We present a formal treatment of retrieval-based models to characterize their generalization ability.
arXiv Detail & Related papers (2022-10-06T00:33:01Z) - Smoothed Embeddings for Certified Few-Shot Learning [63.68667303948808]
We extend randomized smoothing to few-shot learning models that map inputs to normalized embeddings.
Our results are confirmed by experiments on different datasets.
arXiv Detail & Related papers (2022-02-02T18:19:04Z) - Label-Descriptive Patterns and their Application to Characterizing
Classification Errors [31.272875287136426]
State-of-the-art deep learning methods achieve human-like performance on many tasks, but make errors nevertheless.
Characterizing these errors in easily interpretable terms gives insight into whether a model is prone to making systematic errors, but also gives a way to act and improve the model.
In this paper we propose a method that allows us to do so for arbitrary classifiers by mining a small set of patterns that together succinctly describe the input data that is partitioned according to correctness of prediction.
arXiv Detail & Related papers (2021-10-18T19:42:21Z) - Attentional Prototype Inference for Few-Shot Segmentation [128.45753577331422]
We propose attentional prototype inference (API), a probabilistic latent variable framework for few-shot segmentation.
We define a global latent variable to represent the prototype of each object category, which we model as a probabilistic distribution.
We conduct extensive experiments on four benchmarks, where our proposal obtains at least competitive and often better performance than state-of-the-art prototype-based methods.
arXiv Detail & Related papers (2021-05-14T06:58:44Z) - Contrastive Prototype Learning with Augmented Embeddings for Few-Shot
Learning [58.2091760793799]
We propose a novel contrastive prototype learning with augmented embeddings (CPLAE) model.
With a class prototype as an anchor, CPL aims to pull the query samples of the same class closer and those of different classes further away.
Extensive experiments on several benchmarks demonstrate that our proposed CPLAE achieves new state-of-the-art.
arXiv Detail & Related papers (2021-01-23T13:22:44Z) - Pair the Dots: Jointly Examining Training History and Test Stimuli for
Model Interpretability [44.60486560836836]
Any prediction from a model is made by a combination of learning history and test stimuli.
Existing methods to interpret a model's predictions are only able to capture a single aspect of either test stimuli or learning history.
We propose an efficient and differentiable approach to make it feasible to interpret a model's prediction by jointly examining training history and test stimuli.
arXiv Detail & Related papers (2020-10-14T10:45:01Z) - Understanding Classifier Mistakes with Generative Models [88.20470690631372]
Deep neural networks are effective on supervised learning tasks, but have been shown to be brittle.
In this paper, we leverage generative models to identify and characterize instances where classifiers fail to generalize.
Our approach is agnostic to class labels from the training set which makes it applicable to models trained in a semi-supervised way.
arXiv Detail & Related papers (2020-10-05T22:13:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.