XPASC: Measuring Generalization in Weak Supervision
- URL: http://arxiv.org/abs/2206.01444v1
- Date: Fri, 3 Jun 2022 08:16:50 GMT
- Title: XPASC: Measuring Generalization in Weak Supervision
- Authors: Luisa M\"arz, Ehsaneddin Asgari, Fabienne Braune, Franziska
Zimmermann, Benjamin Roth
- Abstract summary: We introduce XPASC (eXPlainability-Association SCore) for measuring the generalization of a model trained with a weakly supervised dataset.
We use XPASC to analyze KnowMAN, an adversarial architecture intended to control the degree of generalization from the labeling functions.
- Score: 5.197057543520865
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Weak supervision is leveraged in a wide range of domains and tasks due to its
ability to create massive amounts of labeled data, requiring only little manual
effort. Standard approaches use labeling functions to specify signals that are
relevant for the labeling. It has been conjectured that weakly supervised
models over-rely on those signals and as a result suffer from overfitting. To
verify this assumption, we introduce a novel method, XPASC
(eXPlainability-Association SCore), for measuring the generalization of a model
trained with a weakly supervised dataset. Considering the occurrences of
features, classes and labeling functions in a dataset, XPASC takes into account
the relevance of each feature for the predictions of the model as well as the
associations of the feature with the class and the labeling function,
respectively. The association in XPASC can be measured in two variants:
XPASC-CHI SQAURE measures associations relative to their statistical
significance, while XPASC-PPMI measures association strength more generally.
We use XPASC to analyze KnowMAN, an adversarial architecture intended to
control the degree of generalization from the labeling functions and thus to
mitigate the problem of overfitting. On one hand, we show that KnowMAN is able
to control the degree of generalization through a hyperparameter. On the other
hand, results and qualitative analysis show that generalization and performance
do not relate one-to-one, and that the highest degree of generalization does
not necessarily imply the best performance. Therefore methods that allow for
controlling the amount of generalization can achieve the right degree of benign
overfitting. Our contributions in this study are i) the XPASC score to measure
generalization in weakly-supervised models, ii) evaluation of XPASC across
datasets and models and iii) the release of the XPASC implementation.
Related papers
- DECIDER: Leveraging Foundation Model Priors for Improved Model Failure Detection and Explanation [18.77296551727931]
We propose DECIDER, a novel approach that leverages priors from large language models (LLMs) and vision-language models (VLMs) to detect failures in image models.
DECIDER consistently achieves state-of-the-art failure detection performance, significantly outperforming baselines in terms of the overall Matthews correlation coefficient.
arXiv Detail & Related papers (2024-08-01T07:08:11Z) - Prospector Heads: Generalized Feature Attribution for Large Models & Data [82.02696069543454]
We introduce prospector heads, an efficient and interpretable alternative to explanation-based attribution methods.
We demonstrate how prospector heads enable improved interpretation and discovery of class-specific patterns in input data.
arXiv Detail & Related papers (2024-02-18T23:01:28Z) - Cross-functional Analysis of Generalisation in Behavioural Learning [4.0810783261728565]
We introduce BeLUGA, an analysis method for evaluating behavioural learning considering generalisation across dimensions of different levels.
An aggregate score measures generalisation to unseen functionalities (or overfitting)
arXiv Detail & Related papers (2023-05-22T11:54:19Z) - Leveraging Instance Features for Label Aggregation in Programmatic Weak
Supervision [75.1860418333995]
Programmatic Weak Supervision (PWS) has emerged as a widespread paradigm to synthesize training labels efficiently.
The core component of PWS is the label model, which infers true labels by aggregating the outputs of multiple noisy supervision sources as labeling functions.
Existing statistical label models typically rely only on the outputs of LF, ignoring the instance features when modeling the underlying generative process.
arXiv Detail & Related papers (2022-10-06T07:28:53Z) - Calibrated Feature Decomposition for Generalizable Person
Re-Identification [82.64133819313186]
Calibrated Feature Decomposition (CFD) module focuses on improving the generalization capacity for person re-identification.
A calibrated-and-standardized Batch normalization (CSBN) is designed to learn calibrated person representation.
arXiv Detail & Related papers (2021-11-27T17:12:43Z) - Regularizing Models via Pointwise Mutual Information for Named Entity
Recognition [17.767466724342064]
We propose a Pointwise Mutual Information (PMI) to enhance generalization ability while outperforming an in-domain performance.
Our approach enables to debias highly correlated word and labels in the benchmark datasets.
For long-named and complex-structure entities, our method can predict these entities through debiasing on conjunction or special characters.
arXiv Detail & Related papers (2021-04-15T05:47:27Z) - Adversarial Feature Augmentation and Normalization for Visual
Recognition [109.6834687220478]
Recent advances in computer vision take advantage of adversarial data augmentation to ameliorate the generalization ability of classification models.
Here, we present an effective and efficient alternative that advocates adversarial augmentation on intermediate feature embeddings.
We validate the proposed approach across diverse visual recognition tasks with representative backbone networks.
arXiv Detail & Related papers (2021-03-22T20:36:34Z) - Causal Feature Selection for Algorithmic Fairness [61.767399505764736]
We consider fairness in the integration component of data management.
We propose an approach to identify a sub-collection of features that ensure the fairness of the dataset.
arXiv Detail & Related papers (2020-06-10T20:20:10Z) - Foreseeing the Benefits of Incidental Supervision [83.08441990812636]
This paper studies whether we can, in a single framework, quantify the benefits of various types of incidental signals for a given target task without going through experiments.
We propose a unified PAC-Bayesian motivated informativeness measure, PABI, that characterizes the uncertainty reduction provided by incidental supervision signals.
arXiv Detail & Related papers (2020-06-09T20:59:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.