Discriminative, Generative and Self-Supervised Approaches for
Target-Agnostic Learning
- URL: http://arxiv.org/abs/2011.06428v1
- Date: Thu, 12 Nov 2020 15:03:40 GMT
- Title: Discriminative, Generative and Self-Supervised Approaches for
Target-Agnostic Learning
- Authors: Yuan Jin, Wray Buntine, Francois Petitjean, Geoffrey I. Webb
- Abstract summary: generative and self-supervised learning models are shown to perform well at the task.
Our derived theorem for the pseudo-likelihood theory also shows that they are related for inferring a joint distribution model.
- Score: 8.666667951130892
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Supervised learning, characterized by both discriminative and generative
learning, seeks to predict the values of single (or sometimes multiple)
predefined target attributes based on a predefined set of predictor attributes.
For applications where the information available and predictions to be made may
vary from instance to instance, we propose the task of target-agnostic learning
where arbitrary disjoint sets of attributes can be used for each of predictors
and targets for each to-be-predicted instance. For this task, we survey a wide
range of techniques available for handling missing values, self-supervised
training and pseudo-likelihood training, and adapt them to a suite of
algorithms that are suitable for the task. We conduct extensive experiments on
this suite of algorithms on a large collection of categorical, continuous and
discretized datasets, and report their performance in terms of both
classification and regression errors. We also report the training and
prediction time of these algorithms when handling large-scale datasets. Both
generative and self-supervised learning models are shown to perform well at the
task, although their characteristics towards the different types of data are
quite different. Nevertheless, our derived theorem for the pseudo-likelihood
theory also shows that they are related for inferring a joint distribution
model based on the pseudo-likelihood training.
Related papers
- Ranking and Combining Latent Structured Predictive Scores without Labeled Data [2.5064967708371553]
This paper introduces a novel structured unsupervised ensemble learning model (SUEL)
It exploits the dependency between a set of predictors with continuous predictive scores, rank the predictors without labeled data and combine them to an ensembled score with weights.
The efficacy of the proposed methods is rigorously assessed through both simulation studies and real-world application of risk genes discovery.
arXiv Detail & Related papers (2024-08-14T20:14:42Z) - Improving Adaptive Conformal Prediction Using Self-Supervised Learning [72.2614468437919]
We train an auxiliary model with a self-supervised pretext task on top of an existing predictive model and use the self-supervised error as an additional feature to estimate nonconformity scores.
We empirically demonstrate the benefit of the additional information using both synthetic and real data on the efficiency (width), deficit, and excess of conformal prediction intervals.
arXiv Detail & Related papers (2023-02-23T18:57:14Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - Prediction of Dilatory Behavior in eLearning: A Comparison of Multiple
Machine Learning Models [0.2963240482383777]
Procrastination, the irrational delay of tasks, is a common occurrence in online learning.
Research focusing on such predictions is scarce.
Studies involving different types of predictors and comparisons between the predictive performance of various methods are virtually non-existent.
arXiv Detail & Related papers (2022-06-30T07:24:08Z) - Masked prediction tasks: a parameter identifiability view [49.533046139235466]
We focus on the widely used self-supervised learning method of predicting masked tokens.
We show that there is a rich landscape of possibilities, out of which some prediction tasks yield identifiability, while others do not.
arXiv Detail & Related papers (2022-02-18T17:09:32Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Selecting the suitable resampling strategy for imbalanced data
classification regarding dataset properties [62.997667081978825]
In many application domains such as medicine, information retrieval, cybersecurity, social media, etc., datasets used for inducing classification models often have an unequal distribution of the instances of each class.
This situation, known as imbalanced data classification, causes low predictive performance for the minority class examples.
Oversampling and undersampling techniques are well-known strategies to deal with this problem by balancing the number of examples of each class.
arXiv Detail & Related papers (2021-12-15T18:56:39Z) - Meta-learning One-class Classifiers with Eigenvalue Solvers for
Supervised Anomaly Detection [55.888835686183995]
We propose a neural network-based meta-learning method for supervised anomaly detection.
We experimentally demonstrate that the proposed method achieves better performance than existing anomaly detection and few-shot learning methods.
arXiv Detail & Related papers (2021-03-01T01:43:04Z) - Learning from Incomplete Features by Simultaneous Training of Neural
Networks and Sparse Coding [24.3769047873156]
This paper addresses the problem of training a classifier on a dataset with incomplete features.
We assume that different subsets of features (random or structured) are available at each data instance.
A new supervised learning method is developed to train a general classifier, using only a subset of features per sample.
arXiv Detail & Related papers (2020-11-28T02:20:39Z) - Instance-Based Learning of Span Representations: A Case Study through
Named Entity Recognition [48.06319154279427]
We present a method of instance-based learning that learns similarities between spans.
Our method enables to build models that have high interpretability without sacrificing performance.
arXiv Detail & Related papers (2020-04-29T23:32:42Z) - Sampling Prediction-Matching Examples in Neural Networks: A
Probabilistic Programming Approach [9.978961706999833]
We consider the problem of exploring the prediction level sets of a classifier using probabilistic programming.
We define a prediction level set to be the set of examples for which the predictor has the same specified prediction confidence.
We demonstrate this technique with experiments on a synthetic dataset and MNIST.
arXiv Detail & Related papers (2020-01-09T15:57:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.