Exploiting non-i.i.d. data towards more robust machine learning
algorithms
- URL: http://arxiv.org/abs/2010.03429v1
- Date: Wed, 7 Oct 2020 14:15:37 GMT
- Title: Exploiting non-i.i.d. data towards more robust machine learning
algorithms
- Authors: Wim Casteels and Peter Hellinckx
- Abstract summary: Machine learning algorithms have increasingly been shown to excel in finding patterns and correlations from data.
In this paper a regularization scheme is introduced that prefers universal causal correlations.
A better performance is obtained on an out-of-distribution test set with respect to a more conventional l-regularization.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the field of machine learning there is a growing interest towards more
robust and generalizable algorithms. This is for example important to bridge
the gap between the environment in which the training data was collected and
the environment where the algorithm is deployed. Machine learning algorithms
have increasingly been shown to excel in finding patterns and correlations from
data. Determining the consistency of these patterns and for example the
distinction between causal correlations and nonsensical spurious relations has
proven to be much more difficult. In this paper a regularization scheme is
introduced that prefers universal causal correlations. This approach is based
on 1) the robustness of causal correlations and 2) the data not being
independently and identically distribute (i.i.d.). The scheme is demonstrated
with a classification task by clustering the (non-i.i.d.) training set in
subpopulations. A non-i.i.d. regularization term is then introduced that
penalizes weights that are not invariant over these clusters. The resulting
algorithm favours correlations that are universal over the subpopulations and
indeed a better performance is obtained on an out-of-distribution test set with
respect to a more conventional l_2-regularization.
Related papers
- Generalized Oversampling for Learning from Imbalanced datasets and
Associated Theory [0.0]
In supervised learning, it is quite frequent to be confronted with real imbalanced datasets.
We propose a data augmentation procedure, the GOLIATH algorithm, based on kernel density estimates.
We evaluate the performance of the GOLIATH algorithm in imbalanced regression situations.
arXiv Detail & Related papers (2023-08-05T23:08:08Z) - Doubly Inhomogeneous Reinforcement Learning [4.334006170547247]
We propose an original algorithm to determine the best data chunks" that display similar dynamics over time and across individuals for policy learning.
Our method is general, and works with a wide range of clustering and change point detection algorithms.
arXiv Detail & Related papers (2022-11-08T03:41:14Z) - Amortized Inference for Causal Structure Learning [72.84105256353801]
Learning causal structure poses a search problem that typically involves evaluating structures using a score or independence test.
We train a variational inference model to predict the causal structure from observational/interventional data.
Our models exhibit robust generalization capabilities under substantial distribution shift.
arXiv Detail & Related papers (2022-05-25T17:37:08Z) - Faster Deterministic Approximation Algorithms for Correlation Clustering
and Cluster Deletion [5.584060970507507]
Correlation clustering is a framework for partitioning datasets based on pairwise similarity and dissimilarity scores.
In this paper we prove new relationships between correlation clustering problems and edge labeling problems related to the principle of strong partitioningic closure.
We develop new approximation algorithms for correlation clustering that have deterministic constant factor approximation guarantees and avoid the canonical linear programming relaxation.
arXiv Detail & Related papers (2021-11-20T22:47:19Z) - Predict then Interpolate: A Simple Algorithm to Learn Stable Classifiers [59.06169363181417]
Predict then Interpolate (PI) is an algorithm for learning correlations that are stable across environments.
We prove that by interpolating the distributions of the correct predictions and the wrong predictions, we can uncover an oracle distribution where the unstable correlation vanishes.
arXiv Detail & Related papers (2021-05-26T15:37:48Z) - Can Active Learning Preemptively Mitigate Fairness Issues? [66.84854430781097]
dataset bias is one of the prevailing causes of unfairness in machine learning.
We study whether models trained with uncertainty-based ALs are fairer in their decisions with respect to a protected class.
We also explore the interaction of algorithmic fairness methods such as gradient reversal (GRAD) and BALD.
arXiv Detail & Related papers (2021-04-14T14:20:22Z) - Clustering-based Unsupervised Generative Relation Extraction [3.342376225738321]
We propose a Clustering-based Unsupervised generative Relation Extraction framework (CURE)
We use an "Encoder-Decoder" architecture to perform self-supervised learning so the encoder can extract relation information.
Our model performs better than state-of-the-art models on both New York Times (NYT) and United Nations Parallel Corpus (UNPC) standard datasets.
arXiv Detail & Related papers (2020-09-26T20:36:40Z) - Decorrelated Clustering with Data Selection Bias [55.91842043124102]
We propose a novel Decorrelation regularized K-Means algorithm (DCKM) for clustering with data selection bias.
Our DCKM algorithm achieves significant performance gains, indicating the necessity of removing unexpected feature correlations induced by selection bias.
arXiv Detail & Related papers (2020-06-29T08:55:50Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z) - Learning Causal Models Online [103.87959747047158]
Predictive models can rely on spurious correlations in the data for making predictions.
One solution for achieving strong generalization is to incorporate causal structures in the models.
We propose an online algorithm that continually detects and removes spurious features.
arXiv Detail & Related papers (2020-06-12T20:49:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.