Self-Supervised Neural Architecture Search for Imbalanced Datasets
- URL: http://arxiv.org/abs/2109.08580v2
- Date: Mon, 20 Sep 2021 16:16:05 GMT
- Title: Self-Supervised Neural Architecture Search for Imbalanced Datasets
- Authors: Aleksandr Timofeev, Grigorios G. Chrysos, Volkan Cevher
- Abstract summary: Neural Architecture Search (NAS) provides state-of-the-art results when trained on well-curated datasets with annotated labels.
We propose a NAS-based framework that bears the threefold contributions: (a) we focus on the self-supervised scenario, where no labels are required to determine the architecture, and (b) we assume the datasets are imbalanced.
- Score: 129.3987858787811
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Neural Architecture Search (NAS) provides state-of-the-art results when
trained on well-curated datasets with annotated labels. However, annotating
data or even having balanced number of samples can be a luxury for
practitioners from different scientific fields, e.g., in the medical domain. To
that end, we propose a NAS-based framework that bears the threefold
contributions: (a) we focus on the self-supervised scenario, i.e., where no
labels are required to determine the architecture, and (b) we assume the
datasets are imbalanced, (c) we design each component to be able to run on a
resource constrained setup, i.e., on a single GPU (e.g. Google Colab). Our
components build on top of recent developments in self-supervised
learning~\citep{zbontar2021barlow}, self-supervised NAS~\citep{kaplan2020self}
and extend them for the case of imbalanced datasets. We conduct experiments on
an (artificially) imbalanced version of CIFAR-10 and we demonstrate our
proposed method outperforms standard neural networks, while using $27\times$
less parameters. To validate our assumption on a naturally imbalanced dataset,
we also conduct experiments on ChestMNIST and COVID-19 X-ray. The results
demonstrate how the proposed method can be used in imbalanced datasets, while
it can be fully run on a single GPU. Code is available
\href{https://github.com/TimofeevAlex/ssnas_imbalanced}{here}.
Related papers
- Higher Order Gauge Equivariant CNNs on Riemannian Manifolds and
Applications [7.322121417864824]
We introduce a higher order generalization of the gauge equivariant convolution, dubbed a gauge equivariant Volterra network (GEVNet)
This allows us to model spatially extended nonlinear interactions within a given field while still maintaining equivariance to global isometries.
In the neuroimaging data experiments, the resulting two-part architecture is used to automatically discriminate between patients with Lewy Body Disease (DLB), Alzheimer's Disease (AD) and Parkinson's Disease (PD) from diffusion magnetic resonance images (dMRI)
arXiv Detail & Related papers (2023-05-26T06:02:31Z) - Neural Architecture Search via Two Constant Shared Weights Initialisations [0.0]
We present a zero-cost metric that highly correlated with the train set accuracy across the NAS-Bench-101, NAS-Bench-201 and NAS-Bench-NLP benchmark datasets.
Our method is easy to integrate within existing NAS algorithms and takes a fraction of a second to evaluate a single network.
arXiv Detail & Related papers (2023-02-09T02:25:38Z) - IMB-NAS: Neural Architecture Search for Imbalanced Datasets [18.45549536555864]
We propose a new and complementary direction for improving performance on long tailed datasets.
We find that an architecture's accuracy obtained on a balanced dataset is not indicative of good performance on imbalanced ones.
To alleviate this compute burden, we aim to efficiently adapt a NAS super-network from a balanced source dataset to an imbalanced target one.
arXiv Detail & Related papers (2022-09-30T23:15:28Z) - Batch-Ensemble Stochastic Neural Networks for Out-of-Distribution
Detection [55.028065567756066]
Out-of-distribution (OOD) detection has recently received much attention from the machine learning community due to its importance in deploying machine learning models in real-world applications.
In this paper we propose an uncertainty quantification approach by modelling the distribution of features.
We incorporate an efficient ensemble mechanism, namely batch-ensemble, to construct the batch-ensemble neural networks (BE-SNNs) and overcome the feature collapse problem.
We show that BE-SNNs yield superior performance on several OOD benchmarks, such as the Two-Moons dataset, the FashionMNIST vs MNIST dataset, FashionM
arXiv Detail & Related papers (2022-06-26T16:00:22Z) - Few-Shot Non-Parametric Learning with Deep Latent Variable Model [50.746273235463754]
We propose Non-Parametric learning by Compression with Latent Variables (NPC-LV)
NPC-LV is a learning framework for any dataset with abundant unlabeled data but very few labeled ones.
We show that NPC-LV outperforms supervised methods on all three datasets on image classification in low data regime.
arXiv Detail & Related papers (2022-06-23T09:35:03Z) - NAS-Bench-Graph: Benchmarking Graph Neural Architecture Search [55.75621026447599]
We propose NAS-Bench-Graph, a tailored benchmark that supports unified, reproducible, and efficient evaluations for GraphNAS.
Specifically, we construct a unified, expressive yet compact search space, covering 26,206 unique graph neural network (GNN) architectures.
Based on our proposed benchmark, the performance of GNN architectures can be directly obtained by a look-up table without any further computation.
arXiv Detail & Related papers (2022-06-18T10:17:15Z) - Discovering Invariant Rationales for Graph Neural Networks [104.61908788639052]
Intrinsic interpretability of graph neural networks (GNNs) is to find a small subset of the input graph's features.
We propose a new strategy of discovering invariant rationale (DIR) to construct intrinsically interpretable GNNs.
arXiv Detail & Related papers (2022-01-30T16:43:40Z) - Data-Free Neural Architecture Search via Recursive Label Calibration [34.84457903882755]
This paper aims to explore the feasibility of neural architecture search given only a pre-trained model without using any original training data.
We start by synthesizing usable data through recovering the knowledge from a pre-trained deep neural network.
We instantiate our proposed framework with three popular NAS algorithms: DARTS, ProxylessNAS and SPOS.
arXiv Detail & Related papers (2021-12-03T18:53:16Z) - Statistical model-based evaluation of neural networks [74.10854783437351]
We develop an experimental setup for the evaluation of neural networks (NNs)
The setup helps to benchmark a set of NNs vis-a-vis minimum-mean-square-error (MMSE) performance bounds.
This allows us to test the effects of training data size, data dimension, data geometry, noise, and mismatch between training and testing conditions.
arXiv Detail & Related papers (2020-11-18T00:33:24Z) - Wiki-CS: A Wikipedia-Based Benchmark for Graph Neural Networks [0.0]
We present Wiki-CS, a novel dataset derived from Wikipedia for benchmarking Graph Neural Networks.
The dataset consists of nodes corresponding to Computer Science articles, with edges based on hyperlinks and 10 classes representing different branches of the field.
We use the dataset to evaluate semi-supervised node classification and single-relation link prediction models.
arXiv Detail & Related papers (2020-07-06T17:25:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.