ScatSimCLR: self-supervised contrastive learning with pretext task
regularization for small-scale datasets
- URL: http://arxiv.org/abs/2108.13939v1
- Date: Tue, 31 Aug 2021 15:58:45 GMT
- Title: ScatSimCLR: self-supervised contrastive learning with pretext task
regularization for small-scale datasets
- Authors: Vitaliy Kinakh, Olga Taran, Svyatoslav Voloshynovskiy
- Abstract summary: We consider a problem of self-supervised learning for small-scale datasets based on contrastive loss between multiple views of the data.
We argue that the number of parameters of the whole system and the number of views can be considerably reduced while preserving the same classification accuracy.
- Score: 5.2424255020469595
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we consider a problem of self-supervised learning for
small-scale datasets based on contrastive loss between multiple views of the
data, which demonstrates the state-of-the-art performance in classification
task. Despite the reported results, such factors as the complexity of training
requiring complex architectures, the needed number of views produced by data
augmentation, and their impact on the classification accuracy are understudied
problems. To establish the role of these factors, we consider an architecture
of contrastive loss system such as SimCLR, where baseline model is replaced by
geometrically invariant "hand-crafted" network ScatNet with small trainable
adapter network and argue that the number of parameters of the whole system and
the number of views can be considerably reduced while practically preserving
the same classification accuracy. In addition, we investigate the impact of
regularization strategies using pretext task learning based on an estimation of
parameters of augmentation transform such as rotation and jigsaw permutation
for both traditional baseline models and ScatNet based models. Finally, we
demonstrate that the proposed architecture with pretext task learning
regularization achieves the state-of-the-art classification performance with a
smaller number of trainable parameters and with reduced number of views.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - MTP: Advancing Remote Sensing Foundation Model via Multi-Task Pretraining [73.81862342673894]
Foundation models have reshaped the landscape of Remote Sensing (RS) by enhancing various image interpretation tasks.
transferring the pretrained models to downstream tasks may encounter task discrepancy due to their formulation of pretraining as image classification or object discrimination tasks.
We conduct multi-task supervised pretraining on the SAMRS dataset, encompassing semantic segmentation, instance segmentation, and rotated object detection.
Our models are finetuned on various RS downstream tasks, such as scene classification, horizontal and rotated object detection, semantic segmentation, and change detection.
arXiv Detail & Related papers (2024-03-20T09:17:22Z) - Mitigating the Effect of Incidental Correlations on Part-based Learning [50.682498099720114]
Part-based representations could be more interpretable and generalize better with limited data.
We present two innovative regularization methods for part-based representations.
We exhibit state-of-the-art (SoTA) performance on few-shot learning tasks on benchmark datasets.
arXiv Detail & Related papers (2023-09-30T13:44:48Z) - Regularization Through Simultaneous Learning: A Case Study on Plant
Classification [0.0]
This paper introduces Simultaneous Learning, a regularization approach drawing on principles of Transfer Learning and Multi-task Learning.
We leverage auxiliary datasets with the target dataset, the UFOP-HVD, to facilitate simultaneous classification guided by a customized loss function.
Remarkably, our approach demonstrates superior performance over models without regularization.
arXiv Detail & Related papers (2023-05-22T19:44:57Z) - Self-Supervised Class Incremental Learning [51.62542103481908]
Existing Class Incremental Learning (CIL) methods are based on a supervised classification framework sensitive to data labels.
When updating them based on the new class data, they suffer from catastrophic forgetting: the model cannot discern old class data clearly from the new.
In this paper, we explore the performance of Self-Supervised representation learning in Class Incremental Learning (SSCIL) for the first time.
arXiv Detail & Related papers (2021-11-18T06:58:19Z) - NDPNet: A novel non-linear data projection network for few-shot
fine-grained image classification [33.71025164816078]
We introduce the non-linear data projection concept into the design of metric-based fine-grained image classification architecture.
Our proposed architecture can be easily embedded into any episodic training mechanisms for end-to-end training from scratch.
arXiv Detail & Related papers (2021-06-13T13:33:09Z) - Analyzing Overfitting under Class Imbalance in Neural Networks for Image
Segmentation [19.259574003403998]
In image segmentation neural networks may overfit to the foreground samples from small structures.
In this study, we provide new insights on the problem of overfitting under class imbalance by inspecting the network behavior.
arXiv Detail & Related papers (2021-02-20T14:57:58Z) - Supervised Contrastive Learning for Pre-trained Language Model
Fine-tuning [23.00300794016583]
State-of-the-art natural language understanding classification models follow two-stages.
We propose a supervised contrastive learning (SCL) objective for the fine-tuning stage.
Our proposed fine-tuning objective leads to models that are more robust to different levels of noise in the fine-tuning training data.
arXiv Detail & Related papers (2020-11-03T01:10:39Z) - On Robustness and Transferability of Convolutional Neural Networks [147.71743081671508]
Modern deep convolutional networks (CNNs) are often criticized for not generalizing under distributional shifts.
We study the interplay between out-of-distribution and transfer performance of modern image classification CNNs for the first time.
We find that increasing both the training set and model sizes significantly improve the distributional shift robustness.
arXiv Detail & Related papers (2020-07-16T18:39:04Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.