Improving Self-supervised Learning for Out-of-distribution Task via
Auxiliary Classifier
- URL: http://arxiv.org/abs/2209.02881v1
- Date: Wed, 7 Sep 2022 02:00:01 GMT
- Title: Improving Self-supervised Learning for Out-of-distribution Task via
Auxiliary Classifier
- Authors: Harshita Boonlia, Tanmoy Dam, Md Meftahul Ferdaus, Sreenatha G.
Anavatti, Ankan Mullick
- Abstract summary: We observe a strong relationship between rotation prediction (self-supervised) accuracy and semantic classification accuracy on OOD tasks.
We introduce an additional auxiliary classification head in our multi-task network along with semantic classification and rotation prediction head.
Our proposed learning method is framed into bi-level optimisation problem where the upper-level is trained to update the parameters for semantic classification and rotation prediction head.
- Score: 6.61825491400122
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In real world scenarios, out-of-distribution (OOD) datasets may have a large
distributional shift from training datasets. This phenomena generally occurs
when a trained classifier is deployed on varying dynamic environments, which
causes a significant drop in performance. To tackle this issue, we are
proposing an end-to-end deep multi-task network in this work. Observing a
strong relationship between rotation prediction (self-supervised) accuracy and
semantic classification accuracy on OOD tasks, we introduce an additional
auxiliary classification head in our multi-task network along with semantic
classification and rotation prediction head. To observe the influence of this
addition classifier in improving the rotation prediction head, our proposed
learning method is framed into bi-level optimisation problem where the
upper-level is trained to update the parameters for semantic classification and
rotation prediction head. In the lower-level optimisation, only the auxiliary
classification head is updated through semantic classification head by fixing
the parameters of the semantic classification head. The proposed method has
been validated through three unseen OOD datasets where it exhibits a clear
improvement in semantic classification accuracy than other two baseline
methods. Our code is available on GitHub
\url{https://github.com/harshita-555/OSSL}
Related papers
- Enhancing Visual Continual Learning with Language-Guided Supervision [76.38481740848434]
Continual learning aims to empower models to learn new tasks without forgetting previously acquired knowledge.
We argue that the scarce semantic information conveyed by the one-hot labels hampers the effective knowledge transfer across tasks.
Specifically, we use PLMs to generate semantic targets for each class, which are frozen and serve as supervision signals.
arXiv Detail & Related papers (2024-03-24T12:41:58Z) - Class-Imbalanced Semi-Supervised Learning for Large-Scale Point Cloud
Semantic Segmentation via Decoupling Optimization [64.36097398869774]
Semi-supervised learning (SSL) has been an active research topic for large-scale 3D scene understanding.
The existing SSL-based methods suffer from severe training bias due to class imbalance and long-tail distributions of the point cloud data.
We introduce a new decoupling optimization framework, which disentangles feature representation learning and classifier in an alternative optimization manner to shift the bias decision boundary effectively.
arXiv Detail & Related papers (2024-01-13T04:16:40Z) - Informative regularization for a multi-layer perceptron RR Lyrae
classifier under data shift [3.303002683812084]
We propose a scalable and easily adaptable approach based on an informative regularization and an ad-hoc training procedure to mitigate the shift problem.
Our method provides a new path to incorporate knowledge from characteristic features into artificial neural networks to manage the underlying data shift problem.
arXiv Detail & Related papers (2023-03-12T02:49:19Z) - Surgical Fine-Tuning Improves Adaptation to Distribution Shifts [114.17184775397067]
A common approach to transfer learning under distribution shift is to fine-tune the last few layers of a pre-trained model.
This paper shows that in such settings, selectively fine-tuning a subset of layers matches or outperforms commonly used fine-tuning approaches.
arXiv Detail & Related papers (2022-10-20T17:59:15Z) - SCAI: A Spectral data Classification framework with Adaptive Inference
for the IoT platform [0.0]
We propose a Spectral data Classification framework with Adaptive Inference.
Specifically, to allocate different computations for different samples while better exploiting the collaboration among different devices.
To the best of our knowledge, this paper is the first attempt to conduct optimization by adaptive inference for spectral detection under the IoT platform.
arXiv Detail & Related papers (2022-06-24T09:22:52Z) - Fine-Grained Visual Classification using Self Assessment Classifier [12.596520707449027]
Extracting discriminative features plays a crucial role in the fine-grained visual classification task.
In this paper, we introduce a Self Assessment, which simultaneously leverages the representation of the image and top-k prediction classes.
We show that our method achieves new state-of-the-art results on CUB200-2011, Stanford Dog, and FGVC Aircraft datasets.
arXiv Detail & Related papers (2022-05-21T07:41:27Z) - Ensemble Classifier Design Tuned to Dataset Characteristics for Network
Intrusion Detection [0.0]
Two new algorithms are proposed to address the class overlap issue in the dataset.
The proposed design is evaluated for both binary and multi-category classification.
arXiv Detail & Related papers (2022-05-08T21:06:42Z) - Prototypical Classifier for Robust Class-Imbalanced Learning [64.96088324684683]
We propose textitPrototypical, which does not require fitting additional parameters given the embedding network.
Prototypical produces balanced and comparable predictions for all classes even though the training set is class-imbalanced.
We test our method on CIFAR-10LT, CIFAR-100LT and Webvision datasets, observing that Prototypical obtains substaintial improvements compared with state of the arts.
arXiv Detail & Related papers (2021-10-22T01:55:01Z) - No Fear of Heterogeneity: Classifier Calibration for Federated Learning
with Non-IID Data [78.69828864672978]
A central challenge in training classification models in the real-world federated system is learning with non-IID data.
We propose a novel and simple algorithm called Virtual Representations (CCVR), which adjusts the classifier using virtual representations sampled from an approximated ssian mixture model.
Experimental results demonstrate that CCVR state-of-the-art performance on popular federated learning benchmarks including CIFAR-10, CIFAR-100, and CINIC-10.
arXiv Detail & Related papers (2021-06-09T12:02:29Z) - Fine-Grained Visual Classification with Efficient End-to-end
Localization [49.9887676289364]
We present an efficient localization module that can be fused with a classification network in an end-to-end setup.
We evaluate the new model on the three benchmark datasets CUB200-2011, Stanford Cars and FGVC-Aircraft.
arXiv Detail & Related papers (2020-05-11T14:07:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.