Leveraging Time Irreversibility with Order-Contrastive Pre-training
- URL: http://arxiv.org/abs/2111.02599v1
- Date: Thu, 4 Nov 2021 02:56:52 GMT
- Title: Leveraging Time Irreversibility with Order-Contrastive Pre-training
- Authors: Monica Agrawal, Hunter Lang, Michael Offin, Lior Gazit, David Sontag
- Abstract summary: We explore an "order-contrastive" method for self-supervised pre-training on longitudinal data.
We prove a finite-sample guarantee for the downstream error of a representation learned with order-contrastive pre-training.
Our results indicate that pre-training methods designed for particular classes of distributions and downstream tasks can improve the performance of self-supervised learning.
- Score: 3.1848820580333737
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Label-scarce, high-dimensional domains such as healthcare present a challenge
for modern machine learning techniques. To overcome the difficulties posed by a
lack of labeled data, we explore an "order-contrastive" method for
self-supervised pre-training on longitudinal data. We sample pairs of time
segments, switch the order for half of them, and train a model to predict
whether a given pair is in the correct order. Intuitively, the ordering task
allows the model to attend to the least time-reversible features (for example,
features that indicate progression of a chronic disease). The same features are
often useful for downstream tasks of interest. To quantify this, we study a
simple theoretical setting where we prove a finite-sample guarantee for the
downstream error of a representation learned with order-contrastive
pre-training. Empirically, in synthetic and longitudinal healthcare settings,
we demonstrate the effectiveness of order-contrastive pre-training in the
small-data regime over supervised learning and other self-supervised
pre-training baselines. Our results indicate that pre-training methods designed
for particular classes of distributions and downstream tasks can improve the
performance of self-supervised learning.
Related papers
- Incremental Self-training for Semi-supervised Learning [56.57057576885672]
IST is simple yet effective and fits existing self-training-based semi-supervised learning methods.
We verify the proposed IST on five datasets and two types of backbone, effectively improving the recognition accuracy and learning speed.
arXiv Detail & Related papers (2024-04-14T05:02:00Z) - Hierarchical Decomposition of Prompt-Based Continual Learning:
Rethinking Obscured Sub-optimality [55.88910947643436]
Self-supervised pre-training is essential for handling vast quantities of unlabeled data in practice.
HiDe-Prompt is an innovative approach that explicitly optimize the hierarchical components with an ensemble of task-specific prompts and statistics.
Our experiments demonstrate the superior performance of HiDe-Prompt and its robustness to pre-training paradigms in continual learning.
arXiv Detail & Related papers (2023-10-11T06:51:46Z) - Towards Robust Continual Learning with Bayesian Adaptive Moment Regularization [51.34904967046097]
Continual learning seeks to overcome the challenge of catastrophic forgetting, where a model forgets previously learnt information.
We introduce a novel prior-based method that better constrains parameter growth, reducing catastrophic forgetting.
Results show that BAdam achieves state-of-the-art performance for prior-based methods on challenging single-headed class-incremental experiments.
arXiv Detail & Related papers (2023-09-15T17:10:51Z) - Improving Adaptive Conformal Prediction Using Self-Supervised Learning [72.2614468437919]
We train an auxiliary model with a self-supervised pretext task on top of an existing predictive model and use the self-supervised error as an additional feature to estimate nonconformity scores.
We empirically demonstrate the benefit of the additional information using both synthetic and real data on the efficiency (width), deficit, and excess of conformal prediction intervals.
arXiv Detail & Related papers (2023-02-23T18:57:14Z) - Self-Distillation for Further Pre-training of Transformers [83.84227016847096]
We propose self-distillation as a regularization for a further pre-training stage.
We empirically validate the efficacy of self-distillation on a variety of benchmark datasets for image and text classification tasks.
arXiv Detail & Related papers (2022-09-30T02:25:12Z) - Task Agnostic Representation Consolidation: a Self-supervised based
Continual Learning Approach [14.674494335647841]
We propose a two-stage training paradigm for CL that intertwines task-agnostic and task-specific learning.
We show that our training paradigm can be easily added to memory- or regularization-based approaches.
arXiv Detail & Related papers (2022-07-13T15:16:51Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Weakly-supervised Temporal Path Representation Learning with Contrastive
Curriculum Learning -- Extended Version [35.86394282979721]
A temporal path(TP) that includes temporal information, e.g., departure time, into the path is of fundamental to enable such applications.
Existing methods fail to achieve the goal since (i) supervised methods require large amounts of task-specific labels when training and thus fail to generalize the obtained TPRs to other tasks.
We propose a Weakly-Supervised Contrastive (WSC) learning model that encodes both the spatial and temporal information of a temporal path into a TPR.
arXiv Detail & Related papers (2022-03-30T07:36:20Z) - Self-supervised driven consistency training for annotation efficient
histopathology image analysis [13.005873872821066]
Training a neural network with a large labeled dataset is still a dominant paradigm in computational histopathology.
We propose a self-supervised pretext task that harnesses the underlying multi-resolution contextual cues in histology whole-slide images to learn a powerful supervisory signal for unsupervised representation learning.
We also propose a new teacher-student semi-supervised consistency paradigm that learns to effectively transfer the pretrained representations to downstream tasks based on prediction consistency with the task-specific un-labeled data.
arXiv Detail & Related papers (2021-02-07T19:46:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.