Self-Guided Masked Autoencoders for Domain-Agnostic Self-Supervised
Learning
- URL: http://arxiv.org/abs/2402.14789v1
- Date: Thu, 22 Feb 2024 18:46:22 GMT
- Title: Self-Guided Masked Autoencoders for Domain-Agnostic Self-Supervised
Learning
- Authors: Johnathan Xie, Yoonho Lee, Annie S. Chen, Chelsea Finn
- Abstract summary: We present Self-guided Masked Autoencoders (SMA), a fully domain-agnostic masked modeling method.
SMA trains an attention based model using a masked modeling objective, by learning masks to sample without any domain-specific assumptions.
We evaluate SMA on three self-supervised learning benchmarks in protein biology, chemical property prediction, and particle physics.
- Score: 58.93724285214628
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Self-supervised learning excels in learning representations from large
amounts of unlabeled data, demonstrating success across multiple data
modalities. Yet, extending self-supervised learning to new modalities is
non-trivial because the specifics of existing methods are tailored to each
domain, such as domain-specific augmentations which reflect the invariances in
the target task. While masked modeling is promising as a domain-agnostic
framework for self-supervised learning because it does not rely on input
augmentations, its mask sampling procedure remains domain-specific. We present
Self-guided Masked Autoencoders (SMA), a fully domain-agnostic masked modeling
method. SMA trains an attention based model using a masked modeling objective,
by learning masks to sample without any domain-specific assumptions. We
evaluate SMA on three self-supervised learning benchmarks in protein biology,
chemical property prediction, and particle physics. We find SMA is capable of
learning representations without domain-specific knowledge and achieves
state-of-the-art performance on these three benchmarks.
Related papers
- Model Evaluation for Domain Identification of Unknown Classes in
Open-World Recognition: A Proposal [0.0]
Open-World Recognition (OWR) is an emerging field that makes a machine learning model competent in rejecting the unknowns.
In this study, we propose an evaluation protocol for estimating a model's capability in separating unknown in-domain (ID) and unknown out-of-domain (OOD)
We experimented with five different domains: garbage, food, dogs, plants, and birds.
arXiv Detail & Related papers (2023-12-09T03:54:25Z) - MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments [72.6405488990753]
Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks.
We propose a single-stage and standalone method, MOCA, which unifies both desired properties.
We achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols.
arXiv Detail & Related papers (2023-07-18T15:46:20Z) - Exploring The Role of Mean Teachers in Self-supervised Masked
Auto-Encoders [64.03000385267339]
Masked image modeling (MIM) has become a popular strategy for self-supervised learning(SSL) of visual representations with Vision Transformers.
We present a simple SSL method, the Reconstruction-Consistent Masked Auto-Encoder (RC-MAE) by adding an EMA teacher to MAE.
RC-MAE converges faster and requires less memory usage than state-of-the-art self-distillation methods during pre-training.
arXiv Detail & Related papers (2022-10-05T08:08:55Z) - Exploring Target Representations for Masked Autoencoders [78.57196600585462]
We show that a careful choice of the target representation is unnecessary for learning good representations.
We propose a multi-stage masked distillation pipeline and use a randomly model as the teacher.
A proposed method to perform masked knowledge distillation with bootstrapped teachers (dBOT) outperforms previous self-supervised methods by nontrivial margins.
arXiv Detail & Related papers (2022-09-08T16:55:19Z) - Unsupervised Domain Adaptation for Monocular 3D Object Detection via
Self-Training [57.25828870799331]
We propose STMono3D, a new self-teaching framework for unsupervised domain adaptation on Mono3D.
We develop a teacher-student paradigm to generate adaptive pseudo labels on the target domain.
STMono3D achieves remarkable performance on all evaluated datasets and even surpasses fully supervised results on the KITTI 3D object detection dataset.
arXiv Detail & Related papers (2022-04-25T12:23:07Z) - Intelligent Masking: Deep Q-Learning for Context Encoding in Medical
Image Analysis [48.02011627390706]
We develop a novel self-supervised approach that occludes targeted regions to improve the pre-training procedure.
We show that training the agent against the prediction model can significantly improve the semantic features extracted for downstream classification tasks.
arXiv Detail & Related papers (2022-03-25T19:05:06Z) - Multi-Domain Adversarial Feature Generalization for Person
Re-Identification [52.835955258959785]
We propose a multi-dataset feature generalization network (MMFA-AAE)
It is capable of learning a universal domain-invariant feature representation from multiple labeled datasets and generalizing it to unseen' camera systems.
It also surpasses many state-of-the-art supervised methods and unsupervised domain adaptation methods by a large margin.
arXiv Detail & Related papers (2020-11-25T08:03:15Z) - Semi-supervised Learning with a Teacher-student Network for Generalized
Attribute Prediction [7.462336024223667]
This paper presents a study on semi-supervised learning to solve the visual attribute prediction problem.
Our method achieves competitive performance on various benchmarks for fashion attribute prediction.
arXiv Detail & Related papers (2020-07-14T02:06:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.