EBMs vs. CL: Exploring Self-Supervised Visual Pretraining for Visual
Question Answering
- URL: http://arxiv.org/abs/2206.14355v1
- Date: Wed, 29 Jun 2022 01:44:23 GMT
- Title: EBMs vs. CL: Exploring Self-Supervised Visual Pretraining for Visual
Question Answering
- Authors: Violetta Shevchenko, Ehsan Abbasnejad, Anthony Dick, Anton van den
Hengel, Damien Teney
- Abstract summary: clean and diverse labeled data is a major roadblock for training models on complex tasks such as visual question answering (VQA)
We review and evaluate self-supervised methods to leverage unlabeled images and pretrain a model, which we then fine-tune on a custom VQA task.
We find that both EBMs and CL can learn representations from unlabeled images that enable training a VQA model on very little annotated data.
- Score: 53.40635559899501
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The availability of clean and diverse labeled data is a major roadblock for
training models on complex tasks such as visual question answering (VQA). The
extensive work on large vision-and-language models has shown that
self-supervised learning is effective for pretraining multimodal interactions.
In this technical report, we focus on visual representations. We review and
evaluate self-supervised methods to leverage unlabeled images and pretrain a
model, which we then fine-tune on a custom VQA task that allows controlled
evaluation and diagnosis. We compare energy-based models (EBMs) with
contrastive learning (CL). While EBMs are growing in popularity, they lack an
evaluation on downstream tasks. We find that both EBMs and CL can learn
representations from unlabeled images that enable training a VQA model on very
little annotated data. In a simple setting similar to CLEVR, we find that CL
representations also improve systematic generalization, and even match the
performance of representations from a larger, supervised, ImageNet-pretrained
model. However, we find EBMs to be difficult to train because of instabilities
and high variability in their results. Although EBMs prove useful for OOD
detection, other results on supervised energy-based training and uncertainty
calibration are largely negative. Overall, CL currently seems a preferable
option over EBMs.
Related papers
- Enhancing Large Vision Language Models with Self-Training on Image Comprehension [99.9389737339175]
We introduce Self-Training on Image (STIC), which emphasizes a self-training approach specifically for image comprehension.
First, the model self-constructs a preference for image descriptions using unlabeled images.
To further self-improve reasoning on the extracted visual information, we let the model reuse a small portion of existing instruction-tuning data.
arXiv Detail & Related papers (2024-05-30T05:53:49Z) - Machine Vision Therapy: Multimodal Large Language Models Can Enhance Visual Robustness via Denoising In-Context Learning [67.0609518552321]
We propose to conduct Machine Vision Therapy which aims to rectify the noisy predictions from vision models.
By fine-tuning with the denoised labels, the learning model performance can be boosted in an unsupervised manner.
arXiv Detail & Related papers (2023-12-05T07:29:14Z) - CTDS: Centralized Teacher with Decentralized Student for Multi-Agent
Reinforcement Learning [114.69155066932046]
This work proposes a novel.
Teacher with Decentralized Student (C TDS) framework, which consists of a teacher model and a student model.
Specifically, the teacher model allocates the team reward by learning individual Q-values conditioned on global observation.
The student model utilizes the partial observations to approximate the Q-values estimated by the teacher model.
arXiv Detail & Related papers (2022-03-16T06:03:14Z) - Revisiting Weakly Supervised Pre-Training of Visual Perception Models [27.95816470075203]
Large-scale weakly supervised pre-training can outperform fully supervised approaches.
This paper revisits weakly-supervised pre-training of models using hashtag supervision.
Our results provide a compelling argument for the use of weakly supervised learning in the development of visual recognition systems.
arXiv Detail & Related papers (2022-01-20T18:55:06Z) - The Lottery Tickets Hypothesis for Supervised and Self-supervised
Pre-training in Computer Vision Models [115.49214555402567]
Pre-trained weights often boost a wide range of downstream tasks including classification, detection, and segmentation.
Recent studies suggest that pre-training benefits from gigantic model capacity.
In this paper, we examine supervised and self-supervised pre-trained models through the lens of the lottery ticket hypothesis (LTH)
arXiv Detail & Related papers (2020-12-12T21:53:55Z) - What Makes for Good Views for Contrastive Learning? [90.49736973404046]
We argue that we should reduce the mutual information (MI) between views while keeping task-relevant information intact.
We devise unsupervised and semi-supervised frameworks that learn effective views by aiming to reduce their MI.
As a by-product, we achieve a new state-of-the-art accuracy on unsupervised pre-training for ImageNet classification.
arXiv Detail & Related papers (2020-05-20T17:59:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.