A picture of the space of typical learnable tasks
- URL: http://arxiv.org/abs/2210.17011v4
- Date: Sat, 22 Jul 2023 03:56:53 GMT
- Title: A picture of the space of typical learnable tasks
- Authors: Rahul Ramesh, Jialin Mao, Itay Griniasty, Rubing Yang, Han Kheng Teoh,
Mark Transtrum, James P. Sethna, Pratik Chaudhari
- Abstract summary: We develop information geometric techniques to understand the representations learned by deep networks when they are trained on different tasks.
We shed light on the following phenomena that relate to the structure of the space of tasks.
We use classification tasks constructed from the CIFAR-10 and Imagenet datasets to study these phenomena.
- Score: 12.374133486467013
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We develop information geometric techniques to understand the representations
learned by deep networks when they are trained on different tasks using
supervised, meta-, semi-supervised and contrastive learning. We shed light on
the following phenomena that relate to the structure of the space of tasks: (1)
the manifold of probabilistic models trained on different tasks using different
representation learning methods is effectively low-dimensional; (2) supervised
learning on one task results in a surprising amount of progress even on
seemingly dissimilar tasks; progress on other tasks is larger if the training
task has diverse classes; (3) the structure of the space of tasks indicated by
our analysis is consistent with parts of the Wordnet phylogenetic tree; (4)
episodic meta-learning algorithms and supervised learning traverse different
trajectories during training but they fit similar models eventually; (5)
contrastive and semi-supervised learning methods traverse trajectories similar
to those of supervised learning. We use classification tasks constructed from
the CIFAR-10 and Imagenet datasets to study these phenomena.
Related papers
- Pre-training Multi-task Contrastive Learning Models for Scientific
Literature Understanding [52.723297744257536]
Pre-trained language models (LMs) have shown effectiveness in scientific literature understanding tasks.
We propose a multi-task contrastive learning framework, SciMult, to facilitate common knowledge sharing across different literature understanding tasks.
arXiv Detail & Related papers (2023-05-23T16:47:22Z) - Task Formulation Matters When Learning Continually: A Case Study in
Visual Question Answering [58.82325933356066]
Continual learning aims to train a model incrementally on a sequence of tasks without forgetting previous knowledge.
We present a detailed study of how different settings affect performance for Visual Question Answering.
arXiv Detail & Related papers (2022-09-30T19:12:58Z) - Fast Inference and Transfer of Compositional Task Structures for
Few-shot Task Generalization [101.72755769194677]
We formulate it as a few-shot reinforcement learning problem where a task is characterized by a subtask graph.
Our multi-task subtask graph inferencer (MTSGI) first infers the common high-level task structure in terms of the subtask graph from the training tasks.
Our experiment results on 2D grid-world and complex web navigation domains show that the proposed method can learn and leverage the common underlying structure of the tasks for faster adaptation to the unseen tasks.
arXiv Detail & Related papers (2022-05-25T10:44:25Z) - Explaining the Effectiveness of Multi-Task Learning for Efficient
Knowledge Extraction from Spine MRI Reports [2.5953185061765884]
We show that a single multi-tasking model can match the performance of task specific models.
We validate our observations on our internal radiologist-annotated datasets on the cervical and lumbar spine.
arXiv Detail & Related papers (2022-05-06T01:51:19Z) - The Effect of Diversity in Meta-Learning [79.56118674435844]
Few-shot learning aims to learn representations that can tackle novel tasks given a small number of examples.
Recent studies show that task distribution plays a vital role in the model's performance.
We study different task distributions on a myriad of models and datasets to evaluate the effect of task diversity on meta-learning algorithms.
arXiv Detail & Related papers (2022-01-27T19:39:07Z) - Divergent representations of ethological visual inputs emerge from
supervised, unsupervised, and reinforcement learning [20.98896935012773]
We compare the representations learned by eight different convolutional neural networks.
We find that the network trained with reinforcement learning differs most from the other networks.
arXiv Detail & Related papers (2021-12-03T17:18:09Z) - Self-Supervised Visual Representation Learning Using Lightweight
Architectures [0.0]
In self-supervised learning, a model is trained to solve a pretext task, using a data set whose annotations are created by a machine.
We critically examine the most notable pretext tasks to extract features from image data.
We study the performance of various self-supervised techniques keeping all other parameters uniform.
arXiv Detail & Related papers (2021-10-21T14:13:10Z) - Factors of Influence for Transfer Learning across Diverse Appearance
Domains and Task Types [50.1843146606122]
A simple form of transfer learning is common in current state-of-the-art computer vision models.
Previous systematic studies of transfer learning have been limited and the circumstances in which it is expected to work are not fully understood.
In this paper we carry out an extensive experimental exploration of transfer learning across vastly different image domains.
arXiv Detail & Related papers (2021-03-24T16:24:20Z) - Multi-Task Learning for Dense Prediction Tasks: A Survey [87.66280582034838]
Multi-task learning (MTL) techniques have shown promising results w.r.t. performance, computations and/or memory footprint.
We provide a well-rounded view on state-of-the-art deep learning approaches for MTL in computer vision.
arXiv Detail & Related papers (2020-04-28T09:15:50Z) - Analyzing Visual Representations in Embodied Navigation Tasks [45.35107294831313]
We use the recently proposed projection weighted Canonical Correlation Analysis (PWCCA) to measure the similarity of visual representations learned in the same environment by performing different tasks.
We then empirically demonstrate that visual representations learned on one task can be effectively transferred to a different task.
arXiv Detail & Related papers (2020-03-12T19:43:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.