VIPriors 2: Visual Inductive Priors for Data-Efficient Deep Learning
Challenges
- URL: http://arxiv.org/abs/2201.08625v1
- Date: Fri, 21 Jan 2022 10:20:52 GMT
- Title: VIPriors 2: Visual Inductive Priors for Data-Efficient Deep Learning
Challenges
- Authors: Attila Lengyel, Robert-Jan Bruintjes, Marcos Baptista Rios, Osman
Semih Kayhan, Davide Zambrano, Nergis Tomen, Jan van Gemert
- Abstract summary: Second edition of "VIPriors: Visual Inductive Priors for Data-Efficient Deep Learning" challenges.
Models are trained from scratch on a reduced number of training samples for various key computer vision tasks.
Results: The provided baselines are outperformed by a large margin in all five challenges.
- Score: 13.085098213230568
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The second edition of the "VIPriors: Visual Inductive Priors for
Data-Efficient Deep Learning" challenges featured five data-impaired
challenges, where models are trained from scratch on a reduced number of
training samples for various key computer vision tasks. To encourage new and
creative ideas on incorporating relevant inductive biases to improve the data
efficiency of deep learning models, we prohibited the use of pre-trained
checkpoints and other transfer learning techniques. The provided baselines are
outperformed by a large margin in all five challenges, mainly thanks to
extensive data augmentation policies, model ensembling, and data efficient
network architectures.
Related papers
- VIPriors 4: Visual Inductive Priors for Data-Efficient Deep Learning Challenges [12.615348941903594]
Fourth edition of the "VIPriors: Visual Inductive Priors for Data-Efficient Deep Learning" workshop features two data-impaired challenges.
These challenges address the problem of training deep learning models for computer vision tasks with limited data.
We aim to stimulate the development of novel approaches that incorporate inductive biases to improve the data efficiency of deep learning models.
arXiv Detail & Related papers (2024-06-26T08:50:51Z) - Augment Before Copy-Paste: Data and Memory Efficiency-Oriented Instance Segmentation Framework for Sport-scenes [7.765333471208582]
In Visual Inductive Priors challenge (VIPriors2023), participants must train a model capable of precisely locating individuals on a basketball court.
We propose memory effIciency inStance framework based on visual inductive prior flow propagation.
Experiments demonstrate our model promising performance even under limited data and memory constraints.
arXiv Detail & Related papers (2024-03-18T08:44:40Z) - Less is More: High-value Data Selection for Visual Instruction Tuning [127.38740043393527]
We propose a high-value data selection approach TIVE, to eliminate redundancy within the visual instruction data and reduce the training cost.
Our approach using only about 15% data can achieve comparable average performance to the full-data fine-tuned model across eight benchmarks.
arXiv Detail & Related papers (2024-03-14T16:47:25Z) - Diffusion-Based Neural Network Weights Generation [80.89706112736353]
D2NWG is a diffusion-based neural network weights generation technique that efficiently produces high-performing weights for transfer learning.
Our method extends generative hyper-representation learning to recast the latent diffusion paradigm for neural network weights generation.
Our approach is scalable to large architectures such as large language models (LLMs), overcoming the limitations of current parameter generation techniques.
arXiv Detail & Related papers (2024-02-28T08:34:23Z) - VIPriors 3: Visual Inductive Priors for Data-Efficient Deep Learning
Challenges [13.085098213230568]
Third edition of the "VIPriors: Visual Inductive Priors for Data-Efficient Deep Learning" workshop featured four data-impaired challenges.
Challenges focused on addressing the limitations of data availability in training deep learning models for computer vision tasks.
arXiv Detail & Related papers (2023-05-31T09:31:54Z) - Offline Q-Learning on Diverse Multi-Task Data Both Scales And
Generalizes [100.69714600180895]
offline Q-learning algorithms exhibit strong performance that scales with model capacity.
We train a single policy on 40 games with near-human performance using up-to 80 million parameter networks.
Compared to return-conditioned supervised approaches, offline Q-learning scales similarly with model capacity and has better performance, especially when the dataset is suboptimal.
arXiv Detail & Related papers (2022-11-28T08:56:42Z) - Multi-dataset Training of Transformers for Robust Action Recognition [75.5695991766902]
We study the task of robust feature representations, aiming to generalize well on multiple datasets for action recognition.
Here, we propose a novel multi-dataset training paradigm, MultiTrain, with the design of two new loss terms, namely informative loss and projection loss.
We verify the effectiveness of our method on five challenging datasets, Kinetics-400, Kinetics-700, Moments-in-Time, Activitynet and Something-something-v2.
arXiv Detail & Related papers (2022-09-26T01:30:43Z) - Self-Supervised Representation Learning: Introduction, Advances and
Challenges [125.38214493654534]
Self-supervised representation learning methods aim to provide powerful deep feature learning without the requirement of large annotated datasets.
This article introduces this vibrant area including key concepts, the four main families of approach and associated state of the art, and how self-supervised methods are applied to diverse modalities of data.
arXiv Detail & Related papers (2021-10-18T13:51:22Z) - Training Deep Networks from Zero to Hero: avoiding pitfalls and going
beyond [59.94347858883343]
This tutorial covers the basic steps as well as more recent options to improve models.
It can be particularly useful in datasets that are not as well-prepared as those in challenges.
arXiv Detail & Related papers (2021-09-06T21:31:42Z) - VIPriors 1: Visual Inductive Priors for Data-Efficient Deep Learning
Challenges [8.50468505606714]
We offer four data-impaired challenges, where models are trained from scratch, and we reduce the number of training samples to a fraction of the full set.
To encourage data efficient solutions, we prohibited the use of pre-trained models and other transfer learning techniques.
arXiv Detail & Related papers (2021-03-05T15:58:17Z) - Improving Deep Learning Models via Constraint-Based Domain Knowledge: a
Brief Survey [11.034875974800487]
This paper presents a first survey of the approaches devised to integrate domain knowledge, expressed in the form of constraints, in Deep Learning (DL) learning models.
We identify five categories that encompass the main approaches to inject domain knowledge: 1) acting on the features space, 2) modifications to the hypothesis space, 3) data augmentation, 4) regularization schemes, 5) constrained learning.
arXiv Detail & Related papers (2020-05-19T15:34:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.