Winning solutions and post-challenge analyses of the ChaLearn AutoDL
challenge 2019
- URL: http://arxiv.org/abs/2201.03801v1
- Date: Tue, 11 Jan 2022 06:21:18 GMT
- Title: Winning solutions and post-challenge analyses of the ChaLearn AutoDL
challenge 2019
- Authors: Zhengying Liu, Adrien Pavao, Zhen Xu, Sergio Escalera, Fabio Ferreira,
Isabelle Guyon, Sirui Hong, Frank Hutter, Rongrong Ji, Julio C. S. Jacques
Junior, Ge Li, Marius Lindauer, Zhipeng Luo, Meysam Madadi, Thomas Nierhoff,
Kangning Niu, Chunguang Pan, Danny Stoll, Sebastien Treguer, Jin Wang, Peng
Wang, Chenglin Wu, Youcheng Xiong, Arbe r Zela, Yang Zhang
- Abstract summary: This paper reports the results and post-challenge analyses of ChaLearn's AutoDL challenge series.
Results show that DL methods dominated, though popular Neural Architecture Search (NAS) was impractical.
A high level modular organization emerged featuring a "meta-learner", "data ingestor", "model selector", "model/learner", and "evaluator"
- Score: 112.36155380260655
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper reports the results and post-challenge analyses of ChaLearn's
AutoDL challenge series, which helped sorting out a profusion of AutoML
solutions for Deep Learning (DL) that had been introduced in a variety of
settings, but lacked fair comparisons. All input data modalities (time series,
images, videos, text, tabular) were formatted as tensors and all tasks were
multi-label classification problems. Code submissions were executed on hidden
tasks, with limited time and computational resources, pushing solutions that
get results quickly. In this setting, DL methods dominated, though popular
Neural Architecture Search (NAS) was impractical. Solutions relied on
fine-tuned pre-trained networks, with architectures matching data modality.
Post-challenge tests did not reveal improvements beyond the imposed time limit.
While no component is particularly original or novel, a high level modular
organization emerged featuring a "meta-learner", "data ingestor", "model
selector", "model/learner", and "evaluator". This modularity enabled ablation
studies, which revealed the importance of (off-platform) meta-learning,
ensembling, and efficient data management. Experiments on heterogeneous module
combinations further confirm the (local) optimality of the winning solutions.
Our challenge legacy includes an ever-lasting benchmark
(http://autodl.chalearn.org), the open-sourced code of the winners, and a free
"AutoDL self-service".
Related papers
- Adapting Vision-Language Models to Open Classes via Test-Time Prompt Tuning [50.26965628047682]
Adapting pre-trained models to open classes is a challenging problem in machine learning.
In this paper, we consider combining the advantages of both and come up with a test-time prompt tuning approach.
Our proposed method outperforms all comparison methods on average considering both base and new classes.
arXiv Detail & Related papers (2024-08-29T12:34:01Z) - Online Analytic Exemplar-Free Continual Learning with Large Models for Imbalanced Autonomous Driving Task [25.38082751323396]
We propose an Analytic Exemplar-Free Online Continual Learning algorithm (AEF-OCL)
The AEF-OCL leverages analytic continual learning principles and employs ridge regression as a classifier for features extracted by a large backbone network.
Experimental results demonstrate that despite being an exemplar-free strategy, our method outperforms various methods on the autonomous driving SODA10M dataset.
arXiv Detail & Related papers (2024-05-28T03:19:15Z) - Challenging Forgets: Unveiling the Worst-Case Forget Sets in Machine Unlearning [9.998859702421417]
Machine unlearning (MU) aims to eliminate the influence of chosen data points on model performance.
Despite various MU methods for data influence erasure, evaluations have largely focused on random data forgetting.
We propose identifying the data subset that presents the most significant challenge for influence erasure, pinpointing the worst-case forget set.
arXiv Detail & Related papers (2024-03-12T06:50:32Z) - Your Vision-Language Model Itself Is a Strong Filter: Towards
High-Quality Instruction Tuning with Data Selection [59.11430077029321]
We introduce a novel dataset selection method, Self-Filter, for vision-language models (VLMs)
In the first stage, we devise a scoring network to evaluate the difficulty of training instructions, which is co-trained with the VLM.
In the second stage, we use the trained score net to measure the difficulty of each instruction, select the most challenging samples, and penalize similar samples to encourage diversity.
arXiv Detail & Related papers (2024-02-19T20:08:48Z) - A Challenge in Reweighting Data with Bilevel Optimization [11.910900792664288]
Bilevel solvers are based on a warm-start strategy where both the parameters of the models and the data weights are learned at the same time.
We show that this joint dynamic may lead to sub-optimal solutions, for which the final data weights are very sparse.
This finding illustrates the difficulty of data reweighting and offers a clue as to why this method is rarely used in practice.
arXiv Detail & Related papers (2023-10-26T13:33:26Z) - Modular Neural Network Approaches for Surgical Image Recognition [0.0]
We introduce and evaluate different architectures of modular learning for Dorsal Capsulo-Scapholunate Septum (DCSS) instability classification.
Our experiments have shown that modular learning improves performances compared to non-modular systems.
In the second part, we present our approach for data labeling and segmentation with self-training applied on shoulder arthroscopy images.
arXiv Detail & Related papers (2023-07-17T22:28:16Z) - MT3: Meta Test-Time Training for Self-Supervised Test-Time Adaption [69.76837484008033]
An unresolved problem in Deep Learning is the ability of neural networks to cope with domain shifts during test-time.
We combine meta-learning, self-supervision and test-time training to learn to adapt to unseen test distributions.
Our approach significantly improves the state-of-the-art results on the CIFAR-10-Corrupted image classification benchmark.
arXiv Detail & Related papers (2021-03-30T09:33:38Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - Puzzle-AE: Novelty Detection in Images through Solving Puzzles [8.999416735254586]
U-Net is proved to be effective for this purpose but overfits on the training data if trained by just using reconstruction error similar to other AE-based frameworks.
We show that training U-Nets based on this task is an effective remedy that prevents overfitting and facilitates learning beyond pixel-level features.
We propose adversarial robust training as an effective automatic shortcut removal.
arXiv Detail & Related papers (2020-08-29T10:53:55Z) - ManyModalQA: Modality Disambiguation and QA over Diverse Inputs [73.93607719921945]
We present a new multimodal question answering challenge, ManyModalQA, in which an agent must answer a question by considering three distinct modalities.
We collect our data by scraping Wikipedia and then utilize crowdsourcing to collect question-answer pairs.
arXiv Detail & Related papers (2020-01-22T14:39:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.