1st Place in ICCV 2023 Workshop Challenge Track 1 on Resource Efficient
Deep Learning for Computer Vision: Budgeted Model Training Challenge
- URL: http://arxiv.org/abs/2311.11470v1
- Date: Wed, 9 Aug 2023 05:38:18 GMT
- Title: 1st Place in ICCV 2023 Workshop Challenge Track 1 on Resource Efficient
Deep Learning for Computer Vision: Budgeted Model Training Challenge
- Authors: Youngjun Kwak, Seonghun Jeong, Yunseung Lee, Changick Kim
- Abstract summary: We describe a resource-aware backbone search framework composed of profile and instantiation phases.
We employ multi-resolution ensembles to boost inference accuracy on limited resources.
Based on our approach, we win first place in International Conference on Computer Vision (ICCV) 2023 Workshop Challenge Track 1 on Resource Efficient Deep Learning for Computer Vision (RCV)
- Score: 15.213786895534225
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The budgeted model training challenge aims to train an efficient
classification model under resource limitations. To tackle this task in
ImageNet-100, we describe a simple yet effective resource-aware backbone search
framework composed of profile and instantiation phases. In addition, we employ
multi-resolution ensembles to boost inference accuracy on limited resources.
The profile phase obeys time and memory constraints to determine the models'
optimal batch-size, max epochs, and automatic mixed precision (AMP). And the
instantiation phase trains models with the determined parameters from the
profile phase. For improving intra-domain generalizations, the multi-resolution
ensembles are formed by two-resolution images with randomly applied flips. We
present a comprehensive analysis with expensive experiments. Based on our
approach, we win first place in International Conference on Computer Vision
(ICCV) 2023 Workshop Challenge Track 1 on Resource Efficient Deep Learning for
Computer Vision (RCV).
Related papers
- Precise Drive with VLM: First Prize Solution for PRCV 2024 Drive LM challenge [8.941623670652389]
This report outlines the methodologies we applied for the PRCV Challenge.
It focuses on cognition and decision-making in driving scenarios.
Our model achieved a score of 0.6064, securing the first prize on the competition's final results.
arXiv Detail & Related papers (2024-11-05T11:00:55Z) - Towards Efficient Pareto Set Approximation via Mixture of Experts Based Model Fusion [53.33473557562837]
Solving multi-objective optimization problems for large deep neural networks is a challenging task due to the complexity of the loss landscape and the expensive computational cost.
We propose a practical and scalable approach to solve this problem via mixture of experts (MoE) based model fusion.
By ensembling the weights of specialized single-task models, the MoE module can effectively capture the trade-offs between multiple objectives.
arXiv Detail & Related papers (2024-06-14T07:16:18Z) - Inter-slice Super-resolution of Magnetic Resonance Images by Pre-training and Self-supervised Fine-tuning [49.197385954021456]
In clinical practice, 2D magnetic resonance (MR) sequences are widely adopted. While individual 2D slices can be stacked to form a 3D volume, the relatively large slice spacing can pose challenges for visualization and subsequent analysis tasks.
To reduce slice spacing, deep-learning-based super-resolution techniques are widely investigated.
Most current solutions require a substantial number of paired high-resolution and low-resolution images for supervised training, which are typically unavailable in real-world scenarios.
arXiv Detail & Related papers (2024-06-10T02:20:26Z) - RL-I2IT: Image-to-Image Translation with Deep Reinforcement Learning [54.40719981158774]
Image-to-Image Translation (I2IT) methods generate images in a single run of a deep learning (DL) model.
We reformulate I2IT as a step-wise decision-making problem via deep reinforcement learning (DRL)
We propose a novel framework that performs RL-based I2IT (RL-I2IT)
arXiv Detail & Related papers (2023-09-24T15:40:40Z) - No Data Augmentation? Alternative Regularizations for Effective Training
on Small Datasets [0.0]
We study alternative regularization strategies to push the limits of supervised learning on small image classification datasets.
In particular, we employ a agnostic to select (semi) optimal learning rate and weight decay couples via the norm of model parameters.
We reach a test accuracy of 66.5%, on par with the best state-of-the-art methods.
arXiv Detail & Related papers (2023-09-04T16:13:59Z) - On Efficient Transformer and Image Pre-training for Low-level Vision [74.22436001426517]
Pre-training has marked numerous state of the arts in high-level computer vision.
We present an in-depth study of image pre-training.
We find pre-training plays strikingly different roles in low-level tasks.
arXiv Detail & Related papers (2021-12-19T15:50:48Z) - Multi-Scale Aligned Distillation for Low-Resolution Detection [68.96325141432078]
This paper focuses on boosting the performance of low-resolution models by distilling knowledge from a high- or multi-resolution model.
On several instance-level detection tasks and datasets, the low-resolution models trained via our approach perform competitively with high-resolution models trained via conventional multi-scale training.
arXiv Detail & Related papers (2021-09-14T12:53:35Z) - Ultra-Data-Efficient GAN Training: Drawing A Lottery Ticket First, Then
Training It Toughly [114.81028176850404]
Training generative adversarial networks (GANs) with limited data generally results in deteriorated performance and collapsed models.
We decompose the data-hungry GAN training into two sequential sub-problems.
Such a coordinated framework enables us to focus on lower-complexity and more data-efficient sub-problems.
arXiv Detail & Related papers (2021-02-28T05:20:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.