Learning Instance and Task-Aware Dynamic Kernels for Few Shot Learning
- URL: http://arxiv.org/abs/2112.03494v1
- Date: Tue, 7 Dec 2021 04:52:36 GMT
- Title: Learning Instance and Task-Aware Dynamic Kernels for Few Shot Learning
- Authors: Rongkai Ma, Pengfei Fang, Gil Avraham, Yan Zuo, Tom Drummond, Mehrtash
Harandi
- Abstract summary: We learn the dynamic kernels of a convolution network as a function of the task at hand, enabling faster generalization.
We empirically show that our model improves performance on few-shot classification and detection tasks.
- Score: 32.3217883750605
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning and generalizing to novel concepts with few samples (Few-Shot
Learning) is still an essential challenge to real-world applications. A
principle way of achieving few-shot learning is to realize a model that can
rapidly adapt to the context of a given task. Dynamic networks have been shown
capable of learning content-adaptive parameters efficiently, making them
suitable for few-shot learning. In this paper, we propose to learn the dynamic
kernels of a convolution network as a function of the task at hand, enabling
faster generalization. To this end, we obtain our dynamic kernels based on the
entire task and each sample and develop a mechanism further conditioning on
each individual channel and position independently. This results in dynamic
kernels that simultaneously attend to the global information whilst also
considering minuscule details available. We empirically show that our model
improves performance on few-shot classification and detection tasks, achieving
a tangible improvement over several baseline models. This includes
state-of-the-art results on 4 few-shot classification benchmarks:
mini-ImageNet, tiered-ImageNet, CUB and FC100 and competitive results on a
few-shot detection dataset: MS COCO-PASCAL-VOC.
Related papers
- AggSS: An Aggregated Self-Supervised Approach for Class-Incremental Learning [17.155759991260094]
This paper investigates the impact of self-supervised learning, specifically image rotations, on various class-incremental learning paradigms.
We observe a shift in the deep neural network's attention towards intrinsic object features as it learns through AggSS strategy.
AggSS serves as a plug-and-play module that can be seamlessly incorporated into any class-incremental learning framework.
arXiv Detail & Related papers (2024-08-08T10:16:02Z) - Intra-task Mutual Attention based Vision Transformer for Few-Shot Learning [12.5354658533836]
Humans possess remarkable ability to accurately classify new, unseen images after being exposed to only a few examples.
For artificial neural network models, determining the most relevant features for distinguishing between two images with limited samples presents a challenge.
We propose an intra-task mutual attention method for few-shot learning, that involves splitting the support and query samples into patches.
arXiv Detail & Related papers (2024-05-06T02:02:57Z) - An Efficient General-Purpose Modular Vision Model via Multi-Task
Heterogeneous Training [79.78201886156513]
We present a model that can perform multiple vision tasks and can be adapted to other downstream tasks efficiently.
Our approach achieves comparable results to single-task state-of-the-art models and demonstrates strong generalization on downstream tasks.
arXiv Detail & Related papers (2023-06-29T17:59:57Z) - USER: Unified Semantic Enhancement with Momentum Contrast for Image-Text
Retrieval [115.28586222748478]
Image-Text Retrieval (ITR) aims at searching for the target instances that are semantically relevant to the given query from the other modality.
Existing approaches typically suffer from two major limitations.
arXiv Detail & Related papers (2023-01-17T12:42:58Z) - Meta-Learning with Self-Improving Momentum Target [72.98879709228981]
We propose Self-improving Momentum Target (SiMT) to improve the performance of a meta-learner.
SiMT generates the target model by adapting from the temporal ensemble of the meta-learner.
We show that SiMT brings a significant performance gain when combined with a wide range of meta-learning methods.
arXiv Detail & Related papers (2022-10-11T06:45:15Z) - Dynamic Kernel Selection for Improved Generalization and Memory
Efficiency in Meta-learning [9.176056742068813]
We present MetaDOCK, a task-specific dynamic kernel selection strategy for designing compressed CNN models.
Our method is based on the hypothesis that for a given set of similar tasks, not all kernels of the network are needed by each individual task.
We show that for the same inference budget, pruned versions of large CNN models obtained using our approach consistently outperform the conventional choices of CNN models.
arXiv Detail & Related papers (2022-06-03T17:09:26Z) - Continual Object Detection via Prototypical Task Correlation Guided
Gating Mechanism [120.1998866178014]
We present a flexible framework for continual object detection via pRotOtypical taSk corrElaTion guided gaTingAnism (ROSETTA)
Concretely, a unified framework is shared by all tasks while task-aware gates are introduced to automatically select sub-models for specific tasks.
Experiments on COCO-VOC, KITTI-Kitchen, class-incremental detection on VOC and sequential learning of four tasks show that ROSETTA yields state-of-the-art performance.
arXiv Detail & Related papers (2022-05-06T07:31:28Z) - Dynamic Relevance Learning for Few-Shot Object Detection [6.550840743803705]
We propose a dynamic relevance learning model, which utilizes the relationship between all support images and Region of Interest (RoI) on the query images to construct a dynamic graph convolutional network (GCN)
The proposed model achieves the best overall performance, which shows its effectiveness of learning more generalized features.
arXiv Detail & Related papers (2021-08-04T18:29:42Z) - Learning to Continuously Optimize Wireless Resource In Episodically
Dynamic Environment [55.91291559442884]
This work develops a methodology that enables data-driven methods to continuously learn and optimize in a dynamic environment.
We propose to build the notion of continual learning into the modeling process of learning wireless systems.
Our design is based on a novel min-max formulation which ensures certain fairness" across different data samples.
arXiv Detail & Related papers (2020-11-16T08:24:34Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.