Meta-DRN: Meta-Learning for 1-Shot Image Segmentation
- URL: http://arxiv.org/abs/2008.00247v1
- Date: Sat, 1 Aug 2020 11:23:37 GMT
- Title: Meta-DRN: Meta-Learning for 1-Shot Image Segmentation
- Authors: Atmadeep Banerjee
- Abstract summary: We propose a novel lightweight CNN architecture for 1-shot image segmentation.
We train our model using 4 meta-learning algorithms that have worked well for image classification and compare the results.
- Score: 0.12691047660244334
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modern deep learning models have revolutionized the field of computer vision.
But, a significant drawback of most of these models is that they require a
large number of labelled examples to generalize properly. Recent developments
in few-shot learning aim to alleviate this requirement. In this paper, we
propose a novel lightweight CNN architecture for 1-shot image segmentation. The
proposed model is created by taking inspiration from well-performing
architectures for semantic segmentation and adapting it to the 1-shot domain.
We train our model using 4 meta-learning algorithms that have worked well for
image classification and compare the results. For the chosen dataset, our
proposed model has a 70% lower parameter count than the benchmark, while having
better or comparable mean IoU scores using all 4 of the meta-learning
algorithms.
Related papers
- SPPNet: A Single-Point Prompt Network for Nuclei Image Segmentation [6.149725843029721]
Single-point prompt network is proposed for nuclei image segmentation.
We replace the original image encoder with a lightweight vision transformer.
The proposed model is evaluated on the MoNuSeg-2018 dataset.
arXiv Detail & Related papers (2023-08-23T16:13:58Z) - MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments [72.6405488990753]
Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks.
We propose a single-stage and standalone method, MOCA, which unifies both desired properties.
We achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols.
arXiv Detail & Related papers (2023-07-18T15:46:20Z) - Improving Zero-shot Generalization and Robustness of Multi-modal Models [70.14692320804178]
Multi-modal image-text models such as CLIP and LiT have demonstrated impressive performance on image classification benchmarks.
We investigate the reasons for this performance gap and find that many of the failure cases are caused by ambiguity in the text prompts.
We propose a simple and efficient way to improve accuracy on such uncertain images by making use of the WordNet hierarchy.
arXiv Detail & Related papers (2022-12-04T07:26:24Z) - Few-Shot Learning of Compact Models via Task-Specific Meta Distillation [16.683801607142257]
We consider a new problem of few-shot learning of compact models.
We propose task-specific meta distillation that simultaneously learns two models in meta-learning.
arXiv Detail & Related papers (2022-10-18T15:06:47Z) - Dynamic Kernel Selection for Improved Generalization and Memory
Efficiency in Meta-learning [9.176056742068813]
We present MetaDOCK, a task-specific dynamic kernel selection strategy for designing compressed CNN models.
Our method is based on the hypothesis that for a given set of similar tasks, not all kernels of the network are needed by each individual task.
We show that for the same inference budget, pruned versions of large CNN models obtained using our approach consistently outperform the conventional choices of CNN models.
arXiv Detail & Related papers (2022-06-03T17:09:26Z) - Adaptive Convolutional Dictionary Network for CT Metal Artifact
Reduction [62.691996239590125]
We propose an adaptive convolutional dictionary network (ACDNet) for metal artifact reduction.
Our ACDNet can automatically learn the prior for artifact-free CT images via training data and adaptively adjust the representation kernels for each input CT image.
Our method inherits the clear interpretability of model-based methods and maintains the powerful representation ability of learning-based methods.
arXiv Detail & Related papers (2022-05-16T06:49:36Z) - Group-Wise Semantic Mining for Weakly Supervised Semantic Segmentation [49.90178055521207]
This work addresses weakly supervised semantic segmentation (WSSS), with the goal of bridging the gap between image-level annotations and pixel-level segmentation.
We formulate WSSS as a novel group-wise learning task that explicitly models semantic dependencies in a group of images to estimate more reliable pseudo ground-truths.
In particular, we devise a graph neural network (GNN) for group-wise semantic mining, wherein input images are represented as graph nodes.
arXiv Detail & Related papers (2020-12-09T12:40:13Z) - Rethinking Few-Shot Image Classification: a Good Embedding Is All You
Need? [72.00712736992618]
We show that a simple baseline: learning a supervised or self-supervised representation on the meta-training set, outperforms state-of-the-art few-shot learning methods.
An additional boost can be achieved through the use of self-distillation.
We believe that our findings motivate a rethinking of few-shot image classification benchmarks and the associated role of meta-learning algorithms.
arXiv Detail & Related papers (2020-03-25T17:58:42Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z) - Self-Learning AI Framework for Skin Lesion Image Segmentation and
Classification [0.0]
To perform medical image segmentation with deep learning models, it requires training on large image dataset with annotation.
To overcome this issue, self-learning annotation scheme was proposed in the two-stage deep learning algorithm.
The classification results of the proposed AI framework achieved training accuracy of 93.8% and testing accuracy of 82.42%.
arXiv Detail & Related papers (2020-01-04T09:31:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.