Self-supervised U-net for few-shot learning of object segmentation in
microscopy images
- URL: http://arxiv.org/abs/2205.10840v1
- Date: Sun, 22 May 2022 14:54:25 GMT
- Title: Self-supervised U-net for few-shot learning of object segmentation in
microscopy images
- Authors: Arnaud Deleruyelle, Cristian Versari, John Klein
- Abstract summary: Self-supervision has proved to significantly increase generalization performances of models trained on few shots.
This paper introduces one such neural pipeline in the context of microscopic image segmentation.
- Score: 4.125187280299247
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: State-of-the-art segmentation performances are achieved by deep neural
networks. Training these networks from only a few training examples is
challenging while producing annotated images that provide supervision is
tedious. Recently, self-supervision, i.e. designing a neural pipeline providing
synthetic or indirect supervision, has proved to significantly increase
generalization performances of models trained on few shots. This paper
introduces one such neural pipeline in the context of microscopic image
segmentation. By leveraging the rather simple content of these images a trainee
network can be mentored by a referee network which has been previously trained
on synthetically generated pairs of corrupted/correct region masks.
Related papers
- Efficient Visualization of Neural Networks with Generative Models and Adversarial Perturbations [0.0]
This paper presents a novel approach for deep visualization via a generative network, offering an improvement over existing methods.
Our model simplifies the architecture by reducing the number of networks used, requiring only a generator and a discriminator.
Our model requires less prior training knowledge and uses a non-adversarial training process, where the discriminator acts as a guide.
arXiv Detail & Related papers (2024-09-20T14:59:25Z) - Image segmentation with traveling waves in an exactly solvable recurrent
neural network [71.74150501418039]
We show that a recurrent neural network can effectively divide an image into groups according to a scene's structural characteristics.
We present a precise description of the mechanism underlying object segmentation in this network.
We then demonstrate a simple algorithm for object segmentation that generalizes across inputs ranging from simple geometric objects in grayscale images to natural images.
arXiv Detail & Related papers (2023-11-28T16:46:44Z) - Adversarially robust segmentation models learn perceptually-aligned
gradients [0.0]
We show that adversarially-trained semantic segmentation networks can be used to perform image inpainting and generation.
We argue that perceptually-aligned gradients promote a better understanding of a neural network's learned representations and aid in making neural networks more interpretable.
arXiv Detail & Related papers (2022-04-03T16:04:52Z) - Is Deep Image Prior in Need of a Good Education? [57.3399060347311]
Deep image prior was introduced as an effective prior for image reconstruction.
Despite its impressive reconstructive properties, the approach is slow when compared to learned or traditional reconstruction techniques.
We develop a two-stage learning paradigm to address the computational challenge.
arXiv Detail & Related papers (2021-11-23T15:08:26Z) - Aesthetics and neural network image representations [0.0]
We analyze the spaces of images encoded by generative neural networks of the BigGAN architecture.
We find that generic multiplicative perturbations of neural network parameters away from the photo-realistic point often lead to networks generating images which appear as "artistic renditions" of the corresponding objects.
arXiv Detail & Related papers (2021-09-16T16:50:22Z) - Towards robust vision by multi-task learning on monkey visual cortex [6.9014416935919565]
We jointly trained a deep network to perform image classification and to predict neural activity in macaque primary visual cortex (V1)
We found that co-training on monkey V1 data leads to increased robustness despite the absence of those distortions during training.
Our results also demonstrated that the network's representations become more brain-like as their robustness improves.
arXiv Detail & Related papers (2021-07-29T21:55:48Z) - aura-net : robust segmentation of phase-contrast microscopy images with
few annotations [2.1574781022415364]
AURA-net is a convolutional neural network (CNN) for the segmentation of phase-contrast microscopy images.
We show that AURA-net outperforms state-of-the-art alternatives in several small (less than 100images) datasets.
arXiv Detail & Related papers (2021-02-02T08:47:14Z) - Counterfactual Generative Networks [59.080843365828756]
We propose to decompose the image generation process into independent causal mechanisms that we train without direct supervision.
By exploiting appropriate inductive biases, these mechanisms disentangle object shape, object texture, and background.
We show that the counterfactual images can improve out-of-distribution with a marginal drop in performance on the original classification task.
arXiv Detail & Related papers (2021-01-15T10:23:12Z) - Understanding the Role of Individual Units in a Deep Neural Network [85.23117441162772]
We present an analytic framework to systematically identify hidden units within image classification and image generation networks.
First, we analyze a convolutional neural network (CNN) trained on scene classification and discover units that match a diverse set of object concepts.
Second, we use a similar analytic method to analyze a generative adversarial network (GAN) model trained to generate scenes.
arXiv Detail & Related papers (2020-09-10T17:59:10Z) - Adversarially-Trained Deep Nets Transfer Better: Illustration on Image
Classification [53.735029033681435]
Transfer learning is a powerful methodology for adapting pre-trained deep neural networks on image recognition tasks to new domains.
In this work, we demonstrate that adversarially-trained models transfer better than non-adversarially-trained models.
arXiv Detail & Related papers (2020-07-11T22:48:42Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.