Deep Consensus Learning
- URL: http://arxiv.org/abs/2103.08475v1
- Date: Mon, 15 Mar 2021 15:51:14 GMT
- Title: Deep Consensus Learning
- Authors: Wei Sun and Tianfu Wu
- Abstract summary: This paper proposes deep consensus learning for layout-to-image synthesis and weakly-supervised image semantic segmentation.
Two deep consensus mappings are exploited to facilitate training the three networks end-to-end.
It obtains compelling layout-to-image synthesis results and weakly-supervised image semantic segmentation results.
- Score: 16.834584070973676
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Both generative learning and discriminative learning have recently witnessed
remarkable progress using Deep Neural Networks (DNNs). For structured input
synthesis and structured output prediction problems (e.g., layout-to-image
synthesis and image semantic segmentation respectively), they often are studied
separately. This paper proposes deep consensus learning (DCL) for joint
layout-to-image synthesis and weakly-supervised image semantic segmentation.
The former is realized by a recently proposed LostGAN approach, and the latter
by introducing an inference network as the third player joining the two-player
game of LostGAN. Two deep consensus mappings are exploited to facilitate
training the three networks end-to-end: Given an input layout (a list of object
bounding boxes), the generator generates a mask (label map) and then use it to
help synthesize an image. The inference network infers the mask for the
synthesized image. Then, the latent consensus is measured between the mask
generated by the generator and the one inferred by the inference network. For
the real image corresponding to the input layout, its mask also is computed by
the inference network, and then used by the generator to reconstruct the real
image. Then, the data consensus is measured between the real image and its
reconstructed image. The discriminator still plays the role of an adversary by
computing the realness scores for a real image, its reconstructed image and a
synthesized image. In experiments, our DCL is tested in the COCO-Stuff dataset.
It obtains compelling layout-to-image synthesis results and weakly-supervised
image semantic segmentation results.
Related papers
- Hamming Similarity and Graph Laplacians for Class Partitioning and
Adversarial Image Detection [2.960821510561423]
We investigate the potential for ReLU activation patterns (encoded as bit vectors) to aid in understanding and interpreting the behavior of neural networks.
We utilize Representational Dissimilarity Matrices (RDMs) to investigate the coherence of data within the embedding spaces of a deep neural network.
We demonstrate that bit vectors aid in adversarial image detection, again achieving over 95% accuracy in separating adversarial and non-adversarial images.
arXiv Detail & Related papers (2023-05-02T22:16:15Z) - Extracting Semantic Knowledge from GANs with Unsupervised Learning [65.32631025780631]
Generative Adversarial Networks (GANs) encode semantics in feature maps in a linearly separable form.
We propose a novel clustering algorithm, named KLiSH, which leverages the linear separability to cluster GAN's features.
KLiSH succeeds in extracting fine-grained semantics of GANs trained on datasets of various objects.
arXiv Detail & Related papers (2022-11-30T03:18:16Z) - SAC-GAN: Structure-Aware Image-to-Image Composition for Self-Driving [18.842432515507035]
We present a compositional approach to image augmentation for self-driving applications.
It is an end-to-end neural network that is trained to seamlessly compose an object represented as a cropped patch from an object image, into a background scene image.
We evaluate our network, coined SAC-GAN for structure-aware composition, on prominent self-driving datasets in terms of quality, composability, and generalizability of the composite images.
arXiv Detail & Related papers (2021-12-13T12:24:50Z) - A Shared Representation for Photorealistic Driving Simulators [83.5985178314263]
We propose to improve the quality of generated images by rethinking the discriminator architecture.
The focus is on the class of problems where images are generated given semantic inputs, such as scene segmentation maps or human body poses.
We aim to learn a shared latent representation that encodes enough information to jointly do semantic segmentation, content reconstruction, along with a coarse-to-fine grained adversarial reasoning.
arXiv Detail & Related papers (2021-12-09T18:59:21Z) - Is Deep Image Prior in Need of a Good Education? [57.3399060347311]
Deep image prior was introduced as an effective prior for image reconstruction.
Despite its impressive reconstructive properties, the approach is slow when compared to learned or traditional reconstruction techniques.
We develop a two-stage learning paradigm to address the computational challenge.
arXiv Detail & Related papers (2021-11-23T15:08:26Z) - USIS: Unsupervised Semantic Image Synthesis [9.613134538472801]
We propose a new Unsupervised paradigm for Semantic Image Synthesis (USIS)
USIS learns to output images with visually separable semantic classes using a self-supervised segmentation loss.
In order to match the color and texture distribution of real images without losing high-frequency information, we propose to use whole image wavelet-based discrimination.
arXiv Detail & Related papers (2021-09-29T20:48:41Z) - Ensembling with Deep Generative Views [72.70801582346344]
generative models can synthesize "views" of artificial images that mimic real-world variations, such as changes in color or pose.
Here, we investigate whether such views can be applied to real images to benefit downstream analysis tasks such as image classification.
We use StyleGAN2 as the source of generative augmentations and investigate this setup on classification tasks involving facial attributes, cat faces, and cars.
arXiv Detail & Related papers (2021-04-29T17:58:35Z) - Synthetic Glacier SAR Image Generation from Arbitrary Masks Using
Pix2Pix Algorithm [12.087729834358928]
Supervised machine learning requires a large amount of labeled data to achieve proper test results.
In this work, we propose to alleviate the issue of limited training data by generating synthetic SAR images with the pix2pix algorithm.
We present different models, perform a comparative study and demonstrate that this approach synthesizes convincing glaciers in SAR images with promising qualitative and quantitative results.
arXiv Detail & Related papers (2021-01-08T23:30:00Z) - Pairwise Relation Learning for Semi-supervised Gland Segmentation [90.45303394358493]
We propose a pairwise relation-based semi-supervised (PRS2) model for gland segmentation on histology images.
This model consists of a segmentation network (S-Net) and a pairwise relation network (PR-Net)
We evaluate our model against five recent methods on the GlaS dataset and three recent methods on the CRAG dataset.
arXiv Detail & Related papers (2020-08-06T15:02:38Z) - Edge Guided GANs with Contrastive Learning for Semantic Image Synthesis [194.1452124186117]
We propose a novel ECGAN for the challenging semantic image synthesis task.
Our ECGAN achieves significantly better results than state-of-the-art methods.
arXiv Detail & Related papers (2020-03-31T01:23:21Z) - Image Embedded Segmentation: Uniting Supervised and Unsupervised
Objectives for Segmenting Histopathological Images [0.0]
This paper presents a new regularization method to train a fully convolutional network for semantic tissue segmentation.
It relies on the benefit of unsupervised learning, in the form of image reconstruction, for network training.
Our experiments demonstrate that it leads to better segmentation results in these datasets, compared to its counterparts.
arXiv Detail & Related papers (2020-01-30T08:09:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.