Self-Ensembling GAN for Cross-Domain Semantic Segmentation
- URL: http://arxiv.org/abs/2112.07999v1
- Date: Wed, 15 Dec 2021 09:50:25 GMT
- Title: Self-Ensembling GAN for Cross-Domain Semantic Segmentation
- Authors: Yonghao Xu, Fengxiang He, Bo Du, Liangpei Zhang, Dacheng Tao
- Abstract summary: This paper proposes a self-ensembling generative adversarial network (SE-GAN) exploiting cross-domain data for semantic segmentation.
In SE-GAN, a teacher network and a student network constitute a self-ensembling model for generating semantic segmentation maps, which together with a discriminator, forms a GAN.
Despite its simplicity, we find SE-GAN can significantly boost the performance of adversarial training and enhance the stability of the model.
- Score: 107.27377745720243
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks (DNNs) have greatly contributed to the performance gains
in semantic segmentation. Nevertheless, training DNNs generally requires large
amounts of pixel-level labeled data, which is expensive and time-consuming to
collect in practice. To mitigate the annotation burden, this paper proposes a
self-ensembling generative adversarial network (SE-GAN) exploiting cross-domain
data for semantic segmentation. In SE-GAN, a teacher network and a student
network constitute a self-ensembling model for generating semantic segmentation
maps, which together with a discriminator, forms a GAN. Despite its simplicity,
we find SE-GAN can significantly boost the performance of adversarial training
and enhance the stability of the model, the latter of which is a common barrier
shared by most adversarial training-based methods. We theoretically analyze
SE-GAN and provide an $\mathcal O(1/\sqrt{N})$ generalization bound ($N$ is the
training sample size), which suggests controlling the discriminator's
hypothesis complexity to enhance the generalizability. Accordingly, we choose a
simple network as the discriminator. Extensive and systematic experiments in
two standard settings demonstrate that the proposed method significantly
outperforms current state-of-the-art approaches. The source code of our model
will be available soon.
Related papers
- A Likelihood Ratio-Based Approach to Segmenting Unknown Objects [4.000869978312742]
Outlier supervision is a widely used strategy for improving OoD detection of the existing segmentation networks.
We propose an adaptive, lightweight unknown estimation module (UEM) for outlier supervision.
Our approach achieves a new state-of-the-art across multiple datasets, outperforming the previous best method by 5.74% average precision points.
arXiv Detail & Related papers (2024-09-10T11:10:32Z) - Supervised Gradual Machine Learning for Aspect Category Detection [0.9857683394266679]
Aspect Category Detection (ACD) aims to identify implicit and explicit aspects in a given review sentence.
We propose a novel approach to tackle the ACD task by combining Deep Neural Networks (DNNs) with Gradual Machine Learning (GML) in a supervised setting.
arXiv Detail & Related papers (2024-04-08T07:21:46Z) - Prompting Diffusion Representations for Cross-Domain Semantic
Segmentation [101.04326113360342]
diffusion-pretraining achieves extraordinary domain generalization results for semantic segmentation.
We introduce a scene prompt and a prompt randomization strategy to help further disentangle the domain-invariant information when training the segmentation head.
arXiv Detail & Related papers (2023-07-05T09:28:25Z) - Unsupervised Domain-adaptive Hash for Networks [81.49184987430333]
Domain-adaptive hash learning has enjoyed considerable success in the computer vision community.
We develop an unsupervised domain-adaptive hash learning method for networks, dubbed UDAH.
arXiv Detail & Related papers (2021-08-20T12:09:38Z) - DANCE: DAta-Network Co-optimization for Efficient Segmentation Model
Training and Inference [85.02494022662505]
DANCE is an automated simultaneous data-network co-optimization for efficient segmentation model training and inference.
It integrates automated data slimming which adaptively downsamples/drops input images and controls their corresponding contribution to the training loss guided by the images' spatial complexity.
Experiments and ablating studies demonstrate that DANCE can achieve "all-win" towards efficient segmentation.
arXiv Detail & Related papers (2021-07-16T04:58:58Z) - Learning a Domain-Agnostic Visual Representation for Autonomous Driving
via Contrastive Loss [25.798361683744684]
Domain-Agnostic Contrastive Learning (DACL) is a two-stage unsupervised domain adaptation framework with cyclic adversarial training and contrastive loss.
Our proposed approach achieves better performance in the monocular depth estimation task compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-03-10T07:06:03Z) - S2-BNN: Bridging the Gap Between Self-Supervised Real and 1-bit Neural
Networks via Guided Distribution Calibration [74.5509794733707]
We present a novel guided learning paradigm from real-valued to distill binary networks on the final prediction distribution.
Our proposed method can boost the simple contrastive learning baseline by an absolute gain of 5.515% on BNNs.
Our method achieves substantial improvement over the simple contrastive learning baseline, and is even comparable to many mainstream supervised BNN methods.
arXiv Detail & Related papers (2021-02-17T18:59:28Z) - EC-GAN: Low-Sample Classification using Semi-Supervised Algorithms and
GANs [0.0]
Semi-supervised learning has been gaining attention as it allows for performing image analysis tasks such as classification with limited labeled data.
Some popular algorithms using Generative Adrial Networks (GANs) for semi-supervised classification share a single architecture for classification and discrimination.
This may require a model to converge to a separate data distribution for each task, which may reduce overall performance.
We propose a novel GAN model namely External GAN (ECGAN) that utilizes GANs and semi-supervised algorithms to improve classification in fully-supervised tasks.
arXiv Detail & Related papers (2020-12-26T05:58:00Z) - Uncertainty-Aware Consistency Regularization for Cross-Domain Semantic
Segmentation [63.75774438196315]
Unsupervised domain adaptation (UDA) aims to adapt existing models of the source domain to a new target domain with only unlabeled data.
Most existing methods suffer from noticeable negative transfer resulting from either the error-prone discriminator network or the unreasonable teacher model.
We propose an uncertainty-aware consistency regularization method for cross-domain semantic segmentation.
arXiv Detail & Related papers (2020-04-19T15:30:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.