ILSGAN: Independent Layer Synthesis for Unsupervised
Foreground-Background Segmentation
- URL: http://arxiv.org/abs/2211.13974v4
- Date: Sat, 7 Oct 2023 09:54:57 GMT
- Title: ILSGAN: Independent Layer Synthesis for Unsupervised
Foreground-Background Segmentation
- Authors: Qiran Zou, Yu Yang, Wing Yin Cheung, Chang Liu, Xiangyang Ji
- Abstract summary: Unsupervised foreground-background segmentation aims at extracting salient objects from cluttered backgrounds.
We propose a simple-yet-effective explicit layer independence modeling approach, termed Independent Layer Synthesis GAN (ILSGAN)
Our ILSGAN achieves strong state-of-the-art generation quality and segmentation performance on complex real-world data.
- Score: 49.61394755739333
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised foreground-background segmentation aims at extracting salient
objects from cluttered backgrounds, where Generative Adversarial Network (GAN)
approaches, especially layered GANs, show great promise. However, without human
annotations, they are typically prone to produce foreground and background
layers with non-negligible semantic and visual confusion, dubbed "information
leakage", resulting in notable degeneration of the generated segmentation mask.
To alleviate this issue, we propose a simple-yet-effective explicit layer
independence modeling approach, termed Independent Layer Synthesis GAN
(ILSGAN), pursuing independent foreground-background layer generation by
encouraging their discrepancy. Specifically, it targets minimizing the mutual
information between visible and invisible regions of the foreground and
background to spur interlayer independence. Through in-depth theoretical and
experimental analyses, we justify that explicit layer independence modeling is
critical to suppressing information leakage and contributes to impressive
segmentation performance gains. Also, our ILSGAN achieves strong
state-of-the-art generation quality and segmentation performance on complex
real-world data. Code is available at: https://github.com/qrzou/ILSGAN
Related papers
- EraseLoRA: MLLM-Driven Foreground Exclusion and Background Subtype Aggregation for Dataset-Free Object Removal [10.015328934927062]
We propose EraseLoRA, a dataset-free framework that replaces attention surgery with background-aware reasoning and test-time adaptation.<n>First, Background-aware Foreground Exclusion (BFE), uses a multimodal large-language models to separate target foreground, non-target foregrounds, and clean background from a single image-mask pair without paired supervision.<n>Second, Background-aware Reconstruction with Subtype Aggregation (BRSA), performs test-time optimization that treats inferred background subtypes as complementary pieces.
arXiv Detail & Related papers (2025-12-25T07:34:38Z) - STAGE: Segmentation-oriented Industrial Anomaly Synthesis via Graded Diffusion with Explicit Mask Alignment [29.572404530614175]
Industrial Anomaly Synthesis (SIAS) plays a pivotal role in enhancing the performance of downstream anomaly segmentation.<n>SIAS methods face several critical limitations.<n>We propose Graded diffusion with Explicit mask alignment, termed STAGE.
arXiv Detail & Related papers (2025-09-08T13:47:01Z) - Segment Concealed Objects with Incomplete Supervision [63.637733655439334]
Incompletely-Supervised Concealed Object (ISCOS) involves segmenting objects that seamlessly blend into their surrounding environments.<n>This task remains highly challenging due to the limited supervision provided by the incompletely annotated training data.<n>In this paper, we introduce the first unified method for ISCOS to address these challenges.
arXiv Detail & Related papers (2025-06-10T16:25:15Z) - Semantic-guided Adversarial Diffusion Model for Self-supervised Shadow Removal [5.083330121710086]
GAN-based training often faces issues such as mode collapse and unstable optimization.
We propose a semantic-guided adversarial diffusion framework for self-supervised shadow removal.
We conduct experiments on multiple public datasets, and the experimental results demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2024-07-01T09:14:38Z) - UGMAE: A Unified Framework for Graph Masked Autoencoders [67.75493040186859]
We propose UGMAE, a unified framework for graph masked autoencoders.
We first develop an adaptive feature mask generator to account for the unique significance of nodes.
We then design a ranking-based structure reconstruction objective joint with feature reconstruction to capture holistic graph information.
arXiv Detail & Related papers (2024-02-12T19:39:26Z) - Unsupervised Camouflaged Object Segmentation as Domain Adaptation [5.304455190825436]
In this work, we investigate a new task, namely unsupervised camouflaged object segmentation (UCOS), where the target objects own a common attribute, i.e., camouflage.
We find that the state-of-the-art unsupervised models struggle in adapting UCOS, due to the domain gap between the properties of generic and camouflaged objects.
We formulate the UCOS as a source-free unsupervised domain adaptation task (UCOS-DA), where both source labels and target labels are absent during the whole model training process.
arXiv Detail & Related papers (2023-08-08T18:46:16Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Progressively Dual Prior Guided Few-shot Semantic Segmentation [57.37506990980975]
Few-shot semantic segmentation task aims at performing segmentation in query images with a few annotated support samples.
We propose a progressively dual prior guided few-shot semantic segmentation network.
arXiv Detail & Related papers (2022-11-20T16:19:47Z) - Deep Spectral Methods: A Surprisingly Strong Baseline for Unsupervised
Semantic Segmentation and Localization [98.46318529630109]
We take inspiration from traditional spectral segmentation methods by reframing image decomposition as a graph partitioning problem.
We find that these eigenvectors already decompose an image into meaningful segments, and can be readily used to localize objects in a scene.
By clustering the features associated with these segments across a dataset, we can obtain well-delineated, nameable regions.
arXiv Detail & Related papers (2022-05-16T17:47:44Z) - Unsupervised Foreground-Background Segmentation with Equivariant Layered
GANs [48.3919579579129]
We propose an unsupervised foreground-background segmentation method via training a segmentation network on the synthetic pseudo segmentation dataset generated from GANs.
To efficiently generate foreground and background layers and overlay them to compose novel images, the construction of such GANs is fulfilled by our proposed Equivariant Layered GAN.
Our methods are evaluated on unsupervised object segmentation datasets including Caltech-UCSD Birds and LSUN Car, achieving state-of-the-art performance.
arXiv Detail & Related papers (2021-04-01T14:13:25Z) - Labels4Free: Unsupervised Segmentation using StyleGAN [40.39780497423365]
We propose an unsupervised segmentation framework for StyleGAN generated objects.
We report comparable results against state-of-the-art supervised segmentation networks.
arXiv Detail & Related papers (2021-03-27T18:59:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.