Background Noise Reduction of Attention Map for Weakly Supervised Semantic Segmentation
- URL: http://arxiv.org/abs/2404.03394v2
- Date: Tue, 9 Apr 2024 02:56:27 GMT
- Title: Background Noise Reduction of Attention Map for Weakly Supervised Semantic Segmentation
- Authors: Izumi Fujimori, Masaki Oono, Masami Shishibori,
- Abstract summary: This paper focuses on addressing the issue of background noise in attention weights within the existing WSSS method based on Conformer, known as TransCAM.
The proposed method successfully reduces background noise, leading to improved accuracy of pseudo labels.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In weakly-supervised semantic segmentation (WSSS) using only image-level class labels, a problem with CNN-based Class Activation Maps (CAM) is that they tend to activate the most discriminative local regions of objects. On the other hand, methods based on Transformers learn global features but suffer from the issue of background noise contamination. This paper focuses on addressing the issue of background noise in attention weights within the existing WSSS method based on Conformer, known as TransCAM. The proposed method successfully reduces background noise, leading to improved accuracy of pseudo labels. Experimental results demonstrate that our model achieves segmentation performance of 70.5% on the PASCAL VOC 2012 validation data, 71.1% on the test data, and 45.9% on MS COCO 2014 data, outperforming TransCAM in terms of segmentation performance.
Related papers
- Set a Thief to Catch a Thief: Combating Label Noise through Noisy Meta Learning [6.68999525326685]
Learning from noisy labels (LNL) aims to train high-performance deep models using noisy datasets.
We propose a novel noisy meta label correction framework STCT, which counterintuitively uses noisy data to correct label noise.
STCT achieves 96.9% label correction and 95.2% classification performance on CIFAR-10 with 80% symmetric noise.
arXiv Detail & Related papers (2025-02-22T05:58:01Z) - Noisy Ostracods: A Fine-Grained, Imbalanced Real-World Dataset for Benchmarking Robust Machine Learning and Label Correction Methods [7.00297060532893]
The Noisy Ostracods dataset is a noisy dataset for genus and species classification of crustacean ostracods.
The noise is open-set, including new classes discovered during curation that were not part of the original annotation.
The Noisy Ostracods dataset is highly imbalanced with a imbalance factor $rho$ = 22429.
arXiv Detail & Related papers (2024-12-03T09:30:57Z) - SSP-RACL: Classification of Noisy Fundus Images with Self-Supervised Pretraining and Robust Adaptive Credal Loss [3.8739860035485143]
Fundus image classification is crucial in the computer aided diagnosis tasks, but label noise significantly impairs the performance of deep neural networks.
We propose a robust framework, Self-Supervised Pre-training with Robust Adaptive Credal Loss (SSP-RACL), for handling label noise in fundus image datasets.
arXiv Detail & Related papers (2024-09-25T02:41:58Z) - Assessing The Impact of CNN Auto Encoder-Based Image Denoising on Image Classification Tasks [0.0]
This study presents a novel approach for defect detection in casting product noisy images, specifically focusing on submersible pump impellers.
The methodology involves utilizing deep learning models such as VGG16, InceptionV3, and other models in both the spatial and frequency domains.
The study achieved remarkable results using VGG16 for noise type classification in the frequency domain, achieving an accuracy of over 99%.
arXiv Detail & Related papers (2024-04-16T15:40:18Z) - Latent Class-Conditional Noise Model [54.56899309997246]
We introduce a Latent Class-Conditional Noise model (LCCN) to parameterize the noise transition under a Bayesian framework.
We then deduce a dynamic label regression method for LCCN, whose Gibbs sampler allows us efficiently infer the latent true labels.
Our approach safeguards the stable update of the noise transition, which avoids previous arbitrarily tuning from a mini-batch of samples.
arXiv Detail & Related papers (2023-02-19T15:24:37Z) - Learning Confident Classifiers in the Presence of Label Noise [5.829762367794509]
This paper proposes a probabilistic model for noisy observations that allows us to build a confident classification and segmentation models.
Our experiments show that our algorithm outperforms state-of-the-art solutions for the considered classification and segmentation problems.
arXiv Detail & Related papers (2023-01-02T04:27:25Z) - Neighborhood Collective Estimation for Noisy Label Identification and
Correction [92.20697827784426]
Learning with noisy labels (LNL) aims at designing strategies to improve model performance and generalization by mitigating the effects of model overfitting to noisy labels.
Recent advances employ the predicted label distributions of individual samples to perform noise verification and noisy label correction, easily giving rise to confirmation bias.
We propose Neighborhood Collective Estimation, in which the predictive reliability of a candidate sample is re-estimated by contrasting it against its feature-space nearest neighbors.
arXiv Detail & Related papers (2022-08-05T14:47:22Z) - Threshold Matters in WSSS: Manipulating the Activation for the Robust
and Accurate Segmentation Model Against Thresholds [16.6833745997519]
Weakly-supervised semantic segmentation (WSSS) has recently gained much attention for its promise to train segmentation models only with image-level labels.
Existing WSSS methods commonly argue that the sparse coverage of CAM incurs the performance bottleneck of WSSS.
This paper provides analytical and empirical evidence that the actual bottleneck may not be sparse coverage but a global thresholding scheme applied after CAM.
arXiv Detail & Related papers (2022-03-30T04:26:14Z) - Treatment Learning Causal Transformer for Noisy Image Classification [62.639851972495094]
In this work, we incorporate this binary information of "existence of noise" as treatment into image classification tasks to improve prediction accuracy.
Motivated from causal variational inference, we propose a transformer-based architecture, that uses a latent generative model to estimate robust feature representations for noise image classification.
We also create new noisy image datasets incorporating a wide range of noise factors for performance benchmarking.
arXiv Detail & Related papers (2022-03-29T13:07:53Z) - Combating Noise: Semi-supervised Learning by Region Uncertainty
Quantification [55.23467274564417]
Current methods are easily distracted by noisy regions generated by pseudo labels.
We propose noise-resistant semi-supervised learning by quantifying the region uncertainty.
Experiments on both PASCAL VOC and MS COCO demonstrate the extraordinary performance of our method.
arXiv Detail & Related papers (2021-11-01T13:23:42Z) - Training Classifiers that are Universally Robust to All Label Noise
Levels [91.13870793906968]
Deep neural networks are prone to overfitting in the presence of label noise.
We propose a distillation-based framework that incorporates a new subcategory of Positive-Unlabeled learning.
Our framework generally outperforms at medium to high noise levels.
arXiv Detail & Related papers (2021-05-27T13:49:31Z) - A Self-Refinement Strategy for Noise Reduction in Grammatical Error
Correction [54.569707226277735]
Existing approaches for grammatical error correction (GEC) rely on supervised learning with manually created GEC datasets.
There is a non-negligible amount of "noise" where errors were inappropriately edited or left uncorrected.
We propose a self-refinement method where the key idea is to denoise these datasets by leveraging the prediction consistency of existing models.
arXiv Detail & Related papers (2020-10-07T04:45:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.