Clean Label Disentangling for Medical Image Segmentation with Noisy
Labels
- URL: http://arxiv.org/abs/2311.16580v1
- Date: Tue, 28 Nov 2023 07:54:27 GMT
- Title: Clean Label Disentangling for Medical Image Segmentation with Noisy
Labels
- Authors: Zicheng Wang, Zhen Zhao, Erjian Guo and Luping Zhou
- Abstract summary: Current methods focusing on medical image segmentation suffer from incorrect annotations, which is known as the noisy label issue.
We propose a class-balanced sampling strategy to tackle the class-imbalanced problem.
We extend our clean label disentangling framework to a new noisy feature-aided clean label disentangling framework.
- Score: 25.180056839942345
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current methods focusing on medical image segmentation suffer from incorrect
annotations, which is known as the noisy label issue. Most medical image
segmentation with noisy labels methods utilize either noise transition matrix,
noise-robust loss functions or pseudo-labeling methods, while none of the
current research focuses on clean label disentanglement. We argue that the main
reason is that the severe class-imbalanced issue will lead to the inaccuracy of
the selected ``clean'' labels, thus influencing the robustness of the model
against the noises. In this work, we come up with a simple but efficient
class-balanced sampling strategy to tackle the class-imbalanced problem, which
enables our newly proposed clean label disentangling framework to successfully
select clean labels from the given label sets and encourages the model to learn
from the correct annotations. However, such a method will filter out too many
annotations which may also contain useful information. Therefore, we further
extend our clean label disentangling framework to a new noisy feature-aided
clean label disentangling framework, which takes the full annotations into
utilization to learn more semantics. Extensive experiments have validated the
effectiveness of our methods, where our methods achieve new state-of-the-art
performance. Our code is available at https://github.com/xiaoyao3302/2BDenoise.
Related papers
- Deep Self-Cleansing for Medical Image Segmentation with Noisy Labels [33.676420623855314]
Medical image segmentation is crucial in the field of medical imaging, aiding in disease diagnosis and surgical planning.
Most established segmentation methods rely on supervised deep learning, in which clean and precise labels are essential for supervision.
We propose a deep self-cleansing segmentation framework that can preserve clean labels while cleansing noisy ones in the training phase.
arXiv Detail & Related papers (2024-09-08T08:33:32Z) - Extracting Clean and Balanced Subset for Noisy Long-tailed Classification [66.47809135771698]
We develop a novel pseudo labeling method using class prototypes from the perspective of distribution matching.
By setting a manually-specific probability measure, we can reduce the side-effects of noisy and long-tailed data simultaneously.
Our method can extract this class-balanced subset with clean labels, which brings effective performance gains for long-tailed classification with label noise.
arXiv Detail & Related papers (2024-04-10T07:34:37Z) - Learning to Segment from Noisy Annotations: A Spatial Correction
Approach [12.604673584405385]
Noisy labels can significantly affect the performance of deep neural networks (DNNs)
We propose a novel Markov model for segmentation noisy annotations that encodes both spatial correlation and bias.
Our approach outperforms current state-of-the-art methods on both synthetic and real-world noisy annotations.
arXiv Detail & Related papers (2023-07-21T00:27:40Z) - BadLabel: A Robust Perspective on Evaluating and Enhancing Label-noise
Learning [113.8799653759137]
We introduce a novel label noise type called BadLabel, which can significantly degrade the performance of existing LNL algorithms by a large margin.
BadLabel is crafted based on the label-flipping attack against standard classification.
We propose a robust LNL method that perturbs the labels in an adversarial manner at each epoch to make the loss values of clean and noisy labels again distinguishable.
arXiv Detail & Related papers (2023-05-28T06:26:23Z) - Category-Adaptive Label Discovery and Noise Rejection for Multi-label
Image Recognition with Partial Positive Labels [78.88007892742438]
Training multi-label models with partial positive labels (MLR-PPL) attracts increasing attention.
Previous works regard unknown labels as negative and adopt traditional MLR algorithms.
We propose to explore semantic correlation among different images to facilitate the MLR-PPL task.
arXiv Detail & Related papers (2022-11-15T02:11:20Z) - Bootstrapping the Relationship Between Images and Their Clean and Noisy
Labels [14.16958860081955]
We propose a new training algorithm that relies on a simple model to learn the relationship between clean and noisy labels.
By learning this relationship, we achieve state-of-the-art performance in asymmetric and instance-dependent label noise problems.
arXiv Detail & Related papers (2022-10-17T08:15:55Z) - Large Loss Matters in Weakly Supervised Multi-Label Classification [50.262533546999045]
We first regard unobserved labels as negative labels, casting the W task into noisy multi-label classification.
We propose novel methods for W which reject or correct the large loss samples to prevent model from memorizing the noisy label.
Our methodology actually works well, validating that treating large loss properly matters in a weakly supervised multi-label classification.
arXiv Detail & Related papers (2022-06-08T08:30:24Z) - Label Noise in Adversarial Training: A Novel Perspective to Study Robust
Overfitting [45.58217741522973]
We show that label noise exists in adversarial training.
Such label noise is due to the mismatch between the true label distribution of adversarial examples and the label inherited from clean examples.
We propose a method to automatically calibrate the label to address the label noise and robust overfitting.
arXiv Detail & Related papers (2021-10-07T01:15:06Z) - Learning to Aggregate and Refine Noisy Labels for Visual Sentiment
Analysis [69.48582264712854]
We propose a robust learning method to perform robust visual sentiment analysis.
Our method relies on an external memory to aggregate and filter noisy labels during training.
We establish a benchmark for visual sentiment analysis with label noise using publicly available datasets.
arXiv Detail & Related papers (2021-09-15T18:18:28Z) - Does label smoothing mitigate label noise? [57.76529645344897]
We show that label smoothing is competitive with loss-correction under label noise.
We show that when distilling models from noisy data, label smoothing of the teacher is beneficial.
arXiv Detail & Related papers (2020-03-05T18:43:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.