Semantic Segmentation with Active Semi-Supervised Representation
Learning
- URL: http://arxiv.org/abs/2210.08403v1
- Date: Sun, 16 Oct 2022 00:21:43 GMT
- Title: Semantic Segmentation with Active Semi-Supervised Representation
Learning
- Authors: Aneesh Rangnekar, Christopher Kanan, Matthew Hoffman
- Abstract summary: We train an effective semantic segmentation algorithm with significantly lesser labeled data.
We extend the prior state-of-the-art S4AL algorithm by replacing its mean teacher approach for semi-supervised learning with a self-training approach.
We evaluate our method on CamVid and CityScapes datasets, the de-facto standards for active learning for semantic segmentation.
- Score: 23.79742108127707
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Obtaining human per-pixel labels for semantic segmentation is incredibly
laborious, often making labeled dataset construction prohibitively expensive.
Here, we endeavor to overcome this problem with a novel algorithm that combines
semi-supervised and active learning, resulting in the ability to train an
effective semantic segmentation algorithm with significantly lesser labeled
data. To do this, we extend the prior state-of-the-art S4AL algorithm by
replacing its mean teacher approach for semi-supervised learning with a
self-training approach that improves learning with noisy labels. We further
boost the neural network's ability to query useful data by adding a contrastive
learning head, which leads to better understanding of the objects in the scene,
and hence, better queries for active learning. We evaluate our method on CamVid
and CityScapes datasets, the de-facto standards for active learning for
semantic segmentation. We achieve more than 95% of the network's performance on
CamVid and CityScapes datasets, utilizing only 12.1% and 15.1% of the labeled
data, respectively. We also benchmark our method across existing stand-alone
semi-supervised learning methods on the CityScapes dataset and achieve superior
performance without any bells or whistles.
Related papers
- Enhancing Visual Continual Learning with Language-Guided Supervision [76.38481740848434]
Continual learning aims to empower models to learn new tasks without forgetting previously acquired knowledge.
We argue that the scarce semantic information conveyed by the one-hot labels hampers the effective knowledge transfer across tasks.
Specifically, we use PLMs to generate semantic targets for each class, which are frozen and serve as supervision signals.
arXiv Detail & Related papers (2024-03-24T12:41:58Z) - Learning from the Best: Active Learning for Wireless Communications [9.523381807291049]
Active learning algorithms identify the most critical and informative samples in an unlabeled dataset and label only those samples, instead of the complete set.
We present a case study of deep learning-based mmWave beam selection, where labeling is performed by a compute-intensive algorithm based on exhaustive search.
Our results show that using an active learning algorithm for class-imbalanced datasets can reduce labeling overhead by up to 50% for this dataset.
arXiv Detail & Related papers (2024-01-23T12:21:57Z) - Hierarchical Point-based Active Learning for Semi-supervised Point Cloud
Semantic Segmentation [48.40853126077237]
It is labour-intensive to acquire large-scale point cloud data with point-wise labels.
Active learning is one of the effective strategies to achieve this purpose but is still under-explored.
This paper develops a hierarchical point-based active learning strategy.
arXiv Detail & Related papers (2023-08-22T03:52:05Z) - TrueDeep: A systematic approach of crack detection with less data [0.0]
We show that by incorporating domain knowledge along with deep learning architectures, we can achieve similar performance with less data.
Our algorithms, developed with 23% of the overall data, have a similar performance on the test data and significantly better performance on multiple blind datasets.
arXiv Detail & Related papers (2023-05-30T14:51:58Z) - LESS: Label-Efficient Semantic Segmentation for LiDAR Point Clouds [62.49198183539889]
We propose a label-efficient semantic segmentation pipeline for outdoor scenes with LiDAR point clouds.
Our method co-designs an efficient labeling process with semi/weakly supervised learning.
Our proposed method is even highly competitive compared to the fully supervised counterpart with 100% labels.
arXiv Detail & Related papers (2022-10-14T19:13:36Z) - Semantic Segmentation with Active Semi-Supervised Learning [23.79742108127707]
We propose a novel algorithm that combines active learning and semi-supervised learning.
Our method obtains over 95% of the network's performance on the full-training set.
arXiv Detail & Related papers (2022-03-21T04:16:25Z) - Learning with Neighbor Consistency for Noisy Labels [69.83857578836769]
We present a method for learning from noisy labels that leverages similarities between training examples in feature space.
We evaluate our method on datasets evaluating both synthetic (CIFAR-10, CIFAR-100) and realistic (mini-WebVision, Clothing1M, mini-ImageNet-Red) noise.
arXiv Detail & Related papers (2022-02-04T15:46:27Z) - Dominant Set-based Active Learning for Text Classification and its
Application to Online Social Media [0.0]
We present a novel pool-based active learning method for the training of large unlabeled corpus with minimum annotation cost.
Our proposed method does not have any parameters to be tuned, making it dataset-independent.
Our method achieves a higher performance in comparison to the state-of-the-art active learning strategies.
arXiv Detail & Related papers (2022-01-28T19:19:03Z) - Improving Semi-Supervised and Domain-Adaptive Semantic Segmentation with
Self-Supervised Depth Estimation [94.16816278191477]
We present a framework for semi-adaptive and domain-supervised semantic segmentation.
It is enhanced by self-supervised monocular depth estimation trained only on unlabeled image sequences.
We validate the proposed model on the Cityscapes dataset.
arXiv Detail & Related papers (2021-08-28T01:33:38Z) - Dense Contrastive Learning for Self-Supervised Visual Pre-Training [102.15325936477362]
We present dense contrastive learning, which implements self-supervised learning by optimizing a pairwise contrastive (dis)similarity loss at the pixel level between two views of input images.
Compared to the baseline method MoCo-v2, our method introduces negligible computation overhead (only 1% slower)
arXiv Detail & Related papers (2020-11-18T08:42:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.