Deep Categorization with Semi-Supervised Self-Organizing Maps
- URL: http://arxiv.org/abs/2006.13682v1
- Date: Wed, 17 Jun 2020 22:00:04 GMT
- Title: Deep Categorization with Semi-Supervised Self-Organizing Maps
- Authors: Pedro H. M. Braga, Heitor R. Medeiros and Hansenclever F. Bassani
- Abstract summary: This article presents a semi-supervised model, called Batch Semi-Supervised Self-Organizing Map (Batch SS-SOM)
The results show that Batch SS-SOM is a good option for semi-supervised classification and clustering.
It performs well in terms of accuracy and clustering error, even with a small number of labeled samples.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Nowadays, with the advance of technology, there is an increasing amount of
unstructured data being generated every day. However, it is a painful job to
label and organize it. Labeling is an expensive, time-consuming, and difficult
task. It is usually done manually, which collaborates with the incorporation of
noise and errors to the data. Hence, it is of great importance to developing
intelligent models that can benefit from both labeled and unlabeled data.
Currently, works on unsupervised and semi-supervised learning are still being
overshadowed by the successes of purely supervised learning. However, it is
expected that they become far more important in the longer term. This article
presents a semi-supervised model, called Batch Semi-Supervised Self-Organizing
Map (Batch SS-SOM), which is an extension of a SOM incorporating some advances
that came with the rise of Deep Learning, such as batch training. The results
show that Batch SS-SOM is a good option for semi-supervised classification and
clustering. It performs well in terms of accuracy and clustering error, even
with a small number of labeled samples, as well as when presented to
unsupervised data, and shows competitive results in transfer learning scenarios
in traditional image classification benchmark datasets.
Related papers
- Continuous Contrastive Learning for Long-Tailed Semi-Supervised Recognition [50.61991746981703]
Current state-of-the-art LTSSL approaches rely on high-quality pseudo-labels for large-scale unlabeled data.
This paper introduces a novel probabilistic framework that unifies various recent proposals in long-tail learning.
We introduce a continuous contrastive learning method, CCL, extending our framework to unlabeled data using reliable and smoothed pseudo-labels.
arXiv Detail & Related papers (2024-10-08T15:06:10Z) - Semi-Supervised One-Shot Imitation Learning [83.94646047695412]
One-shot Imitation Learning aims to imbue AI agents with the ability to learn a new task from a single demonstration.
We introduce the semi-supervised OSIL problem setting, where the learning agent is presented with a large dataset of trajectories.
We develop an algorithm specifically applicable to this semi-supervised OSIL setting.
arXiv Detail & Related papers (2024-08-09T18:11:26Z) - Incremental Self-training for Semi-supervised Learning [56.57057576885672]
IST is simple yet effective and fits existing self-training-based semi-supervised learning methods.
We verify the proposed IST on five datasets and two types of backbone, effectively improving the recognition accuracy and learning speed.
arXiv Detail & Related papers (2024-04-14T05:02:00Z) - Semi-unsupervised Learning for Time Series Classification [1.8811803364757567]
Time series are ubiquitous and inherently hard to analyze and ultimately to label or cluster.
We present SuSL4TS, a deep generative Gaussian mixture model for semi-unsupervised learning to classify time series data.
arXiv Detail & Related papers (2022-07-07T06:59:38Z) - Training image classifiers using Semi-Weak Label Data [26.04162590798731]
In Multiple Instance learning (MIL), weak labels are provided at the bag level with only presence/absence information known.
This paper introduces a novel semi-weak label learning paradigm as a middle ground to mitigate the problem.
We propose a two-stage framework to address the problem of learning from semi-weak labels.
arXiv Detail & Related papers (2021-03-19T03:06:07Z) - Boosting the Performance of Semi-Supervised Learning with Unsupervised
Clustering [10.033658645311188]
We show that ignoring labels altogether for whole epochs intermittently during training can significantly improve performance in the small sample regime.
We demonstrate our method's efficacy in boosting several state-of-the-art SSL algorithms.
arXiv Detail & Related papers (2020-12-01T14:19:14Z) - GPU-based Self-Organizing Maps for Post-Labeled Few-Shot Unsupervised
Learning [2.922007656878633]
Few-shot classification is a challenge in machine learning where the goal is to train a classifier using a very limited number of labeled examples.
We consider the problem of post-labeled few-shot unsupervised learning, a classification task where representations are learned in an unsupervised fashion, to be later labeled using very few annotated examples.
arXiv Detail & Related papers (2020-09-04T13:22:28Z) - Learning to Count in the Crowd from Limited Labeled Data [109.2954525909007]
We focus on reducing the annotation efforts by learning to count in the crowd from limited number of labeled samples.
Specifically, we propose a Gaussian Process-based iterative learning mechanism that involves estimation of pseudo-ground truth for the unlabeled data.
arXiv Detail & Related papers (2020-07-07T04:17:01Z) - Temporal Calibrated Regularization for Robust Noisy Label Learning [60.90967240168525]
Deep neural networks (DNNs) exhibit great success on many tasks with the help of large-scale well annotated datasets.
However, labeling large-scale data can be very costly and error-prone so that it is difficult to guarantee the annotation quality.
We propose a Temporal Calibrated Regularization (TCR) in which we utilize the original labels and the predictions in the previous epoch together.
arXiv Detail & Related papers (2020-07-01T04:48:49Z) - Improving Semantic Segmentation via Self-Training [75.07114899941095]
We show that we can obtain state-of-the-art results using a semi-supervised approach, specifically a self-training paradigm.
We first train a teacher model on labeled data, and then generate pseudo labels on a large set of unlabeled data.
Our robust training framework can digest human-annotated and pseudo labels jointly and achieve top performances on Cityscapes, CamVid and KITTI datasets.
arXiv Detail & Related papers (2020-04-30T17:09:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.