Multi-label Stream Classification with Self-Organizing Maps
- URL: http://arxiv.org/abs/2004.09397v1
- Date: Mon, 20 Apr 2020 15:52:38 GMT
- Title: Multi-label Stream Classification with Self-Organizing Maps
- Authors: Ricardo Cerri, Joel David Costa J\'unior, Elaine Ribeiro de Faria
Paiva and Jo\~ao Manuel Portela da Gama
- Abstract summary: We propose an online incremental method based on self-organizing maps for multi-label stream classification with infinitely delayed labels.
In the classification phase, we use a k-nearest neighbors strategy to compute the winning neurons in the maps.
We predict labels for each instance using the Bayes rule and the outputs of each neuron, adapting the probabilities and conditional probabilities of the classes in the stream.
- Score: 2.055054374525828
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Several learning algorithms have been proposed for offline multi-label
classification. However, applications in areas such as traffic monitoring,
social networks, and sensors produce data continuously, the so called data
streams, posing challenges to batch multi-label learning. With the lack of
stationarity in the distribution of data streams, new algorithms are needed to
online adapt to such changes (concept drift). Also, in realistic applications,
changes occur in scenarios of infinitely delayed labels, where the true classes
of the arrival instances are never available. We propose an online unsupervised
incremental method based on self-organizing maps for multi-label stream
classification with infinitely delayed labels. In the classification phase, we
use a k-nearest neighbors strategy to compute the winning neurons in the maps,
adapting to concept drift by online adjusting neuron weight vectors and dataset
label cardinality. We predict labels for each instance using the Bayes rule and
the outputs of each neuron, adapting the probabilities and conditional
probabilities of the classes in the stream. Experiments using synthetic and
real datasets show that our method is highly competitive with several ones from
the literature, in both stationary and concept drift scenarios.
Related papers
- Hoeffding adaptive trees for multi-label classification on data streams [4.3998370660429655]
Multi-Label Hoeffding Adaptive Tree (MLHAT)
This paper proposes a novel approach to multi-label data stream classification called Multi-Label Hoeffding Adaptive Tree (MLHAT)
arXiv Detail & Related papers (2024-10-26T18:10:22Z) - Continuous Contrastive Learning for Long-Tailed Semi-Supervised Recognition [50.61991746981703]
Current state-of-the-art LTSSL approaches rely on high-quality pseudo-labels for large-scale unlabeled data.
This paper introduces a novel probabilistic framework that unifies various recent proposals in long-tail learning.
We introduce a continuous contrastive learning method, CCL, extending our framework to unlabeled data using reliable and smoothed pseudo-labels.
arXiv Detail & Related papers (2024-10-08T15:06:10Z) - Online Label Shift: Optimal Dynamic Regret meets Practical Algorithms [33.61487362513345]
This paper focuses on supervised and unsupervised online label shift, where the class marginals $Q(y)$ varies but the class-conditionals $Q(x|y)$ remain invariant.
In the unsupervised setting, our goal is to adapt a learner, trained on some offline labeled data, to changing label distributions given unlabeled online data.
We develop novel algorithms that reduce the adaptation problem to online regression and guarantee optimal dynamic regret without any prior knowledge of the extent of drift in the label distribution.
arXiv Detail & Related papers (2023-05-31T05:39:52Z) - All Points Matter: Entropy-Regularized Distribution Alignment for
Weakly-supervised 3D Segmentation [67.30502812804271]
Pseudo-labels are widely employed in weakly supervised 3D segmentation tasks where only sparse ground-truth labels are available for learning.
We propose a novel learning strategy to regularize the generated pseudo-labels and effectively narrow the gaps between pseudo-labels and model predictions.
arXiv Detail & Related papers (2023-05-25T08:19:31Z) - Class-Distribution-Aware Pseudo Labeling for Semi-Supervised Multi-Label
Learning [97.88458953075205]
Pseudo-labeling has emerged as a popular and effective approach for utilizing unlabeled data.
This paper proposes a novel solution called Class-Aware Pseudo-Labeling (CAP) that performs pseudo-labeling in a class-aware manner.
arXiv Detail & Related papers (2023-05-04T12:52:18Z) - Learning with Neighbor Consistency for Noisy Labels [69.83857578836769]
We present a method for learning from noisy labels that leverages similarities between training examples in feature space.
We evaluate our method on datasets evaluating both synthetic (CIFAR-10, CIFAR-100) and realistic (mini-WebVision, Clothing1M, mini-ImageNet-Red) noise.
arXiv Detail & Related papers (2022-02-04T15:46:27Z) - Adaptive Label Smoothing To Regularize Large-Scale Graph Training [46.00927775402987]
We propose the adaptive label smoothing (ALS) method to replace the one-hot hard labels with smoothed ones.
ALS propagates node labels to aggregate the neighborhood label distribution in a pre-processing step, and then updates the optimal smoothed labels online to adapt to specific graph structure.
arXiv Detail & Related papers (2021-08-30T23:51:31Z) - PLM: Partial Label Masking for Imbalanced Multi-label Classification [59.68444804243782]
Neural networks trained on real-world datasets with long-tailed label distributions are biased towards frequent classes and perform poorly on infrequent classes.
We propose a method, Partial Label Masking (PLM), which utilizes this ratio during training.
Our method achieves strong performance when compared to existing methods on both multi-label (MultiMNIST and MSCOCO) and single-label (imbalanced CIFAR-10 and CIFAR-100) image classification datasets.
arXiv Detail & Related papers (2021-05-22T18:07:56Z) - PAL : Pretext-based Active Learning [2.869739951301252]
We propose an active learning technique for deep neural networks that is more robust to mislabeling than the previously proposed techniques.
We use a separate network to score the unlabeled samples for selection.
The resultant technique also produces competitive accuracy in the absence of label noise.
arXiv Detail & Related papers (2020-10-29T21:16:37Z) - Semi-supervised sequence classification through change point detection [22.14406516031776]
We propose a novel framework for semi-supervised learning in such contexts.
In an unsupervised manner, change point detection methods can be used to identify points within a sequence corresponding to likely class changes.
We show that change points provide examples of similar/dissimilar pairs of sequences which, when coupled with labeled, can be used in a semi-supervised classification setting.
arXiv Detail & Related papers (2020-09-24T17:23:13Z) - Knowledge-Guided Multi-Label Few-Shot Learning for General Image
Recognition [75.44233392355711]
KGGR framework exploits prior knowledge of statistical label correlations with deep neural networks.
It first builds a structured knowledge graph to correlate different labels based on statistical label co-occurrence.
Then, it introduces the label semantics to guide learning semantic-specific features.
It exploits a graph propagation network to explore graph node interactions.
arXiv Detail & Related papers (2020-09-20T15:05:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.