Adaptive label thresholding methods for online multi-label
classification
- URL: http://arxiv.org/abs/2112.02301v1
- Date: Sat, 4 Dec 2021 10:34:09 GMT
- Title: Adaptive label thresholding methods for online multi-label
classification
- Authors: Tingting Zhai, Hongcheng Tang, Hao Wang
- Abstract summary: Existing online multi-label classification works cannot handle the online label thresholding problem.
This paper proposes a novel framework of adaptive label thresholding algorithms for online multi-label classification.
- Score: 4.028101568570768
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing online multi-label classification works cannot well handle the
online label thresholding problem and lack the regret analysis for their online
algorithms. This paper proposes a novel framework of adaptive label
thresholding algorithms for online multi-label classification, with the aim to
overcome the drawbacks of existing methods. The key feature of our framework is
that both scoring and thresholding models are included as important components
of the online multi-label classifier and are incorporated into one online
optimization problem. Further, in order to establish the relationship between
scoring and thresholding models, a novel multi-label classification loss
function is derived, which measures to what an extent the multi-label
classifier can distinguish between relevant labels and irrelevant ones for an
incoming instance. Based on this new framework and loss function, we present a
first-order linear algorithm and a second-order one, which both enjoy closed
form update, but rely on different techniques for updating the multi-label
classifier. Both algorithms are proved to achieve a sub-linear regret. Using
Mercer kernels, our first-order algorithm has been extended to deal with
nonlinear multi-label prediction tasks. Experiments show the advantage of our
linear and nonlinear algorithms, in terms of various multi-label performance
metrics.
Related papers
- Dual-Decoupling Learning and Metric-Adaptive Thresholding for Semi-Supervised Multi-Label Learning [81.83013974171364]
Semi-supervised multi-label learning (SSMLL) is a powerful framework for leveraging unlabeled data to reduce the expensive cost of collecting precise multi-label annotations.
Unlike semi-supervised learning, one cannot select the most probable label as the pseudo-label in SSMLL due to multiple semantics contained in an instance.
We propose a dual-perspective method to generate high-quality pseudo-labels.
arXiv Detail & Related papers (2024-07-26T09:33:53Z) - Multi-Label Noise Transition Matrix Estimation with Label Correlations:
Theory and Algorithm [73.94839250910977]
Noisy multi-label learning has garnered increasing attention due to the challenges posed by collecting large-scale accurate labels.
The introduction of transition matrices can help model multi-label noise and enable the development of statistically consistent algorithms.
We propose a novel estimator that leverages label correlations without the need for anchor points or precise fitting of noisy class posteriors.
arXiv Detail & Related papers (2023-09-22T08:35:38Z) - Class-Distribution-Aware Pseudo Labeling for Semi-Supervised Multi-Label
Learning [97.88458953075205]
Pseudo-labeling has emerged as a popular and effective approach for utilizing unlabeled data.
This paper proposes a novel solution called Class-Aware Pseudo-Labeling (CAP) that performs pseudo-labeling in a class-aware manner.
arXiv Detail & Related papers (2023-05-04T12:52:18Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - Evolving Multi-Label Fuzzy Classifier [5.53329677986653]
Multi-label classification has attracted much attention in the machine learning community to address the problem of assigning single samples to more than one class at the same time.
We propose an evolving multi-label fuzzy classifier (EFC-ML) which is able to self-adapt and self-evolve its structure with new incoming multi-label samples in an incremental, single-pass manner.
arXiv Detail & Related papers (2022-03-29T08:01:03Z) - Gradient-based Label Binning in Multi-label Classification [0.0]
In multi-label classification, the ability to model dependencies between labels is crucial to effectively optimize non-decomposable evaluation measures.
The utilization of second-order derivatives, as used by many recent boosting approaches, helps to guide the minimization of non-decomposable losses.
In this work, we address the computational bottleneck of such approach by integrating a novel approximation technique into the boosting procedure.
arXiv Detail & Related papers (2021-06-22T11:48:48Z) - Dual-Refinement: Joint Label and Feature Refinement for Unsupervised
Domain Adaptive Person Re-Identification [51.98150752331922]
Unsupervised domain adaptive (UDA) person re-identification (re-ID) is a challenging task due to the missing of labels for the target domain data.
We propose a novel approach, called Dual-Refinement, that jointly refines pseudo labels at the off-line clustering phase and features at the on-line training phase.
Our method outperforms the state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2020-12-26T07:35:35Z) - Learning Gradient Boosted Multi-label Classification Rules [4.842945656927122]
We propose an algorithm for learning multi-label classification rules that is able to minimize decomposable as well as non-decomposable loss functions.
We analyze the abilities and limitations of our approach on synthetic data and evaluate its predictive performance on multi-label benchmarks.
arXiv Detail & Related papers (2020-06-23T21:39:23Z) - Online Metric Learning for Multi-Label Classification [22.484707213499714]
We propose a novel online metric learning paradigm for multi-label classification.
We first propose a new metric for multi-label classification based on $k$-Nearest Neighbour ($k$NN)
arXiv Detail & Related papers (2020-06-12T11:33:04Z) - Progressive Identification of True Labels for Partial-Label Learning [112.94467491335611]
Partial-label learning (PLL) is a typical weakly supervised learning problem, where each training instance is equipped with a set of candidate labels among which only one is the true label.
Most existing methods elaborately designed as constrained optimizations that must be solved in specific manners, making their computational complexity a bottleneck for scaling up to big data.
This paper proposes a novel framework of classifier with flexibility on the model and optimization algorithm.
arXiv Detail & Related papers (2020-02-19T08:35:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.