Constraint-aware Learning of Probabilistic Sequential Models for Multi-Label Classification
- URL: http://arxiv.org/abs/2507.15156v1
- Date: Sun, 20 Jul 2025 23:31:36 GMT
- Title: Constraint-aware Learning of Probabilistic Sequential Models for Multi-Label Classification
- Authors: Mykhailo Buleshnyi, Anna Polova, Zsolt Zombori, Michael Benedikt,
- Abstract summary: We investigate multi-label classification involving large sets of labels, where the output labels may be known to satisfy some logical constraints.<n>We look at an architecture in which classifiers for individual labels are fed into an expressive sequential model, which produces a joint distribution.
- Score: 0.5624791703748108
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We investigate multi-label classification involving large sets of labels, where the output labels may be known to satisfy some logical constraints. We look at an architecture in which classifiers for individual labels are fed into an expressive sequential model, which produces a joint distribution. One of the potential advantages for such an expressive model is its ability to modelling correlations, as can arise from constraints. We empirically demonstrate the ability of the architecture both to exploit constraints in training and to enforce constraints at inference time.
Related papers
- Correlative and Discriminative Label Grouping for Multi-Label Visual Prompt Tuning [12.052388861361937]
Recent studies have overemphasized co-occurrence relationships among labels, leading to suboptimal models.<n>We propose the Multi-Label Visual Prompt Tuning framework to balance correlative and discriminative relationships among labels.<n>Our proposed approach achieves competitive results and outperforms SOTA methods on multiple pre-trained models.
arXiv Detail & Related papers (2025-04-14T08:52:50Z) - Self-Supervised Open-Ended Classification with Small Visual Language
Models [60.23212389067007]
We present Self-Context Adaptation (SeCAt), a self-supervised approach that unlocks few-shot abilities for open-ended classification with small visual language models.
By using models with approximately 1B parameters we outperform the few-shot abilities of much larger models, such as Frozen and FROMAGe.
arXiv Detail & Related papers (2023-09-30T21:41:21Z) - ChiroDiff: Modelling chirographic data with Diffusion Models [132.5223191478268]
We introduce a powerful model-class namely "Denoising Diffusion Probabilistic Models" or DDPMs for chirographic data.
Our model named "ChiroDiff", being non-autoregressive, learns to capture holistic concepts and therefore remains resilient to higher temporal sampling rate.
arXiv Detail & Related papers (2023-04-07T15:17:48Z) - Leveraging Instance Features for Label Aggregation in Programmatic Weak
Supervision [75.1860418333995]
Programmatic Weak Supervision (PWS) has emerged as a widespread paradigm to synthesize training labels efficiently.
The core component of PWS is the label model, which infers true labels by aggregating the outputs of multiple noisy supervision sources as labeling functions.
Existing statistical label models typically rely only on the outputs of LF, ignoring the instance features when modeling the underlying generative process.
arXiv Detail & Related papers (2022-10-06T07:28:53Z) - From Multi-label Learning to Cross-Domain Transfer: A Model-Agnostic
Approach [1.5076964620370268]
We develop an approach for transfer learning that challenges the long-held assumption that transferability of tasks comes from measurements of similarity between the source and target domains or models.
We show that essentially we can create task-dependence based on source-model capacity.
arXiv Detail & Related papers (2022-07-24T13:37:25Z) - Use All The Labels: A Hierarchical Multi-Label Contrastive Learning
Framework [75.79736930414715]
We present a hierarchical multi-label representation learning framework that can leverage all available labels and preserve the hierarchical relationship between classes.
We introduce novel hierarchy preserving losses, which jointly apply a hierarchical penalty to the contrastive loss, and enforce the hierarchy constraint.
arXiv Detail & Related papers (2022-04-27T21:41:44Z) - Resolving label uncertainty with implicit posterior models [71.62113762278963]
We propose a method for jointly inferring labels across a collection of data samples.
By implicitly assuming the existence of a generative model for which a differentiable predictor is the posterior, we derive a training objective that allows learning under weak beliefs.
arXiv Detail & Related papers (2022-02-28T18:09:44Z) - Learning Debiased and Disentangled Representations for Semantic
Segmentation [52.35766945827972]
We propose a model-agnostic and training scheme for semantic segmentation.
By randomly eliminating certain class information in each training iteration, we effectively reduce feature dependencies among classes.
Models trained with our approach demonstrate strong results on multiple semantic segmentation benchmarks.
arXiv Detail & Related papers (2021-10-31T16:15:09Z) - Dependency Structure Misspecification in Multi-Source Weak Supervision
Models [15.125993628007972]
We study the effects of label model misspecification on test set performance of a downstream classifier.
We derive novel theoretical bounds on the modeling error and empirically show that this error can be substantial.
arXiv Detail & Related papers (2021-06-18T18:15:44Z) - Evolving Multi-label Classification Rules by Exploiting High-order Label
Correlation [2.9822184411723645]
In multi-label classification tasks, each problem instance is associated with multiple classes simultaneously.
The correlation between labels can be exploited at different levels such as capturing the pair-wise correlation or exploiting the higher-order correlations.
This paper aims at exploiting the high-order label correlation within subsets of labels using a supervised learning classifier system.
arXiv Detail & Related papers (2020-07-22T18:13:12Z) - Multi-Label Text Classification using Attention-based Graph Neural
Network [0.0]
A graph attention network-based model is proposed to capture the attentive dependency structure among the labels.
The proposed model achieves similar or better performance compared to the previous state-of-the-art models.
arXiv Detail & Related papers (2020-03-22T17:12:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.