Kernel Density Estimation for Multiclass Quantification
- URL: http://arxiv.org/abs/2401.00490v2
- Date: Tue, 2 Jan 2024 19:52:24 GMT
- Title: Kernel Density Estimation for Multiclass Quantification
- Authors: Alejandro Moreo, Pablo Gonz\'alez, Juan Jos\'e del Coz
- Abstract summary: Quantification is the supervised machine learning task concerned with obtaining accurate predictors of class prevalence.
The distribution-matching (DM) approaches represent one of the most important families among the quantification methods that have been proposed in the literature so far.
We propose a new representation mechanism based on multivariate densities that we model via kernel density estimation (KDE)
- Score: 52.419589623702336
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Several disciplines, like the social sciences, epidemiology, sentiment
analysis, or market research, are interested in knowing the distribution of the
classes in a population rather than the individual labels of the members
thereof. Quantification is the supervised machine learning task concerned with
obtaining accurate predictors of class prevalence, and to do so particularly in
the presence of label shift. The distribution-matching (DM) approaches
represent one of the most important families among the quantification methods
that have been proposed in the literature so far. Current DM approaches model
the involved populations by means of histograms of posterior probabilities. In
this paper, we argue that their application to the multiclass setting is
suboptimal since the histograms become class-specific, thus missing the
opportunity to model inter-class information that may exist in the data. We
propose a new representation mechanism based on multivariate densities that we
model via kernel density estimation (KDE). The experiments we have carried out
show our method, dubbed KDEy, yields superior quantification performance with
respect to previous DM approaches. We also investigate the KDE-based
representation within the maximum likelihood framework and show KDEy often
shows superior performance with respect to the expectation-maximization method
for quantification, arguably the strongest contender in the quantification
arena to date.
Related papers
- Enhancing Knowledge Distillation of Large Language Models through Efficient Multi-Modal Distribution Alignment [10.104085497265004]
We propose Ranking Loss based Knowledge Distillation (RLKD), which encourages consistency of peak predictions between the teacher and student models.
Our method enables the student model to better learn the multi-modal distributions of the teacher model, leading to a significant performance improvement in various downstream tasks.
arXiv Detail & Related papers (2024-09-19T08:06:42Z) - Seeing Unseen: Discover Novel Biomedical Concepts via
Geometry-Constrained Probabilistic Modeling [53.7117640028211]
We present a geometry-constrained probabilistic modeling treatment to resolve the identified issues.
We incorporate a suite of critical geometric properties to impose proper constraints on the layout of constructed embedding space.
A spectral graph-theoretic method is devised to estimate the number of potential novel classes.
arXiv Detail & Related papers (2024-03-02T00:56:05Z) - MISS: Multiclass Interpretable Scoring Systems [13.902264070785986]
We present a machine-learning approach for constructing Multiclass Interpretable Scoring Systems (MISS)
MISS is a fully data-driven methodology for single, sparse, and user-friendly scoring systems for multiclass classification problems.
Results indicate that our approach is competitive with other machine learning models in terms of classification performance metrics and provides well-calibrated class probabilities.
arXiv Detail & Related papers (2024-01-10T10:57:12Z) - Learning Invariant Molecular Representation in Latent Discrete Space [52.13724532622099]
We propose a new framework for learning molecular representations that exhibit invariance and robustness against distribution shifts.
Our model achieves stronger generalization against state-of-the-art baselines in the presence of various distribution shifts.
arXiv Detail & Related papers (2023-10-22T04:06:44Z) - MAUVE Scores for Generative Models: Theory and Practice [95.86006777961182]
We present MAUVE, a family of comparison measures between pairs of distributions such as those encountered in the generative modeling of text or images.
We find that MAUVE can quantify the gaps between the distributions of human-written text and those of modern neural language models.
We demonstrate in the vision domain that MAUVE can identify known properties of generated images on par with or better than existing metrics.
arXiv Detail & Related papers (2022-12-30T07:37:40Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - On Reinforcement Learning and Distribution Matching for Fine-Tuning
Language Models with no Catastrophic Forgetting [5.5302127686575435]
Two main paradigms have emerged to tackle this challenge: Reward Maximization (RM) and, more recently, Distribution Matching (DM)
We show that methods such as KL-control developed for RM can also be construed as belonging to DM.
We leverage connections between the two paradigms to import the concept of baseline into DM methods.
arXiv Detail & Related papers (2022-06-01T20:54:41Z) - A Top-down Supervised Learning Approach to Hierarchical Multi-label
Classification in Networks [0.21485350418225244]
This paper presents a general prediction model to hierarchical multi-label classification (HMC), where the attributes to be inferred can be specified as a strict poset.
It is based on a top-down classification approach that addresses hierarchical multi-label classification with supervised learning by building a local classifier per class.
The proposed model is showcased with a case study on the prediction of gene functions for Oryza sativa Japonica, a variety of rice.
arXiv Detail & Related papers (2022-03-23T17:29:17Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - Who Explains the Explanation? Quantitatively Assessing Feature
Attribution Methods [0.0]
We propose a novel evaluation metric -- the Focus -- designed to quantify the faithfulness of explanations.
We show the robustness of the metric through randomization experiments, and then use Focus to evaluate and compare three popular explainability techniques.
Our results find LRP and GradCAM to be consistent and reliable, while the latter remains most competitive even when applied to poorly performing models.
arXiv Detail & Related papers (2021-09-28T07:10:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.