Unimodal Distributions for Ordinal Regression
- URL: http://arxiv.org/abs/2303.04547v1
- Date: Wed, 8 Mar 2023 13:00:40 GMT
- Title: Unimodal Distributions for Ordinal Regression
- Authors: Jaime S. Cardoso and Ricardo Cruz and Tom\'e Albuquerque
- Abstract summary: We propose two new approaches to incorporate the preference for unimodal distributions into the predictive model.
We analyse the set of unimodal distributions in the probability simplex and establish fundamental properties.
We then propose a new architecture that imposes unimodal distributions and a new loss term that relies on the notion of projection in a set to promote unimodality.
- Score: 2.642698101441705
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many real-world prediction tasks, class labels contain information about
the relative order between labels that are not captured by commonly used loss
functions such as multicategory cross-entropy. Recently, the preference for
unimodal distributions in the output space has been incorporated into models
and loss functions to account for such ordering information. However, current
approaches rely on heuristics that lack a theoretical foundation. Here, we
propose two new approaches to incorporate the preference for unimodal
distributions into the predictive model. We analyse the set of unimodal
distributions in the probability simplex and establish fundamental properties.
We then propose a new architecture that imposes unimodal distributions and a
new loss term that relies on the notion of projection in a set to promote
unimodality. Experiments show the new architecture achieves top-2 performance,
while the proposed new loss term is very competitive while maintaining high
unimodality.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - SimPro: A Simple Probabilistic Framework Towards Realistic Long-Tailed Semi-Supervised Learning [49.94607673097326]
We propose a highly adaptable framework, designated as SimPro, which does not rely on any predefined assumptions about the distribution of unlabeled data.
Our framework, grounded in a probabilistic model, innovatively refines the expectation-maximization algorithm.
Our method showcases consistent state-of-the-art performance across diverse benchmarks and data distribution scenarios.
arXiv Detail & Related papers (2024-02-21T03:39:04Z) - Dr. FERMI: A Stochastic Distributionally Robust Fair Empirical Risk
Minimization Framework [12.734559823650887]
In the presence of distribution shifts, fair machine learning models may behave unfairly on test data.
Existing algorithms require full access to data and cannot be used when small batches are used.
This paper proposes the first distributionally robust fairness framework with convergence guarantees that do not require knowledge of the causal graph.
arXiv Detail & Related papers (2023-09-20T23:25:28Z) - Cooperative Distribution Alignment via JSD Upper Bound [7.071749623370137]
Unsupervised distribution alignment estimates a transformation that maps two or more source distributions to a shared aligned distribution.
This task has many applications including generative modeling, unsupervised domain adaptation, and socially aware learning.
We propose to unify and generalize previous flow-based approaches under a single non-adversarial framework.
arXiv Detail & Related papers (2022-07-05T20:09:03Z) - Learning to Transfer with von Neumann Conditional Divergence [14.926485055255942]
We introduce the recently proposed von Neumann conditional divergence to improve the transferability across multiple domains.
We design novel learning objectives assuming those source tasks are observed either simultaneously or sequentially.
In both scenarios, we obtain favorable performance against state-of-the-art methods in terms of smaller generalization error on new tasks and less catastrophic forgetting on source tasks (in the sequential setup)
arXiv Detail & Related papers (2021-08-07T22:18:23Z) - Mode recovery in neural autoregressive sequence modeling [55.05526174291747]
Recent studies have revealed unexpected and undesirable properties of neural autoregressive sequence models.
We investigate how the modes, or local maxima, of a distribution are maintained throughout the full learning chain.
We conclude that future research must consider the entire learning chain in order to fully understand the potentials and perils.
arXiv Detail & Related papers (2021-06-10T02:17:28Z) - An Online Learning Approach to Interpolation and Extrapolation in Domain
Generalization [53.592597682854944]
We recast generalization over sub-groups as an online game between a player minimizing risk and an adversary presenting new test.
We show that ERM is provably minimax-optimal for both tasks.
arXiv Detail & Related papers (2021-02-25T19:06:48Z) - Deep Ordinal Regression using Optimal Transport Loss and Unimodal Output
Probabilities [3.169089186688223]
We propose a framework for deep ordinal regression based on unimodal output distribution and optimal transport loss.
We empirically analyze the different components of our proposed approach and demonstrate their contribution to the performance of the model.
Experimental results on eight real-world datasets demonstrate that our proposed approach consistently performs on par with and often better than several recently proposed deep learning approaches.
arXiv Detail & Related papers (2020-11-15T19:19:40Z) - Generative Model without Prior Distribution Matching [26.91643368299913]
Variational Autoencoder (VAE) and its variations are classic generative models by learning a low-dimensional latent representation to satisfy some prior distribution.
We propose to let the prior match the embedding distribution rather than imposing the latent variables to fit the prior.
arXiv Detail & Related papers (2020-09-23T09:33:24Z) - Distributional Reinforcement Learning via Moment Matching [54.16108052278444]
We formulate a method that learns a finite set of statistics from each return distribution via neural networks.
Our method can be interpreted as implicitly matching all orders of moments between a return distribution and its Bellman target.
Experiments on the suite of Atari games show that our method outperforms the standard distributional RL baselines.
arXiv Detail & Related papers (2020-07-24T05:18:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.