Universally Rank Consistent Ordinal Regression in Neural Networks
- URL: http://arxiv.org/abs/2110.07470v1
- Date: Thu, 14 Oct 2021 15:44:08 GMT
- Title: Universally Rank Consistent Ordinal Regression in Neural Networks
- Authors: Garrett Jenkinson, Kia Khezeli, Gavin R. Oliver, John Kalantari, Eric
W. Klee
- Abstract summary: Recent methods have resorted to converting ordinal regression into a series of extended binary classification subtasks.
Here we demonstrate that the subtask probabilities form a Markov chain.
We show how to straightforwardly modify neural network architectures to exploit this fact and thereby constrain predictions to be universally rank consistent.
- Score: 4.462334751640166
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Despite the pervasiveness of ordinal labels in supervised learning, it
remains common practice in deep learning to treat such problems as categorical
classification using the categorical cross entropy loss. Recent methods
attempting to address this issue while respecting the ordinal structure of the
labels have resorted to converting ordinal regression into a series of extended
binary classification subtasks. However, the adoption of such methods remains
inconsistent due to theoretical and practical limitations. Here we address
these limitations by demonstrating that the subtask probabilities form a Markov
chain. We show how to straightforwardly modify neural network architectures to
exploit this fact and thereby constrain predictions to be universally rank
consistent. We furthermore prove that all rank consistent solutions can be
represented within this formulation. Using diverse benchmarks and the
real-world application of a specialized recurrent neural network for COVID-19
prognosis, we demonstrate the practical superiority of this method versus the
current state-of-the-art. The method is open sourced as user-friendly PyTorch
and TensorFlow packages.
Related papers
- Robust Stochastically-Descending Unrolled Networks [85.6993263983062]
Deep unrolling is an emerging learning-to-optimize method that unrolls a truncated iterative algorithm in the layers of a trainable neural network.
We show that convergence guarantees and generalizability of the unrolled networks are still open theoretical problems.
We numerically assess unrolled architectures trained under the proposed constraints in two different applications.
arXiv Detail & Related papers (2023-12-25T18:51:23Z) - The Cascaded Forward Algorithm for Neural Network Training [61.06444586991505]
We propose a new learning framework for neural networks, namely Cascaded Forward (CaFo) algorithm, which does not rely on BP optimization as that in FF.
Unlike FF, our framework directly outputs label distributions at each cascaded block, which does not require generation of additional negative samples.
In our framework each block can be trained independently, so it can be easily deployed into parallel acceleration systems.
arXiv Detail & Related papers (2023-03-17T02:01:11Z) - Semantic Strengthening of Neuro-Symbolic Learning [85.6195120593625]
Neuro-symbolic approaches typically resort to fuzzy approximations of a probabilistic objective.
We show how to compute this efficiently for tractable circuits.
We test our approach on three tasks: predicting a minimum-cost path in Warcraft, predicting a minimum-cost perfect matching, and solving Sudoku puzzles.
arXiv Detail & Related papers (2023-02-28T00:04:22Z) - Equivariance with Learned Canonicalization Functions [77.32483958400282]
We show that learning a small neural network to perform canonicalization is better than using predefineds.
Our experiments show that learning the canonicalization function is competitive with existing techniques for learning equivariant functions across many tasks.
arXiv Detail & Related papers (2022-11-11T21:58:15Z) - Regularization-based Pruning of Irrelevant Weights in Deep Neural
Architectures [0.0]
We propose a method for learning sparse neural topologies via a regularization technique which identifies non relevant weights and selectively shrinks their norm.
We tested the proposed technique on different image classification and Natural language generation tasks, obtaining results on par or better then competitors in terms of sparsity and metrics.
arXiv Detail & Related papers (2022-04-11T09:44:16Z) - Deep Neural Networks for Rank-Consistent Ordinal Regression Based On
Conditional Probabilities [3.093890460224435]
Ordinal regression methods for deep neural networks address ordinal response variables.
CORAL method achieves rank consistency among its output layer tasks by imposing a weight-sharing constraint.
We propose a new method for rank-consistent ordinal regression without this limitation.
arXiv Detail & Related papers (2021-11-17T01:10:23Z) - Defensive Tensorization [113.96183766922393]
We propose tensor defensiveization, an adversarial defence technique that leverages a latent high-order factorization of the network.
We empirically demonstrate the effectiveness of our approach on standard image classification benchmarks.
We validate the versatility of our approach across domains and low-precision architectures by considering an audio task and binary networks.
arXiv Detail & Related papers (2021-10-26T17:00:16Z) - Asymptotics of Network Embeddings Learned via Subsampling [4.23373349945751]
We study representation methods using a subsampling approach, such as node2vec, into a single unifying framework.
This provides a theoretical foundation to understand what the embedding vectors represent and how well these methods perform on downstream tasks.
Notably, we observe that typically used loss functions may lead to shortcomings, such as a lack of Fisher consistency.
arXiv Detail & Related papers (2021-07-06T02:54:53Z) - Theoretical Insights Into Multiclass Classification: A High-dimensional
Asymptotic View [82.80085730891126]
We provide the first modernally precise analysis of linear multiclass classification.
Our analysis reveals that the classification accuracy is highly distribution-dependent.
The insights gained may pave the way for a precise understanding of other classification algorithms.
arXiv Detail & Related papers (2020-11-16T05:17:29Z) - Compressive Sensing and Neural Networks from a Statistical Learning
Perspective [4.561032960211816]
We present a generalization error analysis for a class of neural networks suitable for sparse reconstruction from few linear measurements.
Under realistic conditions, the generalization error scales only logarithmically in the number of layers, and at most linear in number of measurements.
arXiv Detail & Related papers (2020-10-29T15:05:43Z) - Deep Ordinal Regression with Label Diversity [19.89482062012177]
We propose that using several discrete data representations simultaneously can improve neural network learning.
Our approach is end-to-end differentiable and can be added as a simple extension to conventional learning methods.
arXiv Detail & Related papers (2020-06-29T08:23:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.