Label Encoding for Regression Networks
- URL: http://arxiv.org/abs/2212.01927v1
- Date: Sun, 4 Dec 2022 21:23:36 GMT
- Title: Label Encoding for Regression Networks
- Authors: Deval Shah, Zi Yu Xue, Tor M. Aamodt
- Abstract summary: We introduce binary-encoded labels (BEL), which generalizes the application of binary classification to regression.
BEL achieves state-of-the-art accuracies for several regression benchmarks.
- Score: 9.386028796990399
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks are used for a wide range of regression problems.
However, there exists a significant gap in accuracy between specialized
approaches and generic direct regression in which a network is trained by
minimizing the squared or absolute error of output labels. Prior work has shown
that solving a regression problem with a set of binary classifiers can improve
accuracy by utilizing well-studied binary classification algorithms. We
introduce binary-encoded labels (BEL), which generalizes the application of
binary classification to regression by providing a framework for considering
arbitrary multi-bit values when encoding target values. We identify desirable
properties of suitable encoding and decoding functions used for the conversion
between real-valued and binary-encoded labels based on theoretical and
empirical study. These properties highlight a tradeoff between classification
error probability and error-correction capabilities of label encodings. BEL can
be combined with off-the-shelf task-specific feature extractors and trained
end-to-end. We propose a series of sample encoding, decoding, and training loss
functions for BEL and demonstrate they result in lower error than direct
regression and specialized approaches while being suitable for a diverse set of
regression problems, network architectures, and evaluation metrics. BEL
achieves state-of-the-art accuracies for several regression benchmarks. Code is
available at https://github.com/ubc-aamodt-group/BEL_regression.
Related papers
- Robust Capped lp-Norm Support Vector Ordinal Regression [85.84718111830752]
Ordinal regression is a specialized supervised problem where the labels show an inherent order.
Support Vector Ordinal Regression, as an outstanding ordinal regression model, is widely used in many ordinal regression tasks.
We introduce a new model, Capped $ell_p$-Norm Support Vector Ordinal Regression(CSVOR), that is robust to outliers.
arXiv Detail & Related papers (2024-04-25T13:56:05Z) - Rethinking Classifier Re-Training in Long-Tailed Recognition: A Simple
Logits Retargeting Approach [102.0769560460338]
We develop a simple logits approach (LORT) without the requirement of prior knowledge of the number of samples per class.
Our method achieves state-of-the-art performance on various imbalanced datasets, including CIFAR100-LT, ImageNet-LT, and iNaturalist 2018.
arXiv Detail & Related papers (2024-03-01T03:27:08Z) - An Ordinal Regression Framework for a Deep Learning Based Severity
Assessment for Chest Radiographs [50.285682227571996]
We propose a framework that divides the ordinal regression problem into three parts: a model, a target function, and a classification function.
We show that the choice of encoding has a strong impact on performance and that the best encoding depends on the chosen weighting of Cohen's kappa.
arXiv Detail & Related papers (2024-02-08T14:00:45Z) - Regularized Linear Regression for Binary Classification [20.710343135282116]
Regularized linear regression is a promising approach for binary classification problems in which the training set has noisy labels.
We show that for large enough regularization strength, the optimal weights concentrate around two values of opposite sign.
We observe that in many cases the corresponding "compression" of each weight to a single bit leads to very little loss in performance.
arXiv Detail & Related papers (2023-11-03T23:18:21Z) - Generating Unbiased Pseudo-labels via a Theoretically Guaranteed
Chebyshev Constraint to Unify Semi-supervised Classification and Regression [57.17120203327993]
threshold-to-pseudo label process (T2L) in classification uses confidence to determine the quality of label.
In nature, regression also requires unbiased methods to generate high-quality labels.
We propose a theoretically guaranteed constraint for generating unbiased labels based on Chebyshev's inequality.
arXiv Detail & Related papers (2023-11-03T08:39:35Z) - Deep Imbalanced Regression via Hierarchical Classification Adjustment [50.19438850112964]
Regression tasks in computer vision are often formulated into classification by quantizing the target space into classes.
The majority of training samples lie in a head range of target values, while a minority of samples span a usually larger tail range.
We propose to construct hierarchical classifiers for solving imbalanced regression tasks.
Our novel hierarchical classification adjustment (HCA) for imbalanced regression shows superior results on three diverse tasks.
arXiv Detail & Related papers (2023-10-26T04:54:39Z) - Learning Label Encodings for Deep Regression [10.02230163797581]
Deep regression networks are widely used to tackle the problem of predicting a continuous value for a given input.
The space of label encodings for regression is large.
This paper introduces Regularized Label Learning (RLEL) for end-to-end training of an entire network and its label encoding.
arXiv Detail & Related papers (2023-03-04T00:11:34Z) - Pseudo-Labeling for Kernel Ridge Regression under Covariate Shift [1.3597551064547502]
We learn a regression function with small mean squared error over a target distribution, based on unlabeled data from there and labeled data that may have a different feature distribution.
We propose to split the labeled data into two subsets, and conduct kernel ridge regression on them separately to obtain a collection of candidate models and an imputation model.
Our estimator achieves the minimax optimal error rate up to a polylogarithmic factor, and we find that using pseudo-labels for model selection does not significantly hinder performance.
arXiv Detail & Related papers (2023-02-20T18:46:12Z) - How Does Pseudo-Labeling Affect the Generalization Error of the
Semi-Supervised Gibbs Algorithm? [73.80001705134147]
We provide an exact characterization of the expected generalization error (gen-error) for semi-supervised learning (SSL) with pseudo-labeling via the Gibbs algorithm.
The gen-error is expressed in terms of the symmetrized KL information between the output hypothesis, the pseudo-labeled dataset, and the labeled dataset.
arXiv Detail & Related papers (2022-10-15T04:11:56Z) - Robust Neural Network Classification via Double Regularization [2.41710192205034]
We propose a novel double regularization of the neural network training loss that combines a penalty on the complexity of the classification model and an optimal reweighting of training observations.
We demonstrate DRFit, for neural net classification of (i) MNIST and (ii) CIFAR-10, in both cases with simulated mislabeling.
arXiv Detail & Related papers (2021-12-15T13:19:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.