TabMixer: Excavating Label Distribution Learning with Small-scale
Features
- URL: http://arxiv.org/abs/2210.13852v1
- Date: Tue, 25 Oct 2022 09:18:15 GMT
- Title: TabMixer: Excavating Label Distribution Learning with Small-scale
Features
- Authors: Weiyi Cong, Zhuoran Zheng and Xiuyi Jia
- Abstract summary: Label distribution learning (LDL) differs from multi-label learning which aims at representing the polysemy of instances by transforming single-label values into descriptive degrees.
Unfortunately, the feature space of the label distribution dataset is affected by human factors and the inductive bias of the feature extractor causing uncertainty in the feature space.
We model the uncertainty augmentation of the feature space to alleviate the problem in LDL tasks.
Our proposed algorithm can be competitive compared to other LDL algorithms on several benchmarks.
- Score: 10.498049147922258
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Label distribution learning (LDL) differs from multi-label learning which
aims at representing the polysemy of instances by transforming single-label
values into descriptive degrees. Unfortunately, the feature space of the label
distribution dataset is affected by human factors and the inductive bias of the
feature extractor causing uncertainty in the feature space. Especially, for
datasets with small-scale feature spaces (the feature space dimension $\approx$
the label space), the existing LDL algorithms do not perform well. To address
this issue, we seek to model the uncertainty augmentation of the feature space
to alleviate the problem in LDL tasks. Specifically, we start with augmenting
each feature value in the feature vector of a sample into a vector (sampling on
a Gaussian distribution function). Which, the variance parameter of the
Gaussian distribution function is learned by using a sub-network, and the mean
parameter is filled by this feature value. Then, each feature vector is
augmented to a matrix which is fed into a mixer with local attention
(\textit{TabMixer}) to extract the latent feature. Finally, the latent feature
is squeezed to yield an accurate label distribution via a squeezed network.
Extensive experiments verify that our proposed algorithm can be competitive
compared to other LDL algorithms on several benchmarks.
Related papers
- Inaccurate Label Distribution Learning with Dependency Noise [52.08553913094809]
We introduce the Dependent Noise-based Inaccurate Label Distribution Learning (DN-ILDL) framework to tackle the challenges posed by noise in label distribution learning.
We show that DN-ILDL effectively addresses the ILDL problem and outperforms existing LDL methods.
arXiv Detail & Related papers (2024-05-26T07:58:07Z) - PAC Learning Linear Thresholds from Label Proportions [13.58949814915442]
Learning from label proportions (LLP) is a generalization of supervised learning.
We show that it is possible to efficiently learn LTFs using LTFs when given access to random bags of some label proportion.
We include an experimental evaluation of our learning algorithms along with a comparison with those of [Saket'21, Saket'22] and random LTFs.
arXiv Detail & Related papers (2023-10-16T05:59:34Z) - All Points Matter: Entropy-Regularized Distribution Alignment for
Weakly-supervised 3D Segmentation [67.30502812804271]
Pseudo-labels are widely employed in weakly supervised 3D segmentation tasks where only sparse ground-truth labels are available for learning.
We propose a novel learning strategy to regularize the generated pseudo-labels and effectively narrow the gaps between pseudo-labels and model predictions.
arXiv Detail & Related papers (2023-05-25T08:19:31Z) - Reinforcement Learning in Low-Rank MDPs with Density Features [12.932032416729774]
MDPs with low-rank transitions are highly representative structures that enable tractable learning.
We investigate sample-efficient learning with density features, i.e., the right matrix, which induce powerful models for state-occupancy distributions.
arXiv Detail & Related papers (2023-02-04T22:46:28Z) - How Does Pseudo-Labeling Affect the Generalization Error of the
Semi-Supervised Gibbs Algorithm? [73.80001705134147]
We provide an exact characterization of the expected generalization error (gen-error) for semi-supervised learning (SSL) with pseudo-labeling via the Gibbs algorithm.
The gen-error is expressed in terms of the symmetrized KL information between the output hypothesis, the pseudo-labeled dataset, and the labeled dataset.
arXiv Detail & Related papers (2022-10-15T04:11:56Z) - Closed-Loop Data Transcription to an LDR via Minimaxing Rate Reduction [27.020835928724775]
This work proposes a new computational framework for learning an explicit generative model for real-world datasets.
In particular, we propose to learn em a closed-loop transcription between a multi-class multi-dimensional data distribution and a linear discriminative representation (LDR) in the feature space.
Our experiments on many benchmark imagery datasets demonstrate tremendous potential of this new closed-loop formulation.
arXiv Detail & Related papers (2021-11-12T10:06:08Z) - Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic
Uncertainty [58.144520501201995]
Bi-Lipschitz regularization of neural network layers preserve relative distances between data instances in the feature spaces of each layer.
With the use of an attentive set encoder, we propose to meta learn either diagonal or diagonal plus low-rank factors to efficiently construct task specific covariance matrices.
We also propose an inference procedure which utilizes scaled energy to achieve a final predictive distribution.
arXiv Detail & Related papers (2021-10-12T22:04:19Z) - Multilevel orthogonal Bochner function subspaces with applications to
robust machine learning [1.533771872970755]
We consider the data as instances of a random field within a relevant Bochner space.
Our key observation is that the classes can predominantly reside in two distinct subspaces.
arXiv Detail & Related papers (2021-10-04T22:01:01Z) - Bidirectional Loss Function for Label Enhancement and Distribution
Learning [23.61708127340584]
Two challenges exist in LDL: how to address the dimensional gap problem during the learning process and how to recover label distributions from logical labels.
This study considers bidirectional projections function which can be applied in LE and LDL problems simultaneously.
Experiments on several real-world datasets are carried out to demonstrate the superiority of the proposed method for both LE and LDL.
arXiv Detail & Related papers (2020-07-07T03:02:54Z) - Self-training Avoids Using Spurious Features Under Domain Shift [54.794607791641745]
In unsupervised domain adaptation, conditional entropy minimization and pseudo-labeling work even when the domain shifts are much larger than those analyzed by existing theory.
We identify and analyze one particular setting where the domain shift can be large, but certain spurious features correlate with label in the source domain but are independent label in the target.
arXiv Detail & Related papers (2020-06-17T17:51:42Z) - OSLNet: Deep Small-Sample Classification with an Orthogonal Softmax
Layer [77.90012156266324]
This paper aims to find a subspace of neural networks that can facilitate a large decision margin.
We propose the Orthogonal Softmax Layer (OSL), which makes the weight vectors in the classification layer remain during both the training and test processes.
Experimental results demonstrate that the proposed OSL has better performance than the methods used for comparison on four small-sample benchmark datasets.
arXiv Detail & Related papers (2020-04-20T02:41:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.