Active Inference-Based Optimization of Discriminative Neural Network
Classifiers
- URL: http://arxiv.org/abs/2306.02447v1
- Date: Sun, 4 Jun 2023 19:30:28 GMT
- Title: Active Inference-Based Optimization of Discriminative Neural Network
Classifiers
- Authors: Faezeh Fallah
- Abstract summary: We propose a novel algorithm to find candidate classification labels of the training samples from their prior probabilities.
The proposed objective function could incorporate the candidate labels, the original reference labels, and the priors of the training samples while still being distribution-based.
- Score: 3.1219977244201056
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Commonly used objective functions (losses) for a supervised optimization of
discriminative neural network classifiers were either distribution-based or
metric-based. The distribution-based losses could compromise the generalization
or cause classification biases towards the dominant classes of an imbalanced
class-sample distribution. The metric-based losses could make the network model
independent of any distribution and thus improve its generalization. However,
they could still be biased towards the dominant classes and could suffer from
discrepancies when a class was absent in both the reference (ground truth) and
the predicted labels. In this paper, we proposed a novel optimization process
which not only tackled the unbalancedness of the class-sample distribution of
the training samples but also provided a mechanism to tackle errors in the
reference labels of the training samples. This was achieved by proposing a
novel algorithm to find candidate classification labels of the training samples
from their prior probabilities and the currently estimated posteriors on the
network and a novel objective function for the optimizations. The algorithm was
the result of casting the generalized Kelly criterion for optimal betting into
a multiclass classification problem. The proposed objective function was the
expected free energy of a prospective active inference and could incorporate
the candidate labels, the original reference labels, and the priors of the
training samples while still being distribution-based. The incorporation of the
priors into the optimization not only helped to tackle errors in the reference
labels but also allowed to reduce classification biases towards the dominant
classes by focusing the attention of the neural network on important but
minority foreground classes.
Related papers
- Bi-discriminator Domain Adversarial Neural Networks with Class-Level
Gradient Alignment [87.8301166955305]
We propose a novel bi-discriminator domain adversarial neural network with class-level gradient alignment.
BACG resorts to gradient signals and second-order probability estimation for better alignment of domain distributions.
In addition, inspired by contrastive learning, we develop a memory bank-based variant, i.e. Fast-BACG, which can greatly shorten the training process.
arXiv Detail & Related papers (2023-10-21T09:53:17Z) - Robust Class-Conditional Distribution Alignment for Partial Domain
Adaptation [0.7892577704654171]
Unwanted samples from private source categories in the learning objective of a partial domain adaptation setup can lead to negative transfer and reduce classification performance.
Existing methods, such as re-weighting or aggregating target predictions, are vulnerable to this issue.
Our proposed approach seeks to overcome these limitations by delving deeper than just the first-order moments to derive distinct and compact categorical distributions.
arXiv Detail & Related papers (2023-10-18T15:49:46Z) - Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - Proposal Distribution Calibration for Few-Shot Object Detection [65.19808035019031]
In few-shot object detection (FSOD), the two-step training paradigm is widely adopted to mitigate the severe sample imbalance.
Unfortunately, the extreme data scarcity aggravates the proposal distribution bias, hindering the RoI head from evolving toward novel classes.
We introduce a simple yet effective proposal distribution calibration (PDC) approach to neatly enhance the localization and classification abilities of the RoI head.
arXiv Detail & Related papers (2022-12-15T05:09:11Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - Adaptive Distribution Calibration for Few-Shot Learning with
Hierarchical Optimal Transport [78.9167477093745]
We propose a novel distribution calibration method by learning the adaptive weight matrix between novel samples and base classes.
Experimental results on standard benchmarks demonstrate that our proposed plug-and-play model outperforms competing approaches.
arXiv Detail & Related papers (2022-10-09T02:32:57Z) - Learning to Adapt Classifier for Imbalanced Semi-supervised Learning [38.434729550279116]
Pseudo-labeling has proven to be a promising semi-supervised learning (SSL) paradigm.
Existing pseudo-labeling methods commonly assume that the class distributions of training data are balanced.
In this work, we investigate pseudo-labeling under imbalanced semi-supervised setups.
arXiv Detail & Related papers (2022-07-28T02:15:47Z) - Prototypical Classifier for Robust Class-Imbalanced Learning [64.96088324684683]
We propose textitPrototypical, which does not require fitting additional parameters given the embedding network.
Prototypical produces balanced and comparable predictions for all classes even though the training set is class-imbalanced.
We test our method on CIFAR-10LT, CIFAR-100LT and Webvision datasets, observing that Prototypical obtains substaintial improvements compared with state of the arts.
arXiv Detail & Related papers (2021-10-22T01:55:01Z) - Instance-based Label Smoothing For Better Calibrated Classification
Networks [3.388509725285237]
Label smoothing is widely used in deep neural networks for multi-class classification.
We take inspiration from both label smoothing and self-distillation.
We propose two novel instance-based label smoothing approaches.
arXiv Detail & Related papers (2021-10-11T15:33:23Z) - Hyperdimensional Computing for Efficient Distributed Classification with
Randomized Neural Networks [5.942847925681103]
We study distributed classification, which can be employed in situations were data cannot be stored at a central location nor shared.
We propose a more efficient solution for distributed classification by making use of a lossy compression approach applied when sharing the local classifiers with other agents.
arXiv Detail & Related papers (2021-06-02T01:33:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.