Uncertainty quantification for multiclass data description
- URL: http://arxiv.org/abs/2108.12857v1
- Date: Sun, 29 Aug 2021 14:42:04 GMT
- Title: Uncertainty quantification for multiclass data description
- Authors: Leila Kalantari, Jose Principe and Kathryn E. Sieving
- Abstract summary: We propose a multiclass data description model based on kernel Mahalanobis distance (MDD-KM)
We report a prototypical classification system based on a hierarchical linear dynamical system with MDD-KM as a component.
- Score: 0.1611401281366893
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this manuscript, we propose a multiclass data description model based on
kernel Mahalanobis distance (MDD-KM) with self-adapting hyperparameter setting.
MDD-KM provides uncertainty quantification and can be deployed to build
classification systems for the realistic scenario where out-of-distribution
(OOD) samples are present among the test data. Given a test signal, a quantity
related to empirical kernel Mahalanobis distance between the signal and each of
the training classes is computed. Since these quantities correspond to the same
reproducing kernel Hilbert space, they are commensurable and hence can be
readily treated as classification scores without further application of fusion
techniques. To set kernel parameters, we exploit the fact that predictive
variance according to a Gaussian process (GP) is empirical kernel Mahalanobis
distance when a centralized kernel is used, and propose to use GP's negative
likelihood function as the cost function. We conduct experiments on the real
problem of avian note classification. We report a prototypical classification
system based on a hierarchical linear dynamical system with MDD-KM as a
component. Our classification system does not require sound event detection as
a preprocessing step, and is able to find instances of training avian notes
with varying length among OOD samples (corresponding to unknown notes of
disinterest) in the test audio clip. Domain knowledge is leveraged to make
crisp decisions from raw classification scores. We demonstrate the superior
performance of MDD-KM over possibilistic K-nearest neighbor.
Related papers
- Reproducible Machine Learning-based Voice Pathology Detection: Introducing the Pitch Difference Feature [1.1455937444848385]
We propose a robust set of features derived from a thorough research of contemporary practices in voice pathology detection.
We combine this feature set, containing data from the publicly available Saarbr"ucken Voice Database (SVD), with preprocessing using the K-Means Synthetic Minority Over-Sampling Technique algorithm.
Our approach has achieved the state-of-the-art performance, measured by unweighted average recall in voice pathology detection.
arXiv Detail & Related papers (2024-10-14T14:17:52Z) - Intra-class Adaptive Augmentation with Neighbor Correction for Deep
Metric Learning [99.14132861655223]
We propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning.
We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining.
Our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
arXiv Detail & Related papers (2022-11-29T14:52:38Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - Bayesian Evidential Learning for Few-Shot Classification [22.46281648187903]
Few-Shot Classification aims to generalize from base classes to novel classes given very limited labeled samples.
State-of-the-art solutions involve learning to find a good metric and representation space to compute the distance between samples.
Despite the promising accuracy performance, how to model uncertainty for metric-based FSC methods effectively is still a challenge.
arXiv Detail & Related papers (2022-07-19T03:58:00Z) - Self-Adaptive Label Augmentation for Semi-supervised Few-shot
Classification [121.63992191386502]
Few-shot classification aims to learn a model that can generalize well to new tasks when only a few labeled samples are available.
We propose a semi-supervised few-shot classification method that assigns an appropriate label to each unlabeled sample by a manually defined metric.
A major novelty of SALA is the task-adaptive metric, which can learn the metric adaptively for different tasks in an end-to-end fashion.
arXiv Detail & Related papers (2022-06-16T13:14:03Z) - Multi-Class Data Description for Out-of-distribution Detection [25.853322158250435]
Deep-MCDD is effective to detect out-of-distribution (OOD) samples as well as classify in-distribution (ID) samples.
By integrating the concept of Gaussian discriminant analysis into deep neural networks, we propose a deep learning objective to learn class-conditional distributions.
arXiv Detail & Related papers (2021-04-02T08:41:51Z) - Federated Deep AUC Maximization for Heterogeneous Data with a Constant
Communication Complexity [77.78624443410216]
We propose improved FDAM algorithms for detecting heterogeneous chest data.
A result of this paper is that the communication of the proposed algorithm is strongly independent of the number of machines and also independent of the accuracy level.
Experiments have demonstrated the effectiveness of our FDAM algorithm on benchmark datasets and on medical chest Xray images from different organizations.
arXiv Detail & Related papers (2021-02-09T04:05:19Z) - A Unified Joint Maximum Mean Discrepancy for Domain Adaptation [73.44809425486767]
This paper theoretically derives a unified form of JMMD that is easy to optimize.
From the revealed unified JMMD, we illustrate that JMMD degrades the feature-label dependence that benefits to classification.
We propose a novel MMD matrix to promote the dependence, and devise a novel label kernel that is robust to label distribution shift.
arXiv Detail & Related papers (2021-01-25T09:46:14Z) - Learning with Out-of-Distribution Data for Audio Classification [60.48251022280506]
We show that detecting and relabelling certain OOD instances, rather than discarding them, can have a positive effect on learning.
The proposed method is shown to improve the performance of convolutional neural networks by a significant margin.
arXiv Detail & Related papers (2020-02-11T21:08:06Z) - GIM: Gaussian Isolation Machines [40.7916016364212]
In many cases, neural network classifiers are exposed to input data that is outside of their training distribution data.
We present a novel hybrid (generative-discriminative) classifier aimed at solving the problem arising when OOD data is encountered.
The proposed GIM's novelty lies in its discriminative performance and generative capabilities, a combination of characteristics not usually seen in a single classifier.
arXiv Detail & Related papers (2020-02-06T09:51:47Z) - Multi-class Gaussian Process Classification with Noisy Inputs [2.362412515574206]
In some situations, the amount of noise can be known before-hand.
We have evaluated the proposed methods by carrying out several experiments, involving synthetic and real data.
arXiv Detail & Related papers (2020-01-28T18:55:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.