Multi-class Classification with Fuzzy-feature Observations: Theory and
Algorithms
- URL: http://arxiv.org/abs/2206.04311v1
- Date: Thu, 9 Jun 2022 07:14:00 GMT
- Title: Multi-class Classification with Fuzzy-feature Observations: Theory and
Algorithms
- Authors: Guangzhi Ma and Jie Lu and Feng Liu and Zhen Fang and Guangquan Zhang
- Abstract summary: We propose a novel framework to address a new realistic problem called multi-class classification with imprecise observations (MCIMO)
First, we give the theoretical analysis of the MCIMO problem based on fuzzy Rademacher complexity.
Then, two practical algorithms based on support vector machine and neural networks are constructed to solve the proposed new problem.
- Score: 36.810603503167755
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The theoretical analysis of multi-class classification has proved that the
existing multi-class classification methods can train a classifier with high
classification accuracy on the test set, when the instances are precise in the
training and test sets with same distribution and enough instances can be
collected in the training set. However, one limitation with multi-class
classification has not been solved: how to improve the classification accuracy
of multi-class classification problems when only imprecise observations are
available. Hence, in this paper, we propose a novel framework to address a new
realistic problem called multi-class classification with imprecise observations
(MCIMO), where we need to train a classifier with fuzzy-feature observations.
Firstly, we give the theoretical analysis of the MCIMO problem based on fuzzy
Rademacher complexity. Then, two practical algorithms based on support vector
machine and neural networks are constructed to solve the proposed new problem.
Experiments on both synthetic and real-world datasets verify the rationality of
our theoretical analysis and the efficacy of the proposed algorithms.
Related papers
- Anomaly Detection using Ensemble Classification and Evidence Theory [62.997667081978825]
We present a novel approach for novel detection using ensemble classification and evidence theory.
A pool selection strategy is presented to build a solid ensemble classifier.
We use uncertainty for the anomaly detection approach.
arXiv Detail & Related papers (2022-12-23T00:50:41Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - Set-valued prediction in hierarchical classification with constrained
representation complexity [4.258263831866309]
We focus on hierarchical multi-class classification problems, where valid sets correspond to internal nodes of the hierarchy.
We propose three methods and evaluate them on benchmark datasets.
arXiv Detail & Related papers (2022-03-13T15:13:19Z) - The SAMME.C2 algorithm for severely imbalanced multi-class
classification [0.0]
A minority class is often considered the more interesting class yet developing a scientific learning algorithm suitable for the observations presents countless challenges.
We suggest a novel multi-class classification algorithm specialized to handle severely imbalanced classes based on the method we refer to as SAMME.C2.
Not only do we provide the resulting algorithm but we also establish scientific and statistical formulation of our proposed SAMME.C2 algorithm.
arXiv Detail & Related papers (2021-12-30T00:20:01Z) - When in Doubt: Improving Classification Performance with Alternating
Normalization [57.39356691967766]
We introduce Classification with Alternating Normalization (CAN), a non-parametric post-processing step for classification.
CAN improves classification accuracy for challenging examples by re-adjusting their predicted class probability distribution.
We empirically demonstrate its effectiveness across a diverse set of classification tasks.
arXiv Detail & Related papers (2021-09-28T02:55:42Z) - Statistical Theory for Imbalanced Binary Classification [8.93993657323783]
We show that optimal classification performance depends on certain properties of class imbalance that have not previously been formalized.
Specifically, we propose a novel sub-type of class imbalance, which we call Uniform Class Imbalance.
These results provide some of the first meaningful finite-sample statistical theory for imbalanced binary classification.
arXiv Detail & Related papers (2021-07-05T03:55:43Z) - Theoretical Insights Into Multiclass Classification: A High-dimensional
Asymptotic View [82.80085730891126]
We provide the first modernally precise analysis of linear multiclass classification.
Our analysis reveals that the classification accuracy is highly distribution-dependent.
The insights gained may pave the way for a precise understanding of other classification algorithms.
arXiv Detail & Related papers (2020-11-16T05:17:29Z) - Conditional Classification: A Solution for Computational Energy
Reduction [2.182419181054266]
We propose a novel solution to reduce the computational complexity of convolutional neural network models.
Our proposed technique breaks the classification task into two steps: 1) coarse-grain classification, in which the input samples are classified among a set of hyper-classes, 2) fine-grain classification, in which the final labels are predicted among those hyper-classes detected at the first step.
arXiv Detail & Related papers (2020-06-29T03:50:39Z) - Saliency-based Weighted Multi-label Linear Discriminant Analysis [101.12909759844946]
We propose a new variant of Linear Discriminant Analysis (LDA) to solve multi-label classification tasks.
The proposed method is based on a probabilistic model for defining the weights of individual samples.
The Saliency-based weighted Multi-label LDA approach is shown to lead to performance improvements in various multi-label classification problems.
arXiv Detail & Related papers (2020-04-08T19:40:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.