Probability-driven scoring functions in combining linear classifiers
- URL: http://arxiv.org/abs/2109.07815v1
- Date: Thu, 16 Sep 2021 08:58:32 GMT
- Title: Probability-driven scoring functions in combining linear classifiers
- Authors: Pawel Trajdos, Robert Burduk
- Abstract summary: This research is aimed at building a new fusion method dedicated to the ensemble of linear classifiers.
The proposed fusion method is compared with the reference method using multiple benchmark datasets taken from the KEEL repository.
The experimental study shows that, under certain conditions, some improvement may be obtained.
- Score: 0.913755431537592
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although linear classifiers are one of the oldest methods in machine
learning, they are still very popular in the machine learning community. This
is due to their low computational complexity and robustness to overfitting.
Consequently, linear classifiers are often used as base classifiers of multiple
ensemble classification systems. This research is aimed at building a new
fusion method dedicated to the ensemble of linear classifiers. The fusion
scheme uses both measurement space and geometrical space. Namely, we proposed a
probability-driven scoring function which shape depends on the orientation of
the decision hyperplanes generated by the base classifiers. The proposed fusion
method is compared with the reference method using multiple benchmark datasets
taken from the KEEL repository. The comparison is done using multiple quality
criteria. The statistical analysis of the obtained results is also performed.
The experimental study shows that, under certain conditions, some improvement
may be obtained.
Related papers
- Linear-time One-Class Classification with Repeated Element-wise Folding [6.116088814650622]
This paper proposes an easy-to-use method for one-class classification: Repeated Element-wise Folding (REF)
REF provides a linear-time alternative for the commonly used computationally much more demanding approaches.
Experiments show that REF can produce similar classification performance or even outperform the more complex algorithms on various benchmark datasets.
arXiv Detail & Related papers (2024-08-21T08:18:39Z) - Synergistic eigenanalysis of covariance and Hessian matrices for enhanced binary classification [72.77513633290056]
We present a novel approach that combines the eigenanalysis of a covariance matrix evaluated on a training set with a Hessian matrix evaluated on a deep learning model.
Our method captures intricate patterns and relationships, enhancing classification performance.
arXiv Detail & Related papers (2024-02-14T16:10:42Z) - Obtaining Explainable Classification Models using Distributionally
Robust Optimization [12.511155426574563]
We study generalized linear models constructed using sets of feature value rules.
An inherent trade-off exists between rule set sparsity and its prediction accuracy.
We propose a new formulation to learn an ensemble of rule sets that simultaneously addresses these competing factors.
arXiv Detail & Related papers (2023-11-03T15:45:34Z) - Convolutional autoencoder-based multimodal one-class classification [80.52334952912808]
One-class classification refers to approaches of learning using data from a single class only.
We propose a deep learning one-class classification method suitable for multimodal data.
arXiv Detail & Related papers (2023-09-25T12:31:18Z) - Anomaly Detection using Ensemble Classification and Evidence Theory [62.997667081978825]
We present a novel approach for novel detection using ensemble classification and evidence theory.
A pool selection strategy is presented to build a solid ensemble classifier.
We use uncertainty for the anomaly detection approach.
arXiv Detail & Related papers (2022-12-23T00:50:41Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - Exploring Category-correlated Feature for Few-shot Image Classification [27.13708881431794]
We present a simple yet effective feature rectification method by exploring the category correlation between novel and base classes as the prior knowledge.
The proposed approach consistently obtains considerable performance gains on three widely used benchmarks.
arXiv Detail & Related papers (2021-12-14T08:25:24Z) - Building an Ensemble of Classifiers via Randomized Models of Ensemble
Members [1.827510863075184]
In this paper, a novel randomized model of base classifier is developed.
In the proposed method, the random operation of the model results from a random selection of the learning set from the family of learning sets of a fixed size.
The DES scheme with the proposed model of competence was experimentally evaluated on the collection of 67 benchmark datasets.
arXiv Detail & Related papers (2021-09-16T10:53:13Z) - Linear Classifier Combination via Multiple Potential Functions [0.6091702876917279]
We propose a novel concept of calculating a scoring function based on the distance of the object from the decision boundary and its distance to the class centroid.
An important property is that the proposed score function has the same nature for all linear base classifiers.
arXiv Detail & Related papers (2020-10-02T08:11:51Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z) - Clustering Binary Data by Application of Combinatorial Optimization
Heuristics [52.77024349608834]
We study clustering methods for binary data, first defining aggregation criteria that measure the compactness of clusters.
Five new and original methods are introduced, using neighborhoods and population behavior optimization metaheuristics.
From a set of 16 data tables generated by a quasi-Monte Carlo experiment, a comparison is performed for one of the aggregations using L1 dissimilarity, with hierarchical clustering, and a version of k-means: partitioning around medoids or PAM.
arXiv Detail & Related papers (2020-01-06T23:33:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.