Classifier Transfer with Data Selection Strategies for Online Support
Vector Machine Classification with Class Imbalance
- URL: http://arxiv.org/abs/2208.05112v1
- Date: Wed, 10 Aug 2022 02:36:20 GMT
- Title: Classifier Transfer with Data Selection Strategies for Online Support
Vector Machine Classification with Class Imbalance
- Authors: Mario Michael Krell, Nils Wilshusen, Anett Seeland, Su Kyoung Kim
- Abstract summary: We focus on data selection strategies which limit the size of the stored training data.
We show that by using the right combination of data selection criteria, it is possible to adapt the classifier and largely increase the performance.
- Score: 1.2599533416395767
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Objective: Classifier transfers usually come with dataset shifts. To overcome
them, online strategies have to be applied. For practical applications,
limitations in the computational resources for the adaptation of batch learning
algorithms, like the SVM, have to be considered.
Approach: We review and compare several strategies for online learning with
SVMs. We focus on data selection strategies which limit the size of the stored
training data [...]
Main Results: For different data shifts, different criteria are appropriate.
For the synthetic data, adding all samples to the pool of considered samples
performs often significantly worse than other criteria. Especially, adding only
misclassified samples performed astoundingly well. Here, balancing criteria
were very important when the other criteria were not well chosen. For the
transfer setups, the results show that the best strategy depends on the
intensity of the drift during the transfer. Adding all and removing the oldest
samples results in the best performance, whereas for smaller drifts, it can be
sufficient to only add potential new support vectors of the SVM which reduces
processing resources.
Significance: For BCIs based on EEG models, trained on data from a
calibration session, a previous recording session, or even from a recording
session with one or several other subjects, are used. This transfer of the
learned model usually decreases the performance and can therefore benefit from
online learning which adapts the classifier like the established SVM. We show
that by using the right combination of data selection criteria, it is possible
to adapt the classifier and largely increase the performance. Furthermore, in
some cases it is possible to speed up the processing and save computational by
updating with a subset of special samples and keeping a small subset of samples
for training the classifier.
Related papers
- Adapt-$\infty$: Scalable Lifelong Multimodal Instruction Tuning via Dynamic Data Selection [89.42023974249122]
Adapt-$infty$ is a new multi-way and adaptive data selection approach for Lifelong Instruction Tuning.
We construct pseudo-skill clusters by grouping gradient-based sample vectors.
We select the best-performing data selector for each skill cluster from a pool of selector experts.
arXiv Detail & Related papers (2024-10-14T15:48:09Z) - Dataset Quantization with Active Learning based Adaptive Sampling [11.157462442942775]
We show that maintaining performance is feasible even with uneven sample distributions.
We propose a novel active learning based adaptive sampling strategy to optimize the sample selection.
Our approach outperforms the state-of-the-art dataset compression methods.
arXiv Detail & Related papers (2024-07-09T23:09:18Z) - Intra-class Adaptive Augmentation with Neighbor Correction for Deep
Metric Learning [99.14132861655223]
We propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning.
We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining.
Our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
arXiv Detail & Related papers (2022-11-29T14:52:38Z) - Explaining Cross-Domain Recognition with Interpretable Deep Classifier [100.63114424262234]
Interpretable Deep (IDC) learns the nearest source samples of a target sample as evidence upon which the classifier makes the decision.
Our IDC leads to a more explainable model with almost no accuracy degradation and effectively calibrates classification for optimum reject options.
arXiv Detail & Related papers (2022-11-15T15:58:56Z) - Adaptive Distribution Calibration for Few-Shot Learning with
Hierarchical Optimal Transport [78.9167477093745]
We propose a novel distribution calibration method by learning the adaptive weight matrix between novel samples and base classes.
Experimental results on standard benchmarks demonstrate that our proposed plug-and-play model outperforms competing approaches.
arXiv Detail & Related papers (2022-10-09T02:32:57Z) - Towards Automated Imbalanced Learning with Deep Hierarchical
Reinforcement Learning [57.163525407022966]
Imbalanced learning is a fundamental challenge in data mining, where there is a disproportionate ratio of training samples in each class.
Over-sampling is an effective technique to tackle imbalanced learning through generating synthetic samples for the minority class.
We propose AutoSMOTE, an automated over-sampling algorithm that can jointly optimize different levels of decisions.
arXiv Detail & Related papers (2022-08-26T04:28:01Z) - Lightweight Conditional Model Extrapolation for Streaming Data under
Class-Prior Shift [27.806085423595334]
We introduce LIMES, a new method for learning with non-stationary streaming data.
We learn a single set of model parameters from which a specific classifier for any specific data distribution is derived.
Experiments on a set of exemplary tasks using Twitter data show that LIMES achieves higher accuracy than alternative approaches.
arXiv Detail & Related papers (2022-06-10T15:19:52Z) - Message Passing Adaptive Resonance Theory for Online Active
Semi-supervised Learning [30.19936050747407]
We propose Message Passing Adaptive Resonance Theory (MPART) for online active semi-supervised learning.
MPART infers the class of unlabeled data and selects informative and representative samples through message passing between nodes on the topological graph.
We evaluate our model with comparable query selection strategies and frequencies, showing that MPART significantly outperforms the competitive models in online active learning environments.
arXiv Detail & Related papers (2020-12-02T14:14:42Z) - Ensemble Wrapper Subsampling for Deep Modulation Classification [70.91089216571035]
Subsampling of received wireless signals is important for relaxing hardware requirements as well as the computational cost of signal processing algorithms.
We propose a subsampling technique to facilitate the use of deep learning for automatic modulation classification in wireless communication systems.
arXiv Detail & Related papers (2020-05-10T06:11:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.