Graph Embedded Intuitionistic Fuzzy Random Vector Functional Link Neural
Network for Class Imbalance Learning
- URL: http://arxiv.org/abs/2307.07881v2
- Date: Fri, 16 Feb 2024 12:58:10 GMT
- Title: Graph Embedded Intuitionistic Fuzzy Random Vector Functional Link Neural
Network for Class Imbalance Learning
- Authors: M.A. Ganaie, M. Sajid, A.K. Malik, M. Tanveer
- Abstract summary: We propose a graph embedded intuitionistic fuzzy RVFL for class imbalance learning (GE-IFRVFL-CIL) model incorporating a weighting mechanism to handle imbalanced datasets.
The proposed GE-IFRVFL-CIL model offers a promising solution to address the class imbalance issue, mitigates the detrimental effect of noise and outliers, and preserves the inherent geometrical structures of the dataset.
- Score: 4.069144210024564
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The domain of machine learning is confronted with a crucial research area
known as class imbalance learning, which presents considerable hurdles in
precise classification of minority classes. This issue can result in biased
models where the majority class takes precedence in the training process,
leading to the underrepresentation of the minority class. The random vector
functional link (RVFL) network is a widely used and effective learning model
for classification due to its good generalization performance and efficiency.
However, it suffers when dealing with imbalanced datasets. To overcome this
limitation, we propose a novel graph embedded intuitionistic fuzzy RVFL for
class imbalance learning (GE-IFRVFL-CIL) model incorporating a weighting
mechanism to handle imbalanced datasets. The proposed GE-IFRVFL-CIL model
offers plethora of benefits: $(i)$ leveraging graph embedding to preserve the
inherent topological structure of the datasets, $(ii)$ employing intuitionistic
fuzzy theory to handle uncertainty and imprecision in the data, $(iii)$ and the
most important, it tackles class imbalance learning. The amalgamation of a
weighting scheme, graph embedding, and intuitionistic fuzzy sets leads to the
superior performance of the proposed models on KEEL benchmark imbalanced
datasets with and without Gaussian noise. Furthermore, we implemented the
proposed GE-IFRVFL-CIL on the ADNI dataset and achieved promising results,
demonstrating the model's effectiveness in real-world applications. The
proposed GE-IFRVFL-CIL model offers a promising solution to address the class
imbalance issue, mitigates the detrimental effect of noise and outliers, and
preserves the inherent geometrical structures of the dataset.
Related papers
- Intuitionistic Fuzzy Universum Twin Support Vector Machine for Imbalanced Data [0.0]
One of the major difficulties in machine learning methods is categorizing datasets that are imbalanced.
We propose intuitionistic fuzzy universum twin support vector machines for imbalanced data (IFUTSVM-ID)
We use an intuitionistic fuzzy membership scheme to mitigate the impact of noise and outliers.
arXiv Detail & Related papers (2024-10-27T04:25:42Z) - Class-Balanced and Reinforced Active Learning on Graphs [13.239043161351482]
Graph neural networks (GNNs) have demonstrated significant success in various applications, such as node classification, link prediction, and graph classification.
Active learning for GNNs aims to query the valuable samples from the unlabeled data for annotation to maximize the GNNs' performance at a lower cost.
Most existing algorithms for reinforced active learning in GNNs may lead to a highly imbalanced class distribution, especially in highly skewed class scenarios.
We propose a novel class-balanced and reinforced active learning framework for GNNs, namely, GCBR. It learns an optimal policy to acquire class-balanced and informative nodes
arXiv Detail & Related papers (2024-02-15T16:37:14Z) - An extended asymmetric sigmoid with Perceptron (SIGTRON) for imbalanced linear classification [0.0]
This article presents a new parameterized sigmoid called SIGTRON, and its companion convex model called SIGTRON-imbalanced classification (SIC) model.
In contrast to the conventional $pi$-weighted cost-sensitive learning model, the SIC model does not have an external $pi$-weight on the loss function.
We show that the proposed SIC model is more adaptive to variations of the dataset.
arXiv Detail & Related papers (2023-12-26T13:14:17Z) - Class-Imbalanced Graph Learning without Class Rebalancing [62.1368829847041]
Class imbalance is prevalent in real-world node classification tasks and poses great challenges for graph learning models.
In this work, we approach the root cause of class-imbalance bias from an topological paradigm.
We devise a lightweight topological augmentation framework BAT to mitigate the class-imbalance bias without class rebalancing.
arXiv Detail & Related papers (2023-08-27T19:01:29Z) - Robust Learning with Progressive Data Expansion Against Spurious
Correlation [65.83104529677234]
We study the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features.
Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process.
We propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance.
arXiv Detail & Related papers (2023-06-08T05:44:06Z) - Stabilizing and Improving Federated Learning with Non-IID Data and
Client Dropout [15.569507252445144]
Label distribution skew induced data heterogeniety has been shown to be a significant obstacle that limits the model performance in federated learning.
We propose a simple yet effective framework by introducing a prior-calibrated softmax function for computing the cross-entropy loss.
The improved model performance over existing baselines in the presence of non-IID data and client dropout is demonstrated.
arXiv Detail & Related papers (2023-03-11T05:17:59Z) - Boosting Differentiable Causal Discovery via Adaptive Sample Reweighting [62.23057729112182]
Differentiable score-based causal discovery methods learn a directed acyclic graph from observational data.
We propose a model-agnostic framework to boost causal discovery performance by dynamically learning the adaptive weights for the Reweighted Score function, ReScore.
arXiv Detail & Related papers (2023-03-06T14:49:59Z) - Do We Really Need a Learnable Classifier at the End of Deep Neural
Network? [118.18554882199676]
We study the potential of learning a neural network for classification with the classifier randomly as an ETF and fixed during training.
Our experimental results show that our method is able to achieve similar performances on image classification for balanced datasets.
arXiv Detail & Related papers (2022-03-17T04:34:28Z) - Class Balancing GAN with a Classifier in the Loop [58.29090045399214]
We introduce a novel theoretically motivated Class Balancing regularizer for training GANs.
Our regularizer makes use of the knowledge from a pre-trained classifier to ensure balanced learning of all the classes in the dataset.
We demonstrate the utility of our regularizer in learning representations for long-tailed distributions via achieving better performance than existing approaches over multiple datasets.
arXiv Detail & Related papers (2021-06-17T11:41:30Z) - AdaGCN:Adaptive Boosting Algorithm for Graph Convolutional Networks on
Imbalanced Node Classification [10.72543417177307]
We propose an ensemble model called AdaGCN, which uses a Graph Convolutional Network (GCN) as the base estimator during adaptive boosting.
Our model also improves state-of-the-art baselines on all of the challenging node classification tasks we consider.
arXiv Detail & Related papers (2021-05-25T02:43:31Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.