Interpretable ML for Imbalanced Data
- URL: http://arxiv.org/abs/2212.07743v1
- Date: Thu, 15 Dec 2022 11:50:31 GMT
- Title: Interpretable ML for Imbalanced Data
- Authors: Damien A. Dablain, Colin Bellinger, Bartosz Krawczyk, David W. Aha,
Nitesh V. Chawla
- Abstract summary: Imbalanced data compounds the black-box nature of deep networks because the relationships between classes may be skewed and unclear.
Existing methods that investigate imbalanced data complexity are geared toward binary classification, shallow learning models and low dimensional data.
We propose a set of techniques that can be used by both deep learning model users to identify, visualize and understand class prototypes, sub-concepts and outlier instances.
- Score: 22.355966235617014
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning models are being increasingly applied to imbalanced data in
high stakes fields such as medicine, autonomous driving, and intelligence
analysis. Imbalanced data compounds the black-box nature of deep networks
because the relationships between classes may be highly skewed and unclear.
This can reduce trust by model users and hamper the progress of developers of
imbalanced learning algorithms. Existing methods that investigate imbalanced
data complexity are geared toward binary classification, shallow learning
models and low dimensional data. In addition, current eXplainable Artificial
Intelligence (XAI) techniques mainly focus on converting opaque deep learning
models into simpler models (e.g., decision trees) or mapping predictions for
specific instances to inputs, instead of examining global data properties and
complexities. Therefore, there is a need for a framework that is tailored to
modern deep networks, that incorporates large, high dimensional, multi-class
datasets, and uncovers data complexities commonly found in imbalanced data
(e.g., class overlap, sub-concepts, and outlier instances). We propose a set of
techniques that can be used by both deep learning model users to identify,
visualize and understand class prototypes, sub-concepts and outlier instances;
and by imbalanced learning algorithm developers to detect features and class
exemplars that are key to model performance. Our framework also identifies
instances that reside on the border of class decision boundaries, which can
carry highly discriminative information. Unlike many existing XAI techniques
which map model decisions to gray-scale pixel locations, we use saliency
through back-propagation to identify and aggregate image color bands across
entire classes. Our framework is publicly available at
\url{https://github.com/dd1github/XAI_for_Imbalanced_Learning}
Related papers
- Class Machine Unlearning for Complex Data via Concepts Inference and Data Poisoning [15.364125432212711]
In current AI era, users may request AI companies to delete their data from the training dataset due to privacy concerns.
Machine unlearning is a new emerged technology to allow model owner to delete requested training data or a class with little affecting on the model performance.
In this paper, we apply the definition of Concept, rather than an image feature or a token of text data, to represent the semantic information of unlearning class.
arXiv Detail & Related papers (2024-05-24T15:59:17Z) - Learn to Unlearn for Deep Neural Networks: Minimizing Unlearning
Interference with Gradient Projection [56.292071534857946]
Recent data-privacy laws have sparked interest in machine unlearning.
Challenge is to discard information about the forget'' data without altering knowledge about remaining dataset.
We adopt a projected-gradient based learning method, named as Projected-Gradient Unlearning (PGU)
We provide empirically evidence to demonstrate that our unlearning method can produce models that behave similar to models retrained from scratch across various metrics even when the training dataset is no longer accessible.
arXiv Detail & Related papers (2023-12-07T07:17:24Z) - Homological Convolutional Neural Networks [4.615338063719135]
We propose a novel deep learning architecture that exploits the data structural organization through topologically constrained network representations.
We test our model on 18 benchmark datasets against 5 classic machine learning and 3 deep learning models.
arXiv Detail & Related papers (2023-08-26T08:48:51Z) - Interpreting Black-box Machine Learning Models for High Dimensional
Datasets [40.09157165704895]
We train a black-box model on a high-dimensional dataset to learn the embeddings on which the classification is performed.
We then approximate the behavior of the black-box model by means of an interpretable surrogate model on the top-k feature space.
Our approach outperforms state-of-the-art methods like TabNet and XGboost when tested on different datasets.
arXiv Detail & Related papers (2022-08-29T07:36:17Z) - A Survey of Learning on Small Data: Generalization, Optimization, and
Challenge [101.27154181792567]
Learning on small data that approximates the generalization ability of big data is one of the ultimate purposes of AI.
This survey follows the active sampling theory under a PAC framework to analyze the generalization error and label complexity of learning on small data.
Multiple data applications that may benefit from efficient small data representation are surveyed.
arXiv Detail & Related papers (2022-07-29T02:34:19Z) - Few-Shot Non-Parametric Learning with Deep Latent Variable Model [50.746273235463754]
We propose Non-Parametric learning by Compression with Latent Variables (NPC-LV)
NPC-LV is a learning framework for any dataset with abundant unlabeled data but very few labeled ones.
We show that NPC-LV outperforms supervised methods on all three datasets on image classification in low data regime.
arXiv Detail & Related papers (2022-06-23T09:35:03Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - Semi-supervised Deep Learning for Image Classification with Distribution
Mismatch: A Survey [1.5469452301122175]
Deep learning models rely on the abundance of labelled observations to train a prospective model.
It is expensive to gather labelled observations of data, making the usage of deep learning models not ideal.
In many situations different unlabelled data sources might be available.
This raises the risk of a significant distribution mismatch between the labelled and unlabelled datasets.
arXiv Detail & Related papers (2022-03-01T02:46:00Z) - Improving Contrastive Learning on Imbalanced Seed Data via Open-World
Sampling [96.8742582581744]
We present an open-world unlabeled data sampling framework called Model-Aware K-center (MAK)
MAK follows three simple principles: tailness, proximity, and diversity.
We demonstrate that MAK can consistently improve both the overall representation quality and the class balancedness of the learned features.
arXiv Detail & Related papers (2021-11-01T15:09:41Z) - Relation-Guided Representation Learning [53.60351496449232]
We propose a new representation learning method that explicitly models and leverages sample relations.
Our framework well preserves the relations between samples.
By seeking to embed samples into subspace, we show that our method can address the large-scale and out-of-sample problem.
arXiv Detail & Related papers (2020-07-11T10:57:45Z) - Visualisation and knowledge discovery from interpretable models [0.0]
We introduce a few intrinsically interpretable models which are also capable of dealing with missing values.
We have demonstrated the algorithms on a synthetic dataset and a real-world one.
arXiv Detail & Related papers (2020-05-07T17:37:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.