Chest X-Rays Image Classification from beta-Variational Autoencoders
Latent Features
- URL: http://arxiv.org/abs/2109.14760v1
- Date: Wed, 29 Sep 2021 23:28:09 GMT
- Title: Chest X-Rays Image Classification from beta-Variational Autoencoders
Latent Features
- Authors: Leonardo Crespi, Daniele Loiacono, Arturo Chiti
- Abstract summary: We investigate and analyse the use of Deep Learning (DL) techniques to extract information from Chest X-Ray (CXR) images.
We trained several beta-Variational Autoencoder (beta-VAE) models on the CheXpert dataset, one of the largest publicly available collection of labeled CXR images.
Latent features have been extracted and used to train other Machine Learning models, able to classify the original images from the features extracted by the beta-VAE.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Chest X-Ray (CXR) is one of the most common diagnostic techniques used in
everyday clinical practice all around the world. We hereby present a work which
intends to investigate and analyse the use of Deep Learning (DL) techniques to
extract information from such images and allow to classify them, trying to keep
our methodology as general as possible and possibly also usable in a real world
scenario without much effort, in the future. To move in this direction, we
trained several beta-Variational Autoencoder (beta-VAE) models on the CheXpert
dataset, one of the largest publicly available collection of labeled CXR
images; from these models, latent features have been extracted and used to
train other Machine Learning models, able to classify the original images from
the features extracted by the beta-VAE. Lastly, tree-based models have been
combined together in ensemblings to improve the results without the necessity
of further training or models engineering. Expecting some drop in pure
performance with the respect to state of the art classification specific
models, we obtained encouraging results, which show the viability of our
approach and the usability of the high level features extracted by the
autoencoders for classification tasks.
Related papers
- Embeddings are all you need! Achieving High Performance Medical Image Classification through Training-Free Embedding Analysis [0.0]
Developing artificial intelligence (AI) and machine learning (ML) models for medical imaging typically involves extensive training and testing on large datasets.
We investigated the feasibility of replacing conventional training procedures with an embedding-based approach.
arXiv Detail & Related papers (2024-12-12T16:59:37Z) - Improving Interpretability and Robustness for the Detection of AI-Generated Images [6.116075037154215]
We analyze existing state-of-the-art AIGI detection methods based on frozen CLIP embeddings.
We show how to interpret them, shedding light on how images produced by various AI generators differ from real ones.
arXiv Detail & Related papers (2024-06-21T10:33:09Z) - Retrieval-Enhanced Contrastive Vision-Text Models [61.783728119255365]
We propose to equip vision-text models with the ability to refine their embedding with cross-modal retrieved information from a memory at inference time.
Remarkably, we show that this can be done with a light-weight, single-layer, fusion transformer on top of a frozen CLIP.
Our experiments validate that our retrieval-enhanced contrastive (RECO) training improves CLIP performance substantially on several challenging fine-grained tasks.
arXiv Detail & Related papers (2023-06-12T15:52:02Z) - Performance of GAN-based augmentation for deep learning COVID-19 image
classification [57.1795052451257]
The biggest challenge in the application of deep learning to the medical domain is the availability of training data.
Data augmentation is a typical methodology used in machine learning when confronted with a limited data set.
In this work, a StyleGAN2-ADA model of Generative Adversarial Networks is trained on the limited COVID-19 chest X-ray image set.
arXiv Detail & Related papers (2023-04-18T15:39:58Z) - DINOv2: Learning Robust Visual Features without Supervision [75.42921276202522]
This work shows that existing pretraining methods, especially self-supervised methods, can produce such features if trained on enough curated data from diverse sources.
Most of the technical contributions aim at accelerating and stabilizing the training at scale.
In terms of data, we propose an automatic pipeline to build a dedicated, diverse, and curated image dataset instead of uncurated data, as typically done in the self-supervised literature.
arXiv Detail & Related papers (2023-04-14T15:12:19Z) - Strong Baseline and Bag of Tricks for COVID-19 Detection of CT Scans [2.696776905220987]
Traditional deep learning frameworks encounter compatibility issues due to variations in slice numbers and resolutions in CT images.
We propose a novel slice selection method for each CT dataset to address this limitation.
In addition to the aforementioned methods, we explore various high-performance classification models, ultimately achieving promising results.
arXiv Detail & Related papers (2023-03-15T09:52:28Z) - Learning disentangled representations for explainable chest X-ray
classification using Dirichlet VAEs [68.73427163074015]
This study explores the use of the Dirichlet Variational Autoencoder (DirVAE) for learning disentangled latent representations of chest X-ray (CXR) images.
The predictive capacity of multi-modal latent representations learned by DirVAE models is investigated through implementation of an auxiliary multi-label classification task.
arXiv Detail & Related papers (2023-02-06T18:10:08Z) - Multi-Branch Deep Radial Basis Function Networks for Facial Emotion
Recognition [80.35852245488043]
We propose a CNN based architecture enhanced with multiple branches formed by radial basis function (RBF) units.
RBF units capture local patterns shared by similar instances using an intermediate representation.
We show it is the incorporation of local information what makes the proposed model competitive.
arXiv Detail & Related papers (2021-09-07T21:05:56Z) - Self-Supervised Representation Learning using Visual Field Expansion on
Digital Pathology [7.568373895297608]
A key challenge in the analysis of such images is their size, which can run into the gigapixels.
We propose a novel generative framework that can learn powerful representations for such tiles by learning to plausibly expand their visual field.
Our model learns to generate different tissue types with fine details, while simultaneously learning powerful representations that can be used for different clinical endpoints.
arXiv Detail & Related papers (2021-09-07T19:20:01Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Image Embedding and Model Ensembling for Automated Chest X-Ray
Interpretation [0.0]
We present and study several machine learning approaches to develop automated Chest X-ray diagnostic models.
In particular, we trained several Convolutional Neural Networks (CNN) on the CheXpert dataset.
We used the trained CNNs to compute embeddings of the CXR images, in order to train two sets of tree-based classifiers from them.
arXiv Detail & Related papers (2021-05-05T14:48:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.