Complex Mixer for MedMNIST Classification Decathlon
- URL: http://arxiv.org/abs/2304.10054v1
- Date: Thu, 20 Apr 2023 02:34:36 GMT
- Title: Complex Mixer for MedMNIST Classification Decathlon
- Authors: Zhuoran Zheng and Xiuyi Jia
- Abstract summary: We develop a Complex Mixer (C-Mixer) with a pre-training framework to alleviate the problem of insufficient information and uncertainty in the label space.
Our method shows surprising potential on both the standard MedMNIST (v2) dataset and the customized weakly supervised datasets.
- Score: 12.402054374952485
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the development of the medical image field, researchers seek to develop
a class of datasets to block the need for medical knowledge, such as
\text{MedMNIST} (v2). MedMNIST (v2) includes a large number of small-sized (28
$\times$ 28 or 28 $\times$ 28 $\times$ 28) medical samples and the
corresponding expert annotations (class label). The existing baseline model
(Google AutoML Vision, ResNet-50+3D) can reach an average accuracy of over 70\%
on MedMNIST (v2) datasets, which is comparable to the performance of expert
decision-making. Nevertheless, we note that there are two insurmountable
obstacles to modeling on MedMNIST (v2): 1) the raw images are cropped to low
scales may cause effective recognition information to be dropped and the
classifier to have difficulty in tracing accurate decision boundaries; 2) the
labelers' subjective insight may cause many uncertainties in the label space.
To address these issues, we develop a Complex Mixer (C-Mixer) with a
pre-training framework to alleviate the problem of insufficient information and
uncertainty in the label space by introducing an incentive imaginary matrix and
a self-supervised scheme with random masking. Our method (incentive learning
and self-supervised learning with masking) shows surprising potential on both
the standard MedMNIST (v2) dataset, the customized weakly supervised datasets,
and other image enhancement tasks.
Related papers
- FEDMEKI: A Benchmark for Scaling Medical Foundation Models via Federated Knowledge Injection [83.54960238236548]
FEDMEKI not only preserves data privacy but also enhances the capability of medical foundation models.
FEDMEKI allows medical foundation models to learn from a broader spectrum of medical knowledge without direct data exposure.
arXiv Detail & Related papers (2024-08-17T15:18:56Z) - Fair Text to Medical Image Diffusion Model with Subgroup Distribution Aligned Tuning [12.064840522920251]
The text to medical image (T2MedI) with latent diffusion model has great potential to alleviate the scarcity of medical imaging data.
However, as the text to nature image models, we show that the T2MedI model can also bias to some subgroups to overlook the minority ones in the training set.
In this work, we first build a T2MedI model based on the pre-trained Imagen model, which has the fixed contrastive language-image pre-training (CLIP) text encoder.
Its decoder has been fine-tuned on medical images from the Radiology Objects in C
arXiv Detail & Related papers (2024-06-21T03:23:37Z) - Robust and Interpretable Medical Image Classifiers via Concept
Bottleneck Models [49.95603725998561]
We propose a new paradigm to build robust and interpretable medical image classifiers with natural language concepts.
Specifically, we first query clinical concepts from GPT-4, then transform latent image features into explicit concepts with a vision-language model.
arXiv Detail & Related papers (2023-10-04T21:57:09Z) - Semi-Supervised Medical Image Segmentation with Co-Distribution
Alignment [16.038016822861092]
This paper proposes Co-Distribution Alignment (Co-DA) for semi-supervised medical image segmentation.
Co-DA aligns marginal predictions on unlabeled data to marginal predictions on labeled data in a class-wise manner.
We show that the proposed approach outperforms existing state-of-the-art semi-supervised medical image segmentation methods.
arXiv Detail & Related papers (2023-07-24T09:08:30Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Multi-Level Global Context Cross Consistency Model for Semi-Supervised
Ultrasound Image Segmentation with Diffusion Model [0.0]
We propose a framework that uses images generated by a Latent Diffusion Model (LDM) as unlabeled images for semi-supervised learning.
Our approach enables the effective transfer of probability distribution knowledge to the segmentation network, resulting in improved segmentation accuracy.
arXiv Detail & Related papers (2023-05-16T14:08:24Z) - Self-Supervised Learning as a Means To Reduce the Need for Labeled Data
in Medical Image Analysis [64.4093648042484]
We use a dataset of chest X-ray images with bounding box labels for 13 different classes of anomalies.
We show that it is possible to achieve similar performance to a fully supervised model in terms of mean average precision and accuracy with only 60% of the labeled data.
arXiv Detail & Related papers (2022-06-01T09:20:30Z) - DeepMCAT: Large-Scale Deep Clustering for Medical Image Categorization [24.100651548850895]
We propose an unsupervised approach for automatically clustering and categorizing large-scale medical image datasets.
We investigated the end-to-end training using both class-balanced and imbalanced large-scale datasets.
arXiv Detail & Related papers (2021-09-30T22:39:57Z) - G-MIND: An End-to-End Multimodal Imaging-Genetics Framework for
Biomarker Identification and Disease Classification [49.53651166356737]
We propose a novel deep neural network architecture to integrate imaging and genetics data, as guided by diagnosis, that provides interpretable biomarkers.
We have evaluated our model on a population study of schizophrenia that includes two functional MRI (fMRI) paradigms and Single Nucleotide Polymorphism (SNP) data.
arXiv Detail & Related papers (2021-01-27T19:28:04Z) - Semi-supervised Medical Image Classification with Relation-driven
Self-ensembling Model [71.80319052891817]
We present a relation-driven semi-supervised framework for medical image classification.
It exploits the unlabeled data by encouraging the prediction consistency of given input under perturbations.
Our method outperforms many state-of-the-art semi-supervised learning methods on both single-label and multi-label image classification scenarios.
arXiv Detail & Related papers (2020-05-15T06:57:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.