AUC-mixup: Deep AUC Maximization with Mixup
- URL: http://arxiv.org/abs/2310.11693v1
- Date: Wed, 18 Oct 2023 03:43:11 GMT
- Title: AUC-mixup: Deep AUC Maximization with Mixup
- Authors: Jianzhi Xv, Gang Li and Tianbao Yang
- Abstract summary: AUC is defined over positive and negative pairs, which makes it challenging to incorporate mixup data augmentation into DAM.
We employ the AUC margin loss and soft labels into the formulation to effectively learn from data generated by mixup.
Our experimental results demonstrate the effectiveness of the proposed AUC-mixup methods on imbalanced benchmark and medical image datasets.
- Score: 47.99058341229214
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While deep AUC maximization (DAM) has shown remarkable success on imbalanced
medical tasks, e.g., chest X-rays classification and skin lesions
classification, it could suffer from severe overfitting when applied to small
datasets due to its aggressive nature of pushing prediction scores of positive
data away from that of negative data. This paper studies how to improve
generalization of DAM by mixup data augmentation -- an approach that is widely
used for improving generalization of the cross-entropy loss based deep learning
methods. %For overfitting issues arising from limited data, the common approach
is to employ mixup data augmentation to boost the models' generalization
performance by enriching the training data. However, AUC is defined over
positive and negative pairs, which makes it challenging to incorporate mixup
data augmentation into DAM algorithms. To tackle this challenge, we employ the
AUC margin loss and incorporate soft labels into the formulation to effectively
learn from data generated by mixup augmentation, which is referred to as the
AUC-mixup loss. Our experimental results demonstrate the effectiveness of the
proposed AUC-mixup methods on imbalanced benchmark and medical image datasets
compared to standard DAM training methods.
Related papers
- DyMix: Dynamic Frequency Mixup Scheduler based Unsupervised Domain Adaptation for Enhancing Alzheimer's Disease Identification [9.506504023554031]
We propose a novel approach called the dynamic frequency mixup scheduler (DyMix) for unsupervised domain adaptation.
Our proposed DyMix adjusts the magnitude of the frequency regions being mixed from the source and target domains.
We demonstrate its outstanding performance in Alzheimer's disease diagnosis compared to state-of-the-art methods.
arXiv Detail & Related papers (2024-10-02T07:18:47Z) - On the Influence of Data Resampling for Deep Learning-Based Log Anomaly Detection: Insights and Recommendations [10.931620604044486]
This study provides an in-depth analysis of the impact of diverse data resampling methods on existingAD approaches.
We assess the performance of theseAD approaches across four datasets with different levels of class imbalance.
We evaluate the effectiveness of the data resampling methods when utilizing optimal resampling ratios of normal to abnormal data.
arXiv Detail & Related papers (2024-05-06T14:01:05Z) - ArSDM: Colonoscopy Images Synthesis with Adaptive Refinement Semantic
Diffusion Models [69.9178140563928]
Colonoscopy analysis is essential for assisting clinical diagnosis and treatment.
The scarcity of annotated data limits the effectiveness and generalization of existing methods.
We propose an Adaptive Refinement Semantic Diffusion Model (ArSDM) to generate colonoscopy images that benefit the downstream tasks.
arXiv Detail & Related papers (2023-09-03T07:55:46Z) - DualMix: Unleashing the Potential of Data Augmentation for Online
Class-Incremental Learning [14.194817677415065]
We show that augmented samples with lower correlation to the original data are more effective in preventing forgetting.
We propose the Enhanced Mixup (EnMix) method that mixes the augmented samples and their labels simultaneously.
To solve the class imbalance problem, we design an Adaptive Mixup (AdpMix) method to calibrate the decision boundaries.
arXiv Detail & Related papers (2023-03-14T12:55:42Z) - Bootstrapping Your Own Positive Sample: Contrastive Learning With
Electronic Health Record Data [62.29031007761901]
This paper proposes a novel contrastive regularized clinical classification model.
We introduce two unique positive sampling strategies specifically tailored for EHR data.
Our framework yields highly competitive experimental results in predicting the mortality risk on real-world COVID-19 EHR data.
arXiv Detail & Related papers (2021-04-07T06:02:04Z) - Federated Deep AUC Maximization for Heterogeneous Data with a Constant
Communication Complexity [77.78624443410216]
We propose improved FDAM algorithms for detecting heterogeneous chest data.
A result of this paper is that the communication of the proposed algorithm is strongly independent of the number of machines and also independent of the accuracy level.
Experiments have demonstrated the effectiveness of our FDAM algorithm on benchmark datasets and on medical chest Xray images from different organizations.
arXiv Detail & Related papers (2021-02-09T04:05:19Z) - Robust Deep AUC Maximization: A New Surrogate Loss and Empirical Studies
on Medical Image Classification [63.44396343014749]
We propose a new margin-based surrogate loss function for the AUC score.
It is more robust than the commonly used.
square loss while enjoying the same advantage in terms of large-scale optimization.
To the best of our knowledge, this is the first work that makes DAM succeed on large-scale medical image datasets.
arXiv Detail & Related papers (2020-12-06T03:41:51Z) - Automatic Data Augmentation via Deep Reinforcement Learning for
Effective Kidney Tumor Segmentation [57.78765460295249]
We develop a novel automatic learning-based data augmentation method for medical image segmentation.
In our method, we innovatively combine the data augmentation module and the subsequent segmentation module in an end-to-end training manner with a consistent loss.
We extensively evaluated our method on CT kidney tumor segmentation which validated the promising results of our method.
arXiv Detail & Related papers (2020-02-22T14:10:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.