End to End Generative Meta Curriculum Learning For Medical Data
Augmentation
- URL: http://arxiv.org/abs/2212.10086v1
- Date: Tue, 20 Dec 2022 08:54:11 GMT
- Title: End to End Generative Meta Curriculum Learning For Medical Data
Augmentation
- Authors: Meng Li, Brian Lovell
- Abstract summary: Current medical image synthetic augmentation techniques rely on intensive use of generative adversarial networks (GANs)
We introduce a novel generative meta curriculum learning method that trains the task-specific model (student) end-to-end with only one additional teacher model.
In contrast to the generator and discriminator in GAN, which compete with each other, the teacher and student collaborate to improve the student's performance on the target tasks.
- Score: 2.471925498075058
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Current medical image synthetic augmentation techniques rely on intensive use
of generative adversarial networks (GANs). However, the nature of GAN
architecture leads to heavy computational resources to produce synthetic images
and the augmentation process requires multiple stages to complete. To address
these challenges, we introduce a novel generative meta curriculum learning
method that trains the task-specific model (student) end-to-end with only one
additional teacher model. The teacher learns to generate curriculum to feed
into the student model for data augmentation and guides the student to improve
performance in a meta-learning style. In contrast to the generator and
discriminator in GAN, which compete with each other, the teacher and student
collaborate to improve the student's performance on the target tasks. Extensive
experiments on the histopathology datasets show that leveraging our framework
results in significant and consistent improvements in classification
performance.
Related papers
- CFTS-GAN: Continual Few-Shot Teacher Student for Generative Adversarial Networks [0.5024983453990064]
Few-shot and continual learning face two well-known challenges in GANs: overfitting and catastrophic forgetting.
This paper proposes a Continual Few-shot Teacher-Student technique for the generative adversarial network (CFTS-GAN) that considers both challenges together.
arXiv Detail & Related papers (2024-10-17T20:49:08Z) - PMT: Progressive Mean Teacher via Exploring Temporal Consistency for Semi-Supervised Medical Image Segmentation [51.509573838103854]
We propose a semi-supervised learning framework, termed Progressive Mean Teachers (PMT), for medical image segmentation.
Our PMT generates high-fidelity pseudo labels by learning robust and diverse features in the training process.
Experimental results on two datasets with different modalities, i.e., CT and MRI, demonstrate that our method outperforms the state-of-the-art medical image segmentation approaches.
arXiv Detail & Related papers (2024-09-08T15:02:25Z) - CLDA: Collaborative Learning for Enhanced Unsupervised Domain Adaptation [15.97351561456467]
Collaborative Learning is a method that updates the teacher's non-salient parameters using the student model and at the same time enhance the student's performance.
CLDA achieves an improvement of +0.7% mIoU for teacher and +1.4% mIoU for student compared to the baseline model in the GTA to Cityscapes.
arXiv Detail & Related papers (2024-09-04T13:35:15Z) - Diffusion-Based Neural Network Weights Generation [80.89706112736353]
D2NWG is a diffusion-based neural network weights generation technique that efficiently produces high-performing weights for transfer learning.
Our method extends generative hyper-representation learning to recast the latent diffusion paradigm for neural network weights generation.
Our approach is scalable to large architectures such as large language models (LLMs), overcoming the limitations of current parameter generation techniques.
arXiv Detail & Related papers (2024-02-28T08:34:23Z) - CLGT: A Graph Transformer for Student Performance Prediction in
Collaborative Learning [6.140954034246379]
We present an extended graph transformer framework for collaborative learning (CLGT) for evaluating and predicting the performance of students.
The experimental results confirm that the proposed CLGT outperforms the baseline models in terms of performing predictions based on the real-world datasets.
arXiv Detail & Related papers (2023-07-30T09:54:30Z) - Bridging Synthetic and Real Images: a Transferable and Multiple
Consistency aided Fundus Image Enhancement Framework [61.74188977009786]
We propose an end-to-end optimized teacher-student framework to simultaneously conduct image enhancement and domain adaptation.
We also propose a novel multi-stage multi-attention guided enhancement network (MAGE-Net) as the backbones of our teacher and student network.
arXiv Detail & Related papers (2023-02-23T06:16:15Z) - EmbedDistill: A Geometric Knowledge Distillation for Information
Retrieval [83.79667141681418]
Large neural models (such as Transformers) achieve state-of-the-art performance for information retrieval (IR)
We propose a novel distillation approach that leverages the relative geometry among queries and documents learned by the large teacher model.
We show that our approach successfully distills from both dual-encoder (DE) and cross-encoder (CE) teacher models to 1/10th size asymmetric students that can retain 95-97% of the teacher performance.
arXiv Detail & Related papers (2023-01-27T22:04:37Z) - Synthetic data generation method for data-free knowledge distillation in
regression neural networks [0.0]
Knowledge distillation is the technique of compressing a larger neural network, known as the teacher, into a smaller neural network, known as the student.
Previous work has proposed a data-free knowledge distillation method where synthetic data are generated using a generator model trained adversarially against the student model.
In this study, we investigate the behavior of various synthetic data generation methods and propose a new synthetic data generation strategy.
arXiv Detail & Related papers (2023-01-11T07:26:00Z) - Unified Framework for Histopathology Image Augmentation and Classification via Generative Models [6.404713841079193]
We propose an innovative unified framework that integrates the data generation and model training stages into a unified process.
Our approach utilizes a pure Vision Transformer (ViT)-based conditional Generative Adversarial Network (cGAN) model to simultaneously handle both image synthesis and classification.
Our experiments show that our unified synthetic augmentation framework consistently enhances the performance of histopathology image classification models.
arXiv Detail & Related papers (2022-12-20T03:40:44Z) - Dynamically Grown Generative Adversarial Networks [111.43128389995341]
We propose a method to dynamically grow a GAN during training, optimizing the network architecture and its parameters together with automation.
The method embeds architecture search techniques as an interleaving step with gradient-based training to periodically seek the optimal architecture-growing strategy for the generator and discriminator.
arXiv Detail & Related papers (2021-06-16T01:25:51Z) - Distilling portable Generative Adversarial Networks for Image
Translation [101.33731583985902]
Traditional network compression methods focus on visually recognition tasks, but never deal with generation tasks.
Inspired by knowledge distillation, a student generator of fewer parameters is trained by inheriting the low-level and high-level information from the original heavy teacher generator.
An adversarial learning process is established to optimize student generator and student discriminator.
arXiv Detail & Related papers (2020-03-07T05:53:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.