Deep Neuroevolution Squeezes More out of Small Neural Networks and Small
Training Sets: Sample Application to MRI Brain Sequence Classification
- URL: http://arxiv.org/abs/2112.12990v1
- Date: Fri, 24 Dec 2021 08:21:52 GMT
- Title: Deep Neuroevolution Squeezes More out of Small Neural Networks and Small
Training Sets: Sample Application to MRI Brain Sequence Classification
- Authors: Joseph N Stember, Hrithwik Shalu
- Abstract summary: Deep Neuroevolution (DNE) holds the promise of providing radiology artificial intelligence (AI) that performs well with small neural networks and small training sets.
We analyzed a training set of 20 patients, each with four sequences/weightings: T1, T1 post-contrast, T2, and T2-FLAIR.
We trained the parameters of a relatively small convolutional neural network (CNN) as follows: First, we randomly mutated the CNN weights. We then measured the CNN training set accuracy, using the latter as the fitness evaluation metric.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Purpose: Deep Neuroevolution (DNE) holds the promise of providing radiology
artificial intelligence (AI) that performs well with small neural networks and
small training sets. We seek to realize this potential via a proof-of-principle
application to MRI brain sequence classification.
Methods: We analyzed a training set of 20 patients, each with four
sequences/weightings: T1, T1 post-contrast, T2, and T2-FLAIR. We trained the
parameters of a relatively small convolutional neural network (CNN) as follows:
First, we randomly mutated the CNN weights. We then measured the CNN training
set accuracy, using the latter as the fitness evaluation metric. The fittest
child CNNs were identified. We incorporated their mutations into the parent
CNN. This selectively mutated parent became the next generation's parent CNN.
We repeated this process for approximately 50,000 generations.
Results: DNE achieved monotonic convergence to 100% training set accuracy.
DNE also converged monotonically to 100% testing set accuracy.
Conclusions: DNE can achieve perfect accuracy with small training sets and
small CNNs. Particularly when combined with Deep Reinforcement Learning, DNE
may provide a path forward in the quest to make radiology AI more human-like in
its ability to learn. DNE may very well turn out to be a key component of the
much-anticipated meta-learning regime of radiology AI algorithms that can adapt
to new tasks and new image types, similar to human radiologists.
Related papers
- Verified Neural Compressed Sensing [58.98637799432153]
We develop the first (to the best of our knowledge) provably correct neural networks for a precise computational task.
We show that for modest problem dimensions (up to 50), we can train neural networks that provably recover a sparse vector from linear and binarized linear measurements.
We show that the complexity of the network can be adapted to the problem difficulty and solve problems where traditional compressed sensing methods are not known to provably work.
arXiv Detail & Related papers (2024-05-07T12:20:12Z) - Convolutional Neural Networks Exploiting Attributes of Biological
Neurons [7.3517426088986815]
Deep neural networks like Convolutional Neural Networks (CNNs) have emerged as front-runners, often surpassing human capabilities.
Here, we integrate the principles of biological neurons in certain layer(s) of CNNs.
We aim to extract image features to use as input to CNNs, hoping to enhance training efficiency and achieve better accuracy.
arXiv Detail & Related papers (2023-11-14T16:58:18Z) - Curriculum Based Multi-Task Learning for Parkinson's Disease Detection [2.1187904593676845]
We leverage severity-based meta-data on the stages of disease to define a curriculum for training a deep convolutional neural network (CNN)
curriculum learning is a training strategy that aims to boost classifier performance by starting with examples that are easier to classify.
Even with our multi-task setting using pre-trained CNNs and transfer learning, PD classification based on T1-weighted (T1-w) MRI was challenging (ROC AUC: 0.59-0.65), but curriculum training boosted performance (by 3.9%) compared to our baseline model.
arXiv Detail & Related papers (2023-02-27T09:58:09Z) - Deep neuroevolution for limited, heterogeneous data: proof-of-concept
application to Neuroblastoma brain metastasis using a small virtual pooled
image collection [0.0]
We seek to address both overfitting and generalizability by applying DNE to a virtually pooled data set consisting of images from various institutions.
Our use case is classifying neuroblastoma brain metastases on MRI. Neuroblastoma is well-suited for our goals because it is a rare cancer.
As in prior DNE work, we used a small training set, consisting of 30 normal and 30 metastasis-containing post-contrast MRI brain scans, with 37% outside images.
arXiv Detail & Related papers (2022-11-26T07:03:37Z) - Direct evaluation of progression or regression of disease burden in
brain metastatic disease with Deep Neuroevolution [0.0]
Deep neuroevolution (DNE) can produce radiology artificial intelligence (AI) that performs well on small training sets.
Here we use DNE for function approximation that predicts progression versus regression of metastatic brain disease.
arXiv Detail & Related papers (2022-03-24T05:29:09Z) - Infant Brain Age Classification: 2D CNN Outperforms 3D CNN in Small
Dataset [0.14063138455565613]
Brain magnetic resonance imaging (MRI) of infants demonstrates a specific pattern of development beyond myelination.
With no standardized criteria, visual estimation of the structural maturity of the brain from MRI before three years of age remains dominated by inter-observer and intra-observer variability.
We explore the general feasibility to tackle this task, and the utility of different approaches, including two- and three-dimensional convolutional neural networks (CNN)
In the best performing approach, we achieved an accuracy of 0.90 [95% CI:0.86-0.94] using a 2D CNN on a central axial thick slab.
arXiv Detail & Related papers (2021-12-27T18:02:48Z) - Classification of diffraction patterns using a convolutional neural
network in single particle imaging experiments performed at X-ray
free-electron lasers [53.65540150901678]
Single particle imaging (SPI) at X-ray free electron lasers (XFELs) is particularly well suited to determine the 3D structure of particles in their native environment.
For a successful reconstruction, diffraction patterns originating from a single hit must be isolated from a large number of acquired patterns.
We propose to formulate this task as an image classification problem and solve it using convolutional neural network (CNN) architectures.
arXiv Detail & Related papers (2021-12-16T17:03:14Z) - BreakingBED -- Breaking Binary and Efficient Deep Neural Networks by
Adversarial Attacks [65.2021953284622]
We study robustness of CNNs against white-box and black-box adversarial attacks.
Results are shown for distilled CNNs, agent-based state-of-the-art pruned models, and binarized neural networks.
arXiv Detail & Related papers (2021-03-14T20:43:19Z) - RIFLE: Backpropagation in Depth for Deep Transfer Learning through
Re-Initializing the Fully-connected LayEr [60.07531696857743]
Fine-tuning the deep convolution neural network(CNN) using a pre-trained model helps transfer knowledge learned from larger datasets to the target task.
We propose RIFLE - a strategy that deepens backpropagation in transfer learning settings.
RIFLE brings meaningful updates to the weights of deep CNN layers and improves low-level feature learning.
arXiv Detail & Related papers (2020-07-07T11:27:43Z) - Neural Additive Models: Interpretable Machine Learning with Neural Nets [77.66871378302774]
Deep neural networks (DNNs) are powerful black-box predictors that have achieved impressive performance on a wide variety of tasks.
We propose Neural Additive Models (NAMs) which combine some of the expressivity of DNNs with the inherent intelligibility of generalized additive models.
NAMs learn a linear combination of neural networks that each attend to a single input feature.
arXiv Detail & Related papers (2020-04-29T01:28:32Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.