Classification of Astronomical Bodies by Efficient Layer Fine-Tuning of
Deep Neural Networks
- URL: http://arxiv.org/abs/2205.07124v1
- Date: Sat, 14 May 2022 20:08:19 GMT
- Title: Classification of Astronomical Bodies by Efficient Layer Fine-Tuning of
Deep Neural Networks
- Authors: Sabeesh Ethiraj, Bharath Kumar Bolla
- Abstract summary: The SDSS-IV dataset contains information about various astronomical bodies such as Galaxies, Stars, and Quasars captured by observatories.
Inspired by our work on deep multimodal learning, we further extended our research in the fine tuning of these architectures to study the effect in the classification scenario.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The SDSS-IV dataset contains information about various astronomical bodies
such as Galaxies, Stars, and Quasars captured by observatories. Inspired by our
work on deep multimodal learning, which utilized transfer learning to classify
the SDSS-IV dataset, we further extended our research in the fine tuning of
these architectures to study the effect in the classification scenario.
Architectures such as Resnet-50, DenseNet-121 VGG-16, Xception, EfficientNetB2,
MobileNetV2 and NasnetMobile have been built using layer wise fine tuning at
different levels. Our findings suggest that freezing all layers with Imagenet
weights and adding a final trainable layer may not be the optimal solution.
Further, baseline models and models that have higher number of trainable layers
performed similarly in certain architectures. Model need to be fine tuned at
different levels and a specific training ratio is required for a model to be
termed ideal. Different architectures had different responses to the change in
the number of trainable layers w.r.t accuracies. While models such as
DenseNet-121, Xception, EfficientNetB2 achieved peak accuracies that were
relatively consistent with near perfect training curves, models such as
Resnet-50,VGG-16, MobileNetV2 and NasnetMobile had lower, delayed peak
accuracies with poorly fitting training curves. It was also found that though
mobile neural networks have lesser parameters and model size, they may not
always be ideal for deployment on a low computational device as they had
consistently lower validation accuracies. Customized evaluation metrics such as
Tuning Parameter Ratio and Tuning Layer Ratio are used for model evaluation.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - Dataset Quantization [72.61936019738076]
We present dataset quantization (DQ), a new framework to compress large-scale datasets into small subsets.
DQ is the first method that can successfully distill large-scale datasets such as ImageNet-1k with a state-of-the-art compression ratio.
arXiv Detail & Related papers (2023-08-21T07:24:29Z) - Layer-wise Linear Mode Connectivity [52.6945036534469]
Averaging neural network parameters is an intuitive method for the knowledge of two independent models.
It is most prominently used in federated learning.
We analyse the performance of the models that result from averaging single, or groups.
arXiv Detail & Related papers (2023-07-13T09:39:10Z) - A Light-weight Deep Learning Model for Remote Sensing Image
Classification [70.66164876551674]
We present a high-performance and light-weight deep learning model for Remote Sensing Image Classification (RSIC)
By conducting extensive experiments on the NWPU-RESISC45 benchmark, our proposed teacher-student models outperforms the state-of-the-art systems.
arXiv Detail & Related papers (2023-02-25T09:02:01Z) - A Gradient Boosting Approach for Training Convolutional and Deep Neural
Networks [0.0]
We introduce two procedures for training Convolutional Neural Networks (CNNs) and Deep Neural Network based on Gradient Boosting (GB)
The presented models show superior performance in terms of classification accuracy with respect to standard CNN and Deep-NN with the same architectures.
arXiv Detail & Related papers (2023-02-22T12:17:32Z) - Classification of Quasars, Galaxies, and Stars in the Mapping of the
Universe Multi-modal Deep Learning [0.0]
Fourth version the Sloan Digital Sky Survey (SDSS-4), Data Release 16 dataset was used to classify the SDSS dataset into galaxies, stars, and quasars using machine learning and deep learning architectures.
We build a novel multi-modal architecture and achieve state-of-the-art results.
arXiv Detail & Related papers (2022-05-22T05:17:31Z) - Supervised Training of Siamese Spiking Neural Networks with Earth's
Mover Distance [4.047840018793636]
This study adapts the highly-versatile siamese neural network model to the event data domain.
We introduce a supervised training framework for optimizing Earth's Mover Distance between spike trains with spiking neural networks (SNN)
arXiv Detail & Related papers (2022-02-20T00:27:57Z) - Efficient deep learning models for land cover image classification [0.29748898344267777]
This work experiments with the BigEarthNet dataset for land use land cover (LULC) image classification.
We benchmark different state-of-the-art models, including Convolution Neural Networks, Multi-Layer Perceptrons, Visual Transformers, EfficientNets and Wide Residual Networks (WRN)
Our proposed lightweight model has an order of magnitude less trainable parameters, achieves 4.5% higher averaged f-score classification accuracy for all 19 LULC classes and is trained two times faster with respect to a ResNet50 state-of-the-art model that we use as a baseline.
arXiv Detail & Related papers (2021-11-18T00:03:14Z) - ANNETTE: Accurate Neural Network Execution Time Estimation with Stacked
Models [56.21470608621633]
We propose a time estimation framework to decouple the architectural search from the target hardware.
The proposed methodology extracts a set of models from micro- kernel and multi-layer benchmarks and generates a stacked model for mapping and network execution time estimation.
We compare estimation accuracy and fidelity of the generated mixed models, statistical models with the roofline model, and a refined roofline model for evaluation.
arXiv Detail & Related papers (2021-05-07T11:39:05Z) - The Heterogeneity Hypothesis: Finding Layer-Wise Differentiated Network
Architectures [179.66117325866585]
We investigate a design space that is usually overlooked, i.e. adjusting the channel configurations of predefined networks.
We find that this adjustment can be achieved by shrinking widened baseline networks and leads to superior performance.
Experiments are conducted on various networks and datasets for image classification, visual tracking and image restoration.
arXiv Detail & Related papers (2020-06-29T17:59:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.