Classification of Quasars, Galaxies, and Stars in the Mapping of the
Universe Multi-modal Deep Learning
- URL: http://arxiv.org/abs/2205.10745v1
- Date: Sun, 22 May 2022 05:17:31 GMT
- Title: Classification of Quasars, Galaxies, and Stars in the Mapping of the
Universe Multi-modal Deep Learning
- Authors: Sabeesh Ethiraj, Bharath Kumar Bolla
- Abstract summary: Fourth version the Sloan Digital Sky Survey (SDSS-4), Data Release 16 dataset was used to classify the SDSS dataset into galaxies, stars, and quasars using machine learning and deep learning architectures.
We build a novel multi-modal architecture and achieve state-of-the-art results.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In this paper, the fourth version the Sloan Digital Sky Survey (SDSS-4), Data
Release 16 dataset was used to classify the SDSS dataset into galaxies, stars,
and quasars using machine learning and deep learning architectures. We
efficiently utilize both image and metadata in tabular format to build a novel
multi-modal architecture and achieve state-of-the-art results. In addition, our
experiments on transfer learning using Imagenet weights on five different
architectures (Resnet-50, DenseNet-121 VGG-16, Xception, and EfficientNet)
reveal that freezing all layers and adding a final trainable layer may not be
an optimal solution for transfer learning. It is hypothesized that higher the
number of trainable layers, higher will be the training time and accuracy of
predictions. It is also hypothesized that any subsequent increase in the number
of training layers towards the base layers will not increase in accuracy as the
pre trained lower layers only help in low level feature extraction which would
be quite similar in all the datasets. Hence the ideal level of trainable layers
needs to be identified for each model in respect to the number of parameters.
For the tabular data, we compared classical machine learning algorithms
(Logistic Regression, Random Forest, Decision Trees, Adaboost, LightGBM etc.,)
with artificial neural networks. Our works shed new light on transfer learning
and multi-modal deep learning architectures. The multi-modal architecture not
only resulted in higher metrics (accuracy, precision, recall, F1 score) than
models using only image data or tabular data. Furthermore, multi-modal
architecture achieved the best metrics in lesser training epochs and improved
the metrics on all classes.
Related papers
- Compute Better Spent: Replacing Dense Layers with Structured Matrices [77.61728033234233]
We identify more efficient alternatives to dense matrices, as exemplified by the success of convolutional networks in the image domain.
We show that different structures often require drastically different initialization scales and learning rates, which are crucial to performance.
We propose a novel matrix family containing Monarch matrices, the Block-Train, which we show performs better than dense for the same compute on multiple tasks.
arXiv Detail & Related papers (2024-06-10T13:25:43Z) - Dataset Quantization [72.61936019738076]
We present dataset quantization (DQ), a new framework to compress large-scale datasets into small subsets.
DQ is the first method that can successfully distill large-scale datasets such as ImageNet-1k with a state-of-the-art compression ratio.
arXiv Detail & Related papers (2023-08-21T07:24:29Z) - NASiam: Efficient Representation Learning using Neural Architecture
Search for Siamese Networks [76.8112416450677]
Siamese networks are one of the most trending methods to achieve self-supervised visual representation learning (SSL)
NASiam is a novel approach that uses for the first time differentiable NAS to improve the multilayer perceptron projector and predictor (encoder/predictor pair)
NASiam reaches competitive performance in both small-scale (i.e., CIFAR-10/CIFAR-100) and large-scale (i.e., ImageNet) image classification datasets while costing only a few GPU hours.
arXiv Detail & Related papers (2023-01-31T19:48:37Z) - Terrain Classification using Transfer Learning on Hyperspectral Images:
A Comparative study [0.13999481573773068]
convolutional neural network (CNN) and the Multi-Layer Perceptron (MLP) have been proven to be an effective method of image classification.
However, they suffer from the issues of long training time and requirement of large amounts of the labeled data.
We propose using the method of transfer learning to decrease the training time and reduce the dependence on large labeled dataset.
arXiv Detail & Related papers (2022-06-19T14:36:33Z) - Learning Rate Curriculum [75.98230528486401]
We propose a novel curriculum learning approach termed Learning Rate Curriculum (LeRaC)
LeRaC uses a different learning rate for each layer of a neural network to create a data-agnostic curriculum during the initial training epochs.
We compare our approach with Curriculum by Smoothing (CBS), a state-of-the-art data-agnostic curriculum learning approach.
arXiv Detail & Related papers (2022-05-18T18:57:36Z) - Classification of Astronomical Bodies by Efficient Layer Fine-Tuning of
Deep Neural Networks [0.0]
The SDSS-IV dataset contains information about various astronomical bodies such as Galaxies, Stars, and Quasars captured by observatories.
Inspired by our work on deep multimodal learning, we further extended our research in the fine tuning of these architectures to study the effect in the classification scenario.
arXiv Detail & Related papers (2022-05-14T20:08:19Z) - Efficient deep learning models for land cover image classification [0.29748898344267777]
This work experiments with the BigEarthNet dataset for land use land cover (LULC) image classification.
We benchmark different state-of-the-art models, including Convolution Neural Networks, Multi-Layer Perceptrons, Visual Transformers, EfficientNets and Wide Residual Networks (WRN)
Our proposed lightweight model has an order of magnitude less trainable parameters, achieves 4.5% higher averaged f-score classification accuracy for all 19 LULC classes and is trained two times faster with respect to a ResNet50 state-of-the-art model that we use as a baseline.
arXiv Detail & Related papers (2021-11-18T00:03:14Z) - LV-BERT: Exploiting Layer Variety for BERT [85.27287501885807]
We introduce convolution into the layer type set, which is experimentally found beneficial to pre-trained models.
We then adopt an evolutionary algorithm guided by pre-training accuracy to find the optimal architecture.
LV-BERT model obtained by our method outperforms BERT and its variants on various downstream tasks.
arXiv Detail & Related papers (2021-06-22T13:20:14Z) - Train your classifier first: Cascade Neural Networks Training from upper
layers to lower layers [54.47911829539919]
We develop a novel top-down training method which can be viewed as an algorithm for searching for high-quality classifiers.
We tested this method on automatic speech recognition (ASR) tasks and language modelling tasks.
The proposed method consistently improves recurrent neural network ASR models on Wall Street Journal, self-attention ASR models on Switchboard, and AWD-LSTM language models on WikiText-2.
arXiv Detail & Related papers (2021-02-09T08:19:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.