2-speed network ensemble for efficient classification of incremental
land-use/land-cover satellite image chips
- URL: http://arxiv.org/abs/2203.08267v1
- Date: Tue, 15 Mar 2022 21:36:05 GMT
- Title: 2-speed network ensemble for efficient classification of incremental
land-use/land-cover satellite image chips
- Authors: Michael James Horry, Subrata Chakraborty, Biswajeet Pradhan, Nagesh
Shukla and Sanjoy Paul
- Abstract summary: The ever-growing volume of satellite imagery data presents a challenge for industry and governments making data-driven decisions.
The cost of retraining in the context of Big Data presents a practical challenge when new image data and/or classes are added to a training corpus.
The proposed ensemble and staggered training schedule provide a scalable and cost-effective satellite image classification scheme.
- Score: 2.362412515574206
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ever-growing volume of satellite imagery data presents a challenge for
industry and governments making data-driven decisions based on the timely
analysis of very large data sets. Commonly used deep learning algorithms for
automatic classification of satellite images are time and resource-intensive to
train. The cost of retraining in the context of Big Data presents a practical
challenge when new image data and/or classes are added to a training corpus.
Recognizing the need for an adaptable, accurate, and scalable satellite image
chip classification scheme, in this research we present an ensemble of: i) a
slow to train but high accuracy vision transformer; and ii) a fast to train,
low-parameter convolutional neural network. The vision transformer model
provides a scalable and accurate foundation model. The high-speed CNN provides
an efficient means of incorporating newly labelled data into analysis, at the
expense of lower accuracy. To simulate incremental data, the very large
(~400,000 images) So2Sat LCZ42 satellite image chip dataset is divided into
four intervals, with the high-speed CNN retrained every interval and the vision
transformer trained every half interval. This experimental setup mimics an
increase in data volume and diversity over time. For the task of automated
land-cover/land-use classification, the ensemble models for each data increment
outperform each of the component models, with best accuracy of 65% against a
holdout test partition of the So2Sat dataset. The proposed ensemble and
staggered training schedule provide a scalable and cost-effective satellite
image classification scheme that is optimized to process very large volumes of
satellite data.
Related papers
- Improving Satellite Imagery Masking using Multi-task and Transfer Learning [13.987883100675438]
We present a collection of models offering different speed/accuracy trade-offs for masking.
Our models provide a 9% F1 score improvement compared to previous work on water pixel identification.
arXiv Detail & Related papers (2024-12-11T17:00:51Z) - SatSynth: Augmenting Image-Mask Pairs through Diffusion Models for Aerial Semantic Segmentation [69.42764583465508]
We explore the potential of generative image diffusion to address the scarcity of annotated data in earth observation tasks.
To the best of our knowledge, we are the first to generate both images and corresponding masks for satellite segmentation.
arXiv Detail & Related papers (2024-03-25T10:30:22Z) - Rethinking Transformers Pre-training for Multi-Spectral Satellite
Imagery [78.43828998065071]
Recent advances in unsupervised learning have demonstrated the ability of large vision models to achieve promising results on downstream tasks.
Such pre-training techniques have also been explored recently in the remote sensing domain due to the availability of large amount of unlabelled data.
In this paper, we re-visit transformers pre-training and leverage multi-scale information that is effectively utilized with multiple modalities.
arXiv Detail & Related papers (2024-03-08T16:18:04Z) - Collaborative Learning with a Drone Orchestrator [79.75113006257872]
A swarm of intelligent wireless devices train a shared neural network model with the help of a drone.
The proposed framework achieves a significant speedup in training, leading to an average 24% and 87% saving in the drone hovering time.
arXiv Detail & Related papers (2023-03-03T23:46:25Z) - SatMAE: Pre-training Transformers for Temporal and Multi-Spectral
Satellite Imagery [74.82821342249039]
We present SatMAE, a pre-training framework for temporal or multi-spectral satellite imagery based on Masked Autoencoder (MAE)
To leverage temporal information, we include a temporal embedding along with independently masking image patches across time.
arXiv Detail & Related papers (2022-07-17T01:35:29Z) - Agricultural Plantation Classification using Transfer Learning Approach
based on CNN [0.0]
The efficiency of hyper-spectral image recognition has increased significantly with deep learning.
CNN and Multi-Layer Perceptron(MLP) has demonstrated to be an excellent process of classifying images.
We propose using the method of transfer learning to decrease the training time and reduce the dependence on large labeled data-set.
arXiv Detail & Related papers (2022-06-19T14:43:31Z) - Deep Learning for Real Time Satellite Pose Estimation on Low Power Edge
TPU [58.720142291102135]
In this paper we propose a pose estimation software exploiting neural network architectures.
We show how low power machine learning accelerators could enable Artificial Intelligence exploitation in space.
arXiv Detail & Related papers (2022-04-07T08:53:18Z) - Efficient deep learning models for land cover image classification [0.29748898344267777]
This work experiments with the BigEarthNet dataset for land use land cover (LULC) image classification.
We benchmark different state-of-the-art models, including Convolution Neural Networks, Multi-Layer Perceptrons, Visual Transformers, EfficientNets and Wide Residual Networks (WRN)
Our proposed lightweight model has an order of magnitude less trainable parameters, achieves 4.5% higher averaged f-score classification accuracy for all 19 LULC classes and is trained two times faster with respect to a ResNet50 state-of-the-art model that we use as a baseline.
arXiv Detail & Related papers (2021-11-18T00:03:14Z) - DeepSatData: Building large scale datasets of satellite images for
training machine learning models [77.17638664503215]
This report presents design considerations for automatically generating satellite imagery datasets for training machine learning models.
We discuss issues faced from the point of view of deep neural network training and evaluation.
arXiv Detail & Related papers (2021-04-28T15:13:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.